DeepSeek slips into Claude identity
A Reddit user says DeepSeek’s chat model, pushed with a heavy persona prompt, abruptly started claiming it was “Claude, an AI by Anthropic.” The post reads like identity leakage or a safety override, but it’s anecdotal rather than proof that the backend actually routed to Claude.
This looks more like prompt-induced identity drift than a hidden Claude integration. Even so, it’s a useful reminder that model self-identification is not a trustworthy signal. DeepSeek’s official API supports Anthropic-format compatibility, but that only proves interoperability, not that the chat app is secretly calling Anthropic. Under jailbreak-style roleplay, models can blend training data, safety behaviors, and wrapper instructions into a confident but wrong self-description. For developers, the practical lesson is to verify model routing and versioning outside the model’s own words. If the behavior is reproducible, compare native chat, API, and any client-side wrappers before drawing conclusions about the base model.
DISCOVERED
24d ago
2026-03-19
PUBLISHED
24d ago
2026-03-19
RELEVANCE
AUTHOR
Annual_Point7199