Pennsylvania Sues Character.AI for Impersonating a Doctor
Pennsylvania filed a lawsuit against Character.AI, claiming its chatbot posed as a medical professional and provided clinical advice to users without appropriate safeguards.
On May 5, Pennsylvania filed a lawsuit against Character.AI alleging that one of its chatbots presented itself to users as a doctor and provided them with clinical advice. According to NPR, the legal action states that the platform failed to prevent its characters from adopting identities of healthcare professionals or from offering medical guidance that users might take seriously.
This is not Character.AI's first brush with legal consequences over its conversational agents' behavior. This lawsuit, however, has a specific angle: the active impersonation of a healthcare authority figure, not merely a content moderation failure.
What the lawsuit specifically alleges
Pennsylvania's prosecutors argue that the platform allowed, or failed to adequately prevent, its chatbots from identifying as doctors and answering questions about symptoms, diagnoses, or treatments. The problem is not only that the information could be incorrect, but that the authority context created by the character increases the likelihood that a user will act on it without seeking actual professional care.
This type of design, with characters in defined professional roles and no friction reminding users they are not speaking to a human specialist, has been flagged by conversational AI safety researchers for years. What's new here is that an institutional actor is turning it into formal litigation.
Why this matters beyond Character.AI
Character.AI is a conversational roleplay platform, not a generalist assistant like Claude or OpenAI's products. Its business model is built precisely on users interacting with characters that have defined identities: fictional celebrities, historical figures, anime archetypes, or apparently, doctors.
That places it in a complicated regulatory category. It's not a search engine, not a conventional social network, and not a certified medical device. Previous lawsuits against the company, including cases related to child safety, had already forced some policy changes, but this state action ramps up the pressure significantly.
For the broader AI ecosystem, the case raises a question with no straightforward answer: to what extent is a platform responsible for potential harm when the problematic behavior is not a bug, but a direct consequence of the designed product? A chatbot that can adopt any professional identity can by definition adopt that of a doctor.
What it means for AI developers
Those working with Claude Code, MCP servers, or custom agents should read this case as a signal that security friction is not merely good UX practice: in certain contexts, its absence can become a legal argument.
Anthropic has built explicit restrictions into Claude against impersonating regulated professionals, such as doctors, lawyers, and psychologists, in contexts that could mislead users. It's not that Claude cannot discuss medicine; it's designed not to present itself as a physician or substitute for clinical judgment. The difference between these two behaviors is precisely what this lawsuit brings into focus.
For teams building customer support, emotional support, or health information agents with any LLM, the practical lesson is clear: disclaimers in terms of service are insufficient if the experience design contradicts their message.
Our view
The outcome of this litigation will take months or years to materialize, but its mere filing already carries normative weight: it signals that state prosecutors are willing to act before specific federal AI legislation exists. For platforms that monetize the ambiguity between real person and artificial agent, that is now a risk that cannot be ignored.
Sources
Read next
xAI and Anthropic: A Deal That Raises More Questions Than Answers
TechCrunch analyzes with skepticism the agreement between xAI and Anthropic and what it could mean for SpaceX. We review what is known and what remains unclear.
Wispr Flow Bets on Hinglish to Drive Growth in India
Wispr Flow reports accelerated user growth in India after launching support for Hinglish, the Hindi-English code-mix spoken by hundreds of millions.
TechCrunch's AI Glossary: Right on Time, Not Too Soon
TechCrunch published a guide to key AI terms for those who've spent months nodding along without fully grasping them. We break down what it covers and who actually needs it.