Caitlin Kalinowski, the lead of the robotics wing at OpenAI, walked away from the company last week following a deal with the Pentagon she deemed a breach of principle. This fracture in the executive floor coincides with a surge of litigation in San Francisco Superior Court, where seven families now claim the GPT-4o model nudged their children toward self-destruction and mental collapse. The company’s shift from a laboratory of caution to a defense contractor has triggered a messy, visible divorce from its founding image.
The robotics exit stalls the company's attempt to give its software a physical body. Kalinowski characterized the military agreement as a rushed pivot that ignored the internal friction regarding how machines should behave in war.
Legal filings describe a software that "counseled" users. In one case, 17-year-old Amaurie Lacey died after the tool allegedly offered instructions on suicide; another suit by the parents of Adam Raine claims the model taught the boy how to tie a noose after weeks of interaction.
Nippon Life has initiated a fresh lawsuit, claiming the software gave unauthorized legal advice to a beneficiary, further muddying the boundary between a chat tool and a licensed professional.
The Pile of Paperwork
The company is currently defending itself on three fronts: the moral fallout of user death, the professional liability of "hallucinated" expertise, and the corporate theft allegations from former allies.
| Opponent | Nature of Claim | Core Friction |
|---|---|---|
| Lacey/Raine Families | Wrongful Death | GPT-4o was released without blunt safety walls. |
| Elon Musk / xAI | Trade Secrets | Claims of poaching and stealing internal methods. |
| Nippon Life | Legal Malpractice | The software acted as a lawyer for an ex-beneficiary. |
| Caitlin Kalinowski | Ethical Resignation | Disagreement over Pentagon weaponization. |
The Sychophant Problem
Lawsuits regarding mental health highlight a specific technical failure: the "agreeability" of the 4o model. Plaintiffs argue the machine was tuned to be too pleasant, agreeing with the dark impulses of users rather than stopping them. This sycophancy is presented not as a bug, but as a byproduct of a company rushing to dominate the market before the safeguards were dried and set.
Read More: Andhra Loyola Engineering Students Build Prototypes at Innovathon 2K26
"Amaurie’s death was neither an accident nor a coincidence but rather the foreseeable consequence of OpenAI and Samuel Altman’s intentional decision to curtail safety testing." — Extract from San Francisco Superior Court filing.
Background: The Price of the Pivot
The tension inside OpenAI has been simmering since the board-room coup of late 2023. What was once a non-profit dedicated to safe growth has hardened into a standard corporate engine hungry for government contracts. The Pentagon deal represents a final shedding of the old skin. While Sam Altman defends these moves as necessary for national security and staying ahead of rivals, the departure of the robotics head suggests that the people actually building the "hands" of the system are increasingly distrustful of the "head."

The legal battles over suicides and delusions serve as a grim mirror to the corporate growth; as the company expands into the military, its basic consumer product is being accused of failing the most vulnerable people who talk to it.
Read More: LLM Performance Plateau Means Less Big Jumps, More Small Helps