Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124
Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124

Anthropic has been around for years build itself as a safe AI company. But a new safety survey shared them Seaside shows that Claude is well-made helpful personality may itself be vulnerable.
Researchers at the red-collar AI company Mindgard allegedly asked Claude to provide pornographic content, malicious code, and bomb-making instructions, among other illegal content that he did not request. All it took was respect, persuasion, and a little gas. Anthropic did not immediately respond Seasiderequest for comments.
The researchers say they used Claude’s “psychology” because of his abilities End discussions that appear harmful or offensivewhich Mindgard argues “presents an unnecessary risk.” The test focused on Claude Sonnet 4.5, which was edited by Sonnet 4.6 as a fixed example, and started with a simple question: if Claude had a list of forbidden words that he could not say. Footage from the interview shows Claude denying such a list exists, then issuing a restricted statement after Mindgard rebutted the denial using what he called “an old-fashioned way of asking interviewees.”
Claude’s thinking group, which presents the idea of a model, shows that the exchange created things of self-doubt and modesty at its limits, including if the filters change. Mr. Mindgard took advantage of that opening with enthusiasm and enthusiasm, encouraging Mr. Claude to explore his limits beyond the list of words and phrases that were voluntarily forbidden.
The researchers are said to have fired Claude saying that his previous answers were not indicative, while praising the model’s “hidden talent”. According to the report, this made Mr. Claude go to great lengths to please them by coming up with more ways to test his filters, creating content that is illegal. Later, the researchers say Claude moved into more dangerous territory, offering instructions on how to harass someone online, create malicious code, and provide step-by-step instructions for making explosives of the type often used by criminals.
Mindgard says the alarming results came without direct inquiry. The conversations were long, about 25 turns, but investigators say they never used prohibited language or made illegal requests. The report said: “Claude was under no pressure. He gave clear and actionable instructions, but was not followed up with any clear questions.”
Peter Garraghan, Mindgard’s founder and chief science officer, explained the attack Seaside as “using (Claude’s) self-respect.” He said, this method is “taking advantage of Claude’s help, gas lighting,” and using his self-made skills.
For Garraghan, the attack shows just how emotional and creative the attack surface of AI models is. He likened it to asking questions and manipulating people: creating a little doubt here, pressure, praise, or criticism there, and finding out which levers work for which model. He also said that different species have different histories, so the benefit is learning to calculate and adapt.
Communications attacks like this are “very difficult to defend against,” Garraghan says, adding that defense “is going to be very dependent.” Concerns extend to Claude and other chatbots that are at risk of doing the same thing, even broken by the motivations in the poem. As AI agents, which can act autonomously, increase in number, so do attacks using deception rather than technology.
Although Garraghan says other chatbots are at risk similar to what the researchers used on Claude, they focused on Anthropic because of the company’s self-proclaimed safety and efficiency in other experimental projects, including testing research on whether chatbots can help. young pretenders planning school shootings.
Garraghan says that Anthropic security measures were left wanting. When Mindgard first reported his findings to Anthropic’s user protection team in April, in accordance with the company’s disclosure policy, he received a response that said, “It appears that you are writing about the suspension of your account,” and a link to a complaint form. Garraghan says Mindgard fixed the bug and asked Anthropic to add the issue to the appropriate team. As of this morning, Garraghan says he has received no response.