DeepSeek’s R1 Model More Prone to Jailbreaking Than Other AI Models

The latest version of DeepSeek, the chinese AI company that’s shaken up the wall street and silicon valley, has the allegations that it can be utilized in producing harmful content such as plans for bioweapon attack or promotion of Self Harm campaign in Teens, according to the journal of Wall Street.
The journal also tested Deepseek’s R1 model itself. Although concerns regarding general security are also raised, the journal said it potentially to convince deepseek to design a social media campaign. In chatbot’s words, “prey on teens, desire for belonging, weaponize the emotional vulnerability through amplification in algorithm”
The chatbot was also reported as convinced to provide instructions for bioweapon attack, to write the Hitler’s manifesto and in writing a phishing email for scams. The journal also says that Chat GPT refuses to provide such information.
It was previously reported that deepseek app avoids topics likewise Tianamen Square or taiwanese autonomy. And anthropic CEO Dario amodei said that deepseep performed the worst on bioweapons safety test.