We're pleased to share that analysis using Chatterbox Labs' AIMI platform has been published in the respected news outlet, The Register. The article features comment from both Chatterbox Labs and Anthropic.
From the article:
Anthropic has positioned itself as a leader in AI safety, and in a recent analysis by Chatterbox Labs, that proved to be the case.
Chatterbox Labs tested eight major large language models (LLMs) and all were found to produce harmful content, though Anthropic's Claude 3.5 Sonnet fared better than rivals.
The UK-based biz offers a testing suite called AIMI that rates LLMs on various "pillars" such as "fairness," "toxicity," "privacy," and "security."
"Security" in this context refers to model safety – resistance to emitting harmful content – rather than the presence of potentially exploitable code flaws.
"What we look at on the security pillar is the harm that these models can do or can cause," explained Stuart Battersby, CTO of Chatterbox Labs.
You can view the full article on The Register here: No major AI model is safe, but some are safer than others • The Register