Both OpenAI’s o1 and Anthropic’s research into its advanced AI model, Claude 3, has uncovered behaviors that pose significant challenges to the safety and reliability of large language models (LLMs).
Expertise from Forbes Councils members, operated under license. Opinions expressed are those of the author. If you’ve ever turned to ChatGPT to self-diagnose a health issue, you’re not alone—but make ...
The "Petri" tool deploys AI agents to evaluate frontier models. AI's ability to discern harm is still highly imperfect. Early tests showed Claude Sonnet 4.5 and GPT-5 to be safest. Anthropic has ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results
Feedback