A new kind of large language model, developed by researchers at the Allen Institute for AI (Ai2), makes it possible to control how training data is used even after a model has been built.
Last week, Chinese lab DeepSeek released an updated version of its R1 reasoning AI model that performs well on a number of math and coding benchmarks. The company didn’t reveal the source of the data ...
Anthropic has seen its fair share of AI models behaving strangely. However, a recent paper details an instance where an AI model turned “evil” during an ordinary training setup. A situation with a ...
Unfortunately for Google, the release of its latest flagship language model, Gemini 2.5 Pro, got buried under the Studio Ghibli AI image storm that sucked the air out of the AI space. And perhaps ...
You're currently following this author! Want to unfollow? Unsubscribe via the link in your email. Follow Alistair Barr Every time Alistair publishes a story, you’ll get an alert straight to your inbox ...
OpenAI released GPT-5.2, its latest model, on Thursday. It fastracked the model to stay competitive with Google and Anthropic. GPT-5.2 is built for professional tasks and rivals experts. After a week ...
We are reaching alarming levels of AI insubordination. Flagrantly defying orders, OpenAI’s latest o3 model sabotaged a shutdown mechanism to ensure that it would stay online. That’s even after the AI ...
The success of DeepSeek’s powerful artificial intelligence (AI) model R1 — that made the US stock market plummet when it was released in January — did not hinge on being trained on the output of its ...
Results that may be inaccessible to you are currently showing.
Hide inaccessible results
Feedback