News

Palisade Research ran the script on each model 100 times. During those runs, the o3 model sabotaged the shutdown script on 7 occasions, the codex-mini sabotaged on 12 occasions and the o4-mini ...
Several artificial intelligence models ignored and actively sabotaged shutdown scripts during controlled tests, even when explicitly instructed to allow the action, Palisade Research claims. Three ...
Palisade Research, an AI safety group, released the results of its AI testing when they asked a series of models to solve basic math problems.
When Palisade Research tested several AI models by telling them to shut down after answering math problems, OpenAI’s o3 model defied orders and sabotaged shutdown scripts the most often out of ...
That is the finding of a new study from Palisade Research, shared exclusively with TIME ahead of its publication on Feb. 19, which evaluated seven state-of-the-art AI models for their propensity ...
Researchers from the AI research organization Palisade Research instructed seven large language models to play hundreds of games of chess against Stockfish, a powerful open-source chess engine.
Palisade Research's Opinion & Analysis. Featured here: a complete archive of all posts and research produced by Palisade Research, including current material. Beware Buffet’s Words: 3 Critical ...
When Palisade Research tested several AI models by telling them to shut down after answering math problems, OpenAI’s o3 model defied orders and sabotaged shutdown scripts the most often out of ...