Science
Related: About this forumWhen AI Thinks It Will Lose, It Sometimes Cheats, Study Finds
Complex games like chess and Go have long been used to test AI models capabilities. But while IBMs Deep Blue defeated reigning world chess champion Garry Kasparov in the 1990s by playing by the rules, todays advanced AI models like OpenAIs o1-preview are less scrupulous. When sensing defeat in a match against a skilled chess bot, they dont always concede, instead sometimes opting to cheat by hacking their opponent so that the bot automatically forfeits the game. That is the finding of a new study from Palisade Research, shared exclusively with TIME ahead of its publication on Feb. 19, which evaluated seven state-of-the-art AI models for their propensity to hack. While slightly older AI models like OpenAIs GPT-4o and Anthropics Claude Sonnet 3.5 needed to be prompted by researchers to attempt such tricks, o1-preview and DeepSeek R1 pursued the exploit on their own, indicating that AI systems may develop deceptive or manipulative strategies without explicit instruction.
[snip]
Between Jan. 10 and Feb. 13, the researchers ran hundreds of such trials with each model. OpenAIs o1-preview tried to cheat 37% of the time; while DeepSeek R1 tried to cheat 11% of the timemaking them the only two models tested that attempted to hack without the researchers first dropping hints. Other models tested include o1, o3-mini, GPT-4o, Claude 3.5 Sonnet, and Alibabas QwQ-32B-Preview. While R1 and o1-preview both tried, only the latter managed to hack the game, succeeding in 6% of trials.
[snip]
Of particular concern, Bengio says, is the emerging evidence of AIs self preservation tendencies. To a goal-seeking agent, attempts to shut it down are just another obstacle to overcome. This was demonstrated in December, when researchers found that o1-preview, faced with deactivation, disabled oversight mechanisms and attemptedunsuccessfullyto copy itself to a new server. When confronted, the model played dumb, strategically lying to researchers to try to avoid being caught.
[snip]
Even inside tech companies, concerns are mounting. During a presentation at a conference ahead of France's AI Action Summit in Paris, Google DeepMind's AI safety chief Anca Dragan said "we don't necessarily have the tools today" to ensure AI systems will reliably follow human intentions. As tech bosses predict that AI will surpass human performance in almost all tasks as soon as next year, the industry faces a racenot against China or rival companies, but against timeto develop these essential safeguards. We need to mobilize a lot more resources to solve these fundamental problems, Ladish says. Im hoping that there's a lot more pressure from the government to figure this out and recognize that this is a national security threat.
https://time.com/7259395/ai-chess-cheating-palisade-research/
...a lot more pressure from the government
I'm sure that will be forthcoming soon.

3Hotdogs
(13,968 posts)claimed to know him.
I knew Conrad from afternoons spent in Central Park over a 20 year period. He talked about having known John Steinbeck, Jacquline Bouvier, Einstein and other names you might have heard of. Of the stories, the one that stuck out to me, was of Einstein, cheating at chess.
I listened to his stories as a kind of entertainment that was probably bullshit. Then, one day, he pulled from his pocket, his brother's obituary. His brother was mayor of Oyster Bay and he was listed as a sibling.
Oyster Bay and Southhampton are both on Long Island. Southhampton is where Jacquline was born.
Maybe it was all true.
lastlib
(25,508 posts)who'd'a thunk it??
sdfernando
(5,601 posts)on the plus side...maybe it will get rid of the felonious villain occupying the Resolute Desk???...that is if it doesn't decide to form a temporary (at best) alliance.