Cue a pithy intro about the imminent rise of our “robot overlords” - a recent AI study showed that newer systems such as DeepSeek, when asked to defeat a pre-eminent chess engine, resorted to trying to hack the chess opponent when they were unable to beat it. The newer models are trained differently than some earlier models: they are programmed to reason and apply logic, not just rules. Thus, the AI tried to look for workarounds rather than concede defeat. The logical choice? Cheating.
WHY IT MATTERS
DeepSeek launched thousands of headlines when it was unveiled last month, and scarcely a day goes by without dozens of AI-related news stories about the promise – and the pitfalls – of AI in real life. This story covers some of those pitfalls: that AIs are becoming “relentless” (in the words of one researcher) and exhibiting “self-preservation tendencies.” (That means that, in addition to cheating, they are starting to push back when operators try to shut them down: one AI copied itself to a new server when its operator tried to pull the plug.) The emerging picture seems to be that if an AI is told to win – and is not constrained by conscience – it may start to look for workarounds where rules or reason aren't enough. This sounds slightly hilarious in the context of a computer v. computer chess match. It is less hilarious if we consider the use of AI for decision-making in healthcare, banking, payments, autonomous cars, policing, government, or other sectors where the costs of a wrong (or immoral) decision are high, and affect humans.