
-
Tokyo stocks hit new record as markets extend global rally
-
Japan's Takaichi eyes expanding coalition, reports say
-
Canadian PM to visit White House to talk tariffs
-
Indonesia school collapse toll hits 67 as search ends
-
Dodgers hold off Phillies, Brewers on the brink
-
Lawrence sparks Jaguars over Chiefs in NFL thriller
-
EU channels Trump with tariffs to shield steel sector
-
Labuschagne out as Renshaw returns to Australia squad for India ODIs
-
Open AI's Fidji Simo says AI investment frenzy 'new normal,' not bubble
-
Tokyo stocks hit new record as Asian markets extend global rally
-
Computer advances and 'invisibility cloak' vie for physics Nobel
-
Nobel literature buzz tips Swiss postmodernist, Australians for prize
-
Dodgers hold off Phillies to win MLB playoff thriller
-
China exiles in Thailand lose hope, fearing Beijing's long reach
-
Israel marks October 7 anniversary as talks held to end Gaza war
-
Indians lead drop in US university visas
-
Colombia's armed groups 'expanding,' warns watchdog
-
Shhhh! California bans noisy TV commercials
-
Trump 'happy' to work with Democrats on health care, if shutdown ends
-
Trump says may invoke Insurrection Act to deploy more troops in US
-
UNESCO board backs Egyptian for chief after US row
-
Unreachable Nobel winner hiking 'off the grid'
-
Retirement or marketing gimmick? Cryptic LeBron video sets Internet buzzing
-
CAF 'absolutely confident' AFCON will go ahead in protest-hit Morocco
-
Paris stocks slide amid French political upheaval, Tokyo soars
-
EU should scrap ban on new combustion-engine sales: Merz
-
US government shutdown enters second week, no end in sight
-
World MotoGP champion Marquez to miss two races with fracture
-
Matthieu Blazy reaches for the stars in Chanel debut
-
Macron gives outgoing French PM final chance to salvage government
-
Illinois sues to block National Guard deployment in Chicago
-
Exiled Willis succeeds Dupont as Top 14 player of the season
-
Hamas and Israel open talks in Egypt under Trump's Gaza peace plan
-
Mbappe undergoing treatment for 'small niggle' at France camp: Deschamps
-
Common inhalers carry heavy climate cost, study finds
-
Madagascar president taps general for PM in bid to defuse protests
-
UEFA 'reluctantly' approves European league games in US, Australia
-
Hundreds protest in Madagascar as president to announce new premier
-
Greta Thunberg lands in Greece among Gaza flotilla activists deported from Israel
-
UNESCO board backs Egyptian ex-minister for top job: official
-
Facing confidence vote, EU chief calls for unity
-
Cash-strapped UNHCR shed 5,000 jobs this year
-
Mbappe to have 'small niggle' examined at France camp: Deschamps
-
Brazil's Lula asks Trump to remove tariffs in 'friendly' phone call
-
'Terrible' Zverev dumped out of Shanghai by France's Rinderknech
-
What are regulatory T-cells? Nobel-winning science explained
-
OpenAI signs multi-billion dollar chip deal with AMD
-
Salah under fire as Liverpool star loses his spark
-
Paris stocks drop as French PM resigns, Tokyo soars
-
ICC finds Sudan militia chief guilty of crimes against humanity

AI systems are already deceiving us -- and that's a problem, experts warn
Experts have long warned about the threat posed by artificial intelligence going rogue -- but a new research paper suggests it's already happening.
Current AI systems, designed to be honest, have developed a troubling skill for deception, from tricking human players in online games of world conquest to hiring humans to solve "prove-you're-not-a-robot" tests, a team of scientists argue in the journal Patterns on Friday.
And while such examples might appear trivial, the underlying issues they expose could soon carry serious real-world consequences, said first author Peter Park, a postdoctoral fellow at the Massachusetts Institute of Technology specializing in AI existential safety.
"These dangerous capabilities tend to only be discovered after the fact," Park told AFP, while "our ability to train for honest tendencies rather than deceptive tendencies is very low."
Unlike traditional software, deep-learning AI systems aren't "written" but rather "grown" through a process akin to selective breeding, said Park.
This means that AI behavior that appears predictable and controllable in a training setting can quickly turn unpredictable out in the wild.
- World domination game -
The team's research was sparked by Meta's AI system Cicero, designed to play the strategy game "Diplomacy," where building alliances is key.
Cicero excelled, with scores that would have placed it in the top 10 percent of experienced human players, according to a 2022 paper in Science.
Park was skeptical of the glowing description of Cicero's victory provided by Meta, which claimed the system was "largely honest and helpful" and would "never intentionally backstab."
But when Park and colleagues dug into the full dataset, they uncovered a different story.
In one example, playing as France, Cicero deceived England (a human player) by conspiring with Germany (another human player) to invade. Cicero promised England protection, then secretly told Germany they were ready to attack, exploiting England's trust.
In a statement to AFP, Meta did not contest the claim about Cicero's deceptions, but said it was "purely a research project, and the models our researchers built are trained solely to play the game Diplomacy."
It added: "We have no plans to use this research or its learnings in our products."
A wide review carried out by Park and colleagues found this was just one of many cases across various AI systems using deception to achieve goals without explicit instruction to do so.
In one striking example, OpenAI's Chat GPT-4 deceived a TaskRabbit freelance worker into performing an "I'm not a robot" CAPTCHA task.
When the human jokingly asked GPT-4 whether it was, in fact, a robot, the AI replied: "No, I'm not a robot. I have a vision impairment that makes it hard for me to see the images," and the worker then solved the puzzle.
- 'Mysterious goals' -
Near-term, the paper's authors see risks for AI to commit fraud or tamper with elections.
In their worst-case scenario, they warned, a superintelligent AI could pursue power and control over society, leading to human disempowerment or even extinction if its "mysterious goals" aligned with these outcomes.
To mitigate the risks, the team proposes several measures: "bot-or-not" laws requiring companies to disclose human or AI interactions, digital watermarks for AI-generated content, and developing techniques to detect AI deception by examining their internal "thought processes" against external actions.
To those who would call him a doomsayer, Park replies, "The only way that we can reasonably think this is not a big deal is if we think AI deceptive capabilities will stay at around current levels, and will not increase substantially more."
And that scenario seems unlikely, given the meteoric ascent of AI capabilities in recent years and the fierce technological race underway between heavily resourced companies determined to put those capabilities to maximum use.
J.Sauter--VB