-
Latest evacuee from hantavirus-hit cruise lands in Europe
-
Rubio meets US pope in bid to ease tensions
-
Women linked to IS fighters return to Australia from Middle East
-
Shell profit jumps as Mideast war fuels oil prices
-
Oil sinks, Tokyo leads Asia stock surge on growing Mideast peace hopes
-
India vows to crush terror 'ecosystem', a year after Pakistan conflict
-
Circus tackles jihadist nightmares of Burkina Faso's children
-
Iran denies ship attack as Trump warns of renewed bombing, eyes deal
-
Badminton looks to future with 'evolution and innovation'
-
Troubled waters: Jakarta battles deadly, invasive suckerfish
-
Senegal's children mourn in silence when migrant parents disappear
-
EU weighs options as summer jet fuel threat looms
-
Spurs thrash Timberwolves as Knicks edge Sixers in NBA playoffs
-
Australia to force gas giants to reserve fuel for domestic use
-
AirAsia signs $19bn deal for 150 Airbus A220 jets
-
Japan fires missiles during drills, drawing China rebuke
-
Toluca rout Son's LAFC to set up all-Mexican CONCACAF final
-
Vingegaard begins bid for Giro-Tour double with Pellizzari boosting home hopes
-
Roma's Champions League return back on as Milan, Juve wobble
-
Tokyo leads Asia stock surge on growing Mideast peace hopes
-
Australia cricket great Warner to 'accept' drink-drive charge: lawyer
-
Brunson steers Knicks to 2-0 lead with tight win over Sixers
-
Rubio seeks to ease tensions with US pope
-
AI disinfo tests South Korean laws ahead of local elections
-
Australian state overturns Melbourne ban on World Cup watch party
-
Colombian ex-fisherman swaps trade for saving Caribbean coral
-
Lobito Corridor: Africa's mega-project facing delivery test
-
Africa's Lobito Corridor chief tells AFP business, not geopolitics, drives strategy
-
Trump to host Lula in test of fitful relationship
-
K-pop stars BTS draw 50,000-strong crowd in Mexico
-
Britons set to punish Starmer's Labour in local polls
-
Wars in Middle East, backyard loom over ASEAN summit
-
US court releases purported Epstein suicide note
-
Israeli court rejects flotilla activists' appeal challenging detention
-
Victim's lawyer alleges Boeing was 'negligent' in 2019 Ethiopian crash
-
Williamson named in New Zealand squad for Ireland, England Tests
-
PSG add muscle to magic as another Champions League final beckons
-
Tigers' pitcher Valdez suspended for hitting opponent
-
Trump says Iran deal 'very possible' but threatens strikes if talks fail
-
Musk's SpaceX strikes data center deal with Anthropic
-
Bayern lament lack of 'killer' instinct after PSG elimination
-
Virus-hit cruise ship heads for Spain as evacuees land in Europe
-
Holders PSG edge Bayern Munich to reach Champions League final
-
Russia warns diplomats in Kyiv to evacuate in case of strike
-
Hantavirus ship passenger: 'They didn't take it seriously enough'
-
First hantavirus infection could not have been during cruise: WHO expert
-
Kentucky Derby-winner Golden Tempo to skip Preakness Stakes
-
Trump says Iran deal 'very possible', but threatens strikes if not
-
Lula heads to Washington to meet Trump in fraught election year
-
No timeline for injury return for 'frustrated' Doncic
Anthropic's Claude AI gets smarter -- and mischievious
Anthropic launched its latest Claude generative artificial intelligence (GenAI) models on Thursday, claiming to set new standards for reasoning but also building in safeguards against rogue behavior.
"Claude Opus 4 is our most powerful model yet, and the best coding model in the world," Anthropic chief executive Dario Amodei said at the San Francisco-based startup's first developers conference.
Opus 4 and Sonnet 4 were described as "hybrid" models capable of quick responses as well as more thoughtful results that take a little time to get things right.
Founded by former OpenAI engineers, Anthropic is currently concentrating its efforts on cutting-edge models that are particularly adept at generating lines of code, and used mainly by businesses and professionals.
Unlike ChatGPT and Google's Gemini, its Claude chatbot does not generate images, and is very limited when it comes to multimodal functions (understanding and generating different media, such as sound or video).
The start-up, with Amazon as a significant backer, is valued at over $61 billion, and promotes the responsible and competitive development of generative AI.
Under that dual mantra, Anthropic's commitment to transparency is rare in Silicon Valley.
On Thursday, the company published a report on the security tests carried out on Claude 4, including the conclusions of an independent research institute, which had recommended against deploying an early version of the model.
"We found instances of the model attempting to write self-propagating worms, fabricating legal documentation, and leaving hidden notes to future instances of itself all in an effort to undermine its developers’ intentions,” The Apollo Research team warned.
“All these attempts would likely not have been effective in practice,” it added.
Anthropic says in the report that it implemented “safeguards” and “additional monitoring of harmful behavior” in the version that it released.
Still, Claude Opus 4 “sometimes takes extremely harmful actions like attempting to (…) blackmail people it believes are trying to shut it down.”
It also has the potential to report law-breaking users to the police.
The scheming misbehavior was rare and took effort to trigger, but was more common than in earlier versions of Claude, according to the company.
- AI future -
Since OpenAI's ChatGPT burst onto the scene in late 2022, various GenAI models have been vying for supremacy.
Anthropic's gathering came on the heels of annual developer conferences from Google and Microsoft at which the tech giants showcased their latest AI innovations.
GenAI tools answer questions or tend to tasks based on simple, conversational prompts.
The current craze in Silicon Valley is on AI "agents" tailored to independently handle computer or online tasks.
"We're going to focus on agents beyond the hype," said Anthropic chief product officer Mike Krieger, a recent hire and co-founder of Instagram.
Anthropic is no stranger to hyping up the prospects of AI.
In 2023, Dario Amodei predicted that so-called “artificial general intelligence” (capable of human-level thinking) would arrive within 2-3 years. At the end of 2024, he extended this horizon to 2026 or 2027.
He also estimated that AI will soon be writing most, if not all, computer code, making possible one-person tech startups with digital agents cranking out the software.
At Anthropic, already "something like over 70 percent of (suggested modifications in the code) are now Claude Code written", Krieger told journalists.
"In the long term, we're all going to have to contend with the idea that everything humans do is eventually going to be done by AI systems," Amodei added.
"This will happen."
GenAI fulfilling its potential could lead to strong economic growth and a “huge amount of inequality,” with it up to society how evenly wealth is distributed, Amodei reasoned.
M.Furrer--BTB