Anthropic unveiled its newest technology of “frontier,” or cutting-edge, AI fashions, Claude Opus 4 and Claude Sonnet 4, throughout its first convention for builders on Thursday in San Francisco. The AI startup, valued at over $61 billion, stated in a weblog put up that the brand new, highly-anticipated Opus mannequin is “the world’s finest coding mannequin,” and “delivers sustained efficiency on long-running duties that require centered effort and 1000’s of steps.” AI brokers powered by the brand new fashions can analyze 1000’s of information sources and carry out complicated actions.
The brand new launch underscores the fierce competitors amongst corporations racing to construct the world’s most superior AI fashions—particularly in areas like software program coding–and implement new methods for pace and effectivity, as Google did this week with its experimental analysis mannequin demo known as Gemini Diffusion. On a benchmark evaluating how effectively totally different massive language fashions carry out on software program engineering duties, Anthropic’s two fashions beat OpenAI’s newest fashions, whereas Google’s finest mannequin lagged behind.

Some early testers have already had entry to the mannequin to attempt it out in real-world duties. In a single instance offered by the corporate, a common supervisor of AI at buying rewards firm Rakuten stated Opus 4 “coded autonomously for practically seven hours” after being deployed on a fancy undertaking.
Dianne Penn, a member of Anthropic’s technical workers, instructed Fortune that “that is truly a really massive change and leap when it comes to what these AI programs can do,” notably because the fashions advance from serving as “copilots,” or assistants, to “brokers,” or digital collaborators that may work autonomously on behalf of the person.
Claude Opus 4 has some new capabilities, she added, together with following directions extra exactly and enchancment in its “reminiscence” capabilities. Traditionally, these programs don’t bear in mind every thing they’ve achieved earlier than, stated Penn, however “we have been deliberate to have the ability to unlock long-term job consciousness.” The mannequin makes use of a file system of kinds to maintain monitor of progress, after which strategically checks on what’s saved in reminiscence with the intention to tackle extra subsequent steps—simply as a human modifications its plans and techniques primarily based on real-world conditions.
Each fashions can alternate between reasoning and utilizing instruments like internet search, they usually may use a number of instruments directly—like looking the net and operating a code check.
“We actually see this can be a race to the highest,” stated Michael Gerstenhaber, AI platform product lead at Anthropic. “We wish to guarantee that AI improves for everyone, that we’re placing strain on all of the labs to extend that in a protected method.” That features exhibiting the corporate’s personal security requirements, he defined.
Claude 4 Opus is launching with stricter security protocols than any earlier Anthropic mannequin. The corporate’s Accountable Scaling Coverage (RSP) is a public dedication that was initially launched in September 2023 and maintained that Anthropic wouldn’t “prepare or deploy fashions able to inflicting catastrophic hurt except we have now applied security and safety measures that can preserve dangers beneath acceptable ranges.” Anthropic was based in 2021 by former OpenAI workers who have been involved that OpenAI was prioritizing pace and scale over security and governance.
In October 2024, the corporate up to date its RSP with a “extra versatile and nuanced strategy to assessing and managing AI dangers whereas sustaining our dedication to not prepare or deploy fashions except we have now applied enough safeguards.”
Till now, Anthropic’s fashions have all been categorized underneath an AI Security Stage 2 (ASL-2) underneath the corporate’s Accountable Scaling Coverage, which “present a baseline stage of protected deployment and mannequin safety for AI fashions.” Whereas an Anthropic spokesperson stated the corporate hasn’t dominated out that its new Claude Opus 4 might meet the ASL-2 threshold, it’s proactively launching the mannequin underneath the stricter ASL-3 security normal—requiring enhanced protections towards mannequin theft and misuse, together with stronger defenses to forestall the discharge of dangerous data or entry to the mannequin’s inside “weights.”
Fashions which are categorized in Anthropic’s third security stage meet extra harmful functionality thresholds, in accordance with the corporate’s accountable scaling coverage, and are highly effective sufficient to pose vital dangers reminiscent of aiding within the growth of weapons or automating AI R&D. Anthropic confirmed that Opus 4 doesn’t require the very best stage of protections, categorized as ASL-4.
“We anticipated that we’d do that once we launched our final mannequin, Claude 3.7 Sonnet,” stated the Anthropic spokesperson. “In that case, we decided that the mannequin didn’t require the protections of the ASL-3 Commonplace. However we acknowledged the very actual risk that given the tempo of progress, close to future fashions would possibly warrant these enhanced measures.”
Within the lead as much as releasing Claude 4 Opus, she defined, Anthropic proactively determined to launch it underneath the ASL-3 Commonplace. “This strategy allowed us to deal with creating, testing, and refining these protections earlier than we would have liked. We’ve dominated out that the mannequin requires ASL-4 safeguards primarily based on our testing.” Anthropic didn’t say what triggered the choice to maneuver to ASL-3.
Anthropic has additionally at all times launched mannequin, or system, playing cards with its launches, which offer detailed data on the fashions’ capabilities and security evaluations. Penn instructed Fortune that Anthropic could be releasing a mannequin card with its new launch of Opus 4 and Sonnet 4, and a spokesperson confirmed it might be launched when the mannequin launches at this time.
Not too long ago, corporations together with OpenAI and Google have delayed releasing mannequin playing cards. In April, OpenAI was criticized for releasing its GPT-4.1 mannequin with out a mannequin card as a result of the corporate stated it was not a “frontier” mannequin and didn’t require one. And in March, Google revealed its Gemini 2.5 Professional mannequin card weeks after the mannequin’s launch, and an AI governance professional criticized it as “meager” and “worrisome.”
This story was initially featured on Fortune.com