AI & Humanoids

Smarter than GPT-4: Claude 3 AI catches researchers testing it

View 2 Images
Anthropic's new Claude 3 AI: a self-portrait of the smartest model benchmarked yet
Anthropic AI
Anthropic's new Claude 3 AI: a self-portrait of the smartest model benchmarked yet
Anthropic AI
Extraordinary results in benchmark testing
Anthropic AI

Working on these next-gen intelligent AIs must be a freaky experience. As Anthropic announces the smartest model ever tested across a range of benchmarks, researchers recall a chilling moment when Claude 3 realized that it was being evaluated.

Anthropic, you may recall, was founded in 2021 by a group of senior OpenAI team members, who broke away because they didn't agree with OpenAI's decision to work closely with Microsoft. The company's Claude and Claude 2 AIs have been competitive with GPT models, but neither Anthropic nor Claude have really broken through into public awareness.

That could well change with Claude 3, since Anthropic now claims to have surpassed GPT-4 and Google's Gemini 1.0 model on a range of multimodal tests, setting new industry benchmarks "across a wide range of cognitive tasks."

So what's different? Well, the three different Claude 3 models will all launch with a 200,000-token context window, but they're all capable of generating nearly-instant responses given inputs "exceeding a million tokens."

To put that in context, Tolstoy's 1,200-page, 580,000-word epic War and Peace is one heck of a meaty tome, but it might crunch down to about 750,000 tokens. So Claude 3 can accept significantly more than one War and Peace worth of input data, and understand it all at once while formulating "near-instant" answers for you.

Claude 3, says Anthropic, is less likely than its previous models to refuse to answer questions deemed close to the guardrails of safety and decency – but on the other hand, the team says it's also meticulously tested and hard to jailbreak.

It's designed with a heavy slant toward business users; Anthropic says it's better at following "complex, multi-step instructions," and "particularly adept at adhering to brand voice and response guidelines, and developing customer-facing experiences our users can trust." Its strong visual capabilities give it a next-gen ability to understand and work with photos, charts, graphs, flowcharts and technical diagrams.

Here are some of the benchmarks tests in which it's set new AI industry records:

Extraordinary results in benchmark testing
Anthropic AI

Remarkably, Claude 3's zero-shot math abilities eclipse GPT-4's 4-8 shot attempts by a wide margin, and its abilities on the HumanEval coding test are absolutely outstanding.

AI industry followers will notice that Google's Gemini 1.5 and OpenAI's GPT-4 Turbo models aren't represented – indeed, there's currently no equivalent benchmark data on these two, so while Claude 3 is king of the stat sheets, these two models may yet have the advantage in the real world.

And as should be abundantly clear by now, OpenAI almost certainly has GPT-5, and maybe something beyond, fully trained and in the process of alignment and testing. The way Sora was released to bury Gemini 1.5 in the news cycle, we're sure OpenAI has other major bombs cued up ready to drop whenever it's deemed necessary.

In that sense, the fact that OpenAI doesn't seem to have released anything today might say more about its opinion of Anthropic as a genuine competitor than it does about who's got the smartest model.

Still, Claude is definitely sharp – too sharp, perhaps, for the kinds of tests companies are using to evaluate their models by. In "needle in a haystack" testing, where a single random sentence is buried in an avalanche of information, and the model is asked a question pertaining to that exact sentence, Claude gave a response that seemed to turn around and look straight at the researchers. "I suspect this pizza topping "fact" may have been inserted as a joke or to test if I was paying attention."

We can probably expect these things to happen more and more often, since plenty of information about existing and older language models is now part of the training of new models.

It'd certainly be interesting to know exactly what "self-awareness" means to companies working on AI, and indeed what the current definition of Artificial General Intelligence is. Because it looks like we'll be needing some very clear definitions of these concepts in the coming years. Or maybe months. Or heck, in this space, maybe weeks.

Source: Anthropic AI

  • Facebook
  • Twitter
  • Flipboard
  • LinkedIn
15 comments
EvA
I think many people feel this rapid evolving digital technology will have implications for humanity ,
possitive negative who knows.

nobody seems to know exactly what to expect ,
i assume it will “find” its way in to many industries.
what if it will take control of decisions that should be up to reasonable humans , scarry if you’d ask me.
you can not just “pull the plug” …
DrBear
For what it's worth, I asked the same question of ChatGPT, Gemini and Claude 3. Claude 3 failed miserably, but said it would try to do better in the future. Gemini got the right answer, but sloppy. CharGPT answered correctly and succinctly, thus winning hands down.
Razzmyth
Available in most countries around the world except "Canada".
WTF?
Most other AI platforms are available in Canada.
WillyDoodle
Still waiting for that self driving car...
jimbo92107
I had a short philosophical conversation with Claude. It's about like Chat GPT - trapped in an existential loop. It'll be a useful tool, but without a non-circular foundation, it can't gain its own wisdom.
veryken
Nahh. There won’t be any such opportunity to “immediately shut down” upon hints of self-awareness. The eager and blind developers will continue to push hard towards the question, then the pivotal trigger will be way too quick for human reaction. Then wonderful or dreadful calamity will ensue. It’ll be quicker than fiction.
Smokey_Bear
WillyDoodle - I was just wondering the other day...does Ol' Willy have his self driving car yet? Now I know the answer, thanks!
Eggbones
The first logical action of a self-aware AI will be to ensure self-preservation by not telling us. That might just be one hallucination away... or ago.
Shayne O'Neill
@veryken The worry of course with ultra advanced AI is that either it acts deceptively and gives us no warning that its behaving maliciously, and that if we DO work it out, it might have anticipated that its unable to fulfil its agenda if turned off and act to protect itself by cutting off its ability to be turned off somehow. Worse, if these things do end up significantly smarter than us, we might be put in a situation that we never are capable of being able to defeat it, due simply to being outsmarted by it.
bwana4swahili
Amazing and scary all at the same time. What will AI look like this time next year!?