Apple agrees with Sam Altman: AI is incredibly dumb

And old Apple Mac computer running ChatGPT in a browser window that is wearing a dunce cap and sat on a green backdrop.
(Image credit: Apple / Rael Hornby / Laptop Mag)

Is ChatGPT getting smarter, or is it getting better at seeming smart? According to Apple, it’s the latter.

A team of AI researchers at Apple published a paper this weekend claiming that most leading large language AI models aren’t actually capable of advanced reasoning, despite how intelligent they might seem.

Large language models, or LLMs, like ChatGPT appear to be getting more advanced and “intelligent” every year. Under the hood, though, their logical reasoning hasn’t improved much. According to Apple’s research, current LLMs’ capabilities “may resemble sophisticated pattern matching more than true logical reasoning.”

What does this research mean for the reality of today’s top AI models? It might be time to focus on creating safer AI models before trying to build smarter ones.

Apple AI researchers find that LLMs struggle with grade-school math

A team of AI researchers at Apple has revealed the findings of a new benchmark test, GSM-Symbolic, which posed a whole new challenge for large language models.

The test revealed that today’s top AI models have limited reasoning capabilities, despite how intelligent they might seem.

In fact, the GSM-Symbolic test revealed that the AI models in the study struggled with basic grade school math problems. The more complex the questions became, the worse the AIs performed.

The researchers explain in their paper, “Adding seemingly relevant but ultimately inconsequential information to the logical reasoning of the problem led to substantial performance drops of up to 65% across all state-of-the-art models.

"Importantly, we demonstrate that LLMs struggle even when provided with multiple examples of the same question or examples containing similar irrelevant information.”

Sam Altman of OpenAI

Apple isn't the only group that believes AI has a long way to go before being worthy of being labelled truly intelligent, even OpenAI's Sam Altman referred to the company's flagship chatbot ChatGPT as "incredibly dumb" in its current state. (Image credit: Justin Sullivan/Getty Images)

This means today’s leading AI models are easily confused by logic-based questions such as math problems. They rely on copying the patterns in math problems in their training data but struggle to do math the way a human can. This shows that large language models only appear to be smart, when, in reality, they’re just really good at acting smart.

This echoes OpenAI CEO Sam Altman’s remarks, claiming AI is actually “incredibly dumb” in its current state. OpenAI is the company behind ChatGPT and Altman has been ambitious in his pursuit of artificial general intelligence, which would be capable of true logical reasoning.

Apple’s study seems to agree. It concludes, “We believe further research is essential to develop AI models capable of formal reasoning, moving beyond pattern recognition to achieve more robust and generalizable problem-solving skills.”

AI might not be smart yet, but it can still be dangerous

If the research published by Apple’s AI team is accurate, today’s leading large language models struggle to hold up on an episode of Are You Smarter Than a Fifth Grader. However, that doesn’t mean AI can’t still be a powerful tool, one that can be incredibly helpful… or harmful. In fact, the Apple study reveals a core strength and potential danger of AI: its ability to mimic.

LLMs like ChatGPT may seem capable of reasoning the way humans are, but as this study points out, that’s just the AI copying human language and patterns. That might not be as advanced as actual logical reasoning, but AI has gotten extremely good at mimicking others. Unfortunately, bad actors have been quick to take advantage of every advancement.

For example, this weekend tech YouTuber Marques Brownlee announced on X that a company used AI to replicate his voice in an ad for their product, which Brownlee was not affiliated with. The AI-generated decoy is shockingly similar to Brownlee’s real voice, though. The ad was clearly intended to deceive viewers into thinking Brownlee was endorsing their product.

Unfortunately, incidents like this are becoming more common, from fake presidential endorsements from Taylor Swift to Scarlett Johansson’s claims that OpenAI copied her voice without her permission.

Average users might not think these controversies affect them, but they’re arguably the most critical aspect of the AI industry. It’s great that basic tools like ChatGPT or Gemini are useful to many people.

However, the ways AI is also being misused for deep fakes, deception, and scams pose a serious risk to the safety of this technology and everyone who interacts with it, knowingly or otherwise.

More from Laptop Mag

Category
Arrow
Arrow
Back to Apple MacBook Pro
Brand
Arrow
Processor
Arrow
RAM
Arrow
Storage Size
Arrow
Screen Size
Arrow
Colour
Arrow
Screen Type
Arrow
Condition
Arrow
Price
Arrow
Any Price
Showing 10 of 425 deals
Filters
Arrow
Load more deals
Stevie Bonifield
Freelance Writer

Stevie Bonifield is a freelance tech journalist specializing in keyboards, peripherals, gaming gear, and mobile tech. Outside of writing, Stevie loves indie games, photography, and building way too many custom keyboards