Jade Small
Jade Small
November 13, 2024 ·  4 min read

Apple Speaks the Truth About AI. It’s Not Good.

Apple’s recent research challenges the popular belief that large language models (LLMs) like ChatGPT and Llama truly possess intelligence. Instead, Apple suggests that these AI systems rely more on memorization than on genuine reasoning, raising concerns about the future of AI and the billions invested by tech giants. So, what exactly did Apple uncover?

A Widespread Disappointment with AI

Despite the excitement surrounding new models, there’s growing disillusionment in the AI community. Many thought the latest LLMs would bring revolutionary advancements, but Apple’s findings reveal that they still struggle with the same issues as earlier models. In essence, their intelligence seems overstated, leaving some to wonder if AI’s potential has been exaggerated.

Proof Piling Up Against LLMs

Recent findings have shown that LLMs fall short in critical areas, such as complex problem-solving and anomaly detection. For example, tests by MIT researchers reveal that these models underperform in tasks like time-series predictions when compared to even older statistical methods. Apple researchers took this a step further, showing that AI performance dramatically declines when models encounter new topics or are asked to follow complex instructions.

Apple’s Bold Statement

Apple researchers didn’t hold back, asserting that “LLMs do not perform genuine reasoning.” Their analysis suggests that these models’ abilities are largely based on recognizing patterns rather than deep understanding. For instance, when tested with simple math problems, minor changes in question phrasing confused the AI, proving its reliance on familiar sequences instead of true comprehension.

Token Bias and Memorization Issues

A core issue is what researchers call “token bias.” For instance, in a test on logical fallacies, the model could answer correctly only when the name “Linda” was used—since that name was common in its training data. However, changing the name to “Bob” led to failure, underscoring how LLMs memorize specific sequences without fully understanding the underlying concepts. This finding raises questions about the reliability of models that depend so heavily on memorized information.

The Role of Complexity in Model Performance

Apple’s experiments showed that as the difficulty of a question increased, AI performance suffered. By creating tests that added progressively harder components, researchers found that the models struggled with more complex reasoning tasks. Even the most advanced LLMs showed diminished performance as task complexity rose, indicating they may not be ready for applications requiring nuanced understanding.

Easily Fooled by Irrelevant Details

The Apple team further revealed that LLMs are easily misled by irrelevant details in problem statements. They tested this by adding unnecessary clauses to questions, finding that LLMs often focused on these extraneous details and produced incorrect answers. This suggests that the models aren’t discerning between essential and non-essential information, a skill vital for genuine intelligence.

Scaling May Not Be the Solution

While some AI proponents believe that simply scaling up model size can solve these issues, others, including Apple, remain skeptical. Gary Marcus, a noted AI skeptic, argues that simply adding more data might improve performance on familiar tasks but doesn’t address the fundamental problem of limited reasoning ability. According to him, larger models may appear smarter due to vast amounts of memorized data, not due to genuine reasoning.

Task Familiarity Versus Complexity

Apple’s research supports the idea that AI performance should be evaluated based on familiarity with new tasks, not complexity. AI scientist François Chollet argues that task complexity doesn’t guarantee intelligence, as models can be trained to solve certain problems without really “thinking” about them. Apple’s work aligns with this idea, suggesting that future AI benchmarks should focus on tasks the models haven’t encountered to test true reasoning abilities.

The Reality Check

As Apple’s findings shake confidence in AI, some in the industry feel we should proceed with caution. Many believe that, until LLMs can handle unfamiliar tasks without pre-trained solutions, they shouldn’t be considered truly intelligent. Instead, Apple’s research suggests we should view LLMs as tools that enhance human intelligence rather than as independent, intelligent systems.

Final Thoughts

Apple’s study brings a fresh, critical perspective to the discussion around AI. For all the buzz, it seems AI may not yet be the solution to achieving true machine intelligence. Until we see models that can tackle new tasks with genuine reasoning, they might be best regarded as impressive mimics, amplifying human intelligence without embodying it.

This content has, in part, been generated with the aid of an artificial intelligence language model. While we strive for accuracy and quality, please note that the information provided may not be entirely error-free or up-to-date. We recommend independently verifying the content and consulting with professionals for specific advice or information. We do not assume any responsibility or liability for the use or interpretation of this content.