A group of Apple researchers has published a paper claiming that large language models (LLMs), the backbone of some of AI's most popular products

Apple Speaks the Truth About AI. It’s Not Good.

submited by
Style Pass
2024-12-01 02:30:05

A group of Apple researchers has published a paper claiming that large language models (LLMs), the backbone of some of AI's most popular products today, like ChatGPT or Llama, can’t genuinely reason, meaning their intelligence claims are highly overstated (or from a cynical perspective, that we are being lied to).

Through a series of tests, they prove that their capacity to reason is most often — or totally — a factor of memorization and not real intelligence.

This adds to the growing trend of disillusionment around LLMs, which could cause a massive shift in investment and directly impact the future of many multi-billion-dollar start-ups. Naturally, it also seriously questions Big Tech's billion-dollar AI expenditures and frontier AI labs’ future, who are dependent on this precise vision being true.

This article is an extract from my newsletter, the place where AI analysts, strategists, and decision-makers use to find answers to the most pressing questions in AI.

Leave a Comment