"The Illusion of Thinking" - LLMs face "complete accuracy collapse" beyond certain complexities....
Apple’s Machine Learning Research group has published another damning piece of analysis of the capabilities of Large Reasoning Models (LRMs) and Large Language Models (LLMs). This important piece of work further exposes the limitations of LRMs and LLMs - both types of model experience “complete collapse” when dealigning with high-complexity tasks.
The paper is available here: https://machinelearning.apple.com/research/illusion-of-thinking
The key takeaway from this is that Generative and/or General AI are far far away from being safe and useful. More limited Machine Learning algorithms designed to analyse specific types of data are clearly useful. But we should be very wary of extrapolating from this to more “general” AI models.