I don't think that the scaling hypothesis gets recognized enough for how radical it is. For decades, AI sought some kind of master algorithm of intelligence. The scaling hypothesis says that there is none: intelligence is the ability to use more compute on more data.
Gary Marcus • Deep Learning Is Hitting a Wall
As we successfully apply simpler, narrow versions of intelligence that benefit from faster computers and lots of data, we are not making incremental progress, but rather picking low-hanging fruit. The jump to general “common sense” is completely different, and there’s no known path from the one to the other.
Erik J. Larson • The Myth of Artificial Intelligence: Why Computers Can’t Think the Way We Do
Erik Hoel • How to navigate the AI apocalypse as a sane person
Azeem Azhar • 🧠 AI’s $100bn question: The scaling ceiling
Hannes Malmberg • How mathematics built the modern world - Works in Progress
Suppose non-conscious algorithms could eventually outperform conscious intelligence in all known data-processing tasks – what, if anything, would be lost by replacing conscious intelligence with superior non-conscious algorithms?