Richard Sutton, father of RL, thinks LLMs are a dead end

Interview by Dwarkesh Patel with Richard Sutton on LLMs, AGI and why he believes AGI will not be born from using LLMs.

In his essay The Bitter Lesson, he argues for not building AI from mimicking human behavior in our systems: "breakthrough progress eventually arrives by an opposing approach based on scaling computation by search and learning."

Solving a Million-Step LLM Task with Zero Errors

Wow! Doing Towers of Hanoi by LLM is something that can be done, by voting on small tasks.

LLMs as mirrors

Wow:

In my view, LLMs are essentially mirrors. They mirror the skill of the operator.

The article goes on to argue that using AI (correctly) is also a matter of skill.

expand_less brightness_auto