Yann LeCun said a number of things that are very dubious, like autoregressive LLMs are a dead end, LLMs do not have an internal world model, and this morning https://www.youtube.com/watch?v=AFi1TPiB058 (in french) that an IA cannot find a strategy to preserve itself against the will of its creator.
As a french, I wish him good luck anyway, I'm all for exploring different avenues of achieving AGI.
What use is it to understand the physical world if all investments are misallocated to the virtual world? Perhaps the AI will detect that there is a housing shortage and politicians will finally believe it because AI said so?
Or is it to accelerate Skynet?
Alternative free to read article: https://sifted.eu/articles/yann-lecun-ami-labs-meta-funding-...
If he's right (that LLMs cannot achieve AGI, but what he's working on can, and does), this would be huge for AI and humanity at large.
Hope it puts to bed the "Europe can't innovate" crowd too.
[dead]
Why world model? To emulate how we became sentient?
A "world" is just senses. In a way the context is one sense. A digital only world is still a world.
I think more success is in a model having high level needs and aspirations that are borne from lower level needs. Model architecture also needs to shift to multiple autonomous systems that interact, in the same ways our brains work - there's a lot under the surface inside our heads, it's not just "us" in there.
We only interact with our environment because of our low level needs, which are primarily: food, water. Secondary: mating. Tertiary: social/tribal credit (which can enable food, water and mating).