Discussion about this post

User's avatar
Rangachari Anand's avatar

Really liked this passage:

I’d guess that it’s actually our human limitations—constraints on working memory, on processing speed, on available energy—as well as our continually changing and complex environments that require us to form more abstract and generalizable internal models.

An analogy I'd like to suggest is to consider the difference in the way birds and airplanes fly. Birds are far more efficient in using available power but airplanes have so much power to spare that it doesn't matter.

Expand full comment
Daniel Visser's avatar

Amazing post, as always :) More evidence (at least in my interpretation) for the "bag of heuristics" explanation can be found in On the Geometry of Deep Learning by Prof. Balestriero and colleagues, which discusses how these models effectively rely on enormous, piecewise-linear tilings to collectively solve tasks – https://arxiv.org/abs/2408.04809.

Expand full comment
43 more comments...

No posts