Discussion about this post

User's avatar
Philip Koop's avatar

ChatGPT labelled that graph 10^3 (Billions) instead of just 10^12, obliging it to start at 10^-1. Did it copy that tic from its source, I wonder?

Geoff Hinton is fond of saying that back in the 80's AI researchers were taking a sui generis approach to each problem domain. He and his colleagues were running around saying that their neural net approach was the only approach needed, it could replace all the others. But at the time, it didn't work. There were only two problems: they didn't have enough data, and they didn't have enough compute. That's not surprising if you need a trillion parameters in practical situations.

The current performance of LLMs is due to remedying these two defects. So far, my experience with co-piloting is underwhelming; they slow me down by nearly as much as they speed me up. Are improvements in the intelligence part of AI extrapolated from this sort of exponential graph? Training a very large LLM is already very expensive; will the next iteration be *exponentially* more expensive?

Expand full comment
Sarora's avatar

For example, title of chapter 2 in that Fogel Book: Technological Change, Cultural Transformation and Political Crises. You've probably covered some of it in Slouching. But more could be said, too, about things that Fogel didn't see coming etc. Roughly 25 years have gone by since.

Expand full comment
12 more comments...

No posts