Regarding the substance of the article, the curve from the 3 data points (1: 100x 2: 25x 3: 25x) could fit lots of different ways besides "growing 30x per generation".
It’s not good that GPUs are this opaque.
People aren't going to do truly uneconomic things just to scale language models exponentially.
GPT-4 needed about 50 gigawatt-hours of energy to train. Using our scaling factor of 30x, we expect GPT-5 to need 1,500, GPT-6 to need 45,000, and GPT-7 to need 1.3 million.
https://news.microsoft.com/europe/features/as-the-world-goes...
(Plus "renewable" is kind of a misnomer, there are still limited resources going into the production and maintenance of those power plants too.)
This should not be a surprise.