We are on this part of the ai takeoff graph. https://waitbutwhy.com/2015/01/artificial-intelligence-revol...
People had no reason to believe one day we would finally understand what causes the thunder. We finally did, and it is not made by Zeus.
I would not be shocked to find out that AGI (using Altman's definition) is more than 50 years away, but I also would not be shocked if it came in 5.
It's really hard to know how scared to be, I think that rationally I should be pretty terrified but I'm not.
We're also seeing lots of optimizations with new models (RoPE/RoPER embedding, Swish/GeLU activation, Flash Attention, etc) but I think some the most interesting gains we'll be seeing soon is with inference-optimized training (-70% parameters for +100% compute) [1] combined with sparsity pruning (-50% size w/ almost no loss in accuracy) [2] and quantization [3] which will lead to significantly smaller models performing well.
[1] https://www.harmdevries.com/post/model-size-vs-compute-overh...
they're also not going to find another 2, 4, 8, 16 ... internets worth of content to parasitise
Not all types of AI need external training data, you can train on how effectively a goal is achieved