It is like a calculator that only worked in one digit, and now it works on 2, the improvement is immense but its still nowhere close to replacing mathematicians since it isn't even working on the same kind of problems.
Edit: In several years we might have a perfect calculator that is better than any human at such tasks, but it still doesn't beat humans at stuff unrelated to calculations. Or in the case of LLMs pattern matching texts, humans don't pattern match texts to plan or mentally simulate scenarios etc, that part isn't covered by LLMs. Human level planning with todays LLM level pattern matching on text would be really useful, we see a lot of humans work that way by using the LLM as a pattern matcher, but there is no progress on automating human level planning so far, LLMs aren't it.
GPT-3.5 was released in March 2022. We are now in June 2024. Over 2 years later.
And on average GPT-4 is about 40% more accurate.
For me, LLMs are very much like self-driving cars. On the journey towards perfect accuracy it gets progressively harder to make advancements.
And for it to replace the status quo it really does need to be perfect. And there is no evidence or research that this is possible.
Ppl dont want to hear that, but you see less and less offers and not only for junior positions.
Hard truth is that like with any tool/automation - the higher performance improves, the less ppl are needed for this kind of work.
Just look at how some parts of manual labor were made redundant.
Why ppl think it wont be the same with mental work is beyond me.