I don't know, I think it's pretty important to a conversation about automating jobs that we not treat jobs as if they're an optimized equivalent exchange. The whole premise there is that the machines are going to start producing a lot of value without any human input. If it happens, it's not going to be equivalent exchange.
If we're not talking about added value, then there's nothing to worry about because then the machines can't automate the jobs, because that would be added value.
> if your work has no value, there is no breakfast to exchange it for
My point more specifically here is that there's a lot of value in the world that can't be exchanged for breakfast, something that (as far as I can tell from your other comments on this thread) you actually agree with, right? There's value that exists outside of traditionally monetarily compensated jobs.
In a (again, entirely theoretical) world where AI gets rid of the need to earn your breakfast, that doesn't mean there's not going to be anything of value to do anymore and that everyone's life is going to be meaningless.
Money is simply a universal equivalent of that value, nothing more. So if you give me a breakfast for help with unloading your truck, it's still an exchange, and my work's value is equal to "one breakfast" even no money was involved.
AI is just a nail gun, there still needs to be somebody who trains/operates it (like with ChatGPT) and who will ask for something in exchange. It may eliminate some jobs (happened before), but it will create others because human society is not only about food (not sure ChatGPT can help here though), but about interactions which are not going to be automated by ChatGPT (think of sex, power).
I feel like your definition of value is jumping around quite a bit. In another comment you described "value" as essentially any kind of social capital or personal reward. Now it's explicitly transactional?
In any case, this is a very narrow definition of value that I don't think matches up with what most people who worry about automation are talking about. It's certainly not what the original thread was talking about when it worried about people without jobs not being able to find meaning in work. What most people think of as useful or meaningful purpose in their lives does not strictly map to transactional value.
> AI is just a nail gun, there still needs to be somebody who trains/operates it (like with ChatGPT) and who will ask for something in exchange.
Then there is no problem! This isn't an issue if AI isn't going to take jobs away.