One of the economic tuning features of an LLM is to nudge the LLM into reaching conclusions and spending the tokens you want it to spend for the question.
presumably everyone running a form of ralph loop against every single workload is a doomsday situation for LLM providers.