>> Predicting language is being able to tell if input is valid or invalid.
If this were the case then the hallucination problem would be solvable.
That hallucination problem is not only going to be hard to detect in any meaningful way but it's going to be harder to eliminate. The very nature of LLM (mixing in noise aka temperature) means that they always risk going off the rails. This is the same thing Lorenz discovered in modeling weather...