Xeons have a much longer shelf life and diverse workloads. If you order hardware specifically for LLM inference and then some new hardware/model combination is much better at that (which it will be, because a lot of people are working on that), you might be in trouble.
It's like setting up a warehouse of GPUs to mine bitcoin while others are switching to ASICs.