At work:
That I don't rent $30,000 a month of PTUs from Microsoft. That I can put more restricted data classifications into it.
> LLM inferencing isn't particularly constrained by Internet latency
But user experience is