> the LLM is deterministic as far as I understand it, it's the sampling at the end that isn't.
I guess it depends how you define the LLM: you could say it was the model/NN and the sampler is an extra added on, but a lot of people would name the model+sampler+system prompt+RLHF tuning (which would include the sampler) as the LLM.
The OP was talking about ChatGPT generating fixed output, not an internal model