Idk I think there's a bit of a difference between a session for some basic website vs machine learning stuff. The base perf cost per user is muuuuuch higher for ML.
Yeah but Google missed the boat when it came to hardware accelerators specifically meant for LLMs (their proprietary TPUs aren't optimized for LLMs) so it's just a matter of whether Google or Microsoft paid Nvidia more. In the current cost cutting climate at Google I doubt the answer is so certain.
The extension with gpt4 as a backend was ime extremely slow as standard. I've not tried it again with the v7 model though which is supposed to be a lot faster