- Groq has exactly 0 dollars in revenue
- Groq requires 576 chips to run a single model
- Groq can do low latency inference, but can't handle batches, and can't run a diversity of different models on each deployment
- Groq quantizes the models, significantly affecting quality to get more speed (and don't communicate this to end users, which is very deceptive)
- Groq can only run inference, cannot train on their systems
- SambaNova has real revenue from big customers
- SambaNova can run any model on a single node at the speed Groq requires
- SambaNova can do low latency inference just like Groq, but can also run large batches and host hundreds of models on a single deployment
- SambaNova does not quantize models unless explicitly stated
- SambaNova can run training at perf competitive with Nvidia, as well as fastest inference in the world at full precision
It really isn't a competition. Groq has done great as garnering hype in recent months, but it is a house of cards.