Oops, you are right. The code is Apache 2.0, license for the model weights is separate.
Translation error? output from GPT: "non-exclusive, global, non-transferable, non-sublicensable, revocable, royalty-free license."
Does anybody know if performance could be greatly increased if only a single language was supported ?
I suspect there’s a high demand for models that are maybe smaller and can run faster if the tradeoff is support for only English.
Is this available in ollama ?
But according to their own evaluation further down, gpt-4o-2024-05-13 outperforms GLM-4V-9B on every task except OCRBench.
Llama-3-8b was garbage for me but damn 70b is good enough
So you’d at minimum be looking at dual 3090 with NVLink for about $4000 or so. Or for the highest performing non-quantized model, you’d be spending about $40,000 for two A100’s.