I thought Mixtral's release was weird when they just pasted a magnet link [0] into Twitter with no information, but at least people could download and analyze it so we got some reasonable third-party commentary in between that and the official announcement. With this one there's nothing at all to go on besides the name and the black box.
(I know that Mistral does a lot more stuff in the open than other companies, just couldn't resist the parallel between this and the blackbox limitations of LLMs in general)
To be fair, this is not a release. This was the previous release https://mistral.ai/news/mixtral-of-experts/
It looks more like not trying very hard to hide things until release, rather than being a black box.
Click/tap on "Direct Chat" in the top tab navigation and you can select "mistral-next" as model.
There's got to be a better name for such a cool product. Maybe MistralX? MistMix?
Also, do coding LLMs use treesitter to "understand" code?
Best models currently: codellama or deepseek coder. 6.7B or 1B depending on how much latency you can tolerate
Treesittee: from looking at the logs of the chat completions requests for Continue or Twinny extensions for VS Code, they both appear to just send a chunk of the document as well as a special placeholder to indicate where the cursor currently is.
If you're wanting something to be your hands so you don't have to type, open-source LLMs and IDE integration is not reliably there yet. Follow the AIDER discord to stay up on the latest in this area.
It's up to the app to put that into the context. Generally, coding LLMs do well if you provide them the source tree, graph, search results, notable files, etc in the context. The is how Sourcegraph's Cody product works, for example.
I've been quite disappointed by French LLMs on Huggingface when I tried a month ago.
Very exciting nevertheless, here’s hoping the bless the OS community once again!
Refusals are a bit "I am just a language model"-y which GPT-4 has gotten away from. Also it's more refuse-y if I broach something rudely (which again I've found GPT-4 to have become much better at.)
Way better at everything than whichever Gemini I've been trying recently (can't tell for sure what I'm using when I use it.) But that one isn't even in contention for any use at all IME.
Overall it felt like I need to try it in daily use to work out if it's a contender with GPT-4 as a daily driver.
It's a preview of their newest prototype model.
To use it, click "Direct Chat" tab and choose "Mistral next"