Mistral AI launches Mixtral-Next

by varunvummadion 2/17/2024, 3:46 AMwith 49 comments

by lolinderon 2/17/2024, 5:44 PM

Mistral's process for releasing new models is extremely low-information. After getting very confused by this link I tried looking for a link that has any better information, and there just isn't one.

I thought Mixtral's release was weird when they just pasted a magnet link [0] into Twitter with no information, but at least people could download and analyze it so we got some reasonable third-party commentary in between that and the official announcement. With this one there's nothing at all to go on besides the name and the black box.

[0] https://news.ycombinator.com/item?id=38570537

by mattpavelleon 2/17/2024, 12:25 PM

For those unfamiliar with the LMSys interface:

Click/tap on "Direct Chat" in the top tab navigation and you can select "mistral-next" as model.

by twobitshifteron 2/17/2024, 2:11 PM

AIExplained on youtube has guessed that Gemini 1.5 pro is taking Mistral’s accurate long content retrieval and Google just scaled it as much as they could. The Gemini 1.5 pro paper has a citation back to the last mistral paper in 2024.

by mrfakenameon 2/17/2024, 5:10 PM

Note that it's actually "Mistral Next" not "Mixtral Next" - so it isn't necessarily a MoE. For example, an early version of Mistral Medium (Miqu) was not a MoE but instead a Llama 70B model. I wonder how many parameters this one has

by bloopernovaon 2/17/2024, 3:36 PM

Slightly related question: what's a good coding LLM to run on a 4070 12GB card?

Also, do coding LLMs use treesitter to "understand" code?

by aunetxon 2/17/2024, 1:52 PM

It's quite funny to use! It is better when speaking French than chat gpt3.5 on my opinion

by justanotherjoeon 2/17/2024, 2:35 PM

wow, this might be the best LLM that i've used in terms of phrasing and presenting the answers.

by AnujNayyaron 2/17/2024, 11:16 AM

No indication that this a MoE (Mistral not Mixtral).

Very exciting nevertheless, here’s hoping the bless the OS community once again!

by ismailmajon 2/17/2024, 12:11 PM

Could it be Mistral Large? This beats GPT-4 on my personal test.

by tmikaeldon 2/17/2024, 12:18 PM

This was linked randomly on Mistrals Discord chat, nothing "official" yet.

It's a preview of their newest prototype model.

To use it, click "Direct Chat" tab and choose "Mistral next"

by ccwilson10on 2/18/2024, 5:55 AM

I used this but, upon asking which model it is, it replied as being a "fine-tuned version of GPT 3.5". Any clue why? In a second chat it replied "You're chatting with one of the fine-tuned versions of the OpenAssistant model!".

by vitorgrson 2/17/2024, 6:18 PM

From my tests, it did better than Gemini Ultra on a few reason/logic questions.

by apapapaon 2/17/2024, 6:32 PM

The Together.AI logo at the bottom is very hard to read... (Dark gray on black)

by redder23on 2/17/2024, 6:01 PM

You can literally type "woke shit" in and you get woke shit out. I am so impressed.

by xeckron 2/17/2024, 6:50 PM

As someone who has only been using GPT-4 since its release, I am pleasantly surprised by how far open LLMs have come.