Mistral vs Mixtral: Comparing the 7B, 8x7B, and 8x22B Large Language Models | by Dmitrii Eliuseev | Apr, 2024


Running the 7B and 22B Models in Google Colab

Image by Img IX, Unsplash

Not so long ago, all IT news channels reported about the new open Mixtral 8x22B model, which outperforms ChatGPT 3.5 on benchmarks like MMLU (massive multitask language understanding) or WinoGrande (commonsense reasoning). This is a great achievement for the…



Source link

Be the first to comment

Leave a Reply

Your email address will not be published.


*