The Mixtral 8x22B and its variants, Mixtral-8x22B-v0.1 and Mixtral 8x22B-Instruct-v0.1, have been officially released and are now available for deployment. This model features a Mixture of Experts architecture and supports multilingual capabilities in English, French, Italian, German, and Spanish. It is designed for high performance with 39 billion active parameters and a 64K token context window. The Mixtral 8x22B is ranked as the number one pretrained model on the Open LLM leaderboard and is licensed under Apache 2.0. It utilizes four A100 GPUs in fp16, with optimized implementations expected soon. The model is also capable of native function calling and excels in math and code tasks.
We just released Mixtral-8x22B-v0.1 and Mixtral-8x22B-Instruct-v0.1: - Free to use under Apache 2.0 license - Outperforms all open models - Native function calling - Masters English, French, Italian, German and Spanish. - Seq_len = 64K https://t.co/3ipqZA0Wkr https://t.co/Zt83yvDPkn
Super excited for our new Mixtral 8x22B model! 🔸capable of function calling 🔸64K tokens context window 🔸fluent in English, French, Italian, German, and Spanish 🔸strong maths and coding capabilities https://t.co/u4roKSijhm
Mixtral 8x22B sets a new standard for performance and efficiency for the AI community. Apache 2.0. https://t.co/kcdL3BYX48
The official Apache 2 Mixtral 8x22B Instruct model is out! 🔥 🌍Multilingual (en/fr/it/de/es) 🧠Math and code capabilities ✏️Native function calling ⚡️39B active params 🤯64k context window Model: https://t.co/pSAu2QKGP8 Announcement: https://t.co/tqKx4epJ5O Let's go! 🚀
Deploy Mixtral 8x22B in one click! Mixtral fast facts: - #1 pretrained model on the Open LLM leaderboard - Mixture of Experts architecture - Apache 2.0 license - Uses 4 A100s in fp16, optimized implementations coming soon! https://t.co/MByCvYdsvY
🚨 New model alert! Mixtral 8x22B is a new sate-of-the-art Mixtral model that can be used for general chat applications. Start using it today: https://t.co/MByCvYdsvY