Mistral AI has released the Mixtral-8x7B model, an unconventional torrent-based release that has shaken up the AI community. The model is a 70B GPT 3.5 class model, outperforming Llama 2-70B and matching or outperforming GPT 3.5. It is supported in the latest release of transformers and is free to use under the Apache 2.0 license. The model masters English, French, Italian, German, and Spanish, with a sequence length of 32K. Mistral AI has also opened beta access to their first platform services, offering chat endpoints for generating text and an embedding endpoint with different performance/price tradeoffs.
Have you heard of Mixture of Experts (MoE) models? 🤔 With the release of @MistralAI Mixtral 8x7B, MoEs are gaining attention, it is also rumored @OpenAI GPT-4 is an MoE👀 But what exactly are MoEs, and how do they work? We created an in-depth blog. https://t.co/a4DL1A8zCe
Big. The @MistralAI API is out. "la plateforme" serves three chat endpoints for generating text following textual instructions and an embedding endpoint. : Mistral-tiny: Affordable, serves Mistral 7B Instruct v0.2, English only, scores 7.6 on MT-Bench. Mistral-small: New… https://t.co/SXOp3MPKTP
Big. @MistralAI just opened beta access to their first platform services. "la plateforme" serves three chat endpoints for generating text following textual instructions and an embedding endpoint. Each endpoint has a different performance/price tradeoff. Mistral-tiny:… https://t.co/JjBmM8cq78
Mistral’s first AI endpoints are here! Things are about to get super interesting in the ecosystem. https://t.co/SVK9kcWdqj https://t.co/lavALHGqyh
As of like fifteen minutes ago, Mixtral-8x7B from @mistralai is fully supported in @huggingface transformers. https://t.co/T5al1h8skT
Proud to announce: Mixtral 8x7B -- Mixtral of Experts - Free to use under Apache 2.0 license - outperforms Llama 2 70B with 6x faster inference. - matches or outperforms GPT3.5 - masters English, French, Italian, German and Spanish. - seq_len = 32K https://t.co/mRMQYtxmx6 1/N https://t.co/BnLzE7fXzR
🤗@MistralAI's new MOE model (Mixtral, what a nice name) is now supported in the latest release of transformers (make sure you have 4.36.0) 🥳🤗 https://t.co/mvnztVuR9u
.@MistralAI platform just dropped https://t.co/4nE9U0tsRi
Good job @MistralAI with Mixtral-8x7B outperforming LLaMA-2-70b https://t.co/25m3ZDlpaw love that they release Apache 2.0 licensed models (in europe!)
interesting: mistral has their largest model, "mistral-medium", on their cloud API. no details on what it is, or if it'll ever be open-sourced, other than that it outperforms mixtral-8x7B by a long shot. https://t.co/VaWhLOBPtt
.@MistralAI just released their blog post on Mixtral MoE, read about it here: https://t.co/5xGmu7l8w7
Transformer MoE Architectures - Why They Are More Efficient Mistral 8x7B MoE model is a solid 70B GPT 3.5 class model. Instead of having every part of the model work on every task, an MoE model splits the work among many specialized sub-models, or "experts." Each expert is good… https://t.co/aoDfAcfXWO https://t.co/wYVYuTE4C7
"Mistral AI bucks release trend by dropping torrent link to new open source LLM" — VentureBeat See the highlights of the story below! 1/10 🧵 https://t.co/ztfs1PUsV7
#AIRevolution: Mistral AI's Mixtral-8x7B Model Takes #SEO To New Levels - Explore Performance Metrics & Demos! #SearchEngineJournal https://t.co/V4RWQRjZ6I
Learn more about Mixtral-8x7B, the new model from @MistralAI, including performance metrics, four demos to try, and what #AI says about #SEO. https://t.co/HFrQZ9QaAc
Mixtral-8x7B outperforms llama-2-70B as per OpenCompass model evaluations https://t.co/nfoPBo55hU
Mistral `mistral-8x7b` on the @vercel AI SDK playground is now chat-tuned and it’s. so. delightful 😍 https://t.co/NvcQHQPs9l
🚀 Skip the wait for Google's Gemini and jump straight into action with @Mistral's Mixtral-8x7B. Released via torrent, no frills attached, it's ready for you to test drive in a vllm-powered interface. Experience the power of MoE AI without the wait: https://t.co/4JO9eM83lb
Mistral AI bucks release trend by dropping torrent link to new open source LLM https://t.co/6It0FsiVpu
We released our tuned Mixtral chat a few hours ago. Play with it through our app or API: https://t.co/qkiR9W526V. Big thanks to @MistralAI ‘s new addition of this MoE model. We are very excited about it.
Mistral 8x7B is now available in LangSmith Playground It uses the the implementation by fireworks AI team that reverse-engineered the architecture from the parameter names. This isn't an official implementation, as the model code hasn’t been released. https://t.co/XrExDMSzYq https://t.co/A4Y2q0Nwbr
You can try the new Mistral 8x7B model here https://t.co/pBctoKTSuV
You can now try @MistralAI mixtral-8x7b on the @Vercel AI Playground and use it with the AI SDK. (h/t @thefireworksai for the experimental implementation) Here's a video comparing it side-by-side to GPT-3.5-Turbo and Llama 2 70b Chat https://t.co/rEWptmjmQY https://t.co/QdC6Jj0oJU
Initial evals for Mistral MoE are out, and it is a solid 70B model that is very similar to GPT 3.5, Gemini Pro, and DeepSeek and slightly better than Llama2-70B. MMLU on the base models is at 0.717 compared to Gemin Pro's 0.718, DeepSeek's ~ 0.717, and GPT 3.5 at 0.7 On other… https://t.co/iCqGUVUTg9 https://t.co/7OrJEig9OL
If you want to try the new Mistral 8x7B model, you can do so here: https://t.co/hxqqzUzjef
Mistral AI's Unconventional Torrent-Based Release of MoE 8x7B LLM Shakes Up the AI Community #AI #AIcommunity #AItechnology #AndreessenHorowitz #artificialintelligence #EricJang #EUAIAct #Fundinground #Gemini #Google #GPT4 #JayScambler #languagemodel https://t.co/HInDfKmedc https://t.co/OHObHHZmTW