Loading...
Groq Inc. has recently launched the Llama3 model, which is now running on their platform using the LPU inference engine. This new model, Llama3-70b, is noted for its high speed, reaching up to 825 trillion transactions per second (T/s). It is comparable in quality to GPT-4-turbo but operates at a faster rate. The Llama3 model is available in different configurations, including a 400 billion parameter version that also boasts impressive speeds.
OMG LLaMA 3 70b on @GroqInc https://t.co/cUJ0btYRWs
Llama 3 running at 825t/sec on Groq. 👀 https://t.co/Niq7vdL6KP
Llama3 is live on @GroqInc! https://t.co/nG1ApI7sHX
Llama 3 400b is gonna be running on groq faster than GPT-4-turbo and at the same level of quality 😳 https://t.co/gPRSEGMWLl
Llama 3 8b running at upwards of 800 T/s on groq 😲 https://t.co/W3I8Or3Uqb
Llama on Speed! 🦙💨 @GroqInc just started running @AIatMeta Llama3-70b; 🚀 wicked fast via the LPU inference engine. Great job 👏 @sundeep. A must try, visit https://t.co/mNqGIVdHnB https://t.co/PhMzjb6dsL
Llama 3 is on @GroqInc now 😄💥 https://t.co/Z8FkmKUKdd
Llama 3 in @GroqInc soon? Look the token speed/s 😳 https://t.co/w9wegzNmOl
Yes, Llama3 on @GroqInc coming.