Mistral Small 3 is the most efficient and versatile model of Mistral. Pre-trained and instructed version, Apache 2.0, 24B, 81% MMLU, 150 token/s. No synthetic data so great base for anything reasoning.
Hey everyone! 👋
Check out Mistral Small 3 – it's setting a new benchmark for "small" LLMs (under 70B)! 🚀 This 24B parameter model from Mistral AI offers performance comparable to much larger models, but with a focus on efficiency.
So here's the key features:
· Powerful & Efficient: State-of-the-art results with low latency (150 tokens/s).
· Locally Deployable: Runs on a single RTX 4090 or a 32GB RAM MacBook (once quantized).
· Knowledge-Dense: Packs a lot of knowledge into a compact size.
· Versatile: Great for fast conversational agents, low-latency function calling, creating subject matter experts (via fine-tuning), and local inference (for privacy).
It's also open-source under the Apache 2.0 License!
Oh that's really cool, congratulations on the release! Small open models are where I feel a lot of the really cool applications are at! Will check this out. ☺️
Okay, Mistral Small 3 looks like a game-changer! The compact design with all that power is impressive. Can’t wait to see how it performs in different setups. Excited to see what’s next!
Impressive work on Mistral Small 3! The balance of efficiency, versatility, and no synthetic data makes it a solid foundation for so many applications. Excited to see how it evolves!
Oh, that's seriously cool! Big congratulations on the release. I truly believe that small open models are the hotspots for a ton of amazing applications. I'm definitely gonna check this out. It sounds super promising.