Mistral AI, a prominent European artificial intelligence startup, introduced a groundbreaking language model known as Mistral Small 3. This model boasts 24 billion parameters and achieves an impressive 81% accuracy on standard benchmarks while processing 150 tokens per second. What sets Mistral Small 3 apart is its ability to match the performance of models three times its size, all while significantly reducing computing costs. The company has made this model available under the Apache 2.0 license, allowing businesses to freely modify and deploy it.
In an exclusive interview with VentureBeat, Guillaume Lample, Mistral’s chief science officer, emphasized that Mistral Small 3 is one of the top models among those with less than 70 billion parameters. This model is on par with Meta’s Llama 3.3 70B, a model three times larger in size. The unveiling of Mistral Small 3 comes at a crucial time when the AI development costs are under intense scrutiny, especially after claims by Chinese startup DeepSeek regarding their competitive model training costs.
Mistral’s success with Mistral Small 3 can be attributed to their focus on efficiency rather than sheer scale. By implementing improved training techniques, Mistral achieved remarkable performance gains without the need for excessive computing power. The model was trained on 8 trillion tokens, a significantly lower number compared to similar models, showcasing its efficiency in AI capabilities.
Noteworthy is the fact that Mistral Small 3 was developed without reinforcement learning or synthetic training data, techniques commonly utilized by competitors. This approach helps in avoiding the embedding of unwanted biases that could be challenging to detect later on. The model’s target audience includes enterprises requiring on-premises deployment for privacy and reliability reasons, such as financial services, healthcare, and manufacturing companies.
As Mistral positions itself as Europe’s AI champion in the global market, the release of Mistral Small 3 signifies a shift towards open-source dominance. With a focus on smaller and more efficient models, Mistral’s strategy could prove to be pivotal as the AI industry evolves. By democratizing access to advanced AI capabilities and reducing computing infrastructure costs, Mistral is paving the way for accelerated adoption across industries.
In the upcoming weeks, Mistral plans to release additional models with enhanced reasoning capabilities, further solidifying their position in the AI landscape. As competition intensifies and efficiency gains emerge, Mistral’s commitment to optimizing smaller models could shape the future of AI development. Stay tuned for more updates on Mistral’s innovative advancements in the AI space.