Mistral (the French AI company) released Mixtral, an open source model that's competitive with GPT-3.5 on most benchmarks and runs on consumer hardware.
This is significant because it represents a real open source alternative to proprietary models. And it changes the calculus for firms thinking about AI infrastructure.
What Mixtral Does
Mixtral is a "mixture of experts" model, which means it selectively activates different parts of the model depending on what it's asked to do. This makes it efficient and fast.
It's competitive with GPT-3.5 (not GPT-4), but it's open source and can run on your own servers. No API fees. No data going to third parties. Complete control.
Why This Matters Now
For Budget-Conscious Firms: If you have infrastructure and technical talent, Mixtral lets you do AI work without paying OpenAI per-token charges. The math works if you're doing high-volume AI processing.
For Privacy-Focused Firms: You can run Mixtral on your own servers, fully under your control, with zero data leaving your network. For healthcare or finance firms with privacy concerns, this is huge.
For Firms Wanting Customization: Mixtral's weights are open, which means you can fine-tune it on your firm's data. You could train it on your precedents, policies, or workflows and get a model that's specific to how you work.
The Realistic Assessment
Mixtral is good, but it's not GPT-4. For high-precision work (legal analysis, complex research), you still want a more capable model. For routine work (email triage, document categorization, initial drafting), Mixtral is perfect.
The infrastructure cost to run it is modest—maybe $1000-$2000 per month in cloud compute for a mid-sized firm. That breaks even quickly if you're doing significant AI processing.
But you need someone to manage it. Running open source models isn't set-and-forget. You need infrastructure expertise, monitoring, updates, and maintenance.
The Hybrid Strategy
The smart move for most firms is hybrid:
- Use Mixtral (or similar open source models) for high-volume, routine work that doesn't require maximum capability.
- Use OpenAI or Anthropic's models through APIs for work that requires more capability.
- This gives you cost control, data privacy where you need it, and maximum capability where it matters.
The Bigger Trend
Mixtral is part of a bigger trend: open source AI is getting better fast. Llama (Meta), Mistral, Stable Diffusion (images)—these are closing the gap with proprietary models.
I expect that in 12-18 months, open source models will be capable enough for 70-80% of professional services work. That changes the vendor dynamics entirely.
Instead of "pick a proprietary API vendor," it becomes "use open source for routine work, pick a proprietary vendor for specialized work."
What to Do Now
You don't need to deploy Mixtral today. But you should understand it. When you're evaluating your AI strategy, understand that open source options exist and are improving.
If you have infrastructure talent on staff, run a pilot with Mixtral on one of your routine AI workflows. See if the quality is acceptable for your use case. If it is, the economics become much better.
And when vendors try to lock you in with claims that "only our model can do what you need," remember that open source alternatives are advancing fast.
Vendor optionality is coming. Mixtral is one sign of it.
Want to discuss AI strategy for your firm?
Book a free 30-minute assessment — no pitch, just practical insights.
Book a Call