Meta just released Llama 3. It's open source. It's genuinely competitive with Claude 3 and GPT-4 on most tasks. And if you work with firms that have data privacy requirements, compliance obligations, or just a philosophical preference for not feeding proprietary data to SaaS vendors, this changes everything.

I've been running Llama 3 through the same workflows I've been testing Claude and GPT-4 on. For professional services, the performance difference is smaller than I expected. The architecture difference is bigger than you need to understand. The strategic implication is massive.

What Llama 3 Actually Is

Meta's Llama has been open source for a while, but Llama 2 was good but not great—useful for research, not for production. Llama 3 changes that. Two sizes: 8B and 70B parameters. The 70B version is competitive with Claude 3 Sonnet on most benchmarks.

More importantly: you can run it yourself. On your infrastructure. Without API calls to a third party. Without your data leaving your network.

That matters if you're:

For those firms, Llama 3 isn't an alternative to GPT-4 or Claude. It's the option that was previously impossible.

The Technical Reality Check

Running Llama 3 yourself requires infrastructure. The 70B version needs a GPU with at least 80GB of memory. That's not off-the-shelf. It's either a serious on-premise investment (a few hundred thousand dollars for hardware and the team to maintain it) or a managed cloud service that costs approximately the same per-token as Claude 3 or GPT-4, except you're paying for compute instead of API calls.

The smaller 8B version runs on more modest hardware, but the quality drops noticeably for complex reasoning tasks.

So when I say "you can run it yourself," I mean: technically true. Practically, most firms will use a managed service to run Llama 3, which brings us back to API calls, just with a different vendor.

The difference: you can switch vendors at any time. You're not locked in to Meta or Anyscale or Together AI. Llama 3 runs anywhere. Proprietary models do not.

Where Llama 3 Wins

Vendor independence: Build on Llama 3, and you're not dependent on any single company's business strategy. If you get unhappy with pricing or service, you move to a different Llama provider.

Compliance: For firms with strict data handling requirements, "the model runs on our infrastructure" is often worth more than the marginal difference in quality.

Cost, at scale: If you're processing millions of tokens per month, the economics of running Llama 3 on your own infrastructure can beat API pricing. Not for small firms. For larger ones, maybe.

Where Llama 3 Has Gaps

Quality on complex reasoning: Claude 3 Opus is still noticeably better on complex legal analysis, multi-step reasoning, and edge cases. Llama 3 is good. It's not better.

Specialization: GPT-4 has had more time in production, more fine-tuning, more optimization for specific tasks. Llama 3 is catching up fast, but it's not there yet.

Ecosystem: More libraries, integrations, and support around GPT-4 and Claude. Llama is growing that ecosystem, but it's behind.

The Strategy Question

For most professional services firms, this is straightforward: use Claude 3 Sonnet or GPT-4 for your production workflows. They work. They're cost-effective. They're reliable.

But if you're a firm where:

...then run a parallel pilot on Llama 3. Test it on the same workflows you're using Claude for. Measure quality and cost. Decide if the trade-offs make sense for your firm.

The Honest Take

Open source AI is maturing fast. In 18 months, open source might be as good as proprietary. In three years, it might be better. The betting strategy is: use proprietary models now because they're best, but don't bet the farm on them. Have a Plan B that uses open source. Llama 3 is a credible Plan B.

For firms that need on-premise AI or have strict vendor independence requirements, Llama 3 is not a Plan B. It's your primary option now. And it's good enough.

Want to discuss AI strategy for your firm?

Book a free 30-minute assessment — no pitch, just practical insights.

Book a Call