Skip to main content

The Open-Source AI Decision

Open-source vs proprietary AI models for enterprise. When open-source wins, when it does not, and how to make the decision with clear eyes.
8 December 2025·8 min read
Mak Khan
Mak Khan
Chief AI Officer
John Li
John Li
Chief Technology Officer
The open-source AI ecosystem has matured dramatically in 2025. Llama 3.1, Mistral Large, Qwen 2.5, DeepSeek. These models are genuinely capable, often approaching proprietary model performance for specific tasks. The question is no longer "is open-source AI good enough?" It is "for which tasks is open-source the right choice, and for which is it not?" That is a different, more useful question.

The Current State

As of late 2025, the open-source AI landscape looks like this:
General capability. The best open-source models (Llama 3.1 405B, Mistral Large) perform within 5-10% of GPT-4 and Claude on most general benchmarks. For many enterprise tasks, that gap is irrelevant.
Specialised capability. For specific tasks (code generation, structured extraction, classification), fine-tuned open-source models can match or exceed proprietary models because they are optimised for the exact task rather than general capability.
Small model efficiency. Smaller open-source models (7B-13B parameters) run on modest hardware and handle straightforward tasks (classification, extraction, formatting) with minimal quality tradeoff. These are the workhorses of cost-effective AI.
Tooling and ecosystem. The ecosystem around open-source AI (inference servers, fine-tuning tools, evaluation frameworks) has matured significantly. Running open-source models in production is no longer a research project.
61%
of enterprises are using or evaluating open-source AI models, up from 23% in 2024
Source: a16z, Open Source AI Survey, 2025

When Open-Source Wins

Data Sovereignty

If your data cannot leave your infrastructure, open-source is the only option. Proprietary models process data on provider servers. Open-source models run wherever you deploy them: your cloud, your data centre, your edge devices.
For NZ enterprises dealing with sensitive health data, government information, or Indigenous data covered by sovereignty frameworks, this is often the deciding factor. The model quality comparison is secondary to the data governance requirement.

Cost at Scale

At high volumes, open-source models are dramatically cheaper. A proprietary API call costs a fraction of a cent, but a million calls a day adds up. Self-hosted open-source models have a fixed infrastructure cost regardless of volume.
Mak's back-of-envelope calculation: for workloads exceeding roughly 50,000 inference requests per day on moderate complexity tasks, self-hosted open-source becomes cheaper than API-based proprietary models. Below that threshold, the operational overhead of self-hosting typically exceeds the cost savings.

Task-Specific Fine-Tuning

Open-source models can be fine-tuned for specific tasks. This is the single biggest technical advantage. A Llama model fine-tuned on your specific document types, your specific classification categories, or your specific extraction templates will outperform a general-purpose proprietary model on those exact tasks.
The investment in fine-tuning (data preparation, training, evaluation) is non-trivial. But for core, high-volume tasks, the performance improvement justifies it.

Vendor Independence

Proprietary AI means dependency on a specific provider. Their pricing changes, their model updates, their API changes, their terms of service changes all affect your business. Open-source models give you control.
This is a strategic consideration, not just a technical one. Organisations building AI as a core capability should think carefully about how much of that capability depends on a single vendor's decisions.

When Proprietary Wins

Frontier Capability

For tasks that require the absolute best available capability, proprietary models are still ahead. Complex reasoning, nuanced creative generation, multi-step planning, and tasks that benefit from very large context windows favour proprietary models.
If your use case requires the top 5% of model capability, proprietary is the right choice today. If it requires the top 20%, open-source is competitive.

Speed to Deployment

Self-hosting open-source models requires infrastructure: GPU servers, inference engines, model management, monitoring. An API call to a proprietary provider requires an API key. For teams that need to move fast, proprietary APIs eliminate an entire category of infrastructure work.
The time-to-value difference is significant. Weeks to months for self-hosted open-source versus hours to days for proprietary API. In a proof-of-concept or early exploration phase, this matters.

Operational Simplicity

Running AI models in production is operational work: uptime, scaling, monitoring, updates, security patching. Proprietary APIs outsource all of this to the provider. Self-hosted models require an operations team or at minimum an engineer who understands GPU infrastructure.
For organisations without existing ML operations capability, the operational overhead of self-hosting is a genuine barrier.

Rapid Model Improvement

Proprietary model providers update their models frequently, and improvements are automatic. Self-hosted models require manual updates: evaluating the new version, testing against your workloads, deploying, and monitoring. If you value always running the latest and best model, proprietary providers deliver this by default.

The Practical Decision Framework

John and I use a structured framework for advising clients:

Step 1: Data Governance

Can your data go to external providers? If no, open-source is the only option. Decision made.

Step 2: Volume Assessment

What is your expected inference volume? Below 10,000 requests per day, proprietary APIs are simpler and competitive on cost. Above 50,000, the economics shift towards self-hosted. Between 10,000 and 50,000, both are viable.

Step 3: Task Complexity

What capability level do your tasks require? Simple tasks (classification, extraction, formatting): open-source is fully competitive. Moderate tasks (summarisation, analysis, structured generation): open-source is competitive with fine-tuning. Complex tasks (reasoning, planning, creative generation): proprietary has an edge.

Step 4: Operational Readiness

Does your team have ML operations capability? If yes, self-hosting is viable. If no, proprietary APIs or managed open-source services (like Azure AI or AWS Bedrock hosting open-source models) are more realistic.

Step 5: Strategic Timeline

Are you building a core AI capability that you will run for years? Consider open-source for vendor independence. Are you experimenting with a use case that may or may not prove out? Use proprietary APIs for speed and simplicity.
The best enterprise AI architectures use both - open-source for high-volume workloads where you want control, proprietary for complex tasks where you want frontier capability. It is not a religion; it is engineering.
Mak Khan
Chief AI Officer

The Multi-Model Reality

The most effective enterprise AI architectures we see use both open-source and proprietary models. Small open-source models handle the high-volume, straightforward tasks. Larger proprietary models handle the complex, low-volume tasks. A routing layer directs each request to the appropriate model based on complexity, cost, and capability requirements.
This multi-model approach gives you the best of both worlds: cost efficiency and control where volume demands it, frontier capability where complexity demands it, and the flexibility to shift the balance as open-source models improve.
The open-source AI decision is not a binary choice. It is a portfolio allocation decision. And like all portfolio decisions, the right answer depends on your specific context, constraints, and objectives.