Data Moats Aren't Enough: The Build-vs-Buy Trap in the Age of AI Agents cover image

The allure of building proprietary technology is strong, especially when fueled by visions of impregnable data moats. But in the era of rapidly evolving AI agents, this mindset is increasingly dangerous. I believe the default for most companies should be to *buy*, or assemble, rather than build, focusing internal efforts on areas that create true, defensible differentiation – and that’s likely a much narrower scope than you think.

The Siren Song of the Data Moat

For years, the mantra was clear: collect as much data as possible, and you'll build an insurmountable competitive advantage. This fueled a build-centric approach. Companies reasoned: 'Our data is unique; off-the-shelf solutions won't cut it; we need to build our own algorithms, our own infrastructure.' This was often true, to a point. For example, early recommendation engines at Netflix or search algorithms at Google had to be custom-built. There simply weren't viable alternatives. And the resulting data advantage did translate into real market power. But the landscape has shifted dramatically.

The rise of powerful, general-purpose AI models and agentic workflows changes the game. These models are pre-trained on vast datasets, often exceeding what any single company could realistically amass. And crucially, they can be fine-tuned on relatively small, targeted datasets to achieve impressive performance. OpenAI's Agents SDK, for example, is designed to facilitate the creation of such workflows [3]. The implication? Your 'unique' data may no longer be as unique – or as valuable – as you thought. It's now a commodity input, not the source of a moat.

Consider the hypothetical example of a regional grocery chain building its own demand forecasting system. Five years ago, this might have been a reasonable decision. Today? It's almost certainly a mistake. Plug into a platform like Amazon Bedrock, leverage a pre-trained model, fine-tune it on your historical sales data, and connect it to your inventory management system via AWS Interconnect [4]. You'll get 80% of the performance at 20% of the cost, freeing up your engineering team to focus on truly differentiating initiatives, like optimizing the in-store customer experience or developing innovative delivery options.

The Cost of Perpetual Building

Building everything in-house isn't just slower; it's also significantly more expensive. And the costs are often hidden. It’s easy to track direct expenses, like salaries and cloud infrastructure. But what about the opportunity cost of diverting resources from core business initiatives? What about the technical debt that accumulates as you maintain and upgrade your custom-built systems? What about the difficulty of attracting and retaining top-tier engineering talent when they're stuck maintaining legacy code instead of working on cutting-edge problems?

NVIDIA's recent announcement focusing on 'cost per token' in AI factories highlights this issue [1]. Building your own AI infrastructure from scratch, even with access to powerful GPUs, can lead to significantly higher operational costs than leveraging existing cloud services or specialized AI platforms. The focus needs to shift from raw compute power to efficient resource utilization and optimized model deployment.

I've seen companies paralyzed by the 'not invented here' syndrome, stubbornly clinging to their custom-built solutions even when superior alternatives exist. They spend years chasing marginal improvements, while competitors who embraced off-the-shelf solutions leap ahead. This is a particularly dangerous trap in the age of AI, where the pace of innovation is accelerating exponentially. Yesterday’s competitive advantage is tomorrow’s liability.

When Building Still Makes Sense: The Core IP Exception

This isn't a blanket condemnation of building. There are still scenarios where it's the right choice. The key is to focus on areas where you can create true, defensible intellectual property – and where that IP directly contributes to your competitive advantage.

Consider a company developing a novel drug discovery platform. In this case, building proprietary AI models to analyze complex biological data might be essential. The algorithms themselves become a core asset, protected by patents and trade secrets. Or, take Anduril Industries, the defense technology company. They don't just buy off-the-shelf drones and sensors; they build their own, tightly integrating them with their Lattice AI platform. This vertical integration allows them to create a highly differentiated product that's difficult for competitors to replicate.

The common thread in these examples is that the custom-built technology isn't just a nice-to-have; it's the *foundation* of the company's competitive advantage. It's something that can't be easily replicated by competitors using off-the-shelf solutions. It's something that creates a lasting strategic lock-in. This level of differentiation must be present to justify the build approach.

Conversely, if you are a company that is relying on open-source models or fine-tuning on existing models, your competitive advantage will not be in the model itself, but in how well you deploy the model or the data you can use for fine-tuning. The model itself is not defensible, but the unique dataset and its application are where the defensibility will come from. Similarly, new features in Adobe Premiere benefit from leveraging GPUs [2], making buying hardware like NVIDIA's, and not necessarily building your own hardware, a defensible position.

The Contrarian Claim: Building Data Pipelines, Not Data Moats

Here's my contrarian claim: The future isn't about building data moats; it's about building *data pipelines*. Instead of hoarding data and trying to build impenetrable fortresses, companies should focus on creating flexible, adaptable pipelines that can ingest, process, and leverage data from a variety of sources. This includes internal data, public datasets, and data provided by third-party partners. The value isn't in owning the data itself, but in the ability to rapidly integrate and analyze it to derive actionable insights.

This requires a different mindset and a different set of skills. It requires investing in robust data infrastructure, developing strong data governance policies, and fostering a culture of data literacy throughout the organization. It also requires embracing open-source tools and cloud-based services that can streamline the data integration process.

Think of Snowflake or Databricks. Their value proposition isn't that they own unique data; it's that they provide a platform for seamlessly integrating and analyzing data from diverse sources. They enable companies to build data pipelines, not data moats. This is the future of data strategy.

The Agentic Shift: From Data to Action

The rise of AI agents further amplifies the importance of data pipelines. These agents are designed to automate complex tasks and workflows, often requiring access to real-time data from multiple sources. Building custom AI agents from scratch is rarely necessary; the focus should be on configuring and integrating them into existing business processes. Companies like Cloudflare are already partnering with OpenAI to power agentic workflows in their Agent Cloud [7], demonstrating the growing importance of this trend.

This means building robust APIs, creating clear data schemas, and implementing rigorous security protocols. It also means fostering a culture of experimentation and iteration, allowing employees to rapidly test and deploy new AI-powered solutions. The goal isn't to build a perfect system from the outset, but to create a flexible, adaptable platform that can evolve alongside the rapidly changing AI landscape. This concept also applies to cybersecurity; focusing on trusted access allows companies to better defend their systems [5].

The New Build-vs-Buy Framework: Ask the Hard Questions

Before embarking on any new technology project, ask yourself these hard questions:

Answering these questions honestly will help you make informed decisions about when to build and when to buy. And in the age of AI agents, that decision could be the difference between success and failure.

A Prediction: The Rise of the 'AI Integrator'

My prediction: We'll see the emergence of a new class of companies – the 'AI Integrator.' These companies won't build their own AI models or infrastructure; they'll specialize in helping other companies integrate AI into their existing business processes. They'll be experts in configuring AI agents, building data pipelines, and ensuring data security. They'll be the plumbers of the AI revolution, connecting the various pieces and ensuring that everything works seamlessly together.

This is where the real value will lie. Not in building the individual components, but in orchestrating them effectively. The companies that embrace this approach will be the winners in the age of AI agents. So, focus on building pipelines, not moats. It’s time to rethink your build-vs-buy strategy – before it’s too late.

Sources

Related Resources

Use these practical resources to move from insight to execution.

Content Notice: This article was created with AI assistance and reviewed for quality. It is intended for informational purposes only and should not be treated as professional advice. We encourage readers to verify claims independently.

Building the Future of Retail?

Junagal partners with operator-founders to build enduring technology businesses.

Start a Conversation

Try Practical Tools

Use our calculators and frameworks to model ROI, unit economics, and execution priorities.