Open Source Will Eat Vertical AI

Every few weeks, another vertical AI startup announces a new mega-round: “$100M to transform [industry] with AI”. The speed at which these startups race through funding rounds, and the sheer scale, is dizzying. Beneath it all sits a question few seem to ask: what is this money actually buying?

The answer, upon closer inspection, is startlingly simple. It's buying go-to-market. Not technology. Not data moats. Not network effects. GTM. Field engineers, sales teams, customer success reps, and the entire apparatus of making enterprise buyers comfortable enough to sign.

Yet, in a world of abundant code and mature procurement, funding sales forces to push undifferentiated software through low friction channels is the entirely wrong thing to do. The model is broken. Vertical software distribution, AI included, is due for an overhaul.

Open source software has always targeted developers. Now, with the abundance of AI generated code, the promise of open source software targeting business users can finally be fulfilled. Keeping software closed source was always an unnecessary constraint on the customer, a relic of the value capture mechanisms of the past. Imagine instead, a world, where anyone can customize an open code base for specific industries using nothing more than a coding agent and their own expertise.

Winner-take-all requires a winner

The software industry spent three decades producing winner-take-all outcomes. The entire VC model depends on it. What if AI applications simply don't generate them?

Every previous technological cycle came with its own set of startup challenges, and venture was the right tool to combat them: winning new surfaces like the browser, rewiring procurement, proving reliability long before the market bought in. SaaS was disruptive to the business model of software. Cloud was foreign, requiring enterprises to abandon the comfort of their own data centers. Mobile demanded entirely new interaction paradigms. These cycles were contrarian, misunderstood, or structurally disruptive. Venture capital thrived because incumbents were slow to respond and startups had the window to build compounding advantages before anyone else showed up.

AI is none of these things. It works. Every CEO wants to adopt it now. For the first time in decades, startups and incumbents are chasing the same territory with the same urgency. It's an execution race. And if all that is being funded is a go-to-market apparatus, the advantage evaporates the moment a well-run competitor watches the first mover navigate the product maze, and then walks the cleared path at half the cost.

Risk capital, to be justified, has to offer asymmetric returns. To paraphrase Howard Marks: for great returns it's not enough to be right, everyone else has to be wrong. In AI, everyone is right. So even in the best case, the results of these investments will be average. Average is not how venture funds work.

The product isn't a product

The cost of building software is collapsing toward zero. This single fact restructures everything about product competition in AI. When anyone can build anything, product advantages don't compound. Without network effects or deep operational complexity, all you are doing as a first mover is charting the path for everyone behind you.

AI application companies positioned themselves as middleware: sitting between foundation models and enterprise customers, filling gaps that raw models couldn't handle. Limited context windows, hallucinations, no access to private data. Early movers like Harvey, Hebbia, and Legora deserve credit for identifying and solving them. But these gaps are closing with every model generation and labs are now delivering comparable experiences with far less boilerplate.

Speed of execution is not a moat. Being first to market, as Thiel would put it, means you are paid to make the expensive mistakes that help everyone else. Unless you hold the last technological advantage in a chain, you are perpetually exposed to whoever builds next. And in AI, "next" arrives every few months.

Scale doesn't save you either. In a world where generating code costs nearly nothing, it becomes viable to build narrowly scoped, domain-expert agents that outperform general-purpose tools on every metric that matters. A specialized regulatory compliance agent will always beat a do-everything legal AI for the customers who care most. 

Even when you do capture a market, there is no pricing power. The knowledge that your software can be replicated shifts leverage to the buyer. They don't need to actually switch. 

It's the distribution, stupid

So the product isn't durable. What about distribution?

Distribution has always been the decisive advantage in vertical software: the company that pushes through procurement barriers, in a category where innovation is exhausted and the channel carries friction, occupies a position that is nearly impossible to dislodge. This is last mover's advantage.

Vertical AI apps are running the same playbook, but neither condition holds. AI is iterating so fast that no one has reached the last innovation in any category. At the same time, the procurement channel has lost its friction entirely. Software ate the world and Covid compressed another decade of digitization into two years. Today, nearly every organization knows how to buy software now. The path is clean. Friction was not just an obstacle for the seller. It was the barrier that kept competitors at bay.

Yet the industry hasn’t adapted. The Field Deployment Engineer model, popularized by Palantir, is now the default sales motion across AI startups. But Palantir operates where friction is abundant: government clearances, compliance, and relationship depth that takes years to accumulate. Eight-figure contracts justify that cost. Most AI startups sell $200K ARR deals into enterprises where the procurement path is wide open.

In a desperate attempt to manufacture winner-take-all effects where none exist organically, the industry has embraced kingmaking: flood a chosen player with capital, install it as the category leader, and let the self-fulfilling prophecy do its work. Once a market is "won" with costly GTM, margins are so thin the company can never become self-sustaining. It expands into adjacent markets, where the war of attrition continues against other well-funded players doing the same thing.

The valuation expectations attached to venture capital assume distribution friction creates lasting advantages. The capital being injected is not funding durable assets. It is funding temporary demand. The business may be real, but the returns won't be. And the need to justify those valuations is what makes these companies fragile: it forces growth at all costs, which forces unprofitable sales motions, which forces the next round, which raises the bar even higher.

Distribution can still be a durable advantage, but it needs to be more capital efficient and hard to replicate.

The counterarguments 

Some will say go-to-market and product are inseparable and that FDEs are really de facto product managers that embed with customers, extract tacit knowledge, and feed those insights back into the software over time. Yet, any feature that proves critical for a particular customer is trivial to copy. The real question is not whether FDEs contribute to product development, but whether that contribution translates into pricing power. It doesn't. Customers that know what they want can increasingly get those modifications built directly with AI coding agents.

Others point to switching costs. Once an enterprise has spent months implementing one vendor's system, trained workflows around its quirks, and built institutional knowledge about its failure modes, the appetite for starting over dwindles. But the calculus is shifting. Many buyers already view AI products as simultaneously expensive and easy to replace. As costs continue falling, the ROI equation for building and customizing internally only becomes more attractive. Switching costs matter less when the alternative is not switching to a competitor but building your own.

The outcome-based pricing thesis is the strongest case: land with a customer, iterate on their specific workflows, then transition to pricing tied to measurable results. If executed well,, this is genuinely disruptive. But the implication is non-trivial. Outcome-based pricing means assuming responsibility for a function that was previously strategic to the buyer. At that point, one is no longer just a vendor, but a direct competitor to the customer.Any provided solution must be costly enough to justify outsourcing while remaining sufficiently non-strategic that the buyer is comfortable handing it over. Navigating this balance at scale will prove challenging.

A new distribution engine

Competing on GTM without matching FDE spend is like trying to ride clean during the 2000s Tour de France. The question is not how to outspend them. It's how to build a distribution architecture that makes their spending irrelevant.

The answer is to rethink how vertical AI software is sold entirely. Release the core product as open source and let customers adopt it on their own terms. Instead of sending FDEs on-site, give customers the ability to open private issues on a repository, have their own coding agents assemble the changes, and rely on an automated process to merge and maintain those contributions upstream. In this model, the product itself is free. The value captured is the ecosystem that is formed around it: agent-first documentation, a contribution model that makes upstream merges seamless, and domain-specific system integrators who are part prompt engineers, part consultants.

The code is published openly for the world to see. Anyone, including competitors, can fork it. This deliberately burns the bridges that proprietary software companies rely on for defensibility. That is how the last mover wins: make the product free, make switching costs zero, and compete on a different dimension, building the best ecosystem for modification, extensions, and domain-specific customization.

A new breed of system integrators emerges in this model. Not the Accentures and Deloittes of the old world, but prompt engineers and AI-native consultants who customize an open source system for specific industries, workflows, and regulatory environments. They make money for themselves by building on top of an existing platform. Think . Salesforce didn't just sell software, it built an ecosystem where thousands of independent builders made a living extending the core product. The same dynamic applies here, except the builders are armed with coding agents, and the platform is open.

Open source business software is not a new idea. The conventional wisdom was sound: enterprises wanted customization but lacked the engineering capacity to maintain forked software, and open source vendors couldn't capture enough value to sustain themselves. But that was a world without abundant AI. Coding agents are the enabling technology that changes the equation. They collapse the customization barrier that keeps enterprises locked into proprietary SaaS.

What makes this model structurally durable is that it is aligned with what customers actually want. Enterprises don’t want to be locked into a vendor’s roadmap or pay for features they’ll never use. They want software that solves a specific problem, can be modified as needs change, and never holds them hostage. The open source model gives them exactly this and aligns with the customer’s interests. Every SaaS company that charges per seat for a product the customer could plausibly build internally is swimming against the current. Take the world as it is: software is abundant, customization is cheap, and buyers have leverage. Build a business that thrives under those conditions rather than one that depends on pretending they don't exist.

Meanwhile, the incumbents are funding your customer education. Every dollar spent on FDEs, onboarding, and enterprise handholding is a dollar spent teaching the market what AI can do, how to evaluate it, and how to integrate it. Being first to market means you are paid to educate customers for everyone who follows.

The entire VC thesis depends on one player capturing a market and extracting monopoly rents. That model assumes proprietary software, friction-laden channels, and winner-take-all dynamics. Without these assumptions the funding logic collapses. And the world is removing them whether the industry likes it or not. Software costs are falling toward zero. Procurement friction has evaporated. AI coding agents are making customization trivial. The founders who recognize this early and build for it will define the next decade of vertical software.





ChatGPT is the new browser and memory is the new cookie

OpenAI launched an Apps SDK on Monday (see here). This effectively means that developers can now build apps inside ChatGPT:
A smartphone screen displaying the ChatGPT interface with a Bookingcom search result Two hotel images are visible one showing a modern indoor lobby with plants and seating the other depicting an outdoor Parisian scene with a bridge and river Text overlays include ChatGPT at the top time 1130 and hotel details like names ratings and prices in USD

This was an obvious play by OpenAI and it was only a matter of time until more dynamic experiences made it into chat. After all its very difficult to do some of the most profitable internet activities (ie ecommerece, gaming etc) with text only. The writing was on the wall here ever since MCP-UI came out and Shopify started incorporating it earlier this year:

Adding apps within the chat experience does 2 things:
1. it helps preserve the user within ChatGPT for longer
2. it gives OpenAI a significant amount of context on user behavior

From a strategic point of view, 2 is incredibly valuable. Users would traditionally leave ChatGPT and go to booking.com once they generated their trip plan. That prevented OpenAI from learning important information about the user - did they book the hotel that was presented to them? did they decide to do all of the suggested activities? This is personalization context that can help OpenAI be more useful to the user in the future.

It wasn't too long ago that ChatGPT gained the ability to reference prior conversations. That particular feature is now commonly referred to as memory. Memory is the personalization layer that helps ChatGPT answer questions better for you in the future, and in a world where model capabilities across labs are converging more and more, what prevents users from switching LLMs is the huge cost of moving their "memories". 

Once ChatGPT remembers the hotel that you like to stay in, the type of food that you order on weekends and the type books you like to read before bed, it is very hard to move away without losing a huge amount of personalization. As far as OpenAI is concerned, the more activities the user can do within the chat app, the harder it becomes to leave. 

In the same way that cookies allowed Google to become a default transaction layer for the internet, memory helps OpenAI own the personalization layer. In a world where Google owns the browser and Apple owns the device, it is very hard to stand out as the aggregator of personalization information across multiple applications - and yet that is exactly what memory + apps allows OpenAI to do, despite not owning the device or the browser. It will be interesting to see if Apple will allow ChatGPT "apps", because it technically violates their mega-app policy. 

The timing here is perhaps notable. There is no real advantage in being a first mover. As the MCP-UI experience shows, there are a lot of hard problems to figure out and there is a big chance that some of this blows up in OpenAI's face. Google will inevitably add this capability to Gemini and it has all the right relationships to do so. Launching early doesn't help OpenAI tremendously, so why now ? 

Perhaps one of the most obvious reasons to get going here is that OpenAI's devices efforts are very real and the sooner they can stand up an ecosystem (that spans all the apps customers already use today), the sooner they can have a viable standalone device.

Notably, there are more platform plays coming to ChatGPT soon. Sam Altman brought up "Sign In With ChatGPT" in May:

What better way to ensure that you have context on your users than allowing them to bring their OpenAI identity with them everywhere they go? 

So..there it is. The inklings of a new platform emerging. Whether app publishers will be willing to do this is a totally different question. Giving your app away in this new modality has its risks and seeing how OpenAI is trying to do everything everywhere I would be cautious to move in this direction. Ultimately, users will decide for them. If chat is the way people browse the internet from now on, app publishers won't have a choice.







Search, not AI, is still king

Search, not AI, is still the foundation of knowledge work. 

The Google search box was not just a necessary precursor to Artificial Intelligence, good search already looks a lot like it. Take the question “what is the boiling point of alcohol?”. AI can work through the physics and derive the answer, or it can just…look it up.

In a perfect world, where every question has a ready answer, you search. In the real world, some answers exist and others have to be worked out. Or, you let AI do the thinking.

Thinking, rather than searching is what can cause bad AI outputs (or hallucinations). A published source is usually more reliable than an answer an LLM derives from scratch. At worst, retrieval might miss a relevant document, but derivation can fabricate one.

This is clearest in serious knowledge work, where correctness is critical and answers take multiple reasoning steps. Small mistakes compound. The more “thinking” AI does, the higher the risk of hallucination.

The best way to lower this risk is to lean harder on search and less on derivation. Reasoning was never the problem. That’s the fun part of the job. 

Yet in most AI chatbots, reasoning and search are fused into one black box that answers questions instead of helping users do sensemaking. Wouldn’t it make more sense for search and AI generation to be separate?

The search for reasoning “context” can’t look like a Google search box, where you type in an intent and then groom through results. It should look more like a loop of search, review, synthesize, search again. It is more of a recipe than a search goal. It’s search that lets you specify how to search, not what to search.

Search, taken to an extreme, can look like an ETL loop that mines data for signal. Letting users define the search strategy in detail creates a much better leverage point for human-machine interaction. It keeps the fact foraging deterministic, and lets AI reasoning do what it does best. 

Bigger models help, but they’re not the whole story. Reliability in AI will come from these type of boring UX changes that make everything more deterministic.

Innovation and survival

These days innovation is commonly thought of as the job of the small Sillicon Valley startups that have no respect for the status quo. The "Lean Startup" has become the de-facto rule-of-the-land and every enterpreneur has mention "leanness" on their slide decks to be taken seriously. To innovate, you need to be fast, take creative risk and iterate.

Yet real veterans of the craft, like Amazon, prove that innovation has nothing to do with being lean or small or iteration. Amazon has repeatedly entered new markets and out-innovated incumbents, seemingly violating every rule of the "playbook".

The phrase “innovation playbook” sounds like a contradiction to begin with. If innovation were a defined process, wouldn't everyone "innovate"? In practice, it's closer to a decision-making framework, best understood in contrast to alternatives.

Sillicon Valley thought leadership has gotten innovation completely wrong. The “Lean” framework is, at its core, a way to operate under extreme constraints. Amazon, as I will show below based on my personal experience, has a much richer understanding of what it really takes to do it. 

The contrast

Sillicon Valley preaches: “build an MVP, ship fast, measure and validate, iterate, stay efficient.” Amazon aims for a Minimum Lovable Product. Where others “measure and validate,” Amazon “dives deep” and “invents and simplifies.” “Customer obsession" often comes at the expense of efficiency. It’s not a deliberate rebuke to Valley dogma, but it does invert the usual iterative paradigm.

1) MLP vs. MVP

A groundbreaking product can't be beta-tested into existence, it has to start with an uncompromising vision. To illustrate the point,  MVP is a tactic for sharpening focus in a resource constrained world, while MLP is all about focusing on a great user experience. Two very different decision-making frameworks with serious ramifications.

When the first Kindle was built, Jeff Bezos insisted on cellular connectivity even though the default pattern was syncing through a PC. An MVP-focused team might have cut wireless to ship on time. Wireless was the riskiest part, and much of the team saw it as nonessential to basic viability. Bezos kept it because the “magic moment” depended on it.

2) Process vs Outcome

As anyone who’s worked on a complex project knows, ambiguity shows up no matter how crisp the vision. Amazon equips its people with tools that reward analysis and judgment, rather than mindless execution.

Instead of “ship quick and fail fast,” Amazonians “insist on the highest standards.” Instead of “incorporate user feedback,” Amazon “invents and simplifies.” Instead of “iterating” it “dives deep.” 

Intellectual rigor is the point, and it’s enforced through a writing-first process best exemplified by the PRFAQ—a ritual that can feel brutally nitpicky. A good PRFAQ forces clarity of thought. Ideas get shaken around until they land as a cohesive whole. That same writing culture shows up in technical designs, security reviews, and operational readiness. The bar is incredibly high.

3) Speed vs. quality

Quality is non-negotiable, but it doesn’t have to slow execution. The one-way door vs. two-way door framework keeps decisions fast when they’re reversible and careful when they aren’t. “Disagree and commit” turns debate into forward motion.

Teams inside Amazon still run segments, VOC, user studies, alphas/betas, and A/B tests. They’re tools for refinement, not for steering the product away from its vision. A/B tests pick the button that converts better; CSAT confirms the experience worked as intended. Occasionally data surfaces an outlier that deserves a strong response, but the vision doesn’t swing with every experiment. Product-market fit is not "found", it's forged. As Jeff Bezos likes to point out, when the data and the anecdotes disagree, trust the anecdotes.

4) Nimble at scale

A common theme in growing organizations is better communication and code re-use across teams. The Amazon approach is counterintuitive: in the short term, even teams building similar things should often duplicate. Coordination costs and external interfaces drag more than duplication does. It's important to realize that duplication is expensive. Yet, the investment is fully worth it. Costs can be consolidated later, organizational and technical inefficiencies can be fixed, but disappointed customers rarely come back.

What does this all mean? 

One striking conclusion is that this kind of innovation is expensive. Many of the tools above spend money to buy clarity, quality, and speed. The question is: can you innovate like this as a resource-constrained startup? A company chasing the next milestone ahead of a funding round rarely has the slack to apply this level of rigor and dedication to the vision. 

What is absolutely true about startups—and what justifies the “Lean” framework—is that they operate under extreme information asymmetry. Startups tend to proliferate around technology cycles that enable new markets. New markets come with information gaps that you can only close by putting product in front of users and learning from their behavior. In that model, speed and cost efficiency are essential.

But that’s the key point: lean is optimized for discovery under constraints, not for invention under conviction. Lean assumes the primary risk is building the wrong thing, versus knowing what to build and doing it the right way. Use the right framework for the right job.