As frontier models converge for a growing class of work, raw AI capability explains less of the difference between products. That is a narrower claim than it sounds. Model quality still matters. Accuracy matters. Latency matters. Long-context reasoning, code generation, multimodal performance, tool use, and safety behavior can decide whether a product works at all. But for many teams, the strategic question is no longer, “Can we get access to capable AI?” It is, “Can we turn that capability into a workflow users care about, trust, repeat, and eventually cannot replace?”
That is where product sense stops being a soft skill. In the AI era, product sense becomes product judgment: the operational judgment that decides where intelligence belongs, what context it needs, which data should enter the system, when humans stay accountable, and what durable asset the product compounds over time. AI strategy is expanding beyond model access into last-mile judgment, not moving away from models.
The feature that does not change the work
Picture a product team with access to the same frontier models as its competitors. The team ships an AI assistant into an existing B2B product. It can summarize records, answer questions, draft emails, and generate reports. The demo looks good. Users try it once. Then they go back to the old workflow.
The model did not fail. The product never changed the work. It sat beside the workflow instead of inside it. It did not know the user’s permissions, the current customer state, the exception history, the source-of-truth data, or the moment when a decision had to be made. It created another surface to check. This is a common failure mode when AI features are deployed without redesigning the work: impressive capability, weak product judgment.
A better team asks a different question. Not “Where can we add AI?” but “What recurring workflow is expensive, frequent, painful, or strategically important enough that intelligence should live there?” This shift matters because AI now touches nearly every part of the product lifecycle: discovery, design, prototyping, coding, testing, analytics, feedback, competitive analysis, sales, marketing, and operations. Open-weight models are improving. Enterprises increasingly use multiple models by use case. In a 2025 a16z CIO survey, 37% of respondents reported using five or more models in production. That statistic does not prove models are interchangeable. It does show that enterprise AI stacks are becoming multi-model, routed, and use-case specific. In that world, the ability to orchestrate models around proprietary context, workflow constraints, and measurable outcomes becomes a distinct source of value. The model becomes a component. The moat moves to the system around it.
Product sense is not taste
People often reduce product sense to taste: good UX instinct, customer empathy, prioritization, and a feel for what should be built. Those still matter. But AI stretches product sense into something more operational.
In an AI product, product judgment means deciding which workflows deserve intelligence, which context improves or degrades the answer, when the AI should act or stay silent, where human verification must remain, and which part of the value chain the company must own. This is why context has become a product decision, not just an engineering concern. Anthropic’s work on context engineering frames the problem clearly: agents perform better when systems curate what enters memory, what gets retrieved, what gets summarized, and what stays out. That is product judgment.
More context is not automatically better. A financial AI assistant that remembers every conversation may feel powerful until it uses stale preferences, exposes sensitive information, or treats a passing comment as a permanent instruction. The product has to decide what should be remembered, forgotten, hidden, surfaced, and audited. The same logic shows up in Notion’s AI strategy. A generic chatbot can answer questions. An agent with workspace context can help because it understands the documents, projects, owners, permissions, and history inside the user’s real environment. The value comes from connected context, not from the chat box.
Claude in Excel is another useful example. The opportunity is not “AI spreadsheet chat.” The opportunity is intelligence inside a painful, high-frequency workflow. A user is already modeling, reconciling, checking formulas, cleaning data, and preparing analysis in Excel. Asking that user to move into a separate AI interface adds friction. Meeting the user inside the spreadsheet changes the work. This is the product-judgment move: find the work as it really happens, then decide where intelligence reduces effort or improves judgment.
Data alone is not the moat
The easy counterargument is that product judgment matters less than proprietary data. Sometimes that is true. Incumbents with decades of enterprise data, distribution, compliance trust, and system-of-record status can have advantages a sharper startup cannot overcome by taste alone. In commerce, Amazon and Google have behavioral and advertising data that can shape agentic shopping experiences. In vertical AI, domain-specific interaction data and evals can make a smaller specialized system outperform a general one.
But “we have data” is not a strategy. Data creates advantage only when it is relevant, fresh, governed, accessible at the right moment, and tied to a feedback loop that improves the product. Raw data ownership does not help if the product cannot decide which data matters, how to retrieve it, how to protect it, and how to convert it into a better outcome. That is why the strongest AI products combine product judgment with structural assets.
A shopping agent does not win because it can generate a pleasant recommendation. It wins if it understands intent better than a browsing interface: price range, fit, size, availability, support, delivery timing, brand preference, return risk, and the shopper’s past behavior. Retailers can use clickstream data, carts, purchases, ignored results, and returns to understand which products matter in which contexts. The insight is not “add AI to shopping.” It is that shopping is full of hidden constraints. The product has to read them.
The same principle applies in enterprise software. Products that embed into core workflows, connect to proprietary data, and build workflow-native integrations create advantages that are harder to copy than a polished assistant. Products that generate valuable usage data can improve their evals, tune workflows, reduce errors, and create better automation loops. Taste without asset creation is fragile. A good AI wrapper can become a durable company, but only if it compounds into something model providers or incumbents cannot easily absorb: trusted execution, workflow position, proprietary data, distribution, domain-specific evals, implementation knowledge, or an agent-readable system.
The new strategic question: what must we own?
A company building with AI now faces a more precise build-versus-buy decision. It can buy models, infrastructure, and open-weight systems. It can route tasks across vendors, expose tools through MCP servers, build custom agents, wire proprietary data into an existing model, or hire forward-deployed teams to turn customer needs into working systems. The strategic question is not “Which AI should we use?” It is “Which part of the value chain must we own for this product to become defensible?”
For some products, the answer is the data layer. An agent-readable data layer lets AI clients read, query, reason about, write to, and connect data without scraping a human UI. That can become a structural advantage as agents become a normal way users interact with software. For others, the answer is workflow integration. If models are good enough for the task, the company that owns the workflow, permissions, history, and user trust may win. For regulated industries, the answer often includes governance. Healthcare AI adoption can fail when a system disrupts clinical workflow, adds documentation burden, or creates unclear accountability. In finance, legal, healthcare, and insurance, trust, auditability, retention, privacy, and policy enforcement are product features. They decide whether the product can be deployed.
For enterprise AI, the answer may be implementation. The spread of forward-deployed models shows a practical truth: many customers do not have an AI technology problem. They have a business implementation problem. They need the system to fit their workflows, data locality, latency requirements, compliance rules, approval paths, and outcome targets.
This also cuts the other way. Incumbents already own many workflows. Model providers are moving into memory, retrieval, evals, tool-use frameworks, and enterprise services. If they own more of the surrounding system, the last mile does not automatically belong to startups or application companies. Last-mile judgment is a test of where advantage can actually compound, not a slogan for the application layer. The value comes from how models, tools, data, agents, human judgment, and business processes work together inside the customer’s actual environment.
Product judgment also creates speed
There is another reason product judgment matters more as AI gets faster: bad paths get cheaper too. AI lets teams generate prototypes, write code, test variants, analyze feedback, and produce collateral faster. That speed helps only if the team knows what to ignore.
Strong product judgment compresses decision time. It helps a team eliminate the chatbot bolted onto the side of the product, reject the custom model when existing models are sufficient, avoid collecting context that creates privacy risk, and choose a narrow workflow agent with clear escalation rules instead of a general agent with vague authority. Product judgment does not compete with speed. It makes speed useful. Without judgment, AI accelerates activity. With judgment, it accelerates learning.
The last-mile AI strategy test
Before investing in an AI product bet, ask five questions. These questions separate an AI feature from a potential product advantage.
1. Workflow
What recurring workflow are we changing? The workflow should be frequent, expensive, painful, risky, or strategically important enough to matter. If the AI does not change how work gets done, it will compete with every other assistant for attention.
Weak answer: “Users can ask questions about their data.” Strong answer: “Analysts spend three hours every Monday reconciling forecast changes across spreadsheets, CRM exports, and finance assumptions. The product identifies discrepancies, explains variance, and prepares the review packet inside the spreadsheet workflow.”
2. Context
What does the AI know here that a generic model would not know? This can include user history, company data, files, permissions, behavior, domain rules, transaction state, workflow timing, or live system data. But context needs boundaries. The product must decide what to include, persist, retrieve, summarize, hide, or forget.
Weak answer: “We connect the model to all company data.” Strong answer: “For this renewal workflow, the AI sees the contract terms, support history, usage trend, approval policy, and current account owner. It does not see unrelated employee records, stale notes, or private conversations that cannot improve the decision.” A generic assistant has capability. A product with the right context has relevance.
3. Judgment
Where do humans still need taste, accountability, verification, or override authority? This question matters most in high-stakes workflows. AI can expand analysis, draft options, find patterns, and recommend actions. But accountability stays with people and institutions.
Weak answer: “The system decides automatically unless the user stops it.” Strong answer: “The system drafts the recommendation, cites the inputs, flags uncertainty, and routes exceptions to the accountable owner before anything changes in the system of record.” A decision product needs guardrails, escalation paths, audit trails, and clear authority. It should know when to act, when to ask, when to explain, and when to stop. If no one can reconstruct why the system made a recommendation, the product is not ready for serious work.
4. Asset
What durable advantage compounds over time? The answer cannot be “better UX” alone. The product needs an asset that gets stronger with use. That asset could be proprietary data, usage loops, domain-specific evals, workflow integration, trust, distribution, implementation knowledge, agent-readable infrastructure, or customer-specific outcome history.
Weak answer: “Our interface is easier to use than the generic assistant.” Strong answer: “Every completed workflow improves our exception library, eval set, approval logic, and customer-specific outcome history, making the next recommendation more accurate and easier to audit.” This is where many AI wrappers fail. They solve a visible pain, but they do not accumulate anything structural. A competitor with better distribution, deeper data, or model-level access can copy the surface.
5. Outcome
What measurable customer result improves? The answer should be concrete: time saved, revenue gained, errors reduced, decisions improved, compliance strengthened, work removed, cycle time shortened, or margin improved.
Weak answer: “The product makes the team more productive.” Strong answer: “The product cuts the weekly reconciliation cycle from three hours to thirty minutes, reduces formula errors, and gives the finance lead an auditable packet before the forecast meeting.” This question protects teams from demo theater. A product can look intelligent and still fail to change the economics of the customer’s work.
A strong AI strategy can explain exactly which workflow changes, why the product has privileged context, where human judgment remains essential, what asset compounds, and how the customer outcome improves.
The gateway question
The companies that win with AI will not be the ones that simply advertise AI. Many of the strongest AI products will disappear into the work. They will automate a painful step, improve a decision, expose the right context, or make a workflow agent-readable.
The gateway question is simple: what would have to become true for this AI feature to be missed if it disappeared? If the answer is “users would lose a convenient shortcut,” the product may still be useful, but convenience is easy to copy. On the other hand, if users would lose a workflow they now trust, a decision process they can audit, a data loop that improves with every use, or an outcome they can measure, the product is closer to a moat.
Before asking which model to use, ask which workflow deserves intelligence. Before adding more context, ask which context improves the outcome and which context creates risk. Before automating a decision, ask where human judgment, accountability, and override authority must remain. Before claiming a data moat, ask whether the data is fresh, governed, relevant, and used in a feedback loop. AI makes building easier. That does not make strategy easier. It moves strategy closer to the user, closer to the workflow, and closer to the judgment calls that decide what should exist in the first place.