March 20, 2026·28 min read·4 views·6 providers

Semantic Layer War: Who Owns the AI-Data Interface

LLMs are reshaping BI: warehouses, dbt, and BI vendors race to own the semantic layer. Deep analysis of architectures, pipelines, migration paths, and near

Key Finding

The Open Semantic Interchange (OSI) initiative has achieved unusually broad industry participation (25+ organizations including direct competitors) and represents a genuine attempt at semantic portability standards, with MetricFlow as the Apache 2.0 reference implementation.

high confidenceSupported by Perplexity, Grok, Gemini-Lite, Gemini, Anthropic
Justin Furniss
Justin Furniss

@Parallect.ai and @SecureCoders. Founder. Hacker. Father. Seeker of all things AI

perplexitygrok-premiumgemini-liteopenai-minigeminianthropic

The Semantic Layer War: How LLMs Are Reshaping BI Architecture and Who Will Own the AI-Data Interface

A Definitive Cross-Provider Analysis — March 2026


Executive Summary

  • The semantic layer has crossed the threshold from BI convenience to AI infrastructure prerequisite. All six providers independently confirm that LLMs querying raw warehouse schemas without semantic grounding fail catastrophically in production — with accuracy as low as 6-17% on enterprise schemas versus 86%+ on academic benchmarks. The semantic layer is the single highest-leverage intervention for improving AI reliability, with documented accuracy improvements of 5-10x when semantic context is applied.

  • The market has fractured into three competing architectural bets — warehouse-native (Snowflake, Databricks), transformation-native (dbt/MetricFlow), and universal/headless (Cube, AtScale) — and no single approach will win outright. The Open Semantic Interchange (OSI) initiative, launched in 2025 with unusually broad industry participation, is the most credible attempt at preventing winner-take-all lock-in, but its v1.0 specification is still being finalized and real interoperability remains unproven.

  • The dbt-Fivetran merger (October 2025) is the most consequential consolidation event in the modern data stack, creating a combined ~$600M ARR entity spanning ingestion, transformation, and semantic definition. Community anxiety about open-source governance is real and documented — the tension between dbt's open-source roots and commercial monetization pressure is the most important unresolved strategic question in the ecosystem.

  • Cloud warehouses are executing an aggressive platform absorption strategy that threatens standalone BI tools. Snowflake's Semantic Views and Databricks' Unity Catalog Metric Views embed semantic definitions natively, offering operational simplicity at the cost of vendor lock-in. The "thin visualization client" scenario for legacy BI is no longer hypothetical — it is the stated direction of both Snowflake and Databricks product strategy.

  • The winning enterprise architecture pattern in 2026 is "compose, not build or buy monolithically": warehouse-native semantics for stable core KPIs, transformation-layer semantics (dbt) for governed metric definitions, BI-layer semantics for rapid iteration, and OSI as the interchange format — with MCP as the AI access protocol connecting all layers to LLM agents. Organizations that delay semantic layer investment to wait for architectural clarity are accumulating governance debt that compounds with every AI deployment.


Cross-Provider Consensus

Finding 1: LLM Accuracy Without Semantic Layers Is Unacceptably Low for Production Use

Providers in agreement: Perplexity, Grok, OpenAI-Mini, Gemini, Anthropic (5/6 providers) Confidence: HIGH

All five providers cite empirical evidence of dramatic accuracy gaps between academic benchmarks and enterprise production environments. Perplexity cites a 93% accuracy drop (86% on Spider benchmark → 6% on enterprise schemas). OpenAI-Mini cites the same Promethium benchmark showing 6-10% accuracy for vanilla GPT-4 on real enterprise schemas. Anthropic cites AtScale's TPC-DS benchmark showing 80% error rates without semantic grounding. Gemini-Lite describes this as the "context blindness" problem. Grok cites Looker+Gemini reducing errors by ~66% with semantic grounding. The convergence across providers using different source materials makes this the highest-confidence finding in the entire report.

Practitioner caveat (Anthropic, Grok): Vendor-reported accuracy numbers (AtScale's 92.5%, App Orchid's 99.8%) should be treated skeptically — they are self-reported on controlled test sets. Independent benchmarks like LiveSQLBench show even state-of-the-art models (Gemini-2.5-pro) achieving only 28-35% on colloquial real-world queries.


Finding 2: Three Dominant Architectural Patterns Have Emerged

Providers in agreement: Perplexity, Grok, Gemini-Lite, OpenAI-Mini, Gemini, Anthropic (6/6 providers) Confidence: HIGH

Every provider independently converges on the same three-way taxonomy: (1) BI-native/embedded semantics, (2) warehouse-native/platform semantics, (3) universal/headless semantics. The framing varies slightly — Gemini-Lite adds "transformation-native" as a fourth category, Anthropic separates dbt from universal headless — but the underlying architectural reality is consistent across all sources. This taxonomy is the most reliable framework for evaluating vendor positioning.


Finding 3: The Open Semantic Interchange (OSI) Is a Genuine Industry Inflection Point

Providers in agreement: Perplexity, Grok, Gemini-Lite, Gemini, Anthropic (5/6 providers) Confidence: HIGH

All five providers confirm OSI's launch, its broad coalition (Snowflake, Salesforce, dbt Labs, Databricks, AtScale, Cube, Omni, Sigma, ThoughtSpot, plus 20+ others), and the significance of dbt Labs open-sourcing MetricFlow under Apache 2.0 as the reference implementation. The coalition breadth is historically unusual for data infrastructure standards efforts. However, providers uniformly note that OSI v1.0 is still being finalized and real tool-to-tool interoperability has not yet been demonstrated at scale.


Finding 4: dbt Has Massive Developer Adoption and Is Repositioning as Semantic Infrastructure

Providers in agreement: Perplexity, Grok, OpenAI-Mini, Gemini, Anthropic (5/6 providers) Confidence: HIGH

Multiple providers cite consistent adoption figures: 25,000+ companies using dbt, 80,000+ data teams (Anthropic), 5,000+ paying customers, $100M+ ARR (OpenAI-Mini). The dbt Semantic Layer (MetricFlow) reached GA in October 2024 and has 200+ enterprise adopters as of early 2025 (Gemini). The open-sourcing of MetricFlow under Apache 2.0 at Coalesce 2025 is confirmed by all providers as a strategic pivot toward community governance.


Finding 5: Warehouse-Native Semantics Create Operational Simplicity at the Cost of Vendor Lock-in

Providers in agreement: Perplexity, Grok, Gemini-Lite, OpenAI-Mini, Gemini, Anthropic (6/6 providers) Confidence: HIGH

Universal agreement that Snowflake Semantic Views and Databricks Unity Catalog Metric Views offer compelling operational advantages (zero overhead, native security, query optimization) but create significant migration risk. Perplexity cites Gartner data showing 78% of enterprises manage data across 10+ heterogeneous platforms, making pure warehouse-native semantics problematic for most large organizations. All providers agree this is the central lock-in trade-off data teams must evaluate.


Finding 6: The dbt-Fivetran Merger Is the Most Significant Modern Data Stack Consolidation Event

Providers in agreement: Perplexity, Grok, Anthropic (3/6 providers) Confidence: HIGH (for the fact; MEDIUM for implications)

Three providers confirm the October 2025 merger, the ~$600M combined ARR figure, and the strategic rationale (80-90% customer overlap). Anthropic provides the most detailed community sentiment analysis, noting real anxiety about open-source governance and monetization pressure. The implications for the semantic layer landscape — whether the combined entity accelerates or constrains open standards — remain genuinely uncertain.


Finding 7: The "Compose" Architecture Pattern Is Winning Over Monolithic Approaches

Providers in agreement: Grok, Gemini-Lite, OpenAI-Mini, Anthropic (4/6 providers) Confidence: MEDIUM

Four providers independently identify the composable architecture as the emerging best practice: warehouse for compute/storage, dbt/MetricFlow for governed metric definitions, BI tool for visualization and iteration, catalog for governance context. However, the specific composition varies by provider and the evidence base is primarily practitioner opinion rather than controlled studies.


Finding 8: Data Catalogs (Atlan, Alation, Collibra) Complement Rather Than Replace Semantic Layers

Providers in agreement: Perplexity, Grok, OpenAI-Mini, Gemini (4/6 providers) Confidence: HIGH

All four providers draw a consistent distinction: catalogs provide the "map" (what data exists, lineage, ownership, discovery) while semantic layers provide the "rules" (how to calculate metrics, which joins are valid, what business terms mean). The best production deployments use both in combination. Gemini uniquely highlights Atlan's positioning as a "metadata lakehouse" that bridges both functions.


Unique Insights by Provider

Perplexity

  • The five-level context hierarchy for LLM accuracy. Perplexity provides the most detailed quantification of how accuracy scales with context richness: raw schema alone (10-20%) → adding relationships (20-40%) → data catalog integration (40-70%) → semantic layer with vetted metrics (70-90%) → tribal knowledge/validated patterns (90-99%). This is the most actionable accuracy framework in the entire report and is not replicated by other providers. It gives data teams a concrete investment roadmap rather than a binary "use a semantic layer or don't."

  • The "agentic semantic layer" as a distinct emerging category. Perplexity is the only provider to specifically call out ThoughtSpot's Spotter Semantics (announced March 2026) as representing a new category beyond traditional semantic layers — one that enforces deterministic query generation for AI agents rather than merely providing metadata context. This distinction between "semantic layer as documentation" versus "semantic layer as enforcement mechanism" is architecturally significant.

  • Unstructured data governance as the next semantic frontier. Perplexity uniquely identifies the gap between current semantic layers (which assume structured data) and the emerging need to govern unstructured assets (documents, images, audio) flowing through RAG systems. This is a genuine capability gap that no current semantic layer vendor has solved.


Grok

  • AtScale's Snowflake Ventures strategic investment (December 2025). Grok is the only provider to specifically highlight that Snowflake Ventures led a strategic equity round in AtScale, signaling potential full acquisition. This is a significant M&A signal — Snowflake investing in a universal semantic layer competitor suggests either a hedge against its own warehouse-native approach or a precursor to acquisition. This deserves specific monitoring.

  • OSI working group composition detail. Grok provides the most complete list of OSI working group participants, including AWS, Collibra, DataHub, Domo, Firebolt, Informatica, Instacart, JPMC, Preset, Starburst, Alation, Atlan, BlackRock, Cube, Elementum AI, Hex, Honeydew, Mistral AI, Omni, RelationalAI, Salesforce, Select Star, Sigma, and ThoughtSpot. The breadth of this list — including direct competitors — is historically unusual and strengthens the case that OSI represents genuine industry consensus rather than vendor-driven positioning.


Gemini-Lite

  • The "metrics-first migration" as the canonical approach for legacy BI transitions. Gemini-Lite is the most explicit in articulating the migration sequence: define "contested metrics" (Revenue, Churn) in a central semantic layer before moving the visualization layer. This sequencing insight — semantics first, visualization second — is the key differentiator between successful and failed BI migrations, and is stated more clearly here than in other providers.

  • The OSI "3-year disappearance" prediction. Gemini-Lite makes the boldest structural prediction: the term "semantic layer" will disappear as a distinct category within 3 years, absorbed into "Data Intelligence Platforms." The winners will be platforms that allow semantic definitions to be exported and interoperable rather than locked inside proprietary interfaces. This is a more aggressive consolidation prediction than other providers offer.


OpenAI-Mini

  • Specific funding data for the BI startup ecosystem. OpenAI-Mini provides the most complete funding picture: Omni $69M at $650M valuation (March 2025), Sigma $200M Series D (2024), Hex $28M (2023) plus $70M Series C (Anthropic also cites this). This funding data, cross-referenced with Ramp's BI spend data (Hex growing from 9% to 25% of startup BI spend since 2023), provides the clearest picture of where enterprise dollars are actually flowing.

  • Prompt-to-SQL injection as an underappreciated security risk. OpenAI-Mini is the only provider to specifically cite academic research on "Prompt-to-SQL Injections" — where malicious or poorly sanitized prompts trick LLMs into running unauthorized queries. This security failure mode is distinct from accuracy failures and represents a governance risk that semantic layers partially (but not fully) mitigate.


Gemini (Full)

  • ServiceNow's acquisition of Pyramid Analytics (February 2026). Gemini is the only provider to report this specific M&A event, framing it as evidence that enterprise workflow platforms are acquiring semantic/analytics capabilities to enable "DecisionOps" — moving from passive dashboards to autonomous enterprise intelligence. This acquisition pattern (workflow platform + analytics semantics) may be more significant than warehouse-native consolidation for the broader enterprise software market.

  • Streaming Semantic Layers as an emerging requirement. Gemini uniquely identifies the gap between batch-updated semantic layers (sufficient for human BI users checking dashboards daily) and the sub-second freshness requirements of AI agents querying data thousands of times per hour. Tools like Streamkap integrating CDC to update semantic definitions in real-time represent a new infrastructure category that no current major semantic layer vendor has fully addressed.

  • Workday's 5x AI accuracy improvement via Atlan context layer. Gemini provides the most specific named enterprise case study for context-layer-driven accuracy improvement: Workday achieved a 5x improvement in AI response accuracy by grounding agents in Atlan's shared semantic layer enriched with "decision traces." This is a more specific and credible data point than most vendor-reported accuracy claims.

  • App Orchid's 99.8% Text-to-SQL accuracy via ontology-driven semantic layer. While this figure should be treated with appropriate skepticism (vendor-reported, controlled test set), Gemini is the only provider to cite this specific benchmark, which represents the theoretical upper bound of what semantic grounding can achieve.


Anthropic

  • The Ramp Velocity Report BI spend data. Anthropic is the only provider to cite Ramp's actual enterprise spending data showing Hex growing from 9% to 25% of startup BI spend since 2023, while Looker, Tableau, and Power BI still command 34% of total BI dollars with their share increasing 6 percentage points. This is the most credible independent market share data in the report — based on actual transaction data rather than surveys or vendor claims — and it reveals a nuanced picture: AI-native tools are growing fast among startups, but incumbents are holding or gaining share in total enterprise dollars.

  • The MetricFlow self-hosting limitation as a critical adoption barrier. Anthropic uniquely identifies that MetricFlow defines metrics but doesn't serve them — to query MetricFlow metrics from an application, you need dbt Cloud's paid Semantic Layer API. There's no self-hosted API endpoint. This is a significant practical limitation that vendor documentation obscures and that directly affects the "open-source" positioning of dbt's semantic layer.

  • Gartner's MCP-without-semantics failure prediction. Anthropic cites a specific Gartner prediction: "by 2028, 60% of agentic analytics projects relying solely on MCP will fail due to the lack of a consistent semantic layer." This is the clearest analyst statement connecting MCP adoption to semantic layer requirements and provides a concrete failure timeline.

  • The 95% AI pilot failure rate context. Anthropic cites MIT Project NANDA finding that ~95% of generative AI pilots show no measurable P&L impact. This provides crucial context for the entire semantic layer investment case: the semantic layer is not just a technical nicety but the primary intervention separating the 5% of AI deployments that work from the 95% that don't.


Contradictions and Disagreements

Contradiction 1: dbt's Positioning — Transform Tool vs. Semantic Layer Owner

Side A (Grok, Gemini-Lite): dbt is positioning as a transform tool that provides canonical governed metrics for both BI and AI — not a full BI replacement or pure warehouse layer. The emphasis is on "define once, query everywhere" with dbt as the upstream source feeding other tools.

Side B (Anthropic, Perplexity): dbt is positioning as the semantic layer for AI — actively building infrastructure for AI agents to query against dbt-defined semantics via MCP, and the dbt-Fivetran merger signals ambitions to own the entire data pipeline from ingestion through semantic definition.

Assessment: This is a genuine strategic ambiguity, not a factual disagreement. dbt Labs appears to be deliberately maintaining both framings simultaneously — "we're a transform tool" for the data engineering audience and "we're the semantic layer for AI" for the AI/BI audience. The Fivetran merger makes the broader platform ambition harder to deny. Practitioners should watch dbt's pricing and feature decisions in 2026 for signals about which framing wins internally.


Contradiction 2: Will the Semantic Layer Consolidate or Remain Federated?

Side A (Gemini-Lite, OpenAI-Mini): The semantic layer will consolidate into 2-3 dominant approaches by 2028, with the term potentially disappearing as a distinct category absorbed into "Data Intelligence Platforms." The market will follow the pattern of previous data infrastructure consolidations.

Side B (Anthropic, Perplexity): The semantic layer will not consolidate into 2-3 winners. Instead, OSI will enable a federated model where multiple semantic layers coexist and interoperate. The "semantic layer" as a distinct product category will partially dissolve but will persist as a capability embedded in three places simultaneously (warehouse, transformation layer, BI tool).

Assessment: Both predictions are plausible and may describe different market segments. Enterprise organizations standardized on a single cloud platform will likely consolidate to warehouse-native semantics. Multi-cloud, multi-BI organizations will maintain federated approaches. The resolution may be segment-specific rather than market-wide.


Contradiction 3: Accuracy Claims — Vendor Numbers vs. Independent Benchmarks

Side A (OpenAI-Mini, Grok citing vendor claims): AtScale reports 92.5% accuracy with semantic grounding. App Orchid reports 99.8% accuracy. Looker+Gemini reduces errors by ~66%. dbt Semantic Layer achieves 83% accuracy on complex business questions.

Side B (Anthropic, Perplexity citing independent benchmarks): LiveSQLBench shows Gemini-2.5-pro achieving only 28-35% on real-world colloquial queries. Promethium benchmarks show 6-10% accuracy for vanilla GPT-4 on enterprise schemas. Uber's internal text-to-SQL achieves only 50% overlap with ground truth tables. Production accuracy for open-ended questions against heterogeneous enterprise systems is 10-20%.

Assessment: These numbers are not directly comparable — vendor accuracy claims are measured on controlled, curated test sets designed to showcase their tools, while independent benchmarks test against real-world enterprise complexity. Both sets of numbers are likely "true" in their respective contexts. Data teams should demand independent benchmark results on their own schemas before accepting vendor accuracy claims. The gap between 92.5% (vendor-reported) and 28-35% (independent benchmark) is the most important number to interrogate in any vendor evaluation.


Contradiction 4: Whether Semantic Layers Should Live in the BI Tool or Upstream

Side A (Omni's argument, cited by Anthropic): The semantic layer belongs in the BI tool because business logic needs to live where people learn, investigate, and make decisions. Upstream semantic layers are too rigid for the pace of business change — the business learns something new, but the data model doesn't catch up for weeks or months.

Side B (dbt Labs, Snowflake, Databricks, cited by Perplexity, Grok, Gemini): The semantic layer must live upstream (in the transformation layer or warehouse) to serve both human BI users AND AI agents simultaneously from a single source of truth. BI-native semantics are inaccessible to AI agents without special connectors.

Assessment: This is a genuine architectural tension with no clean resolution. The emerging consensus (Anthropic, Grok) is a layered approach: stable core KPIs upstream, iterative/exploratory metrics in the BI layer, with OSI as the synchronization mechanism. But this "both/and" answer introduces its own complexity — which layer wins when definitions conflict?


Contradiction 5: The dbt-Fivetran Merger — Strategic Acceleration vs. Governance Risk

Side A (Perplexity): The merger creates a streamlined end-to-end data workflow, reduces operational overhead, and the open-sourcing of MetricFlow demonstrates commitment to open standards. The merger is net positive for the ecosystem.

Side B (Anthropic, Grok): Community anxiety is real and justified. Recent monetization moves (VS Code Extension freemium limits), the BSL licensing history, and the pattern of acquired open-source projects becoming commercial products create legitimate governance risk. The merger concentrates too much of the modern data stack under a single commercial entity.

Assessment: Both perspectives are grounded in evidence. The open-sourcing of MetricFlow is a genuine positive signal; the monetization pressure from a ~$600M ARR entity with investor expectations is a genuine risk. This contradiction will resolve based on dbt Labs' specific product and licensing decisions in 2026-2027 — watch for changes to dbt Core's Apache 2.0 license as the key indicator.


Detailed Synthesis

The Semantic Layer's Transformation from BI Feature to AI Infrastructure

The semantic layer's elevation from a BI consistency tool to mandatory AI infrastructure is the defining architectural shift of 2025-2026. This transition was not driven by vendor marketing but by the empirical failure of LLMs deployed against raw warehouse schemas. The evidence is unambiguous across all six providers: without semantic grounding, LLMs fail catastrophically in production environments. The gap between academic benchmark performance (86% accuracy on Spider) and real enterprise performance (6-17% on production schemas) represents what Anthropic calls "the most important number in the semantic layer debate" [Anthropic]. This accuracy collapse occurs because enterprise schemas are vastly more complex than academic test sets — ambiguous column names, implicit business rules, multi-hop joins, and domain-specific terminology that LLMs cannot infer from schema structure alone [Perplexity].

The mechanism by which semantic layers restore accuracy is well-documented. Perplexity provides the most granular framework: raw schema alone produces 10-20% accuracy; adding relationship information (foreign keys, valid joins) improves this to 20-40%; integrating data catalog definitions pushes to 40-70%; adding a semantic layer with vetted metrics and governance policies achieves 70-90%; and incorporating tribal knowledge (validated query patterns, business rule overrides) reaches 90-99% [Perplexity]. This five-level hierarchy is the most actionable accuracy roadmap available to data teams — it frames semantic layer investment not as a binary choice but as a progressive capability build.

The practical implications are stark. Anthropic cites MIT Project NANDA's finding that approximately 95% of generative AI pilots show no measurable P&L impact, with Gartner forecasting that more than 40% of agentic AI projects will be abandoned by 2027 [Anthropic]. The semantic layer is the primary intervention separating successful AI deployments from failed ones — not model selection, not prompt engineering, not infrastructure optimization.

The Three-Way Architectural Fragmentation

The semantic layer market has fractured into three competing architectural philosophies, each backed by different vendors with different commercial interests and different trade-off profiles [Perplexity, Grok, Gemini-Lite, OpenAI-Mini, Gemini, Anthropic].

Warehouse-Native Semantics represent the most aggressive platform consolidation play. Snowflake's Semantic Views — introduced as a native schema-level object in 2025 — allow organizations to store business concepts directly in the database, with Cortex Analyst reading these definitions to generate governed SQL without data leaving Snowflake's security boundary [Gemini, Anthropic]. The "Semantic View Autopilot" feature automatically analyzes verified queries to expand semantic model coverage iteratively, reducing the upfront definition burden [Perplexity]. Databricks' Unity Catalog Metric Views take a parallel approach, making metrics first-class data assets with built-in lineage, access controls, and precomputed materialized views for query optimization [Gemini, Anthropic]. Both approaches offer compelling operational advantages: zero external infrastructure, native security inheritance, and query optimization that leverages the warehouse's existing engine [Perplexity].

The trade-off is vendor lock-in of a particularly severe kind — not just tool lock-in but ontology lock-in. Organizations that define all their metrics as Snowflake Semantic Views face not just a data migration but a business logic migration if they need to change platforms [Perplexity, Grok]. Gartner data cited by Perplexity shows 78% of enterprises manage data across 10+ heterogeneous platforms, meaning pure warehouse-native semantics immediately create fragmentation for the majority of large organizations.

Transformation-Layer Semantics represent dbt's bet that metrics belong in version-controlled code alongside data transformations. MetricFlow, now open-sourced under Apache 2.0, provides a warehouse-agnostic query generation engine that compiles metric requests into optimized SQL for any supported warehouse [Grok, Anthropic]. The developer appeal is strong: metrics undergo the same pull request reviews, CI/CD pipelines, and testing rigor as data transformations [OpenAI-Mini, Gemini]. The dbt community of 80,000+ data teams and 25,000+ companies represents the largest population of practitioners actively building semantic infrastructure [Anthropic, OpenAI-Mini].

However, Anthropic identifies a critical limitation that vendor documentation obscures: MetricFlow defines metrics but doesn't serve them. Querying MetricFlow metrics from an application requires dbt Cloud's paid Semantic Layer API — there is no self-hosted API endpoint [Anthropic]. This means the "open-source" framing of MetricFlow is partially misleading: the definition engine is open, but the serving infrastructure is commercial. Additionally, MetricFlow doesn't pre-aggregate or cache, meaning every query hits the warehouse directly — adding cost and latency for high-volume or customer-facing use cases [Anthropic].

Universal/Headless Semantics provide the highest degree of flexibility and portability. Cube, with 19,268 GitHub stars as of January 2026 and recognition as a Representative Vendor in Gartner's February 2026 Market Guide for Agentic Analytics, offers REST, GraphQL, and SQL APIs that allow any application — BI tools, notebooks, AI agents, custom applications — to consume the same metric definitions [Anthropic, Grok]. AtScale targets enterprise-scale deployments with MDX/DAX-compatible interfaces, deep Microsoft Excel and Power BI integration, and the ability to handle complex business logic (53-week calendars, currency conversions) that simpler semantic layers cannot express [Gemini, OpenAI-Mini]. GigaOm named AtScale a Leader and Fast Mover in the 2025 Semantic Layer Radar [Grok].

The universal approach's limitation is operational complexity — it requires deploying, monitoring, and maintaining an additional system. Anthropic notes that Cube's MCP integration gap is a current weakness: Cube exposes REST, SQL, and GraphQL APIs but none speak MCP natively, requiring custom wrapper development for AI agent connectivity [Anthropic].

The OSI Initiative: Genuine Standard or Strategic Hedge?

The Open Semantic Interchange initiative deserves careful analysis because it is simultaneously the most promising development for reducing lock-in and the most strategically ambiguous [Perplexity, Grok, Gemini-Lite, Gemini, Anthropic].

The coalition is genuinely unprecedented. Grok provides the most complete participant list: Snowflake, Salesforce, dbt Labs, AtScale, Cube, Omni, Sigma, ThoughtSpot, AWS, Collibra, DataHub, Domo, Firebolt, Informatica, Instacart, JPMC, Preset, Starburst, Alation, Atlan, BlackRock, Elementum AI, Hex, Honeydew, Mistral AI, RelationalAI, and Select Star [Grok]. The inclusion of direct competitors (Snowflake and Databricks, dbt and Cube, Looker and ThoughtSpot) in the same working group is historically unusual for data infrastructure standards efforts and suggests genuine industry consensus that semantic fragmentation is a shared problem.

The strategic logic is clear: Perplexity's SiliconANGLE analysis notes that Snowflake, Databricks, Informatica, ThoughtSpot, Starburst, Collibra, and Alation banded together through OSI as a response to SAP, ServiceNow, Salesforce, and Oracle attempting to entrench their own semantic spheres of influence through proprietary data sharing agreements [Perplexity]. OSI is partly a defensive coalition against enterprise software vendor lock-in, not purely altruistic standardization.

The critical uncertainty is whether OSI will achieve real interoperability or become another "open standard" that exists on paper while vendors compete on proprietary extensions. The OSI v0.1 specification was released on GitHub in January 2026 [Gemini], but v1.0 is still being finalized and no provider cites evidence of actual tool-to-tool semantic portability in production. Past data infrastructure standardization efforts (SQL dialects, PMML, ONNX) suggest that standards reduce but do not eliminate fragmentation — vendors comply with the minimum required for interoperability claims while differentiating on proprietary capabilities above the standard.

The dbt-Fivetran Merger: Implications for the Ecosystem

The October 2025 merger of dbt Labs and Fivetran is the most consequential consolidation event in the modern data stack and deserves more analytical attention than most providers give it [Anthropic, Grok, Perplexity].

The strategic logic is straightforward: 80-90% of Fivetran's customers already use dbt, making this a formalization of an existing ecosystem relationship rather than a cross-sell play [Anthropic]. The combined entity — approaching $600M ARR — spans data ingestion (Fivetran), transformation (dbt), semantic definition (MetricFlow), reverse ETL (Census, acquired May 2025), and SQL tooling (SQLMesh/SQLGlot via Tobiko Data, acquired September 2025) [Anthropic]. This is a vertically integrated data pipeline company, not just a BI tool.

The community anxiety is real and documented. Anthropic provides the most nuanced analysis: recent monetization moves including freemium limits on the dbt VS Code Extension signal commercial pressure, and the history of open-source projects acquired by commercial entities (Tableau/Salesforce, Looker/Google) is not encouraging [Anthropic]. Grok notes that the community is specifically concerned about whether dbt Core's Apache 2.0 license will be maintained and whether MetricFlow's open-source commitment will survive commercial pressure [Grok].

The merger also raises questions about competitive dynamics. The combined Fivetran/dbt entity now competes with Snowflake and Databricks for "data pipeline ownership" while simultaneously depending on them as the compute layer. This creates a complex co-opetition dynamic that will shape product roadmap decisions in ways that are difficult to predict from the outside.

The LLM-to-Data Pipeline in Production

The Model Context Protocol (MCP) has emerged as the critical technical bridge between LLMs and semantic layers, but its adoption creates new risks without semantic governance [Anthropic, Grok, Gemini].

MCP solves the technical problem of giving LLMs access to governed business logic rather than raw data. Snowflake's managed MCP server (generally available) allows secure connection to Cortex Agents, Cortex Analyst, and Cortex Search, ensuring data doesn't leave Snowflake's governance boundary [Anthropic]. AtScale's MCP Server enables natural language access via Slack and Google Meet for enterprise deployments [Anthropic]. However, Anthropic cites Gartner's prediction that "by 2028, 60% of agentic analytics projects relying solely on MCP will fail due to the lack of a consistent semantic layer" — MCP is a transport protocol, not a semantic governance mechanism [Anthropic].

The security dimension is underappreciated. OpenAI-Mini is the only provider to specifically cite academic research on "Prompt-to-SQL Injections" — where malicious prompts trick LLMs into running unauthorized queries [OpenAI-Mini]. This represents a governance risk distinct from accuracy failures: even a semantically grounded LLM can be manipulated if the semantic layer doesn't enforce row-level security and query scope restrictions.

Gemini uniquely identifies the streaming semantic layer gap: current semantic layers are batch-updated, which is sufficient for human BI users checking dashboards daily, but AI agents querying data thousands of times per hour require sub-second schema freshness [Gemini]. When underlying database schemas change (column renames, table additions, data type changes), agents operating on stale semantic definitions fail silently — generating syntactically valid but semantically incorrect queries. This is an infrastructure gap that no current major semantic layer vendor has fully addressed.

The Enterprise Decision Framework

The evidence strongly supports the "compose, not build or buy monolithically" pattern [Grok, Gemini-Lite, OpenAI-Mini, Anthropic]. The specific composition depends on organizational context:

For organizations standardized on a single cloud platform with limited multi-tool requirements, warehouse-native semantics (Snowflake Semantic Views or Databricks Metric Views) offer the lowest operational overhead and strongest governance integration. The lock-in risk is real but manageable if the organization has genuinely committed to that platform for the long term.

For organizations with multiple BI tools, multi-cloud data environments, or requirements to expose metrics to custom applications and AI agents, universal/headless semantics (Cube, AtScale) or transformation-layer semantics (dbt MetricFlow) provide superior flexibility. The operational overhead is real but justified by the portability benefit.

For organizations in transition — migrating from legacy BI to AI-native analytics — the "metrics-first migration" pattern is the most reliable approach [Gemini-Lite]: define contested metrics (Revenue, Churn, Active Users) in a central semantic layer before moving the visualization layer. This sequencing prevents the most common migration failure mode — replicating metric fragmentation in the new tool rather than eliminating it.

Real-world migration evidence supports this approach. Anthropic cites Aviatrix migrating from Looker to Omni in three weeks, doubling BI user adoption by eliminating the LookML bottleneck [Gemini]. Incident.io reduced custom SQL from 70% to 5% of queries by migrating to Omni with dbt integration [Gemini]. BuzzFeed consolidated 2,000 Looker dashboards to 400 Omni dashboards in under three months, reducing ad-hoc reporting time by 80% [Perplexity, Grok]. The pattern across successful migrations is consistent: use the migration as an opportunity to rationalize metric definitions, not just replicate existing dashboards.

Market Dynamics and the Path to Consolidation

The BI market is valued at approximately $38-41 billion in 2025-2026 and projected to reach $56-62 billion by 2030-2031 at an 8-9% CAGR [Anthropic, Gemini, Grok]. The AI-native analytics subsegment is growing at 2-3x the overall market rate. Anthropic's Ramp Velocity data provides the most credible independent market share signal: Hex has grown from 9% to 25% of startup BI spend since 2023, while Looker, Tableau, and Power BI still command 34% of total BI dollars with their share increasing 6 percentage points [Anthropic]. This reveals a bifurcated market: AI-native tools are capturing new deployments among startups and tech-forward organizations, while incumbents maintain and grow their share of total enterprise dollars.

The M&A landscape is accelerating. Beyond the dbt-Fivetran merger, Gemini uniquely reports ServiceNow's acquisition of Pyramid Analytics in February 2026 — a signal that enterprise workflow platforms are acquiring semantic/analytics capabilities to enable autonomous enterprise intelligence [Gemini]. Grok highlights Snowflake Ventures' strategic investment in AtScale as a potential acquisition precursor [Grok]. Omni's acquisition of Explo (embedded analytics) in October 2025 signals consolidation of the BI and embedded analytics markets [Gemini].

The three-year prediction that emerges from synthesizing all providers: the semantic layer will not consolidate into 2-3 winners in the traditional sense. Instead, it will become a capability embedded in multiple layers simultaneously — warehouse, transformation, BI — with OSI as the interchange format. The "semantic layer vendor" category will partially dissolve as the capability becomes table stakes for all data platforms. The winners will be platforms that make multi-layer semantic governance feel seamless rather than fragmented, and organizations that invest in governing their business definitions regardless of which tool houses them.


Evidence Explorer

Select a citation or claim to explore evidence.

Go Deeper

Follow-up questions based on where providers disagreed or confidence was low.

Independent benchmark study comparing LLM accuracy across semantic layer architectures (warehouse-native vs. transformation-native vs. universal/headless) on identical enterprise schemas, using a standardized evaluation methodology like LiveSQLBench rather than vendor-controlled test sets.

The most critical gap in the current research is the absence of independent, apples-to-apples accuracy comparisons across architectural approaches. Vendor claims range from 83% to 99.8% accuracy, while independent benchmarks show 28-35% for state-of-the-art models on real-world queries. A rigorous independent study on 3-5 representative enterprise schemas would be the single most valuable piece of research for data teams making architectural decisions.

OSI v1.0 specification analysis and real-world interoperability testing — specifically, whether a metric defined in dbt MetricFlow can be consumed without modification by Snowflake Cortex Analyst, Databricks Genie, and a Cube-powered application simultaneously.

OSI is the most consequential standards initiative in the semantic layer space, but no provider cites evidence of actual tool-to-tool portability in production. The gap between "OSI specification exists" and "OSI enables real portability" is the most important unresolved question for organizations evaluating lock-in risk. Testing this on a representative set of metric definitions (revenue, churn, DAU) would provide concrete guidance.

dbt-Fivetran merger impact analysis — specifically tracking changes to dbt Core's Apache 2.0 licensing, MetricFlow's open-source governance model, and pricing for dbt Cloud's Semantic Layer API over the 12 months following the merger.

Multiple providers flag community anxiety about open-source governance post-merger, but the actual impact will only be visible through specific product and licensing decisions. A structured monitoring framework tracking these indicators would provide early warning of governance drift and help organizations assess whether their dbt-centric semantic layer investments are at risk.

Enterprise case study research on semantic layer migration failures — specifically, organizations that attempted to implement warehouse-native semantics (Snowflake Semantic Views or Databricks Metric Views) and encountered the lock-in or heterogeneous environment problems identified in this report.

The current research is heavily weighted toward success stories (BuzzFeed/Omni, Aviatrix/Omni, Incident.io/Omni). The failure modes of warehouse-native semantic approaches — particularly for organizations that later needed to migrate platforms or integrate multi-cloud data — are underrepresented. Understanding what breaks in practice would significantly improve the enterprise decision framework.

Streaming semantic layer landscape analysis — mapping which vendors (if any) have production-ready solutions for real-time semantic definition updates via CDC, and what the actual freshness requirements are for different AI agent use cases (conversational analytics vs. automated decision-making vs. real-time monitoring).

Gemini identifies streaming semantic layers as an emerging infrastructure gap, but this finding is from a single provider with limited corroboration. If AI agents genuinely require sub-second semantic freshness (as opposed to the minutes-to-hours acceptable for human BI), this represents either a significant unaddressed market opportunity or a requirement that current architectures cannot meet — both of which have major implications for vendor selection.

Key Claims

Cross-provider analysis with confidence ratings and agreement tracking.

12 claims · sorted by confidence
1

Snowflake Semantic Views and Databricks Unity Catalog Metric Views represent production-ready warehouse-native semantic layers that eliminate external infrastructure overhead but create significant vendor lock-in.

high·Perplexity, Grok, Gemini-Lite, OpenAI-Mini, Gemini, Anthropic(NONE (universal agreement on both the capability and the lock-in risk) disagrees)·
2

The Open Semantic Interchange (OSI) initiative has achieved unusually broad industry participation (25+ organizations including direct competitors) and represents a genuine attempt at semantic portability standards, with MetricFlow as the Apache 2.0 reference implementation.

high·Perplexity, Grok, Gemini-Lite, Gemini, Anthropic(NONE (though real interoperability in production has not yet been demonstrated) disagrees)·
3

LLMs querying enterprise schemas without semantic grounding achieve only 6-17% accuracy in production, versus 86%+ on academic benchmarks — a gap of 70-80 percentage points.

high·Perplexity, OpenAI-Mini, Anthropic, Grok(NONE (though vendor-reported numbers with semantic grounding are much higher and should be treated skeptically) disagrees)·
4

dbt has 25,000+ companies and 80,000+ data teams using it, with $100M+ ARR and 5,000+ paying customers — making it the largest developer community in the semantic layer space.

high·OpenAI-Mini, Anthropic, Grok, Perplexity·
5

The dbt-Fivetran merger created a combined entity approaching $600M ARR spanning ingestion, transformation, semantic definition, reverse ETL, and SQL tooling — the most significant modern data stack consolidation event to date.

high·Anthropic, Grok, Perplexity·
6

MetricFlow's "open-source" positioning is partially misleading — the definition engine is Apache 2.0, but the serving infrastructure (Semantic Layer API) requires a paid dbt Cloud plan with no self-hosted alternative.

high·Anthropic(NONE (other providers don't address this specific limitation, but none contradict it) disagree)·
7

The "compose" architecture pattern — warehouse for compute, dbt/MetricFlow for governed metrics, BI tool for visualization, catalog for governance context — is the emerging best practice for organizations with multi-tool environments.

medium·Grok, Gemini-Lite, OpenAI-Mini, Anthropic(Gemini-Lite (argues this complexity is unnecessary for single-platform organizations; warehouse-native is simpler and sufficient) disagrees)·
8

The BI market will reach $56-62 billion by 2030-2031 at an 8-9% CAGR, with the AI-native analytics subsegment growing at 2-3x the overall market rate.

medium·Gemini, Grok, Anthropic(NONE (figures are consistent across providers using different market research sources) disagrees)·
9

Adding a semantic layer with vetted metrics and governance policies improves LLM accuracy from 10-20% (raw schema) to 70-90%, with tribal knowledge pushing this to 90-99%.

medium·Perplexity, OpenAI-Mini(Anthropic (independent benchmarks show even state-of-the-art models achieving only 28-35% on real-world colloquial queries, suggesting the upper bound is lower than vendor claims) disagree)·
10

Gartner predicts that by 2028, 60% of agentic analytics projects relying solely on MCP will fail due to the lack of a consistent semantic layer.

medium·Anthropic(NONE (single-provider claim, not independently verified) disagree)·
11

Hex has grown from 9% to 25% of startup BI spend since 2023, while Looker, Tableau, and Power BI still command 34% of total BI dollars with their share increasing 6 percentage points — revealing a bifurcated market where AI-native tools capture new deployments but incumbents hold enterprise dollars.

medium·Anthropic(NONE (single-provider claim based on Ramp transaction data, but methodology is credible) disagree)·
12

Streaming semantic layers — updating semantic definitions in real-time as underlying schemas change via CDC — are an emerging infrastructure requirement for AI agent deployments that no current major semantic layer vendor has fully addressed.

low·Gemini(NONE (single-provider claim, limited corroboration) disagree)·

Topics

semantic layerLLMs and BIdbt semantic layerwarehouse-native semanticsAI-data interfacedata governance for AIBI architecture 2026

Share this research

Read by 4 researchers

Share:

Research synthesized by Parallect AI

Multi-provider deep research — every angle, synthesized.

Start your own research