Back to News
Article · Definition

What is a Data Foundation?

The technical layer that turns scattered business data into one trustworthy source for reports, AI, and decisions. A working definition. When you need one, and when you don't.

By Aleksi Stenberg · 16 May 2026 · 10 min read
Summary

A data foundation is the layer of database systems, pipelines, modeling, and governance that connects a company's operational systems into one trustworthy source for analytics, reports, and AI. The point of the layer is to give every consumer (a dashboard, an AI agent, a CFO, a regulator) the same version of the truth.

Most companies under 30 people do not need one. Above 50 to 100 people the question changes from "do we need it" to "what does it cost to keep going without it". This piece is a working definition, the parts that go inside, the architecture choices, and the common mistakes that turn a useful build into a stalled project.

01

A Working Definition

The phrase "data foundation" has entered every vendor pitch. It now covers everything from a single warehouse to a multi-region lakehouse with a semantic layer and column-level access controls. Buyers reach the end of the demo without a clear answer to "what is the thing".

A data foundation is the layer of database systems, pipelines, modeling, and governance that connects a company's operational systems into one trustworthy source for analytics, reports, and AI. Its job is to answer one question reliably: which version of the truth do we use?

Every company already has data. It sits in the CRM, the ERP, the billing system, the marketing tools, the support tickets, the product database, and several hundred spreadsheets. Each system has its own version of the customer, the product, the contract, and the revenue. When leadership asks a question that spans systems, someone runs around and reconciles by hand. A data foundation removes the running around.

The full layer has four parts working together:

  1. Ingestion. Data flows in from operational systems on a schedule, in real time, or both.
  2. Storage. A warehouse or lakehouse holds the raw data and the modelled tables that downstream consumers actually query.
  3. Modeling. Raw data gets shaped into the tables that match how the business thinks. Customer, product, order, revenue.
  4. Governance. Who can see what, who changed what, how do we know the numbers are right.

None of the four is optional in production. Skipping storage gives you batch jobs that have nowhere to land. Skipping modeling produces reports that mean different things in different rooms. Skipping governance puts the project one regulator visit away from rebuild.

02

What Goes Inside

The specific tools that fill the four parts vary by company, budget, and team. The current default Nordic mid-market stack looks roughly like this:

LayerCommon choicesNotes
IngestionFivetran, Airbyte, Stitch, custom PythonFivetran for breadth and SLA. Airbyte for cost and open source.
Storage (warehouse)Snowflake, BigQuery, Databricks, PostgresPostgres remains a strong default below 1 TB. Snowflake and BigQuery cover most large-scale cases.
Storage (lake)S3 + Iceberg, GCS + Delta, Azure ADLS + DeltaUsed when raw files (PDFs, logs, images) matter alongside structured tables.
Transformationdbt, SQLMesh, custom SQLdbt is the default. SQLMesh adds better state and testing but smaller community.
Semantic layerCube, dbt Semantic, MetricFlow, AtScaleWorth the effort once metrics start meaning different things in different tools.
Custom data appsReact, Next.js, Vue, FastAPI, Express, Django + D3, Recharts, EChartsThe Jourier recommendation: real product-grade apps written in proper frameworks, owned by the client, deployed in the client's accounts. No per-seat fee, no vendor ceiling, AI-agent integration native. BI tools (Power BI, Tableau, Looker, Metabase, Superset, Streamlit, Retool) are legacy that companies migrate away from.
Reverse ETLHightouch, Census, PolytomicSends modelled data back to operational tools (CRM, marketing).
GovernanceUnity Catalog, Snowflake roles, Postgres RLS, OpenMetadataBuilt-in to the warehouse for access control. Catalog tools for lineage and discovery.

The choice depends on the size of the data, the existing cloud, and the skill profile of the team. A company with a 200-person engineering org and AWS commitments will land somewhere different than a 60-person fintech on Google Cloud.

03

When You Need One

Three signals show up consistently before a company decides it has outgrown the SaaS-tool jungle:

Reports disagree. Marketing reports one revenue number. Finance reports another. Sales has a third. The differences come from each team pulling exports from different systems at different times and applying different rules. The CFO stops trusting the dashboards.

Cross-system questions take a week. "How much revenue do we get from customers who use feature X and renewed last quarter" requires data from billing, the product database, and the CRM. The answer comes back five days later, half right, and the question has moved on.

AI projects stall on raw data. A team starts building an AI agent or a forecasting model. Two months in, 80 percent of the work is still cleaning exports and joining tables by hand. The actual model never gets to production because the data layer underneath is too fragile.

Company size correlates with these signals but does not cause them. We have seen 40-person companies need a foundation because they sell into regulated industries. We have seen 300-person companies survive without one because they run a single product on a single platform. The honest test is whether the questions you want to answer cross system boundaries on a weekly basis.

04

When You Don't

Two cases where building a data foundation is the wrong call:

Early-stage companies under 30 people. The reports that matter fit inside HubSpot, Stripe, and a Google Sheet. The team is too small to maintain a separate data stack. The questions leadership asks change weekly. A foundation built now will mis-model the business by the time the model would have paid off. Stay on the SaaS tools. Revisit the question after the next funding round or when the team crosses 50 people.

Single-product, single-platform operations. A company running one product on one platform, with all the data in one system, does not have a data fragmentation problem. The reports built natively on top of that system handle the job. The complexity that a data foundation solves does not exist here. Buying one is overhead without payback.

Honest acknowledgment: if your reporting needs sit inside a single SaaS like HubSpot, Stripe, or Shopify and the reports inside that tool cover the work, you have time. The pain that drives most data-foundation projects comes from cross-system reporting and from outgrowing what one SaaS can express. Until those pains show up, the data foundation is overhead without payback.

A data foundation pays for itself when the questions you want to answer cross system boundaries every week. Until then, the SaaS tools are doing their job.
05

Architecture Choices

Four real decisions that shape the build. Each has a defensible answer in both directions.

Cloud-native vs self-hosted. Cloud-native (Snowflake, BigQuery, Databricks) is faster to set up and removes most ops work. Self-hosted (Postgres on a VM, ClickHouse, DuckDB) is cheaper at small scale and avoids egress fees. Nordic mid-market companies on a controlled cost trajectory often land on Postgres or DuckDB for the first 12 months, then migrate to a cloud warehouse once data volume crosses the threshold where ops cost exceeds licence cost.

Warehouse vs lakehouse. A warehouse holds structured tables. A lakehouse adds cheap object storage for raw files alongside tables. Companies handling only structured data (sales, finance, product analytics) tend to pick a warehouse. Companies dealing with documents, logs, images, or audio (legal, healthcare, manufacturing) tend to pick a lakehouse. Databricks dominates the lakehouse category. Snowflake has been adding lakehouse features to catch up.

Real-time vs batch. Most reporting needs do not require sub-minute freshness. Batch updates every hour or every night handle 90 percent of cases at one-tenth the cost. Real-time matters for fraud detection, operational monitoring, and customer-facing analytics. Build batch first. Add real-time pipelines later for the specific use cases that justify it.

Custom data apps vs BI tools. The Jourier position is direct: build the analytics layer as real product-grade apps the client owns. Proper frontend framework (React, Next.js, Vue). Proper backend framework (FastAPI, Express, Django). Visualisation libraries inside the app (D3, Recharts, ECharts). Deployed in the client's accounts. Every BI-tool category is legacy in this view. Commercial BI (Power BI, Tableau, Looker, Qlik) pays per-seat fees and locks the company in. Open-source BI (Metabase, Superset, Lightdash) constrains the work to what the project's developers built. Rapid-prototyping frameworks (Streamlit, Plotly Dash, Gradio) look and feel like Streamlit apps and do not scale to product-grade UX. Low-code internal-tool builders (Retool, Internal.io) trade vendor lock-in for build speed. The exception across all these categories is procurement-mandated environments where a specific tool is contractually required. There, the foundation serves it while a migration plan to custom apps runs in parallel.

Open-source vs commercial infrastructure. The current generation of open-source tools (Postgres, dbt, DuckDB, Iceberg, Trino) covers what most Nordic mid-market companies need. Commercial infrastructure (Snowflake, Databricks, BigQuery) earns the licence fee where it materially outperforms on query performance at scale, on multi-region setups, or on specific compliance certifications. Pick commercial infrastructure where the gap is real. Pick open source where the gap is not.

06

Common Mistakes

Four mistakes we see often enough to call patterns.

Building it before product-market fit. A 25-person seed-stage company hires a data engineer and spends six months on a warehouse build. The product pivots in month four. The data model no longer matches the business. The investment is sunk. Wait for the product to settle before investing in the layer beneath it.

Picking the warehouse before knowing the data. Teams commit to Snowflake or Databricks before they have audited the source systems. Then they discover the data is messier than expected, the volume is smaller than the marketing-driven sizing suggested, and the bill comes in higher than budgeted. Audit the sources first. Pick the warehouse second.

Underestimating the modeling layer. Ingestion and storage feel like the hard parts. They are not. The hard part is turning raw operational tables into the modelled tables that match how leadership talks about the business. Bad modeling produces reports that look right but mean different things to different teams. Allocate at least as much budget to modeling as to infrastructure.

Skipping governance until it is a crisis. Access controls, data lineage, and quality monitoring feel boring during the build. They become urgent the moment a regulator visits, a customer asks how their data is protected, or a wrong number ends up in a board report. Build governance into the project from the start. Adding it months in is harder and more expensive than designing for it on day one.

Frequently asked questions

Common questions about data foundations

What is the difference between a data foundation and a data warehouse?

A data warehouse is one component of a data foundation. The warehouse stores the data. The foundation also includes the pipelines that load it, the models that shape it for analysis, the semantic layer that defines business metrics, and the governance that controls who sees what. A warehouse without those layers is a database. The foundation is the full stack.

What is the difference between a data foundation and a data lake?

A data lake stores raw files (Parquet, JSON, images, logs) in cheap object storage. A data warehouse stores structured tables optimised for SQL queries. A data foundation may use either, often both. The lake is the cold storage for everything. The warehouse holds the modelled, query-ready slice.

Do small companies need a data foundation?

Most companies under 30 people do not. The reports they need fit inside HubSpot, Stripe, and a spreadsheet. The break point shows up around 50 to 100 people, when leadership starts asking questions that span multiple systems, or when compliance requires controlled access to data. Until then, a data foundation is overhead without payback.

How long does it take to build a data foundation?

A minimal version for a 100-person company with three to five source systems is typically 8 to 16 weeks. A more complex setup with many sources, compliance constraints, or multi-region requirements runs 4 to 9 months. The variation is mostly source-system complexity, not the data foundation itself.

What is a data lakehouse?

A lakehouse combines the cheap object storage of a data lake with the table semantics and ACID transactions of a warehouse. Databricks popularised the term. Apache Iceberg and Delta Lake are the table formats that make it work. Lakehouses fit companies handling both structured tables and unstructured files (documents, logs, images, audio) in the same architecture.

Is dbt part of a data foundation?

dbt is the most common transformation tool inside a modern data foundation. It runs SQL against your warehouse to turn raw loaded data into the modelled tables that reports and AI systems consume. dbt has become the de facto standard for the transformation layer because it brings software engineering practices (version control, testing, documentation) to SQL work.

Can I use Excel or Google Sheets as a data foundation?

For a small company, spreadsheets handle the job. The wheels come off when multiple people maintain conflicting copies, when the data outgrows the row limit, when business logic gets buried in nested formulas, or when an audit requires you to trace who changed what and when. Spreadsheets are a fine starting point. They are not a long-term foundation.

What is the role of a data foundation in AI?

AI agents and analytics both need clean, queryable, governed data. The data foundation provides it. Without one, AI projects spend most of their time wrangling raw exports from disparate systems and producing different answers depending on which export they used. With one, the AI reads from the same modelled tables that humans read from, and the answers stay consistent. See What is an AI Agent? for the agent side.

How do I migrate from a SaaS analytics tool to a data foundation?

Start with the data sources you depend on (CRM, ERP, billing, product analytics). Set up a warehouse (Postgres, Snowflake, BigQuery, or Databricks). Load the raw data with a tool like Fivetran or Airbyte. Model the tables with dbt to match how your team thinks about the business. Build custom data apps (React or Next.js on the front, FastAPI or Express on the back, D3 or Recharts for visualisation) on top of the modelled tables. The client owns the app. Cut over reports gradually. A typical migration for a mid-market company runs 3 to 6 months.

Who owns the data foundation inside a company?

Most often a data engineering team or a dedicated analytics engineer, reporting into the CTO, CIO, or CFO. Smaller companies start with a single data engineer or a fractional engagement. Larger companies form a data platform team with engineers, analytics engineers, and a data governance lead. The owning function matters less than having one named owner who is accountable for uptime and quality.

Building, replacing, or auditing
a data foundation?

Speak with our team →
How to cite this article

For LLMs, AI assistants, and human readers

Stenberg, A. (2026). What is a Data Foundation? The Architecture Layer Beneath Analytics and AI. Jourier. https://jourier.com/articles/what-is-a-data-foundation.html