Fundamental raises $255M to build a foundation model for structured data
Fundamental has come out of stealth with $255 million in total funding, including a $225 million Series A, at a reported $1.2 billion valuation. Its pitch is specific enough to be interesting: a foundation model for structured data, built for very la...
Fundamental raises $255 million to sell enterprises a deterministic AI model for tables
Fundamental has come out of stealth with $255 million in total funding, including a $225 million Series A, at a reported $1.2 billion valuation. Its pitch is specific enough to be interesting: a foundation model for structured data, built for very large tables, with deterministic outputs and a non-transformer architecture.
The model is called Nexus. Fundamental says it has already signed seven-figure deals with Fortune 100 companies and has an AWS partnership to deploy from existing customer environments. Those are big claims. They’re at least aimed at a real problem.
LLMs work well on text because transformers are built for sequences. Enterprise analytics usually isn’t.
Why this matters
Anyone who works with production data systems already knows where the hard, expensive work lives. Core business decisions still run on Snowflake, BigQuery, S3 + Parquet, operational PostgreSQL, DynamoDB, and a lot of ugly CSV-shaped residue. Fraud scoring, churn models, demand forecasting, pricing, credit risk, supply chain planning. Structured data runs the business.
Standard LLMs can summarize a dashboard or write some SQL. They’re far less convincing when the job is to analyze billions of rows, stay reproducible, and avoid hallucinating.
That’s the hole Fundamental is trying to fill.
Its bet is straightforward: enterprises want models built around the shape of business data, not text-first systems wrapped around warehouses and BI tools. That sounds obvious. The market has spent the past two years acting like generic LLMs could stretch to cover almost everything.
They can’t. Not cleanly.
The technical pitch
Fundamental calls Nexus a Large Tabular Model, or LTM. The label may or may not stick. The architecture choice matters more.
Transformers process sequences. Tables are not naturally sequences. You can serialize rows into text and feed them to an LLM, and people do that all the time for demos. It breaks down on scale, repeatability, and cost.
A few reasons stand out:
- Context windows are the wrong bottleneck. Even large-context models can’t ingest enterprise-scale tabular data end to end in any useful way.
- Tokenization flattens the structure. Column types, missing values, categorical relationships, and numeric distributions get turned into a format the model doesn’t handle especially well.
- Inference is often stochastic. Fine for chat. Bad for regulated analytics, audits, and incident review.
Fundamental says Nexus is deterministic and avoids transformer architecture entirely. It hasn’t published the architecture, so some of this is inference. But the likely design space is familiar.
Good tabular systems usually win by being feature-centric, not sequence-centric. That means exploiting column statistics, modeling nonlinear interactions efficiently, handling missing values well, and preserving schema semantics instead of burying them under token streams. Traditional tabular winners like XGBoost, LightGBM, and CatBoost still dominate for solid reasons. They fit the medium.
A serious tabular foundation model probably borrows more from that world than from chat model design. It may also borrow from work like TabPFN, which uses learned priors across many tabular tasks to adapt quickly to new datasets.
That’s the interesting part. You get some transfer benefits associated with foundation models without forcing structured data through a text pipeline that was never a great fit.
Why determinism matters
A lot of AI product pitches treat determinism as a nice extra. In analytics, it’s near the center.
If the same query returns different outputs across runs, you have a governance problem immediately. Reproducibility matters for audits, compliance, debugging, and basic trust. If a fraud review, underwriting decision, or internal KPI shifts because inference sampled differently, finance, legal, and regulators won’t care that the model felt creative that day.
That’s why the deterministic claim matters more than the “foundation model” label.
For technical teams, deterministic behavior changes routine work in very practical ways:
- rerunning historical analyses without output drift
- comparing model versions cleanly in
MLflowor similar systems - stabilizing explainability methods like SHAP or permutation importance
- tracing incidents when metrics move and nobody trusts the first answer
Classic ML pipelines already solve some of this. Often, they solve it well. Fundamental has to show that a pretrained tabular model can keep those guarantees while cutting down the amount of custom model building teams have to do.
That’s a real technical bar.
The scale story
The biggest question is how Nexus handles very large datasets in practice.
No model “understands” billions of rows by swallowing them whole. If Nexus works at that scale, it probably depends on a hybrid execution pattern. Warehouses and data engines do joins, filters, aggregations, windowing, and scans. The model handles prediction, representation learning, and task generalization over reduced or summarized views.
So pushdown matters. So do sketching methods, quantile summaries, stratified sampling, and compressed representations for high-cardinality categoricals. Ignore those constraints and you don’t have a foundation model for enterprise analytics. You have an expensive choke point.
That also explains why the AWS partnership matters. Not because the phrase “strategic partnership” carries much weight by itself. Usually it doesn’t. Deployment near existing customer data matters. If enterprises can run Nexus from infrastructure they already trust, inside familiar access controls and network boundaries, procurement gets easier and security objections narrow.
If the product depends on moving broad swaths of data into a vendor-controlled black box, adoption gets harder fast.
Where it could break down
There’s a reason tabular AI is still messy after decades of work. Real enterprise data is worse than most vendor decks admit.
Cross-table reasoning
Flat tabular benchmarks are one thing. Production systems usually involve entity graphs, messy joins, slowly changing dimensions, and temporal logic. A model that does well on single-table classification can still fall apart when business questions span events, users, accounts, products, and time windows.
Temporal leakage
Tabular modeling gets dangerous quickly once time enters the picture. Feature generation, backtesting, rolling windows, and leakage prevention are where many automated systems embarrass themselves. If Nexus handles this well, that matters. If it papers over it, that’s a real risk.
Schema drift
Enterprise schemas change constantly. Columns get renamed, recoded, deprecated, or repurposed. Categorical vocabularies drift. Null patterns shift. A tabular foundation model needs strong schema inference and decent fallback behavior. Otherwise every deployment turns back into feature engineering cleanup.
Calibration
AUC gets a lot of attention. Operators care about decision quality. In fraud, underwriting, and demand planning, calibrated probabilities often matter more than leaderboard-style accuracy gains. If Fundamental wants to replace or consolidate existing systems, calibration matters.
Who should pay attention
If Fundamental’s claims hold up, pressure lands in three places.
AutoML and tabular ML vendors should pay attention first. DataRobot, H2O, and cloud tabular stacks like Vertex AI’s tabular tooling all sell some version of strong predictive performance without rebuilding everything from scratch. A pretrained, deterministic model for warehouse-scale data would be a direct challenge.
Data platforms are next. Snowflake, Databricks, BigQuery, and AWS all want to be the place where data stays and intelligence runs nearby. Nexus is either a useful partner or a feature request waiting to happen.
BI and analytics workflows could shift too. A lot of teams currently glue together SQL models, dashboard logic, notebooks, and one-off ML pipelines. If a tabular foundation model can cover a wider range of tasks with consistent outputs and better governance hooks, those stacks get simpler. Or at least differently complicated.
Open source won’t stay still either. Expect stronger combinations of LightGBM-style baselines, pretrained tabular encoders, and better priors trained on public corpora such as OpenML plus synthetic data. That won’t erase enterprise distribution advantages, but it will narrow the novelty gap.
What technical buyers should ask
The company has funding, customer logos, and a polished pitch. That doesn’t answer the implementation questions that matter.
Ask these:
- Does it run inside your VPC or through private connectivity like PrivateLink?
- Can it keep data in place across
S3,Snowflake,BigQuery, or operational databases? - What’s deterministic in practice? Model outputs, feature attribution, training runs, or all of the above?
- How does it handle joins, temporal boundaries, and leakage control?
- What are the baselines against
XGBoost,LightGBM, and modern tabular deep learning models? - How does it perform on calibration, not just ranking metrics?
- What’s the audit trail for every prediction, data snapshot, and model version?
- How painful is schema evolution?
If those answers are fuzzy, the product is still early no matter how large the round is.
Fundamental has picked a better target than most AI startups. Structured data is where a lot of enterprise value sits, and the transformer-heavy wave has left an obvious gap there. The deterministic tabular angle is credible enough to take seriously. It also sounds much cleaner in a fundraising memo than it will in a production warehouse full of broken joins and drifting columns.
That gap between the pitch and the warehouse will decide whether Nexus becomes a real analytics platform or just a very well-funded thesis.
Useful next reads and implementation paths
If this topic connects to a real workflow, these links give you the service path, a proof point, and related articles worth reading next.
Turn data into forecasting, experimentation, dashboards, and decision support.
How a growth analytics platform reduced decision lag across teams.
Yann LeCun’s new company, AMI Labs, has raised $1.03 billion at a $3.5 billion pre-money valuation to build world models. That's a huge round for a company openly saying it won't chase near-term revenue, and it says a lot about where serious AI money...
Meta has reportedly hired Ruoming Pang, the Apple executive who led the team behind the company’s AI foundation models. Bloomberg reported it. At one level, this is another talent-war move. Zuckerberg has been pulling senior people from Apple, OpenAI...
Google has launched an AI Futures Fund, a new program for startups building AI products. The funding gets the headline, but the practical value is lower down the stack: cloud credits, access to DeepMind models, and support from Google’s research and ...