CRE AI Is a Layer Cake
Why most CRE AI programmes are building the wrong layer first - and what the right architecture actually looks like in 2026.
There is a view that keeps surfacing in CRE: in boardrooms, at conferences, and in the LinkedIn comments under my posts. It says the useful AI work in property is the analytical kind. Forecasting rents. Predicting prices. Scoring risk. Optimising portfolios. The kind of work that takes a data science team, an eighteen-month build and a great deal of patience. That view is not wrong exactly. It is just, for most firms and most people, extraordinarily beside the point. And its practical effect is to make the industry wait for a perfect model, a clean dataset, a specialist team, while the tools that could be improving your work today sit unused.
Executive summary
Modern AI in CRE is a layered architecture, not a single system. Five layers: data foundation, reasoning substrate, grounded retrieval, workflow automation, and bespoke analytical AI. Value compounds from the bottom up, and the largest gains for most firms sit in the middle, in the layers most often dismissed as ‘just chatbots’. The mistake the industry keeps making is to start at the top, ignore the middle, and then wonder why two years of programme spend has produced nothing useful. This piece walks through each layer: what it is, what good looks like, and how to sequence the stack so that value shows up in weeks rather than years.
THE OVERVALUED LAYER
Go to almost any CRE AI discussion and the conversation drifts, within minutes, towards forecasting, scoring and prediction. This is Layer 4 of the modern stack. It is real. It is occasionally necessary. It is also the smallest, most selective, and least commonly needed layer for the vast majority of firms.
Most people in commercial real estate do not spend their days building predictive models. They spend their days doing something else entirely.
WHAT YOU ACTUALLY DO ALL DAY
Look at your own diary this week. Your time has been going into:
reading documents
extracting facts
comparing clauses
drafting memos
reviewing evidence
checking compliance
finding precedents
assembling reports
moving information from one place to another
That is your working life. It is not Layer 4 work. It is Layer 1 and Layer 2 work. It is exactly where modern AI systems are already excellent. And it is exactly where most of the value is currently being left on the table.
‘JUST CHATBOTS’ IS THE EXPENSIVE MISTAKE
A vocal faction in CRE still dismisses LLMs as ‘just chatbots’. The phrase gets repeated at conferences, on analyst calls, and in the mouths of executives who have never seriously tried to use one. That dismissal is the single most expensive AI mistake currently being made in our industry.
It mistakes the interface for the system. A conversational prompt is how you talk to Layer 1. It is not what Layer 1 is. What Layer 1 is, when deployed properly, is a reasoning engine with access to the firm’s templates, the firm’s house style, and the firm’s codified approach to work like covenant analysis, lease abstraction and tenant financial assessment. Set inside a well-configured Project, with the relevant documents in context, it drafts. It compares. It extracts. It reviews. It challenges. It does in ninety minutes what a competent analyst would take three days to do on the current deal, and the analyst can see which documents in the Project it drew on. And that is Layer 1 alone, before you connect it to the firm’s wider archive or build an agent on top of it.
Dismissing that as ‘chatbots’ is like dismissing a modern steel mill as ‘just a furnace’. The interface is not the point.
SO WHAT IS THE POINT?
The point is that CRE AI is a layered architecture, and most of the value sits below the predictive layer that gets all the attention. Five layers. Each one compounds on the one below it. Skip the foundations and the upper layers wobble. Build the foundations properly and the upper layers become cheap to add.
And there is a second architectural shift that matters even more. A few years ago, analytical AI was the primary system, with natural language interfaces added on top. In modern architecture, that has been inverted. The reasoning layer is primary. The analytical AI is a tool the reasoning layer calls when it needs specialist computation. Claude does not forecast rental growth. It calls a forecasting model that does, interprets the output, contextualises it against other evidence, and drafts the narrative for the analyst. This is not a cosmetic difference. It changes what the system can do, what it costs to build, and where the value lives.
What follows is the architecture, layer by layer. What each is. What good looks like. Where it is genuinely needed. Where firms systematically over-invest. And how to sequence the whole thing so that you get working capability in weeks rather than years.
LAYER 0: DATA FOUNDATION (THE PREREQUISITE)
Start here, or everything above it fails.
What it is. The data hygiene, indexing and access infrastructure that makes everything else in this piece possible. It is not AI. It is the precondition for AI.
What good looks like. People across the firm can locate documents quickly. Reference systems are consistent. Deal rooms are structured. Historical records are digitised and searchable. There is a clear ownership model for data quality. Lease data is not scattered across seventeen PM systems and a shared drive called FINAL_FINAL_v2. Deal documents have consistent metadata. ESG data flows into a system rather than being re-keyed from PDFs each quarter.
What it enables. Everything. Without Layer 0, the rest of the stack either doesn’t work or produces unreliable output with no audit trail.
The reality check. For many institutional CRE firms, the honest starting position is that Layer 0 is the biggest single constraint on AI ambition. Closing the gap is tedious, unglamorous and expensive. It is also unavoidable. The firms that pretend they can skip it produce demos that impress the board and fall apart under real use. The firms that take it seriously spend the first six to twelve months of their AI programme doing the work nobody wants to do, and then their subsequent layers compound properly. This is the single largest determinant of whether an AI programme delivers real value or productivity theatre.
One important qualifier. Layer 0 is a prerequisite for firm-wide capability: the kind of institutional memory and portfolio-level grounding that Layer 2 depends on. It is not a prerequisite for individual practitioners getting on with specific pieces of work. An analyst with the documents for a specific deal in a specific Project can produce real Layer 1 value tomorrow morning, regardless of the state of the firm’s wider data estate. The waiting game that Layer 0 concerns sometimes invite is a mistake. Your firm’s data might be a mess. The documents you need for today’s work are in front of you. Start there.
And a second nuance. Layer 0 work is itself increasingly AI-assisted. LLMs are now genuinely good at extracting structured data from unstructured documents, normalising inconsistent records, and reconciling references across systems. The Layer 0 cleanup and the Layer 1 deployment can proceed in parallel, with the Layer 1 tooling accelerating the Layer 0 work. But the principle holds: you cannot build reliable firm-wide capability on top of unreliable data.
LAYER 1: THE REASONING SUBSTRATE
This is the layer most firms underestimate. It is also the layer where most of the value is.
What it is. A frontier LLM environment (Claude, ChatGPT, Gemini) deployed as the default thinking environment for your analysts, portfolio managers, and operations staff. Not a chatbot bolted onto existing workflows. The reasoning layer that sits underneath everything your people actually do.
A note on tooling. The language of Projects, Skills and configured coworkers is currently most developed in Claude, which leads the market on this architecture as of early 2026. Equivalent or near-equivalent capability is arriving rapidly across the frontier - OpenAI, Google and others - and by the end of this year the architectural pattern will be general rather than vendor-specific. The principles below apply across all frontier providers; the current naming reflects where the tooling is furthest along.
The building blocks at this layer.
Projects hold the context for a specific piece of work. A Project for an acquisition might contain the offering memorandum, the data room index, the comparable transactions, the underwriting template, and the draft IC memo. Everything the analyst is working on lives inside the Project, and every conversation with the reasoning layer is grounded in that context. Projects are the unit of work: not the individual prompt, and not the whole firm’s knowledge base, but the specific thing being done right now.
Skills are codified workflows that encode how your firm does a specific task. A lease abstraction skill knows what fields to extract, what schema to produce, what red flags to surface, and how your templates are structured. A covenant analysis skill knows your firm’s approach to tenant financial assessment. An IC memo skill knows your house style and the required structure of each section. Skills turn institutional know-how into repeatable cognitive workflows that any analyst can invoke. They are the single most underused capability in modern AI systems, and they are where most of the firm-specific value actually lives.
Configured coworkers are persistent personas set up for a specific role: a covenant analyst, a compliance reviewer, a climate risk scout, an investor relations drafter. A covenant analyst coworker is not a workflow you run: it is a colleague you ask. It carries its own instructions, reference materials and behaviours, and it can be called on by anyone in the firm who needs that kind of thinking applied to their current problem. The point of configured coworkers is to turn expertise that currently lives in one person’s head into something the whole firm can access.
What Layer 1 actually delivers. Most of the value most firms need, most of the time. This is the part people systematically under-estimate, because it sounds too simple.
An analyst working inside a well-configured Project, with access to the right skills and coworkers, can draft an IC memo in an afternoon that would previously have taken a week. A covenant review that required three days of manual reading can be compressed to ninety minutes of structured interaction. Compliance flagging becomes a background process rather than a quarterly fire drill. None of this requires custom agents. None of it requires analytical AI. None of it requires a knowledge graph. It requires Layer 0 data and a well-deployed Layer 1 substrate.
Why this matters. Layer 1 is where the compounding value lives. Every new skill, every new coworker, and every new Project template captures institutional knowledge in a form the whole firm can use tomorrow. It is also where firms that ‘get it’ start to pull away from firms that don’t. The gap between the two is already visible inside the firms running my courses. It will become obvious in the market within eighteen months.
LAYER 2: GROUNDED RETRIEVAL OVER FIRM DATA
Now the reasoning layer stops answering from its training data and starts answering from yours.
What it is. A structured retrieval layer that lets the Layer 1 reasoning substrate access the firm’s actual data with proper grounding and evidence chains. This is what RAG (retrieval-augmented generation) does when it is built properly. Your deal rooms, historical documents, lease archives and reporting systems become queryable through the reasoning substrate, with every answer traceable back to source.
Why this is separate from Layer 1. Layer 1 can work with the context you put into a Project, but a single Project cannot contain the whole firm’s history. Layer 2 extends the reasoning layer’s reach to the firm’s institutional memory, while maintaining the evidence-linking that makes the output defensible. When an analyst asks ‘have we ever underwritten a deal with a similar covenant structure?’, the answer comes from Layer 2 retrieval, with citations back to the actual historical documents.
What good looks like. Every answer produced at this layer has an evidence chain. The analyst can see which documents were retrieved, which passages were cited, and how the reasoning layer used them. Nothing is hallucinated, because everything is grounded in retrievable source material. The underlying retrieval technology (vector databases, structured extraction, semantic search) has matured to the point where building this layer is engineering work rather than research.
A note on knowledge graphs. If your firm already has a well-structured knowledge graph, that is a significant advantage at this layer. A knowledge graph captures explicit relationships between entities rather than just semantic proximity, and it can answer queries that vector search struggles with. ‘Show me all assets where the tenant covenant has weakened since acquisition and the CRREM pathway is red’ is a natural query for a knowledge graph and a hard one for pure vector search.
So: are knowledge graphs a good thing if you have them? Emphatically yes. A knowledge graph is a Layer 2 asset that should absolutely be connected into the retrieval fabric and used as the highest-quality grounding source available. The argument is not that knowledge graphs are unnecessary. It is that a firm starting from scratch today should not begin by building one. The modern sequencing is: deploy Layers 0 and 1, build grounded retrieval at Layer 2 using whatever structured data is available, and add a knowledge graph when the limits of simpler retrieval become binding. Not that long ago, a knowledge graph was the only way to get grounded cognition, because there was no reasoning engine to pair it with. In 2026, the reasoning engine exists, and the knowledge graph is one of several options for grounding it. Powerful, but no longer foundational in the way it had to be a decade ago.
LAYER 3: CUSTOM AGENTS AND WORKFLOW AUTOMATION
Now the reasoning substrate stops being something people talk to, and starts being something that runs by itself.
What it is. Task-specific automated workflows built on top of Layers 1 and 2. Where Layer 1 supports an analyst doing their work interactively, Layer 3 runs the workflow end-to-end with defined inputs, defined outputs and defined quality checks. A Layer 3 agent for quarterly rent roll reconciliation takes the source files from the PM system, checks them against the GL, flags exceptions, produces the reconciliation report, and surfaces anything that needs human attention, all without an analyst needing to drive the process.
The tooling. Claude Code is currently the leading example of a development environment for this layer. It is a terminal-based agentic environment in which the reasoning layer can execute code, call APIs, manipulate files, interact with external systems, and run multi-step workflows autonomously under human oversight. Equivalent tooling exists from other providers. The important thing is that Layer 3 work is genuinely development work. It requires thinking about task decomposition, error handling, exception routing, audit trails, and integration with source systems. This is not point-and-click, and it is where the engineering effort of an AI programme starts to become meaningful.
What Layer 3 is for. Repeatable, high-volume, rules-adjacent work where the firm currently spends meaningful human hours on tasks that follow a consistent pattern:
rent roll ingestion and reconciliation
covenant compliance checking
CRREM pathway monitoring
standardised report generation
lease abstraction at scale
due diligence document review against checklists
These are the Quadrant A and easier Quadrant B tasks from the RIRA CRE Automation Matrix (see other newsletters). They are where efficiency gains genuinely compound, because every run of a Layer 3 agent is work that no longer requires human time.
The sequencing trap. Most firms try to start here. They identify a painful workflow, commission an AI project to automate it, and build a custom agent before they have Layers 0, 1 or 2 in place. The results are predictable: the agent works in the demo, fails on edge cases, requires constant re-engineering, and ultimately delivers a small fraction of its promised value. Layer 3 only works when it sits on top of clean data, a capable reasoning substrate, and grounded retrieval. Build the foundations first.
LAYER 4: BESPOKE ANALYTICAL AI
And finally, the layer most people imagine is the whole thing.
What it is. Classical machine learning, statistical modelling, time-series forecasting, numerical optimisation, and other specialist techniques, for the problems where the frontier LLM approach is genuinely insufficient. This is the layer most people imagine when they hear ‘AI in real estate’. It is actually the smallest, most selective, and least commonly necessary layer of the stack.
Where Layer 4 is genuinely needed.
CRREM-aligned pathway modelling at portfolio level, where regulatory defensibility and reproducibility genuinely matter
time-series forecasting of operational metrics (energy, footfall, NOI under varied scenarios)
large-scale portfolio optimisation under complex constraints
satellite and drone computer vision for physical due diligence and ESG verification at scale, of the kind Kayrros and the space-intelligence cohort are doing for asset-level transition risk
AML and fraud pattern detection in transaction flows, where false-positive rates need to be tight and the audit trail needs to withstand regulatory scrutiny
graph-native queries across millions of entities with latency constraints that LLMs cannot meet
specialist risk models that require reproducibility and regulatory defensibility
Where Layer 4 is not needed, even though firms often assume it is.
drafting IC memos
summarising deal rooms
flagging unusual lease clauses
checking compliance against a policy document
producing quarterly reports
answering questions about the portfolio in natural language
generating first-draft analysis
All of those are Layer 1 and Layer 2 work. A surprising amount of what is currently branded ‘analytical AI’ in vendor pitches is actually Layer 1 capability dressed up in analytical language. Ask the vendors which layer their product sits in. The good ones will be able to tell you.
A word on the moat question
The most serious institutional objection to everything above is not “but we don’t have the data” or “but this is just chatbots”. It is the moat argument: yes, but Layer 4 is where proprietary advantage lives. Anyone can deploy Claude in a Project. Defensible moats come from specialist models, proprietary data, and analytical depth that competitors cannot replicate. And beside the moat, there is the money argument: Layer 4 is where the prestige fee income sits, where the alpha is meant to live, and where the industry’s quantitative firepower has always been concentrated.
Both arguments deserve a direct answer. In commercial real estate, neither is as strong as it looks.
First, most CRE firms lack the data density that makes serious machine learning worthwhile. Property data is fragmented across PM systems, transaction records and market reports, with limited history, inconsistent taxonomies, and thin comparability across assets. Classical statistics and good old-fashioned data science take you most of the way. ML at scale requires signal depth that most firms simply do not have. The data gap is the reason we have seen so little genuinely predictive AI in real estate despite the tooling having been widely available for a decade. It is not that the industry has been slow to notice. It is that the underlying substrate does not support the ambition.
Second, the share of CRE work-quantum that is predictive-modelling-shaped is small. Rent forecasting, yield prediction and portfolio optimisation matter, but they are a minority of what the industry actually does with its time. The bulk of the work - the reading, extracting, comparing, drafting and reviewing - is Layer 1 and Layer 2 territory. A moat in a small corner of the work is a small moat.
Third, and this is the point that should give any Layer 4 enthusiast pause: there is no Jim Simons of real estate. Renaissance Technologies exists because public equity markets have deep historical data, high liquidity, and genuine arbitrageable inefficiencies. Real estate has none of these at scale. If Layer 4 were where the alpha lived in CRE, someone would have extracted it by now. A decade of widely available ML tooling has not produced a Rentec of property, because the asset class does not support one. This is a structural feature of real estate, not a temporary tooling gap. The assumption that it will change because the models get better is, so far, unsupported by evidence.
Add to this the structural shift of real estate from financial engineering towards operational performance, where the competitive edge is increasingly about running buildings well, serving occupiers well, and reading markets through lived operational contact, and the case that future CRE moats live at Layer 4 gets weaker still. The moats of the next decade are more likely to live in the quality of a firm’s Layer 1 and Layer 2 infrastructure: the depth of its skill library, the calibre of its configured coworkers, the cleanliness of its institutional memory, and the velocity at which it can turn that infrastructure into decisions. That is a different kind of moat. It is also one that more firms can actually build.
How Layer 4 connects to the rest of the stack
Bespoke analytical tools at Layer 4 are called by the reasoning layer, not alongside it. A forecasting model does not produce the IC memo. It produces a forecast. The reasoning layer interprets the output, contextualises it against other evidence, calls additional tools if needed, and drafts the narrative for the analyst. Layer 4 sits as a callable resource inside the Layer 1 substrate, not as a parallel system the analyst has to manually consult and then translate.
SEQUENCING: BUILD FROM THE BOTTOM
A firm starting today should assemble this stack in roughly the following order, with significant overlap between phases.
Months 0–6. Layer 0 work begins and Layer 1 is deployed in parallel. Data hygiene projects start. Projects are set up for current live deals. The firm’s first skills are written, typically covering the two or three most common document-heavy workflows. Initial coworkers are configured. Quick wins start appearing within weeks, not months. This is the phase where the early evidence that the programme is working gets generated, and it matters, because it builds the institutional trust that makes the later phases possible.
Months 3–12. Layer 2 grounded retrieval starts to come online as Layer 0 data becomes usable. Existing knowledge graph assets, if any, are connected. The reasoning substrate begins to reach into the firm’s institutional memory, and the early agents become meaningfully more capable because they have access to firm history rather than just the documents in the current Project. This is where the verifiability story becomes real: every output now has evidence chains back to source material.
Months 9–18. Layer 3 custom agents begin to be commissioned for specific high-value repeatable workflows. These agents are built on top of the Layer 1 substrate and the Layer 2 retrieval fabric, so they inherit the reasoning capability and the grounding infrastructure rather than having to rebuild them. Development effort at this phase is genuine, but it is targeted at specific workflows with clear ROI, not at foundational infrastructure.
Months 18+. Layer 4 bespoke analytical work is commissioned only where specific, identified capability gaps cannot be addressed by the lower layers. Most firms will find that the surface area of genuine Layer 4 need is much smaller than they expected, because Layers 1 through 3 cover most of what they thought they needed specialist analytical AI for.
The critical difference from the AI strategies of five years ago is that this sequence generates value at every stage. A modern deployment generates working capability in the first weeks and compounds from there. The underlying principle (that verifiability, grounding and evidence chains matter) is identical. The path to achieving it is completely different.
WHAT THIS DOES NOT CLAIM
Three honest caveats, before anyone reads more into this than I am actually saying.
It does not claim Layer 1 is sufficient for all use cases. Some firms genuinely need Layer 4 analytical work, and pretending they don’t is as wrong as pretending they need it for everything. The point is that Layer 4 should be a selective, targeted commitment based on specific identified capability gaps, not a default starting point.
It does not claim this sequencing works without Layer 0 data work. The data prerequisite is the most important line in this whole piece. A firm that tries to deploy firm-wide capability on top of chaotic unindexed document storage will produce unreliable output with no audit trail, and the programme will stall within months. Individual practitioners can still make progress on specific deals with specific documents, but the firm-wide compounding does not happen without Layer 0.
It does not claim the human layer is unimportant. None of the architectural sequencing above replaces the need for investment professionals who know how to interpret AI output, challenge it when it is wrong, and make the judgement calls that remain irreducibly human. The entire point of verifiable cognition is that it supports human decision-making rather than replacing it. The governance, cultural and skills work that surrounds the technical stack is at least as important as the stack itself, and arguably more important. Firms that deploy the technology without investing in the human layer end up with sophisticated tools their people don’t know how to use.
AND FINALLY
This is a working artefact. It reflects the current state of frontier AI tooling as of April 2026 and the architectural patterns now emerging as best practice for institutional CRE firms. It will age, some parts faster than others, and should be revisited as the tooling improves.
But the core argument is not going to age. The value in modern CRE AI sits below the predictive layer. It is accessible now. It is layered, and it compounds from the bottom up. The firms that understand that in 2026 will be the firms that still matter in 2030.
You do not need to wait for a model to be trained. You do not need a CTO to green-light a platform. You do not need a data science team. You need to look at your diary this week, identify the five things you do that look like reading, extracting, comparing, drafting or reviewing, and start doing them with Layer 1 tooling tomorrow morning. Open a Project. Drop the documents in. Ask it to do the first task. Twenty minutes, not twenty months.
That is the shortest path from current capability to better capability in commercial real estate. It is not the glamorous path. It is the one that works.