THE BLOG

Antony Slumbers Antony Slumbers

Where’s the New Business?

A framework for finding where AI creates genuinely new ventures - not just better processes

Over the past three months we’ve built a collection of frameworks for addressing how to leverage AI in CRE: RIRA for strategy, the CRE Automation Matrix for analysis, and the Prompting Framework for execution. Together they answer: how are we creating value, what kind of work is this, and how do we get it done?

But there’s a question they don’t force you to confront.

When participants work through RIRA’s ‘Imagine’ phase - with its three horizons of Efficiency (H1), Capability (H2), and Transformation (H3) - they almost universally gravitate towards H1 and H2. Faster lease abstracts. Better evidence-linked analysis. Smarter compliance checks. All valuable. All necessary.

But none of them are new businesses.

H1 makes your current work faster. H2 makes it better. But H3 builds something that doesn’t exist yet.

And almost nobody gets to H3, not because they lack ambition, but because H1 and H2 are natural extensions of work you already understand. You can picture your workflows done more efficiently. You can picture them done with better evidence. But picturing a product or service that serves a customer who doesn’t yet know they need it, that earns revenue from a budget line that doesn’t currently exist - that requires a fundamentally different kind of thinking. It’s uncomfortable. It implicates your own business model.

Which is exactly why you need a provocation tool to get there.

The H3 Provocation Framework
The H3 Provocation Framework is a set of five questions designed to surface where AI creates genuinely new businesses, products, or market structures in commercial real estate. It sits within RIRA’s Imagine phase as a dedicated instrument for Horizon 3 thinking. It doesn’t replace the “Faster taxis / Better taxis / Uber?” diagnostic, rather it provides the provocations that generate H3 hypotheses in the first place.

WHAT H3s LOOK LIKE IN CRE

Before I describe the five provocations, let me show you what they can produce. I’ve run the framework against four of the most common products and services in commercial real estate - diligence, advisory, occupancy, and portfolio management - to illustrate the kinds of hypotheses that emerge. These aren’t predictions. They’re outputs of a generative process, and the same process can be pointed at any activity across the CRE landscape.

Continuous Diligence.
Due diligence exists as an industry because maintaining a verified, current information state across hundreds of documents requires sustained human attention over weeks. That’s a cognitive and temporal constraint. If AI removes it - keeping an asset’s entire information state continuously current and verified - the question “shall we commission a diligence report?” stops making sense. You don’t commission what already exists. The value migrates from “who can do diligence well” (a cottage industry of lawyers, surveyors, and consultants charging £50–100k per transaction) to “who maintains the verified state.” That’s a data infrastructure play, not a professional services play. Transaction speed becomes a product. The seller offering a continuously-diligenced asset with verified evidence chains commands a premium because they’ve collapsed the buyer’s time and risk.

Evidence-Based Advisory.
AI is about to flood the market with plausible-sounding investment analysis. The supply of professional-looking memos will explode. What won’t explode is the supply of trustworthy analysis - recommendations backed by evidence chains, auditable reasoning, and explicit assumption registers. The value migrates from “trust me, I’m experienced” to “trust the evidence chain, and I’ll interpret what it means.” The human adviser’s role shifts from analyst-packager to interpreter of verified outputs and owner of the final judgement call - pure Quadrant D work (see The CRE Automation Matrix Framework for details). The advisory firm that builds verifiable decision support first doesn’t just have a better product - they’ve made every competitor’s narrative-based approach look unaccountable by comparison.

Occupier-as-a-Service.
What if occupiers didn’t buy space but bought outcomes - guaranteed workspace performance delivered against SLAs rather than lease terms? Air quality, temperature responsiveness, service levels - all continuously monitored, verified, and priced on delivery. The purist version breaks the institutional capital stack (lenders can’t underwrite an SLA the way they underwrite a 15-year FRI lease). But the realistic transition is a hybrid: a conventional lease providing the contracted income floor, with a verified performance premium on top. The hotel sector already proved this model - variable income becomes investable once you have enough verified performance data to make it predictable. PBSA and BTR are moving the same way. Over time, as the data layer matures and investors learn to underwrite operational capability rather than just tenant covenant, the proportions shift. Value migrates from “location plus specification” to “verified performance delivery.”

Portfolio Intelligence as Product.
Large portfolio owners sit on enormous operational data - tenant behaviour, maintenance patterns, energy consumption, lease events - used solely for internal management. If the intelligence derived from that data became a product - benchmarking services, predictive models, optimisation algorithms - you’ve created a revenue stream that doesn’t currently exist in CRE’s business model. The moat is proprietary data combined with verification infrastructure that’s hard to replicate. The REIT doesn’t just earn rental income - it earns knowledge income. And unlike rental income, knowledge products compound: the more data you accumulate, the better the models become, the more customers they attract.

The Common Pattern
Notice what these four share. None of them improve an existing workflow. They each create something that didn’t previously exist as a product or service. They each involve a value migration - a shift in where the premium concentrates. And they each emerged from asking a specific, slightly uncomfortable question about the current state of things.

I chose diligence, advisory, occupancy, and portfolio management because they’re familiar to almost everyone in the industry. But the framework isn’t limited to these. Point it at development appraisal, debt origination, tenant representation, facilities management, fund reporting - any activity where AI is about to change the underlying economics - and it will surface H3 candidates specific to that domain. The five provocations are a lens, not a list.

One more thing worth saying plainly: the timelines on these are uncertain. Continuous diligence may be five years away from mainstream adoption, or three, or seven. The point of the framework isn’t to predict when. It’s to discern the direction of travel - to see where value is migrating so you can start positioning now rather than reacting later. The firms that recognised flexible workspace was a structural shift, not a fad, had years of advantage over those that waited for proof. The same dynamic applies here, across a far wider set of CRE activities.

Those questions are the framework.

THE FIVE PROVOCATIONS

The five provocations follow a narrative arc:
What becomes free →
What falls apart →
What gets built →
Who wins →
Who pays.


Each builds on the last, and the sequence matters.

1. Constraint Collapse - “What becomes free - and whose business breaks?”
Every workflow has a binding constraint - not always cost. Sometimes it’s time, cognitive bandwidth, scale, or access. AI is about to remove some of these entirely. The question isn’t “what gets cheaper?” It’s “whose revenue depends on this constraint existing?” If the answer includes your firm, this is where you need to be paying attention. The continuous diligence hypothesis emerged directly from this question: the constraint wasn’t that diligence was expensive, it was that maintaining a continuously verified information state was operationally impossible regardless of how much you spent. Remove that constraint and the entire episodic diligence model becomes a solution to a problem that no longer exists.

2. The Unbundling - “What are you actually selling - and which part is about to become worthless?”
Every service you charge for is a bundle of components, and you’ve probably never itemised them because the bundle is just “what we do.” AI will replicate some of those components to a verifiable standard. The ones it can’t replicate are where your future premium concentrates. The evidence-based advisory hypothesis came from this: an investment advisory mandate bundles market knowledge, analytical packaging, relationship access, and strategic judgement into a single percentage fee. AI commoditises the first two. The question is whether clients keep paying the same fee for the last two - or start buying judgement separately, possibly from someone who was never an “advisor” before but who now owns the best evidence infrastructure.

3. The New Entrant - “If someone started from scratch today, would they build what you’ve got?”
A well-funded team with no legacy systems, no existing relationships, but full access to frontier AI enters your market. They don’t need to respect how things currently work. What do they build? Who do they sell to? And the question that should keep you up at night: why can’t you build it first? Usually the honest answer is “because our current business gets in the way” - and that’s exactly the answer that should worry you most. The occupier-as-a-service hypothesis is a new entrant question: if you were starting a property management business today, would you build what property management currently looks like? Or would you build an AI-orchestrated operating platform with continuous monitoring, predictive maintenance, and verified performance delivery - that happens to manage buildings?

4. The Control Point - “After this shift, whose signature still matters?”
Value in CRE concentrates around accountability - whoever signs the recommendation, the valuation, the approval. Their signature carries weight because they’re standing behind judgements that can’t easily be verified any other way. When AI changes what can be verified and evidenced, that signature may carry less weight - or shift to someone else entirely. Whoever holds the accountability after the migration holds the value. Consider: a Red Book valuation requires a RICS-qualified surveyor’s signature because comparable selection, adjustment logic, and market judgement need a professional to stand behind them. If AI produces the analysis with full evidence chains, auditable reasoning, and verified comparables, the surveyor’s role shifts from “produce the valuation” to “validate the machine’s output.” That’s a different job, with different economics - and it may not need the same provider.

5. The Customer - “Who has this problem right now - and what are they currently paying to solve it badly?”
The most transformative new businesses don’t create demand from nothing. They serve demand that’s already there but currently met by expensive, slow, or inadequate solutions. The test isn’t “would someone hypothetically pay for this?” It’s “who’s already spending money or losing money because this doesn’t exist yet?” Every institutional buyer who has ever lost a competitive bid because their diligence took two weeks longer than the other side’s is paying the cost of the continuous diligence gap - they’re just paying it in lost deals rather than invoices. Every tenant who signed a lease based on a glossy brochure and discovered the building doesn’t perform is paying the cost of the missing performance verification layer. The customer already exists. They just don’t know yet that what you’re building is the solution.

HOW TO USE THIS

Work through the five provocations in sequence after a standard RIRA pass has produced its H1 and H2 outputs for a specific workflow. Those outputs become the foundation - you’re pushing further, not replacing them. Give each question time to breathe. The uncomfortable answers - the ones that implicate your own firm’s business model - are the ones worth pursuing.

Then classify what emerges. Using AI to produce lease abstracts faster is an H1. Building a lease abstraction system with evidence chains as a firm standard is an H2. Creating a continuously verified portfolio information state that becomes a product you sell to buyers, lenders, and insurers - that’s an H3. The distinction matters because H2s improve your current business while H3s build a new one. Both are valuable. Confusing them leads to underinvestment in the H3 and over-claiming on the H2.

Most ideas will turn out to be H2s. That’s fine. That’s expected. The framework has teeth precisely because it filters honestly. But the one or two ideas that survive all five provocations and still look like genuine H3s - those are worth serious investment of time and thought.

THE COMPLETE STACK

This completes RIRA. The framework has always had three horizons in its Imagine phase, but until now H3 has been a diagnostic label - something you aspire to - rather than a generative method. The five provocations give it teeth. They turn “are we building Uber or a faster taxi?” from a question you ask at the end into a structured process you work through to find out.

The stack remains three frameworks, each answering a different question. RIRA asks how we’re creating value - and now has the tools to push that question all the way to transformation, not just efficiency and capability. The Automation Matrix asks what kind of work this is and how it should be automated. The Prompting Framework asks how we actually get work done. Strategy, analysis, execution.

If your AI strategy starts and ends with doing current work faster, you’re playing an H1 game that everybody else will also be playing within two years. The firms that engineered verifiability into their cognitive work will be pulling ahead. And the firms that ran these provocations honestly, saw the value migration early, and started building - they’ll be playing a different game entirely.

The provocations don’t tell you which game to play. They help you see the games that are available.

Run the five provocations on your most profitable workflow - and note which answer makes you uncomfortable.

Read More
Antony Slumbers Antony Slumbers

AI and Office Space Demand

How much space will you need in 3, 5, 10 years?

A new genre of research is emerging. Call it AI-impact modelling.

The approach goes something like this. Take every occupation in the economy. Decompose each into its constituent tasks - data entry, document drafting, scheduling, analysis, negotiation, physical labour. Score each task for AI susceptibility on a scale of 0 to 1. Weight the scores by how much time each occupation spends on each task. Apply an adoption discount to reflect real-world barriers. Out comes a number: the “practical AI impact” for every job in the economy.

The methodology traces back to Eloundou et al.’s influential 2023/24 paper “GPTs are GPTs,” which assessed GPT-4’s capability against occupational tasks. Since then, the ILO, the OECD, Felten et al., and most recently the UCL/Cardiff GAISI study have all built on or extended the framework. It has become the default approach for estimating AI’s labour market impact.

And the numbers are eye-catching. Depending on whose model you look at, somewhere between 25% and 45% of the average knowledge worker’s tasks are susceptible to AI automation. Apply discount factors for adoption barriers, layer on displacement assumptions and timeline projections, and you get headline figures about millions of FTE capacity freed, hundreds of thousands of office jobs at risk, and tens of millions of square feet of space demand evaporating.

These models are proliferating because there’s a commercial appetite for them. Occupiers want to know how much space they’ll need. Investors want to know whether their assets are exposed. Consultants want to sell certainty into an uncertain market. And a multi-hundred-occupation spreadsheet with scenario matrices and regional breakdowns certainly looks like certainty.

It isn’t.

WHAT THE EVIDENCE ACTUALLY SHOWS

Before I explain why the models don’t work, let me acknowledge that something real is happening. The GAISI study (UCL, Cardiff, Oxford, Surrey - August 2025) found that job postings in AI-exposed occupations were 5.5% lower in Q2 2025 than pre-ChatGPT trends would predict - roughly 84,000 fewer postings per month. The pay premium for AI-exposed tasks fell ~12% between 2017 and 2024. Displacement effects appear, at present, to outweigh productivity-driven gains in labour demand.

But notice what this actually measures: hiring decisions. Firms choosing not to fill roles, restructuring teams, rethinking workflows. The real signal is in the aggregate behaviour of employers, not in theoretical task scores.

WHY TASK BASED MODELS DON’T WORK

Here’s the fundamental problem.

The capability scores are already stale. Every major task-exposure framework was built to assess capabilities circa 2023 - primarily GPT-4. Even GAISI, published in 2025, used GPT-4o and Gemini 2.5 Pro for its task ratings. GPT-4o was not a reasoning model, and while Gemini 2.5 Pro is more capable, both have already been superseded. Current frontier systems are qualitatively different, and in 18 months they’ll be different again. Using a 2023 capability snapshot to project 2031 or 2036 labour markets is like using a 2015 smartphone to predict what your phone would do today.

The adoption discount is a guess wearing a lab coat. These models typically apply a uniform discount - say 0.6 - to all occupations, reflecting the gap between what AI cando and what organisations will deploy. But adoption varies enormously. A fintech processing team will hit 85% AI integration years before an NHS trust reaches 30%. A flat discount applied uniformly creates false precision while masking the variation that actually matters.

The precision is theatrical. When a model tells you that a particular occupation faces a practical impact of 34.332% and will have 12.70284 hours freed per week - that level of specificity implies extraordinary confidence. The real uncertainty band is ±15 percentage points at best. Six decimal places on a qualitative assumption is not data science. It is aesthetics.

And jobs don’t work like this. This is the deepest problem. The entire framework treats occupations as fixed containers with automatable components. You take today’s task mix, score each piece, and calculate what percentage gets “freed up.” But that’s not how AI changes work. What happens in practice is that roles get redesigned, merged, split, and reinvented. A financial analyst who used to spend 40% of their time gathering data doesn’t just get that 40% back as spare capacity. The role transforms. Expectations rise. New responsibilities appear that didn’t previously exist.

The model is treating the subject as a photograph when it is actually a video.
Changing all the time.

In practice, you might as well just guess. You’d get a different number, but it wouldn’t be meaningfully less accurate. And the spreadsheet version is arguably worse than an honest guess - because the aesthetic of precision triggers all manner of anchoring biases in anyone not deeply sophisticated in their data analysis. Hand a board a number with six decimal places and they’ll treat it as a fact. That’s not a useful tool. It’s a crutch that could lead to a string of bad decisions.

THE QUESTION THESE MODELS GET WRONG

But there’s a deeper problem than imprecision, and it comes from recent work by MIT’s David Autor and Neil Thompson.

Their June 2025 paper “Expertise” argues that the standard task-based approach asks the wrong question entirely. It’s not how much of a job AI can automate that matters. It’s which parts.

The insight is this: when automation removes the low-expertise tasks from a job - the routine data gathering, the basic processing, the scheduling - the remaining work concentrates around higher-expertise activities. The job becomes harder, more specialised, and better paid. But fewer people can do it, so employment in that role falls.

When automation removes the high-expertise tasks - the complex analysis, the specialist judgement - the job gets easier. More people can do it. Employment expands, but wages drop.

Autor and Thompson document this empirically across 300+ US occupations over four decades. Their accounting clerk example is instructive: computers eliminated much of routine bookkeeping between 1980 and 2018, yet while employment fell by a third, real wages rose by nearly 40%. The remaining work - reconciliations, exception handling, judgement calls - was harder and more valuable.

Inventory clerks went the other way. Automation removed their most skilled work - technical analysis, compliance checks - leaving basic counting and stocking. The job became easier, wages fell, and employment expanded as more people could now do it.

As Neil Thompson put it: taxi drivers once relied on deep knowledge of local streets as a real differentiator. GPS automated that expertise. The result was a more commoditised service - lower wages, many more drivers.

The same task. The same “automation exposure.” Completely opposite outcomes - for wages, employment, and the kind of space those workers need.

This matters enormously for office space demand, because the task-based models produce a single number per occupation and treat it as if it tells you something about future space requirements. It doesn’t. You need to know the direction of the expertise shift.

If AI removes the grunt work from professional roles - the research, the routine drafting, the data processing - the remaining workforce is smaller, more expert, more senior, better paid, and likely needs better space. Less total square footage, but higher quality. More collaborative, more client-facing, more premium.

If AI removes the high-expertise components and commoditises the role, you get more people doing simpler work for lower pay. The space requirement shifts toward cheaper, more generic, possibly non-office settings. Same occupation. Same AI exposure score on the spreadsheet. Completely different implications for your portfolio.

WE NEED A BIGGER PIE

There’s another variable that every task-based model ignores: growth.

The bottom line is an absolute certainty: AI means we will need fewer people to achieve our current level of output.

The question is whether the pie stays the same size. If an economy grows - new products, new services, expanded demand - then productivity gains get absorbed into expansion rather than headcount reduction. Jevons paradox: when something becomes cheaper and more efficient, demand for it can increase.

But growth doesn’t always bring jobs. Bloomberg recently highlighted an unprecedented “jobless boom” in the US - GDP growing while non-farm payroll growth flatlines. The divergence has never been this persistent this far into an expansion. If AI reinforces that pattern, even economic growth won’t necessarily translate into office demand.

For a low-growth economy like the UK, the arithmetic is harder still. Productivity gains without matching demand growth translate directly into headcount reduction. This macroeconomic context is absent from every task-based model I’ve seen, and it’s arguably the most consequential variable in the system.

HOW TO ACTUALLY THINK ABOUT THIS

So if spreadsheet determinism doesn’t work, what does?

The answer is to stop pretending we can model 10 years of AI-driven change with decimal places and instead be honest about what’s knowable at different time horizons.

Three years out is an engineering problem.

Most relevant variables are already observable. Lease structures are signed. Supply is largely consented. Hybrid patterns have stabilised. AI adoption is still early for most firms.

Three-year demand is a function of things you can measure today: lease expiry profiles, observed attendance rates, sector-level employment trends, and the hiring slowdowns GAISI is already documenting. The methodology should be empirical, not modelled.

Five years is where it gets genuinely hard - but not impossible.

AI capabilities in 2031 will be radically different from today, but organisational adaptation will still be incomplete. Too far for extrapolation, too near for speculation.

The right framework here is not task decomposition. It’s two variables: productivityand growth. How much more productive will AI make your people - actually, not theoretically? And is demand for your output expanding or flat?

These interact. And they produce radically different space outcomes.

Take a company with 500 people currently occupying 60,000 sq ft. Run it through four scenarios over a five-year horizon:

Scenario 1: Low growth, high AI productivity gain Output +10%. Productivity +40%.The toughest scenario for space. Headcount drops to ~390, space requirement to ~47,000 sq ft - a 22% reduction. And per Autor/Thompson, AI has stripped out the routine work, leaving a smaller, more senior team. They don’t just need less space. They need different space - more collaborative, more client-facing. This firm is handing back a floor and upgrading what remains.

Scenario 2: Low growth, low AI productivity gain Output +10%. Productivity +15%.Nothing dramatic. Headcount drifts to ~480, space to ~57,500 sq ft - barely a 4% reduction. The firm absorbs AI gradually, trims a few roles through attrition, renews at roughly the same size. Most organisations today are probably closer to this than they’d like to admit.

Scenario 3: High growth, high AI productivity gain Output doubles (a fast-growing tech or professional services firm, say). Productivity +40%. Jevons paradox in action. Growth absorbs the productivity gain and then some. Headcount rises to ~715, space to ~86,000 sq ft - a 43% expansion. But the type of space changes dramatically: fewer individual desks, more project space, more client-facing environments.

Scenario 4: High growth, low AI productivity gain Output doubles. Productivity +15%.Growth without much AI transformation. Headcount expands to ~870, space surges past 104,000 sq ft. Scaling the old way - more people, more desks. But this may be the least likely scenario over five years, because firms growing this fast tend to invest most aggressively in AI.

The point is not that these numbers are precise - they’re napkin maths. The productivity ranges (15-40%) reflect early-adopter evidence: coding teams reporting 30-50% gains, legal and customer service operations seeing 2-3x throughput. Many firms will be lower, particularly over five years. Over ten - which is more typical of CRE investment horizons - these ranges become conservative. But they show what task-based models fundamentally miss: the same AI productivity gain can mean 22% less space or 43% more space, depending on whether a firm is growing. No occupation-level task score tells you which one you’re in. Only your own strategic context does.

These scenarios assume constant density at 120 sq ft per person. In practice, collaborative space runs larger per head than desk-based processing space - which narrows the reduction in Scenario 1 but reinforces the point about character over quantity.

And the Autor/Thompson expertise lens adds a third variable: what kind of space? In three of four scenarios, routine processing work diminishes and expert, collaborative work grows. The square footage number alone misses this. You might need less space in total but higher specification - or more space of a fundamentally different kind.

NOTE: There are other variables these scenarios don’t capture - attendance patterns, peak-load coordination, the logic of signing a five-day lease for three-day usage, whether co-location becomes more or less important as teams shift to creating and curating AI agents. These are real and consequential. But they are also deeply specific to each organisation’s culture, stage, and ways of working - and largely unknowable over a 5-10 year horizon. They reinforce the point: the answers live inside each company’s particular circumstances, not in a macro model.

Ten years is scenario territory.

Anyone offering a 10-year office demand forecast with specific numbers is selling confidence they don’t have. The honest approach is qualitative scenarios - genuinely different futures. In one, AI delivers transformative productivity and new sectors emerge. In another, productivity gains outstrip demand growth, creating structural surplus in commodity stock. In a third, the Jevons paradox catalyses broader expansion. In a fourth, the US-style jobless boom becomes structural - output growing, employment stagnant.

No spreadsheet helps you choose between these. What helps is understanding the forces, tracking the signals, and building resilience across multiple futures.

THE SUPPLY TRAP

There’s a temptation, particularly for landlords and investors, to look at the current supply picture and take comfort. There is already a well-documented shortage of the kind of space the best occupiers want - high-specification, ESG-compliant, well-located, operationally excellent. In many UK markets, occupiers are being pushed to second-best options because the best simply isn’t available.

But relying on supply constraints to fill your building is increasingly risky. AI doesn’t just change how much space firms need - it makes hybrid and distributed working progressively easier and more effective. As thetools for asynchronous collaboration, AI-synthesised meeting outputs, and rich shared context improve, the compulsion to be in a specific physical space diminishes. The occupier who accepted your second-best building because their first choice was full might, in three years’ time, decide they don’t need a central office of that size at all.

Supply constraints are a short-term tailwind. They are not a strategy.

WANT BEATS NEED

This brings us to what I think is the most important strategic question in office real estate right now - and it has nothing to do with task decomposition models.

The #SpaceAsAService era has already demonstrated that selling to someone who needs your product is fundamentally different from creating something someone actively wants. The entire trajectory of operational intensity in the office sector - the rise of flexible space, managed offices, hospitality-influenced design, amenity-rich environments - reflects this shift. Just building to spec is no longer enough.

AI accelerates this. If the processing, analysis, and document production that currently fill desks get absorbed by AI, the purpose of the office shifts further toward collaboration, client relationships, mentoring, culture-building, and the irreducibly human work that remains. Space that serves those purposes will be wanted regardless of macro demand trends. Space that’s still designed for rows of people doing information processing will not.

For developers and investors, this means every asset decision needs to start with a question that no macro model can answer: when this building is ready, what will companies want from their workspace? Not what will they need - because need is declining. What will they choose? What will they pay a premium for? What will make them pick your building over the increasingly viable alternative of not having an office at all?

You can still win big in a shrinking market if you’re building something people choose rather than something they’re forced to occupy. But it requires understanding what work is becoming - not just how much of it there is.

Relying on aggregate demand to fill your building is a bet on a macro variable you can’t control and increasingly can’t predict. Building something people actively desire is a bet on quality, experience, and understanding the future of work.

That’s a better bet.

Read More
Antony Slumbers Antony Slumbers

CRE: Constraints, Moats and Value

AI is removing the constraints that built CRE’s moats. What replaces them?

Two rather contradictory things happened on Wednesday, 11th February, 2026. First, shares in CBRE, JLL and Cushman & Wakefield fell precipitously on Wall Street, erasing circa 12-14% of their value. The consensus reason?

“We believe investors are rotating out of high-fee, labor-intensive business models viewed as potentially vulnerable to AI-driven disruption”

And then, on exactly the same day, PropTech giant Yardi put out a research report alongside the Association of Real Estate Funds (AREF) ‘examining how the real estate investment industry is responding to operational pressures through technology, data and digital transformation.
Matt Glenny, commercial and investment management director at Yardi, describes the findings this way:

The report gives us a real insight into the perception of the real estate industry compared to a lot of the hype that we hear. Just 45% of respondents see AI as the biggest tech disruptor of the next three to five years

He adds that…

88% of respondents are trialling AI in some form to automate or enhance performance of manual tasks such as reading and summarising documents and automating meeting minutes. Predictive analytics and the adoption of AI agents are viewed as future aspirations rather than a current consideration.

So, the market and the industry are pointing in opposite directions.

Who is right?

Sangeet Paul Choudary, senior fellow at the University of California, Berkeley has written:

New tech collapses old constraints. Once a constraint disappears, the logic of competitive advantage and business model design must be reimagined from first principles.

If this is so then we need to look at what constraints AI is removing, what new ones it will impose, and how defensible are the moats underpinning current profitability within the CRE professional services industry.

What follows demonstrates the ‘Release’ phase of our RIRA Framework. Starting with…

CONSTRAINTS AI REMOVES IN CRE

Analytical Cost
Tasks that previously required teams of analysts working for days - lease abstraction, comp analysis, market research, financial modelling, due diligence document review - can now be performed in minutes. This makes existing work cheaper and, perhaps more importantly, it makes previously uneconomic work viable. The “non-consumption” opportunity is real: firms can run 50 scenario analyses where they previously ran 3, or screen 200 properties where they previously screened 20.

Knowledge Access
Information asymmetry - knowing things that clients and competitors don’t - has historically been a core source of CRE value. AI though will collapse this by making market data, comparable transactions, planning documents, and regulatory frameworks accessible to anyone with a well-structured prompt. The gatekeeping function that sustained brokerage margins will erode when the information itself becomes abundant. Proprietary data will exist, and still have considerable value, but far more information will become accessible than many realise.

Coordination
This is perhaps the least appreciated constraint that is being removed by AI. CRE transactions involve coordinating across multiple parties - lawyers, surveyors, lenders, agents, asset managers - each operating within their own systems and timelines. AI can interpret across these disparate data sources without requiring standardisation or consensus, compressing transaction timelines and reducing the friction that historically justified intermediary fees.

Talent Scaling
This is a fundamental constraint that AI is dismantling. A boutique firm with three sharp principals and well-designed AI workflows can now produce analytical output that previously required the likes of CBRE’s headcount. The scale advantage that the big companies enjoyed, hundreds of researchers, analysts, and junior professionals feeding the advisory machine, becomes less decisive when AI handles the analytical infrastructure underneath the senior relationship.

Geographic Knowledge
Deep local market knowledge - planning regimes, tenant dynamics, micro-location factors - has been a genuine moat for regional specialists. AI agents with access to planning portals, EPC data, Land Registry records, and local comparable evidence can approximate (not yet replicate) much of this knowledge, reducing the premium that locality alone commands.

CONSTRAINTS AI INTRODUCES

Verification
This becomes the new bottleneck. The more convincing AI outputs become, the harder hallucinations are to spot, and the greater the temptation to skip verification. This is a genuinely new constraint - the cost of ensuring an AI-generated IC pack is accurate may exceed the cost of generating it. Firms that build robust evidence architecture gain a real advantage; firms that don’t take on hidden liability risk.

Liability and Accountability
These create a structural constraint. When a junior analyst produces a flawed valuation, there’s a clear chain of professional accountability. When an AI system produces one, the liability question is genuinely unresolved. Professional indemnity, regulatory compliance, and fiduciary duty all sit awkwardly with AI-generated outputs.

However, this means that liability can become a moat. In knowledge work with ambiguous outcomes, absorbing liability remains a function that pure AI tools cannot perform.

Data Quality and Provenance
These constraints intensify. AI systems are only as good as their inputs, and CRE data remains fragmented, inconsistent, and often proprietary. The firms that have clean, structured, proprietary datasets gain compounding advantage. Those relying on the same publicly available data as everyone else will find their AI outputs commoditised immediately.

Governance and Workflow Design
These become a constraint on deployment speed. It’s not enough to have capable AI - you need checkpoints, audit trails, escalation protocols, and quality assurance routines. Building this infrastructure requires architectural thinking that most CRE professionals haven’t had to do before. The firms that solve the governance problem will be able to deploy AI at scale; the rest remain stuck in pilot purgatory.

Cognitive Dependency
This risk emerges as a longer-term constraint (though I believe a major one in the making). As professionals offload analytical tasks to AI, the institutional knowledge that informed those tasks atrophies. Junior professionals who never learn to build a model from scratch may lack the judgement to evaluate whether an AI-generated model is sound. Designing workflows and systemic methodologies to counter cognitive dependency will become a major requirement in knowledge businesses. I address this at length in my ‘Cognitive Sovereignty - Use It or Lose It’ article:

WHY CONSTRAINT MAPPING MUST BE WORKFLOW-SPECIFIC

These are generic constraints that are being removed or introduced across the CRE industry. They’re structural features of what AI does to knowledge work.

But the weighting differs enormously, and analysing them in relation to particular workflows is vital.

Consider the contrast between investment capital markets and property management. In capital markets, the information asymmetry constraint has historically been central to value creation, knowing about off-market deals, having proprietary comp data, and understanding institutional investor appetite. AI’s collapse of that constraint is existential to the business model.

In property management, information asymmetry was never the primary value driver: the constraints that matter are physical presence, emergency response capability, tenant relationship management, and regulatory compliance around health and safety. AI removes some PM constraints (reactive maintenance scheduling, compliance tracking, tenant communication at scale) but leaves the physical and relational constraints largely untouched.

Or compare valuation with development advisory. Standardised valuations face near-total constraint removal, AI can assemble comparables, model cash flows, and produce RICS-formatted reports with diminishing human input. Against that, on the new constraint side, there will be a need for professional sign-off and PI liability.

Other areas of the industry will be largely untouched by AI. Development advisory, for example, faces an irreducible constraint in planning committee politics, local authority relationships, and community engagement. The constraint AI does remove for development (planning application analysis, precedent search, viability modelling) is important but secondary to the human-political constraint that gates every project.

Then there’s leasing, where we’re seeing constraint-swaps. In the US, in multi-family, pricing has had constraints removed but that in turn has created a new legal constraint and we’re seeing several companies being sued over the ‘coordination’ of algorithmic pricing. That constraint doesn’t appear in capital markets or valuation work at all. It’s a constraint born from the interaction of AI capability with a specific regulatory environment (which arguably should have been anticipated).

So the constraint map - which constraints dominate, which are secondary, and critically, how they interact - has to be workflow-specific. And the interactions are where the real insight lives. Which is why our RIRA Framework has ‘Release’ as the first ‘R’ - you need to understand the new constraints that are appropriate to your specific workflow, or sector.

What is absolutely clear though is that AI is fundamentally changing the environment in which our businesses operate, and this in turn means that many of the moats we’ve relied on in the past might no longer exist.

THE FALSE MOATS

The Yardi/AREF report, together with Linkedin commentary on the stock price erasure, and pronouncements during earnings calls, make it clear that many in the industry believe CRE has special protection from several ‘moats’ that make it distinctive from other industries, and more defensible against AI.

These need to be addressed:

“Our relationships are our moat.”
Relationships are real, but their defensive value depends heavily on what those relationships enable.

If a client relationship primarily gives access to deal flow and information, AI erodes the information component. The relationship remains valuable, but the fee it commandscompresses because the analytical work underneath it requires less human capital.

The relationship is a moat for the individual, not necessarily for the firm’s margin structure. And crucially: relationships don’t scale, which means they can’t offset margin compression across the broader business.

“Our data is our moat.”
This is half-true, which makes it dangerous. The major companies do have proprietary transaction data, but much of it sits in unstructured formats, inconsistent taxonomies, and siloed systems. Having data and being able to use data as a strategic asset are very different capabilities.

Meanwhile, AI dramatically lowers the cost of assembling comparable datasets from public sources - Land Registry, planning portals, EPC databases, Companies House. The proprietary data advantage is real but narrower than assumed, and it’s a wasting asset if competitors build structured alternatives faster.

“Complex deals require human judgement.”
The definition of “complex” has been a moving target for decades (spreadsheets killed some complexity, databases killed more). What counted as requiring senior judgement five years ago (multi-scenario cash flow analysis, market positioning reports, tenant covenant assessments) is already within AI’s capability.

The residual, true strategic judgement under genuine uncertainty, is real but represents a much smaller slice of billable activity than the industry likes to admit. Is it more than 15%? Most of what brokers call “judgement” involves pattern recognition applied to structured data, which is exactly the capability profile where AI is advancing fastest.

“Regulatory complexity protects us.”
CRE is indeed heavily regulated; RICS standards, planning law, lease code compliance, environmental regulations. But regulation protects the activity, not the incumbent. AI systems that can navigate regulatory requirements more reliably than human professionals don’t erode the regulatory constraint - they erode the premium charged for managing it. Regulation becomes a moat only if firms combine it with liability absorption and governance credibility.

“Our brand and institutional trust are our moat.”
For trophy transactions - the £500m office acquisition, the major corporate relocation - institutional brand matters. Clients want CBRE or JLL on the cover page for their board. But this protects the top 5-10% of transactions by value. For the long tail of mid-market deals, where the analytical quality matters more than the logo, brand becomes less decisive as AI enables smaller firms to produce institutional-grade output.

“Scale is our moat.”
This is the one Wall Street is actively repricing. Scale was a moat when it provided analytical capability that couldn’t be replicated without hundreds of researchers. AI inverts this as it makes scale available to the lean.

It’s not hard to envisage a five-person firm with well-architected AI workflows producing the analytical throughput that required a 40-person team three years ago.

When technology collapses the constraint that justified your architecture, the architecture itself becomes a liability. The big companies’ headcount, which was their competitive advantage, becomes their cost structure problem.

THE MEANING FOR VALUE CREATION

So if AI is removing constraints, and creating new ones, and the industry’s moats are leaking fast, what does this mean for value creation?

The CRE industry is mistaking a stable playing field for a permanent one. Furthermore, it is thinking about how AI will support its existing operating models, rather than thinking about how to fundamentally change the game.

Last week the CEO of one of the major players said on an earnings call:

that AI would benefit the business over the longer term, adding that its transaction and investment work would be “most protected” from disruption.

Which encapsulates this thinking precisely.

The share price crash is the market, brutally and without nuance, beginning to price a structural revaluation into this sector. What the industry now needs to do is double down on understanding where value is migrating to. As analytical and transactional work commoditises, value will migrate to whoever resolves the new constraints - verification, liability, governance, and coordination.

Exactly what this means in terms of products and services is still to be determined. But by diving deep into the first ‘R’ in the RIRA framework, the map starts to become more like the territory, allowing us to better agree with or dismiss most of the ‘noise’ around CRE and AI today.

What is clear is that we have to think harder about where AI is taking us, and explain this precisely to the markets. The competitive advantage will accrue to those who can demonstrate why AI is a feature, not a bug. How many will manage that?

OVER TO YOU

What to do about all of this? What’s the strategic question you should be asking that you’re currently not?

Read More
Antony Slumbers Antony Slumbers

The PropTech Apocalypse

And why you need RIRA

Summary: Over $300 billion in software market value was erased in a single day. This piece explains why - and what it means for commercial real estate. Using the RIRA framework applied to an investment deal workflow, I show how AI is swapping old constraints for new ones, where value is migrating, and why the firms that own the data layer and the execution layer will capture the next cycle’s profits. It’s a long read. It’s meant to be.

Over the last three weeks I’ve introduced three frameworks - the CRE Automation Matrix, Prompting for CRE Professionals and the RIRA (Release, Imagine, Redesign, Actualise) Framework itself. This week I want to show you RIRA in action, as an example for how it can be incorporated throughout your business.

But first, let’s talk about what happened on Wall Street at the start of February.

$300 Billion Erased

February opened with a software massacre.

On Tuesday the 3rd, according to Forbes, over $300 billion in market value was erased from the software sector in one trading day

“Several of the most entrenched enterprise software companies fell sharply in a single day. Salesforce, ServiceNow, Adobe, and Workday each dropped around 7 percent. Intuit fell nearly 11 percent. At the same time, valuation multiples across the sector compressed violently. The average forward earnings multiple for software companies collapsed from roughly 39x to about 21x in just a few months.”

Wall Street called it the “SaaS-pocalypse.

This was less a random correction and more a fundamental repricing. The market is waking up to something the venture capital world has been whispering about for months: the software model that dominated enterprise technology for two decades is structurally breaking.

The Thesis: Why Software is Being Hollowed Out

The argument is as follows, and it maps onto commercial real estate with uncomfortable precision.

For twenty years, businesses bought software to help humans organise their work. Dashboards. To-do lists. Complex menus. You paid for a “seat” - a licence for your employee to log into a system. The software didn’t actually do anything. It was a digital filing cabinet. Your employee still had to log in, click buttons, enter data, send emails.

You paid twice. Once for the software. Once for the salary of the person operating it.

The thesis is straightforward: if AI can now execute tasks, paying for software that merely organises human tasks is becoming a bad investment. The industry is converging on this from multiple directions - Phil Fersht at HFS Research calls it “Services-as-Software,” a16z frames it as “software eating labour,” I’ve described it as “Outcomes-as-a-Service.” The labels differ but the signal is the same: not tools that assist the user, but systems that replace the user’s labour. Not dashboards that show you what needs doing, but agents that do it.

This creates what I’ve been calling the “barbell effect.” If the middle layer of software - the screens humans look at to manage tasks - is being hollowed out, value migrates to two poles. At one end: the data layer. AI agents cannot function without accurate, structured, proprietary data. Your lease data, payment history, building performance records - these become more valuable, not less. At the other end: the agent layer. Systems that actually execute work - negotiate a lease, dispatch a vendor, rebalance an HVAC system, screen an acquisition.

The middle - standalone CRMs, basic dashboards, static reporting tools, “thin wrapper” applications that put a digital skin on a manual process - gets crushed.

Clayton Christensen’s ‘conservation of attractive profits’ is the key: when a layer commoditises, profits don’t die - they move to adjacent layers where differentiation still matters.

What looked like “$300 billion erased” is the market repricing where future margins will sit: away from software that organises human work, and towards the layers that own the data and execute outcomes.

The question for every CRE professional is: which side of the barbell are you building on?

The question for every PropTech company is: are we facing our own apocalypse, and what can we do about it?

This is where RIRA comes in. As a strategy-to-execution framework for rebundling value creation in commercial real estate under conditions where AI materially changes the unit economics of analysis and content production, it helps you develop new products and services that leverage, rather than are crushed by, AI.

RIRA in Action: The Investment Deal

To demonstrate the process of using RIRA I am going to use the example of ‘The Investment Deal’. How might the way we deal with deals develop in a heavily AI-mediated world?

RIRA, which stands for ‘Release-Imagine-Redesign-Actualise’, is a way of looking at a workflow through four lenses, where each lens depends on the previous one, and the power comes from following the thread all the way through.

I’ve chosen an investment and capital markets workflow because the constraint-swaps at each stage are the most vivid, the evidence is the most concrete, and the implications are the most strategically significant. At the end, I’ll sketch how the same logic applies everywhere else.

Release: The Analyst Bottleneck Dissolves - and a New One Appears

The Release lens is about what constraints AI removes, and which new constraints it introduces.

Every acquisitions professional knows this workflow intimately. An Offering Memorandum arrives. A junior analyst extracts data from the PDF. Types it into Excel. Massages the assumptions. Builds an ARGUS model. Runs comps. Drafts an Investment Committee memo. Repeat.

A typical CRE acquisition consumes 30–50 analyst hours on data extraction alone. Deal volume is capped by how many OMs a human team can physically process. Firms miss opportunities not because the deals aren’t there, but because they lack the manpower to review them.

AI tools are already releasing this constraint. We are seeing tools (vendor-reported) that generate full five-year cash flow models and IRR analyses in under five minutes with (allegedly) 99%+ extraction accuracy. Others that achieve 95% faster approvals with 400% better risk detection than human reviewers. Firms using these tools report handling 3–4× more deals with the same staff. The precise numbers warrant scrutiny - vendor benchmarks rarely survive contact with messy, real-world portfolios - but the direction of travel is clear.

But, and this is the core insight of the Release stage, AI never just removes a constraint. It swaps it for a new one.

The old bottleneck was throughput: we couldn’t screen enough deals. The new bottleneck is explainability. An Investment Committee cannot approve a £100 million acquisition based on “trust me” from a model. Dealpath, in a recent survey, reports that 93% of institutional firms cite significant AI adoption barriers, with 43% lacking internal expertise. The binding constraint shifts from speed to trust. If the AI cannot show its working - if the outputs aren’t auditable, traceable, and defensible - the speed is worthless.

Data quality becomes an equally hard constraint. If the underlying data is fragmented or dirty, the AI will confidently produce bad returns. “Clean data” becomes the new prerequisite for speed.

Diagnostic question: If we could screen ten times as many opportunities tomorrow, what breaks first - governance, explainability, data quality, decision rights, or IC bandwidth? That’s the new binding constraint, and it’s where execution will stall unless you redesign around it.

Imagine: The 24/7 Deal Hunter

The Imagine lens is about what becomes possible, given the new technologies at your disposal.

If the marginal cost of intelligence drops to near zero, what becomes possible?

Across three horizons: H1 for what becomes more efficient, H2 for what new capabilities can we develop, and H3 for what transformation is now possible?

Think taxis - they could be faster, or maybe they could accept credit cards or… well they could become Uber.

Imagine an investment firm where no human performs data extraction. A sourcing agent continuously monitors thousands of inputs - public listings, court filings for distress, planning/zoning changes, demographic shifts, permit applications. It doesn’t just search. It reasons. It holds the firm’s specific investment thesis - “value-add multifamily in growth markets with strong transport links” - and when it identifies a potential asset, it performs a preliminary underwrite. It ingests the rent roll regardless of format, maps it to the firm’s chart of accounts, pulls real-time tax and insurance comps, and runs a ten-year DCF. The human principal receives a deal memo only for assets that clear the IRR hurdle.

If we could combine market intelligence, debt capacity analysis, and portfolio fit scoring in a single system, and run it autonomously, 24/7/365, we’d have enabled something pretty special. Are we there yet? No, but the likes of JLL (building on the skyline technology they bought a few years ago) and Cherre, building on their years of data wrangling, are getting there.

The critical distinction: this isn’t “a better dashboard.” It’s the elimination of the entire analyst-receives-OM → manual-data-entry → build-ARGUS-model → manual-comp-research → IC-memo workflow. The human enters at the judgement stage, not the extraction stage. The role of the analyst doesn’t disappear - it transforms from “data processor” to “decision validator.”

Design question: If the marginal cost (and time) of analysis and execution tends toward zero, what should we now do that we currently don’t - because it was previously uneconomic or too slow? Where should the system run 24/7, and where should humans enter purely for judgement, not extraction?

Redesign: Dismantling the Excel Industrial Complex

The Redesign lens is about building an evidence-based architecture for new workflows.

The Imagine stage showed us the destination. The Redesign stage is the engineering - how do we physically rebuild our data and workflows to get there?

The core problem has a name that every CRE professional will recognise: the Excel Industrial Complex.

Today, an investment model is a spreadsheet. The moment that file is saved, it is dead. It is a snapshot in time that does not update when interest rates change, a tenant vacates, or the market moves. When you want to test a new scenario, someone manually changes the inputs and saves a new version. Knowledge lives in filenames: “Model_v7_FINAL_FINAL_JB_edits.xlsx.”

The Redesign phase replaces this with what I’d call a Live Financial Twin - an investment model that is a living software object connected to the market.

Underneath that sits the real asset: the canonical deal graph - a structured schema that captures asset terms, counterparties, risks, covenants, and underwriting assumptions as first-class objects, not cells in a spreadsheet. Crucially, it preserves model lineage: every assumption and output is traceable back to source evidence (the specific clause, rent roll line, comp, rate series, or email confirmation) with an auditable chain of custody.

This is where you’d map the workflow against the CRE Automation Matrix — identifying which components are candidates for full automation, which need human-in-the-loop gates, and which remain irreducibly human judgement.

The architecture would have three layers. First, an ingestion layer: OCR and NLP tools ingest documents - rent rolls, tax bills, service charge packs - and convert them into structured data objects, not spreadsheets. Second, live connections: the model is connected via API to real-time market feeds. If the ten-year gilt yield moves, the model automatically updates the cost of debt and recalculates the IRR. You don’t wait for an analyst to notice. Third, write-back capability: when a decision is made - “approve budget,” “proceed to exclusivity” - it doesn’t just sit in meeting minutes. It writes back to the accounting system and updates the forecast automatically.

But notice the new constraint this introduces. If the model is live and automated, you need verification architecture that didn’t previously exist. You need what some are calling “Auditor Agents“ - secondary models whose sole job is cross-checking the maths and logic of the primary underwriting agent. You need human-in-the-loop gates at specific decision points. You need an audit trail that can satisfy both your IC and your regulators.

This is where the barbell becomes concrete. The value accrues to whoever owns the “canonical deal graph“ - the structured, permissioned, auditable data layer - and its agent interfaces. Not to whoever has the prettiest dashboard.

Engineering question: Where are we relying on humans to bridge systems - and what would it take to replace those bridges with a canonical deal graph and auditable model lineage tied to source evidence?

Actualise: From Effort to Edge

The Actualise lens is about embedding and capturing value - turning capability into competitive advantage and revenue.

The first three stages created capability. The Actualise stage answers the question that matters most: how do we capture the value, and how do we stop it leaking to competitors?

Start with what’s changed about what you’re selling. The traditional investment advisory model charges for effort - analyst hours, transaction fees, AUM-based management fees that are proxies for headcount. But if the workflow you’ve redesigned can screen ten times the deals with the same team, the effort-based model actively penalises you for being good at this. You did the work in three minutes; you can’t credibly bill for thirty hours.

The business model has to shift from charging for effort to charging for outcomes: time-to-decision, risk-adjusted conviction, deal throughput, yield improvement. This isn’t a cosmetic change. It requires rethinking fee structures, incentive alignment, and how you demonstrate value to LPs and Investment Committees. But it’s also where the margin expansion lives. The firms that cling to effort-based billing in an AI-accelerated world are subsidising their competitors’ transition.

Next, recognise where the moats actually form. If you’ve built the architecture described in the Redesign stage - the canonical deal graph, live financial models, auditable lineage - you now own something that compounds. Every deal screened enriches the dataset. Every assumption validated sharpens the model. Every IC decision recorded becomes training signal. A competitor starting today doesn’t have your technology and they lack your years of structured deal intelligence. This is the data flywheel, and it is the deepest source of defensibility in the new landscape. It is conservation of attractive profits made concrete: the proprietary data layer becomes the locus of value.

Finally, look for the profit pools you didn’t plan for. Well-architected data infrastructure generates optionality. A firm that built audit-grade, real-time performance data for underwriting purposes discovers it can also price transition risk more accurately than peers, or license anonymised benchmarking data, or offer LPs portfolio-level transparency that compresses their cost of capital. These adjacent opportunities weren’t the original design objective but they are emergent properties of having built the data layer properly. The firms that capture them are the ones that recognised their data infrastructure as a platform, not just a workflow tool.

Business model question: Where are we still charging for effort (hours, seats, process friction) when the value we deliver is actually an outcome (risk reduced, yield improved, time-to-decision shortened)? And which datasets are we generating today - currently treated as operational exhaust - that could become compounding advantage if captured, structured, and reused?

RIRA AS FLYWHEEL

I’ve presented these stages sequentially because that’s how writing works - you read one section after another. But RIRA isn’t a linear process. In practice, you’ll find that the Redesign stage surfaces constraints you didn’t identify in Release. The Actualise stage reveals profit pools that send you back to Imagine with new questions. The framework is iterative by design — each pass through it sharpens your understanding of where value is forming and where it’s migrating. The power isn’t in completing four stages; it’s in the quality of the loops between them.

For instance, when you reach the Redesign stage and discover that your data architecture requires auditor agents and human-in-the-loop gates, that’s a new constraint — which sends you back to Release with a sharper question. And when Actualise reveals that your underwriting data could be licensed as a benchmarking product, that’s a new possibility — which sends you back to Imagine with a business model you hadn’t considered. The framework tightens with each pass.

And because the framework adapts to whatever constraints a given domain surfaces, the same iterative logic applies well beyond investment deals.

The RIRA Pattern Holds Everywhere

I chose the investment deal because the constraint-swaps are vivid, the evidence base is deep, and the stakes are high enough to hold your attention. But the logic isn’t specific to capital markets. It’s structural.

Every CRE workflow has a binding constraint that AI is about to release. Every released constraint produces a new one that nobody was planning for. Every new possibility demands redesigned architecture. And every redesigned architecture creates value that migrates somewhere - the only question is whether it migrates to you or away from you.

Take leasing. The released constraint is responsiveness - business hours, language barriers, the sheer friction of human-to-human scheduling. EliseAI already automates 90% of prospect interactions across 70% of the top 50 US rental operators. That’s a real number, today. But follow the thread. The new constraint isn’t technological, it’s regulatory. When algorithms optimise pricing and availability across an entire portfolio, you’ve crossed from operational efficiency into territory that regulators recognise as market coordination. The US Department of Justice’s lawsuit against RealPage is the canary. The firms that saw leasing automation purely as a cost play, without redesigning their compliance and governance architecture, are now exposed to a constraint they never anticipated. Release → new constraint → architectural consequence. The pattern holds.

You can run this logic through facilities management, asset management, development, valuation, property management - any workflow where humans currently bridge between systems, interpret data, or make routine decisions under time pressure. The binding constraints differ. The new constraints that emerge will differ. The architectural responses will differ. But the sequence is invariant, and that’s what makes RIRA a framework rather than just a case study.

If you’re wondering how RIRA applies to your specific workflow - leasing operations, portfolio strategy, facilities management, tenant experience - that’s precisely what we work through, step by step, in my forthcoming course. The investment deal example gives you the logic. Applying it to your own context is where the value compounds. It’s also where the non-obvious constraints - the ones that only surface when Release, Imagine, Redesign and Actualise interact - tend to catch firms out. Contact me for details.

The Takeaway
The $300 billion that was erased from software stocks is a signal, not an anomaly. Value is moving. It’s moving from tools that organise human work to systems that execute it. From dashboards to data layers. From seats to outcomes.

Christensen was right: when one layer commoditises, attractive profits don’t die - they change address.

And in this cycle, they’re migrating towards data ownership and execution.

The RIRA framework is a way of thinking clearly about where it’s moving in your specific context - and making sure you’re building on the right side of the barbell when it arrives.

And a way for PropTechs to avoid their own apocalypse. Or at least see it coming.

Read More
Antony Slumbers Antony Slumbers

You need to prompt with skill

No matter who you are, quality prompting has a high ROI

A Prompting Framework for CRE Professionals:
In the first part of this series we looked at RIRA- a strategy-to-execution framework for rebundling value creation in commercial real estate under conditions where AI materially changes the unit economics of analysis and content production. And last week we looked at ‘The CRE Automation Matrix Framework’ - a tool for understanding where value is moving to, what to automate, and how. This week we introduce ‘A Prompting Framework for CRE Professionals’, showing how effective prompting is a repertoire matched to domain expertise, AI fluency and task mode.

Together these frameworks form a set of foundational tools for maximising the impact, and leverage, of AI on any business.

THE EVIDENCE
Anthropic, makers of the frontier AI model Claude recently published their fourth ‘Economic index Report’, in which they show that people who can articulate complex, nuanced requirements in a prompt get complex, nuanced outputs. AI amplifies existing expertise rather than substituting for it. The ROI gap between a team that knows how to prompt well and one that doesn’t is enormous.

So if you have domain expertise, and prompting expertise, you will gain the most from current AI.

What This Framework Is

The ‘Prompting Framework for CRE Professionals’ is a practical framework for choosing how to prompt based on two realities of professional work:

(1) your domain expertise is relative to the problem you’re facing (you can be an “expert” in offices and a “novice” in data centres)
(2) your task is either Discovery (sense-making, shaping the problem, exploring options) or Execution (producing a deliverable under constraints).

The framework maps these two variables into four prompting modes - A, B, C, and D - and adds an expert loop (“Mode Switching”) that deliberately alternates between divergent analysis and convergent drafting to produce better work with fewer avoidable errors.

Why it matters: the Strategic Problem AND What Goes Wrong Without It

Most AI misuse in commercial real estate is not “people using the wrong model”. It’s people using the wrong prompting mode.

Without a mode choice:

  • You treat Discovery like Execution: You ask for an “IC pack” before you’ve clarified the investment thesis, risk boundaries, and decision criteria.

    The Result: a polished document that is directionally wrong - high credibility veneer, low decision usefulness.

  • You treat Execution like Discovery: You stay in brainstorming indefinitely, generating option overload and never landing a decision-ready artefact.

    The Result: time sink, “AI theatre”, and stakeholder impatience.

  • Novices over-delegate: A junior analyst asks for market comps, regulatory interpretation, or lease risk conclusions without guardrails.

    The Result: plausible-sounding inaccuracies that slip into deliverables and create reputational or compliance risk.

  • Experts under-specify: A senior person assumes the model “knows what I mean” and provides thin context.

    The Result: generic output, hidden assumptions, and brittle analysis.

  • Teams don’t share a prompting language: Everyone prompts differently, so outputs are inconsistent, review is slower, and governance becomes hand-wavy (“Just be careful with AI”).

THE FRAMEWORK EXPLAINED

Core Insight (Domain Expertise × AI Fluency × Task Mode)
The intent of the framework is to remind you that “good prompting” is not a universal template. It is a repertoire you select based on who you are in that moment (expert/novice), what you are doing (discover/execute), and how comfortable you are with AI as a collaborator.

How to Use It

Before you type anything substantial, force a 10-second classification:

  1. Am I effectively a novice or expert for this problem?

  2. Am I in Discovery or Execution mode?

  3. What is my AI fluency - do I know how to specify constraints, request assumption registers, and force checkability?

Typical mistakes:

  • Confusing seniority with domain expertise for the task.

  • Skipping “task mode” and asking for outputs too early.

  • Treating AI fluency as irrelevant (“I’m a real estate person, not a prompt engineer”).

CRE example: You are an experienced office investor exploring PBSA for the first time. You are an expert in underwriting discipline but a novice in this asset’s demand drivers and operational risks. You should start in A (Guided Exploration) and quickly move into B (Scaffolded Production) for structured outputs, rather than pretending you are in D.

Quadrant A: Guided Exploration (Discovery mode × Domain novice)

For:
Domain novices learning the territory.

Technique 1: Outcome-first querying

  • Intent: Anchor exploration in the decision you eventually need to make.

  • How to use it: Ask for the shape of an answer before the content. Start with “What would a good answer look like?” or “What decision does this inform, and what inputs would it require?”

  • Typical mistakes: Asking for facts and figures before you know which facts matter; accepting an answer that isn’t decision-linked.

  • CRE example: “I’m evaluating a retail park acquisition. Outline the top 12 questions an IC needs answered and why each matters (cashflow, leasing, capex, ESG/MEES, risk). Then tell me what data I need to answer them.”

Technique 2: Terminology unpacking

  • Intent: Reduce ambiguity and prevent false confidence.

  • How to use it: Ask for definitions, distinctions, and “if you hear X, what does it imply?”

  • Typical mistakes: Using terms loosely (“prime”, “stabilised”, “reversionary”) and letting AI mirror your vagueness.

  • CRE example: “Explain the difference between ERV, passing rent, and headline rent in a UK multi-let office; list common traps when comparing deals.”

Technique 3: Map-before-territory

  • Intent: Build a conceptual map before you go deep.

  • How to use it: Ask for frameworks, taxonomies, checklists, and causal diagrams that organise a domain.

  • Typical mistakes: Diving into detailed drafting without a map; treating a single framework as “the truth”.

  • CRE example: “Create a taxonomy of data-centre underwriting risks: power, planning, customer concentration, capex, supply chain, grid constraints, and exit liquidity - then explain how each category shows up in diligence.”

Quadrant B: Scaffolded Production (Execution mode × Domain novice)

For:
Domain novices who need to produce work product.

Technique 1: Template generation

  • Intent: Reduce cognitive load by producing a strong structure that you fill with verified inputs.

  • How to use it: Request a deliverable skeleton with headings, subheadings, and placeholder prompts for required data.

  • Typical mistakes: Letting AI fill placeholders with invented content; confusing a template with a conclusion.

  • CRE example: “Generate an IC pack structure for a UK office acquisition. For each section, include: required inputs, typical analyses, and ‘red flag’ questions. Use placeholders rather than invented numbers.”

Technique 2: Checklist-driven execution

  • Intent: Shift quality from “hope” to “process control”.

  • How to use it: Ask the model to create a completion checklist and a review checklist (what to verify; what to cite; what cannot be assumed).

  • Typical mistakes: Only checking the final narrative, not the underlying assumptions and calculations.

  • CRE example: “Create a QA checklist for a lease abstract summary used in underwriting: items to verify, clauses that must be escalated, and what constitutes ‘insufficient evidence’.”

Technique 3: Role-based guardrails

  • Intent: Force the model to behave like a constrained assistant, not a confident commentator.

  • How to use it: Assign a role plus rules: ask clarifying questions, provide an assumptions register, avoid fabricating, flag uncertainties.

  • Typical mistakes: Overly theatrical roles (“be a genius investor”) instead of operational constraints; failing to require uncertainty labelling.

  • CRE example prompt pattern (short):
    Act as a CRE analyst. If key inputs are missing, ask for them. Use placeholders rather than guessing.

    Provide:
    (1) Assumptions register,
    (2) Data required list,
    (3) Draft output.

    Do not invent rents, yields, or regulatory claims.

Quadrant C: Accelerated Analysis (Discovery mode × Domain expert)

For:
Domain experts exploring new angles or complex problems.

Technique 1: Adversarial stress-testing

  • Intent: Use the model as a structured critic to surface fragility.

  • How to use it: Ask for a red-team review: “If this is wrong, how would it fail?” Force concrete failure pathways.

  • Typical mistakes: Asking for “pros and cons” (too generic); not tying critique to evidence requirements.

  • CRE example: “Red-team this investment thesis: ‘Grade A offices will rebound due to flight to quality’. Identify 8 ways it fails, what early signals would show failure, and what mitigations exist.”

Technique 2: Comparative framework generation

  • Intent: Avoid single-lens thinking by generating multiple analytical frames.

  • How to use it: Request 3–5 alternative frameworks (e.g., demand-side, regulatory, financing, operational) and compare what each highlights.

  • Typical mistakes: Treating frameworks as decoration rather than decision tools.

  • CRE example: “Analyse a workplace strategy decision using three frames: productivity/performance, talent/retention, and cost/risk. Where do they conflict?”

Technique 3: Assumption surfacing

  • Intent: Make hidden assumptions explicit so they can be tested.

  • How to use it: Ask the model to list assumptions, classify them (critical vs minor), and propose validation methods.

  • Typical mistakes: Allowing assumptions to remain embedded in prose.

  • CRE example: “List all assumptions embedded in a ‘stabilised income’ valuation narrative and propose which can be evidenced vs which require judgement.”

Technique 4: Perspective multiplication

  • Intent: Create decision robustness by simulating stakeholder lenses.

  • How to use it: Ask for outputs from specific perspectives: IC chair, lender credit, occupier, planner, ESG committee, asset manager.

  • Typical mistakes: Using caricatures; failing to produce actionable questions.

  • CRE example: “From a lender’s credit committee perspective, what are the five deal-breakers in this business plan? What evidence would satisfy each?”

Quadrant D: Precision Delegation (Execution mode × Domain expert)

For:
Domain experts in production mode.

Technique 1: Front-loaded context & constraints

  • Intent: Replace ambiguity with bounded execution.

  • How to use it: Provide the scenario, the decision context, constraints, definitions, what not to do, and your preferred format.

  • Typical mistakes: Under-specifying constraints; assuming the model knows your house style; forgetting confidentiality boundaries.

  • CRE example: “Draft the ‘Risks & Mitigations’ section in our IC style: short bullets, no marketing language, each risk must have an owner and a mitigation plan.”

Technique 2: Artefact-specific commands

  • Intent: Tell the model exactly what artefact you are producing (memo, IC section, lease clause summary, board slide) and the acceptance criteria.

  • How to use it: “Produce X, for audience Y, with length Z, and include A/B/C elements.”

  • Typical mistakes: Asking for “a report” instead of a defined artefact.

  • CRE example: “Write a 250-word executive summary for an IC: thesis, key numbers placeholders, top 3 risks, recommendation.”

Technique 3: Negative specification

  • Intent: Prevent known failure behaviours.

  • How to use it: Explicitly prohibit: invented numbers, uncited claims, legal conclusions, unbounded scope, buzzwords.

  • Typical mistakes: Only saying what you want, not what you don’t want.

  • CRE example: “Do not reference specific regulations unless I provide text. If uncertain, flag as ‘needs legal check’.”

Technique 4: Output format locking

  • Intent: Make outputs reusable and reviewable.

  • How to use it: Fix headings, bullet structure, and required sections (assumptions; data gaps; confidence).

  • Typical mistakes: Letting format drift across iterations; mixing analysis and narrative.

  • CRE example: “Use exactly these headings: Background; Investment case; Key sensitivities; Risks; Mitigations; Open questions.”

Technique 5: Iterative refinement

  • Intent: Improve quality through controlled revision rather than re-prompting from scratch.

  • How to use it: Run short cycles: “Revise only section 3; keep the rest unchanged; incorporate these three changes.”

  • Typical mistakes: Rewriting everything (introduces new errors); failing to track what changed.

  • CRE example: “Revise the leasing section to reflect a 12-month void assumption and add two mitigations; do not change other sections.”

The Expert’s Meta-Skill: Mode Switching (C → D → C → D)

Intent:
Experts get better outcomes by alternating between:

  • C (Diverge): generate options, risks, questions, frames

  • D (Converge/Draft): lock format, draft deliverable

  • C (Stress-test): red team, assumption checks, scenario attacks

  • D (Refine): tighten, align, finalise, QA

Typical mistake: Staying in one quadrant because it feels productive (either endless ideation or premature drafting).

CRE example: In underwriting, use C to expand risk hypotheses, D to draft the IC narrative, C to stress-test with lender/IC lenses, D to produce the final pack with an assumption register and clear decision asks.

WHY BOTHER
Perhaps this reads as prescriptive. In a way it is. Until it becomes muscle memory. Then you can iterate and ‘free solo’ your way to exceptional outcomes. The discipline is in knowing why you are doing X in Y order,

This is workflow control. You already manage risk with checklists, templates, and IC disciplines. This framework simply extends that discipline to AI-assisted work.

Use it as a structured assistant: to generate your own work frameworks, surface assumptions, draft formats, and run adversarial critique - tasks where reliability comes from constraints and review.

Do make the effort for ‘Mode switching’, because this replaces rework. It’s a slow way to get where you need to fast. Without pushing in then zooming out you miss so much, and so often that means endless versions before a final outcome. This foreshadows that by stress testing as you go.

In Quadrant D, which requires artefact-specific commands and format locking, you can upload your own house style and templates and jump straight to outputs that are ready to go.

And if you are worried that your messy data means none of this will work very well, this is exactly why you need assumption surfacing and evidence plans. The framework makes uncertainty explicit. Once known you can fill in the gaps.

Here’s a 10-minute quick start you can try out today.

  1. Pick one live task you’re doing this week (IC pack, lease summary, board note).

  2. Decide Discovery vs Execution. Decide whether you’re a novice or expert for that task.

  3. Run one prompt from the matching quadrant:

    • A: ask for a map and definitions

    • B: ask for a template with placeholders + checklist

    • C: ask for assumption register + red team

    • D: lock format + constraints + negative spec

  1. Add one mandatory control: “No invented numbers; use placeholders; list missing inputs.”

  2. Save the best prompt as your team’s starting template.


Working with this framework will lead to compounding returns. Referencing back to the Anthropic study, how we prompt the AI determines how effective it can be.

A closing caution: this framework makes AI collaboration more effective, but “more effective” applied to the wrong task simply accelerates waste. Prompting discipline without task discipline is efficiency theatre with better production values.

This is the third in a stack of frameworks, and sequence matters. RIRA asks the strategic question: how are we creating value—faster taxis, better taxis, or Uber? The CRE Automation Matrix asks the analytical question: what kind of work is this, and what’s the right human-AI interaction model? This Prompting Framework answers the operational question: how do we actually get that work done?

Work the stack top-down. If you haven’t passed the “faster taxis or Uber?” test, no prompting technique will rescue you from building the wrong thing more efficiently. Start with strategy, filter through analysis, then reach for this framework when you’re ready to execute properly.

Read More