3 minute read

Under active development — findings and figures may change.

AI is everywhere in academic research. Kobak et al. (2025, Science Advances) tracked words that language models overuse — “delve,” “nuanced,” “meticulous” — across 14 million biomedical abstracts and found at least 13.5% of 2024 papers were processed by an LLM. The same pattern shows up in the real estate literature, as a quick replication on 100K real estate papers indexed by OpenAlex shows.

That is the writing layer in the research process. The more consequential shift is deeper. A growing number of papers rely on AI not for drafting but for execution — work that could not exist without machine learning carrying out the core analysis. Bartik, Gupta and Milo (2025) read thousands of municipal zoning codes and built regulation measures that no research team could produce by hand. Calainho, van de Minne and Francke (2024) replaced linear hedonic models with ML on 30,000 New York transactions and showed systematic gains in out-of-sample accuracy. Shen and Ross (2021) extracted a description-quality measure from MLS listing text that captures soft information about property quality invisible to structured data. Leow and Lindenthal (2025) applied the Gu-Kelly-Xiu ML asset-pricing framework to REIT factor returns and showed substantial forecast improvements over OLS.

In each case, AI enables a measurement or prediction the research requires. Remove it and the paper disappears. But the role is still that of a skilled research assistant: executing tasks specified by a human. The PI — the person deciding what to study and why — remains human.

Core research competencies: AI vs human (self-assessment)

AI outperforms human researcher in many dimension (speaking for myself, obviously). The question is whether it can shine higher up the value chain. Can LLMs suggest research topics that are genuinely innovative and plausibly doable — functioning more as a PI than as an RA? Do humans still have a competitive edge?

The new working paper tests this. I mapped the full published corpus of Real Estate Economics (1,676 articles, 1973–2026) and real-estate-relevant subsets of JREFE, JUE, AER, JF, and RFS into a shared semantic embedding space. The result is a coordinate system for the field — not a literature review, but a map. Against that map, I generated 1,499 research ideas under eight conditions, varying what context the model received: nothing, the full corpus, individual cluster seeds, methods borrowed from economics and finance, methods from psychology. Each idea was scored on atypicality (a measure of unusual knowledge combination that retroactively predicts citations) and mapped back into the research space.

The figure below shows where generated ideas land. Grey dots are the full corpus; blue dots are REE papers; red dots are AI-generated ideas. Condition A is naïve generation from training data alone. Condition F draws on methods and paradigms from economics and finance journals.

Naïve generation (Condition A)
A: Naïve — no context provided
Econ/finance paradigm transfer (Condition F)
F: Paradigm transfer from economics & finance

Methodological scaffolding moves ideas outward into less-explored territory. Topical scaffolding alone does not. The best ideas — particularly those generated through method transfer from economics and finance — score comparably to the median published paper on the citation-predictive criterion. Some land squarely on papers published twenty years ago, having rediscovered questions the field already answered. But that is also true of human research proposals.

There is an uncomfortable regularity in how AI gets adopted: if a system offers a plausible-looking shortcut for a task it was never designed for, people will happily use it anyway, and it takes a lot of effort to convince them of the limits. Researchers will use LLMs to generate research ideas. They already do. The useful question is not whether this is a misguided idea but what the machines actually serve them when they try — and under what conditions the output is worth anything. That is what this paper is about.

Working paper (PDF)Interactive idea explorer