Large Language Models: An Applied Econometric Framework

Jens Ludwig et al.

Annual Review of Economics2026https://doi.org/10.1146/annurev-economics-120925-105620preprint
AJG 3ABDC A*
Weight
0.46

Abstract

Large language models (LLMs) enable researchers to analyze text at unprecedented scale and minimal cost. Researchers can now revisit old questions and tackle novel ones with rich data. We provide an econometric framework for realizing this potential in two empirical uses. For prediction problems—forecasting outcomes from text—valid conclusions require “no training leakage” between the LLM's training data and the researcher's sample, which can be enforced through careful model choice and research design. For estimation problems—automating the measurement of economic concepts for downstream analysis—valid downstream inference requires combining LLM outputs with a small validation sample to deliver consistent and precise estimates. Absent a validation sample, researchers cannot assess possible errors in LLM outputs, and consequently seemingly innocuous choices (which model, which prompt) can produce dramatically different parameter estimates. When used appropriately, LLMs are powerful tools that can expand the frontier of empirical economics.

4 citations

Open via your library →

Cite this paper

https://doi.org/https://doi.org/10.1146/annurev-economics-120925-105620

Or copy a formatted citation

@article{jens2026,
  title        = {{Large Language Models: An Applied Econometric Framework}},
  author       = {Jens Ludwig et al.},
  journal      = {Annual Review of Economics},
  year         = {2026},
  doi          = {https://doi.org/https://doi.org/10.1146/annurev-economics-120925-105620},
}

Paste directly into BibTeX, Zotero, or your reference manager.

Flag this paper

Large Language Models: An Applied Econometric Framework

Flags are reviewed by the Arbiter methodology team within 5 business days.


Evidence weight

0.46

Balanced mode · F 0.40 / M 0.15 / V 0.05 / R 0.40

F · citation impact0.37 × 0.4 = 0.15
M · momentum0.60 × 0.15 = 0.09
V · venue signal0.50 × 0.05 = 0.03
R · text relevance †0.50 × 0.4 = 0.20

† Text relevance is estimated at 0.50 on the detail page — for your query’s actual relevance score, open this paper from a search result.