Do large language models (really) need statistical foundations?

Weijie Su

Annals of Applied Statistics2026https://doi.org/10.1214/26-aoas2151article
AJG 2ABDC A
Weight
0.37

Abstract

Large language models (LLMs) represent a new paradigm for processing unstructured data, with applications across an unprecedented range of domains. In this paper we address, through two arguments, whether the development and application of LLMs would genuinely benefit from foundational contributions from the statistics discipline. First, we argue affirmatively, beginning with the observation that LLMs are inherently statistical models due to their profound data dependency and stochastic generation processes, where statistical insights are naturally essential for handling variability and uncertainty. Second, we argue that the persistent black-box nature of LLMs—stemming from their immense scale, architectural complexity, and development practices often prioritizing empirical performance over theoretical interpretability—renders closed-form or purely mechanistic analyses generally intractable, thereby necessitating statistical approaches due to their flexibility and often demonstrated effectiveness. To substantiate these arguments, the paper outlines several research areas—including alignment, watermarking, uncertainty quantification, evaluation, and data mixture optimization—where statistical methodologies are critically needed and are already beginning to make valuable contributions. We conclude with a discussion suggesting that statistical research concerning LLMs will likely form a diverse “mosaic” of specialized topics, rather than deriving from a single unifying theory, and highlight the importance of timely engagement by our statistics community in LLM research.

1 citation

Open via your library →

Cite this paper

https://doi.org/https://doi.org/10.1214/26-aoas2151

Or copy a formatted citation

@article{weijie2026,
  title        = {{Do large language models (really) need statistical foundations?}},
  author       = {Weijie Su},
  journal      = {Annals of Applied Statistics},
  year         = {2026},
  doi          = {https://doi.org/https://doi.org/10.1214/26-aoas2151},
}

Paste directly into BibTeX, Zotero, or your reference manager.

Flag this paper

Do large language models (really) need statistical foundations?

Flags are reviewed by the Arbiter methodology team within 5 business days.


Evidence weight

0.37

Balanced mode · F 0.40 / M 0.15 / V 0.05 / R 0.40

F · citation impact0.16 × 0.4 = 0.06
M · momentum0.53 × 0.15 = 0.08
V · venue signal0.50 × 0.05 = 0.03
R · text relevance †0.50 × 0.4 = 0.20

† Text relevance is estimated at 0.50 on the detail page — for your query’s actual relevance score, open this paper from a search result.