ChatGPT is incredible (at being average)

Ihor Rudko & Aysan Bashirpour Bonab

Ethics and Information Technology2025https://doi.org/10.1007/s10676-025-09845-2article
AJG 1ABDC B
Weight
0.59

Abstract

In this article, we examine a peculiar issue apropos large language models (LLMs) and generative AI more broadly: the frequently overlooked phenomenon of output homogenization. It describes the tendency of chatbots to structure their outputs in a highly recognizable manner, which often amounts to the aggregation of verbal, visual, and narrative clichés, trivialities, truisms, predictable argumentations, and similar. We argue that the most appropriate conceptual lens through which said phenomenon can be framed is that of Frankfurtian bullshit. In this respect, existing attempts at applying the BS framework to LLMs are insufficient, as those are chiefly presented in opposition to the so-called algorithmic hallucinations. Here, we contend that further conceptual rupture from the original metaphor of Frankfurt (1986) is needed, distinguishing between the what-BS, which manifests in falsehoods and factual inconsistencies of LLMs, and the how-BS, which reifies in the dynamics of output homogenization. We also discuss how issues of algorithmic biases and model collapse can be framed as critical instances of the how-BS. The homogenization problem, then, is more significant than it initially appears, potentially exhibiting a powerful structuring effect on individuals, organizations, institutions, and society at large. We discuss this in the concluding section of the article.

13 citations

Open via your library →

Cite this paper

https://doi.org/https://doi.org/10.1007/s10676-025-09845-2

Or copy a formatted citation

@article{ihor2025,
  title        = {{ChatGPT is incredible (at being average)}},
  author       = {Ihor Rudko & Aysan Bashirpour Bonab},
  journal      = {Ethics and Information Technology},
  year         = {2025},
  doi          = {https://doi.org/https://doi.org/10.1007/s10676-025-09845-2},
}

Paste directly into BibTeX, Zotero, or your reference manager.

Flag this paper

ChatGPT is incredible (at being average)

Flags are reviewed by the Arbiter methodology team within 5 business days.


Evidence weight

0.59

Balanced mode · F 0.40 / M 0.15 / V 0.05 / R 0.40

F · citation impact0.60 × 0.4 = 0.24
M · momentum0.82 × 0.15 = 0.12
V · venue signal0.50 × 0.05 = 0.03
R · text relevance †0.50 × 0.4 = 0.20

† Text relevance is estimated at 0.50 on the detail page — for your query’s actual relevance score, open this paper from a search result.