Generative Artificial Intelligence (GenAI) and a Taxonomy of Undesirable Results
Eric E. Cohen & Gregory J. Gerard
Abstract
This paper makes the case that the term “hallucinations” is inadequate for describing undesirable results from Generative Artificial Intelligence (GenAI) output. We explain hallucinations, how the term is used, and why it is inadequate. We propose a taxonomy for categorizing these undesirable results, facilitating clearer communication about them. This taxonomy aids in understanding, troubleshooting, and developing resilience against undesirable results, which is important for corporate governance as GenAI integrates both specifically into enterprise resource planning systems, and also software, generally. We give several examples illustrating how our taxonomy provides the necessary language to describe and address undesirable results. We discuss the implications of using the taxonomy widely, particularly in the context of text-to-text Large Language model output, and consider its relevance across different GenAI modalities, as common GenAI multimodal solutions expand.
Evidence weight
Balanced mode · F 0.40 / M 0.15 / V 0.05 / R 0.40
| F · citation impact | 0.50 × 0.4 = 0.20 |
| M · momentum | 0.50 × 0.15 = 0.07 |
| V · venue signal | 0.50 × 0.05 = 0.03 |
| R · text relevance † | 0.50 × 0.4 = 0.20 |
† Text relevance is estimated at 0.50 on the detail page — for your query’s actual relevance score, open this paper from a search result.