Legal Overfitting
Ian Ayres & Yair Listokin
Abstract
A central concern of machine learning is overfitting—which occurs when a prediction model includes too many explanatory variables and predicts noise. The problem with overfitting is that it leads the model to poorer out-of-sample predictions because it misattributes causal significance to irrelevant variables. We argue that a similar phenomenon reduces the quality of precedential reasoning. A judge or lawyer trying to reconcile prior authoritative opinions that are decided with some noise to make the best prediction about the outcome of a new case may misattribute causal significance to factors that do not offer precedential guidance. Machine learning has developed a series of estimation and diagnostic techniques to reduce the likelihood of overfitting. For example, it is standard to train models on a subset of the available data and then test how well the model predicts out-of-sample. This article argues that the quality of precedential reasoning would be improved if judges used analogous techniques in deciding cases. Lawyers might also use these techniques to improve their ability to predict what the law is. We present a decision tree estimation of copyright law’s fair use defense to illustrate the phenomenon of overfitting and how it might be limited.
Evidence weight
Balanced mode · F 0.40 / M 0.15 / V 0.05 / R 0.40
| F · citation impact | 0.50 × 0.4 = 0.20 |
| M · momentum | 0.50 × 0.15 = 0.07 |
| V · venue signal | 0.50 × 0.05 = 0.03 |
| R · text relevance † | 0.50 × 0.4 = 0.20 |
† Text relevance is estimated at 0.50 on the detail page — for your query’s actual relevance score, open this paper from a search result.