Shortcomings of deep learning for distributional predictors: a note

Bonnie B. Smith et al.

Biostatistics2025https://doi.org/10.1093/biostatistics/kxaf051article
ABDC A
Weight
0.50

Abstract

A number of domains in biomedical research use data with a large number of predictors all representing the same type of measurement. Often, an important summary is the within-person distribution of these predictors. Here we focus on settings where the mean relationship between outcome and predictors is fully captured by this distribution and, more generally, on problems where the goal is to learn a mapping that is invariant under permutations of the input vector. We compare unstructured neural networks, which do not explicitly incorporate the permutation invariance property, versus networks that we call ordered predictors neural networks. We show in simulations that the unstructured deep learning approach can yield higher prediction errors, compared to the approach that explicitly leverages the invariance to simplify the learning task. Additionally, in the context of neural Bayes estimation, in which neural networks are used to construct point estimators, we show that ordered predictors neural networks can yield substantially more precise estimators. We therefore recommend that, when permutation invariance is known or suspected to hold, investigators use a learning or statistical modeling approach that can leverage the invariance, rather than an unstructured deep learning approach.

Open via your library →

Cite this paper

https://doi.org/https://doi.org/10.1093/biostatistics/kxaf051

Or copy a formatted citation

@article{bonnie2025,
  title        = {{Shortcomings of deep learning for distributional predictors: a note}},
  author       = {Bonnie B. Smith et al.},
  journal      = {Biostatistics},
  year         = {2025},
  doi          = {https://doi.org/https://doi.org/10.1093/biostatistics/kxaf051},
}

Paste directly into BibTeX, Zotero, or your reference manager.

Flag this paper

Shortcomings of deep learning for distributional predictors: a note

Flags are reviewed by the Arbiter methodology team within 5 business days.


Evidence weight

0.50

Balanced mode · F 0.40 / M 0.15 / V 0.05 / R 0.40

F · citation impact0.50 × 0.4 = 0.20
M · momentum0.50 × 0.15 = 0.07
V · venue signal0.50 × 0.05 = 0.03
R · text relevance †0.50 × 0.4 = 0.20

† Text relevance is estimated at 0.50 on the detail page — for your query’s actual relevance score, open this paper from a search result.