Deep and interpretable regression models for ordinal outcomes

Lucas Kook, Lisa Herzog, Torsten Hothorn, Oliver Dürr, Beate Sick*

*Corresponding author for this work

Publication: Scientific journalJournal articlepeer-review

Abstract

Outcomes with a natural order commonly occur in prediction problems and often the available input data are a mixture of complex data like images and tabular predictors. Deep Learning (DL) models are state-of-the-art for image classification tasks but frequently treat ordinal outcomes as unordered and lack interpretability. In contrast, classical ordinal regression models consider the outcome's order and yield interpretable predictor effects but are limited to tabular data. We present ordinal neural network transformation models (ONTRAMs), which unite DL with classical ordinal regression approaches. ONTRAMs are a special case of transformation models and trade off flexibility and interpretability by additively decomposing the transformation function into terms for image and tabular data using jointly trained neural networks. The performance of the most flexible ONTRAM is by definition equivalent to a standard multi-class DL model trained with cross-entropy while being faster in training when facing ordinal outcomes. Lastly, we discuss how to interpret model components for both tabular and image data on two publicly available datasets.

Original languageEnglish
Article number108263
JournalPattern Recognition
Volume122
DOIs
Publication statusPublished - Feb 2022
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2021

Keywords

  • Deep learning
  • Distributional regression
  • Interpretability
  • Ordinal regression
  • Transformation models

Cite this