Let's agree to disagree: Learning highly debatable multirater labelling

CH Sudre, BG Anson, S Ingala, CD Lane… - … Image Computing and …, 2019 - Springer
CH Sudre, BG Anson, S Ingala, CD Lane, D Jimenez, L Haider, T Varsavsky, R Tanno
Medical Image Computing and Computer Assisted Intervention–MICCAI 2019: 22nd …, 2019Springer
Classification and differentiation of small pathological objects may greatly vary among
human raters due to differences in training, expertise and their consistency over time. In a
radiological setting, objects commonly have high within-class appearance variability whilst
sharing certain characteristics across different classes, making their distinction even more
difficult. As an example, markers of cerebral small vessel disease, such as enlarged
perivascular spaces (EPVS) and lacunes, can be very varied in their appearance while …
Abstract
Classification and differentiation of small pathological objects may greatly vary among human raters due to differences in training, expertise and their consistency over time. In a radiological setting, objects commonly have high within-class appearance variability whilst sharing certain characteristics across different classes, making their distinction even more difficult. As an example, markers of cerebral small vessel disease, such as enlarged perivascular spaces (EPVS) and lacunes, can be very varied in their appearance while exhibiting high inter-class similarity, making this task highly challenging for human raters. In this work, we investigate joint models of individual rater behaviour and multi-rater consensus in a deep learning setting, and apply it to a brain lesion object-detection task. Results show that jointly modelling both individual and consensus estimates leads to significant improvements in performance when compared to directly predicting consensus labels, while also allowing the characterization of human-rater consistency.
Springer
Showing the best result for this search. See all results