Sean Meling Murray, 2018
Deep learning models have been wildly successful at a range of predictive tasks, but they have lacked a principled way of expressing uncertainty about their predictions.
Uncertainty approximations are critical in many practical applications, such as self-driving cars or AI-assisted diagnostic medicine.
These papers introduce a mathematically principled approach to uncertainty estimation in different kinds of neural networks
Applies MC dropout to a binary diagnostic setting in health care
Gal et. al. (2016)
Leibig et. al. (2017)
Monte Carlo Dropout
Is the approximated uncertainty obtained from MC dropout in convolutional neural networks a useful quantity for a multi-class setting?
Additionally, we examine if there is any information to be gained from establishing a connection between the uncertainty and the runner-up prediction.
Neural networks are essentially a cascade of
matrix multiplications followed by non-linear transformations.
Image source: https://www.digitaltrends.com/cool-tech/what-is-an-artificial-neural-network/
CNNs are a specialized kind of network that take advantage of the grid-like topology of image and time-series data.
Feature map
Image,input
Kernel, filter, feature, receptive field
2D map that indicates the presence of a feature in the input
Dropout is a regularisation technique used to prevent overfitting in neural networks.
Image source: deeplearningbook.org
Temporarily removes all connections from a node to the rest of the network at training time.
At test time dropout is turned off and weights are scaled to match expected output, can be viewed as model averaging.
Correct | n | uncertainty |
---|---|---|
0 | 2084 | 0.1802 |
1 | 7916 | 0.1021 |
(Note: Validation accuracy was 76.14%.)
Correct | n |
---|---|
0 | 921 |
1 | 9071 |
If we had used runner up predictions in the cases where the model misclassifies:
Accuracy goes from
Suggests that the runner up prediction contains useful information.
Runner up correct |
n |
prob1 |
prob2 |
uncertainty |
---|---|---|---|---|
0 | 921 | 0.5085 | 0.2156 | 0.1780 |
1 | 1163 | 0.5740 | 0.2644 | 0.1819 |
Correct | n |
---|---|
0 | 2084 |
1 | 7916 |
Leibig et. al. (2017) explore MC dropout in a binary setting:
Define a referral threshold based on uncertainty, and measure accuracy of predictions after referral to a human expert.
In a multi-class setting, we propose incorporating the runner up prediction for use in referrals.
ratio of two different measures of uncertainty
softmax prediction
softmax runner up
uncertainty associated with prediction
0
1
Infinity
Idea:
referral criteria | correct | incorrect |
---|---|---|
uncertainty >= .18 | 1556 | 1064 |
tau <= 1 | 673 | 907 |
false negatives