Carolina Cuesta-Lázaro
Arnau Quera-Bofarull
(Joseph Bullock)
2 months PhD placement at IBEX innovations, as part of the CDT program
Carolina / Arnau
Cosmology
Joe
Particle Physics
Detect
collimator
Segment
Open beam
Bone
Soft-tissue
Kazeminia, S., Karimi, et al (2015)
Extracts features from a high dimensional feature space, once trained on a particular dataset.
DEEP LEARNING
http://cs229.stanford.edu/proj2017/final-reports/5241462.pdf
Švihlík et al (2014)
http://cs.swansea.ac.uk/~csadeline/projects_astro.html
Badrinarayanan et al (2015)
Semantics (what)
MaxPooling
Translation Invariance
Location (where)
Receptive field grows very slowly !
Not tractable to learn long range correlations
1 | 0 |
---|---|
2 | 3 |
1 | 1 | 0 | 0 |
---|---|---|---|
1 | 1 | 0 | 0 |
2 | 2 | 3 | 3 |
2 | 2 | 3 | 3 |
Upsample
1 | 0 |
---|---|
2 | 3 |
1 | 0 | 0 | 0 |
---|---|---|---|
0.6 | 0.5 | 0 | 0 |
0.3 | 1 | 2 | 3 |
1.5 | 2 | 1.4 | 0.3 |
Other layers
...
2 | 1 |
---|---|
0.5 | 1.4 |
2 | 0 | 0 | 1 |
---|---|---|---|
0 | 0 | 0 | 0 |
0 | 0 | 0 | 1.4 |
0 | 0.5 | 0 | 0 |
1 | 2 |
---|---|
1 | 0 |
1 | 0 | 2 |
---|---|---|
2 | 1 | 1 |
1 | 0 | 0 |
Learned filter
1 | 1 | 0 | 0 |
---|---|---|---|
0 | 0 | 0 | 0 |
0 | 0 | 0 | 0 |
0 | 0 | 0 | 0 |
Stride 2 : Filter moves 2 pixels in the output for every pixel in the input
Feature map
1 | 1+4 | 2 | 2 |
---|---|---|---|
0 | 0+2 | 0 | 0 |
0 | 0 | 0 | 0 |
0 | 0 | 0 | 0 |
Sum where overlap!
Badrinarayanan et al (2015)
Skip connections to combine
information at different scales
Deconvolutions to
upsample
Up-weight the loss function
at the boundaries
Trained on 30 images for cell segmentation !
Ronneberger et al (2015)
Increases convexity and smoothness of loss function
Pyramid of
Dilated Convolutions
Fast
Chen et al (2018)
Chen et al (2018)
Upsampling
Duplicate
Indices unpooling
Deconvolution
Interpolated
Preserve detail
Encoder - Decoder
Dilated Convolutions
Pyramid of convolutions
Architecture
Skip Connections
Asymmetric
Image credit : www.grabcad.com
Coursera
Small (compared to typical DL datasets) and unbalanced.
How can we do Deep Learning with ~150 images?
We generate new images by applying combinations of:
We typically divide our dataset into three subsets:
First attempts focused on a very simplified SegNet model.
Going deeper has limits ( limited image size, GPU memory bottleneck, overfitting).
$$ L := - \sum_{i=1}^3 \color{royalblue}y_i \color{cadetblue}\log (f(x)_i) \color{black}+ \color{orange} \lambda \color{black}\sum_j \color{Maroon}\theta_j^2$$
Loss function = Cross-entropy + Regularisation
Regularisation hyperparameter $$\lambda = 5 \cdot 10^{-4}$$
Network parameters
Danger! K-fold cross-validation needed.
Metrics | F1-Score | AUC | Accuracy | Confidence |
---|---|---|---|---|
Weighted Average | 0.92 | 0.98 | 92% | 97% |
Calibration?
TP Rate
FP Rate
True Label
Predicted Label
$$\text{Confidence}(X) = \frac{1}{| X|} \sum_{i \in X}p_i$$
Good calibration: Confidence ~ Accuracy
% of samples
Probability
The network outputs 3 probability maps.
Soft tissue probability map
Apply threshold to probability map -> impose confidence level
Probability
Kazeminia, S., Karimi, et al (2015)
Cryptocurrency Times