Final Presentation
Final Presentation
Article Review
Introduction
Deep neural networks have achieved state-of-the-art
performance on a wide variety of machine learning tasks.
While most research emphasizes Bayesian deep learning for uncertainty, authors
results highlight the potential of non-Bayesian methods, hoping to inspire further
exploration in this direction.
Deep Ensembles: A Simple Recipe For
Predictive Uncertainty Estimation
Problem setup and High-level
summary
Where:
is a function that evaluates the quality of the predictive distribution
relative to an event .
denotes the true distribution on (y, x) tuples.
A proper scoring rule is the one where with equality if and only if
NNs can then be trained according to measure that encourages calibration of predictive
uncertainty by minimizing the loss
Ensembles: training and prediction
Authors focus only on the randomization based approach as it is better suited for distributed, parallel
computation.
They came to a conclusion that training on entire dataset with random initialization was better than
bagging for deep ensembles.
Experimental Results
Evaluation metrics and experimental setup
Evaluation Metrics and Experimental Details
Evaluation Metrics:
Both Classification & Regression:
Negative Log Likelihood (NLL) for evaluating predictive uncertainty.
Classification:
Classification Accuracy.
Brier Score (BS): and 0 otherwise.
Experimental Setup:
Batch size: 100.
Optimizer: Adam with a fixed learning rate of 0.1.
Adversarial Training: set ϵ to 0.01 times the range of the training data for each dimension.
Weight Initialization: Default in Torch.
Classification on MNIST