This paper considers the generation of prediction intervals (PIs) by neural networks for quantifying uncertainty in regression tasks. It is axiomatic that high-quality PIs should be as narrow as possible, whilst capturing a specified portion of data. We derive a loss function directly from this axiom that requires no distributional assumption. We show how its form derives from a likelihood principle, that it can be used with gradient descent, and that model uncertainty is accounted for in ensembled form. Benchmark experiments show the method outperforms current state-of-the-art uncertainty quantification methods, reducing average PI width by over 10%.
High-Quality Prediction Intervals for Deep Learning: A Distribution-Free, Ensembled Approach
Tim Pearce,A. Brintrup,Mohamed H. Zaki,A. Neely
Published 2018 in International Conference on Machine Learning
ABSTRACT
PUBLICATION RECORD
- Publication year
2018
- Venue
International Conference on Machine Learning
- Publication date
2018-02-20
- Fields of study
Mathematics, Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-30 of 30 references · Page 1 of 1