Loss Surfaces, Mode Connectivity, and Fast Ensembling of DNNs

T. Garipov,Pavel Izmailov,Dmitrii Podoprikhin,D. Vetrov,A. Wilson

Published 2018 in Neural Information Processing Systems

ABSTRACT

The loss functions of deep neural networks are complex and their geometric properties are not well understood. We show that the optima of these complex loss functions are in fact connected by simple curves over which training and test accuracy are nearly constant. We introduce a training procedure to discover these high-accuracy pathways between modes. Inspired by this new geometric insight, we also propose a new ensembling method entitled Fast Geometric Ensembling (FGE). Using FGE we can train high-performing ensembles in the time required to train a single model. We achieve improved performance compared to the recent state-of-the-art Snapshot Ensembles, on CIFAR-10, CIFAR-100, and ImageNet.

PUBLICATION RECORD

  • Publication year

    2018

  • Venue

    Neural Information Processing Systems

  • Publication date

    2018-02-27

  • Fields of study

    Mathematics, Computer Science

  • Identifiers
  • External record

    Open on Semantic Scholar

  • Source metadata

    Semantic Scholar

CITATION MAP

EXTRACTION MAP

CLAIMS

  • No claims are published for this paper.

CONCEPTS

  • No concepts are published for this paper.

REFERENCES

Showing 1-26 of 26 references · Page 1 of 1

CITED BY

Showing 1-100 of 880 citing papers · Page 1 of 9