A Loss Function for Generative Neural Networks Based on Watson's Perceptual Model

Steffen Czolbe,Oswin Krause,Ingemar J. Cox,C. Igel

Published 2020 in Neural Information Processing Systems

ABSTRACT

To train Variational Autoencoders (VAEs) to generate realistic imagery requires a loss function that reflects human perception of image similarity. We propose such a loss function based on Watson's perceptual model, which computes a weighted distance in frequency space and accounts for luminance and contrast masking. We extend the model to color images, increase its robustness to translation by using the Fourier Transform, remove artifacts due to splitting the image into blocks, and make it differentiable. In experiments, VAEs trained with the new loss function generated realistic, high-quality image samples. Compared to using the Euclidean distance and the Structural Similarity Index, the images were less blurry; compared to deep neural network based losses, the new approach required less computational resources and generated images with less artifacts.

PUBLICATION RECORD

  • Publication year

    2020

  • Venue

    Neural Information Processing Systems

  • Publication date

    2020-06-26

  • Fields of study

    Mathematics, Computer Science, Engineering

  • Identifiers
  • External record

    Open on Semantic Scholar

  • Source metadata

    Semantic Scholar

CITATION MAP

EXTRACTION MAP

CLAIMS

  • No claims are published for this paper.

CONCEPTS

  • No concepts are published for this paper.

REFERENCES

Showing 1-30 of 30 references · Page 1 of 1

CITED BY

Showing 1-63 of 63 citing papers · Page 1 of 1