Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
The Unreasonable Effectiveness of Deep Features as a Perceptual Metric
11.609
Zitationen
5
Autoren
2018
Jahr
Abstract
While it is nearly effortless for humans to quickly assess the perceptual similarity between two images, the underlying processes are thought to be quite complex. Despite this, the most widely used perceptual metrics today, such as PSNR and SSIM, are simple, shallow functions, and fail to account for many nuances of human perception. Recently, the deep learning community has found that features of the VGG network trained on ImageNet classification has been remarkably useful as a training loss for image synthesis. But how perceptual are these so-called "perceptual losses"? What elements are critical for their success? To answer these questions, we introduce a new dataset of human perceptual similarity judgments. We systematically evaluate deep features across different architectures and tasks and compare them with classic metrics. We find that deep features outperform all previous metrics by large margins on our dataset. More surprisingly, this result is not restricted to ImageNet-trained VGG features, but holds across different deep architectures and levels of supervision (supervised, self-supervised, or even unsupervised). Our results suggest that perceptual similarity is an emergent property shared across deep visual representations.
Ähnliche Arbeiten
ImageNet classification with deep convolutional neural networks
2017 · 75.547 Zit.
Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks
2017 · 21.389 Zit.
Nonlinear total variation based noise removal algorithms
1992 · 15.414 Zit.
High-Resolution Image Synthesis with Latent Diffusion Models
2022 · 12.213 Zit.
Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network
2017 · 12.076 Zit.