Learning from Simulated and Unsupervised Images through Adversarial Training

Simulated+Unsupervised (S+U) learning is proposed to learn a model to improve the realism of a simulator’s output using unlabeled real data, while preserving the annotation information from the simulator

Released in: Learning from Simulated and Unsupervised Images through Adversarial Training

Source: Learning from Simulated and Unsupervised Images through Adversarial Training

Contributor:

Summary

With recent progress in graphics, it has become more tractable to train models on synthetic images, potentially avoiding the need for expensive annotations. However, learning from synthetic images may not achieve the desired performance due to a gap between synthetic and real image distributions. To reduce this gap, authors propose Simulated+Unsupervised (S+U) learning, where the task is to learn a model to improve the realism of a simulator’s output using unlabeled real data, while preserving the annotation information from the simulator.

Authors develop a method for S+U learning that uses an adversarial network similar to Generative Adversarial Networks (GANs), but with synthetic images as inputs instead of random vectors. They make several key modifications to the standard GAN algorithm to preserve annotations, avoid artifacts, and stabilize training: (i) a ‘self-regularization’ term, (ii) a local adversarial loss, and (iii) updating the discriminator using a history of refined images.

They show that this enables generation of highly realistic images, which we demonstrate both qualitatively and with a user study. Authors quantitatively evaluate the generated images by training models for gaze estimation and hand pose estimation. This shows a significant improvement over using synthetic images, and achieve state-of-the-art results on the MPIIGaze dataset without any labeled real data.

2017

Year Released

Key Links & Stats

mjdietzx / SimGAN

Learning from Simulated and Unsupervised Images through Adversarial Training

MIT License

Learning from Simulated and Unsupervised Images through Adversarial Training

@misc{https://doi.org/10.48550/arxiv.1612.07828, doi = {10.48550/ARXIV.1612.07828}, url = {https://arxiv.org/abs/1612.07828}, author = {Shrivastava, Ashish and Pfister, Tomas and Tuzel, Oncel and Susskind, Josh and Wang, Wenda and Webb, Russ}, keywords = {Computer Vision and Pattern Recognition (cs.CV), Machine Learning (cs.LG), Neural and Evolutionary Computing (cs.NE), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {Learning from Simulated and Unsupervised Images through Adversarial Training}, publisher = {arXiv}, year = {2016}, copyright = {arXiv.org perpetual, non-exclusive license} }

ML Tasks

  1. General
  2. Image Generation

ML Platform

  1. Tensorflow

Modalities

  1. General

Verticals

  1. General

CG Platform

  1. Not Applicable

Related organizations

Apple Inc