This is a list of papers I've worked on organized by topic. I've tried to give a brief summary for each paper.


Conditional image synthesis with auxiliary classifier gans
A Odena, C Olah, J Shlens
ICML 2017
This was arguably the first paper in which GANs were made to work on the ImageNet dataset. We gave a way to use label information to improve image synthesis performance.

Self-attention generative adversarial networks
H Zhang, I Goodfellow, D Metaxas, A Odena
ICML 2019 (Long Talk)
We propose several tweaks to the GAN training procedure that dramatically improve image synthesis performance. BigGAN is based on this work. Source code is available here.

Is Generator Conditioning Causally Related to GAN Performance?
A Odena, J Buckman, C Olsson, T B Brown, C Olah, C Raffel, I Goodfellow
ICML 2018
We show that the conditioning of the input-output jacobian of GAN generators is predictive of many GAN training pathologies. We then give evidence that the relationship is causal by conducting an intervention that clips the range of the jacobian singular values.

Discriminator Rejection Sampling
S Azadi, C Olsson, T Darrell, I Goodfellow, A Odena
ICLR 2019
We show that GAN discriminators can be used after training is finished to perform rejection sampling on GAN generators.

Skill Rating for Generative Models
C Olsson, S Bhupatiraju, T B Brown, A Odena, I Goodfellow
We show how to use chess-style tournament ranking to evaluate GANs and other generative models.

Open Questions about Generative Adversarial Networks
A Odena
Distill (Commentary)
I give a set of Open Problems that I think machine learning researchers working on GANs ought to think about.

Semi-Supervised Learning

Semi-supervised learning with generative adversarial networks
A Odena
Workshop on Data-Efficient Machine Learning (ICML 2016)
I invented (concurrent with this paper) a techique for using GANs to do semi-supervised learning.

Realistic Evaluation of Deep Semi-Supervised Learning Algorithms
A Oliver*, A Odena*, C Raffel*, ED Cubuk, IJ Goodfellow
NeurIPS 2018 (Spotlight)
We argue that existing methods for evaluating semi-supervised learning techniques are flawed and propose a new framework for doing these evaluations. Source code is available here.

Machine Learning and Computer Systems

TensorFuzz: Debugging neural networks with coverage-guided fuzzing
A Odena, C Olsson, D Anderson, I Goodfellow
ICML 2019 (Long Talk)
We apply the notions of coverage-guided-fuzzing and property-based-testing to neural networks. We show that approximate-nearest-neighbors algorithms can give useful coverage metrics in this context. Source code is available here.

Faster Asynchronous SGD
A Odena
Workshop on Optimization Methods for the Next Generation of Machine Learning (ICML) 2016
I speed up asynchronous SGD by quantifying gradient update staleness in terms of moving averages of gradient statistics.

Miscellaneous Machine Learning

Deconvolution and checkerboard artifacts
A Odena, V Dumoulin, C Olah
We show that the ubiquitous "deconvolution" operation used in image-upsampling produces strange checkerboard-artifacts. We then propose a simple fix.

Changing Model Behavior at Test-Time Using Reinforcement Learning
A Odena, D Lawson, C Olah
ICLR 2017 (Workshop Track)
I show how to change the test-time resource-usage of neural networks on a per-input basis using reinforcement learning.