2 Generative Adversarial Networkproceedings.mlr.press/v80/tao18b/tao18b.pdf · Chenyang Tao 1Liqun...

10
χ 2 Generative Adversarial Network Chenyang Tao 1 Liqun Chen 1 Ricardo Henao 1 Jianfeng Feng 2 Lawrence Carin 1 Abstract To assess the difference between real and syn- thetic data, Generative Adversarial Networks (GANs) are trained using a distribution discrep- ancy measure. Three widely employed measures are information-theoretic divergences, integral probability metrics, and Hilbert space discrep- ancy metrics. We elucidate the theoretical connec- tions between these three popular GAN training criteria and propose a novel procedure, called χ 2 - GAN, that is conceptually simple, stable at train- ing and resistant to mode collapse. Our procedure naturally generalizes to address the problem of simultaneous matching of multiple distributions. Further, we propose a resampling strategy that sig- nificantly improves sample quality, by repurpos- ing the trained critic function via an importance weighting mechanism. Experiments show that the proposed procedure improves stability and con- vergence, and yields state-of-art results on a wide range of generative modeling tasks. 1. Introduction Learning to sample from complicated distributions has at- tracted considerable recent interest, with many important applications (Zhu et al., 2017; Ledig et al., 2017; Yu et al., 2017; Hu et al., 2017). Likelihood-free models avoid the need to explicitly assume a particular parametrization of the data-generating distribution p G (x). Such models implic- itly define a distribution via generator G(z; θ): Z→X , and a latent random variable Z with pre-specified distribu- tion q(z). Samples from the generator are produced by first drawing Z q(z) and then feeding it through the generator. To match p G (x) to the true data distribution p d (x), one es- timates a discrepancy measure, d(p d ,p G ). In the GAN framework, the discrepancy is first estimated by maxi- mizing an auxiliary variational functional V (p d ,p G ; D): 1 Electrical & Computer Engineering, Duke University, Durham, NC 27708, USA 2 ISTBI, Fudan University, Shanghai, China. Cor- respondence to: Chenyang Tao <[email protected]>. Proceedings of the 35 th International Conference on Machine Learning, Stockholm, Sweden, PMLR 80, 2018. Copyright 2018 by the author(s). P×P→ R between distributions p d (x) and p G (x) sat- isfying d(p d ,p G ) = max D V (p d ,p G ; D), where P is the space of probability distributions and V (p d ,p G ; D) is esti- mated using samples from the two distributions. Function D(x; ω), parameterized by ω and known as the critic func- tion, is intended to maximally discriminate between samples of the two distributions. One seeks to match the generator distribution p G (x) to the unknown true distribution p d (x) by solving a minimax game between the critic and the gen- erator: min G max D V (p d ,p G ; D). Following ideas from the original GAN (Goodfellow et al., 2014), which optimizes the Jensen-Shannon divergence, much recent work has focused on information-theoretic divergences, such as the KL-divergence (Sønderby et al., 2017). Many other studies have investigated the gener- alized f -divergence (Csisz ´ ar, 1963), Div f (p d k p G ) , R f p d (x) p G (x) p G (x) dx, where f (·): R R is a convex function satisfying f (1) = 0, that summarizes the local dis- crepancy between p d (x) and p G (x). Nowozin et al. (2016) proposed an algorithm based on the variational formula- tion of Div f (p d k p G ), Uehara et al. (2016) explored in depth its density-ratio formulation, and Nock et al. (2017) further generalized it from an information-geometric per- spective. Interestingly, Mao et al. (2017) showed that a specific type of f -divergence, namely the χ 2 -divergence, can be directly optimized for GAN learning, by recasting it as a least-squares regression problem. However, Arjovsky & Bottou (2017) showed that when us- ing divergence-based objectives, the parameter updates for the generator can be either uninformative or numerically unstable, and divergence-based objectives may not be contin- uous wrt the generator parameters. These issues motivated development of GAN formulations based on Integral Prob- ability Metrics (IPMs) (M¨ uller, 1997). IPM models seek to optimize an objective of the form V IPM (p d ,p G ; D)= E Xp d [D(X; ω)] - E X 0 p G [D(X 0 ; ω)], where E Xp d [·] denotes the expectation wrt to distributions p d (x). When the critic D(x; ω) is chosen from a unit ball of Lipschitz-1 functions (kDk Lip 1), the IPM reduces to the Wasserstein- 1 or earth mover’s distance (Rubner et al., 2000). In this case, the challenge with d IPM (p d ,p G ) lies in constraining the critic function Lipschitz constant (Arjovsky et al., 2017; Gulrajani et al., 2017; Miyato et al., 2017).

Transcript of 2 Generative Adversarial Networkproceedings.mlr.press/v80/tao18b/tao18b.pdf · Chenyang Tao 1Liqun...

  • χ2 Generative Adversarial Network

    Chenyang Tao 1 Liqun Chen 1 Ricardo Henao 1 Jianfeng Feng 2 Lawrence Carin 1

    AbstractTo assess the difference between real and syn-thetic data, Generative Adversarial Networks(GANs) are trained using a distribution discrep-ancy measure. Three widely employed measuresare information-theoretic divergences, integralprobability metrics, and Hilbert space discrep-ancy metrics. We elucidate the theoretical connec-tions between these three popular GAN trainingcriteria and propose a novel procedure, called χ2-GAN, that is conceptually simple, stable at train-ing and resistant to mode collapse. Our procedurenaturally generalizes to address the problem ofsimultaneous matching of multiple distributions.Further, we propose a resampling strategy that sig-nificantly improves sample quality, by repurpos-ing the trained critic function via an importanceweighting mechanism. Experiments show that theproposed procedure improves stability and con-vergence, and yields state-of-art results on a widerange of generative modeling tasks.

    1. IntroductionLearning to sample from complicated distributions has at-tracted considerable recent interest, with many importantapplications (Zhu et al., 2017; Ledig et al., 2017; Yu et al.,2017; Hu et al., 2017). Likelihood-free models avoid theneed to explicitly assume a particular parametrization ofthe data-generating distribution pG(x). Such models implic-itly define a distribution via generator G(z; θ) : Z → X ,and a latent random variable Z with pre-specified distribu-tion q(z). Samples from the generator are produced by firstdrawing Z ∼ q(z) and then feeding it through the generator.

    To match pG(x) to the true data distribution pd(x), one es-timates a discrepancy measure, d(pd, pG). In the GANframework, the discrepancy is first estimated by maxi-mizing an auxiliary variational functional V (pd, pG;D) :

    1Electrical & Computer Engineering, Duke University, Durham,NC 27708, USA 2ISTBI, Fudan University, Shanghai, China. Cor-respondence to: Chenyang Tao .

    Proceedings of the 35 th International Conference on MachineLearning, Stockholm, Sweden, PMLR 80, 2018. Copyright 2018by the author(s).

    P × P → R between distributions pd(x) and pG(x) sat-isfying d(pd, pG) = maxD V (pd, pG;D), where P is thespace of probability distributions and V (pd, pG;D) is esti-mated using samples from the two distributions. FunctionD(x;ω), parameterized by ω and known as the critic func-tion, is intended to maximally discriminate between samplesof the two distributions. One seeks to match the generatordistribution pG(x) to the unknown true distribution pd(x)by solving a minimax game between the critic and the gen-erator: minG maxD V (pd, pG;D).

    Following ideas from the original GAN (Goodfellow et al.,2014), which optimizes the Jensen-Shannon divergence,much recent work has focused on information-theoreticdivergences, such as the KL-divergence (Sønderby et al.,2017). Many other studies have investigated the gener-alized f -divergence (Csiszár, 1963), Divf (pd ‖ pG) ,∫f(pd(x)pG(x)

    )pG(x) dx, where f(·) : R → R is a convex

    function satisfying f(1) = 0, that summarizes the local dis-crepancy between pd(x) and pG(x). Nowozin et al. (2016)proposed an algorithm based on the variational formula-tion of Divf (pd ‖ pG), Uehara et al. (2016) explored indepth its density-ratio formulation, and Nock et al. (2017)further generalized it from an information-geometric per-spective. Interestingly, Mao et al. (2017) showed that aspecific type of f -divergence, namely the χ2-divergence,can be directly optimized for GAN learning, by recasting itas a least-squares regression problem.

    However, Arjovsky & Bottou (2017) showed that when us-ing divergence-based objectives, the parameter updates forthe generator can be either uninformative or numericallyunstable, and divergence-based objectives may not be contin-uous wrt the generator parameters. These issues motivateddevelopment of GAN formulations based on Integral Prob-ability Metrics (IPMs) (Müller, 1997). IPM models seekto optimize an objective of the form VIPM(pd, pG;D) =EX∼pd [D(X;ω)] − EX′∼pG [D(X ′;ω)], where EX∼pd [·]denotes the expectation wrt to distributions pd(x). Whenthe critic D(x;ω) is chosen from a unit ball of Lipschitz-1functions (‖D‖Lip ≤ 1), the IPM reduces to the Wasserstein-1 or earth mover’s distance (Rubner et al., 2000). In thiscase, the challenge with dIPM(pd, pG) lies in constrainingthe critic function Lipschitz constant (Arjovsky et al., 2017;Gulrajani et al., 2017; Miyato et al., 2017).

  • Chi-square Generative Adversarial Network

    Separately, Reproducing Kernel Hilbert Space (RKHS) the-ory has motivated development of a powerful set of methodsto handle probability problems (Muandet et al., 2017). Inparticular, the embedding of probability measures via ker-nels (Sriperumbudur et al., 2010) has attracted significantinterest. Let κ(·, ·) be a positive definite function known asthe kernel function. The kernel embedding of distributionp(x) is given by νp(x) , EX∼p[κ(x,X)] . The MaximalMean Discrepancy MMD(pd, pG) , ‖νpd − νpG‖H, de-fines a distance metric on distributions pd(x) and pG(x),where ‖ · ‖H is the kernel-induced norm in H, a Hilbertspace. MMD readily translates into an algorithm that doesnot require the adversarial game for generative modeling (Liet al., 2015; Dziugaite et al., 2015). However, RKHS-basedgenerative models have high computational cost, while inthe case of generative models, also struggling when deal-ing with complex distributions (Bińkowski et al., 2018). Inpractice, good performance can be achieved with carefulhyperparameter tuning and by introducing auxiliary lossterms to the objective (Zhang et al., 2017; Li et al., 2017b).

    From a pragmatic perspective, GANs rarely converge tothe desired equilibrium (Arora & Zhang, 2018), insteadsettling for a sub-optimal local solution, where samplesproduced by the trained generative model often lack di-versity (Salimans et al., 2016). To alleviate these issues,most existing studies focus on seeking more stable architec-tures (Radford et al., 2016), enforcing heuristically-derivedor theoretically-inspired regularized objectives (Salimanset al., 2016; Warde-Farley & Bengio, 2017; Mescheder et al.,2017b; Roth et al., 2017), and procedures that leverage care-fully designed optimization paths (Karras et al., 2018).

    We present new theoretical insights on GAN-based gener-ative modeling, the cause of some of its difficulties, andprincipled solutions to address associated challenges. Ourkey contributions include: (i) We present theory connectingthree major generative modeling frameworks: divergence-,IPM- and kernel-based approaches. (ii) A novel, conceptu-ally simple procedure is introduced, termed χ2-GAN, thatis stable at training and embraces sample diversity duringgeneration. (iii) It is demonstrated that our formulationnaturally generalizes to problems requiring simultaneousmatching of multiple distributions. (iv) We propose to fullyexploit the learned critic function, by repurposing it as aweighting mechanism in a resampling procedure, leveraginguseful information from the critic to improve sample quality.

    2. Learning χ2 GANs2.1. Distribution Mixture and Generative Modeling

    Consider joint random variables (X,Y ), where X is drawnfrom mixture distribution [p(x) + q(x)]/2, and Y is a ran-dom variable identifying the mixture component from which

    X is drawn; Y = 1 if X is drawn from p(x), with Y = −1if X is drawn from q(x). We denote the joint density for(X,Y ) as µ(x, y; p, q); to avoid notational clutter, we of-ten omit its dependency on (p, q) when the context is clear.Further, let µ(x) and µ(y) be the marginals of µ(x, y).

    It can be readily verified that X and Y from µ(x, y) arestatistically independent if and only if p(x) and q(x) match.For subsequent generative modeling, p(x) is the true datadistribution, pd(x), and q(x) is the generator distribution,pG(x). We can therefore cast the problem of learning the pa-rameters of a generative model G(x; θ) as seeking to matchour generator distribution to that of the data, by minimizingthe statistical dependency between the data variable, X , andits label, Y .

    More generally, random variables (X,Y ) are manifestedwith X drawn from mixture model 1M

    ∑Mm=1 pm(x) and

    with Y ∈ {1, . . . ,M} identifying the mixture componentfrom which X is drawn. Extending the ideas discussedabove, we can jointly match M > 2 distributions {pm(x)},by minimizing the statistical dependency between data Xand mixture-component label Y .

    2.2. Covariance Operator and Statistical Dependency

    Let f(x) : X → R and g(y) : Y → R be two square-integrable functions defined over the domains of X and Y ,respectively. The covariance wrt the joint density µ(x, y) is

    cov(f, g) = EX,Y∼µ[(f(X)− E[f ])(g(Y )− E[g])],

    where E[f ] = EX [f(X)] is the marginal mean of f(x),similarly defined for E[g]. When X and Y are statisticallyindependent, i.e. µ(x, y) = µ(x)µ(y), cov(f, g) = 0 forevery choice of f(x) and g(y). We show below that thestatistical independency between X and Y is implied whencov(f, g) = 0 holds for every f(x) and g(y) chosen from asufficiently rich function space.

    Let HX and HY be two Hilbert spaces for functions de-fined on X and Y , equipped with inner products 〈·, ·〉HXand 〈·, ·〉HY respectively. In an RKHS, the kernel functiondefines the inner product, i.e., 〈·, ·〉H = κ(·, ·). The cross-covariance operator wrt the triplet {µ(x, y),HX ,HY } isdefined as the operator CXY : HX → HY satisfying

    〈g, CXY f〉HY = cov(f(X), g(Y )),

    for all f(x) ∈ HX and g(y) ∈ HY . The existence ofCXY is a direct result of the Riesz representation theorem(Yoshida, 1974). One can further define the covarianceoperator CXX : HX → HX as

    〈f1, CXXf2〉HX = cov(f1(x), f2(x)), (1)

    and similarly for CY Y .

  • Chi-square Generative Adversarial Network

    Further assume HX and HY are separable and let E ={ei}∞i=1 and F = {fi}∞i=1 be their respective CompleteOrthonormal Systems (CONS). The Hilbert-Schmidt norm‖ · ‖HS of an operator A : HX → HY is defined as

    ‖A‖2HS :=∞∑

    i,j=1

    〈Aei, fj〉2HY , (2)

    which is independent of the choice of the CONS (Adams &Fournier, 2003). It can be readily verified that for the cross-covariance operator ‖CXY ‖2HS = 0 implies cov(f, g) = 0for f(x) ∈ HX and g(y) ∈ HY . To understand when doesit imply independence, let us further denote the space ofsquare integrable functions wrt random variable X as L2X .The following result, adapted from Theorem 4 of Grettonet al. (2005a), states that whenHX andHY are sufficientlyrich, a vanishing ‖CXY ‖2HS implies the statistical indepen-dence of X and Y , and vice versa.

    Theorem 1. If HX and HY are dense in L2X and L2Y , re-spectively, then ‖CXY ‖2HS = 0 if and only if X and Y arestatistically independent.

    Proofs for all theoretical results are found in the Supplemen-tary Material (SM).

    The cross-covariance operator norm ‖CXY ‖2HS in (2) isat the core of the Hilbert-Schmidt Independence Crite-ria (HSIC) (Gretton et al., 2012); we are interested insuch criteria to assess the independence of X and Y dis-cussed in Sec. 2.1. In an RKHS, let K̃(n)X be the Grammatrix of samples {xi}ni=1, whose entries are defined as[K̃

    (n)X ]ij = κ(xi, xj), then the centralized Gram matrix is

    given by K(n)X = HnK̃(n)X Hn, where Hn , In −

    1n1n1

    Tn

    is the centralizing matrix, with a similar definition for K(n)Y ;1n is an n-dimensional vector of all ones, and In is the n×nidentity matrix). The empirical estimator for HSIC resultsin an elegant population-wise expression:

    ‖Ĉ(n)XY ‖2HS ,

    1

    n2tr(K

    (n)X K

    (n)Y

    ), (3)

    where tr(·) is the trace operation. Despite its simple mathe-matical expression, it is difficult to use HSIC as an optimiza-tion objective, because computations require quadratic time,O(n2), and (3) depends on the inner product of the Hilbertspace, which is not invariant to the kernel function spaceused. As a result, HSIC is known to be highly sensitive tothe choice of the kernel Hilbert space.

    2.3. Normalization of Cross-covariance Operator

    To circumvent the dependence on the form of the inner prod-uct, we consider the normalized cross-covariance operatorVXY : HX → HY (Baker, 1973), with slight abuse of

    notation, given by

    VXY = C−1/2Y Y CXY C

    −1/2XX .

    Intuitively, operators C−1/2XX and C−1/2Y Y defined as (3) via

    (1) normalize the covariances in function space, thus VXYcan be understood as the cross-correlation operator. Thefollowing Lemma formalizes this intuition.

    Lemma 2. For ‖f‖2HX = ‖g‖2HY = 1, we have

    〈g, VXY f〉HY = corr(C−1/2XX f(X), C

    −1/2Y Y g(Y )

    ).

    The next two theoretical results expand on the invariance of‖VXY ‖2HS to the choice of the function space.Proposition 3. If E = {ei}∞i is a complete orthonormalsystem inHX , then C−1/2XX E = {C

    −1/2XX ei}∞i is a complete

    orthonormal system in L2X .

    Corollary 4. Let E = {ei}∞i and F = {fj}∞j be therespective CONS forHX andHY , we have

    ‖VXY ‖2HS =∑i,j

    {corr

    (C−1/2XX ei, C

    −1/2Y Y fj

    )}2. (4)

    The above result shows that ‖VXY ‖2HS only depends onC−1/2XX E and C

    −1/2Y Y F , and not directly on the inner prod-

    uct of the Hilbert space. Further, in RKHS it is knownthat ‖VXY ‖2HS is invariant when the kernel functions arecharacteristic e.g., Gaussian kernels.

    Fukumizu et al. (2007) proposed a regularized empiricalestimator for ‖VXY ‖2HS in (4), given by

    ‖V̂ (n)XY ‖2HS = tr(R

    (n)X R

    (n)Y ), (5)

    where R(n)X = K(n)X (K

    (n)X + �nIn)

    −1 (similarly defined forR

    (n)Y ) and �n > 0 is the regularization parameter.

    Use of a metric like ‖V̂ (n)XY ‖2HS appears promising as a meansof assessing the independence of X and Y . However, inaddition to the challenge of selecting kernels for the RKHSs,(5) comes with significant computational overhead; the ma-trix inversion in R(n)X requiresO(n3) operations. While thisestimator is shown to be consistent, provided �n → 0 and�3nn→∞, empirical results indicate that �n must be care-fully tuned to avoid degenerate solutions. Further, despite itstheoretical invariance, empirical estimates vary significantlywrt the choice of kernels, while also not performing well inhigh-dimensional settings.

    2.4. χ2 Generative Adversarial Net

    Key to circumvent the excessive computational burden of(5) is that rather than explicitly computing C−1/2XX E and

  • Chi-square Generative Adversarial Network

    C−1/2Y Y F in (4), we can instead use some pre-specified func-

    tion spaces Ẽ and F̃ to estimate ‖VXY ‖2HS. Next, we derivespecific results on evaluating ‖VXY ‖2HS for the mixture dis-tribution µ(x, y) defined in Sec. 2.1.

    Let σ2µ(f) denote the variance of a function f(x) wrtrandom variable X with marginal µ(x). Let φ(x) =E[Y |X = x] be the conditional expectation of Y givenX , gχ(x) =

    φ(x)σµ(φ)

    is the variance normalized φ(x), andψ(x) = Pr(Y = +1|X = x) is the conditional probabilityof Y = +1 given X , i.e., the critic used in the originalGAN, D(x;ω). The χ2 mutual information of µ(x, y) isdefined as

    MIχ(µ) =∫∫ (

    µ(x, y)

    µ(x)µ(y)− 1)2

    µ(x)µ(y) dx dy,

    where µ(x) and µ(y) denote the marginals. The χ2 distancebetween p(x) and q(x) is thus defined as

    Disχ(p, q) =

    √∫(p(x)− q(x))2

    p(x)+q(x)2

    dx.

    The next proposition, connecting divergence, IPM andHilbert spaced based discrepancies constitutes our mainresult.

    Proposition 5. The following quantities are identical:

    (i) ‖VY X‖2HS, (ii) MIχ(µ),(iii) {Disχ(p, q)}4/(16‖φ(X)‖2µ), (iv) {corr(φ(X), Y )}2,(v) {corr(ψ(X), Y )}2,(vi) (EX1∼p[gχ(X1)]− EX2∼q[gχ(X2)])2/4.

    The equivalence between (i)-(iii) unveils the connectionsbetween the RKHS independence metric ‖VY X‖2HS, theinformation theoretic divergence metric MIχ(µ) and thevariance-constrained IPM metric Disχ(p, q) (Mroueh &Sercu, 2017). On the other hand, (iv)-(vi) provide us withpractical empirical estimators for these metrics. The key in-sight from Proposition 5 is that we only need to compute thecritic ψ(x) or φ(x) to estimate ‖VXY ‖2HS, or equivalently,the χ2 mutual information of the mixture distribution in Sec.2.1. Consequently, to formulate an optimization procedurefor generative modeling, we can minimize ‖VXY ‖2HS wrtthe generator. We call the framework χ2-GAN due to itsinherent connection to the χ2 metric.

    We now detail the construction of χ2-GAN based on es-timator (iv) from Proposition 5. Let pG(x) be the datagenerating distribution implicitly defined by (a deep neural)generator G(z; θ) parametrized by θ, Z ∼ q(z), and letD = {D(x;ω);ω ∈ Ω} be a parameterized family fromwhich we choose the critic D(x;ω). The (least-squares)

    Algorithm 1 χ2 GAN.Input: data {xi}, batchsize b, decay ρ, learning rate δ.for t = 1, 2, 3, . . . do

    1. Sample minibatch {xi ∼ pd(x), zi ∼ p(z)}bi=12. Update the critic D(x;ω) to improve (minimize)∑

    i(D(xi;ω)− 1)2 + (D(G(zi; θt−1);ω) + 1)2

    3. Update the variance σ2µ for D(x;ωt)4. Update the correlation estimatect(θt−1) = (1− ρ)ct−1

    + ρ2b∑i

    (D(xi;ωt)−D(G(zi;θt−1);ωt)

    σµ(D(x;ωt))

    )5. Update the generator G(z; θ)

    θt = θt−1 + δGradClip(∇θ(ct(θt−1))2

    )end for

    loss functions for the critic and generator are given by

    LD(ω|θ) = EX,Y∼µ[(D(X;ω)− Y )2],

    LG(θ|ω) =(EX∼pd [D(X;ω)]− EX∼pG [D(X;ω)]

    σµ(D(X;ω))

    )2.

    where µ(x, y) = 12pd(x)+12pG(x) as in Sec. 2.1. Note that

    µ(x, y) and σµ(·) are implicitly dependent on θ via pG(x).

    We further denote ω∗(θ) , arg minω LD(ω|θ), the optimalcritic parameters conditioned on the generator. When D isdense in L2X , we have

    D(x;ω∗(θ)) = φ(x), and LG(θ|ω∗(θ)) = 4‖VXY ‖2HS,

    which follows from Proposition 5. To match pG(x) to pd(x),we solve minθ LG(θ|ω∗(θ)). We propose to decouple theabove optimization scheme into the following GAN-likeiterations:

    ωt ← arg minω LD(ω|θt−1),θt ← arg minθ∈∆(θt−1) LG(θ|ωt, θt−1),

    where ∆(θt−1) denotes the trust region for the generatorupdate. In LG(θ|ωt, θt−1), we have replaced σµ(D(x;ω))with its stale estimate σµt−1(D(x;ω)). We regularize theupdate of the generator to make sure LG(θ|ωt) remains agood approximation to ‖VXY ‖2HS. This can be implementedwith proximal gradient descent, or simply via gradient clip-ping as summarized in Algorithm 1.

    2.5. Joint Matching of Multiple Distributions

    We now discuss the generalization to multi-component mix-tures. Since the M components in 1M

    ∑Mm=1 pm(x) are

    mutually exclusive, the space of L2Y has dimension M − 1.We use the following M − 1 empirical basis functions to

  • Chi-square Generative Adversarial Network

    Figure 1. Relations between likelihood-free generative models.

    span L2Y :

    γm(yl) =

    {1, m = l

    −(M − 1)−1, m 6= l ,

    for m ∈ {1, . . . ,M − 1} and l ∈ {1, . . . ,M}. Wecollect {γm(y)}M−1m=1 into a vector function Γ(y) =[γ1(y), · · · , γM−1(y)]T , with corresponding (M − 1) ×(M − 1) covariance matrix ΣY with elements [ΣY ]ij =1/(M − 1) if i = j and [ΣY ]ij = −1/(M − 1)2 oth-erwise. Note that EY [Γ(Y )] = 0, and this constructionexactly recovers our Y ∈ {±1} binary labeling whenM = 2. For the data side, we use critics ψm(x) =Pr(Y = ym|X = x) as the empirical basis, and simi-larly use a compact M − 1 dimensional vector representa-tion Ψ̃(x) = [ψ1(x), . . . , ψ(M−1)(x)]T . We use Ψ(x) =Ψ̃(x) − EX [Ψ̃(X)] for mean centering and denote its co-variance matrix as ΣX . The objective for multi-distributionmatching is then

    Vχ2 =∥∥∥EX,Y∼µ [Σ−1/2X Ψ(X)TΓ(Y )Σ−1/2Y ]∥∥∥2Fro ,

    where ‖ · ‖Fro denotes the Frobenius norm. In practice,we use cross-entropy loss to estimate Ψ̃(x), and leveragea moving average estimator to track the expectation andcovariance of Ψ(x), otherwise it is similar to Algorithm 1.The complete algorithm and additional remarks are foundin the SM.

    2.6. Importance Resampling

    Current practice in modeling with GANs discards the criticafter training, with the learned generator used for sampling.However, the generator distribution rarely reaches the de-sired target distribution on real-world complex datasets,while the trained critic contains useful local informationthat does not get incorporated into the generator duringtraining. Consider two cases: i) the generator does not haveenough capacity to characterize the target distribution, and

    Algorithm 2 Importance resampling.Input: Generator G(z; θ), critic D(x;ω), sample size n.1. Sample candidates: {xi = G(zi; θ)}ni=1, zi ∼ p(z).2. Compute importance weights:

    wi = w̃i(∑nj=1 w̃j)

    −1,

    where w̃i = ζ(D(xi;ω)).3. Sample j ∼ Cat(w1, . . . , wn).Return: xj = G(zj ; θ)

    approaches a solution that covers the support of the targetwithout properly capturing its topology; ii) the target hasdisjoint support regions and the generator covers each ofthem, but with inconsistent amounts of probability mass.In either case, as described below, the critic can be repur-posed to provide additional information to assist samplegeneration (after training).

    To fully harness the information from the critic to improvesample quality, we propose to resample the generator. Recallthat in importance sampling, one uses a proposal distributionq(x) to sample, and then one reweights samples by theirimportance weights w(x) , p(x)/q(x), to compute theexpectation wrt to the target distribution p(x), i.e.,

    EX∼p[f(X)] = EX∼q [w(X)f(X)] ≈1

    n

    n∑i=1

    p(xi)

    q(xi)f(xi),

    where {xi}ni=1 are iid samples from q(x). We proposetreating the data-generating distribution pG(x) as the pro-posal distribution, and collecting the importance weightsw(x) = pd(x)/pG(x) from the critic. For divergence-basedgenerative models (f -GAN, χ2-GAN, etc.), importanceweights can often be directly computed from the critic viasome simple transformation ζ(·) (see SM for details). Forother generative models, an auxiliary log-density ratio criticcan be trained with cross-entropy loss to track the impor-tance weights. We summarize the importance resamplingprocedure in Algorithm 2.

    3. Related WorkThe proposed χ2 GAN connects three popular likelihood-free generative modeling frameworks (see Figure 1). Itis derived from the theory of RKHS independence analy-sis and it can be shown that the popular MMD objective(Gretton et al., 2012) is an unnormalized version of our χ2

    objective (see SM). χ2 GAN optimizes a divergence criteriawith an IPM loss, using a critic trained with a stable least-squares loss, similar to that of LS-GAN (Mao et al., 2017).Regular divergence-based GANs directly optimize the di-vergence between pd(x) and pG(x), while χ2-GAN insteadoptimizes the divergence between µ(x, y) and µ(x)µ(y),which characterizes the independence between the sample

  • Chi-square Generative Adversarial Network

    Figure 2. Toy model comparison for GAN, MMD-GAN, WGAN-GP and χ2-GAN. Distributions are visualized with KDE plots.

    and associated label, i.e. the mutual information. This al-lows easy generalization beyond matching two distributions.Fisher GAN (Mroueh & Sercu, 2017) is closest to our χ2-GAN, which builds on the IPM framework and like ours,also normalizes the critic with its second moment. The keydifferences are that Fisher GAN relies on a more sophisti-cated augmented Lagrangian to optimize the same objectivefor both the critic and generator, while χ2 GAN decouplesthe critic and generator objectives, requiring simpler (un-constrained) stochastic-gradient-descent-type updates.

    For RKHS-based generative modeling the choice of kernelis crucial. Theoretical properties and empirical performancehave been analyzed for a number of popular kernels, such asinverse multi-quadratic (Gorham & Mackey, 2017), Plum-mer (Unterthiner et al., 2018), rational quadratic (Bińkowskiet al., 2018) and energy distance (Liu, 2017). Although de-rived from a kernel formulation, our training procedure is inpractice kernel-free.

    Modern generative modeling has deep roots in statisticaltesting. Prior studies have primarily focused on two sampletests (Gretton et al., 2012). Our study builds on the work ofindependence testing (Gretton et al., 2005b), which general-izes two sample tests and can be extended to settings withmultiple generators and critics (Durugkar et al., 2017).

    Simultaneous matching of multiple distributions is a keytechnique needed in many machine learning applications(Zhao et al., 2017). In the GAN context, generalization tothe JSD metric have been explored to address this challenge(Gan et al., 2017; Li et al., 2017a; Pu et al., 2018b). Howeverfor RKHS and IPM-based generative models, currently thereis no generalization, and one has to build M(M−1)2 pairwisecritics, which can be prohibitively expensive whenM growslarge. Our χ2 GAN represents the first attempt to bridge thisgap. Our theory implies using a quadratic number of criticsis unnecessary, and instead χ2 GAN computes an M − 1dimensional critic.

    Importance sampling is a classic technique used in Monte

    Table 1. Quantitative results on MNIST. † is estimated using AIS;‡ is reported in (Hu et al., 2018); § are likelihood-based models.

    Model log p(x) ≥ IS

    NF§ (Rezende & Mohamed, 2015) -85.1 -PixelRNN§ (Oord et al., 2016) -79.2 -AVB§ (Mescheder et al., 2017a) -79.5 -ASVAE§ (Pu et al., 2017) -81.14 -sVAE-r§ (Pu et al., 2018a) -79.26 † 9.12

    GAN (Goodfellow et al., 2014) -114.25 † 8.34 ‡

    WGAN-GP (Gulrajani et al., 2017) -79.92 † 8.45 ‡

    DCGAN (Radford et al., 2016) -79.47 † 8.93χ2 GAN (ours) -78.85 † 9.01

    Carlo methods (Liu, 2008). One of its key applications is toevaluate the quality of statistical models (Neal, 2001; Wuet al., 2017). Recently, this idea has been used in likelihood-based generative models to sharpen the variational bound(Burda et al., 2016), and has been extended to improvethe training of likelihood-assisted GAN variants (Hu et al.,2018). To the best of the authors’ knowledge, resamplingthe generator as proposed here has not yet been exploredbefore and, not being exclusive to our formulation, can beeasily used with other methods.

    4. ExperimentsWe consider a wide range of synthetic and real-worldtasks to experimentally validate χ2 GAN, and bench-mark it against other state-of-the-art solutions. All ex-periments are implemented with Tensorflow and run ona single NVIDIA TITAN X GPU. Details of the exper-imental setup are in the SM, and code for our experi-ments are available from https://www.github.com/chenyang-tao/chi2gan.

    4.1. Toy Distributions

    We compare χ2 GAN with one representative model fromeach category, namely the original GAN, WGAN-GP andMMD GAN, on three toy distributions. The same modelarchitecture is used for all models, with the exception ofMMD GAN which does not need an explicit critic function.The generation results are summarized in Figure 2.

    All models except MMD perform well on the baseline Swissroll experiment. The mixture-of-Gaussians experiments testalgorithm robustness to mode collapse. The original GANdemonstrates its vulnerability dealing with distributionswith disjoint modes, and MMD learns an overly smootheddistribution, even with carefully tuned kernel hyperparame-ter. Both WGAN-GP and χ2 GAN successfully learn goodapproximations to the target distribution, with the lattershowing faster convergence and more-stable training.

    https://www.github.com/chenyang-tao/chi2ganhttps://www.github.com/chenyang-tao/chi2gan

  • Chi-square Generative Adversarial Network

    Table 2. Unsupervised Inception Score on CIFAR-10

    Model IS

    ALI (Dumoulin et al., 2017) 5.34 ± .05DCGAN (Radford et al., 2016) 6.16 ± .07MMD-GAN (Li et al., 2017b) 6.17 ± .07WGAN-GP (Gulrajani et al., 2017) 6.56 ± .05ASVAE (Pu et al., 2017) 6.89 ± .05sVAE-r (Pu et al., 2018a) 6.96 ± .066χ2-GAN (ours) 7.47 ± 0.105WGAN-GP ResNet 7.86 ± .07Fisher-GAN ResNet (Mroueh & Sercu, 2017) 7.90 ± .05χ2-GAN ResNet (ours) 7.88 ± .10

    4.2. Image Datasets

    We trained χ2 GAN on a number of popular image datasetsto demonstrate its ability to learn complex distributions forreal-world applications. For supervised generation tasks,we condition the generator on the label of an image. Toquantitatively evaluate model performance, we consider thefollowing metrics in our experiments: (1) Inception score(IS) (Salimans et al., 2016) for datasets associated with one-hot labels; (2) AIS score (Wu et al., 2017) to estimate thelog likelihood.

    We only report the quantitative results for the raw generatordistribution in the main text, and results for the importanceresampled generator are found separately in SM. Two net-work architectures were considered in these experiments:DCGAN and ResNet. In all experiments we have usedXaiver initialization and Adam optimizer. All images shownare random samples and not cherry picked. We note betterquantitative results have been reported in the literature usingspecific techniques orthogonal to our main contributions(Karras et al., 2018; Warde-Farley & Bengio, 2017; Miyatoet al., 2017; Salimans et al., 2018), see SM for a discussion.

    MNIST We used the binarized MNIST in this experimentand compared with the results from prior results in Table1. Our χ2 GAN achieves an AIS score of −78.85 natsand an IS score of 9.01. These results lead the chart forall likelihood-free generative models we considered, andthey are comparable to, or even better than those from thebest-performing likelihood-based models.

    Cifar10 For this dataset, we experimented with both un-supervised and supervised generation tasks. Quantitativeresults are summarized in Tables 2 and 3. For both tasks, χ2

    GAN consistently achieved state-of-the-art results obtainedwith the network architectures considered. Most notably,our χ2 GAN significantly outperformed DCGAN, MMDGAN and WGAN-GP in the unsupervised generation taskwith the DCGAN architecture. We also provide quantitativeresults with the Fréchet Inception Distance (FID) (Heuselet al., 2017) in SM. See Figure 3 for qualitative assessment.

    CelebA We provide a comparison of DCGAN, Fisher GAN

    Table 3. Supervised Inception Score on CIFAR-10

    Model IS

    SteinGAN (Wang & Liu, 2016) 6.16 ± .07DCGAN (Radford et al., 2016) 6.58 ± .05AC-GAN (Odena et al., 2017) 8.25 ± .07SGAN (Huang et al., 2017) 8.59 ± .12Fisher-GAN ResNet (Mroueh & Sercu, 2017) 8.16 ± .12WGAN-GP ResNet (Gulrajani et al., 2017) 8.42 ± .10χ2-GAN ResNet (ours) 8.44 ± .10

    Table 4. Unsupervised Inception Score on ImageNet

    Model IS

    DCGAN (Radford et al., 2016) 5.965PixelCNN++ (Salimans et al., 2017) 7.65ASVAE (Pu et al., 2017) 11.14

    χ2-GAN (ours) 11.34

    and our χ2 GAN on the face generation task in Figure 4. Wetrained DCGAN and χ2 GAN to generate the face samples,and collected Fisher GAN’s samples from the original paper.All models used the DCGAN architecture. We observe thatχ2 GAN produced (subjectively) more compelling samples,capturing facial details, illumination and more realistic tex-tures compared with its counterparts. Additional samplesare shown in the SM, and we find no evidence of modecollapse. We also provide additional experimental evidencein the SM to verify χ2-GAN learns generalizable featuresrather than remembering training examples.

    We also use the face-generation task to demonstrate the ef-ficacy of importance resampling. In Figure 4 we comparethe accepted and rejected samples1. Those less-compellingsamples produced by the generator are immediately identi-fied based on the importance score. Quantitive results forimportance resampling can be found in the SM.

    ImageNet We also considered ImageNet to evaluate thescalability of the models on large datasets. All images areresized to 64×64, and the quantitative results are reported inTable 4. With the simple DCGAN architecture, our χ2 GANsignificantly outperformed more-sophisticated PixelCNN++GAN, even surpassing the performance of likelihood-basedASVAE. See the SM for generated samples.

    Stability, robustness, convergence and sample diversityIn all our experiments on the image datasets, χ2 GANdemonstrated stable training dynamics. It showed a similarconvergence rate 2 compared with WGAN-GP in terms ofiterations, but much cheaper per-iteration cost. χ2 GANalso demonstrated robustness as we varied model architec-tures, network normalization schemes and hyperparameters

    1We used a less well-trained model and picked our samplesbased on the importance weights to highlight the difference.

    2Wrt IS score and visual inspection, see SM.

  • Chi-square Generative Adversarial Network

    Figure 3. Generated images on Cifar-10. Left: unsupervised generation; right: supervised generation.

    Figure 4. Face generation quality comparison. From left to right: DCGAN, Fisher GAN, χ2 GAN, high importance weight samples andlow importance weight samples.

    Figure 5. MNIST image translation. First column is the inputimage from test set, subsequent columns are sampled translations.

    (results not shown). We also do not find any evidence formode collapse in our experiments.

    4.3. Matching Multiple Distributions

    To demonstrate the flexibility of the χ2 GAN frameworkto match multiple distributions, we consider the followingtranslation task: Given paired examples {(xi, zi)}ni=1, sam-ple from all possible translations z for a new observationx, and vice versa. More explicitly, consider the distributiontriplet: p0(x, z) = pd(x, z), p1(x, z) = pd(x)q1(z|x; θ1),and p2(x, z) = pd(z)q2(x|z; θ2), where q1(z|x; θ1) andq2(x|z; θ2) are the translation models. When translationsare faithful to the data distribution, we have p0(x, z) =p1(x, z) = p2(x, z). Here (x, z) are paired data, and weconsider the problem of image-to-image translation.

    Rotated MNIST In this experiment we pair each MNISTdigit with a random sample of the same type, rotated by90°. Our translation results are presented in Figure 5. It isobserved that χ2 GAN translations achieved both fidelity

    Figure 6. Edges2shoes translation.

    and diversity for this task.

    Edges-to-shoes We evaluate the performance on the more-realistic edges-to-shoes dataset, where the model learns totranslate between shoes and sketches. As shown in Figure6, χ2 GAN learned to produce faithful translations.

    5. ConclusionsWe have developed a framework that unifies prior theoret-ical frameworks for likelihood-free generative modelling,and based on this we proposed a novel algorithm namedχ2 GAN. Our approach is conceptually simple, and can bereadily generalized to match multiple distributions. Em-pirical evidence verified that this new method offers com-petitive performance on a wide range of generation tasks.For future work, we intend to investigate its connections tothe likelihood-based generative models, and to seek novelapplications by integrating it with other machine learningalgorithms.

  • Chi-square Generative Adversarial Network

    AcknowledgementsThe authors would like to thank the anonymous reviewersfor their insightful comments. This research was supportedin part by DARPA, DOE, NIH, ONR and NSF. J Feng ispartially supported by the key project of Shanghai Science& Technology Innovation Plan (No. 16JC1420402). Theauthors would also like to thank Prof. C Leng, Dr. Y Zhang,Dr. Y Pu and K Bai for fruitful discussions.

    ReferencesAdams, R. A. and Fournier, J. J. Sobolev spaces, volume

    140. Academic press, 2003.Arjovsky, M. and Bottou, L. Towards principled methods

    for training generative adversarial networks. In NIPSWorkshop. 2017.

    Arjovsky, M., Chintala, S., and Bottou, L. Wassersteingenerative adversarial networks. In ICML, 2017.

    Arora, S. and Zhang, Y. Do GANs actually learn the distri-bution? an empirical study. In ICLR, 2018.

    Baker, C. R. Joint measures and cross-covariance operators.Transactions of the American Mathematical Society, 186:273–289, 1973.

    Bińkowski, M., Sutherland, D. J., Arbel, M., and Gretton,A. Demystifying mmd gans. In ICLR, 2018.

    Burda, Y., Grosse, R., and Salakhutdinov, R. Importanceweighted autoencoders. In ICLR, 2016.

    Csiszár, I. Eine informationstheoretische Ungleichung undihre Anwendung auf den Beweis der Ergodizitat vonMarkoffschen Ketten. Publ. Math. Inst. Hungar. Acad, 8:85–108, 1963.

    Dumoulin, V., Belghazi, I., Poole, B., Lamb, A., Arjovsky,M., Mastropietro, O., and Courville, A. Adversariallylearned inference. In ICLR, 2017.

    Durugkar, I., Gemp, I., and Mahadevan, S. Generativemulti-adversarial networks. In ICLR, 2017.

    Dziugaite, G. K., Roy, D. M., and Ghahramani, Z. Traininggenerative neural networks via maximum mean discrep-ancy optimization. In UAI, 2015.

    Fukumizu, K., Gretton, A., Sun, X., and Schölkopf, B. Ker-nel measures of conditional dependence. In NIPS, 2007.

    Gan, Z., Chen, L., Wang, W., Pu, Y., Zhang, Y., Liu, H.,Li, C., and Carin, L. Triangle generative adversarialnetworks. In NIPS, 2017.

    Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B.,Warde-Farley, D., Ozair, S., Courville, A., and Bengio, Y.Generative adversarial nets. In NIPS, 2014.

    Gorham, J. and Mackey, L. Measuring sample quality withkernels. In ICML, 2017.

    Gretton, A., Bousquet, O., Smola, A., and Scholkopf, B.

    Measuring statistical dependence with Hilbert-Schmidtnorms. In ALT, 2005a.

    Gretton, A., Herbrich, R., Smola, A., Bousquet, O., andSchölkopf, B. Kernel methods for measuring indepen-dence. The Journal of Machine Learning Research, 6:2075–2129, 2005b.

    Gretton, A., Borgwardt, K. M., Rasch, M. J., Schölkopf,B., and Smola, A. A kernel two-sample test. Journal ofMachine Learning Research, 13(Mar):723–773, 2012.

    Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., andCourville, A. Improved training of Wasserstein GANs.arXiv preprint arXiv:1704.00028, 2017.

    Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., andHochreiter, S. Gans trained by a two time-scale updaterule converge to a local nash equilibrium. In NIPS, 2017.

    Hu, Z., Yang, Z., Liang, X., Salakhutdinov, R., and Xing,E. P. Toward controlled generation of text. In ICML,2017.

    Hu, Z., Yang, Z., Salakhutdinov, R., and Xing, E. P. Onunifying deep generative models. In ICLR, 2018.

    Huang, X., Li, Y., Poursaeed, O., Hopcroft, J., and Belongie,S. Stacked generative adversarial networks. In CVPR,2017.

    Karras, T., Aila, T., Laine, S., and Lehtinen, J. Progressivegrowing of GANs for improved quality, stability, andvariation. In ICLR, 2018.

    Ledig, C., Theis, L., Huszár, F., Caballero, J., Cunningham,A., Acosta, A., Aitken, A., Tejani, A., Totz, J., Wang, Z.,et al. Photo-realistic single image super-resolution usinga generative adversarial network. In CVPR, 2017.

    Li, C., Liu, H., Chen, C., Pu, Y., Chen, L., Henao, R.,and Carin, L. Alice: Towards understanding adversariallearning for joint distribution matching. In NIPS, 2017a.

    Li, C.-L., Chang, W.-C., Cheng, Y., Yang, Y., and Póczos,B. Mmd gan: Towards deeper understanding of momentmatching network. In NIPS, 2017b.

    Li, Y., Swersky, K., and Zemel, R. Generative momentmatching networks. In ICML, 2015.

    Liu, J. S. Monte Carlo strategies in scientific computing.Springer Science & Business Media, 2008.

    Liu, L. On the two-sample statistic approach to generativeadversarial networks. Master’s thesis, Princeton Univer-sity, 2017.

    Mao, X., Li, Q., Xie, H., Lau, R. Y., Wang, Z., andPaul Smolley, S. Least squares generative adversarialnetworks. In ICCV. 2017.

    Mescheder, L., Nowozin, S., and Geiger, A. Adversarialvariational Bayes: Unifying variational autoencoders andgenerative adversarial networks. In ICML, 2017a.

    Mescheder, L., Nowozin, S., and Geiger, A. The numerics

  • Chi-square Generative Adversarial Network

    of gans. In NIPS, 2017b.Miyato, T., Kataoka, T., Koyama, M., and Yoshida, Y. Spec-

    tral normalization for generative adversarial networks. InICML Implicit Models Workshop, 2017.

    Mroueh, Y. and Sercu, T. Fisher GAN. In NIPS. 2017.Mroueh, Y., Sercu, T., and Goel, V. McGAN: Mean and

    covariance feature matching gan. In ICML, 2017.Muandet, K., Fukumizu, K., Sriperumbudur, B., Schölkopf,

    B., et al. Kernel mean embedding of distributions: Areview and beyond. Foundations and Trends® in MachineLearning, 10(1-2):1–141, 2017.

    Müller, A. Integral probability metrics and their generatingclasses of functions. Advances in Applied Probability, 29(2):429–443, 1997.

    Neal, R. M. Annealed importance sampling. Statistics andcomputing, 11(2):125–139, 2001.

    Nock, R., Cranko, Z., Menon, A. K., Qu, L., andWilliamson, R. C. f-gans in an information geometricnutshell. In NIPS, 2017.

    Nowozin, S., Cseke, B., and Tomioka, R. f-GAN: Traininggenerative neural samplers using variational divergenceminimization. In NIPS, 2016.

    Odena, A., Olah, C., and Shlens, J. Conditional imagesynthesis with auxiliary classifier GANs. In ICML, 2017.

    Oord, A. v. d., Kalchbrenner, N., and Kavukcuoglu,K. Pixel recurrent neural networks. arXiv preprintarXiv:1601.06759, 2016.

    Pu, Y., Wang, W., Henao, R., Chen, L., Gan, Z., Li, C., andCarin, L. Adversarial symmetric variational autoencoder.In NIPS, 2017.

    Pu, Y., Chen, L., Dai, S., Wang, W., Li, C., and Carin, L.Symmetric variational autoencoder and connections toadversarial learning. In AISTATS, 2018a.

    Pu, Y., S., D., Gan, Z., Wang, W., G., W., Y., Z., Henao,R., and Carin, L. Joint gan: Multi-domain joint distribu-tion learning with generative adversarial nets. In ICML,2018b.

    Radford, A., Metz, L., and Chintala, S. Unsupervised rep-resentation learning with deep convolutional generativeadversarial networks. In ICLR, 2016.

    Rezende, D. J. and Mohamed, S. Variational inference withnormalizing flows. In ICML, 2015.

    Roth, K., Lucchi, A., Nowozin, S., and Hofmann, T. Stabi-lizing training of generative adversarial networks throughregularization. In NIPS, 2017.

    Rubner, Y., Tomasi, C., and Guibas, L. J. The earth mover’sdistance as a metric for image retrieval. Internationaljournal of computer vision, 40(2):99–121, 2000.

    Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V.,

    Radford, A., and Chen, X. Improved techniques fortraining GANs. In NIPS, 2016.

    Salimans, T., Karpathy, A., Chen, X., and Kingma, D. P. Pix-elcnn++: Improving the pixelcnn with discretized logisticmixture likelihood and other modifications. In ICLR,2017.

    Salimans, T., Zhang, H., Radford, A., and Metaxas, D. Im-proving gans using optimal transport. In ICLR, 2018.

    Sønderby, C. K., Caballero, J., Theis, L., Shi, W., andHuszár, F. Amortised map inference for image super-resolution. In ICLR, 2017.

    Sriperumbudur, B. K., Gretton, A., Fukumizu, K.,Schölkopf, B., and Lanckriet, G. R. Hilbert space embed-dings and metrics on probability measures. Journal ofMachine Learning Research, 11(Apr):1517–1561, 2010.

    Uehara, M., Sato, I., Suzuki, M., Nakayama, K., and Mat-suo, Y. Generative adversarial nets from a density ratioestimation perspective. arXiv preprint arXiv:1610.02920,2016.

    Unterthiner, T., Nessler, B., Klambauer, G., Heusel, M.,Ramsauer, H., and Hochreiter, S. Coulomb GANs: Prov-ably optimal nash equilibria via potential fields. In ICLR,2018.

    Wang, D. and Liu, Q. Learning to draw samples: Withapplication to amortized MLE for generative adversariallearning. arXiv:1611.01722, 2016.

    Warde-Farley, D. and Bengio, Y. Improving generativeadversarial networks with denoising feature matching. InICLR, 2017.

    Wu, Y., Burda, Y., Salakhutdinov, R., and Grosse, R. On thequantitative analysis of decoder-based generative models.In ICLR. 2017.

    Yoshida, K. Functional analysis. Grundlehren der mathe-matischen Wissenschaften. Springer-Verlag, 1974.

    Yu, L., Zhang, W., Wang, J., and Yu, Y. Seqgan: Sequencegenerative adversarial nets with policy gradient. In AAAI,2017.

    Zhang, Y., Gan, Z., Fan, K., Chen, Z., Henao, R., Shen,D., and Carin, L. Adversarial feature matching for textgeneration. In ICML, 2017.

    Zhao, H., Zhang, S., Wu, G., Costeira, J. P., Moura, J. M.,and Gordon, G. J. Multiple source domain adapta-tion with adversarial training of neural networks. arXivpreprint arXiv:1705.09684, 2017.

    Zhu, J.-Y., Park, T., Isola, P., and Efros, A. A. Unpairedimage-to-image translation using cycle-consistent adver-sarial networks. In ICCV, 2017.