Convergence of Non-Convex Non-Concave GANs Using Sinkhorn Divergence
نویسندگان
چکیده
Sinkhorn divergence is a symmetric normalization of entropic regularized optimal transport. It smooth and continuous metrized weak-convergence with excellent geometric properties. We use it as an alternative for the minimax objective function in formulating generative adversarial networks. The optimization defined objective, under non-convex non-concave condition. This work focuses on optimization’s convergence stability. propose first order sequential stochastic gradient descent ascent (SeqSGDA) algorithm. Under some mild approximations, learning converges to local points. Using structural similarity index measure (SSIM), we supply non-asymptotic analysis algorithm’s rate. Empirical evidences show rate, which inversely proportional number iterations, when tested tiny colour datasets Cats CelebA deep convolutional networks ResNet neural architectures. entropy regularization parameter $\varepsilon $ approximated SSIM tolerance $\epsilon . determine that iteration complexity return -stationary point be $\mathcal {O}\left ({\kappa \, \log (\epsilon ^{-1})}\right)$ , where $\kappa value depends divergence’s convexity step ratio SeqSGDA
منابع مشابه
On Convergence Rate of Concave-Convex Procedure
Concave-Convex Procedure (CCCP) has been widely used to solve nonconvex d.c.(difference of convex function) programs occur in learning problems, such as sparse support vector machine (SVM), transductive SVM, sparse principal componenent analysis (PCA), etc. Although the global convergence behavior of CCCP has been well studied, the convergence rate of CCCP is still an open problem. Most of d.c....
متن کاملNon-Negative Matrix Factorization with Sinkhorn Distance
Non-negative Matrix Factorization (NMF) has received considerable attentions in various areas for its psychological and physiological interpretation of naturally occurring data whose representation may be parts-based in the human brain. Despite its good practical performance, one shortcoming of original NMF is that it ignores intrinsic structure of data set. On one hand, samples might be on a m...
متن کاملDual Iterative Hard Thresholding: From Non-convex Sparse Minimization to Non-smooth Concave Maximization
Iterative Hard Thresholding (IHT) is a class of projected gradient descent methods for optimizing sparsity-constrained minimization models, with the best known efficiency and scalability in practice. As far as we know, the existing IHT-style methods are designed for sparse minimization in primal form. It remains open to explore duality theory and algorithms in such a non-convex and NP-hard prob...
متن کاملOn the Convergence of the Concave-Convex Procedure
The concave-convex procedure (CCCP) is a majorization-minimization algorithm that solves d.c. (difference of convex functions) programs as a sequence of convex programs. In machine learning, CCCP is extensively used in many learning algorithms like sparse support vector machines (SVMs), transductive SVMs, sparse principal component analysis, etc. Though widely used in many applications, the con...
متن کاملParameter Convergence in systems with Convex/Concave Parameterization
2 Statement of the Problem A large class of problems in parameter estimation concerns systems where parameters occur nonlinearly. In [1]-[5], a stability framework for identification and control of such systems has been established. We address the issue of parameter convergence in such systems in this paper. Sufficient conditions under which parameter estimates converge to their true values are...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Access
سال: 2021
ISSN: ['2169-3536']
DOI: https://doi.org/10.1109/access.2021.3074943