Convergence of Non-Convex Non-Concave GANs Using Sinkhorn Divergence

نویسندگان

چکیده

Sinkhorn divergence is a symmetric normalization of entropic regularized optimal transport. It smooth and continuous metrized weak-convergence with excellent geometric properties. We use it as an alternative for the minimax objective function in formulating generative adversarial networks. The optimization defined objective, under non-convex non-concave condition. This work focuses on optimization’s convergence stability. propose first order sequential stochastic gradient descent ascent (SeqSGDA) algorithm. Under some mild approximations, learning converges to local points. Using structural similarity index measure (SSIM), we supply non-asymptotic analysis algorithm’s rate. Empirical evidences show rate, which inversely proportional number iterations, when tested tiny colour datasets Cats CelebA deep convolutional networks ResNet neural architectures. entropy regularization parameter $\varepsilon $ approximated SSIM tolerance $\epsilon . determine that iteration complexity return -stationary point be $\mathcal {O}\left ({\kappa \, \log (\epsilon ^{-1})}\right)$ , where $\kappa value depends divergence’s convexity step ratio SeqSGDA

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

On Convergence Rate of Concave-Convex Procedure

Concave-Convex Procedure (CCCP) has been widely used to solve nonconvex d.c.(difference of convex function) programs occur in learning problems, such as sparse support vector machine (SVM), transductive SVM, sparse principal componenent analysis (PCA), etc. Although the global convergence behavior of CCCP has been well studied, the convergence rate of CCCP is still an open problem. Most of d.c....

متن کامل

Non-Negative Matrix Factorization with Sinkhorn Distance

Non-negative Matrix Factorization (NMF) has received considerable attentions in various areas for its psychological and physiological interpretation of naturally occurring data whose representation may be parts-based in the human brain. Despite its good practical performance, one shortcoming of original NMF is that it ignores intrinsic structure of data set. On one hand, samples might be on a m...

متن کامل

Dual Iterative Hard Thresholding: From Non-convex Sparse Minimization to Non-smooth Concave Maximization

Iterative Hard Thresholding (IHT) is a class of projected gradient descent methods for optimizing sparsity-constrained minimization models, with the best known efficiency and scalability in practice. As far as we know, the existing IHT-style methods are designed for sparse minimization in primal form. It remains open to explore duality theory and algorithms in such a non-convex and NP-hard prob...

متن کامل

On the Convergence of the Concave-Convex Procedure

The concave-convex procedure (CCCP) is a majorization-minimization algorithm that solves d.c. (difference of convex functions) programs as a sequence of convex programs. In machine learning, CCCP is extensively used in many learning algorithms like sparse support vector machines (SVMs), transductive SVMs, sparse principal component analysis, etc. Though widely used in many applications, the con...

متن کامل

Parameter Convergence in systems with Convex/Concave Parameterization

2 Statement of the Problem A large class of problems in parameter estimation concerns systems where parameters occur nonlinearly. In [1]-[5], a stability framework for identification and control of such systems has been established. We address the issue of parameter convergence in such systems in this paper. Sufficient conditions under which parameter estimates converge to their true values are...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Access

سال: 2021

ISSN: ['2169-3536']

DOI: https://doi.org/10.1109/access.2021.3074943