GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium

Authors

  • Naresh Babu Bynagari Director of Sales, Career Soft Solutions Inc, 145 Talmadge rd Edison NJ 08817, Middlesex, USA

Keywords:

GANs Trained
Time-Scale Update
Local Nash Equilibrium

Abstract

When it comes to the formation of real-looking images using some complex models, Generative Adversarial Networks do not disappoint. The complex models involved are often the types with infeasible maximum likelihoods. Be that as it may, there is not yet any proof for the convergence of GANs training. This paper proposes a TTUR (a two-time scale update rule) for training the Generative Adversarial Networks with a descent of stochastic gradient based on haphazard loss functions. The two time-scale update rule has separate learning rates for the generator and the discriminator. With the aid of the stochastic approximation theory, this paper demonstrates that the TTUR reaches a point of convergence under the influence of mild assumption to a kind of remote and stationary state known as Nash equilibrium. This unification or meeting point principle also applies to the widespread Adam optimization. This is a form or replacement optimization algorithm designed into stochastic gradient descent and used for tutoring the deep learning models in the system. For the Adam optimization theory, this paper evinces that it is in line with the dynamics of a weighty ball in a frictional state. Thus, we prove that it favours flat minina in the objective perspective of things. To carry out an evaluation of how GANs perform during the image creation process, this paper presents what we have termed the 'Fréchet Inception Distance", also known as FID—a concept known to dwell on the resemblance between the images created and the real ones in a way that is more improved compared to the Inception Score. Experimentally, the TTUR helps in the bettering of DCGANs and Improved Wasserstein GANs (WGAN-GP). This makes it perform better than the traditional CelebA GAN training, LSUN Bedrooms, CIFAR-10, SVHN and the One Billion Word Benchmark.

Downloads

Download data is not yet available.

Metrics

Metrics Loading ...

References

Abbeel, P., and Mordatch, I. Learning with an opponent learning awareness. In Proceedings of the International Conference on Autonomous Agents and Multi-Agent Systems, 2018.

Bottou, L. and Bousquet, O. The tradeoffs of large scale learning. In Advances in Neural Information Processing Systems, pp. 161–168, 2008.

Bynagari, N. B. (2015). Machine Learning and Artificial Intelligence in Online Fake Transaction Alerting. Engineering International, 3(2), 115-126. https://doi.org/10.18034/ei.v3i2.566 DOI: https://doi.org/10.18034/ei.v3i2.566

Bynagari, N. B. (2016). Industrial Application of Internet of Things. Asia Pacific Journal of Energy and Environment, 3(2), 75-82. https://doi.org/10.18034/apjee.v3i2.576 DOI: https://doi.org/10.18034/apjee.v3i2.576

Bynagari, N. B. (2017). Prediction of Human Population Responses to Toxic Compounds by a Collaborative Competition. Asian Journal of Humanity, Art and Literature, 4(2), 147-156. https://doi.org/10.18034/ajhal.v4i2.577 DOI: https://doi.org/10.18034/ajhal.v4i2.577

Bynagari, N. B. (2018). On the ChEMBL Platform, a Large-scale Evaluation of Machine Learning Algorithms for Drug Target Prediction. Asian Journal of Applied Science and Engineering, 7, 53–64. Retrieved from https://upright.pub/index.php/ajase/article/view/31

Chambolle, A. and Pock, T. A first-order primal-dual algorithm for convex problems with applications to imaging. Journal of Mathematical Imaging and Vision, 40(1):120–145, 2011. DOI: https://doi.org/10.1007/s10851-010-0251-1

Foerster, J., Chen, R. Y., Al-Shedivat, M., Whiteson, S., Abbeel, P., and Mordatch, I. Learning with an opponent learning awareness. In Proceedings of the International Conference on Autonomous Agents and Multi-Agent Systems, 2018.

Ganapathy, A. (2018). Cascading Cache Layer in Content Management System. Asian Business Review, 8(3), 177-182. https://doi.org/10.18034/abr.v8i3.542 DOI: https://doi.org/10.18034/abr.v8i3.542

Ganapathy, A. (2018). UI/UX Automated Designs in the World of Content Management Systems. Asian Journal of Applied Science and Engineering, 7(1), 43-52.

Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., and Courville, A. C. Improved training of Wasserstein GANs. In Advances in Neural Information Processing Systems, pp. 5767–5777, 2017.

Harker, P. T. and Pang, J.-S. Finite-dimensional variational inequality and nonlinear complementarity problems: a survey of theory, algorithms and applications. Mathematical Programming, 48(1-3):161–220, 1990. DOI: https://doi.org/10.1007/BF01582255

Hirsch, M. W. (1989). Convergent activation dynamics in continuous time networks. Neural Networks, 2(5):331–349. DOI: https://doi.org/10.1016/0893-6080(89)90018-X

Juditsky, A., Nemirovski, A., and Tauvel, C. Solving variational inequalities with stochastic mirror-prox algorithm. Stochastic Systems, 1(1):17–58, 2011. DOI: https://doi.org/10.1287/10-SSY011

Kim, S.-J., Magnani, A., and Boyd, S. Robust Fisher discriminant analysis. In Advances in Neural Information Processing Systems, 2006.

Korpelevich, G. The extragradient method for finding saddle points and other problems. Matecon, 12:747–756, 1976.

Krizhevsky, A. and Hinton, G. (2009). Learning multiple layers of features from tiny images. Technical report, Citeseer.

Magnanti, T. L. and Perakis, G. Averaging schemes for variational inequalities and systems of equations. Mathematics of Operations Research, 22(3):568–587, 1997. DOI: https://doi.org/10.1287/moor.22.3.568

Nash, J. Non-cooperative games. Annals of Mathematics, pp. 286–295, 1951. DOI: https://doi.org/10.2307/1969529

Nedic, A. and Ozdaglar, A. Subgradient methods for saddle-point problems. Journal of Optimization Theory and Applications, 142(1):205–228, 2009. DOI: https://doi.org/10.1007/s10957-009-9522-7

Paruchuri, H. (2018). AI Health Check Monitoring and Managing Content Up and Data in CMS World. Malaysian Journal of Medical and Biological Research, 5(2), 141-146. https://doi.org/10.18034/mjmbr.v5i2.554 DOI: https://doi.org/10.18034/mjmbr.v5i2.554

Paruchuri, H., & Asadullah, A. (2018). The Effect of Emotional Intelligence on the Diversity Climate and Innovation Capabilities. Asia Pacific Journal of Energy and Environment, 5(2), 91-96. https://doi.org/10.18034/apjee.v5i2.561 DOI: https://doi.org/10.18034/apjee.v5i2.561

Polyak, B. T. and Juditsky, A. B. (1992) Acceleration of stochastic approximation by averaging. SIAM J. Control and Optimization, 30(4):838–855, doi: 10.1137/0330046. DOI: https://doi.org/10.1137/0330046

Rosen, J. B. Existence and uniqueness of equilibrium points for concave n-person games. Econometrica, 33(3):520–534, 1965. DOI: https://doi.org/10.2307/1911749

Vadlamudi, S. (2016). What Impact does Internet of Things have on Project Management in Project based Firms?. Asian Business Review, 6(3), 179-186. https://doi.org/10.18034/abr.v6i3.520 DOI: https://doi.org/10.18034/abr.v6i3.520

Vadlamudi, S. (2018). Agri-Food System and Artificial Intelligence: Reconsidering Imperishability. Asian Journal of Applied Science and Engineering, 7(1), 33-42.

Vaishnavh Nagarajan and J. Zico Kolter. Gradient descent and optimization is locally stable. In Advances in Neural Information Processing Systems, pp. 5585–5595, 2017.

Zhiming Zhou, Han Cai, Shu Rong, Yuxuan Song, Kan Ren, Weinan Zhang, Jun Wang, and Yong Yu. Activation maximizing generative adversarial nets. In Proc. International Conference on Learning Representations (ICLR), 2018.

Downloads

Published

2019-04-25

How to Cite

Bynagari, N. B. (2019). GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. Asian Journal of Applied Science and Engineering, 8(1), 25–34. https://doi.org/10.18034/ajase.v8i1.9

Issue

Section

Articles