Learning-Rate Annealing Methods for Deep Neural Networks
نویسندگان
چکیده
Deep neural networks (DNNs) have achieved great success in the last decades. DNN is optimized using stochastic gradient descent (SGD) with learning rate annealing that overtakes adaptive methods many tasks. However, there no common choice regarding scheduled-annealing for SGD. This paper aims to present empirical analysis of based on experimental results major data-sets image classification one key applications DNNs. Our experiment involves recent deep network models combination a variety methods. We also propose an combining sigmoid function warmup shown overtake both and other existing schedules accuracy most cases
منابع مشابه
Distributed Newton Methods for Deep Neural Networks
Deep learning involves a difficult non-convex optimization problem with a large number of weights between any two adjacent layers of a deep structure. To handle large data sets or complicated networks, distributed training is needed, but the calculation of function, gradient, and Hessian is expensive. In particular, the communication and the synchronization cost may become a bottleneck. In this...
متن کاملFast algorithms for learning deep neural networks
With the increase in computation power and data availability in recent times, machine learning and statistics have seen an enormous development and widespread application in areas such as computer vision, computational biology and others. A focus of current research are deep neural nets: nested functions consisting of a hierarchy of layers of thousands of weights and nonlinear, hidden units. Th...
متن کاملBayesian Incremental Learning for Deep Neural Networks
In industrial machine learning pipelines, data often arrive in parts. Particularly in the case of deep neural networks, it may be too expensive to train the model from scratch each time, so one would rather use a previously learned model and the new data to improve performance. However, deep neural networks are prone to getting stuck in a suboptimal solution when trained on only new data as com...
متن کاملRepresentational Distance Learning for Deep Neural Networks
Deep neural networks (DNNs) provide useful models of visual representational transformations. We present a method that enables a DNN (student) to learn from the internal representational spaces of a reference model (teacher), which could be another DNN or, in the future, a biological brain. Representational spaces of the student and the teacher are characterized by representational distance mat...
متن کاملNeural networks: different problems require different learning rate adaptive methods
In a previous study, a new adaptive method (AM) was developed to adjust the learning rate in artificial neural networks: the generalized no-decrease adaptive method (GNDAM). The GNDAM is fundamentally different from other traditional AMs. Instead of using the derivative sign of a given weight to adjust its learning rate, this AM is based on a trial and error heuristic where global learning rate...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Electronics
سال: 2021
ISSN: ['2079-9292']
DOI: https://doi.org/10.3390/electronics10162029