Fast convergence rates of deep neural networks for classification
نویسندگان
چکیده
Abstract We derive the fast convergence rates of a deep neural network (DNN) classifier with rectified linear unit (ReLU) activation function learned using hinge loss. consider three cases for true model: (1) smooth decision boundary, (2) conditional class probability, and (3) margin condition (i.e., probability inputs near boundary is small). show that DNN loss achieves rate convergences all provided architecture number layers, nodes sparsity) carefully selected. An important implication architectures are very flexible use in various without much modification. In addition, we by minimizing cross-entropy, under conditions noise exponent large. Even though they strong, explain these two not too absurd image classification problems. To confirm our theoretical explanation, present results small numerical study conducted to compare cross-entropy.
منابع مشابه
Fast algorithms for learning deep neural networks
With the increase in computation power and data availability in recent times, machine learning and statistics have seen an enormous development and widespread application in areas such as computer vision, computational biology and others. A focus of current research are deep neural nets: nested functions consisting of a hierarchy of layers of thousands of weights and nonlinear, hidden units. Th...
متن کاملFast Metric Learning For Deep Neural Networks
Similarity metrics are a core component of many information retrieval and machine learning systems. In this work we propose a method capable of learning a similarity metric from data equipped with a binary relation. By considering only the similarity constraints, and initially ignoring the features, we are able to learn target vectors for each instance using one of several appropriately designe...
متن کاملSelective Classification for Deep Neural Networks
Selective classification techniques (also known as reject option) have not yet been considered in the context of deep neural networks (DNNs). These techniques can potentially significantly improve DNNs prediction performance by trading-off coverage. In this paper we propose a method to construct a selective classifier given a trained neural network. Our method allows a user to set a desired ris...
متن کاملDeep Recurrent Neural Networks for Supernovae Classification
We apply deep recurrent neural networks, which are capable of learning complex sequential information, to classify supernovae. The observational time and filter fluxes are used as inputs to the network, but since the inputs are agnostic additional data such as host galaxy information can also be included. Using the Supernovae Photometric Classification Challenge (SPCC) data, we find that deep n...
متن کاملClassification regions of deep neural networks
The goal of this paper is to analyze the geometric properties of deep neural network classifiers in the input space. We specifically study the topology of classification regions created by deep networks, as well as their associated decision boundary. Through a systematic empirical investigation, we show that state-of-the-art deep nets learn connected classification regions, and that the decisio...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Neural Networks
سال: 2021
ISSN: ['1879-2782', '0893-6080']
DOI: https://doi.org/10.1016/j.neunet.2021.02.012