Learning Compact Features via In-Training Representation Alignment

نویسندگان

چکیده

Deep neural networks (DNNs) for supervised learning can be viewed as a pipeline of the feature extractor (i.e., last hidden layer) and linear classifier output that are trained jointly with stochastic gradient descent (SGD) on loss function (e.g., cross-entropy). In each epoch, true is estimated using mini-batch sampled from training set model parameters then updated gradients. Although latter provides an unbiased estimation former, they subject to substantial variances derived size number mini-batches, leading noisy jumpy updates. To stabilize such undesirable variance in estimating gradients, we propose In-Training Representation Alignment (ITRA) explicitly aligns distributions two different mini-batches matching SGD process. We also provide rigorous analysis desirable effects representation learning: (1) extracting compact representation; (2) reducing over-adaption via adaptively weighting mechanism; (3) accommodating multi-modalities. Finally, conduct large-scale experiments both image text classifications demonstrate its superior performance strong baselines.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Learning Via Compact Data Representation

We present an unsupervised learning methodology derived from compact data encoding and demonstrate how to construct models of polysemy, priming, semantic disambiguation and learning using this theoretical basis. The model is capable of simulating human-like performance on artificial grammar learning.

متن کامل

Image alignment via kernelized feature learning

Machine learning is an application of artificial intelligence that is able to automatically learn and improve from experience without being explicitly programmed. The primary assumption for most of the machine learning algorithms is that the training set (source domain) and the test set (target domain) follow from the same probability distribution. However, in most of the real-world application...

متن کامل

Image Classification via Sparse Representation and Subspace Alignment

Image representation is a crucial problem in image processing where there exist many low-level representations of image, i.e., SIFT, HOG and so on. But there is a missing link across low-level and high-level semantic representations. In fact, traditional machine learning approaches, e.g., non-negative matrix factorization, sparse representation and principle component analysis are employed to d...

متن کامل

A Novel Graph-based Compact Representation of Word Alignment

In this paper, we propose a novel compact representation called weighted bipartite hypergraph to exploit the fertility model, which plays a critical role in word alignment. However, estimating the probabilities of rules extracted from hypergraphs is an NP-complete problem, which is computationally infeasible. Therefore, we propose a divide-and-conquer strategy by decomposing a hypergraph into a...

متن کامل

Learning Features of Representation in Conceptual Context

When people categorize an object, they often encode a certain number of its properties for later classification. In Schyns and Murphy (1993), we suggested that the way people group objects into categories could induce the learning of new dimensions of ca egorization--i.e., dimensions that did not exist prior to the experience with the categorization system. In this research, we examine whether ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2023

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v37i7.26044