Soft rank neighbor embeddings
نویسندگان
چکیده
Correlation-based multidimensional scaling is proposed for reconstructing pairwise dissimilarity or score relationships in a Euclidean space. Pearson correlation between pairs of objects in source and target space can be directly maximized by gradient methods, while gradient optimization of Spearman rank correlation profits from a numerically soft formulation introduced in this work. Scale and shift invariance properties of correlation help circumventing typical distance concentration problems.
منابع مشابه
Fast Nearest Neighbor Preserving Embeddings
We show an analog to the Fast Johnson-Lindenstrauss Transform for Nearest Neighbor Preserving Embeddings in `2. These are sparse, randomized embeddings that preserve the (approximate) nearest neighbors. The dimensionality of the embedding space is bounded not by the size of the embedded set n, but by its doubling dimension λ. For most large real-world datasets this will mean a considerably lowe...
متن کاملSecond-Order Word Embeddings from Nearest Neighbor Topological Features
We introduce second-order vector representations of words, induced from nearest neighborhood topological features in pre-trained contextual word embeddings. We then analyze the effects of using second-order embeddings as input features in two deep natural language processing models, for named entity recognition and recognizing textual entailment, as well as a linear model for paraphrase recogni...
متن کاملSpherical Nearest Neighbor Classification: Application to Hyperspectral Data
The problem of feature transformation arises in many fields of information processing, including machine learning, data compression, computer vision and geosciences applications. Here we discuss an approach that seeks a hyperspherical coordinate system preserving geodesic distances in the high dimensional hyperspectral data space. A lower dimensional hyperspherical manifold is computed using a ...
متن کاملBuilding Earth Mover's Distance on Bilingual Word Embeddings for Machine Translation
Following their monolingual counterparts, bilingual word embeddings are also on the rise. As a major application task, word translation has been relying on the nearest neighbor to connect embeddings cross-lingually. However, the nearest neighbor strategy suffers from its inherently local nature and fails to cope with variations in realistic bilingual word embeddings. Furthermore, it lacks a mec...
متن کاملLearning Label Embeddings for Nearest-Neighbor Multi-class Classification with an Application to Speech Recognition
We consider the problem of using nearest neighbor methods to provide a conditional probability estimate, P (y|a), when the number of labels y is large and the labels share some underlying structure. We propose a method for learning label embeddings (similar to error-correcting output codes (ECOCs)) to model the similarity between labels within a nearest neighbor framework. The learned ECOCs and...
متن کامل