Lexical semantics enhanced neural word embeddings
نویسندگان
چکیده
Current breakthroughs in natural language processing have benefited dramatically from neural models, through which distributional semantics can leverage data representations to facilitate downstream applications. Since embeddings use context prediction on word co-occurrences yield dense vectors, they are inevitably prone capture more semantic association than similarity. To improve vector space models deriving similarity, we post-process deep metric learning, inject lexical-semantic relations, including syn/antonymy and hypo/hypernymy, into a space. We introduce hierarchy-fitting, novel specialization approach modelling similarity nuances inherently stored the IS-A hierarchies. Hierarchy-fitting attains state-of-the-art results common- rare-word benchmark datasets for embeddings. It also incorporates an asymmetric distance function specialize hypernymy's directionality explicitly, it significantly improves vanilla multiple evaluation tasks of detecting hypernymy without negative impacts judgement. The demonstrate efficacy hierarchy-fitting specializing with relations late fusion, potentially expanding its applicability aggregating heterogeneous various knowledge resources learning multimodal spaces.
منابع مشابه
Lexical Coherence Graph Modeling Using Word Embeddings
Coherence is established by semantic connections between sentences of a text which can be modeled by lexical relations. In this paper, we introduce the lexical coherence graph (LCG), a new graph-based model to represent lexical relations among sentences. The frequency of subgraphs (coherence patterns) of this graph captures the connectivity style of sentence nodes in this graph. The coherence o...
متن کاملDict2vec : Learning Word Embeddings using Lexical Dictionaries
Learning word embeddings on large unlabeled corpus has been shown to be successful in improving many natural language tasks. The most efficient and popular approaches learn or retrofit such representations using additional external data. Resulting embeddings are generally better than their corpus-only counterparts, although such resources cover a fraction of words in the vocabulary. In this pap...
متن کاملTemporal Word Analogies: Identifying Lexical Replacement with Diachronic Word Embeddings
This paper introduces the concept of temporal word analogies: pairs of words which occupy the same semantic space at different points in time. One well-known property of word embeddings is that they are able to effectively model traditional word analogies (“word w1 is to word w2 as word w3 is to word w4”) through vector addition. Here, I show that temporal word analogies (“wordw1 at time tα is ...
متن کاملRight-truncatable Neural Word Embeddings
This paper proposes an incremental learning strategy for neural word embedding methods, such as SkipGrams and Global Vectors. Since our method iteratively generates embedding vectors one dimension at a time, obtained vectors equip a unique property. Namely, any right-truncated vector matches the solution of the corresponding lower-dimensional embedding. Therefore, a single embedding vector can ...
متن کاملClinical Abbreviation Disambiguation Using Neural Word Embeddings
This study examined the use of neural word embeddings for clinical abbreviation disambiguation, a special case of word sense disambiguation (WSD). We investigated three different methods for deriving word embeddings from a large unlabeled clinical corpus: one existing method called Surrounding based embedding feature (SBE), and two newly developed methods: Left-Right surrounding based embedding...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Knowledge Based Systems
سال: 2022
ISSN: ['1872-7409', '0950-7051']
DOI: https://doi.org/10.1016/j.knosys.2022.109298