Language as a latent sequence: Deep latent variable models for semi-supervised paraphrase generation
نویسندگان
چکیده
This paper explores deep latent variable models for semi-supervised paraphrase generation, where the missing target pair unlabelled data is modelled as a sequence. We present novel unsupervised model named variational sequence auto-encoding reconstruction (VSAR), which performs inference given an observed text. To leverage information from text pairs, we additionally introduce supervised call dual directional learning (DDL), designed to integrate with our proposed VSAR model. Combining DDL (DDL+VSAR) enables us conduct learning. Still, combined suffers cold-start problem. further combat this issue, propose improved weight initialisation solution, leading two-stage training scheme knowledge-reinforced-learning (KRL). Our empirical evaluations suggest that yields competitive performance against state-of-the-art baselines on complete data. Furthermore, in scenarios only fraction of labelled pairs are available, consistently outperforms strong baseline (DDL) by significant margin (p<.05; Wilcoxon test). code publicly available at https://github.com/jialin-yu/latent-sequence-paraphrase.
منابع مشابه
Semi-Supervised and Latent-Variable Models of Natural Language Semantics
This thesis focuses on robust analysis of natural language semantics. A primary bottleneck for semantic processing of text lies in the scarcity of high-quality and large amounts of annotated data that provide complete information about the semantic structure of natural language expressions. In this dissertation, we study statistical models tailored to solve problems in computational semantics, ...
متن کاملSemi-supervised latent variable models for sentence-level sentiment analysis
We derive two variants of a semi-supervised model for fine-grained sentiment analysis. Both models leverage abundant natural supervision in the form of review ratings, as well as a small amount of manually crafted sentence labels, to learn sentence-level sentiment classifiers. The proposed model is a fusion of a fully supervised structured conditional model and its partially supervised counterp...
متن کاملSupervised Spectral Latent Variable Models
We present a probabilistic structured prediction method for learning input-output dependencies where correlations between outputs are modeled as low-dimensional manifolds constrained by both geometric, distance preserving output relations, and predictive power of inputs. Technically this reduces to learning a probabilistic, input conditional model, over latent (manifold) and output variables us...
متن کاملParaphrase Generation from Latent-Variable PCFGs for Semantic Parsing
One of the limitations of semantic parsing approaches to open-domain question answering is the lexicosyntactic gap between natural language questions and knowledge base entries – there are many ways to ask a question, all with the same answer. In this paper we propose to bridge this gap by generating paraphrases of the input question with the goal that at least one of them will be correctly map...
متن کاملLatent Variable Models for Hippocampal Sequence Analysis
VIRTUAL TUNING CURVES we only train the HMMs on spikes from PBEs; to determine if the inferred states encode position data, we compute virtual tuning curves in two ways: (A) by decoding RUN data using the PBE-only HMM, and then using the true position data to estimate a map from states to position, and (B) by using the Bayesian decoder to estimate position during PBEs, and to learn a map from t...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: AI open
سال: 2023
ISSN: ['2666-6510']
DOI: https://doi.org/10.1016/j.aiopen.2023.05.001