Language as a latent sequence: Deep latent variable models for semi-supervised paraphrase generation

نویسندگان

چکیده

This paper explores deep latent variable models for semi-supervised paraphrase generation, where the missing target pair unlabelled data is modelled as a sequence. We present novel unsupervised model named variational sequence auto-encoding reconstruction (VSAR), which performs inference given an observed text. To leverage information from text pairs, we additionally introduce supervised call dual directional learning (DDL), designed to integrate with our proposed VSAR model. Combining DDL (DDL+VSAR) enables us conduct learning. Still, combined suffers cold-start problem. further combat this issue, propose improved weight initialisation solution, leading two-stage training scheme knowledge-reinforced-learning (KRL). Our empirical evaluations suggest that yields competitive performance against state-of-the-art baselines on complete data. Furthermore, in scenarios only fraction of labelled pairs are available, consistently outperforms strong baseline (DDL) by significant margin (p<.05; Wilcoxon test). code publicly available at https://github.com/jialin-yu/latent-sequence-paraphrase.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Semi-Supervised and Latent-Variable Models of Natural Language Semantics

This thesis focuses on robust analysis of natural language semantics. A primary bottleneck for semantic processing of text lies in the scarcity of high-quality and large amounts of annotated data that provide complete information about the semantic structure of natural language expressions. In this dissertation, we study statistical models tailored to solve problems in computational semantics, ...

متن کامل

Semi-supervised latent variable models for sentence-level sentiment analysis

We derive two variants of a semi-supervised model for fine-grained sentiment analysis. Both models leverage abundant natural supervision in the form of review ratings, as well as a small amount of manually crafted sentence labels, to learn sentence-level sentiment classifiers. The proposed model is a fusion of a fully supervised structured conditional model and its partially supervised counterp...

متن کامل

Supervised Spectral Latent Variable Models

We present a probabilistic structured prediction method for learning input-output dependencies where correlations between outputs are modeled as low-dimensional manifolds constrained by both geometric, distance preserving output relations, and predictive power of inputs. Technically this reduces to learning a probabilistic, input conditional model, over latent (manifold) and output variables us...

متن کامل

Paraphrase Generation from Latent-Variable PCFGs for Semantic Parsing

One of the limitations of semantic parsing approaches to open-domain question answering is the lexicosyntactic gap between natural language questions and knowledge base entries – there are many ways to ask a question, all with the same answer. In this paper we propose to bridge this gap by generating paraphrases of the input question with the goal that at least one of them will be correctly map...

متن کامل

Latent Variable Models for Hippocampal Sequence Analysis

VIRTUAL TUNING CURVES we only train the HMMs on spikes from PBEs; to determine if the inferred states encode position data, we compute virtual tuning curves in two ways: (A) by decoding RUN data using the PBE-only HMM, and then using the true position data to estimate a map from states to position, and (B) by using the Bayesian decoder to estimate position during PBEs, and to learn a map from t...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: AI open

سال: 2023

ISSN: ['2666-6510']

DOI: https://doi.org/10.1016/j.aiopen.2023.05.001