Interpolation of stochastic grammar and word bigram models in natural language understanding

نویسندگان

  • Sven C. Martin
  • Andreas Kellner
  • Thomas Portele
چکیده

The paper shows the effects of combining a stochastic grammar with a word bigram language model by log-linear interpolation. It is divided into three main parts: The first part derives the stochastic grammar model and gives a sound theoretical motivation to incorporate word dependencies such as bigrams. The second part describes two different algorithmic approaches to the combination of both models by log-linear interpolation. The third part reports attribute error rate (AER) results measured on the Philips corpus of train time table inquiries that show a reduction of up to 9% relative. 1. STOCHASTIC MODEL OF NATURAL LANGUAGE UNDERSTANDING The Philips Natural Language Understanding (NLU) module is used in automated inquiry systems (AIS), such as train table enquiries [2], to analyze the word sequence of a user utterance. It does not try to find parse trees that cover the whole word sequence but breaks up the sequence into chunks, where each chunk belongs to a semantically meaningful concept. A stochastic context–free grammar is used to derive the word chunk from a concept. The chunking is useful since the spontaneous speech that occurs in dialogue applications is very ungrammatical. Thus, a robust NLU model concentrates on the useful parts of a user utterance. Other recent works also employ some kind of chunking, e.g. [6, 9]. The stochastic model of the Philips NLU module was developed by H. Aust in [1, p. 81]. Here, we show that this model can be derived from Bayes’ decision rule. This derivation gives a sound theoretical motivation to incorporate word dependencies such as bigrams. Bayes’ decision rule finds the most likely concept sequence K̂ = k̂1, . . . , k̂s, given the sequence O = o1, . . . , ot of acoustic observations. The derivation of the concept sequence K does not directly depend on the acoustic observations O but on a word sequence W = w1, . . . , wN derived from O as an intermediate result: K̂ = argmax K p(K|O)

منابع مشابه

Permugram language models

In natural languages, the words within an utterance are often correlated over large distances. Long-spanning con-textual eeects of this type cannot be eeciently and robustly captured by the traditional N-gram approaches of stochastic language modelling. We present a new kind of stochastic grammar | the permugram model. A permu-gram model is obtained by linear interpolation of a large number of ...

متن کامل

Experience with a Stack Decoder-Based HMM CSR and Back-Off N-Gram Language Models

Stochastic language models are more useful than nonstochastic models because they contribute more information than a simple acceptance or rejection of a word sequence. Back-off N-gram language models [ I l l are an effective class of word based stochastic language model. The first part of this paper describes our experiences using the back-off language models in our time-synchronous decoder CSR...

متن کامل

Using a stochastic context-free grammar as a language model for speech recognition

This paper describes a number of experiments in adding new grammatical knowledge to the Berkeley Restaurant Project (BeRP), our medium-vocabulary (1300 word), speaker-independent, spontaneous continuous-speech understanding system (Jurafsky et al. 1994). We describe an algorithm for using a probabilistic Earley parser and a stochastic context-free grammar (SCFG) to generate word transition prob...

متن کامل

Parsing N-Best Lists of Handwritten Sentences

This paper investigates the application of a probabilistic parser for natural language on the list of the Nbest sentences produced by an off-line recognition system for cursive handwritten sentences. For the generation of the N-best sentence list an HMM-based recognizer including a bigram language model is used. The parsing of the sentences is achieved by a bottom-up chart parser for stochastic...

متن کامل

A word graph interface for a flexible concept based speech understanding framework

In this paper, we introduce a word graph interface between speech and natural language processing systems within a flexible speech understanding framework based on stochastic concept modeling augmented with background ”filler” models. Each concept represents a set of phrases ( written as a context free grammar (CFG)) with the same meaning, and is compiled into a stochastic recursive transition ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

متن کامل
عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2000