1 / 11

Phrase-based Statistical Language Generation using Graphical Models & Active Learning

Phrase-based Statistical Language Generation using Graphical Models & Active Learning. Two directions of research in statistical NLG : Statistics is introduced in the generation process Statistics is introduced at the generation decision lev el

udell
Download Presentation

Phrase-based Statistical Language Generation using Graphical Models & Active Learning

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Phrase-based Statistical Language Generation using Graphical Models & Active Learning

  2. Two directions of research in statistical NLG : • Statistics is introduced in the generation process • Statistics is introduced at the generation decision level • Traditional methods rely on handcrafted generator to define the generation decision space • BAGEL : an NLG system that can be trained from aligned data • BAGEL shifts the effort from model design and implementation to data annotation • BAGEL uses a stack based semantic representation to reduce the number of semantic concepts to be searched

  3. Phrase-based generation from semantic stacks • First maps the input dialogue to a set of stacks of semantic concepts and then aligns it with a word sequence • Concepts used in stack: • Dialogue act type, i.e., either inform(an utterance providing information about the object under discussion) or reject(request of the user cannot be met) • Attributes of the object • Values for the attributes • Special symbols for denoting irrelevance

  4. Generator goal : find the most likely combination given an unordered set of mandatory semantic stacks from an input dialogue • BAGEL models contiguous words belonging to the same semantic stack as a phrase

  5. Dynamic Bayesian networks for NLG • Provides a principled framework for predicting elements in a large structured space • Suitable for modelling linguistic variation • BAGEL must derive the optimal sequence of semantic stacks that will appear in the utterance given a set of mandatory stacks • Mandatory stacks that are visited in a current sequence are marked • Any sequence not including all mandatory input stacks in the final frame are discarded

  6. Certainty-based active learning • Domain-specific data not readily available • Active learning is used to optimize data in which the next semantic input to annotate is determined by the current model • BAGEL’s active learning training process: • Generate an utterance for each semantic input using the current model • Annotate the k semantic inputs { } yielding the lowest realization probability • Retrain the model with the additional k data points

  7. Score evaluation • Results are averaged over a 10-fold cross-validation using unseen data • Performance improves with • Adding dependency on future stacks • Backing off to partial stacks(with sparse data only)

  8. Score evaluation • Increasing the training set one utterance at a time using active learning significantly outperforms random sampling (k=1) • Increasing the number of utterances to be queried to 10 at a time results to smaller performance increase

  9. Subjective evaluation • Subjects are asked to evaluate the informativeness and naturalness of the utterances • Utterances are the same as those used for cross validation

  10. Score evaluation

  11. Discussions • BAGEL uses domain specific data in the generation process without any syntactic annotation • No manual work involved beyond the semantic annotation • Future work: • Use of committee based active learning • Can syntactic information improve performance in more complex domains?

More Related