350 likes | 367 Views
Explore cutting-edge models such as ESIM, KIM, DMAN, MT-DNN, and BERT+SRL for improved natural language understanding. Learn about global knowledge integration, external knowledge enrichment, discourse markers, and multi-task learning for enhanced NLU performance.
E N D
An Introduction toNatural Language Inference Jianhao Shen 1801111357 2019/05/29
Outline • Introduction • Enhanced sequential inference model(ESIM) • Neural Natural Language Inference Models Enhanced withExternal Knowledge(KIM) • Discourse Marker Augmented Network with Reinforcement Learning for Natural Language Inference(DMAN) • Multi-Task Deep Neural Networks for Natural Language Understanding(MT-DNN) • I Know What You Want: Semantic Learning for Text Comprehension(BERT+SRL)
Example (Premise)
Natural Language Inference(NLI) • Does premise P entails hypothesis H, they are contradict to each other, or they have no relation? • Intuitive notion of inference • One step inference, not long chains of deduction • NLI is a necessary condition for real NLU • Reasoning/inference • Global knowledge/common sense
Approach • Convert sentences into logic forms • Hard(Semantic parsing is still very difficult) • Data-driven • Large dataset, e.g. SNLI(2015) and MultiNLI(2017)
Leaderboard ESIM 88.0 ELMo ELMo GPT BERT
Outline • Introduction • Enhanced sequential inference model(ESIM) • Neural Natural Language Inference Models Enhanced withExternal Knowledge(KIM) • Discourse Marker Augmented Network with Reinforcement Learning for Natural Language Inference(DMAN) • Multi-Task Deep Neural Networks for Natural Language Understanding(MT-DNN) • I Know What You Want: Semantic Learning for Text Comprehension(BERT+SRL)
ESIM • Key idea • Co-attention • Local inference • Inference composition Chen et al., Enhanced LSTM for Natural Language Inference ACL17
Co-attention • Soft alignment Two dogs are running through a field. There are animals outdoors.
Local inference • Represent each word using another sentence • Enhancement of local inference information • Element-wise product for similarity • Element-wise difference for “contradiction”
Inference composition • BiLSTM with enhanced local information • Compute both max and average pooling • MLP for classification
Ablation analysis 6 hours 40 hours
Outline • Introduction • Enhanced sequential inference model(ESIM) • Neural Natural Language Inference Models Enhanced withExternal Knowledge(KIM) • Discourse Marker Augmented Network with Reinforcement Learning for Natural Language Inference(DMAN) • Multi-Task Deep Neural Networks for Natural Language Understanding(MT-DNN) • I Know What You Want: Semantic Learning for Text Comprehension(BERT+SRL)
KIM • Global knowledge/common sense help inference • P: A lady standing in a wheat field. • H: The person standing in a corn field. Chen et al., Neural Natural Language Inference Models Enhanced with External Knowledge ACL18
External Knowledge • Only consider lexical-level semantic knowledge: • represent relation between words wi and wj as rij • Use relations in WordNet: • Synonymy(同义词) • Antonymy(反义词) • Hypernymy(上位词) • Hyponymy(下位词) • Co-hyponyms(拥有同上位词但不同义)e.g. [dog, wolf]=1 • TransE: No improvement
Knowledge-Enriched Co-Attention External knowledge: Word pairs with semantic relationship may be aligned together
Result WordNet Baseline: 85.8
Outline • Introduction • Enhanced sequential inference model(ESIM) • Neural Natural Language Inference Models Enhanced withExternal Knowledge(KIM) • Discourse Marker Augmented Network with Reinforcement Learning for Natural Language Inference(DMAN) • Multi-Task Deep Neural Networks for Natural Language Understanding(MT-DNN) • I Know What You Want: Semantic Learning for Text Comprehension(BERT+SRL)
DM Prediction • Discourse Marker connects two sentences and expresses relationship between them • but, although • because, so • if, when, still, before • …… • No need to label & large data! • BookCorpus • (S1, S2, m) • 6,527,128 pairs for 8 DM
DMAN rp rh Pan et al., Discourse Marker Augmented Network with Reinforcement Learning for Natural Language Inference ACL18
Outline • Introduction • Enhanced sequential inference model(ESIM) • Neural Natural Language Inference Models Enhanced withExternal Knowledge(KIM) • Discourse Marker Augmented Network with Reinforcement Learning for Natural Language Inference(DMAN) • Multi-Task Deep Neural Networks for Natural Language Understanding(MT-DNN) • I Know What You Want: Semantic Learning for Text Comprehension(BERT+SRL)
MT-DNN • Multi-task learning(MTL) • Not enough data in one task • Regularization • Language Model Pretraining • Universal language representation • Large (unsupervised) data Liu X, He P, Chen W, et al. Multi-Task Deep Neural Networks for Natural Language Understanding[J]. arXiv preprint arXiv:1901.11504, 2019.
Outline • Introduction • Enhanced sequential inference model(ESIM) • Neural Natural Language Inference Models Enhanced withExternal Knowledge(KIM) • Discourse Marker Augmented Network with Reinforcement Learning for Natural Language Inference(DMAN) • Multi-Task Deep Neural Networks for Natural Language Understanding(MT-DNN) • I Know What You Want: Semantic Learning for Text Comprehension(BERT+SRL)
Semantic Role Labeling • SRL • Who did what to whom, when where and why • Shallow semantic • NLI/MRC • Text comprehension and inference • Deep semantic
SRL model • spaCy: POS tags • BiLSTM: BIO encoding
Summary • ESIM • Co-attention • Local inference • Inference composition • Global Knowledge • WordNet • Discourse Marker • No need to label & large data • Multi-task + LM Pretraining • Semantic role labeling