1 / 75

Constrained Conditional Models Towards Better Semantic Analysis of Text

Constrained Conditional Models Towards Better Semantic Analysis of Text. Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign. With thanks to: Collaborators: Ming-Wei Chang , Gourab Kundu, Lev Ratinov, Rajhans Samdani,

zizi
Download Presentation

Constrained Conditional Models Towards Better Semantic Analysis of Text

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Constrained Conditional Models Towards Better Semantic Analysis of Text Dan Roth Department of Computer Science University of Illinois at Urbana-Champaign With thanks to: Collaborators:Ming-Wei Chang, Gourab Kundu, Lev Ratinov, Rajhans Samdani, Vivek Srikumar, Many others Funding: NSF; DHS; NIH; DARPA. DASH Optimization (Xpress-MP) May 2013 KU Leuven, Belguim

  2. Nice to Meet You

  3. Learning and Inference in NLP • Natural Language Decisions are Structured • Global decisions in which several local decisions play a role but there are mutual dependencies on their outcome. • It is essential to make coherent decisions in a way that takes the interdependencies into account. Joint, Global Inference. • TODAY: • How to support real, high level, natural language decisions • How to learn models that are used, eventually, to make global decisions • A framework that allows one to exploit interdependencies among decision variables both in inference (decision making) and in learning. • Inference: A formulation for incorporating expressive declarative knowledge in decision making. • Learning: Ability to learn simple models; amplify its power by exploiting interdependencies.

  4. Comprehension (ENGLAND, June, 1989) - Christopher Robin is alive and well. He lives in England. He is the same person that you read about in the book, Winnie the Pooh. As a boy, Chris lived in a pretty home called Cotchfield Farm. When Chris was three years old, his father wrote a poem about him. The poem was printed in a magazine for others to read. Mr. Robin then wrote a book. He made up a fairy tale land where Chris lived. His friends were animals. There was a bear called Winnie the Pooh. There was also an owl and a young pig, called a piglet. All the animals were stuffed toys that Chris owned. Mr. Robin made them come to life with his words. The places in the story were all near Cotchfield Farm. Winnie the Pooh was written in 1925. Children still love to read about Christopher Robin and his animal friends. Most people don't know he is a real person who is grown now. He has written two books of his own. They tell what it is like to be famous. 1. Christopher Robin was born in England. 2. Winnie the Pooh is a title of a book. 3. Christopher Robin’s dad was a magician. 4. Christopher Robin must be at least 65 now. This is an Inference Problem

  5. Learning and Inference • Global decisions in which several local decisions play a role but there are mutual dependencies on their outcome. • In current NLP we often think about simpler structured problems: Parsing, Information Extraction, SRL, etc. • As we move up the problem hierarchy (Textual Entailment, QA,….) not all component models can be learned simultaneously • We need to think about (learned) models for different sub-problems • Knowledge relating sub-problems (constraints) becomes more essential and may appear only at evaluation time • Goal: Incorporate models’ information, along with prior knowledge (constraints) in making coherent decisions • Decisions that respect the local models as well as domain & context specific knowledge/constraints.

  6. Outline • Natural Language Processing with Constrained Conditional Models • A formulation for global inference with knowledge modeled as expressive structural constraints • Some examples • Extended semantic role labeling • Preposition based predicates and their arguments • Multiple simple models, Latent representations and Indirect Supervision • Amortized Integer Linear Programming Inference • Exploiting Previous Inference Results • Can the k-th inference problem be cheaper than the 1st?

  7. Three Ideas Underlying Constrained Conditional Models Modeling • Idea 1: Separate modeling and problem formulation from algorithms • Similar to the philosophy of probabilistic modeling • Idea 2: Keep models simple, make expressive decisions (via constraints) • Unlike probabilistic modeling, where models become more expressive • Idea 3: Expressive structured decisions can be supported by simply learned models • Global Inference can be used to amplify simple models (and even allow training with minimal supervision). Inference Learning

  8. Inference with General Constraint Structure [Roth&Yih’04,07]Recognizing Entities and Relations Dole ’s wife, Elizabeth , is a native of N.C. E1E2E3 R23 R12 Improvement over no inference: 2-5% Y = argmaxy score(y=v) [[y=v]] = = argmaxscore(E1 = PER)¢[[E1 = PER]] +score(E1 = LOC)¢[[E1 = LOC]] +… score(R1 = S-of)¢[[R1 = S-of]] +….. Subject to Constraints Note:Non Sequential Model An Objective function that incorporates learned models with knowledge (constraints) A constrained Conditional Model Key Questions: How to guide the global inference? How to learn? Why not Jointly? Models could be learned separately; constraints may come up only at decision time.

  9. Penalty for violating the constraint. Weight Vector for “local” models How far y is from a “legal” assignment Features, classifiers; log-linear models (HMM, CRF) or a combination Constrained Conditional Models (Soft) constraints component How to solve? This is an Integer Linear Program Solving using ILP packages gives an exact solution. Cutting Planes, Dual Decomposition & other search techniques are possible How to train? Training is learning the objective function Decouple? Decompose? How to exploit the structure to minimize supervision?

  10. Placing in context: a crash course in structured prediction Structured Prediction: Inference Joint features on inputs and outputs Set of allowed structures Feature Weights (estimated during learning) • Inference: given input x(a document, a sentence), predictthe best structure y= {y1,y2,…,yn} 2Y(entities & relations) • Assign values to the y1,y2,…,yn, accounting for dependencies among yis • Inference is expressed as a maximization of a scoring function y’ = argmaxy2 YwTÁ (x,y) • Inference requires, in principle, touching all y 2 Y at decision time, when we are given x 2 X and attempt to determine the best y 2 Y for it, given w • For some structures, inference is computationally easy. • Eg: Using the Viterbi algorithm • In general, NP-hard(can be formulated as an ILP)

  11. Structured Prediction: Learning Learning: given a set of structured examples {(x,y)} finda scoring function w that minimizes empirical loss. Learning is thus driven by the attempt to find a weight vector w such thatfor each given annotated example (xi, yi):

  12. Structured Prediction: Learning Score of annotated structure Score of any other structure Penalty for predicting other structure 8 y • Learning: given a set of structured examples {(x,y)} finda scoring function w that minimizes empirical loss. • Learning is thus driven by the attempt to find a weight vector w such thatfor each given annotated example (xi, yi): • We call these conditions the learning constraints. • In most learning algorithms used today, the update of the weight vector w is done in an on-line fashion, • Think about it as Perceptron; this procedure applies to Structured Perceptron, CRFs, Linear Structured SVM • W.l.o.g. (almost) we can thus write the generic structured learning algorithm as follows:

  13. In the structured case, the prediction (inference) step is often intractable and needs to be done many times Structured Prediction: Learning Algorithm • For each example (xi, yi) • Do: (with the current weight vector w) • Predict: perform Inference with the current weight vector • yi’ = argmaxy2 YwTÁ( xi ,y) • Check the learning constraints • Is the score of the current prediction better than of (xi, yi)? • If Yes – a mistaken prediction • Update w • Otherwise: no need to update w on this example • EndFor

  14. Solution I: decompose the scoring function to EASY and HARD parts Structured Prediction: Learning Algorithm EASY: could be feature functions that correspond to an HMM, a linear CRF, or even ÁEASY (x,y) = Á(x), omiting dependence on y, corresponding to classifiers. May not be enough if the HARD part is still part of each inference step. • For each example (xi, yi) • Do: • Predict: perform Inference with the current weight vector • yi’ = argmaxy2 YwEASYTÁEASY ( xi ,y) + wHARDTÁHARD ( xi ,y) • Check the learning constraint • Is the score of the current prediction better than of (xi, yi)? • If Yes – a mistaken prediction • Update w • Otherwise: no need to update w on this example • EndDo

  15. Solution II: Disregard some of the dependencies: assume a simple model. Structured Prediction: Learning Algorithm • For each example (xi, yi) • Do: • Predict: perform Inference with the current weight vector • yi’ = argmaxy2 YwEASYTÁEASY ( xi ,y) + wHARDTÁHARD ( xi ,y) • Check the learning constraint • Is the score of the current prediction better than of (xi, yi)? • If Yes – a mistaken prediction • Update w • Otherwise: no need to update w on this example • EndDo

  16. Structured Prediction: Learning Algorithm Solution III: Disregard some of the dependencies during learning; take into account at decision time This is the most commonly used solution in NLP today • For each example (xi, yi) • Do: • Predict: perform Inference with the current weight vector • yi’ = argmaxy2 YwEASYTÁEASY ( xi ,y) + wHARDTÁHARD ( xi ,y) • Check the learning constraint • Is the score of the current prediction better than of (xi, yi)? • If Yes – a mistaken prediction • Update w • Otherwise: no need to update w on this example • EndDo • yi’ = argmaxy2 YwEASYTÁEASY ( xi ,y) + wHARDTÁHARD ( xi ,y)

  17. (Soft) constraints component is more general since constraints can be declarative, non-grounded statements. Examples: CCM Formulations CCMs can be viewed as a general interface to easily combine declarative domain knowledge with data driven statistical models • Formulate NLP Problems as ILP problems (inference may be done otherwise) • 1. Sequence tagging (HMM/CRF + Global constraints) • 2. Sentence Compression (Language Model + Global Constraints) • 3. SRL (Independent classifiers + Global Constraints) Sequential Prediction HMM/CRF based: Argmax ¸ij xij Sentence Compression/Summarization: Language Model based: Argmax ¸ijk xijk Linguistics Constraints Cannot have both A states and B states in an output sequence. Linguistics Constraints If a modifier chosen, include its head If verb is chosen, include its arguments

  18. Archetypical Information Extraction Problem: E.g., Concept Identification and Typing, Event Identification, etc. Semantic Role Labeling I left my pearls to my daughter in my will . [I]A0left[my pearls]A1[to my daughter]A2[in my will]AM-LOC . • A0 Leaver • A1 Things left • A2 Benefactor • AM-LOC Location I left my pearls to my daughter in my will .

  19. Algorithmic Approach I left my nice pearls to her I left my nice pearls to her I left my nice pearls to her [ [ [ [ [ ] ] ] ] ] candidate arguments • Identify argument candidates • Pruning [Xue&Palmer, EMNLP’04] • Argument Identifier • Binary classification • Classify argument candidates • Argument Classifier • Multi-class classification • Inference • Use the estimated probability distribution given by the argument classifier • Use structural and linguistic constraints • Infer the optimal global output Ileftmy nice pearlsto her Use the pipeline architecture’s simplicity while maintaining uncertainty: keep probability distributions over decisions & use global inference at decision time.

  20. Semantic Role Labeling (SRL) I left my pearls to my daughter in my will .

  21. Semantic Role Labeling (SRL) I left my pearls to my daughter in my will .

  22. Semantic Role Labeling (SRL) I left my pearls to my daughter in my will . One inference problem for each verb predicate.

  23. No duplicate argument classes Reference-Ax Continuation-Ax Many other possible constraints: Unique labels No overlapping or embedding Relations between number of arguments; order constraints If verb is of type A, no argument of type B Constraints Any Boolean rule can be encoded as a set of linear inequalities. If there is an Reference-Ax phrase, there is an Ax If there is an Continuation-x phrase, there is an Ax before it Universally quantified rules Learning Based Java: allows a developer to encode constraints in First Order Logic; these are compiled into linear inequalities automatically.

  24. SRL: Posing the Problem Demo: http://cogcomp.cs.illinois.edu/

  25. y1 y1 y2 y2 y3 y3 y4 y4 y5 y5 y6 y6 y7 y7 y8 y8 Context: Constrained Conditional Models Conditional Markov Random Field Constraints Network y* = argmaxywiÁ(x; y) • Linear objective functions • Often Á(x,y) will be local functions, or Á(x,y) = Á(x) - i½i dC(x,y) • Expressive constraints over output variables • Soft, weighted constraints • Specified declaratively as FOL formulae • Clearly, there is a joint probability distribution that represents this mixed model. • We would like to: • Learn a simple model or several simple models • Make decisions with respect to a complex model Key difference from MLNs which provide a concise definition of a model, but the whole joint one.

  26. Constrained Conditional Models—Before a Summary • Constrained Conditional Models – ILP formulations – have been shown useful in the context of many NLP problems • [Roth&Yih, 04,07: Entities and Relations; Punyakanok et. al: SRL …] • Summarization; Co-reference; Information & Relation Extraction; Event Identifications; Transliteration; Textual Entailment; Knowledge Acquisition; Sentiments; Temporal Reasoning, Dependency Parsing,… • Some theoretical work on training paradigms [Punyakanok et. al., 05 more; Constraints Driven Learning, PR, Constrained EM…] • Some work on Inference, mostly approximations, bringing back ideas on Lagrangian relaxation, etc. • Good summary and description of training paradigms: [Chang, Ratinov & Roth, Machine Learning Journal 2012] • Summary of work & a bibliography: http://L2R.cs.uiuc.edu/tutorials.html

  27. Outline • Natural Language Processing with Constrained Conditional Models • A formulation for global inference with knowledge modeled as expressive structural constraints • Some examples • Extended semantic role labeling • Preposition based predicates and their arguments • Multiple simple models, Latent representations and Indirect Supervision • Amortized Integer Linear Programming Inference • Exploiting Previous Inference Results • Can the k-th inference problem be cheaper than the 1st?

  28. Verb SRL is not sufficient John, a fast-rising politician, slept on the train to Chicago. • Relation: sleep • Sleeper: John, a fast-rising politician • Location: on the train to Chicago. What was John’s destination? train to Chicago gives answer without verbs! Who was John? John, a fast-rising politician gives answer without verbs!

  29. Examples of preposition relations Queen of England City of Chicago

  30. Predicates expressed by prepositions Ambiguity & Variability live at Conway House at:1 Index of definition on Oxford English Dictionary Location stopped at 9 PM at:2 drive at 50 mph at:5 Temporal look at the watch at:9 Numeric cooler in the evening in:3 the camp on the island on:7 ObjectOfVerb arrive on the 9th on:17

  31. Preposition relations [Transactions of ACL, ‘13] • An inventory of 32 relations expressed by preposition • Prepositions are assigned labels that act as predicate in a predicate-argument representation • Semantically related senses of prepositions merged • Substantial inter-annotator agreement • A new resource: Word sense disambiguation data, re-labeled • SemEval 2007 shared task [Litkowski 2007] • ~16K training and 8K test instances • 34 prepositions • Small portion of the Penn Treebank [Dalhmeier, et al 2009] • only 7 prepositions, 22 labels

  32. Computational Questions • How do we predict the preposition relations? [EMNLP, ’11] • Capturing the interplay with verb SRL? • Very small jointly labeled corpus, cannot train a global model! • What about the arguments? [Trans. Of ACL, ‘13] • Annotation only gives us the predicate • How do we train an argument labeler?

  33. Predicting preposition relations Does not take advantage of the interactions between preposition and verb relations • Multiclass classifier • Uses sense disambiguation features • Depend on words syntactically connected to preposition [Hovy, et al. 2009] • Additional features based on NER, gazetteers, word clusters

  34. Coherence of predictions Predicate arguments from different triggers should be consistent Joint constraints linking the two tasks. Destination  A1 Location Destination Predicate: head.02 A0 (mover): The bus A1 (destination): for Nairobi in Kenya The bus was heading for Nairobi in Kenya.

  35. Variable ya,t indicates whether candidate argument a is assigned a label t. • ca,t is the corresponding score Joint inference Verb arguments Preposition relations Argument candidates Re-scaling parameters (one per label) Preposition relation label Constraints: Preposition Each argument label Verb SRL constraints Only one label per preposition Joint constraints (as linear inequalities)

  36. Desiderata for joint prediction Joint constraints between tasks, easy with ILP forumation Use small amount of joint data to re-scale scores to be in the same numeric range Joint Inference – no (or minimal) joint learning Intuition: The correct interpretation of a sentence is the one that gives a consistent analysis across all the linguistic phenomena expressed in it Should account for dependencies between linguistic phenomena Should be able to use existing state of the art models minimal use of expensive jointly labeled data

  37. Joint prediction helps [EMNLP, ‘11] Predicted together All results on Penn Treebank Section 23

  38. Example Weatherford said market conditions led to the cancellation of the planned exchange. • Independent preposition SRL mistakenly labels the toas a Location • Verb SRL identifies to the cancellation of the planned exchange as an A2for the verb led • Constraints (verbnet) prohibits an A2to be labeled as a Location, joint inference correctly switches the prediction to EndCondition

  39. Preposition relations and arguments Enforcing consistency between verb argument labels and preposition relations can help improve both • How do we predict the preposition relations? [EMNLP, ’11] • Capturing the interplay with verb SRL? • Very small jointly labeled corpus, cannot train a global model! • What about the arguments? [Trans. Of ACL, ‘13] • Annotation only gives us the predicate • How do we train an argument labeler?

  40. Indirect Supervision • In many cases, we are interested in a mapping from X to Y, but Y cannot be expressed as a simple function of X , hence cannot be learned well only as a function of X. • Consider the following sentences: S1: Druce will face murder charges, Conte said. S2: Conte said Druce will be charged with murder . • Are S1 and S2 a paraphrase of each other? • There is a need for an additional set of variables to justify this decision. • There is no supervision of these variables, and typically no evaluation of these, but learning to assign values to them support better prediction of Y. • Discriminative form of EM [Chang et. al ICML10’, NAACL’10] [[Yu & Joachims’09]

  41. Preposition arguments Poor care led to her death frompneumonia. • Cause(death, pneumonia) Governor Object Score candidates and select one governor and object

  42. Types are an abstraction that capture common properties of groups of entities. Relations depend on argument types Poor care led to her death frompneumonia. • Cause(death, pneumonia) • Poor care led to her death fromthe flu. How do we generalize to unseen words in the same “type”? Our primary goal is to model preposition relations and their arguments But the relation prediction strongly depends also on the semantic type of the arguments.

  43. WordNet IS-A hierarchy as types pneumonia => respiratory disease => disease => illness => ill health => pathological state => physical condition => condition => state => attribute => abstraction => entity Picking the right level in this hierarchy can generalize pneumonia and flu In addition to WordNet hypernyms, we also cluster verbs, nouns and adjectives using the dependency based word similarity of (Lin, 1998) and treat cluster membership as types. More general, but less discriminative Picking incorrectly will over-generalize

  44. Why are types important? Some semantic relations hold only for certain types of entities

  45. Predicate-argument structure of prepositions Supervision Prediction y Latent Structure Cause Relation r(y) flu Object Governor death h(y) experience disease Object type Governor type Poor care led to her death fromflu.

  46. Inference takes into account constrains among parts of the structure; formulated as an ILP Latent inference Standard inference: Find an assignment to the full structure Latent inference: Given an example with annotated r(y*) Given that we have constraints between r(y) and h(y) this process completes the structure in the best possible way to support correct prediction of the variables being supervised.

  47. Learning algorithm Generalization of Latent Structure SVM [Yu & Joachims’09] & Indirect Supervision learning [Chang et. al. ’10] • Initialize weight vector using multi-class classifier for predicates • Repeat • Use latent inference with current weight to “complete” all missing pieces • Train weight vector (e.g., with Structured SVM) • During training, a weight vector w is penalized more if it makes a mistake on r(y)

  48. Supervised Learning Score of annotated structure Score of any other structure Penalty for predicting other structure Completion of the hidden structure done in the inference step (guided by constraints) Completion of the hidden structure done in the inference step (guided by constraints) Penalty for making a mistake must not be the same for the labeled r(y) and inferred h(y) parts Learning (updating w) is driven by:

  49. Performance on relation labeling Model size: 2.21 non-zero weights Learned to predict both predicates and arguments Model size: 5.41 non-zero weights Using types helps. Joint inference with word sense helps more More components constrain inference results and improve inference

  50. Preposition relations and arguments Enforcing consistency between verb argument labels and preposition relations can help improve both Knowing the existence of a hidden structure lets us “complete” it and helps us learn • How do we predict the preposition relations? [EMNLP, ’11] • Capturing the interplay with verb SRL? • Very small jointly labeled corpus, cannot train a global model! • What about the arguments? [Trans. Of ACL, ‘13] • Annotation only gives us the predicate • How do we train an argument labeler?

More Related