1 / 12

Improving Machine Learning Approaches to Coreference Resolution

Improving Machine Learning Approaches to Coreference Resolution. Vincent Ng and Claire Cardie Cornell Univ. ACL 2002 slides prepared by Ralph Grishman. Goal. Improve on Soon et al. by better preprocessing (chunking, names, …) better search procedure for antecedent

leoma
Download Presentation

Improving Machine Learning Approaches to Coreference Resolution

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Improving Machine Learning Approaches to Coreference Resolution Vincent Ng and Claire Cardie Cornell Univ. ACL 2002 slides prepared by Ralph Grishman

  2. Goal Improve on Soon et al. by • better preprocessing (chunking, names, …) • better search procedure for antecedent • better selection of positive examples • more features • more features • more features ...

  3. Better search for antecedent • Soon et al. Use decision tree as binary classifier, take nearest antecedent classified as +ve • Ng&Cardie use same sort of classifier, but count +ve and -ve examples at each leaf, and use that to compute a probability • Ng&Cardie then take highest ranking antecedent (if probability > 0.5)

  4. Better choice of positive examples • Soon et al. always use most recent antecedent • For Ng&Cardie, if anaphor is not a pronoun, they use most recent antecedent that is not a pronoun

  5. More features #1 • Soon et al. Have a ‘same string’ feature • Ng&Cardie split this up into 3 features, for pronominals, nominals, and names

  6. First improvements: F scores

  7. More features Added 41 more features: • lexical • grammatical • semantic

  8. Lexical features (examples) • Non-empty overlap of words of two NPs • Prenominal modifiers of one NP are a subset of prenominal modifiers of other

  9. Grammatical features (examples) • NPs are in predicate nominal construct • One NP spans the other • NP1 is a quoted string • One of the NPs is a title

  10. Semantic features (examples) For nominals with different heads • direct or indirect hypernym relation in WordNet • distance of hypernym relation • sense number for hypernym relation

  11. Selecting features • Full feature set yielded very low precision on nominal anaphors • overtraining: too many features for too little data • So they (manually) eliminated many features which led to low precision (on training data) • no ‘development set’ separate from training and test sets

  12. Adding features: F scores

More Related