1 / 36

A computational study of cross-situational techniques for learning word-to-meaning mappings

A computational study of cross-situational techniques for learning word-to-meaning mappings. Jeffrey Mark Siskind Presented by David Goss-Grubbs March 5, 2006. The Problem: Mapping Words to Concepts. Child hears John went to school Child sees GO( John , TO( school )) Child must learn

kalil
Download Presentation

A computational study of cross-situational techniques for learning word-to-meaning mappings

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. A computational study of cross-situational techniques for learning word-to-meaning mappings Jeffrey Mark Siskind Presented by David Goss-Grubbs March 5, 2006

  2. The Problem: Mapping Words to Concepts • Child hears John went to school • Child sees GO(John, TO(school)) • Child must learn • John John • went  GO(x, y) • to  TO(x) • school  school

  3. Two Problems • Referential uncertainty • MOVE(John, feet) • WEAR(John, RED(shirt)) • Determining the correct alignment • John TO(x) • walked  school • to  John • school  GO(x, y)

  4. Helpful Constraints • Partial Knowledge • Cross-situational inference • Covering constraints • Exclusivity

  5. Partial Knowledge • Child hears Mary lifted the block • Child sees • CAUSE(Mary, GO(block, UP)) • WANT(Mary, block) • BE(block, ON(table)) • If the child knows lift contains CAUSE, the second two hypotheses can be ruled out.

  6. Cross-situational inference • John lifted the ball CAUSE(John, GO(ball, UP)) • Mary lifted the block  CAUSE(Mary, GO(block, UP)) • Thus, lifted  {UP, GO(x, y), GO(x, UP), CAUSE(x, y), CAUSE(x, GO(y, z)), CAUSE(x, GO(y, UP))}

  7. Covering constraints • Assume: all components of an utterance’s meaning come from the meanings of words in that utterance. • If it is known that CAUSE is not part of the meaning of John, the or ball, it must be part of the meaning of lifted. • (But what about constructional meaning?)

  8. Exclusivity • Assume: any portion of the meaning of an utterance comes from no more than one of its words. • If John walked WALK(John) andJohn  JohnThen walked can be no more thanwalked  WALK(x)

  9. Three more problems • Bootstrapping • Noisy Input • Homonymy

  10. Bootstrapping • Lexical acquisition is much easier if some of the language is already known • Some of Siskind’s strategies (e.g. cross-situational learning) work without such knowledge • Others (e.g. exclusivity) require it. • The algorithm starts off slow, then speeds up

  11. Noise • Only a subset of all possible meanings will be available to the algorithm • If none of them contain the correct meaning, cross-situational learning would cause those words never to be acquired • Some portion of the input must be ignored. • (A statistical approach is rejected – it is not clear why)

  12. Homonymy • Similar to noisy input, cross-situational techniques would fail to find a consistent mapping for homonymous words. • When an inconsistency is found, a split is made. • If the split is corroborated, a new sense is created; otherwise it is noise.

  13. The problem, formally stated • From: a sequence of utterances • Each utterance is an unordered collection of words • Each utterance is paired with a set of conceptual expressions • To: a lexicon • The lexicon maps each word to a set of conceptual expressions, one for each sense of the word

  14. Composition • Select one sense for each word • Find all ways of combining these conceptual expressions • The meaning of an utterance is derived only from the meaning of its component words. • Every conceptual expression in the meanings of the words must appear in the final conceptual expression (copies are possible)

  15. The simplified algorithm: no noise or homonymy • Two learning stages • Stage 1: The set of conceptual symbols • E.g. {CAUSE, GO, UP} • Stage 2: The conceptual expression • CAUSE(x, GO(y, UP))

  16. Stage 1: Conceptual symbol set • Maintain sets of necessary and possible conceptual symbols for each word • Initialize the former to the empty set and the latter to the universal set • Utterances will increase the necessary set and decrease the possible set, until they converge on the actual conceptual symbol set

  17. Stage 2: Conceptual expression • Maintain a set of possible conceptual expressions for each word • Initialize to the set of all expressions that can be composed from the actual conceptual symbol set • New utterances will decrease the possible conceptual expression set until only one remains

  18. Example

  19. Selecting the meaning John took the ball • CAUSE(John, GO(ball, TO(John))) • WANT(John, ball) • CAUSE(John, GO(PART-OF (LEFT(arm), John), TO(ball))) • Second is eliminated because no CAUSE • Third is eliminated because no word has LEFT or PART-OF

  20. Updated table

  21. Stage 2

  22. Noise and Homonymy • Noisy or homonymous data can corrupt the lexicon • Adding an incorrect element to the set of necessary elements • Taking a correct element away from the set of possible elements • This may or may not create an inconsistent entry

  23. Extended algorithm • Necessary and possible conceptual symbols are mapped to senses rather than words • Words are mapped to their senses • Each sense has a confidence factor

  24. Sense assignment • For each utterance, find the cross-product of all the senses • Choose the “best” consistent sense assignment • Update the entries for those senses as before • Add to a sense’s confidence factor each time it is used in a preferred assignment

  25. Inconsistent utterances • Add the minimal number of new senses until the utterance is no longer inconsistent – three possibilities • If the current utterance is noise, new senses are bad (and will be ignored) • There really are new senses • The original senses were bad, and the right senses are only now being added. • On occasion, remove senses with low confidence factors

  26. Four simulations • Vary the task along five parameters • Vocabulary growth rate by size of corpus • Number of required exposures to a word by size of corpus • How high can it scale?

  27. Method (1 of 2) • Construct a random lexicon • Vary it by three parameters • Vocabulary size • Homonymy rate • Conceptual-symbol inventory size

  28. Method (2 of 2) • Construct a series of utterances, each paired with a set of meaning hypotheses • Vary this by the following parameters • Noise rate • Degree of referential uncertainty • Cluster size (5) • Similarity probability (.75)

  29. Sensitivity analysis

  30. Vocabulary size

  31. Degree of referential uncertainty

  32. Noise rate

  33. Conceptual-symbol inventory size

  34. Homonymy rate

  35. Vocabulary Growth

  36. Number of exposures

More Related