390 likes | 493 Views
Contextual Vocabulary Acquisition ( for CSE 663). William J. Rapaport Department of Computer Science & Engineering, Department of Philosophy, Department of Linguistics and Center for Cognitive Science rapaport@cse.buffalo.edu http://www.cse.buffalo.edu/~rapaport.
E N D
Contextual Vocabulary Acquisition(for CSE 663) William J. Rapaport Department of Computer Science & Engineering, Department of Philosophy, Department of Linguistics and Center for Cognitive Science rapaport@cse.buffalo.edu http://www.cse.buffalo.edu/~rapaport
Contextual Vocabulary Acquisition • CVA = active, deliberate acquisition of a meaning for a word in a text by reasoning from “context” • CVA = what you do when: • You’re reading • You come to an unfamiliar word • It’s important for understanding the passage • No one’s around to ask • Dictionary doesn’t help • No dictionary • Too lazy to look it up :-) • Word not in dictionary • Definition of no use • Too hard • Inappropriate • So, you “figure out” a meaning for the word “from context” • “figure out” = compute (infer) a hypothesis about what the word might mean in that text • “context” = ??
What Does ‘Brachet’ Mean?(From Malory’s Morte D’Arthur [page # in brackets]) 1. There came a white hart running into the hall with a white brachet next to him, and thirty couples of black hounds came running after them. [66] • As the hart went by the sideboard,the white brachet bit him.[66] • The knight arose, took up the brachet androde away with the brachet.[66] • A lady came in and cried aloud to King Arthur,“Sire, the brachet is mine”.[66] • There was the white brachet which bayed at him fast.[72] 18. The hart lay dead; a brachet was biting on his throat,and other hounds came behind.[86]
What Is the “Context” for CVA? • “context” ≠ textual context • surrounding words; “co-text” of word • “context” = wide context = • “internalized” co-text … • ≈ reader’s interpretive mental model of textual “co-text” • … “integrated” via belief revision … • infer new beliefs from internalized co-text + prior knowledge • remove inconsistent beliefs • … with reader’s prior knowledge: • “world” knowledge • language knowledge • previous hypotheses about word’s meaning • but not including external sources (dictionary, humans) “Context” for CVA is in reader’s mind, not in the text
Prior Knowledge Text PK1 PK2 PK3 PK4
Prior Knowledge Text T1 PK1 PK2 PK3 PK4
Integrated KB Text T1 internalization PK1 PK2 PK3 PK4 I(T1)
B-R Integrated KB Text T1 internalization PK1 PK2 PK3 PK4 I(T1) inference P5
B-R Integrated KB Text T1 internalization PK1 PK2 PK3 PK4 I(T1) T2 inference P5 I(T2) P6
B-R Integrated KB Text T1 internalization PK1 PK2 PK3 PK4 I(T1) T2 inference T3 P5 I(T2) P6 I(T3)
B-R Integrated KB Text T1 internalization PK1 PK2 PK3 PK4 I(T1) T2 inference T3 P5 I(T2) P6 I(T3)
Note: All “contextual” reasoning is done in this “context”: B-R Integrated KB (the reader’s mind) Text T1 internalization PK1 PK2 PK3 PK4 P7 I(T1) T2 inference T3 P5 I(T2) P6 I(T3)
Overview of CVA Project • Background: • People do “incidental” CVA • Possibly best explanation of how we learn vocabulary • Given # of words high-school grad knows (~45K),& # of years to learn them (~18) = ~2.5K words/year • But only taught ~10% in 12 school years • Students are taught “deliberate” CVAin order to improve their vocabulary • CVA project: From Algorithm to Curriculum • Implemented computational theory of how tofigure out (compute) a meaning for an unfamiliar wordfrom “wide context” • Convert algorithms to an improved, teachable curriculum
Computational CVA • Implemented in SNePS (Shapiro 1979; Shapiro & Rapaport 1992) • Intensional, propositional semantic-networkknowledge-representation, reasoning, & acting system • Indexed by node: From any node, can describe rest of network • Serves as model of the reader (“Cassie”) • KB: SNePS representation of reader’s prior knowledge • I/P: SNePS representation of word in its co-text • Processing (“simulates”/“models”/is?! reading): • Uses logical inference, generalized inheritance, belief revisionto reason about text integrated with reader’s prior knowledge • N & V definition algorithms deductively search this “belief-revised, integrated” KB (the context) for slot fillers for definition frame… • O/P: Definition frame • slots (features): classes, structure, actions, properties, etc. • fillers (values): info gleaned from context (= integrated KB)
Cassie learns what “brachet” means:Background info about: harts, animals, King Arthur, etc.No info about: brachetsInput: formal-language (SNePS) version of simplified EnglishA hart runs into King Arthur’s hall.• In the story, B12 is a hart.• In the story, B13 is a hall.• In the story, B13 is King Arthur’s.• In the story, B12 runs into B13.A white brachet is next to the hart.• In the story, B14 is a brachet.• In the story, B14 has the property “white”.• Therefore, brachets are physical objects.(deduced while reading; PK: Cassie believes that only physical objects have color)
--> (defineNoun "brachet") Definition of brachet: Class Inclusions: phys obj, Possible Properties: white, Possibly Similar Items: animal, mammal, deer, horse, pony, dog, I.e., a brachet is a physical object that can be white and that might be like an animal, mammal, deer, horse, pony, or dog
A hart runs into King Arthur’s hall.A white brachet is next to the hart.The brachet bites the hart’s buttock.[PK: Only animals bite]--> (defineNoun "brachet") Definition of brachet: Class Inclusions: animal, Possible Actions: bite buttock, Possible Properties: white, Possibly Similar Items: mammal, pony,
A hart runs into King Arthur’s hall. A white brachet is next to the hart. The brachet bites the hart’s buttock. The knight picks up the brachet. The knight carries the brachet. [PK: Only small things can be picked up/carried] --> (defineNoun "brachet") Definition of brachet: Class Inclusions: animal, Possible Actions: bite buttock, Possible Properties: small, white, Possibly Similar Items: mammal, pony,
A hart runs into King Arthur’s hall.A white brachet is next to the hart.The brachet bites the hart’s buttock.The knight picks up the brachet.The knight carries the brachet.The lady says that she wants the brachet. [PK: Only valuable things are wanted]--> (defineNoun "brachet") Definition of brachet: Class Inclusions: animal, Possible Actions: bite buttock, Possible Properties: valuable, small, white, Possibly Similar Items: mammal, pony,
A hart runs into King Arthur’s hall.A white brachet is next to the hart.The brachet bites the hart’s buttock.The knight picks up the brachet.The knight carries the brachet.The lady says that she wants the brachet. The brachet bays at Sir Tor. [PK: Only hunting dogs bay] --> (defineNoun "brachet") Definition of brachet: Class Inclusions: hound, dog, Possible Actions: bite buttock, bay, hunt, Possible Properties: valuable, small, white, I.e. A brachet is a hound (a kind of dog) that can bite, bay, and hunt, and that may be valuable, small, and white.
General Comments • Cassie’s behavior human protocols • Cassie’s definition OED’s definition: = A brachet is “a kind of hound which hunts by scent”
Fragment of reader’s prior knowledge: m3 = In “real life”, white is a color Member(Lex(white),Lex(color),LIFE) m6 = In “real life”, harts are deer AKO(Lex(hart),Lex(deer),LIFE) m8 = In “real life”, deer are mammals AKO(Lex(deer),Lex(mammal),LIFE) m11 = In “real life”, halls are buildings AKO(Lex(hall),Lex(building),LIFE) m12 = In “real life”, b1 is named “King Arthur” Name(b1,”King Arthur”,LIFE) m14 = In “real life”, b1 is a king Isa(ISA,b1,Lex(king),LIFE) (etc.)
m16 = if v3 has property v2 & v2 is a color & v3 v1 then v1 is a class of physical objects all(x,y,z)({Is1(z,y),Member1(y,lex(color)),Member1(z,x)} &=> {AKO1(x,lex(physical object))})
Reading the story: m17 = In the story, b2 is a hart ISA(b2,lex(hart),STORY) m24 = In the story, the hart runs into b3 Does(b2,into(b3,lex(run)),STORY) (b3 is King Arthur’s hall) – not shown (harts are deer) – not shown
A fragment of the entire network showing the reader’s mental context consisting of prior knowledge, the story, & inferences. The definition algorithm searches this entire network,abstracts parts of it, & produces a hypothesized meaning for ‘brachet’.
Noun Algorithm • Generate initial hypothesis by “syntactic manipulation” • Algebra: Solve an equation for unknown value X • Syntax: “Solve” a sentence for unknown word X • “A white brachet (X) is next to the hart”→ X (a brachet) is something that is next to the hart and that can be white I.e., “define” node X in terms of immediately connected nodes • Then find or infer from wide context: • Basic-level class memberships (e.g., “dog”, rather than “animal”) • else most-specific-level class memberships • else names of individuals • Properties of Xs (else, of individual Xs) (e.g., size, color, …) • Structure of Xs (else …) (part-whole, physical structure…) • Acts that Xs perform (else …) or that can be done to/with Xs • Agents that do things to/with Xs • … or to whom things can be done with Xs • … or that own Xs • Possible synonyms, antonyms I.e., “define” word X in terms of some (but not all) other connected nodes
Verb Algorithm • Generate initial hypothesis by syntactic/algebraic manipulation • Then find or infer from wide context: • Class membership (e.g., Conceptual Dependency) • What kind of act is X-ing (e.g., walking is a kind of moving) • What kinds of acts are X-ings (e.g., sauntering is a kind of walking) • Properties/manners of X-ing (e.g., moving by foot, slow walking) • Transitivity/subcategorization information • Return class membership of agent, object, indirect object, instrument • Possible synonyms, antonyms • Causes & effects • [Also: preliminary work on adjective algorithm]
Belief Revision • To revise definitions of words used inconsistently with current meaning hypothesis • SNeBR (ATMS; Martins & Shapiro 1988, Johnson 2006): • If inference leads to a contradiction, then: • SNeBR asks user to remove culprit(s) • & automatically removes consequences inferred from culprit
Revision & Expansion • Removal & revision being automated via SNePSwD by ranking all propositions with kn_cat: most intrinsic info re: language; fundamental background info certain (“before” is transitive) story info in text (“King Lot rode to town”) life background info w/o variables or inference (“dogs are animals”) story-comp info inferred from text (King Lot is a king, rode on a horse) life-rule.1 everyday commonsense background info (BearsLiveYoung(x) Mammal(x)) life-rule.2 specialized background info (x smites y x kills y by hitting y) least certain questionable already-revised life-rule.2; not part of input
Belief Revision: ‘smite’ • Misunderstood word: • Initially believe that ‘smite’ means:“kill by hitting” • Read “King Lot smote down King Arthur” • Infer that King Arthur is dead • Then read: “King Arthur drew his sword Excalibur” • Contradiction! • Weaken definition to: “hit and possibly kill” • Then read more passages in which smiting ⇒ killing • Hypothesize that ‘smite’ means “hit”
Belief Revision: ‘dress’ • Well-entrenched word… • Believe ‘dress’ means “put clothes on” • Commonsense belief: • Spears don’t wear clothing • … used in new sense: • Read “King Claudius dressed his spear” • Infer that spear wears clothing • Contradiction! • Modify definition to: “put clothes on OR something else” • Read “King Arthur dressed his troops before battle” • Infer that ‘dress’ means: “put clothes on OR prepare for battle” • Eventually: Induce more general definition: • “prepare” (for the day, for battle, for eating…)
A Computational Theory of CVA • A word does not have a unique meaning. • A word does not have a “correct” meaning. • Author’s intended meaning for word doesn’t need to be known by readerin order for reader to understand word in context • Even familiar/well-known words can acquire new meanings in new contexts. • Neologisms are usually learned only from context • Every co-text can give some clue to a meaning for a word. • Generate initial hypothesis via syntactic/algebraic manipulation • But co-text must be integrated with reader’s prior knowledge • Large co-text + large PK more clues • Lots of occurrences of word allow asymptotic approach to stable meaning hypothesis • CVA is computable • CVA is “open-ended”, hypothesis generation. • CVA ≠ guess missing word (“cloze”); CVA ≠ word-sense disambiguation • Some words are easier to compute meanings for than others (N < V < Adj/Adv) • CVA can improve general reading comprehension (through active reasoning) • CVA can & should be taught in schools
From Algorithm to Curriculum • State of the art in vocabulary learning from context: • Mauser 1984: “context” = definition! • Clarke & Nation 1980: a “strategy” (algorithm?): • Determine part of speech of word • Look at grammatical context • Who does what to whom? • Look at surrounding textual context • Search for clues (as we do) • Guess the word; check your guess
CVA: From Algorithm to Curriculum • “guess the word” = “then a miracle occurs” • Surely, computer scientists can “be more explicit”! • And so should teachers!
From Algorithm to Curriculum (cont’d) • We have explicit, GOF (symbolic) AI theory of how to do CVA Teachable! • Goal: • Not: teach people to “think like computers” • But: explicate computable & teachable methods to hypothesize word meanings from context • AI as computational psychology: • Devise computer programs that faithfully simulate(human) cognition • Can tell us something about (human) mind • Joint work with Michael Kibby (UB Reading Clinic) • We are teaching a machine, to see if what we learn in teaching it can help us teach students better
If time, continue w/ holism & CRA slides • Else goto case frames: • http://www.cse.buffalo.edu/~rapaport/CVA/CaseFrames/case-frames/ • Then goto SNePSLOG predicates: • http://www.cse.buffalo.edu/~rapaport/663/F08/snepslogcaseframes
CVA as Computational Philosophy & Philosophical Computation • CVA & holistic semantic theories: • Semantic networks: • “Meaning” of a node is its location in the entire network • Holism: • Meaning of a word is its relationships to all other words in the language • Problems (Fodor & Lepore): • No 2 people ever share a belief • No 2 people ever mean the same thing • No 1 person ever means the same thing at different times • No one can ever change his/her mind • Nothing can be contradicted • Nothing can be translated • CVA offers principled way to restrict “entire network”to a useful subnetwork • That subnetwork can be shared across people, individuals, languages,… • Can also account for language/concept change • Via “dynamic”/“incremental” semantics
CVA as Computational Philosophy & Philosophical Computation (cont’d) • CVA and the Chinese Room • How would Searle-in-the-Room figure out the meaning of an unknown squiggle? • By CVA techniques! • Searle’s CR argument from semantics: • Computer programs are purely syntactic • Cognition is semantic • Syntax alone does not suffice for semantics No purely syntactic computer program can exhibit semantic cognition • “Syntactic Semantics” (Rapaport 1985ff) • Syntax does suffice for the kind of semantics needed for NLU in the CR • All input—linguistic, perceptual, etc.—is encoded in a single network(or: in a single, real neural network: the brain!) • Relations—including semantic ones—among nodes of such a networkare manipulated syntactically • Hence computationally (CVA helps make this precise)
Summary • Contextual Vocabulary Acquisition project is: • Computational philosophy • And computational psychology! • Philosophical computation • With applications to: • Computational linguistics • Reading education