1 / 67

Organization and Emergence of Semantic Knowledge: A Parallel-Distributed Processing Approach

Organization and Emergence of Semantic Knowledge: A Parallel-Distributed Processing Approach. James L. McClelland Department of Psychology and Center for Mind, Brain, and Computation Stanford University. Some Phenomena in Conceptual Development. Progressive differentiation of concepts

rgonzales
Download Presentation

Organization and Emergence of Semantic Knowledge: A Parallel-Distributed Processing Approach

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Organization and Emergence ofSemantic Knowledge: A Parallel-Distributed Processing Approach James L. McClelland Department of Psychology and Center for Mind, Brain, and Computation Stanford University

  2. Some Phenomena in Conceptual Development • Progressive differentiation of concepts • Illusory correlations and U-shaped developmental trajectories • Conceptual reorganization • Domain- and property-specific constraints on generalization • Acquired sensitivity to an object’s causal properties • What underlies these phenomena?

  3. Naïve Domain Theories? • Mechanisms of learning are thought to be too weak • They learn by contiguity and generalize by similarity… • But generalization is domain dependent. • .. so it is proposed instead that development begins with initial constraints, in the form of innately pre-specified proto-theories that guide inference and learning.

  4. An Alternative View:Sensitivity to Coherent Covariation • Coherent Covariation: • The tendency of properties of objects to co-occur in clusters. • e.g. • Has wings • Can fly • Is light • Or • Has roots • Has rigid cell walls • Can grow tall

  5. Our Answer in More Detail • Domain general mechanisms sensitive to experience underlie the development and elaboration of conceptual knowledge. • These mechanisms exploit the principles of parallel-distributed processing. • Models built on these principles are sensitive to coherent covariation. • This sensitivity is the main cause of all of the phenomena.

  6. Principles of Parallel Distributed Processing /h/ /i/ /n/ /t/ • Processing occurs via interactions among neuron-like processing units via weighted connections. • A representation is a pattern of activation. • The knowledge is in the connections. • Learning occurs through gradual connection adjustment, driven by experience. • Both representation and processing are affected. H I N T

  7. Principles of Parallel Distributed Processing /h/ /i/ /n/ /t/ • Processing occurs via interactions among neuron-like processing units via weighted connections. • A representation is a pattern of activation. • The knowledge is in the connections. • Learning occurs through gradual connection adjustment, driven by experience. • Both representation and processing are affected. H I N T

  8. Principles of Parallel Distributed Processing /h/ /i/ /n/ /t/ • Processing occurs via interactions among neuron-like processing units via weighted connections. • A representation is a pattern of activation. • The knowledge is in the connections. • Learning occurs through gradual connection adjustment, driven by experience. • Both representation and processing are affected. H I N T

  9. Principles of Parallel Distributed Processing /h/ /i/ /n/ /t/ • Processing occurs via interactions among neuron-like processing units via weighted connections. • A representation is a pattern of activation. • The knowledge is in the connections. • Learning occurs through gradual connection adjustment, driven by experience. • Both representation and processing are affected. H I N T

  10. Differentiation in Developmentand in a simple PDP network

  11. The Rumelhart Model

  12. Quillian’sHierarchicalPropositional Model

  13. The Rumelhart Model: Target output for ‘robin can’ input

  14. The Training Data: All propositions true of items at the bottom levelof the tree, e.g.: Robin can {fly, move, grow}

  15. aj wij ai neti=Sajwij wki Forward Propagation of Activation

  16. Back Propagation of Error (d) aj wij ai di ~ Sdkwki wki dk ~ (tk-ak) Error-correcting learning: At the output layer: Dwki = edkai At the prior layer: Dwij = edjaj …

  17. The Rumelhart Model

  18. What Drives Progressive Differentiation? • Waves of differentiation reflect coherent covariation of properties across items. • Patterns of coherent covariation are reflected in the principal components of the property covariance matrix. • Figure shows attribute loadings on the first three principal components: • 1. Plants vs. animals • 2. Birds vs. fish • 3. Trees vs. flowers • Same color = features covary in component • Diff color = anti-covarying features

  19. “Now wait just a minute…” • Didn’t you tell the network the taxonomic organization directly? • Pine ISA Tree, Plant • Robin ISA Bird, Animal • Yes we did. • We do think names kids hear for things affect their conceptual representations. • But labels aren’t necessary as long as an item’s properties exhibit coherent covariation.

  20. 12345678910111213141516 Properties Coherent Incoherent CoherenceTraining Patterns Items No labels are provided Each item and each property occurs with equal frequency

  21. 12345678910111213141516 Properties Coherent Incoherent ISCANHAS Contexts: Items Note: coherence is present between, not within training experiences!

  22. Effects of Coherence on Learning CoherentProperties Incoherent Properties

  23. Effect of Coherence on Representation

  24. Effects of Coherent Variation on Learning in Connectionist Models • Attributes that vary together create the acquired concepts that populate the taxonomic hierarchy, and determine which properties are central and which are incidental to a given concept. • Labeling of these concepts or their properties is in no way necessary. • But it is easy to learn names for such concepts. • Arbitrary properties (those that do not co-vary with others) are very difficult to learn. • And it is harder to learn names for concepts that are only differentiated by such arbitrary properties.

  25. Where are we on that list of Phenomena? • Progressive differentiation of concepts • Illusory correlations and U-shaped developmental trajectories • Conceptual reorganization • Domain- and property-specific constraints on generalization • Acquired sensitivity to an object’s causal properties

  26. Illusory Correlations • Rochel Gelman found that children think that all animals have feet. • Even animals that look like small furry balls and don’t seem to have any feet at all.

  27. A typical property thata particular object lacks e.g., pine has leaves An infrequent, atypical property

  28. Conceptual Reorganization (Carey, 1985) • Carey demonstrates that young children ‘discover’ the unity of plants and animals as living things with many shared properties only around the age of 10. • She suggests that the coalescence of the concept of living thing depends on learning about diverse aspects of plants and animals including • Nature of life sustaining processes • What it means to be dead vs. alive • Reproductive properties • Can reorganization occur in a connectionist net?

  29. Conceptual Reorganization in the Model • Suppose superficial appearance information, which is not coherent with much else, is always available… • And there is a pattern of coherent covariation across information that is contingently available in different contexts. • The model forms initial representations based on superficial appearances. • Later, it discovers the shared structure that cuts across the different contexts, reorganizing its representations.

  30. Organization of Conceptual Knowledge Early and Late in Development

  31. Inference and Generalizationin the PDP Model • A semantic representation for a new item can be derived by error propagation from given information, using knowledge already stored in the weights. • Crucially: • The similarity structure, and hence the pattern of generalization depends on the knowledge already stored in the weights.

  32. Start with a neutral representation on the representation units. Use backprop to adjust the representation to minimize the error.

  33. The result is a representation similar to that of the average bird…

  34. Use the representation to infer what this new thing can do.

  35. Inference and Generalizationin the PDP Model • A semantic representation for a new item can be derived by error propagation from given information, using knowledge already stored in the weights. • Crucially: • The similarity structure, and hence the pattern of generalization, depends on the knowledge already stored in the weights.

  36. Domain Specificity • What constraints are required for development and elaboration of domain-specific knowledge? • Are domain specific constraints required? • Or are there general principles that allow for acquisition of conceptual knowledge of all different types?

  37. Differential Importance (Marcario, 1991) • 3-4 yr old children see a puppet and are told he likes to eat, or play with, a certain object (e.g., top object at right) • Children then must choose another one that will “be the same kind of thing to eat” or that will be “the same kind of thing to play with”. • In the first case they tend to choose the object with the same color. • In the second case they will tend to choose the object with the same shape.

  38. Can the knowledge that one kind of property is important for one type of thing while another is important for a different type of thing be learned?

  39. Adjustments to Training Environment • Among the plants: • All trees are large • All flowers are small • Either can be bright or dull • Among the animals: • All birds are bright • All fish are dull • Either can be small or large • In other words: • Size covaries with properties that differentiate different types of plants • Brightness covaries with properties that differentiate different types of animals

  40. Testing Feature Importance • After partial learning, model is shown eight test objects: • Four “Animals”: • All have skin • All combinations of bright/dull and large/small • Four “Plants”: • All have roots • All combinations of bright/dull and large/small • Representations are generated by usingback-propagation to representation.

  41. Similarities of Obtained Representations Brightness is relevant for Animals Size is relevant for Plants

  42. In Rogers and McClelland (2004) we also address: • Conceptual differentiation in prelinguistic infants. • Many of the phenomena addressed by classic work on semantic knowledge from the 1970’s: • Basic level • Typicality • Frequency • Expertise • Disintegration of conceptual knowledge in semantic dementia • How the model can be extended to capture causal properties of objects and explanations. • What properties a network must have to be sensitive to coherent covariation.

  43. A Coherence Requires Convergence A A

  44. language Semantic Representation in the Brain • Damage to temporal pole is associated with semantic dementia, a domain-general loss of semantic information • Imaging and lesion studies suggest that other brain areas are associated with more specific types of information. • We suggest that the temporal pole serves as the convergent semantic representation in the brain. • With bi-directional connections to regions containing modality specific information. • The interface with language occurs via connections between language areas and temporal pole.

  45. In summary • Sensitivity to coherent co-variation in experience can explain many aspects of conceptual development. • PDP networks subject to a domain-general architectural constraint provide the necessary mechanisms. • Our simulations do not prove domain general learning methods will turn out to be fully sufficient. • There is still room for domain- or content-specific constraints • And the framework is fully compatible with their integration. • But our findings suggest it may be worth exploring how far we can go without them.

  46. Thanks for your attention!

More Related