1 / 31

The PDP Approach to Understanding the Mind and Brain

The PDP Approach to Understanding the Mind and Brain. J. McClelland Cognitive Core Class Lecture March 7, 2011. Decartes’ Legacy. Mechanistic approach to sensation and action Divine inspiration creates mind This leads to four dissociations: Mind / Brain

johniet
Download Presentation

The PDP Approach to Understanding the Mind and Brain

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. The PDP Approach to Understanding the Mind and Brain J. McClelland Cognitive Core Class Lecture March 7, 2011

  2. Decartes’ Legacy • Mechanistic approach to sensation and action • Divine inspiration creates mind • This leads to four dissociations: • Mind / Brain • Higher Cognitive Functions / Sensory-motor systems • Human / Animal • Descriptive / Mechanistic

  3. Early Computational Models of Human Cognition (1950-1980) • The computer contributes to the overthrow of behaviorism. • Computer simulation models emphasize strictly sequential operations, using flow charts. • Simon announces that computers can ‘think’. • Symbol processing languages are introduced allowing some success at theorem proving, problem solving, etc. • Minsky and Pappert kill off Perceptrons. • Cognitive psychologists distinguish between algorithm and hardware. • Neisser deems physiology to be only of ‘peripheral interest’ • Psychologists investigate mental processes as sequences of discrete stages.

  4. Ubiquity of the Constraint SatisfactionProblem • In sentence processing • I saw the grand canyon flying to New York • I saw the sheep grazing in the field • In comprehension • Margie was sitting on the front steps when she heard the familiar jingle of the “Good Humor” truck. She remembered her birthday money and ran into the house. • In reaching, grasping, typing…

  5. Graded and variable nature of neuronal responses

  6. Lateral Inhibition in Eye of Limulus (Horseshoe Crab)

  7. The Interactive Activation Model

  8. dog goat hammer dog goat hammer Distributed Representations in the Brain:Overlapping Patterns for Related Concepts (Kiani et al, 2007) • Many hundreds of single neurons recorded in monkey IT. • 1000 different photographs were presented twice each to each neuron. • Hierarchical clustering based on the distributed representation of each picture: • The pattern of activation over all the neurons

  9. Kiani et al, J Neurophysiol 97: 4296–4309, 2007.

  10. The Rumelhart Model The QuillianModel

  11. DER’s Goals for the Model • Show how learning could capture the emergence of hierarchical structure • Show how the model could make inferences as in the Quillian model

  12. Early Later LaterStill Experie nce

  13. Start with a neutral representation on the representation units. Use backprop to adjust the representation to minimize the error.

  14. The result is a representation similar to that of the average bird…

  15. Use the representation to infer what this new thing can do.

  16. Questions About the Rumelhart Model • Does the model offer any advantages over other approaches? • Do distributed representations really buy us anything? • Can the mechanisms of learning and representation in the model tell us anything about • Development? • Effects of neuro-degeneration?

  17. Phenomena in Development • Progressive differentiation • Overgeneralization of • Typical properties • Frequent names • Emergent domain-specificity of representation • Basic level advantage • Expertise and frequency effects • Conceptual reorganization

  18. Disintegration in Semantic Dementia • Loss of differentiation • Overgeneralization

  19. Animals Plants Birds Fish Flowers Trees The Hierarchical Naïve Bayes Classifier Model (with R. Grosse and J. Glick) • The world consists of things that belong to categories. • Each category in turn may consist of things in several sub-categories. • The features of members of each category are treated as independent • P({fi}|Cj) = Pi p(fi|Cj) • Knowledge of the features is acquired for the most inclusive category first. • Successive layers of sub-categories emerge as evidence accumulates supporting the presence of co-occurrences violating the independence assumption. Living Things …

  20. A One-Class and a Two-Class Naïve Bayes Classifier Model

  21. Accounting for the network’s feature attributions with mixtures of classes at different levels of granularity Regression Beta Weight Epochs of Training Property attribution model: P(fi|item) = akp(fi|ck) + (1-ak)[(ajp(fi|cj) + (1-aj)[…])

  22. Should we replace the PDP model with the Naïve Bayes Classifier? • It explains a lot of the data, and offers a succinct abstract characterization • But • It only characterizes what’s learned when the data actually has hierarchical structure • So it may be a useful approximate characterization in some cases, but can’t really replace the real thing.

  23. Structure Extracted by a Structured Statistical Model

  24. Predictions • Similarity ratings (and patterns of inference) will violate the hierarchical structure • Patterns of inference will vary by context

  25. Experiments • Size, predator/prey, and other properties affect similarity across birds, fish, and mammals • Property inferences show clear context specificity • Future experiments will examine whether inferences (even of biological properties) violate a hierarchical tree for items like weasels, pandas, and beavers

  26. The Nature of Cognition, and the Place of PDP in Cognitive Theory? • Many view human cognition as inherently • Structured • Systematic • Rule-governed • In this framework, PDP models are seen as • Mere implementations of higher-level, rational, or ‘computational level’ models • … that don’t work as well as models that stipulate explicit rules or structures

  27. The Alternative • We argue instead that cognition (and the domains to which cognition is applied) is inherently • Quasi-regular • Semi-systematic • Context sensitive • On this view, highly structured models: • Are Procrustian beds into which natural cognition fits uncomfortably • Won’t capture human cognitive abilities as well as models that allow a more graded and context sensitive conception of structure

  28. Levels of Analysis • Marr (1982) suggested we should analyze cognitive tasks at three levels: • Computation: what are the goals, what information is available, how could the information be used to achieve the goals; what is the best that can be done with the given information? • Algorithms and representations: How is information represented? What algorithms are used in manipulating representations? • Implementation: How are the algorithms and representations implemented in neural circuitry? • PDP models often closely approximate (and can in many cases exactly match) idealized competence models (including structured probabilistic models). • Which is the approximation? • The PDP approach encourages computational level analysis but asks many questions about it: • How do we know what task – which computations – an organism is actually trying to carry out? • Is performance constrained by tasks the organism was trying to perform when it evolved or that it has performed habitually? Such constraints may be ‘wired into’ the processing mechanism, constraining its performance and preventing optimality for a given task. • The approach leads us to ask: How does the architecture and/or type of processing machinery constrain the problem and its solution? Perhaps performance is being optimized within such constraints? • The PDP approach also blurs the distinction between the algorithmic and implementation levels • PDP models generally do not concern themselves with the minute details of neural implementation, and their performance often approximates performance that would be achieved by an explicit algorithm – thus they appear to lie between Marr’s algorithmic and implementation levels • PDP models do not deny that there are temporally extended cognitive processes, e.g. in problem solving and planning, that involve many steps and that can often be usefully characterized in terms of a sequence of discrete states (but leave open the possibility that insight and creativity short-circuit such processes). • The automatic and intuition-based nature of PDP models may, however, be very relevant even in our most advanced forms of cognition.

More Related