100 likes | 107 Views
Explore the fundamentals of Bayes' Theorem, Bayesian models of cognition, distributions, and cognitive modeling applications. Learn how to calculate probabilities, interpret likelihoods, and make informed decisions based on prior knowledge. Discover the relationship between prior, likelihood, posterior probabilities. Dive into examples like the hit-and-run case to grasp practical implementations. Enhance your understanding of Bayesian theory through demonstrations and applications in cognitive science.
E N D
Honors 207, Intro to Cognitive Science David Allbritton An introduction to Bayes' Theorem and Bayesian models of human cognition Bayesian Models
Bayes Theorem: An Introduction • What is Bayes' Theorem? • What does it mean? • An application: Calculating a probability • What are distributions? • Bayes' Theorem using distributions • An application to cognitive modelling: perception
What is Bayes' Theorem? • A theorem of probability theory • Concerns “conditional probabilities” • Example 1: Probability class is cancelled if it snows • P (A | B) • “Probability of A given B,” where A = “class is cancelled” and B = “it snows.” • Example 2: Probability that it snowed, if class was cancelled • P (B | A) • Bayes' Theorem tells how these two conditional probabilities are related
P (B | A) * P (A) P (A | B) = --------------------------- P (B) likelihood * prior posterior probability = --------------------------- normalizing constant What is Bayes Theorem? (cont.)
What does it mean? • The “prior” is the probability of A as estimated before you knew anything about B; the prior probability of A. • The “likelihood” is the new information you have that will change your estimate of the probability of A; it is the likelihood of B if A were true. • The “normalizing constant” just turns the resulting quantity into a probability (a number between 0 and 1); it is not that interesting to us.
An application: Calculating a probability "A cab was involved in a hit and run accident. There are two cab companies in town, the green (85% of the cabs in the city) and the blue (15%). A witness said that the cab in the accident was blue. Tests showed that the witness is 80% reliable in identifying cabs." Question: What is the probability that the cab in the accident was blue? A = The cab is blue. B = The witness says the cab is blue. P (A) = prior probability that the cab is blue P (B) = overall probability that the witness says the cab is blue P (B | A) = likelihood the witness says the cab is blue when it really is blue P (A | B) = posterior probability the cab is blue given that the witness says it is P(A) = .15 P(B|A) = .8 P(B) = P(B|A)*P(A) + P(B|~A)*P(~A) = .8*.15 + .2*.85 = .29 P (A | B) = [ P(B | A) * P(A) ] / P(B) = .8 * .15 / .29 = .41
What are distributions? • Demonstration: • http://condor.depaul.edu/~dallbrit/extra/psy241/chi-square-demonstration3.xls • Terms • Probability density function: area under the curve = 1 • Normal distribution, Gaussian distribution • Standard deviation • Uniform distribution
Bayes Theorem using distributions • Posterior, prior and likelihood are not single values, but functions over distributions • P(theta | phi) = P(phi | theta) * P(theta) / P(phi) • Posterior dist. = C * likelihood dist. * prior dist. • C is just a normalizing constant to make the posterior distribution sum to 1, and can be ignored here • Because the posterior is a distribution, need a decision rule to use it to choose a value to output
An application to cognitive modeling: perception • Task: guess the distal angle theta that produced the observed proximal angle phi • The viewing angle is unknown, so theta is unknown • p (theta | phi) = posterior; result of combining our prior knowledge and perceptual information • p (phi | theta) = likelihood; probabilities for various values of theta to produce the observed value of phi • p (theta) = prior; probabilities of various values of theta that were known before phi was observed • Gain function = rewards for correct guess, penalties for incorrect guesses • Decision rule = function of posterior and gain function
More information: • http://en.wikipedia.org/wiki/Bayes'_theorem • http://yudkowsky.net/bayes/bayes.html