Probabilistic Models in Human and Machine Intelligence

26
Probabilistic Models in Human and Machine Intelligence

description

Probabilistic Models in Human and Machine Intelligence. A Very Brief History of Cog Sci and AI. 1950’s-1980’s The mind is a von Neumann computer architecture Symbolic models of cognition 1980’s-1990’s The mind is a massively parallel neuron-like networks of simple processors - PowerPoint PPT Presentation

Transcript of Probabilistic Models in Human and Machine Intelligence

Page 1: Probabilistic Models in Human and Machine Intelligence

Probabilistic Models in Human and Machine Intelligence

Page 2: Probabilistic Models in Human and Machine Intelligence

Machine Learning @ CU

Intro courses

CSCI 5622: Machine Learning

CSCI 5352: Network Analysis and Modeling

CSCI 7222: Probabilistic Models

Other courses

cs.colorado.edu/~mozer/Teaching/Machine_Learning_Courses

Page 3: Probabilistic Models in Human and Machine Intelligence

A Very Brief History of Cog Sci and AI 1950’s-1980’s The mind is like a modern digital computer Symbols are basic elements of representation, e.g., John, father Symbol manipulation is basic operation, e.g.,

(father Y X) & (father Z X) -> (sibling Y Z) 1980’s-1990’s The mind is a massively parallel neuron-like networks of simple processors Numerical vectors are basic elements of representation Numerical computing is basic operation:

 y = f ( Σi wi xi) Late 1990’s - ? The mind operates according to laws of probability and statistical inference Invades cog sci, AI (planning, natural language processing), ML Formalizes the statistical intuitions underlying neural nets

Page 4: Probabilistic Models in Human and Machine Intelligence

Relation of Probabilistic Models to Symbolic and Subsymbolic Models

Subsymbolicmodels

Symbolicmodels

Probabilisticmodels

rule learning(small # examples)

structuredrepresentations

statistical learning (large # examples)

feature-vectorrepresentations

Page 5: Probabilistic Models in Human and Machine Intelligence

What Is Probability?

Frequentist notion

Relative frequency obtained if event were observed many times (e.g., coin flip)

Subjective notion

Degree of belief in some hypothesis

Analogous to neural net activation

Long philosophical battle between these two views

Subjective notion makes sense for cog sci and AI given that probabilities represent mental states

Page 6: Probabilistic Models in Human and Machine Intelligence

Do People ReasonAccording To The Laws Of Probability?

The probability of breast cancer is 1% for a woman at 40 who participates in routine screening. If a woman has breast cancer, the probability is 80% that she will have a positive mammography. If a woman does not have breast cancer, the probability is 9.6% that she will also have a positive mammography.

A woman in this age group had a positive mammography in a routine screening? What is the probability that she actually has breast cancer?

A. A. greater than 90%B. between 70% and 90%C. between 50% and 70%D. between 30% and 50%E. between 10% and 30%F. less than 10%

Is this typical or the exception?

Perhaps high-level reasoning isn’t Bayesian but underlying mechanisms of learning, inference, memory, language, and perception are.

95 / 100 doctors

correct answer

Page 7: Probabilistic Models in Human and Machine Intelligence

Griffiths and Tenenbaum (2006)Optimal Predictions in Everyday Cognition

If you were assessing an insurance case for an 18-year-old man, what would you predict for his lifespan?

If you phoned a box office to book tickets and had been on hold for 3 minutes, what would you predict for the total time you would be on hold?

If your friend read you her favorite line of poetry, and told you it was line 5 of a poem, what would you predict for the total length of the poem?

If you opened a book about the history of ancient Egypt to a page listing the reigns of the pharaohs, and noticed that in 4000 BC a particular pharaoh had been ruling for 11 years, what would you predict for the total duration of his reign?

Page 8: Probabilistic Models in Human and Machine Intelligence

Griffiths and Tenenbaum Conclusion

Average responses reveal a “close correspondence between peoples’ implicit probabilistic models and the statistics of the world.”

People show a statistical sophistication and optimality of reasoning generally assumed to be absent in the domain of higher-order cognition.

Page 9: Probabilistic Models in Human and Machine Intelligence

Griffiths and Tenenbaum Bayesian ModelIf an individual has lived for tcur=50 years, how many years ttotal do you expect them to live?

Page 10: Probabilistic Models in Human and Machine Intelligence

What Does Optimality Entail?

Individuals have complete, accurate knowledge about the domain priors.

Fairly sophisticated computation involving Bayesian integral

Page 11: Probabilistic Models in Human and Machine Intelligence
Page 12: Probabilistic Models in Human and Machine Intelligence

From The Economist (1/5/2006)

“[Griffiths and Tenenbuam]…put the idea of a Bayesian brain to a quotidian test. They found that it passed with flying colors.”

“The key to successful Bayesian reasoning is … in having an appropriate prior… With the correct prior, even a single piece of data can be used to make meaningful Bayesian predictions.”

Page 13: Probabilistic Models in Human and Machine Intelligence

My Caution

Bayesian formalism is sufficiently broad that nearly any theory can be cast in Bayesian terms

E.g., adding two numbers as Bayesian inference

Emphasis on how cognition conforms to Bayesian principles often directs attention away from important memory and processing limitations.

Page 14: Probabilistic Models in Human and Machine Intelligence

Value Of Probabilistic Models InCognitive Science

Elegant theories

Statistical inference produces strong constraints on theories

Key claims of theories are explicit

Can minimize assumptions via Bayesian model averagingPrincipled mathematical account

Wasn’t true of symbolic or neural net theories

Currency of probability provides strong constraints(vs. neural net activation)

Page 15: Probabilistic Models in Human and Machine Intelligence

Latent Dirichlet Allocation(a.k.a. Topic Model)

Problem

Given a set of text documents, can we infer the topics that are covered by the set, and can we assign topics to individual documents

Unsupervised learning problem

Technique

Exploit statistical regularities in data

E.g., documents that are on the topic of education will likely contain a set of words such as ‘teacher’, ‘student’, ‘lesson’, etc.

Page 16: Probabilistic Models in Human and Machine Intelligence

Generative Model of Text

Each document is a collection of topics (e.g., education, finance, the arts)

Each topic is characterized by a set of words that are likely to appear

The string of words in a document is generated by:

1) Draw a topic from the probability distribution associated with a document

2) Draw a word from the probability distribution associated with a topic

Bag of words approach

Page 17: Probabilistic Models in Human and Machine Intelligence

Inferring (Learning) Topics

Input: set of unlabeled documentsLearning task

Infer distribution over topics for each document

Infer distribution over words for each topicDistribution over topics can be helpful for classifying or clustering documents

Page 18: Probabilistic Models in Human and Machine Intelligence
Page 19: Probabilistic Models in Human and Machine Intelligence
Page 20: Probabilistic Models in Human and Machine Intelligence

Topic Modeling Of Hotel Reviews

Dan Knights, Rob Lindsey @ JD Powers

Page 21: Probabilistic Models in Human and Machine Intelligence

Phrase Discovery0.17 new york 0.31 shuttle 0.27 non 0.19 minutes0.16 new 0.23 lax 0.14 requested 0.13 waited0.14 ny 0.16 flight 0.14 smoke 0.11 300.14 vegas 0.12 early 0.12 room 0.10 200.12 strip 0.11 sheraton 0.11 given 0.10 150.11 york 0.09 sheraton gateway 0.09 smelled 0.10 450.10 coaster 0.09 proximity 0.08 reserved 0.10 check0.10 nyny 0.09 flights 0.08 change 0.10 min0.08 roller 0.08 catch 0.07 told 0.10 waiting0.08 las 0.08 morning 0.07 cigarette 0.09 arrived0.07 it's 0.07 bus 0.07 assigned 0.09 wait0.07 bars 0.07 pick 0.07 request 0.09 late0.07 las vegas 0.07 shuttles 0.07 called 0.09 100.07 fun 0.07 terminal 0.07 asked 0.08 arrival0.06 drinks 0.06 layover 0.07 reservation 0.08 bell0.06 mgm grand 0.06 international 0.06 advance 0.08 late night0.06 you're 0.06 driver 0.06 resolve 0.08 pm0.06 mgm 0.06 closeness 0.06 cigarette smoke 0.07 luggage0.06 arcade 0.06 minutes 0.05 guaranteed 0.07 took forever0.06 chin 0.06 pickup 0.05 smokers 0.07 told0.06 italian 0.06 drop 0.05 prior 0.06 called0.05 city 0.05 ride 0.05 upgrade 0.06 took care0.05 island 0.05 marriott 0.05 ended 0.06 400.05 skyline 0.05 terminals 0.05 checked 0.06 cleaned0.05 big apple 0.05 convenience 0.05 smell 0.06 checkout0.05 luxor 0.05 to/from 0.05 asking 0.05 took long

Rob Lindsey @ JD Powers

Page 22: Probabilistic Models in Human and Machine Intelligence

Value Of Probabilistic Models In AI and ML

AI and ML fundamentally have to deal with uncertainty in the world, and uncertainty is well described in the language of random events.

Clean, explicit means of incorporating prior knowledge Probability is the optimal thing to compute, in the sense that any other strategy will lead to lower expected returns

e.g., “I bet you $1 that roll of die will produce number < 3. How much are you willing to wager?”

Provides unified framework for re-casting many existing algorithms

Allows you to see interrelationship among algorithms

Allows you to develop new algorithms

Page 23: Probabilistic Models in Human and Machine Intelligence
Page 24: Probabilistic Models in Human and Machine Intelligence

Important Technical IssuesRepresenting structured data

grammars

relational schemas (e.g., paper authors, topics)Hierarchical models

different levels of abstraction

Nonparametric models

flexible models that grow in complexity as the data justifiesInference: Exact vs. approximate

Markov chain Monte Carlo, particle filters, variational approximations

Page 25: Probabilistic Models in Human and Machine Intelligence
Page 26: Probabilistic Models in Human and Machine Intelligence

Rationality in Cognitive ScienceSome theories in cognitive science are based on premise that human performance is optimal

Rational theories, ideal observer theories

Ignores biological constraints

Probably true in some areas of cognition (e.g., vision)More interesting: bounded rationality

Optimality is assumed to be subject to limitations on processing hardware and capacity, representation, experience with the world.