90 likes | 104 Views
This lecture explores the learning problem in machine learning, discussing different approaches such as Bayesian learning, decision tree learning, and neural nets. It also covers the reasons why learning is possible and the importance of bias in learning.
E N D
Lecture 26Introduction to Machine Learning CSE 573 Artificial Intelligence I Henry Kautz Fall 2001 CSE 573
Road Map • What is the learning problem? • Why is learning possible? • Approaches to machine learning • Bayesian learning • Decision tree learning • Neural nets • Support vector machines • Nearest neighbor methods • Speed-up learning ... CSE 573
Why Machine Learning • Flood of data WalMart – 25 TerabytesWWW – 1,000 Terabytes • Speed of computation versus slowness of programming highly complex systems (telephone switching systems) = 1 line code @ day @ programmer • Desire for customization a browser that browses by itself? • Sheer ignorance how the heck do you identify gene splice sites? CSE 573
The Learning Problem • Learning = improving with experience at • performing some task • with respect to some performance measure • based on experience chess giving out credit cards CSE 573
Kinds of Learning • Supervised vs Unsupervised • Active vs Passive • Classification vs Action • Empirical vs Analytic • Linear vs Non-linear CSE 573
Terminology – Classification Learning • Instance – described by list of attributes (features) • Target function – some function of the instances we would like to learn • value of chess board • whether or not a credit card holder will default • Concept learning – target is just + or - • Hypothesis space – space of all candidate functions that could be learned – may or may not include the actual target function (if not, is only approximate) • Training set – set of instances labeled with the value of the target function • Test set – labeled data used to measure accuracy of learning CSE 573
Why is Learning Possible? • What can we conclude from: [broccoli, no], [hamburger, yes], [asparagus, no], [cake, yes], [cauliflower, no], [bread, yes] • Experience alone never justifies any conclusion about any unseen instance. • Learning occurs when PREJUDICE meets DATA! CSE 573
Bias • The nice word for prejudice is “bias”. • The world is simple • Occam’s razor “It is needless to do more when less will suffice” – William of Occam, died 1349 of the Black plague • MDL – Minimum description length • Concepts can be approximated by conjunctions of predicates • ... by linear functions • ... by short decision trees • ... by something in the hypothesis space that I choose apriori! CSE 573
Next • Bayesian learning • Decision tree learning • Neural net learning CSE 573