500 likes | 559 Views
Explore the theory and practice of learning in humans, animals, and machines to discover major issues, solutions, and how they work. This course dives into how people, animals, and machines learn, with insights from experts in the field.
E N D
Human, Animal, and Machine Learning Vasile Rus http://www.cs.memphis.edu/~vrus/teaching/cogsci
Overview • General Info about this course/seminar • Why Learning? • The ultimate goal: learning to learn • The crux of intelligent systems that enables them to increase adaptivity to environment and therefore chances of survival
General Information • Web Page: http://ww.cs.memphis.edu/~vrus/teaching/cogsci/ • Check the page as often as possible • It is the main way of getting latest info
General Information • Instructor • Vasile Rus, PhD • Office: 320 Dunn Hall/FIT403c • Phone: x5259 • E-mail: vrus@memphis.edu
What is Learning? • the cognitive process of acquiring skill or knowledge (WordNet) • the process of acquiring knowledge, skills, attitudes, or values, through study, experience, or teaching (Wikipedia)
Machine Learning? • Processes and algorithms that allow computers to "learn“, i.e. improve their knowledge and performance from experience • applications to • search engines • medical diagnosis • bioinformatics and cheminformatics • detecting credit card fraud • and many more: stock market analysis, classifying DNA sequences, speech and handwriting recognition, object recognition in computer vision, game playing and robot locomotion
Goals of this Course • Learn about theory and practice of learning in humans, animals, and machines: • What are the major issues? • What are the major solutions? • How well do they work ? • How do they work ?
Goals of this Course • At the end you should: • Agree that learning is subtle and interesting! • Be able to model a problem as a machine learning problem, run machine learning algorithms on some data to solve the problem
Questions the Course Will Answer • How people learn? • What are the primitive and patterns of learning? • How animals learn? • Can we make machines learn?
Today • Motivation • Course Goals • The Syllabus
Syllabus – Student Session • Week 1: Introduction to Machine Learning • Week 2: The WEKA Machine Learning environment • Week 3: Concept Learning • Week 4: Decision Trees Learning • Week 5: Linear Regression and Perceptrons • Week 6: Hypotheses Spaces and Evaluating Hypotheses • Week 7: Graphical Models: Naïve Bayes, Bayes Nets • Week 8: SPRING BREAK
Syllabus (cont’d) • Week 9: Graphical Models: Hidden Markov Models • Week 10: Graphical Models: LDA • Week 11: Computational Learning Theory • Week 13: Support Vector Machines • Week 14: Instance-based Learning • Week 15: Project Presentations • Week 16: Project Presentations
Syllabus – Plenary Session • Jan 22: NO TALK • Jan 29 • Art Graesser, The University of Memphis: “How Are Theoretical Principles of Learning Incorporated in Intelligent Pedagogical Agents?” • Feb 5 • Razvan Bunescu, Ohio University: “Machine learning approaches to word sense disambiguation and (co)reference resolution” • Feb 12 • Andrew Olney, The University of Memphis: “Building a BrainTrust” • Feb 19 • Giuseppe diFabbrizio, Amazon, Inc.: “Learning to interact - A machine learning approach to dialog management” • Feb 26 • Kim Oller, The University of Memphis: “Evolutionary-Developmental Biology (Evo-Devo) as an influence on current thinking about human and animal language development and evolution” • March 5: NO TALK • Marc 12 • SPRING BREAK
Syllabus – Plenary Session • March 19 • Panayiota Kendeou, University of Minnesota: “The Knowledge Revision Components (KReC) framework: We cannot escape the past but we can reduce its impact” • March 26 • Nobal Niraula, The University of Memphis: “A Machine Learning Approach to Anaphora Resolution in Dialogue based Intelligent Tutoring Systems” • April 2 • Phil Pavlik, The University of Memphis: “Results of Data Mining Student Vocabulary Learning” • April 9 • Dan Stefanescu, The University of Memphis: “Short Text Similarity based on Parsing and Information Content” • April 16 • Michael Johnson, Marquette University: Title Pending • April 23 NO TALK • April 30 NO TALK
To be successful you need to • Read the syllabus • Understand the structure of the seminar • Read the general policies • Attend sessions and participate by asking questions or/and contributing with related remarks • Explore the seminar website • Don't limit yourself to what is asked in the seminar
Grading • Assignments • 4-5 (or more) • 35% of final grade • Project • 40% • Quizzes: 20% • Participation and Presentation: 5%
Grading 2.5 above or below the cut-off will earn you a + or – in front of your grade. For example: 89 has a letter equivalent of B+ Exception: 90-91 will give you A-, 92 to 96 will give you A, anything above 97 means A+.
Other Issues • Attendance can help you when on borderline • General announcements are posted on the web site frequently! • Please check it out as often as possible • If you notice any inconsistencies on the website (broken links, misspellings, etc.) please notify me • Thank you!
Bibliography • Tom Mitchell: Machine Learning, McGraw Hill, 1997, ISBN 0070428077. • RECOMMENDED • Ian H. Witten and Eibe Frank: Data Mining:Practical Machine Learning Tools and Techniques, Morgan Kaufmann, 2005, ISBN 0-12-088407-0. • Any graduate textbook on learning
Office Hours and Extra Help • During the following times I'll be available in my office • Mondays: 11:00AM-12:00PM • Wednesdays: 1:20-2:20 PM • By appointment • You must send me an email to set up an appointment • If you just knock on my door without notice the chances are that I'll be busy • Please use the office hours!
Assignment Submission • Submissions: • You will have on average one-two weeks from the date the work is assigned • Late submissions are not accepted • In exceptional cases you may have a 48-hour grace period at the cost of 50% of the grade (you should ask for it before the due date) • Should be submitted electronically AND on paper
Plagiarism • Plagiarism is not tolerated. If caught, you'll be given grade 0 (zero) and disciplinary actions will be taken • It's OK to help some of your friends who may have problems • This is actually a good learning tool • but it is not OK to share answers. If they need, help/discuss with them but never show them your solution • I may (and I will) ask you to demonstrate and explain your solutions
Project • Preferably an interdisciplinary team • A common project OR • Something of your choice
Machine Learning • The study of automated processes, algorithms, and systems that learn from experience: improve their knowledge and performance
A Typical Learning Task • Learning the Sound ‘R’
Learning Sound ‘R’ • at least 32 different R sounds to consider as separate distinct sounds • http://mommyspeechtherapy.com/?p=1116 • Vocalic R: • R can be vowel-like too • depending on the location of the R relative to a vowel, the R will change pronunciation • In words like car, fear, for, the R sound comes after the vowels; each vowel is pronounced differently and so is the R • the R takes on the characteristic of the vowel depending on context and combination. The six different vocalic combinations, [ar, air, ear, er, or, ire], are collectively called vocalic R, r-controlled vowels, or vowel R • How do children learn to pronounce R in English?
Machine Learning • Standard process but no standard algorithm • Machine Learning task • Learning from experience E with respect to some class of tasks T and performance measure P • Learning is successful if P increases after learning from E
Checkers Playing • Task T: play checkers • Performance P: percent of games won against opponents • Training experience E: playing practice games against itself
Examples of Learning Tasks • Recognize spoken words • Sphinx system (Lee, 1989) learns speaker specific strategies for recognizing phonemes and words • Neural networks • Hidden Markov models • Learning to drive an autonomous vehicle • Many methods
Examples of Learning Tasks • Playing backgammon • Very competitive • Reinforcement learning
ML Process • Specify: T, P, E • Specify: the type of knowledge to be learned • Specify: Representation of the target knowledge • Specify: Learning mechanism/algorithm
Checkers: ML Process – Step #1 • Task T: learn how to play checkers • Performance P: #wins in world tournament • Experience E: see next slide
Checkers: ML Process – Step #1 • Step #1: Choose training experience • Type of feedback: direct or indirect • If indirect feedback is available then there is an issue of credit assignment • In our checkers playing problem, we only know the final result • Control over training experience • Does the learner control the training examples or does a teacher provide them? • Distribution of training examples • Ideally, training examples should have same distribution as testing, future examples
Checkers: ML Process – Step #2 • Choose type of knowledge • Target function • Target function: • Find a best search strategy in the space of legal moves that yields best move sequences • ChooseMove : B → M, maps a legal state of the board to a move • V : B → R, maps a legal board state to a real value or score (higher values mean better states)
Checkers: ML Process – Step #2 • Value of target function V(b) for a board state b is • V(b)=100, if a final board state is a win • V(b)=-100, if a final board state is a loss • V(b)=0, if a final board state is drawn • V(b)=V(b’) where b’ is the best final board state starting at b and playing optimally • Even V(b) is hard to compute and so an operation description of V is needed • Operational description of ideal function V is an approximation of it
Checkers: ML Process – Step #4 • We need an algorithm for finding weights of a linear function that minimizes E • Incrementally refine weights as new training examples become available • Robust to errors in the training estimated values • LMS algorithm
4 Modules of Learning Systems • Performance System • Applies the learned function • Critic • Generates training examples • Generalizer • The learning algorithm • Experiment Generator • Generates problem instances/samples
Key Issues in ML • What algorithms can learn functions from examples and how well can they do it? • Which algorithms perform well for what types of problems and representations? • How does noise impact learning? • How much training data is sufficient? • Mantra#2: the more data the better but … more data many times is not possible or is expensive • How can prior knowledge be used in learning?
Key Issues in ML • What are the limits of learnability? • Under what conditions is successful learning possible, less possible, or impossible? • Nature of learning problems • Under what conditions is a particular learning algorithm assured of learning successfully? • Nature of learning algorithms • How can learners alter their representations to improve?
Gist of ML • Experts do their job and many times it is hard to make them articulate a procedure/function that captures their expertise • Basic idea: • Have a mechanism that can learn from examples and hope for the best • Improve your knowledge from more examples
Summary • Intro to Machine Learning
Next Time • Intro to Weka • Classification and clustering