Click here to load reader
View
34
Download
0
Embed Size (px)
DESCRIPTION
The Discipline and Future of Machine Learning. Tom M. Mitchell E. Fredkin Professor and Department Head March 2007. The Discipline of Machine Learning. The defining question: - PowerPoint PPT Presentation
Tom M. MitchellE. Fredkin Professor and Department HeadMarch 2007 The Discipline and Futureof Machine Learning
The Discipline of Machine Learning
The defining question: How can we build computer systems that automatically improve with experience, and what are the fundamental laws that govern all learning processes?
A process learns with respect to if it Improves its performance Pat task Tthrough experience E
Machine Learning - PracticeSpeech RecognitionControl learning Reinforcement learning Supervised learning Bayesian networks Hidden Markov models Unsupervised clustering Explanation-based learning ....Extracting facts from text
Machine Learning - TheoryPAC Learning Theory# examples (m)representational complexity (H)error rate (e)failure probability (d)Other theories for Reinforcement skill learning Semi-supervised learning Active student querying also relating: # of mistakes during learning convergence rate asymptotic performance bias, variance VC dimension(for supervised concept learning)
The Discipline of Machine LearningMachine Learning: How can we build computer systems that automatically improve with experience, and what are the fundamental laws that govern all learning processes?
Computer Science:How can we build machines that solve problems, and which problems are inherently tractable/intractable?
Statistics:What can be learned from data with a set of modeling assumptions, while taking into account the data-collection process?
Animal learning (Cognitive science, Psychology, Neuroscience)Machine learningStatistics Computer scienceAdaptive Control TheoryandRoboticsEvolutionEconomics
ML and CSMachine learning already the preferred approach toSpeech recognition, Natural language processingComputer visionMedical outcomes analysisMany robot control problems
The ML niche will growWhy?All softwareML software
ML and Empirical SciencesEmpirical science is a learning process, subject to automation and to studyimprove performance P (accuracy)at task T (predict which gene knockouts will impact the aromatic AA pathway, and how)with experience E (active experimentation)
Functional genomic hypothesis generation and experimentation by a robot scientist, King et al., Nature, 427(6971), 247-252Which protein ORFs influence which enzymes in the AAA pathway
Our current state:The problem of tabula-rasa function approximation is solved (in an 80-20 sense): Given: Class of hypotheses H = {h: X Y}Labeled examples {}Determine:The h from H that best approximates f
Its time to move onEnrich the function approx problem definitionUse function approx as building blockWork on new problems
Some Current Research QuestionsWhen/how can unlabeled data be useful in function approximation?
How can assumed sparsity of relevant features be exploited in high dimensional nonparametric learning?
How can information learned from one task be transferred to simplify learning another?
What algorithms can learn control strategies from delayed rewards and other inputs?
What are the best active learning strategies for different learning problems?
To what degree can one preserve data privacy while obtaining the benefits of data mining?
The Future of Machine Learning
A Quick Look Back19601970198019902000Samuels checker learnerPerceptronsWinstons symbolic concept learnerRule learningDecision tree learningNeural networksExplanation-based learningDimensionality reductionBayes netsPAC learning theoryArchitectures for learning and problem solvingReinforcement learningSemi-supervised learningNon-parametric methodsStatistical perspective on learningHMMsSVMsTheories of grammar inductionLarge scale dataminingSpeech applicationsRobot controlPrivacy preserving data miningTransfer learningVersion SpacesTheories of perceptron capacity and learnabilityEvolutionary and revolutionary changesWhat might lead to the next revolution?
Use Machine Learning to help understand Human Learning(and vice versa)
Models of Learning Processes# of examplesError rateReinforcement learningExplanations
Learning from examplesComplexity of learners representationProbability of successPrior probabilitiesLoss functions
# of examplesError rateReinforcement learningExplanations
Human supervisionLecturesQuestions, HomeworksAttention, motivationSkills vs. PrinciplesImplicit vs. Explicit learningMemory, retention, forgettingHebbian learning, consolidation Machine Learning:Human Learning:
Reinforcement Learning[Sutton and Barto 1981; Samuel 1957]Observed immediate rewardLearned sum of future rewards
Reinforcement Learning in MLr =100V=1000V=72V=81V=90 = .9S0S2S1S3
Reinforcement Learning in MLVariants of RL have been used for a variety of practical control learning problems Temporal Difference learningQ learning Learning MDPs, POMDPs
Theoretical results tooAssured convergence to optimal V(s) under certain conditionsAssured convergence for Q(s,a) under certain conditions
Dopamine As Reward Signal[Schultz et al., Science, 1997]t
Dopamine As Reward Signal[Schultz et al., Science, 1997]t
Dopamine As Reward Signal[Schultz et al., Science, 1997]t
RL Models for Human Learning[Seymore et al., Nature 2004]
[Seymore et al., Nature 2004]
Human and Machine LearningAdditional overlaps:
Learning of perceptual representationsDimensionality reduction methods, low level perceptsLewicky et al.: optimal sparse codes of natural scenes yield gabor filters found in primate visual cortex. Similar result for auditory cortex.
Learning with redundant sensory inputCoTraining methods, Sensory redundancy hypothesis in developmentDe Sa & Ballard; Coen: co-clustering voice/video yields phonemesMitchell & Perfetti: co-training in second language learning
Learning and explanationsExplanation-based learning, teaching concepts & skills, chunkingVanLehn et al: explanation-based learning accounts for some human learning behaviors.Chi: students learn best when forced to explainNewell; Anderson: chunking/knowledge-compilation models
2. Never-ending learning
Never-Ending LearningCurrent machine learning systems: Learn one functionAre shut down after they learn itStart from scratch when programmed to learn the next function
Lets study and construct learning processes that:Learn many different thingsFormulate their own next learning taskUse what they have already learned to help learn the next thing
Example: Never-ending learning robotImagine a robot with three goals: (1) avoid collisions, (2) recharge when battery low, and (3) find and collect trash
What is stopping us from giving it some trash examples, then letting it learn for a year?
What must it start with to formulate and solve relevant learning subtasks?Learn to recognize trash in sceneLearn where to search for trash, and whenLearn how close to get to find out whether trash is thereLearn to manipulate trashTransfer what it learned about paper trash to help with bottle trashDiscover relevant subcategories of trash (e.g., plastic versus glass bottles), and of other objects in the environment
Core Questions for Never-Ending Learning AgentWhat function or fact to learn next?Self-reflection on performance, credit assignment
What representation for this target function or fact?Choice of input-output representation for target functionE.g., classify whether its trash
How to obtain (which type of) training experience?Primarily self-supervised, but occasional teacher inputE.g., classify whether its trash
Guided by what prior knowledge?Transfer learning, but transfer between what?XPaperTrash help learn XPlasticTrash ?State(t) x Action(t) State(t+1) help learn XPlasticTrash ?
Example: Never-ending language learnerRead the Web project: Create 24x7 web agent that each day:Extracts more facts from the web into structured databaseLearns to extract facts better than yesterday
Starting point:Ontology of hundreds of categories and relationsand 6-10 training examples of each Never-ending learning architectureState of art language processing primitivesLearning mechanismsTop level task:Populate a database of these categories and relations by reading the web, and improve continually[Carlson, Cohen, Fahlman, Hong, Nyberg, Wang, ...]
Q: how can it obtain useful training experience (i.e., self-supervise)?A: redundancy
Bootstrapping: Learning to extract named entitiesI arrived in Pittsburgh on Saturday.location?x1: I arrived in _________ on Saturday.x2: Pittsburgh
Bootstrap learning to extract named entities[Riloff and Jones, 1999], [Collins and Singer, 1999], ...
Co-TrainingAnswer1Classifier1Answer2 Classifier2I flew to New York today.New YorkI flew to ____ todayIdea: Train Classifier1 and Classifier2 to:1. Correctly classify labeled examples2. Agree on classification of unlabeled
Co-Training Theory [Blum&Mitchell 98; Dasgupta 04, ...]Final Accuracy# unlabeled examplesConditional dependence among inputs# labeled examplesNumber of redundant inputs want inputs less dependent, increased number of redundant inputs, disagreement over unlabeled examples can bound true error
Example Bootstrap learning algorithms:
Classifying web pages [Blum&Mitchell 98; Slattery 99]Classifying email [Kiritchenko&Matwin 01; Chan et al. 04]Named entity extraction [Collins&Singer 99; Jones 05]Wrapper induction [Muslea et al., 01; Mohapatra et al. 04]Word sense disambiguation [Yarowsky 96]Discovering new word senses [Pant