original

24
Kansas State University Department of Computing and Information Sciences 732: Machine Learning and Pattern Recognition Friday, 12 January 2007 William H. Hsu Department of Computing and Information Sciences, KSU http://www.kddresearch.org http:// www.cis.ksu.edu/~bhsu Readings: Chapter 1, Mitchell A Brief Survey of Machine Learning CIS 732/830: Lecture 0 CIS 732/830: Lecture 0

Upload: butest

Post on 18-May-2015

379 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Friday, 12 January 2007

William H. Hsu

Department of Computing and Information Sciences, KSUhttp://www.kddresearch.org

http://www.cis.ksu.edu/~bhsu

Readings:

Chapter 1, Mitchell

A Brief Survey of Machine Learning

CIS 732/830: Lecture 0CIS 732/830: Lecture 0

Page 2: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Lecture OutlineLecture Outline

• Course Information: Format, Exams, Homework, Programming, Grading

• Class Resources: Course Notes, Web/News/Mail, Reserve Texts

• Overview– Topics covered

– Applications

– Why machine learning?

• Brief Tour of Machine Learning– A case study

– A taxonomy of learning

– Intelligent systems engineering: specification of learning problems

• Issues in Machine Learning– Design choices

– The performance element: intelligent systems

• Some Applications of Learning– Database mining, reasoning (inference/decision support), acting

– Industrial usage of intelligent systems

Page 3: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Course Information and AdministriviaCourse Information and Administrivia

• Instructor: William H. Hsu– E-mail: [email protected]

– Office phone: (785) 532-6350 x29; home phone: (785) 539-7180; ICQ 28651394

– Office hours: Wed/Thu/Fri (213 Nichols); by appointment, Tue

• Grading– Assignments (8): 32%, paper reviews (13): 10%, midterm: 20%, final: 25%

– Class participation: 5%

– Lowest homework score and lowest 3 paper summary scores dropped

• Homework– Eight assignments: written (5) and programming (3) – drop lowest

– Late policy: due on Fridays

– Cheating: don’t do it; see class web page for policy

• Project Option– CIS 690/890 project option for graduate students

– Term paper or semester research project; sign up by 09 Feb 2007

Page 4: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Class ResourcesClass Resources

• Web Page (Required)– http://www.kddresearch.org/Courses/Spring-2007/CIS732

– Lecture notes (MS PowerPoint XP, PDF)

– Homeworks (MS PowerPoint XP, PDF)

– Exam and homework solutions (MS PowerPoint XP, PDF)

– Class announcements (students responsibility to follow) and grade postings

• Course Notes at Copy Center (Required)

• Course Web Group– Mirror of announcements from class web page

– Discussions (instructor and other students)

– Dated research announcements (seminars, conferences, calls for papers)

• Mailing List (Automatic)– [email protected]

– Sign-up sheet

– Reminders, related research announcements

Page 5: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Course OverviewCourse Overview

• Learning Algorithms and Models– Models: decision trees, linear threshold units (winnow, weighted majority),

neural networks, Bayesian networks (polytrees, belief networks, influence diagrams, HMMs), genetic algorithms, instance-based (nearest-neighbor)

– Algorithms (e.g., for decision trees): ID3, C4.5, CART, OC1

– Methodologies: supervised, unsupervised, reinforcement; knowledge-guided

• Theory of Learning– Computational learning theory (COLT): complexity, limitations of learning

– Probably Approximately Correct (PAC) learning

– Probabilistic, statistical, information theoretic results

• Multistrategy Learning: Combining Techniques, Knowledge Sources

• Data: Time Series, Very Large Databases (VLDB), Text Corpora

• Applications– Performance element: classification, decision support, planning, control

– Database mining and knowledge discovery in databases (KDD)

– Computer inference: learning to reason

Page 6: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Why Machine Learning?Why Machine Learning?

• New Computational Capability

– Database mining: converting (technical) records into knowledge

– Self-customizing programs: learning news filters, adaptive monitors

– Learning to act: robot planning, control optimization, decision support

– Applications that are hard to program: automated driving, speech recognition

• Better Understanding of Human Learning and Teaching

– Cognitive science: theories of knowledge acquisition (e.g., through practice)

– Performance elements: reasoning (inference) and recommender systems

• Time is Right

– Recent progress in algorithms and theory

– Rapidly growing volume of online data from various sources

– Available computational power

– Growth and interest of learning-based industries (e.g., data mining/KDD)

Page 7: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Rule and Decision Tree LearningRule and Decision Tree Learning

• Example: Rule Acquisition from Historical Data

• Data– Patient 103 (time = 1): Age 23, First-Pregnancy: no, Anemia: no, Diabetes: no,

Previous-Premature-Birth: no, Ultrasound: unknown, Elective C-Section: unknown, Emergency-C-Section: unknown

– Patient 103 (time = 2): Age 23, First-Pregnancy: no, Anemia: no, Diabetes: yes, Previous-Premature-Birth: no, Ultrasound: abnormal, Elective C-Section: no, Emergency-C-Section: unknown

– Patient 103 (time = n): Age 23, First-Pregnancy: no, Anemia: no, Diabetes: no, Previous-Premature-Birth: no, Ultrasound: unknown, Elective C-Section: no,

Emergency-C-Section: YES

• Learned Rule– IF no previous vaginal delivery, AND abnormal 2nd trimester ultrasound,

AND malpresentation at admission, AND no elective C-SectionTHEN probability of emergency C-Section is 0.6

– Training set: 26/41 = 0.634

– Test set: 12/20 = 0.600

Page 8: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Neural Network LearningNeural Network Learning

• Autonomous Learning Vehicle In a Neural Net (ALVINN): Pomerleau et al– http://www.ri.cmu.edu/projects/project_160.html

– Drives 70mph on highways

Page 9: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Relevant DisciplinesRelevant Disciplines

• Artificial Intelligence

• Bayesian Methods

• Cognitive Science

• Computational Complexity Theory

• Control Theory

• Information Theory

• Neuroscience

• Philosophy

• Psychology

• Statistics

MachineLearningSymbolic Representation

Planning/Problem SolvingKnowledge-Guided Learning

Bayes’s TheoremMissing Data Estimators

PAC FormalismMistake Bounds

Language LearningLearning to Reason

OptimizationLearning Predictors

Meta-Learning

Entropy MeasuresMDL Approaches

Optimal Codes

ANN ModelsModular Learning

Occam’s RazorInductive Generalization

Power Law of PracticeHeuristic Learning

Bias/Variance FormalismConfidence IntervalsHypothesis Testing

Page 10: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Specifying A Learning ProblemSpecifying A Learning Problem

• Learning = Improving with Experience at Some Task– Improve over task T,

– with respect to performance measure P,

– based on experience E.

• Example: Learning to Play Checkers– T: play games of checkers

– P: percent of games won in world tournament

– E: opportunity to play against self

• Refining the Problem Specification: Issues– What experience?

– What exactly should be learned?

– How shall it be represented?

– What specific algorithm to learn it?

• Defining the Problem Milieu– Performance element: How shall the results of learning be applied?

– How shall the performance element be evaluated? The learning system?

Page 11: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Example: Learning to Play CheckersExample: Learning to Play Checkers

• Type of Training Experience

– Direct or indirect?

– Teacher or not?

– Knowledge about the game (e.g., openings/endgames)?

• Problem: Is Training Experience Representative (of Performance Goal)?

• Software Design

– Assumptions of the learning system: legal move generator exists

– Software requirements: generator, evaluator(s), parametric target function

• Choosing a Target Function

– ChooseMove: Board Move (action selection function, or policy)

– V: Board R (board evaluation function)

– Ideal target V; approximated target

– Goal of learning process: operational description (approximation) of V

Page 12: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

A Target Function forA Target Function forLearning to Play CheckersLearning to Play Checkers

• Possible Definition– If b is a final board state that is won, then V(b) = 100

– If b is a final board state that is lost, then V(b) = -100

– If b is a final board state that is drawn, then V(b) = 0

– If b is not a final board state in the game, then V(b) = V(b’) where b’ is the best final board state that can be achieved starting from b and playing optimally until the end of the game

– Correct values, but not operational

• Choosing a Representation for the Target Function– Collection of rules?

– Neural network?

– Polynomial function (e.g., linear, quadratic combination) of board features?

– Other?

• A Representation for Learned Function–

– bp/rp = number of black/red pieces; bk/rk = number of black/red kings; bt/rt = number of black/red pieces threatened (can be taken on next turn)

bwbwbwbwbwbww bV 6543210 rtbtrkbkrpbp ˆ

Page 13: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

A Training Procedure for A Training Procedure for Learning to Play CheckersLearning to Play Checkers

• Obtaining Training Examples

– the target function

– the learned function

– the training value

• One Rule For Estimating Training Values:

• Choose Weight Tuning Rule

– Least Mean Square (LMS) weight update rule:

REPEAT

• Select a training example b at random

• Compute the error(b) for this training example

• For each board feature fi, update weight wi as follows:

where c is a small, constant factor

to adjust the learning rate

bV̂

bV

bVtrain

bVbV Successortrainˆ

bVbV berror ˆ train

berrorfcww iii

Page 14: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Design Choices forDesign Choices forLearning to Play CheckersLearning to Play Checkers

Completed Design

Determine Type ofTraining Experience

Gamesagainst experts

Gamesagainst self

Table ofcorrect moves

DetermineTarget Function

Board valueBoard move

Determine Representation ofLearned Function

Polynomial Linear functionof six features

Artificial neuralnetwork

DetermineLearning Algorithm

Gradientdescent

Linearprogramming

Page 15: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Some Issues in Machine LearningSome Issues in Machine Learning

• What Algorithms Can Approximate Functions Well? When?

• How Do Learning System Design Factors Influence Accuracy?

– Number of training examples

– Complexity of hypothesis representation

• How Do Learning Problem Characteristics Influence Accuracy?

– Noisy data

– Multiple data sources

• What Are The Theoretical Limits of Learnability?

• How Can Prior Knowledge of Learner Help?

• What Clues Can We Get From Biological Learning Systems?

• How Can Systems Alter Their Own Representation?

Page 16: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Interesting ApplicationsInteresting Applications

Reasoning (Inference, Decision Support)Cartia ThemeScapes - http://www.cartia.com

6500 news storiesfrom the WWWin 1997

Planning, Control

Normal

Ignited

Engulfed

Destroyed

Extinguished

Fire Alarm

Flooding

DC-ARM - http://www-kbs.ai.uiuc.edu

Database MiningNCSA D2K - http://alg.ncsa.uiuc.edu

Page 17: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Lecture OutlineLecture Outline

• Read: Chapter 2, Mitchell; Section 5.1.2, Buchanan and Wilkins

• Suggested Exercises: 2.2, 2.3, 2.4, 2.6

• Taxonomy of Learning Systems

• Learning from Examples

– (Supervised) concept learning framework

– Simple approach: assumes no noise; illustrates key concepts

• General-to-Specific Ordering over Hypotheses

– Version space: partially-ordered set (poset) formalism

– Candidate elimination algorithm

– Inductive learning

• Choosing New Examples

• Next Week

– The need for inductive bias: 2.7, Mitchell; 2.4.1-2.4.3, Shavlik and Dietterich

– Computational learning theory (COLT): Chapter 7, Mitchell

– PAC learning formalism: 7.2-7.4, Mitchell; 2.4.2, Shavlik and Dietterich

Page 18: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

What to Learn?What to Learn?

• Classification Functions– Learning hidden functions: estimating (“fitting”) parameters

– Concept learning (e.g., chair, face, game)

– Diagnosis, prognosis: medical, risk assessment, fraud, mechanical systems

• Models– Map (for navigation)

– Distribution (query answering, aka QA)

– Language model (e.g., automaton/grammar)

• Skills– Playing games

– Planning

– Reasoning (acquiring representation to use in reasoning)

• Cluster Definitions for Pattern Recognition– Shapes of objects

– Functional or taxonomic definition

• Many Problems Can Be Reduced to Classification

Page 19: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

How to Learn It?How to Learn It?

• Supervised

– What is learned? Classification function; other models

– Inputs and outputs? Learning:

– How is it learned? Presentation of examples to learner (by teacher)

• Unsupervised

– Cluster definition, or vector quantization function (codebook)

– Learning:

– Formation, segmentation, labeling of clusters based on observations, metric

• Reinforcement

– Control policy (function from states of the world to actions)

– Learning:

– (Delayed) feedback of reward values to agent based on actions selected; model

updated based on reward, (partially) observable state

xfxfx, ˆionapproximatexamples

xfx,xdx 21 codebook discretemetric distancensobservatio

as:pnir,si policy1 sequence rdstate/rewa i :

Page 20: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

(Supervised) Concept Learning(Supervised) Concept Learning

• Given: Training Examples <x, f(x)> of Some Unknown Function f

• Find: A Good Approximation to f

• Examples (besides Concept Learning)– Disease diagnosis

• x = properties of patient (medical history, symptoms, lab tests)

• f = disease (or recommended therapy)

– Risk assessment

• x = properties of consumer, policyholder (demographics, accident history)

• f = risk level (expected cost)

– Automatic steering

• x = bitmap picture of road surface in front of vehicle

• f = degrees to turn the steering wheel

– Part-of-speech tagging

– Fraud/intrusion detection

– Web log analysis

– Multisensor integration and prediction

Page 21: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

A Learning ProblemA Learning Problem

UnknownFunction

x1

x2

x3

x4

y = f (x1, x2, x3, x4 )

• xi: ti, y: t, f: (t1 t2 t3 t4) t

• Our learning function: Vector (t1 t2 t3 t4 t) (t1 t2 t3 t4) t

Example x1 x2 x3 x4 y0 0 1 1 0 01 0 0 0 0 02 0 0 1 1 13 1 0 0 1 14 0 1 1 0 05 1 1 0 0 06 0 1 0 1 0

Page 22: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Hypothesis Space:Hypothesis Space:Unrestricted CaseUnrestricted Case

• | A B | = | B | | A |

• |H4 H | = | {0,1} {0,1} {0,1} {0,1} {0,1} | = 224 = 65536 function values

• Complete Ignorance: Is Learning Possible?– Need to see every possible input/output pair

– After 7 examples, still have 29 = 512 possibilities (out of 65536) for fExample x1 x2 x3 x4 y

0 0 0 0 0 ?1 0 0 0 1 ?2 0 0 1 0 03 0 0 1 1 14 0 1 0 0 05 0 1 0 1 06 0 1 1 0 07 0 1 1 1 ?8 1 0 0 0 ?9 1 0 0 1 1

10 1 0 1 0 ?11 1 0 1 1 ?12 1 1 0 0 013 1 1 0 1 ?14 1 1 1 0 ?15 1 1 1 1 ?

Page 23: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Training ExamplesTraining Examplesfor Concept for Concept EnjoySportEnjoySport

Example Sky AirTemp

Humidity Wind Water Forecast EnjoySport

0 Sunny Warm Normal Strong Warm Same Yes1 Sunny Warm High Strong Warm Same Yes2 Rainy Cold High Strong Warm Change No3 Sunny Warm High Strong Cool Change Yes

• Specification for Examples– Similar to a data type definition

– 6 attributes: Sky, Temp, Humidity, Wind, Water, Forecast

– Nominal-valued (symbolic) attributes - enumerative data type

• Binary (Boolean-Valued or H -Valued) Concept

• Supervised Learning Problem: Describe the General Concept

Page 24: original

Kansas State University

Department of Computing and Information SciencesCIS 732: Machine Learning and Pattern Recognition

Representing HypothesesRepresenting Hypotheses

• Many Possible Representations

• Hypothesis h: Conjunction of Constraints on Attributes

• Constraint Values

– Specific value (e.g., Water = Warm)

– Don’t care (e.g., “Water = ?”)

– No value allowed (e.g., “Water = Ø”)

• Example Hypothesis for EnjoySport

– Sky AirTemp Humidity Wind Water Forecast

<Sunny ? ? Strong ? Same>

– Is this consistent with the training examples?

– What are some hypotheses that are consistent with the examples?