Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267...

59
Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin
  • date post

    22-Dec-2015
  • Category

    Documents

  • view

    215
  • download

    0

Transcript of Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267...

Page 1: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Chapter 12Machine Learning

ID: 116 117Name: Qun Yu (page 1-33) Kai Zhu (page 34-59)Class: CS267 Fall 2008Instructor: Dr. T.Y.Lin

Page 2: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Introduction

Machine Learning is a key part of A.I. research.

Rough Set Theory can be used for some problems in Machine Learning.

Will discuss 2 cases:1) Learning from Examples2) An Imperfect Teacher

Page 3: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Case1:Learning from Examples

We assume: 2 agents: a knower, a learner Knower’s Knowledge set U U is unchanged, will NOT increase during the

learning process. It is called CWA(Closed World Assumption)

Knower knows everything about U Learner has ability to learn U, in other words,

learner knows some attributes of objects in U

Page 4: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Example For Case 1 KR-System

U a b c d e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

4 0 0 1 2 1

5 2 1 0 2 1

6 0 0 1 2 2

7 2 0 0 1 0

8 0 1 2 2 1

9 2 1 0 2 2

10 2 0 0 1 0

Attributes of learner’s knowledge : B={a,b,c,d}

Attributes of knower’s knowledge : e

Page 5: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Example For Case 1 KR-System

U a b c d e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

4 0 0 1 2 1

5 22 11 00 22 1

6 0 0 1 2 2

7 2 0 0 1 0

8 0 1 2 2 1

9 22 11 00 22 2

10 2 0 0 1 0

3 concepts of knower’s knowledge:

X0={3,7,10}

X1={1,2,4,5,8}

X2={6,9}

5 concepts of learner’s knowledge:

Y0={1,2}

Y1={3,7,10}

Y2={4,6}

YY3={5,9}={5,9}

Y4={8}

Page 6: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Which Objects are learnable?X0={3,7,10}

X1={1,2,4,5,8}

X2={6,9}

Y0={1,2}

Y1={3,7,10}

Y2={4,6}

Y3={5,9}

Y4={8}

BX0 = Y1 = {3,7,10} = X0 = BX0

X0 is exactly Y-definable and can be learned fully.

BX1 = Y0 Y4 = {1,2,8}

BX1 = Y0 Y2 Y3 Y4 = {1,2,4,5,6,8,9}

X1 is roughly Y-definable, so learner can learn object {1,2,8}, not sure about {4,5,6,9}.

BX2 = Ø

BX1 = Y2 Y3 = {4,5,6,9}

X2 is internally Y-indefinable, so it is NOT learnable.

Learnable:Learnable: The knower’s knowledge can be expressed in terms of learner’s knowledge.

Page 7: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Quality of Learning?BX0 = Y1 = {3,7,10} = X0 = BX0

BX1 = Y0 Y4 = {1,2,8}

BX1 = Y0 Y2 Y3 Y4 = {1,2,4,5,6,8,9}

BX2 = Ø

BX1 = Y2 Y3 = {4,5,6,9}

X0 X1 X2

Positive objects 3,7,10 1,2,8 Ø

Border-line objects Ø 3,7,10 4,5,6,9

Negative objects 1,2,4,5,6,8,9 4,5,6,9 1,2,3,7,8,10

POSB{e}=POSB(X0) POSB(X1)

POSB(X2) = {1,2,3,7,8,10} (6 objects)

U={1,2,3,4,5,6,7,8,9,10} (10 objects)

{e}= 6/10 = 0.6

Page 8: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Simplifier table…step 1U a b c d e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

4 0 0 1 2 1

5 22 11 00 22 1

6 0 0 1 2 2

7 2 0 0 1 0

8 0 1 2 2 1

9 22 11 00 22 2

10 2 0 0 1 0

Group the duplicate rows :

U a b c d e

3,7,10

2 0 0 1 0

1,2 1 2 0 1 1

4 0 0 1 2 1

5 22 11 00 22 1

8 0 1 2 2 1

6 0 0 1 2 2

9 22 11 00 22 2

Page 9: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Simplifier table…step 2Remove inconsistent rows :

U a b c d e

3,7,10

2 0 0 1 0

1,2 1 2 0 1 1

8 0 1 2 2 1

U a b c d e

3,7,10

2 0 0 1 0

1,2 1 2 0 1 1

4 0 0 1 2 1

5 22 11 00 22 1

8 0 1 2 2 1

6 0 0 1 2 2

9 22 11 00 22 2

Page 10: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Simplifier table…step 3

{a}, {b}

Find attributes reduct :U a b c d e

3,7,10

2 0 0 1 0

1,2 1 2 0 1 1

8 0 1 2 2 1

U b c d e

3,7,10

0 0 1 0

1,2 2 0 1 1

8 1 2 2 1U a c d e

3,7,10

2 0 1 0

1,2 1 0 1 1

8 0 2 2 1

U a b d e

3,7,10

2 0 1 0

1,2 1 2 1 1

8 0 1 2 1U a b c e

3,7,10

2 0 0 0

1,2 1 2 0 1

8 0 1 2 1

Delete a:

Delete b:

Delete c:

Delete d:

Page 11: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Simplifier table…step 4Find value reduct & decision rules:

U a e

3,7,10

2 0

1,2 1 1

8 0 1

U b e

3,7,10

0 0

1,2 2 1

8 1 1a2e0

a1e1

a0e1

a2e0

a1Va0e1

b0e0

b2e1

b1e1

b0e0

b2Vb1e1

Page 12: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

All objects are necessary in learning process?

From the KR-System table, we can see some objects are same by attributes {a,b,c,d,e}:

1 = 2

3 = 7 =10

D1={1,2}

D2={3,7,10}

Therefore, only one object, either 1 or 2 in D1 is necessary in learning process. So is D2.

4,5,6,8,9 are necessary.

U a b c d e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

4 0 0 1 2 1

5 2 1 0 2 1

6 0 0 1 2 2

7 2 0 0 1 0

8 0 1 2 2 1

9 2 1 0 2 2

10 2 0 0 1 0

Page 13: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Prove from Decision Rule…(1)

U a b c d e

1 1 2 0 1 1

3 2 0 0 1 0

4 0 0 1 2 1

5 22 11 00 22 1

6 0 0 1 2 2

8 0 1 2 2 1

9 22 11 00 22 2

Remove objects 2,7 and 10 (keep 1 and 3) as Table1,Table1 will provide the same decision rules.

Table1

a2 e0

a1 e1

a0 e1

b0 e0

b2 e1

b1 e1

OR

Page 14: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Prove from Decision Rule…(2)

U a b c d e

2 1 2 0 1 1

4 0 0 1 2 1

5 22 11 00 22 1

6 0 0 1 2 2

8 0 1 2 2 1

9 22 11 00 22 2

10 2 0 0 1 0

Remove objects 1,3 and 7 (keep 2 and 10)as Table 2,Table2 will provide the same decision rules.

Table 2

a2 e0

a1 e1

a0 e1

b0 e0

b2 e1

b1 e1

OR

Page 15: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Prove from Decision Rule…(3)

U a b c d e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

5 22 11 00 22 1

6 0 0 1 2 2

7 2 0 0 1 0

9 22 11 00 22 2

10 2 0 0 1 0

Remove objects 4 and 8 as Table3,

The whole new decision algorithm will be:

Table 3

a2 e0

a1 e1

a0 e2

Now concept X2 Now is internally definable. Object 6 is learnable now. So decision algorithm changed.

b2c0 e1

b0c0 e0

b0c1 e2

b2d1 e1

b0d1 e0

b0d2 e2

OR OR

Page 16: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Prove from Decision Rule…(4)

U a b c d e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

4 0 0 1 2 1

5 22 11 00 22 1

6 0 0 1 2 2

7 2 0 0 1 0

8 0 1 2 2 1

10 2 0 0 1 0

Table 4

Remove objects 9 as Table 4,

The whole new decision algorithm will be:

b0 e0

b2 e1

a0b1 e1

a2b1 e1Now object 5 is positive object of X1. So decision algorithm changed.

a2d1 e0

a1d1 e1

a2d2 e1

a0d2 e1

OR

Page 17: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Case2: An Imperfect TeacherWe assume:

2 agents: a knower, a learner

Knower’s Knowledge set U

U is unchanged, will NOT increase during the learning process. It is called CWA(Closed World Assumption)

Knower knows everything about U

Learner has ability to learn U, in other words, learner knows some attributes of objects in U

Page 18: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Example 1 For Case 2 KR-System

U a b c

1 0 2 +

2 0 1 +

3 1 0 +

4 0 1 0

5 1 0 0

6 1 1 -

7 2 1 -

8 0 1 -

9 1 0 -

Attributes of learner’s knowledge : B={a,b}

Attributes of knower’s knowledge : c

Attribute value 0: knower can’t classify this object.

Page 19: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Example 1 For Case 2KR-System

3 concepts of knower’s knowledge:

X0={4,5} ignorance region

X+={1,2,3}

X-={6,7,8,9}

X*=X+ X- competence region5 concepts of learner’s knowledge:

Y0={1}

Y1={2,4,8}

Y2={3,5,9}

YY3={6}={6}

Y4={7}

U a b c

1 0 2 +

2 0 1 +

3 1 0 +

4 0 1 0

5 1 0 0

6 11 11 -

7 2 1 -

8 0 1 -

9 1 0 -

Page 20: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Can the learner be able to discover the ignorance region?

BX+ = Y0= {1}

BX+ = Y0 Y1 Y2 = {1,2,3,4,5,8,9}

X+ is roughly Y-definable, so learner can learn object {1}, not sure about {2,3,4,5,8,9}.

BX- = Y3 Y4 = {6,7}

BX- = Y1 Y2 Y3 Y4 = {2,3,4,5,6,7,8,9}

X- is roughly Y-definable, so learner can learn object {6,7}, not sure about {2,3,4,5,8,9}.

BX0 = Ø

BX0 = Y1 Y2 = {2,3,4,5,8,9}

X0 is internally Y-indefinable, so it is NOT learnable.

X0={4,5} ignorance region

X+={1,2,3}

X-={6,7,8,9}

X*=X+ X- competence region

Y0={1}

Y1={2,4,8}

Y2={3,5,9}

Y3={6}

Y4={7} Learner can’t discover the ignorance region X0

Page 21: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

The ignorance region influences the learner’s ability to learn?

In this example, the answer is NO, because the ignorance region(X0) is not learnable.

Hence, we can prove it:

BNB(X*)= BNB(X+) BNB(X-)

BX+ = Y0= {1}

BX+ = Y0 Y1 Y2 = {1,2,3,4,5,8,9}

BX- = Y3 Y4 = {6,7}

BX- = Y1 Y2 Y3 Y4 = {2,3,4,5,6,7,8,9}

X*=X+ X- competence region

X0={4,5} ignorance regionX* X*

put 4 into X+

X*put 4 into X-

Positive objects

1,6,7 1,6,7 1,6,7

Border-line objects

2,3,4,5,8,9 2,3,4,5,8,9 2,3,4,5,8,9

Negative objects

Ø Ø Ø

Border-line region of the competence region X* remain unchanged, therefore, it doesn't matter knower knows it or NOT.

Page 22: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Put 4 into X+

X+={1,2,3,4}

X-={6,7,8,9}

X0={5}

BX+ = Y0= {1}

BX+ = Y0 Y1 Y2 = {1,2,3,4,5,8,92,3,4,5,8,9}

BX- = Y3 Y4 = {6,7}

BX- = Y1 Y2 Y3 Y4 = {2,3,4,52,3,4,5,6,7,8,98,9}

BNB(X*)= BNB(X+) BNB(X-)

={2,3,4,5,8,9} {2,3,4,5,8,9}

={2,3,4,5,8,9}

X0={4,5} ignorance

region

X+={1,2,3}

X-={6,7,8,9}

X*=X+ X- competence

region

Y0={1}

Y1={2,4,8}

Y2={3,5,9}

Y3={6}

Y4={7}

Calculation Sheet:Border-line objects

Page 23: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Simplifier table…step 1Remove inconsistent rows :

U a b c

1 0 2 +

6 1 1 -

7 2 1 -

U a b c

1 0 2 +

2 0 1 +

3 1 0 +

4 0 1 0

5 1 0 0

6 11 11 -

7 2 1 -

8 0 1 -

9 1 0 -

Learner can’t discover the ignorance region {4,5} and the ignorance region won’t affect the learning process.

Page 24: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Simplifier table…step 2

{a}, {b}

Find attributes reduct :

Delete a:

Delete b:

U a b c

1 0 2 +

6 1 1 -

7 2 1 -U a c

1 0 +

6 1 -

7 2 -

U b c

1 2 +

6 1 -

7 1 -

Page 25: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Simplifier table…step 3Find value reduct & decision rules:

a0c+

a1c-

a2c-

b2c+

b1c-

U a c

1 0 +

6 1 -

7 2 -

U b c

1 2 +

6 1 -

7 1 -

Page 26: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Example 2 For Case 2 KR-System

U a b c

1 0 2 +

2 0 1 +

3 1 0 +

4 0 1 +

5 1 0 0

6 1 1 0

7 2 1 -

8 0 1 -

9 1 0 -

Attributes of learner’s knowledge : B={a,b}

Attributes of knower’s knowledge : c

Attribute value 0: knower can’t classify this object.

Page 27: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Example 2 For Case 2KR-System

3 concepts of knower’s knowledge:

X0={5,6} ignorance region

X+={1,2,3,4}

X-={7,8,9}

X*=X+ X- competence region5 concepts of learner’s knowledge:

Y0={1}

Y1={2,4,8}

Y2={3,5,9}

YY3={6}={6}

Y4={7}

U a b c

1 0 2 +

2 0 1 +

3 1 0 +

4 0 1 +

5 1 0 0

6 11 11 0

7 2 1 -

8 0 1 -

9 1 0 -

Page 28: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Can the learner be able to discover the ignorance region?

BX+ = Y0= {1}

BX+ = Y0 Y1 Y2 = {1,2,3,4,5,8,9}

X+ is roughly Y-definable, so learner can learn object {1}, not sure about {2,3,4,5,8,9}.

BX- = Y4 = {7}

BX- = Y1 Y2 Y4 = {2,3,4,5,7,8,9}

X- is roughly Y-definable, so learner can learn object {7}, not sure about {2,3,4,5,8,9}.

BX0 = Y3= {6}

BX0 = Y2 Y3 = {3,5,6,9}

X0 is roughly Y-definable, so learner can learn object {6}, not sure about {3,5,9}.

X+={1,2,3,4}

X-={7,8,9}

X*=X+ X- competence region

X0={5,6} ignorance region

Y0={1}

Y1={2,4,8}

Y2={3,5,9}

Y3={6}

Y4={7} Learner can discover the ignorance region X0

Page 29: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

The ignorance region influences the learner’s ability to learn?

In this example, the answer is YES, because the ignorance region(X0) is roughly learnable, object 6 is important. Hence, we can prove it:

BX+ = Y0= {1}

BX+ = Y0 Y1 Y2 = {1,2,3,4,5,8,9}

BX- = Y4 = {7}

BX- = Y1 Y2 Y4 = {2,3,4,5,7,8,9}

X*=X+ X- competence region

X0={5,6} ignorance region

X* X* put 6 into X+

X*put 5 into X-

Positive objects

1, 7 1,6,7 1, 7

Border-line objects

2,3,4,5,8,9 2,3,4,5,8,9 2,3,4,5,8,9

Negative objects

6 Ø 6

Object 6 will fall in positive region if knower knows it (move from X0 to X+).

Therefore, X0 affects the learning process.BNB(X*)= BNB(X+) BNB(X-)

Page 30: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Put 6 into X+

X+={1,2,3,4,6}

X-={7,8,9}

X0={5}

BX+ = Y0= {1,6}

BX+ = Y0 Y1 Y2 Y3 = {1,2,3,4,5,2,3,4,5,6,8,98,9}

BX- = Y4 = {7}

BX- = Y1 Y2 Y4 = {2,3,4,52,3,4,5,7,8,98,9}

POSB(X*)= POSB(X+) POSB(X-)

={1,6} {7}

={1,6,7}

X0={5,6} ignorance

region

X+={1,2,3,4}

X-={7,8,9}

X*=X+ X- competence

region

Y0={1}

Y1={2,4,8}

Y2={3,5,9}

Y3={6}

Y4={7}

Calculation Sheet:Positive objects

Page 31: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Simplifier table…step 1Remove inconsistent rows :

U a b c

1 0 2 +

6 1 1 0

7 2 1 -

Object 6 is learnable, therefore learner can discover the ignorance region {4,5} and the ignorance region will affect the learning process.

U a b c

1 0 2 +

2 0 1 +

3 1 0 +

4 0 1 +

5 1 0 0

6 11 11 0

7 2 1 -

8 0 1 -

9 1 0 -

Page 32: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Simplifier table…step 2

{a}

Find attributes reduct :

Delete a:

Delete b:

U a b c

1 0 2 +

6 1 1 0

7 2 1 -U a c

1 0 +

6 1 0

7 2 -

U b c

1 2 +

6 1 0

7 1 -

A is indispensable

Page 33: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Simplifier table…step 3Find value reduct & decision rules:

a0c+

a1c0

a2c-

U a c

1 0 +

6 1 0

7 2 -

Object 6 is learnable, so learner can discover the knower’s ignorance.

Page 34: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Chapter 12 Machine LearningChapter 12 Machine LearningInductive LearningInductive Learning

Presented by: Kai ZhuPresented by: Kai Zhu

Professor: Dr. T.Y. LinProfessor: Dr. T.Y. Lin

Class ID: 117Class ID: 117

Page 35: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

In the pervious chapters, we assumed that the set of In the pervious chapters, we assumed that the set of instances U is constant and unchanged during the instances U is constant and unchanged during the

learning process. learning process.

In any real life situations however this is not the In any real life situations however this is not the case and new instances can be added to the set U. case and new instances can be added to the set U.

Page 36: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Example Example

Lets consider the K-R system Lets consider the K-R system given in Example 1 in the given in Example 1 in the

pervious section in chapter 12pervious section in chapter 12

Page 37: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Table 9Table 9

U a b c d e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

4 0 0 1 2 1

5 2 1 0 2 1

6 0 0 1 2 2

7 2 0 0 1 0

8 0 1 2 2 1

9 2 1 0 2 2

10 2 0 0 1 0

Page 38: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Consistent table:Consistent table:

U a b c d e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

7 2 0 0 1 0

8 0 1 2 2 1

10 2 0 0 1 0

Inconsistent table:Inconsistent table:U a b c d e

4 0 0 1 2 1

5 2 1 0 2 1

6 0 0 1 2 2

9 2 1 0 2 2

(e)=card(1,2,3,7,8,10)/card(1,2,3,4,5,6,7,8,9,10)=6/10=0.6

Page 39: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

After remove duplicate After remove duplicate and inconsistentand inconsistent

U a b c d e

1 1 2 0 1 1

2 2 0 0 1 0

3 0 1 2 2 1

Page 40: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

After computing, get core After computing, get core and reduct values tableand reduct values table

U a b c d e

1 - - - - 1

2 - - - - 0

3 - - - - 1

U a b c d e

1(1) 1 x x x 1

1(2) x 2 x x 1

2(1) 2 x x x 0

2(2) x 0 x x 0

3(1) 0 x x x 1

3(2) x 1 x x 1

3(3) x x 2 x 1

3(4) x x x 2 1

Core tableCore table

Reduct tableReduct table

Page 41: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

From reduct table, there are 16 From reduct table, there are 16 Simplified Tables. One of them : Simplified Tables. One of them :

corresponding decision corresponding decision algorithms:algorithms:

U a b c d e

1(1) 1 x x x 1

2(1) 2 x x x 0

3(1) 0 x x x 1

a1a1 e1 e1a2 a2 e0e0a0 a0 e1e1

Page 42: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Table 10Table 10

U a b c d

e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

4 0 0 1 2 1

5 2 1 0 2 1

6 0 0 1 2 2

7 2 0 0 1 0

8 0 1 2 2 1

9 2 1 0 2 2

10 2 0 0 1 0

11 0 1 2 2 1

Page 43: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Consistent table:Consistent table:

Inconsistent table:Inconsistent table:

(e)=card(1,2,3,7,8,10,11)/card(1,2,3,4,5,6,7,8,9,10,11)=7/11=0.636

U a b c d

e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

7 2 0 0 1 0

8 0 1 2 2 1

10 2 0 0 1 0

11 0 1 2 2 1

U a b c d

e

4 0 0 1 2 1

5 2 1 0 2 1

6 0 0 1 2 2

9 2 1 0 2 2

Page 44: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

After remove duplicate After remove duplicate and inconsistentand inconsistent

U a b c de

1 1 2 0 1 1

2 2 0 0 1 0

3 0 1 2 2 1

Page 45: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

After computing, get core After computing, get core and reduct values tableand reduct values table

U a b c d e

1 - - - - 1

2 - - - - 0

3 - - - - 1

U a b c d e

1(1) 1 x x x 1

1(2) x 2 x x 1

2(1) 2 x x x 0

2(2) x 0 x x 0

3(1) 0 x x x 1

3(2) x 1 x x 1

3(3) x x 2 x 1

3(4) x x x 2 1

Core tableCore table

Reduct tableReduct table

Page 46: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

From reduct table, there are 16 From reduct table, there are 16 Simplified Tables. One of them : Simplified Tables. One of them :

corresponding decision corresponding decision algorithms:algorithms:

U a b c d e

1(1) 1 x x x 1

2(1) 2 x x x 0

3(1) 0 x x x 1

a1 a1 e1 e1a2 a2 e0e0a0 a0 e1e1

Page 47: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

It is obvious that in table 10 the new instance does It is obvious that in table 10 the new instance does not change the decision algorithm, that means that not change the decision algorithm, that means that

the learned concepts will remain the same.the learned concepts will remain the same.

But the quality of learning But the quality of learning (e) changed. (from 0.6 (e) changed. (from 0.6 to 0.636)to 0.636)

Page 48: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Table 11Table 11

U a b c d

e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

4 0 0 1 2 1

5 2 1 0 2 1

6 0 0 1 2 2

7 2 0 0 1 0

8 0 1 2 2 1

9 2 1 0 2 2

10 2 0 0 1 0

11 1 2 0 1 0

Page 49: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Consistent table:Consistent table:

Inconsistent table:Inconsistent table:

(e)=card(3,7,8,10)/card(1,2,3,4,5,6,7,8,9,10,11)=4/11=0.363

U a b c d

e

3 2 0 0 1 0

7 2 0 0 1 0

8 0 1 2 2 1

10 2 0 0 1 0

U a b c d

e

1 1 2 0 1 1

2 1 2 0 1 1

4 0 0 1 2 1

5 2 1 0 2 1

6 0 0 1 2 2

9 2 1 0 2 2

11 1 2 0 1 0

Page 50: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

After remove duplicate After remove duplicate and inconsistentand inconsistent

U a b c de

1 2 0 0 1 0

2 0 1 2 2 1

Page 51: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

After computing, get core After computing, get core and reduct values tableand reduct values table

Core tableCore table

Reduct tableReduct table

a b c d

e

- - - - 0

- - - - 1

U a b c d

e

1(1) 2 x x x 0

1(2) x 0 x x 0

1(3) x x 0 x 0

1(4) x x x 1 0

2(1) 0 x x x 1

2(2) x 1 x x 1

2(3) x x 2 x 1

2(4) x x x 2 1

Page 52: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

From reduct table, there are 16 From reduct table, there are 16 Simplified Tables. One of them : Simplified Tables. One of them :

corresponding decision corresponding decision algorithms:algorithms:

U a b c d

e

1(1) 2 x x x 0

2(1) 0 x x x 1

a2 a2 e0e0a0 a0 e1e1 (Compare to(Compare to

a1a1 e1 e1a2 a2 e0e0a0 a0 e1)e1)

Page 53: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Table 12Table 12

U a b c d

e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

4 0 0 1 2 1

5 2 1 0 2 1

6 0 0 1 2 2

7 2 0 0 1 0

8 0 1 2 2 1

9 2 1 0 2 2

10 2 0 0 1 0

11 1 0 0 1 3

Page 54: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Consistent table:Consistent table:

Inconsistent table:Inconsistent table:

(e)=card(1,2,3,7,8,10,11)/card(1,2,3,4,5,6,7,8,9,10)=6/10=0.636

U a b c d

e

1 1 2 0 1 1

2 1 2 0 1 1

3 2 0 0 1 0

7 2 0 0 1 0

8 0 1 2 2 1

10 2 0 0 1 0

11 1 0 0 1 3

U a b c d

e

4 0 0 1 2 1

5 2 1 0 2 1

6 0 0 1 2 2

9 2 1 0 2 2

Page 55: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

After remove duplicate After remove duplicate and inconsistentand inconsistent

U a b c de

1 1 2 0 1 1

2 2 0 0 1 0

3 0 1 2 2 1

4 1 0 0 1 3

Page 56: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

After computing, get core After computing, get core and reduct values tableand reduct values table

Core tableCore table

Reduct tableReduct table

U a b c d

e

1 - 2 - - 1

2 2 - - - 0

3 - - - - 1

4 1 0 - - 3

U a b c d

e

1(1) x 2 x x 1

2(1) 2 x x x 0

3(1) 0 x x x 1

3(2) x 1 x x 1

3(3) x x 2 x 1

3(4) x x x 2 1

4(1) 1 0 x x 3

Page 57: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

From reduct table, there are 4 From reduct table, there are 4 Simplified Tables. One of them : Simplified Tables. One of them :

corresponding decision corresponding decision algorithms:algorithms:

U a b c d

e

1(1) x 2 x x 1

2(1) 2 x x x 0

3(1) 0 x x x 1

4(1) 1 0 x x 3b2 b2 e1e1a2 a2 e0e0a0a0e1e1

a1b0a1b0e3e3

(Compare to(Compare toa1a1 e1 e1a2 a2 e0e0a0 a0 e1)e1)

Page 58: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

To sum up, as we have seen from the above To sum up, as we have seen from the above examples, adding a new instance to the universe examples, adding a new instance to the universe

we could face three possibilities:we could face three possibilities:

1. The new instance confirms actual knowledge. 1. The new instance confirms actual knowledge. (table10)(table10)

2. The new instance contradicts the actual 2. The new instance contradicts the actual knowledge. (table11)knowledge. (table11)

3. The new instance is a completely new case.3. The new instance is a completely new case.

(table12)(table12)

Page 59: Chapter 12 Machine Learning ID: 116 117 Name: Qun Yu (page 1-33) Kai Zhu (page 34-59) Class: CS267 Fall 2008 Instructor: Dr. T.Y.Lin.

Thank youThank you