Regresi Used

35
Introduction to Linear Regression and Correlation Analysis

description

oke

Transcript of Regresi Used

Page 1: Regresi Used

Introduction to Linear Regression and Correlation Analysis

Page 2: Regresi Used

Introduction to Regression Analysis• Regression analysis is used to:

– Predict the value of a dependent variable based on the value of at least one independent variable

– Explain the impact of changes in an independent variable on the dependent variable

Dependent variable: the variable we wish to explain

Independent variable: the variable used to explain the dependent variable

Page 3: Regresi Used

Simple Linear Regression Model

• Only one independent variable, x

• Relationship between x and y is described by a linear function

• Changes in y are assumed to be caused by changes in x

Page 4: Regresi Used

Types of Regression Models

Positive Linear Relationship

Negative Linear Relationship

Relationship NOT Linear

No Relationship

Page 5: Regresi Used

εxββy 10 Linear component

Population Linear Regression

The population regression model:

Population y intercept

Population SlopeCoefficient

Random Error term, or residualDependent

Variable

Independent Variable

Random Error component

Page 6: Regresi Used

Population Linear Regression(continued)

Random Error for this x value

y

x

Observed Value of y for xi

Predicted Value of y for xi

εxββy 10

xi

Slope = β1

Intercept = β0

εi

Page 7: Regresi Used

xbby 10i

The sample regression line provides an estimate of the population regression line

Estimated Regression Model

Estimate of the regression

intercept

Estimate of the regression slope

Estimated (or predicted) y value

Independent variable

The individual random error terms ei have a mean of zero

Page 8: Regresi Used

Least Squares Criterion

• b0 and b1 are obtained by finding the values of b0 and b1 that minimize the sum of the squared residuals

210

22

x))b(b(y

)y(ye

Page 9: Regresi Used

The Least Squares Equation

• The formulas for b1 and b0 are:

algebraic equivalent:

n

xx

n

yxxy

b 22

1 )(

21 )(

))((

xx

yyxxb

xbyb 10

and

Page 10: Regresi Used

• b0 is the estimated average value of

y when the value of x is zero

• b1 is the estimated change in the

average value of y as a result of a one-unit change in x

Interpretation of the Slope and the Intercept

Page 11: Regresi Used

Finding the Least Squares Equation

• The coefficients b0 and b1 will usually be found using computer software, such as Excel or Minitab

• Other regression measures will also be computed as part of computer-based regression analysis

Page 12: Regresi Used

Simple Linear Regression Example

• A real estate agent wishes to examine the relationship between the selling price of a home and its size (measured in square feet)

• A random sample of 10 houses is selected–Dependent variable (y) = house price in

$1000s

– Independent variable (x) = square feet

Page 13: Regresi Used

Sample Data for House Price Model

House Price in $1000s(y)

Square Feet (x)

245 1400

312 1600

279 1700

308 1875

199 1100

219 1550

405 2350

324 2450

319 1425

255 1700

Page 14: Regresi Used

Regression Using Excel

• Tools / Data Analysis / Regression

Page 15: Regresi Used

Excel OutputRegression Statistics

Multiple R 0.76211

R Square 0.58082

Adjusted R Square 0.52842

Standard Error 41.33032

Observations 10

ANOVA  df SS MS F

Significance F

Regression 1 18934.934818934.934

811.084

8 0.01039

Residual 8 13665.5652 1708.1957

Total 9 32600.5000      

 Coefficien

ts Standard Error t StatP-

value Lower 95%Upper 95%

Intercept 98.24833 58.03348 1.692960.1289

2 -35.57720232.0738

6

Square Feet 0.10977 0.03297 3.329380.0103

9 0.03374 0.18580

The regression equation is:

feet) (square 0.10977 98.24833 price house

Page 16: Regresi Used

0

50

100

150

200

250

300

350

400

450

0 500 1000 1500 2000 2500 3000

Square Feet

Ho

use

Pri

ce (

$100

0s)

Graphical Presentation

• House price model: scatter plot and regression line

feet) (square 0.10977 98.24833 price house

Slope = 0.10977

Intercept = 98.248

Page 17: Regresi Used

Explained and Unexplained Variation

• Total variation is made up of two parts:

SSR SSE SST Total sum of Squares

Sum of Squares

Regression

Sum of Squares Error

2)yy(SST 2)yy(SSE 2)yy(SSR

where:

= Average value of the dependent variable

y = Observed values of the dependent variable

= Estimated value of y for the given x valuey

y

Page 18: Regresi Used

• SST = total sum of squares

– Measures the variation of the yi values around their mean y

• SSE = error sum of squares

– Variation attributable to factors other than the relationship between x and y

• SSR = regression sum of squares

– Explained variation attributable to the relationship between x and y

(continued)

Explained and Unexplained Variation

Page 19: Regresi Used

(continued)

Xi

y

x

yi

SST = (yi - y)2

SSE = (yi - yi )2

SSR = (yi - y)2

_

_

_

Explained and Unexplained Variation

y

y

y_y

Page 20: Regresi Used

• The coefficient of determination is the portion of the total variation in the dependent variable that is explained by variation in the independent variable

• The coefficient of determination is also called R-squared and is denoted as R2

Coefficient of Determination, R2

SST

SSRR 2

1R0 2 where

Page 21: Regresi Used

Coefficient of determination

Coefficient of Determination, R2

squares of sum total

regressionby explained squares of sum

SST

SSRR 2

(continued)

Note: In the single independent variable case, the coefficient of determination is

where:R2 = Coefficient of determination

r = Simple correlation coefficient

22 rR

Page 22: Regresi Used

R2 = +1

Examples of Approximate R2 Values

y

x

y

x

R2 = 1

R2 = 1

Perfect linear relationship between x and y:

100% of the variation in y is explained by variation in x

Page 23: Regresi Used

Examples of Approximate R2 Values

y

x

y

x

0 < R2 < 1

Weaker linear relationship between x and y:

Some but not all of the variation in y is explained by variation in x

Page 24: Regresi Used

Examples of Approximate R2 Values

R2 = 0

No linear relationship between x and y:

The value of Y does not depend on x. (None of the variation in y is explained by variation in x)

y

xR2 = 0

Page 25: Regresi Used

Excel OutputRegression Statistics

Multiple R 0.76211

R Square 0.58082

Adjusted R Square 0.52842

Standard Error 41.33032

Observations 10

ANOVA  df SS MS F

Significance F

Regression 1 18934.934818934.934

811.084

8 0.01039

Residual 8 13665.5652 1708.1957

Total 9 32600.5000      

 Coefficien

ts Standard Error t StatP-

value Lower 95%Upper 95%

Intercept 98.24833 58.03348 1.692960.1289

2 -35.57720232.0738

6

Square Feet 0.10977 0.03297 3.329380.0103

9 0.03374 0.18580

58.08%58.08% of the variation in house prices is explained by

variation in square feet

0.5808232600.5000

18934.9348

SST

SSRR2

Page 26: Regresi Used

Comparing Standard Errors

y

y y

x

x

x

y

x

1bs small

1bs large

s small

s large

Variation of observed y values from the regression line

Variation in the slope of regression lines from different possible samples

Page 27: Regresi Used

House Price in $1000s

(y)

Square Feet (x)

245 1400

312 1600

279 1700

308 1875

199 1100

219 1550

405 2350

324 2450

319 1425

255 1700

(sq.ft.) 0.1098 98.25 price house

Estimated Regression Equation:

Example: House Prices

Predict the price for a house with 2000 square feet

Page 28: Regresi Used

317.85

0)0.1098(200 98.25

(sq.ft.) 0.1098 98.25 price house

Example: House PricesPredict the price for a house with 2000 square feet:

The predicted price for a house with 2000 square feet is 317.85($1,000s) = $317,850

(continued)

Page 29: Regresi Used

Residual Analysis

• Purposes–Examine for linearity assumption–Examine for constant variance for all

levels of x –Evaluate normal distribution

assumption

• Graphical Analysis of Residuals–Can plot residuals vs. x–Can create histogram of residuals to

check for normality

Page 30: Regresi Used

Residual Analysis for Linearity

Not Linear Linear

x

resi

dua

ls

x

y

x

y

x

resi

dua

ls

Page 31: Regresi Used

Residual Analysis for Constant Variance

Non-constant variance Constant variance

x x

y

x x

y

resi

dua

ls

resi

dua

ls

Page 32: Regresi Used

House Price Model Residual Plot

-60

-40

-20

0

20

40

60

80

0 1000 2000 3000

Square Feet

Re

sid

ua

ls

Excel Output

RESIDUAL OUTPUT

Predicted House Price Residuals

1 251.92316 -6.923162

2 273.87671 38.12329

3 284.85348 -5.853484

4 304.06284 3.937162

5 218.99284 -19.99284

6 268.38832 -49.38832

7 356.20251 48.79749

8 367.17929 -43.17929

9 254.6674 64.33264

10 284.85348 -29.85348

Page 33: Regresi Used

Summary• Introduced correlation analysis• Discussed correlation to measure the

strength of a linear association• Introduced simple linear regression analysis• Calculated the coefficients for the simple

linear regression equation• measures of variation (R2 and sε)• Addressed assumptions of regression and

correlation

Page 34: Regresi Used

Summary

• Described inference about the slope

• Addressed estimation of mean values and prediction of individual values

• Discussed residual analysis

(continued)

Page 35: Regresi Used

R software regression

• yx=c(245,1400,312,1600,279,1700,308,1875,199,1100,219,1550,405,2350,324,2450,319,1425,255,1700)

• mx=matrix(yx,10,2, byrow=T)• hprice=mx[,1]• sqft=mx[,2]• reg1=lm(hprice~sqft)• summary(reg1)• plot(reg1)