140
Chapter 7: Introduction to linear regression OpenIntro Statistics, 2nd Edition

Chapter 7: Introduction to linear regression

  • Upload
    vuphuc

  • View
    238

  • Download
    2

Embed Size (px)

Citation preview

Page 1: Chapter 7: Introduction to linear regression

Chapter 7: Introduction to linear regression

OpenIntro Statistics, 2nd Edition

Page 2: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

1 Line fitting, residuals, and correlation

2 Fitting a line by least squares regression

3 Types of outliers in linear regression

4 Inference for linear regression

OpenIntro Statistics, 2nd Edition

Chp 7: Intro. to linear regression

Page 3: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Modeling numerical variables

In this unit we will learn to quantify the relationship between twonumerical variables, as well as modeling numerical responsevariables using a numerical or categorical explanatory variable.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 2 / 56

Page 4: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Poverty vs. HS graduate rate

The scatterplot below shows the relationship between HS graduaterate in all 50 US states and DC and the % of residents who live belowthe poverty line (income below $23,050 for a family of 4 in 2012).

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

Response variable?

% in poverty

Explanatory variable?

% HS grad

Relationship?

linear, negative, moderatelystrong

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 3 / 56

Page 5: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Poverty vs. HS graduate rate

The scatterplot below shows the relationship between HS graduaterate in all 50 US states and DC and the % of residents who live belowthe poverty line (income below $23,050 for a family of 4 in 2012).

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

Response variable?

% in poverty

Explanatory variable?

% HS grad

Relationship?

linear, negative, moderatelystrong

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 3 / 56

Page 6: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Poverty vs. HS graduate rate

The scatterplot below shows the relationship between HS graduaterate in all 50 US states and DC and the % of residents who live belowthe poverty line (income below $23,050 for a family of 4 in 2012).

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

Response variable?

% in poverty

Explanatory variable?

% HS grad

Relationship?

linear, negative, moderatelystrong

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 3 / 56

Page 7: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Poverty vs. HS graduate rate

The scatterplot below shows the relationship between HS graduaterate in all 50 US states and DC and the % of residents who live belowthe poverty line (income below $23,050 for a family of 4 in 2012).

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

Response variable?

% in poverty

Explanatory variable?

% HS grad

Relationship?

linear, negative, moderatelystrong

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 3 / 56

Page 8: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Poverty vs. HS graduate rate

The scatterplot below shows the relationship between HS graduaterate in all 50 US states and DC and the % of residents who live belowthe poverty line (income below $23,050 for a family of 4 in 2012).

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

Response variable?

% in poverty

Explanatory variable?

% HS grad

Relationship?

linear, negative, moderatelystrong

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 3 / 56

Page 9: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Poverty vs. HS graduate rate

The scatterplot below shows the relationship between HS graduaterate in all 50 US states and DC and the % of residents who live belowthe poverty line (income below $23,050 for a family of 4 in 2012).

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

Response variable?

% in poverty

Explanatory variable?

% HS grad

Relationship?

linear, negative, moderatelystrong

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 3 / 56

Page 10: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Quantifying the relationship

Correlation describes the strength of the linear associationbetween two variables.

It takes values between -1 (perfect negative) and +1 (perfectpositive).

A value of 0 indicates no linear association.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 4 / 56

Page 11: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Quantifying the relationship

Correlation describes the strength of the linear associationbetween two variables.

It takes values between -1 (perfect negative) and +1 (perfectpositive).

A value of 0 indicates no linear association.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 4 / 56

Page 12: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Quantifying the relationship

Correlation describes the strength of the linear associationbetween two variables.

It takes values between -1 (perfect negative) and +1 (perfectpositive).

A value of 0 indicates no linear association.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 4 / 56

Page 13: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Guessing the correlation

Which of the following is the best guess for the correlation between %in poverty and % HS grad?

(a) 0.6

(b) -0.75

(c) -0.1

(d) 0.02

(e) -1.5

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 5 / 56

Page 14: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Guessing the correlation

Which of the following is the best guess for the correlation between %in poverty and % HS grad?

(a) 0.6

(b) -0.75

(c) -0.1

(d) 0.02

(e) -1.5

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 5 / 56

Page 15: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Guessing the correlation

Which of the following is the best guess for the correlation between% in poverty and % HS grad?

(a) 0.1

(b) -0.6

(c) -0.4

(d) 0.9

(e) 0.5

8 10 12 14 16 18

6

8

10

12

14

16

18

% female householder, no husband present

% in

pov

erty

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 6 / 56

Page 16: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Guessing the correlation

Which of the following is the best guess for the correlation between% in poverty and % HS grad?

(a) 0.1

(b) -0.6

(c) -0.4

(d) 0.9

(e) 0.5

8 10 12 14 16 18

6

8

10

12

14

16

18

% female householder, no husband present

% in

pov

erty

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 6 / 56

Page 17: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Assessing the correlation

Which of the following is has the strongest correlation, i.e. correlationcoefficient closest to +1 or -1?

(a) (b)

(c) (d)

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 7 / 56

Page 18: Chapter 7: Introduction to linear regression

Line fitting, residuals, and correlation

Assessing the correlation

Which of the following is has the strongest correlation, i.e. correlationcoefficient closest to +1 or -1?

(a) (b)

(c) (d)

(b)→correlationmeans linearassociation

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 7 / 56

Page 19: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression

1 Line fitting, residuals, and correlation

2 Fitting a line by least squares regressionEyeballing the lineResidualsBest lineThe least squares lineRecap: Interpreting the slope and the interceptPrediction & extrapolationConditions for the least squares lineR2

Categorical explanatory variables

3 Types of outliers in linear regression

4 Inference for linear regression

OpenIntro Statistics, 2nd Edition

Chp 7: Intro. to linear regression

Page 20: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Eyeballing the line

Eyeballing the line

Which of the follow-ing appears to bethe line that best fitsthe linear relation-ship between % inpoverty and % HSgrad? Choose one.

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

(a)(b)(c)(d)

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 8 / 56

Page 21: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Eyeballing the line

Eyeballing the line

Which of the follow-ing appears to bethe line that best fitsthe linear relation-ship between % inpoverty and % HSgrad? Choose one.

(a)

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

(a)(b)(c)(d)

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 8 / 56

Page 22: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Residuals

Residuals

Residuals are the leftovers from the model fit: Data = Fit + Residual

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 9 / 56

Page 23: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Residuals

Residuals (cont.)

Residual

Residual is the difference between the observed (yi) and predicted yi.

ei = yi − yi

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

y

5.44

yy

−4.16

y

DC

RI

% living in poverty inDC is 5.44% morethan predicted.

% living in poverty inRI is 4.16% less thanpredicted.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 10 / 56

Page 24: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Residuals

Residuals (cont.)

Residual

Residual is the difference between the observed (yi) and predicted yi.

ei = yi − yi

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

y

5.44

yy

−4.16

y

DC

RI

% living in poverty inDC is 5.44% morethan predicted.

% living in poverty inRI is 4.16% less thanpredicted.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 10 / 56

Page 25: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Residuals

Residuals (cont.)

Residual

Residual is the difference between the observed (yi) and predicted yi.

ei = yi − yi

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

y

5.44

yy

−4.16

y

DC

RI

% living in poverty inDC is 5.44% morethan predicted.

% living in poverty inRI is 4.16% less thanpredicted.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 10 / 56

Page 26: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Best line

A measure for the best line

We want a line that has small residuals:

1. Option 1: Minimize the sum of magnitudes (absolute values) ofresiduals

|e1| + |e2| + · · · + |en|

2. Option 2: Minimize the sum of squared residuals – least squares

e21 + e2

2 + · · · + e2n

Why least squares?1. Most commonly used2. Easier to compute by hand and using software3. In many applications, a residual twice as large as another is

usually more than twice as bad

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 11 / 56

Page 27: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Best line

A measure for the best line

We want a line that has small residuals:1. Option 1: Minimize the sum of magnitudes (absolute values) of

residuals|e1| + |e2| + · · · + |en|

2. Option 2: Minimize the sum of squared residuals – least squares

e21 + e2

2 + · · · + e2n

Why least squares?1. Most commonly used2. Easier to compute by hand and using software3. In many applications, a residual twice as large as another is

usually more than twice as bad

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 11 / 56

Page 28: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Best line

A measure for the best line

We want a line that has small residuals:1. Option 1: Minimize the sum of magnitudes (absolute values) of

residuals|e1| + |e2| + · · · + |en|

2. Option 2: Minimize the sum of squared residuals – least squares

e21 + e2

2 + · · · + e2n

Why least squares?1. Most commonly used2. Easier to compute by hand and using software3. In many applications, a residual twice as large as another is

usually more than twice as bad

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 11 / 56

Page 29: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Best line

A measure for the best line

We want a line that has small residuals:1. Option 1: Minimize the sum of magnitudes (absolute values) of

residuals|e1| + |e2| + · · · + |en|

2. Option 2: Minimize the sum of squared residuals – least squares

e21 + e2

2 + · · · + e2n

Why least squares?

1. Most commonly used2. Easier to compute by hand and using software3. In many applications, a residual twice as large as another is

usually more than twice as bad

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 11 / 56

Page 30: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Best line

A measure for the best line

We want a line that has small residuals:1. Option 1: Minimize the sum of magnitudes (absolute values) of

residuals|e1| + |e2| + · · · + |en|

2. Option 2: Minimize the sum of squared residuals – least squares

e21 + e2

2 + · · · + e2n

Why least squares?1. Most commonly used

2. Easier to compute by hand and using software3. In many applications, a residual twice as large as another is

usually more than twice as bad

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 11 / 56

Page 31: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Best line

A measure for the best line

We want a line that has small residuals:1. Option 1: Minimize the sum of magnitudes (absolute values) of

residuals|e1| + |e2| + · · · + |en|

2. Option 2: Minimize the sum of squared residuals – least squares

e21 + e2

2 + · · · + e2n

Why least squares?1. Most commonly used2. Easier to compute by hand and using software

3. In many applications, a residual twice as large as another isusually more than twice as bad

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 11 / 56

Page 32: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Best line

A measure for the best line

We want a line that has small residuals:1. Option 1: Minimize the sum of magnitudes (absolute values) of

residuals|e1| + |e2| + · · · + |en|

2. Option 2: Minimize the sum of squared residuals – least squares

e21 + e2

2 + · · · + e2n

Why least squares?1. Most commonly used2. Easier to compute by hand and using software3. In many applications, a residual twice as large as another is

usually more than twice as bad

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 11 / 56

Page 33: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Best line

The least squares line

y = β0 + β1x��

����predicted y��

��intercept

@@@R

slope

HHHHHj

explanatory variable

Notation:Intercept:

Parameter: β0Point estimate: b0

Slope:Parameter: β1Point estimate: b1

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 12 / 56

Page 34: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression The least squares line

Given...

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

% HS grad % in poverty(x) (y)

mean x = 86.01 y = 11.35sd sx = 3.73 sy = 3.1

correlation R = −0.75

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 13 / 56

Page 35: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression The least squares line

Slope

Slope

The slope of the regression can be calculated as

b1 =sy

sxR

In context...b1 =

3.13.73

× −0.75 = −0.62

InterpretationFor each additional % point in HS graduate rate, we would expect the% living in poverty to be lower on average by 0.62% points.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 14 / 56

Page 36: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression The least squares line

Slope

Slope

The slope of the regression can be calculated as

b1 =sy

sxR

In context...b1 =

3.13.73

× −0.75 = −0.62

InterpretationFor each additional % point in HS graduate rate, we would expect the% living in poverty to be lower on average by 0.62% points.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 14 / 56

Page 37: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression The least squares line

Slope

Slope

The slope of the regression can be calculated as

b1 =sy

sxR

In context...b1 =

3.13.73

× −0.75 = −0.62

InterpretationFor each additional % point in HS graduate rate, we would expect the% living in poverty to be lower on average by 0.62% points.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 14 / 56

Page 38: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression The least squares line

Intercept

Intercept

The intercept is where the regression line intersects the y-axis. Thecalculation of the intercept uses the fact the a regression line alwayspasses through (x, y).

b0 = y − b1x

0 20 40 60 80 1000

10

20

30

40

50

60

70

% HS grad

% in

pov

erty

intercept

b0 = 11.35 − (−0.62) × 86.01

= 64.68

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 15 / 56

Page 39: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression The least squares line

Intercept

Intercept

The intercept is where the regression line intersects the y-axis. Thecalculation of the intercept uses the fact the a regression line alwayspasses through (x, y).

b0 = y − b1x

0 20 40 60 80 1000

10

20

30

40

50

60

70

% HS grad

% in

pov

erty

intercept

b0 = 11.35 − (−0.62) × 86.01

= 64.68

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 15 / 56

Page 40: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression The least squares line

Intercept

Intercept

The intercept is where the regression line intersects the y-axis. Thecalculation of the intercept uses the fact the a regression line alwayspasses through (x, y).

b0 = y − b1x

0 20 40 60 80 1000

10

20

30

40

50

60

70

% HS grad

% in

pov

erty

intercept

b0 = 11.35 − (−0.62) × 86.01

= 64.68

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 15 / 56

Page 41: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression The least squares line

Which of the following is the correct interpretation of the intercept?

(a) For each % point increase in HS graduate rate, % living in povertyis expected to increase on average by 64.68%.

(b) For each % point decrease in HS graduate rate, % living inpoverty is expected to increase on average by 64.68%.

(c) Having no HS graduates leads to 64.68% of residents livingbelow the poverty line.

(d) States with no HS graduates are expected on average to have64.68% of residents living below the poverty line.

(e) In states with no HS graduates % living in poverty is expected toincrease on average by 64.68%.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 16 / 56

Page 42: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression The least squares line

Which of the following is the correct interpretation of the intercept?

(a) For each % point increase in HS graduate rate, % living in povertyis expected to increase on average by 64.68%.

(b) For each % point decrease in HS graduate rate, % living inpoverty is expected to increase on average by 64.68%.

(c) Having no HS graduates leads to 64.68% of residents livingbelow the poverty line.

(d) States with no HS graduates are expected on average to have64.68% of residents living below the poverty line.

(e) In states with no HS graduates % living in poverty is expected toincrease on average by 64.68%.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 16 / 56

Page 43: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression The least squares line

More on the intercept

Since there are no states in the dataset with no HS graduates, theintercept is of no interest, not very useful, and also not reliable sincethe predicted value of the intercept is so far from the bulk of the data.

0 20 40 60 80 1000

10

20

30

40

50

60

70

% HS grad

% in

pov

erty

intercept

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 17 / 56

Page 44: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression The least squares line

Regression line

% in poverty = 64.68 − 0.62 % HS grad

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 18 / 56

Page 45: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Recap: Interpreting the slope and the intercept

Interpretation of slope and intercept

Intercept: When x = 0, y isexpected to equal theintercept.

Slope: For each unit in x, y isexpected to increase /decrease on average by theslope.

Note: These statements are not causal, unless the study is a randomized controlled

experiment.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 19 / 56

Page 46: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Prediction & extrapolation

Prediction

Using the linear model to predict the value of the responsevariable for a given value of the explanatory variable is calledprediction, simply by plugging in the value of x in the linear modelequation.There will be some uncertainty associated with the predictedvalue.

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 20 / 56

Page 47: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Prediction & extrapolation

Extrapolation

Applying a model estimate to values outside of the realm of theoriginal data is called extrapolation.

Sometimes the intercept might be an extrapolation.

0 20 40 60 80 1000

10

20

30

40

50

60

70

% HS grad

% in

pov

erty

intercept

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 21 / 56

Page 48: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Prediction & extrapolation

Examples of extrapolation

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 22 / 56

Page 49: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Prediction & extrapolation

Examples of extrapolation

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 23 / 56

Page 50: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Prediction & extrapolation

Examples of extrapolation

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 24 / 56

Page 51: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions for the least squares line

1. Linearity

2. Nearly normal residuals

3. Constant variability

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 25 / 56

Page 52: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions for the least squares line

1. Linearity

2. Nearly normal residuals

3. Constant variability

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 25 / 56

Page 53: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions for the least squares line

1. Linearity

2. Nearly normal residuals

3. Constant variability

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 25 / 56

Page 54: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions: (1) Linearity

The relationship between the explanatory and the responsevariable should be linear.

Methods for fitting a model to non-linear relationships exist, butare beyond the scope of this class. If this topic is of interest, anOnline Extra is available on openintro.org covering newtechniques.Check using a scatterplot of the data, or a residuals plot.

x x

ysu

mm

ary(

g)$r

esid

uals

x

ysu

mm

ary(

g)$r

esid

uals

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 26 / 56

Page 55: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions: (1) Linearity

The relationship between the explanatory and the responsevariable should be linear.Methods for fitting a model to non-linear relationships exist, butare beyond the scope of this class. If this topic is of interest, anOnline Extra is available on openintro.org covering newtechniques.

Check using a scatterplot of the data, or a residuals plot.

x x

ysu

mm

ary(

g)$r

esid

uals

x

ysu

mm

ary(

g)$r

esid

uals

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 26 / 56

Page 56: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions: (1) Linearity

The relationship between the explanatory and the responsevariable should be linear.Methods for fitting a model to non-linear relationships exist, butare beyond the scope of this class. If this topic is of interest, anOnline Extra is available on openintro.org covering newtechniques.Check using a scatterplot of the data, or a residuals plot.

x x

ysu

mm

ary(

g)$r

esid

uals

x

ysu

mm

ary(

g)$r

esid

uals

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 26 / 56

Page 57: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Anatomy of a residuals plot

% HS grad

% in

pov

erty

80 85 90

5

10

15

−5

0

5

N RI:

% HS grad = 81 % in poverty = 10.3% in poverty = 64.68 − 0.62 ∗ 81 = 14.46

e = % in poverty − % in poverty

= 10.3 − 14.46 = −4.16

� DC:

% HS grad = 86 % in poverty = 16.8% in poverty = 64.68 − 0.62 ∗ 86 = 11.36

e = % in poverty − % in poverty

= 16.8 − 11.36 = 5.44

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 27 / 56

Page 58: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Anatomy of a residuals plot

% HS grad

% in

pov

erty

80 85 90

5

10

15

−5

0

5

N RI:

% HS grad = 81 % in poverty = 10.3% in poverty = 64.68 − 0.62 ∗ 81 = 14.46

e = % in poverty − % in poverty

= 10.3 − 14.46 = −4.16

� DC:

% HS grad = 86 % in poverty = 16.8% in poverty = 64.68 − 0.62 ∗ 86 = 11.36

e = % in poverty − % in poverty

= 16.8 − 11.36 = 5.44

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 27 / 56

Page 59: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions: (2) Nearly normal residuals

The residuals should be nearly normal.

This condition may not be satisfied when there are unusualobservations that don’t follow the trend of the rest of the data.Check using a histogram or normal probability plot of residuals.

residuals

freq

uenc

y

−4 −2 0 2 4 6

02

46

810

12

−2 −1 0 1 2

−4

−2

02

4

Normal Q−Q Plot

Theoretical Quantiles

Sam

ple

Qua

ntile

s

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 28 / 56

Page 60: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions: (2) Nearly normal residuals

The residuals should be nearly normal.This condition may not be satisfied when there are unusualobservations that don’t follow the trend of the rest of the data.

Check using a histogram or normal probability plot of residuals.

residuals

freq

uenc

y

−4 −2 0 2 4 6

02

46

810

12

−2 −1 0 1 2

−4

−2

02

4

Normal Q−Q Plot

Theoretical Quantiles

Sam

ple

Qua

ntile

s

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 28 / 56

Page 61: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions: (2) Nearly normal residuals

The residuals should be nearly normal.This condition may not be satisfied when there are unusualobservations that don’t follow the trend of the rest of the data.Check using a histogram or normal probability plot of residuals.

residuals

freq

uenc

y

−4 −2 0 2 4 6

02

46

810

12

−2 −1 0 1 2

−4

−2

02

4

Normal Q−Q Plot

Theoretical Quantiles

Sam

ple

Qua

ntile

s

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 28 / 56

Page 62: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions: (3) Constant variability

80 85 90

68

1012

1416

18

% HS grad

% in

pov

erty

80 90

−4

04

The variability of pointsaround the least squares lineshould be roughly constant.

This implies that the variabilityof residuals around the 0 lineshould be roughly constant aswell.

Also called homoscedasticity.

Check using a histogram ornormal probability plot ofresiduals.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 29 / 56

Page 63: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions: (3) Constant variability

80 85 90

68

1012

1416

18

% HS grad

% in

pov

erty

80 90

−4

04

The variability of pointsaround the least squares lineshould be roughly constant.

This implies that the variabilityof residuals around the 0 lineshould be roughly constant aswell.

Also called homoscedasticity.

Check using a histogram ornormal probability plot ofresiduals.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 29 / 56

Page 64: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions: (3) Constant variability

80 85 90

68

1012

1416

18

% HS grad

% in

pov

erty

80 90

−4

04

The variability of pointsaround the least squares lineshould be roughly constant.

This implies that the variabilityof residuals around the 0 lineshould be roughly constant aswell.

Also called homoscedasticity.

Check using a histogram ornormal probability plot ofresiduals.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 29 / 56

Page 65: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Conditions: (3) Constant variability

80 85 90

68

1012

1416

18

% HS grad

% in

pov

erty

80 90

−4

04

The variability of pointsaround the least squares lineshould be roughly constant.

This implies that the variabilityof residuals around the 0 lineshould be roughly constant aswell.

Also called homoscedasticity.

Check using a histogram ornormal probability plot ofresiduals.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 29 / 56

Page 66: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Checking conditions

What condition is this linear modelobviously violating?

(a) Constant variability

(b) Linear relationship

(c) Normal residuals

(d) No extreme outliersx x

yg$residuals

x

yg$residuals

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 30 / 56

Page 67: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Checking conditions

What condition is this linear modelobviously violating?

(a) Constant variability

(b) Linear relationship

(c) Normal residuals

(d) No extreme outliersx x

yg$residuals

x

yg$residuals

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 30 / 56

Page 68: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Checking conditions

What condition is this linear modelobviously violating?

(a) Constant variability

(b) Linear relationship

(c) Normal residuals

(d) No extreme outliersx x

yg$residuals

x

yg$residuals

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 31 / 56

Page 69: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Conditions for the least squares line

Checking conditions

What condition is this linear modelobviously violating?

(a) Constant variability

(b) Linear relationship

(c) Normal residuals

(d) No extreme outliersx x

yg$residuals

x

yg$residuals

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 31 / 56

Page 70: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression R2

R2

The strength of the fit of a linear model is most commonlyevaluated using R2.

R2 is calculated as the square of the correlation coefficient.

It tells us what percent of variability in the response variable isexplained by the model.

The remainder of the variability is explained by variables notincluded in the model or by inherent randomness in the data.

For the model we’ve been working with, R2 = −0.622 = 0.38.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 32 / 56

Page 71: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression R2

R2

The strength of the fit of a linear model is most commonlyevaluated using R2.

R2 is calculated as the square of the correlation coefficient.

It tells us what percent of variability in the response variable isexplained by the model.

The remainder of the variability is explained by variables notincluded in the model or by inherent randomness in the data.

For the model we’ve been working with, R2 = −0.622 = 0.38.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 32 / 56

Page 72: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression R2

R2

The strength of the fit of a linear model is most commonlyevaluated using R2.

R2 is calculated as the square of the correlation coefficient.

It tells us what percent of variability in the response variable isexplained by the model.

The remainder of the variability is explained by variables notincluded in the model or by inherent randomness in the data.

For the model we’ve been working with, R2 = −0.622 = 0.38.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 32 / 56

Page 73: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression R2

R2

The strength of the fit of a linear model is most commonlyevaluated using R2.

R2 is calculated as the square of the correlation coefficient.

It tells us what percent of variability in the response variable isexplained by the model.

The remainder of the variability is explained by variables notincluded in the model or by inherent randomness in the data.

For the model we’ve been working with, R2 = −0.622 = 0.38.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 32 / 56

Page 74: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression R2

R2

The strength of the fit of a linear model is most commonlyevaluated using R2.

R2 is calculated as the square of the correlation coefficient.

It tells us what percent of variability in the response variable isexplained by the model.

The remainder of the variability is explained by variables notincluded in the model or by inherent randomness in the data.

For the model we’ve been working with, R2 = −0.622 = 0.38.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 32 / 56

Page 75: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression R2

Interpretation of R2

Which of the below is the correct interpretation of R = −0.62, R2 = 0.38?

(a) 38% of the variability in the % of HGgraduates among the 51 states isexplained by the model.

(b) 38% of the variability in the % ofresidents living in poverty among the 51states is explained by the model.

(c) 38% of the time % HS graduates predict% living in poverty correctly.

(d) 62% of the variability in the % ofresidents living in poverty among the 51states is explained by the model.

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 33 / 56

Page 76: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression R2

Interpretation of R2

Which of the below is the correct interpretation of R = −0.62, R2 = 0.38?

(a) 38% of the variability in the % of HGgraduates among the 51 states isexplained by the model.

(b) 38% of the variability in the % ofresidents living in poverty among the 51states is explained by the model.

(c) 38% of the time % HS graduates predict% living in poverty correctly.

(d) 62% of the variability in the % ofresidents living in poverty among the 51states is explained by the model.

80 85 90

6

8

10

12

14

16

18

% HS grad

% in

pov

erty

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 33 / 56

Page 77: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Categorical explanatory variables

Poverty vs. region (east, west)

poverty = 11.17 + 0.38 × west

Explanatory variable: region, reference level: eastIntercept: The estimated average poverty percentage in easternstates is 11.17%

This is the value we get if we plug in 0 for the explanatory variable

Slope: The estimated average poverty percentage in westernstates is 0.38% higher than eastern states.

Then, the estimated average poverty percentage in westernstates is 11.17 + 0.38 = 11.55%.This is the value we get if we plug in 1 for the explanatory variable

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 34 / 56

Page 78: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Categorical explanatory variables

Poverty vs. region (east, west)

poverty = 11.17 + 0.38 × west

Explanatory variable: region, reference level: eastIntercept: The estimated average poverty percentage in easternstates is 11.17%

This is the value we get if we plug in 0 for the explanatory variable

Slope: The estimated average poverty percentage in westernstates is 0.38% higher than eastern states.

Then, the estimated average poverty percentage in westernstates is 11.17 + 0.38 = 11.55%.This is the value we get if we plug in 1 for the explanatory variable

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 34 / 56

Page 79: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Categorical explanatory variables

Poverty vs. region (east, west)

poverty = 11.17 + 0.38 × west

Explanatory variable: region, reference level: eastIntercept: The estimated average poverty percentage in easternstates is 11.17%

This is the value we get if we plug in 0 for the explanatory variable

Slope: The estimated average poverty percentage in westernstates is 0.38% higher than eastern states.

Then, the estimated average poverty percentage in westernstates is 11.17 + 0.38 = 11.55%.This is the value we get if we plug in 1 for the explanatory variable

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 34 / 56

Page 80: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Categorical explanatory variables

Poverty vs. region (east, west)

poverty = 11.17 + 0.38 × west

Explanatory variable: region, reference level: eastIntercept: The estimated average poverty percentage in easternstates is 11.17%

This is the value we get if we plug in 0 for the explanatory variable

Slope: The estimated average poverty percentage in westernstates is 0.38% higher than eastern states.

Then, the estimated average poverty percentage in westernstates is 11.17 + 0.38 = 11.55%.

This is the value we get if we plug in 1 for the explanatory variable

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 34 / 56

Page 81: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Categorical explanatory variables

Poverty vs. region (east, west)

poverty = 11.17 + 0.38 × west

Explanatory variable: region, reference level: eastIntercept: The estimated average poverty percentage in easternstates is 11.17%

This is the value we get if we plug in 0 for the explanatory variable

Slope: The estimated average poverty percentage in westernstates is 0.38% higher than eastern states.

Then, the estimated average poverty percentage in westernstates is 11.17 + 0.38 = 11.55%.This is the value we get if we plug in 1 for the explanatory variable

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 34 / 56

Page 82: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Categorical explanatory variables

Poverty vs. region (northeast, midwest, west, south)

Which region (northeast, midwest, west, or south) is the referencelevel?

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.50 0.87 10.94 0.00

region4midwest 0.03 1.15 0.02 0.98region4west 1.79 1.13 1.59 0.12

region4south 4.16 1.07 3.87 0.00

(a) northeast

(b) midwest

(c) west

(d) south

(e) cannot tell

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 35 / 56

Page 83: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Categorical explanatory variables

Poverty vs. region (northeast, midwest, west, south)

Which region (northeast, midwest, west, or south) is the referencelevel?

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.50 0.87 10.94 0.00

region4midwest 0.03 1.15 0.02 0.98region4west 1.79 1.13 1.59 0.12

region4south 4.16 1.07 3.87 0.00

(a) northeast

(b) midwest

(c) west

(d) south

(e) cannot tell

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 35 / 56

Page 84: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Categorical explanatory variables

Poverty vs. region (northeast, midwest, west, south)

Which region (northeast, midwest, west, or south) has the lowestpoverty percentage?

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.50 0.87 10.94 0.00

region4midwest 0.03 1.15 0.02 0.98region4west 1.79 1.13 1.59 0.12

region4south 4.16 1.07 3.87 0.00

(a) northeast

(b) midwest

(c) west

(d) south

(e) cannot tell

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 36 / 56

Page 85: Chapter 7: Introduction to linear regression

Fitting a line by least squares regression Categorical explanatory variables

Poverty vs. region (northeast, midwest, west, south)

Which region (northeast, midwest, west, or south) has the lowestpoverty percentage?

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.50 0.87 10.94 0.00

region4midwest 0.03 1.15 0.02 0.98region4west 1.79 1.13 1.59 0.12

region4south 4.16 1.07 3.87 0.00

(a) northeast

(b) midwest

(c) west

(d) south

(e) cannot tell

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 36 / 56

Page 86: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

1 Line fitting, residuals, and correlation

2 Fitting a line by least squares regression

3 Types of outliers in linear regression

4 Inference for linear regression

OpenIntro Statistics, 2nd Edition

Chp 7: Intro. to linear regression

Page 87: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Types of outliers

How do outliers influence the leastsquares line in this plot?

To answer this question think ofwhere the regression line would bewith and without the outlier(s).Without the outliers the regressionline would be steeper, and liecloser to the larger group ofobservations. With the outliers theline is pulled up and away fromsome of the observations in thelarger group.

−20

−10

0

−5

0

5

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 37 / 56

Page 88: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Types of outliers

How do outliers influencethe least squares line in thisplot? 0

5

10

−2

0

2

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 38 / 56

Page 89: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Types of outliers

How do outliers influencethe least squares line in thisplot?

Without the outlier there isno evident relationshipbetween x and y.

0

5

10

−2

0

2

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 38 / 56

Page 90: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Some terminology

Outliers are points that lie away from the cloud of points.

Outliers that lie horizontally away from the center of the cloud arecalled high leverage points.

High leverage points that actually influence the slope of theregression line are called influential points.

In order to determine if a point is influential, visualize theregression line with and without the point. Does the slope of theline change considerably? If so, then the point is influential. Ifnot, then itOs not an influential point.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 39 / 56

Page 91: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Some terminology

Outliers are points that lie away from the cloud of points.

Outliers that lie horizontally away from the center of the cloud arecalled high leverage points.

High leverage points that actually influence the slope of theregression line are called influential points.

In order to determine if a point is influential, visualize theregression line with and without the point. Does the slope of theline change considerably? If so, then the point is influential. Ifnot, then itOs not an influential point.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 39 / 56

Page 92: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Some terminology

Outliers are points that lie away from the cloud of points.

Outliers that lie horizontally away from the center of the cloud arecalled high leverage points.

High leverage points that actually influence the slope of theregression line are called influential points.

In order to determine if a point is influential, visualize theregression line with and without the point. Does the slope of theline change considerably? If so, then the point is influential. Ifnot, then itOs not an influential point.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 39 / 56

Page 93: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Some terminology

Outliers are points that lie away from the cloud of points.

Outliers that lie horizontally away from the center of the cloud arecalled high leverage points.

High leverage points that actually influence the slope of theregression line are called influential points.

In order to determine if a point is influential, visualize theregression line with and without the point. Does the slope of theline change considerably? If so, then the point is influential. Ifnot, then itOs not an influential point.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 39 / 56

Page 94: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Influential points

Data are available on the log of the surface temperature and the logof the light intensity of 47 stars in the star cluster CYG OB1.

3.6 3.8 4.0 4.2 4.4 4.6

4.0

4.5

5.0

5.5

6.0

log(temp)

log(

light

inte

nsity

)

w/ outliersw/o outliers

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 40 / 56

Page 95: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Types of outliers

Which of the below best de-scribes the outlier?

(a) influential

(b) high leverage

(c) none of the above

(d) there are no outliers

−20

0

20

40

−2

0

2

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 41 / 56

Page 96: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Types of outliers

Which of the below best de-scribes the outlier?

(a) influential

(b) high leverage

(c) none of the above

(d) there are no outliers

−20

0

20

40

−2

0

2

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 41 / 56

Page 97: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Types of outliers

Does this outlier influencethe slope of the regressionline?

−5

0

5

10

15

−5

0

5

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 42 / 56

Page 98: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Types of outliers

Does this outlier influencethe slope of the regressionline?

Not much...−5

0

5

10

15

−5

0

5

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 42 / 56

Page 99: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Recap

Which of following is true?

(a) Influential points always change the intercept of the regressionline.

(b) Influential points always reduce R2.

(c) It is much more likely for a low leverage point to be influential,than a high leverage point.

(d) When the data set includes an influential point, the relationshipbetween the explanatory variable and the response variable isalways nonlinear.

(e) None of the above.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 43 / 56

Page 100: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Recap

Which of following is true?

(a) Influential points always change the intercept of the regressionline.

(b) Influential points always reduce R2.

(c) It is much more likely for a low leverage point to be influential,than a high leverage point.

(d) When the data set includes an influential point, the relationshipbetween the explanatory variable and the response variable isalways nonlinear.

(e) None of the above.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 43 / 56

Page 101: Chapter 7: Introduction to linear regression

Types of outliers in linear regression

Recap (cont.)

R = 0.08,R2 = 0.0064

−2

0

2

4

6

8

10

−2

0

2

R = 0.79,R2 = 0.6241

0

5

10

−2

0

2

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 44 / 56

Page 102: Chapter 7: Introduction to linear regression

Inference for linear regression

1 Line fitting, residuals, and correlation

2 Fitting a line by least squares regression

3 Types of outliers in linear regression

4 Inference for linear regressionUnderstanding regression output from softwareHT for the slopeCI for the slope

OpenIntro Statistics, 2nd Edition

Chp 7: Intro. to linear regression

Page 103: Chapter 7: Introduction to linear regression

Inference for linear regression Understanding regression output from software

Nature or nurture?

In 1966 Cyril Burt published a paper called “The genetic determination ofdifferences in intelligence: A study of monozygotic twins reared apart?” Thedata consist of IQ scores for [an assumed random sample of] 27 identicaltwins, one raised by foster parents, the other by the biological parents.

70 80 90 100 110 120 130

70

80

90

100

110

120

130

biological IQ

fost

er IQ

R = 0.882

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 45 / 56

Page 104: Chapter 7: Introduction to linear regression

Inference for linear regression Understanding regression output from software

Which of the following is false?

Coefficients:

Estimate Std. Error t value Pr(>|t|)

(Intercept) 9.20760 9.29990 0.990 0.332

bioIQ 0.90144 0.09633 9.358 1.2e-09

Residual standard error: 7.729 on 25 degrees of freedom

Multiple R-squared: 0.7779, Adjusted R-squared: 0.769

F-statistic: 87.56 on 1 and 25 DF, p-value: 1.204e-09

(a) Additional 10 points in the biological twin’s IQ is associated withadditional 9 points in the foster twin’s IQ, on average.

(b) Roughly 78% of the foster twins’ IQs can be accurately predictedby the model.

(c) The linear model is fosterIQ = 9.2 + 0.9 × bioIQ.(d) Foster twins with IQs higher than average IQs tend to have

biological twins with higher than average IQs as well.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 46 / 56

Page 105: Chapter 7: Introduction to linear regression

Inference for linear regression Understanding regression output from software

Which of the following is false?

Coefficients:

Estimate Std. Error t value Pr(>|t|)

(Intercept) 9.20760 9.29990 0.990 0.332

bioIQ 0.90144 0.09633 9.358 1.2e-09

Residual standard error: 7.729 on 25 degrees of freedom

Multiple R-squared: 0.7779, Adjusted R-squared: 0.769

F-statistic: 87.56 on 1 and 25 DF, p-value: 1.204e-09

(a) Additional 10 points in the biological twin’s IQ is associated withadditional 9 points in the foster twin’s IQ, on average.

(b) Roughly 78% of the foster twins’ IQs can be accurately predictedby the model.

(c) The linear model is fosterIQ = 9.2 + 0.9 × bioIQ.(d) Foster twins with IQs higher than average IQs tend to have

biological twins with higher than average IQs as well.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 46 / 56

Page 106: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope

Assuming that these 27 twins comprise a representative sample of alltwins separated at birth, we would like to test if these data provideconvincing evidence that the IQ of the biological twin is a significantpredictor of IQ of the foster twin. What are the appropriate hypothe-ses?

(a) H0 : b0 = 0; HA : b0 , 0

(b) H0 : β0 = 0; HA : β0 , 0

(c) H0 : b1 = 0; HA : b1 , 0(d) H0 : β1 = 0; HA : β1 , 0

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 47 / 56

Page 107: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope

Assuming that these 27 twins comprise a representative sample of alltwins separated at birth, we would like to test if these data provideconvincing evidence that the IQ of the biological twin is a significantpredictor of IQ of the foster twin. What are the appropriate hypothe-ses?

(a) H0 : b0 = 0; HA : b0 , 0

(b) H0 : β0 = 0; HA : β0 , 0

(c) H0 : b1 = 0; HA : b1 , 0(d) H0 : β1 = 0; HA : β1 , 0

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 47 / 56

Page 108: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope (cont.)

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

We always use a t-test in inference for regression.Remember: Test statistic, T = point estimate−null value

SE

Point estimate = b1 is the observed slope.

SEb1 is the standard error associated with the slope.

Degrees of freedom associated with the slope is df = n − 2,where n is the sample size.Remember: We lose 1 degree of freedom for each parameter we estimate, and in simple

linear regression we estimate 2 parameters, β0 and β1.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 48 / 56

Page 109: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope (cont.)

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

We always use a t-test in inference for regression.

Remember: Test statistic, T = point estimate−null valueSE

Point estimate = b1 is the observed slope.

SEb1 is the standard error associated with the slope.

Degrees of freedom associated with the slope is df = n − 2,where n is the sample size.Remember: We lose 1 degree of freedom for each parameter we estimate, and in simple

linear regression we estimate 2 parameters, β0 and β1.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 48 / 56

Page 110: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope (cont.)

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

We always use a t-test in inference for regression.Remember: Test statistic, T = point estimate−null value

SE

Point estimate = b1 is the observed slope.

SEb1 is the standard error associated with the slope.

Degrees of freedom associated with the slope is df = n − 2,where n is the sample size.Remember: We lose 1 degree of freedom for each parameter we estimate, and in simple

linear regression we estimate 2 parameters, β0 and β1.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 48 / 56

Page 111: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope (cont.)

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

We always use a t-test in inference for regression.Remember: Test statistic, T = point estimate−null value

SE

Point estimate = b1 is the observed slope.

SEb1 is the standard error associated with the slope.

Degrees of freedom associated with the slope is df = n − 2,where n is the sample size.Remember: We lose 1 degree of freedom for each parameter we estimate, and in simple

linear regression we estimate 2 parameters, β0 and β1.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 48 / 56

Page 112: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope (cont.)

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

We always use a t-test in inference for regression.Remember: Test statistic, T = point estimate−null value

SE

Point estimate = b1 is the observed slope.

SEb1 is the standard error associated with the slope.

Degrees of freedom associated with the slope is df = n − 2,where n is the sample size.Remember: We lose 1 degree of freedom for each parameter we estimate, and in simple

linear regression we estimate 2 parameters, β0 and β1.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 48 / 56

Page 113: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope (cont.)

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

We always use a t-test in inference for regression.Remember: Test statistic, T = point estimate−null value

SE

Point estimate = b1 is the observed slope.

SEb1 is the standard error associated with the slope.

Degrees of freedom associated with the slope is df = n − 2,where n is the sample size.

Remember: We lose 1 degree of freedom for each parameter we estimate, and in simple

linear regression we estimate 2 parameters, β0 and β1.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 48 / 56

Page 114: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope (cont.)

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

We always use a t-test in inference for regression.Remember: Test statistic, T = point estimate−null value

SE

Point estimate = b1 is the observed slope.

SEb1 is the standard error associated with the slope.

Degrees of freedom associated with the slope is df = n − 2,where n is the sample size.Remember: We lose 1 degree of freedom for each parameter we estimate, and in simple

linear regression we estimate 2 parameters, β0 and β1.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 48 / 56

Page 115: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope (cont.)

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

T =0.9014 − 0

0.0963= 9.36

df = 27 − 2 = 25

p − value = P(|T | > 9.36) < 0.01

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 49 / 56

Page 116: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope (cont.)

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

T =0.9014 − 0

0.0963= 9.36

df = 27 − 2 = 25

p − value = P(|T | > 9.36) < 0.01

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 49 / 56

Page 117: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope (cont.)

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

T =0.9014 − 0

0.0963= 9.36

df = 27 − 2 = 25

p − value = P(|T | > 9.36) < 0.01

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 49 / 56

Page 118: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

Testing for the slope (cont.)

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

T =0.9014 − 0

0.0963= 9.36

df = 27 − 2 = 25

p − value = P(|T | > 9.36) < 0.01

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 49 / 56

Page 119: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

% College graduate vs. % Hispanic in LA

What can you say about the relationship between % college graduateand % Hispanic in a sample of 100 zip code areas in LA?

Education: College graduate

0.0

0.2

0.4

0.6

0.8

1.0

No dataFreeways

Race/Ethnicity: Hispanic

0.0

0.2

0.4

0.6

0.8

1.0

No dataFreeways

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 50 / 56

Page 120: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

% College educated vs. % Hispanic in LA - another look

What can you say about the relationship between of % college gradu-ate and % Hispanic in a sample of 100 zip code areas in LA?

% Hispanic

% C

olle

ge g

radu

ate

0% 25% 50% 75% 100%

0%

25%

50%

75%

100%

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 51 / 56

Page 121: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

% College educated vs. % Hispanic in LA - linear model

Which of the below is the best interpretation of the slope?

Estimate Std. Error t value Pr(>|t|)(Intercept) 0.7290 0.0308 23.68 0.0000%Hispanic -0.7527 0.0501 -15.01 0.0000

(a) A 1% increase in Hispanic residents in a zip code area in LA isassociated with a 75% decrease in % of college grads.

(b) A 1% increase in Hispanic residents in a zip code area in LA isassociated with a 0.75% decrease in % of college grads.

(c) An additional 1% of Hispanic residents decreases the % ofcollege graduates in a zip code area in LA by 0.75%.

(d) In zip code areas with no Hispanic residents, % of collegegraduates is expected to be 75%.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 52 / 56

Page 122: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

% College educated vs. % Hispanic in LA - linear model

Which of the below is the best interpretation of the slope?

Estimate Std. Error t value Pr(>|t|)(Intercept) 0.7290 0.0308 23.68 0.0000%Hispanic -0.7527 0.0501 -15.01 0.0000

(a) A 1% increase in Hispanic residents in a zip code area in LA isassociated with a 75% decrease in % of college grads.

(b) A 1% increase in Hispanic residents in a zip code area in LA isassociated with a 0.75% decrease in % of college grads.

(c) An additional 1% of Hispanic residents decreases the % ofcollege graduates in a zip code area in LA by 0.75%.

(d) In zip code areas with no Hispanic residents, % of collegegraduates is expected to be 75%.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 52 / 56

Page 123: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

% College educated vs. % Hispanic in LA - linear model

Do these data provide convincing evidence that there is a statisticallysignificant relationship between % Hispanic and % college graduatesin zip code areas in LA?

Estimate Std. Error t value Pr(>|t|)(Intercept) 0.7290 0.0308 23.68 0.0000

hispanic -0.7527 0.0501 -15.01 0.0000

How reliable is this p-value if these zip code areas are not randomlyselected?

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 53 / 56

Page 124: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

% College educated vs. % Hispanic in LA - linear model

Do these data provide convincing evidence that there is a statisticallysignificant relationship between % Hispanic and % college graduatesin zip code areas in LA?

Estimate Std. Error t value Pr(>|t|)(Intercept) 0.7290 0.0308 23.68 0.0000

hispanic -0.7527 0.0501 -15.01 0.0000

Yes, the p-value for % Hispanic is low, indicating that the data provideconvincing evidence that the slope parameter is different than 0.

How reliable is this p-value if these zip code areas are not randomlyselected?

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 53 / 56

Page 125: Chapter 7: Introduction to linear regression

Inference for linear regression HT for the slope

% College educated vs. % Hispanic in LA - linear model

Do these data provide convincing evidence that there is a statisticallysignificant relationship between % Hispanic and % college graduatesin zip code areas in LA?

Estimate Std. Error t value Pr(>|t|)(Intercept) 0.7290 0.0308 23.68 0.0000

hispanic -0.7527 0.0501 -15.01 0.0000

Yes, the p-value for % Hispanic is low, indicating that the data provideconvincing evidence that the slope parameter is different than 0.

How reliable is this p-value if these zip code areas are not randomlyselected?

Not very...

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 53 / 56

Page 126: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Confidence interval for the slope

Remember that a confidence interval is calculated as point estimate±ME andthe degrees of freedom associated with the slope in a simple linear regressionis n−2. Which of the below is the correct 95% confidence interval for the slopeparameter? Note that the model is based on observations from 27 twins.

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

(a) 9.2076 ± 1.65 × 9.2999(b) 0.9014 ± 2.06 × 0.0963

(c) 0.9014 ± 1.96 × 0.0963

(d) 9.2076 ± 1.96 × 0.0963

n = 27 df = 27 − 2 = 25

95% : t?25 = 2.06

0.9014 ± 2.06 × 0.0963

(0.7 , 1.1)

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 54 / 56

Page 127: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Confidence interval for the slope

Remember that a confidence interval is calculated as point estimate±ME andthe degrees of freedom associated with the slope in a simple linear regressionis n−2. Which of the below is the correct 95% confidence interval for the slopeparameter? Note that the model is based on observations from 27 twins.

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

(a) 9.2076 ± 1.65 × 9.2999(b) 0.9014 ± 2.06 × 0.0963

(c) 0.9014 ± 1.96 × 0.0963

(d) 9.2076 ± 1.96 × 0.0963

n = 27 df = 27 − 2 = 25

95% : t?25 = 2.06

0.9014 ± 2.06 × 0.0963

(0.7 , 1.1)

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 54 / 56

Page 128: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Confidence interval for the slope

Remember that a confidence interval is calculated as point estimate±ME andthe degrees of freedom associated with the slope in a simple linear regressionis n−2. Which of the below is the correct 95% confidence interval for the slopeparameter? Note that the model is based on observations from 27 twins.

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

(a) 9.2076 ± 1.65 × 9.2999(b) 0.9014 ± 2.06 × 0.0963

(c) 0.9014 ± 1.96 × 0.0963

(d) 9.2076 ± 1.96 × 0.0963

n = 27 df = 27 − 2 = 25

95% : t?25 = 2.06

0.9014 ± 2.06 × 0.0963

(0.7 , 1.1)

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 54 / 56

Page 129: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Confidence interval for the slope

Remember that a confidence interval is calculated as point estimate±ME andthe degrees of freedom associated with the slope in a simple linear regressionis n−2. Which of the below is the correct 95% confidence interval for the slopeparameter? Note that the model is based on observations from 27 twins.

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

(a) 9.2076 ± 1.65 × 9.2999(b) 0.9014 ± 2.06 × 0.0963

(c) 0.9014 ± 1.96 × 0.0963

(d) 9.2076 ± 1.96 × 0.0963

n = 27 df = 27 − 2 = 25

95% : t?25 = 2.06

0.9014 ± 2.06 × 0.0963

(0.7 , 1.1)

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 54 / 56

Page 130: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Confidence interval for the slope

Remember that a confidence interval is calculated as point estimate±ME andthe degrees of freedom associated with the slope in a simple linear regressionis n−2. Which of the below is the correct 95% confidence interval for the slopeparameter? Note that the model is based on observations from 27 twins.

Estimate Std. Error t value Pr(>|t|)(Intercept) 9.2076 9.2999 0.99 0.3316

bioIQ 0.9014 0.0963 9.36 0.0000

(a) 9.2076 ± 1.65 × 9.2999(b) 0.9014 ± 2.06 × 0.0963

(c) 0.9014 ± 1.96 × 0.0963

(d) 9.2076 ± 1.96 × 0.0963

n = 27 df = 27 − 2 = 25

95% : t?25 = 2.06

0.9014 ± 2.06 × 0.0963

(0.7 , 1.1)

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 54 / 56

Page 131: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Recap

Inference for the slope for a single-predictor linear regressionmodel:

Hypothesis test:

T =b1 − null value

SEb1

df = n − 2

Confidence interval:b1 ± t?df=n−2SEb1

The null value is often 0 since we are usually checking for anyrelationship between the explanatory and the response variable.

The regression output gives b1, SEb1 , and two-tailed p-value forthe t-test for the slope where the null value is 0.

We rarely do inference on the intercept, so we’ll be focusing onthe estimates and inference for the slope.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 55 / 56

Page 132: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Recap

Inference for the slope for a single-predictor linear regressionmodel:

Hypothesis test:

T =b1 − null value

SEb1

df = n − 2

Confidence interval:b1 ± t?df=n−2SEb1

The null value is often 0 since we are usually checking for anyrelationship between the explanatory and the response variable.

The regression output gives b1, SEb1 , and two-tailed p-value forthe t-test for the slope where the null value is 0.

We rarely do inference on the intercept, so we’ll be focusing onthe estimates and inference for the slope.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 55 / 56

Page 133: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Recap

Inference for the slope for a single-predictor linear regressionmodel:

Hypothesis test:

T =b1 − null value

SEb1

df = n − 2

Confidence interval:b1 ± t?df=n−2SEb1

The null value is often 0 since we are usually checking for anyrelationship between the explanatory and the response variable.

The regression output gives b1, SEb1 , and two-tailed p-value forthe t-test for the slope where the null value is 0.

We rarely do inference on the intercept, so we’ll be focusing onthe estimates and inference for the slope.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 55 / 56

Page 134: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Recap

Inference for the slope for a single-predictor linear regressionmodel:

Hypothesis test:

T =b1 − null value

SEb1

df = n − 2

Confidence interval:b1 ± t?df=n−2SEb1

The null value is often 0 since we are usually checking for anyrelationship between the explanatory and the response variable.

The regression output gives b1, SEb1 , and two-tailed p-value forthe t-test for the slope where the null value is 0.

We rarely do inference on the intercept, so we’ll be focusing onthe estimates and inference for the slope.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 55 / 56

Page 135: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Recap

Inference for the slope for a single-predictor linear regressionmodel:

Hypothesis test:

T =b1 − null value

SEb1

df = n − 2

Confidence interval:b1 ± t?df=n−2SEb1

The null value is often 0 since we are usually checking for anyrelationship between the explanatory and the response variable.

The regression output gives b1, SEb1 , and two-tailed p-value forthe t-test for the slope where the null value is 0.

We rarely do inference on the intercept, so we’ll be focusing onthe estimates and inference for the slope.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 55 / 56

Page 136: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Recap

Inference for the slope for a single-predictor linear regressionmodel:

Hypothesis test:

T =b1 − null value

SEb1

df = n − 2

Confidence interval:b1 ± t?df=n−2SEb1

The null value is often 0 since we are usually checking for anyrelationship between the explanatory and the response variable.

The regression output gives b1, SEb1 , and two-tailed p-value forthe t-test for the slope where the null value is 0.

We rarely do inference on the intercept, so we’ll be focusing onthe estimates and inference for the slope.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 55 / 56

Page 137: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Caution

Always be aware of the type of data you’re working with: randomsample, non-random sample, or population.

Statistical inference, and the resulting p-values, are meaninglesswhen you already have population data.

If you have a sample that is non-random (biased), inference onthe results will be unreliable.

The ultimate goal is to have independent observations.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 56 / 56

Page 138: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Caution

Always be aware of the type of data you’re working with: randomsample, non-random sample, or population.

Statistical inference, and the resulting p-values, are meaninglesswhen you already have population data.

If you have a sample that is non-random (biased), inference onthe results will be unreliable.

The ultimate goal is to have independent observations.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 56 / 56

Page 139: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Caution

Always be aware of the type of data you’re working with: randomsample, non-random sample, or population.

Statistical inference, and the resulting p-values, are meaninglesswhen you already have population data.

If you have a sample that is non-random (biased), inference onthe results will be unreliable.

The ultimate goal is to have independent observations.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 56 / 56

Page 140: Chapter 7: Introduction to linear regression

Inference for linear regression CI for the slope

Caution

Always be aware of the type of data you’re working with: randomsample, non-random sample, or population.

Statistical inference, and the resulting p-values, are meaninglesswhen you already have population data.

If you have a sample that is non-random (biased), inference onthe results will be unreliable.

The ultimate goal is to have independent observations.

OpenIntro Statistics, 2nd Edition Chp 7: Intro. to linear regression 56 / 56