Multiple Regression Analysis: Inference

1 / 34

# Multiple Regression Analysis: Inference - PowerPoint PPT Presentation

Multiple Regression Analysis: Inference. Assumptions of the Classical Linear Model (CLM). Given the Gauss-Markov assumptions, OLS is BLUE. Beyond the Gauss-Markov assumptions, we need another assumption to conduct tests of hypotheses (inference).

I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.

## PowerPoint Slideshow about 'Multiple Regression Analysis: Inference' - zulema

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript

Assumptions of the Classical Linear Model (CLM)

Given the Gauss-Markov assumptions, OLS is BLUE.

Beyond the Gauss-Markov assumptions, we need another assumption to conduct tests of hypotheses (inference).

Assume that u is independent of x1, x2, …, xk and u is normally distributed with zero mean and variance σ²:

u ~ N(0, σ²).

CLM Assumptions (continued . . .)

Under CLM, OLS is BLUE; OLS is the minimum variance unbiased estimator.

y|x ~ N(ß0 + ß1x1 +…+ ßkxk, σ²)

Normal Sampling Distributions

Under the CLM assumptions, conditional on the sample values of the explanatory variables

so that is distributed normally because it is a linear combination of the right-hand side variables.

The t Test

Under the CLM assumptions, the expression

follows a t distribution (versus a standard normal distribution), because we have to estimate σ² by .

Note the degrees of freedom: n – k – 1 .

The t Test

• Knowing the sampling distribution allows us to carry out hypothesis tests.
• Example: H0: ßj = 0
• If we accept the null hypothesis, then we conclude that xj has no effect on y, controlling for other x’s.

Steps of the t Test

• Form the relevant hypothesis.
• - one-sided hypothesis
• - two-sided hypothesis
• 2. Calculate the t statistic.
• 3. Find the critical value, c.
• - Given a significance level, α, we look up the corresponding percentile in a t distribution with n – k – 1 degrees of freedom and call it c, the critical value.
• 4. Apply rejection rule to determine whether or not to accept the null hypothesis.

Types of Hypotheses and Significance Levels

Hypothesis: null vs. alternative

- one-sided H0: ßj = 0 and H1: ßj < 0 or H1: ßj >0

- two-sided H0: ßj = 0 and H1: ßj 0

Significance level (α)

- If we want to have only a 5% probability of rejecting Ho, if it really is true, then we say our significance level is 5%.

- α values are generally 0.01, 0.05, or 0.10

- α values are dictated by sample size

Critical Value c

What do you need to find c?

1. t-distribution table (Appendix Table B.3, p. 723 Hirschey

2. Significance level

3. Degrees of freedom

- n – k – 1, where n is the # of observations, k is the # of RHS variables, and 1 is for the constant.

One-Sided Alternatives

yi = ß0 + ß1x1i+ … + ßkxki+ ui

H0: ßj = 0 H1: ßj> 0

Fail to reject

reject

(1 - a)

a

c

0

Critical value c: the (1 – α)th percentile in a t-dist with n – k – 1 DF.

t-statistic:

Results: Reject H0 if t-statistic > c; fail to reject Ho if t-statistic < c

One-Sided Alternatives

yi = ß0 + ß1x1i+ … + ßkxki+ ui

H0: ßj = 0 H1: ßj < 0

Fail to reject

reject

(1 - α)

α

-c

0

Critical value c: the (1 – α)th percentile in a t-dist with n – k – 1 DF.

t-statistic:

Results: Reject Ho if t-statistic < -c; fail to reject Ho if t-statistic > -c

Two-Sided Alternative

yi = ß0 + ß1X1i+ … + ßkXki+ ui

H0: ßj = 0 H1:

fail to reject

reject

reject

(1 - α)

α/2

α /2

-c

c

0

Critical value: the (1 – α/2)th percentile in a t-dist with n – k – 1 DF.

t-statistic:

Results: Reject H0 if |t-statistic|> c; fail to reject H0 if |t-statistic|< c

Summary for H0: ßi = 0

• unless otherwise stated, the alternative is assumed to be two-sided.
• if we reject the null hypothesis, we typically say “xj is statistically significant at the α% level.”
• if we fail to reject the null hypothesis, we typically say “xj is statistically insignificant at the α% level.”

Testing Other Hypotheses

• A more general form of the t-statistic recognizes that we may want to test H0: ßj = aj
• In this case, the appropriate t-statistic is

where aj = 0 for the conventional t-test

t-Test: Example

• Tile Example
• Q = 17.513 – 0.296P + 0.0661 + 0.036A
• (-0.35) (-2.91) (2.56) (4.61)
• t-statistics are in parentheses
• Questions:
• (a) How do we calculate the standard errors?
• (b) Which coefficients are statistically different from zero?

Confidence Intervals

Another way to use classical statistical testing is to construct a confidence interval using the same critical value as was used for a two-sided test.

A (1 – α)% confidence interval is defined as

where c is the percentile

in a distribution.

Computing p-values for t Tests

An alternative to the classical approach is to ask, “what is the smallest significance level at which the null hypothesis would be rejected?”

Compute the t-statistic, and then obtain the probability of getting a larger value than this calculated value.

The p-value is this probability.

Example: Regression Relation Between Units Sold and Personal Selling expenditures for Electronic Data Processing (EDP), Inc.

• Units sold = -1292.3 + 0.09289 PSE
• (396.5) + (0.01097)
• What are the associated t-statistics for the intercept and slope parameter estimates?
• t-stat for = - 3.26 p-value 0.009
• t-stat for = 8.47 p-value 0.000
• If p-value < α, then reject H0: ßi = 0
• If p-value > α, then fail to reject H0: ßi = 0
• (c) What conclusion about the statistical significance of the estimated parameters do you reach, given these p-values?

Testing a Linear Combination of Parameter Estimates

Let’s suppose that, instead of testing whether ß1 is equal to a constant, you want to test to see if it is equal to another parameter, that is H0: ß1 = ß2.

Use the same basic procedure for forming a t-statistic.

Overall Significance

H0: ß1 = ß2 = … = ßk = 0

Use of F-statistic

F Distribution with 4 and 30 degrees of freedom (for a regression model with four X variables based on 35 observations).

The F Statistic

Reject H0 at a

significance level

if F > c

fail to reject

Appendix Tables B.2, pp.720-722. Hirschey

reject

a

(1 - a)

0

c

F

Example:

UNITSt = -117.513 – 0.296Pt + 0.036ADt + 0.006PSEt

(-0.35) (-2.91) (2.56) (4.61)

PSEt = Selling Expenses UNITSt = # of units Sold

s standard error of the regression is 123.9

R² = 0.97 n = 32 = 0.958

• Calculate the F-statistic.
• What are the degrees-of-freedom associated with the F- statistic?
• What is the cutoff value of this F-statistic when α = 0.05? When α = 0.01?

General Linear Restrictions

The basic form of the F-statistic will work for any set of linear restrictions.

First estimate the unrestricted (UR) model and then estimate the restricted (R) model.

In each case, make note of the SSE.

Test of General Linear Restrictions

- This F-statistic is measuring the relative increase in SSE, when moving from the unrestricted (UR) model to the restricted (R) model.

- q = number of restrictions

Example:

Unrestricted Model

Restricted Model (under H0); note q = 1

F-Statistic Summary

• Just as with t-statistics, p-values can be calculated by looking up the percentile in the appropriate F distribution.
• If q = 1, then F = t², and the p-values will be the same.

Summary: Inferences

• t-Test
• (a) one-sided vs. two-sided hypotheses
• (b) tests associated with a constant value
• (c) tests associated with linear combinations of parameters
• (d) p-values of t-tests
• Confidence intervals for estimated coefficients
• F-test
• p-values of F-tests