1 / 75

Regression Models

Regression Models. Chapter 4. To accompany Quantitative Analysis for Management , Tenth Edition , by Render, Stair, and Hanna Power Point slides created by Jeff Heyl. © 2009 Prentice-Hall, Inc. . Learning Objectives. After completing this chapter, students will be able to:.

bert
Download Presentation

Regression Models

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Regression Models Chapter 4 To accompanyQuantitative Analysis for Management, Tenth Edition,by Render, Stair, and Hanna Power Point slides created by Jeff Heyl © 2009 Prentice-Hall, Inc.

  2. Learning Objectives After completing this chapter, students will be able to: • Identify variables and use them in a regression model • Develop simple linear regression equations from sample data and interpret the slope and intercept • Compute the coefficient of determination and the coefficient of correlation and interpret their meanings • Interpret the F-test in a linear regression model • List the assumptions used in regression and use residual plots to identify problems

  3. Learning Objectives After completing this chapter, students will be able to: • Develop a multiple regression model and use it to predict • Use dummy variables to model categorical data • Determine which variables should be included in a multiple regression model • Transform a nonlinear function into a linear one for use in regression • Understand and avoid common mistakes made in the use of regression analysis

  4. Chapter Outline 4.1Introduction 4.2Scatter Diagrams 4.3Simple Linear Regression 4.4Measuring the Fit of the Regression Model 4.5Using Computer Software for Regression 4.6Assumptions of the Regression Model

  5. Chapter Outline 4.7 Testing the Model for Significance 4.8 Multiple Regression Analysis 4.9 Binary or Dummy Variables 4.10 Model Building 4.11 Nonlinear Regression 4.12 Cautions and Pitfalls in Regression Analysis

  6. Introduction • Regression analysis is a very valuable tool for a manager • Regression can be used to • Understand the relationship between variables • Predict the value of one variable based on another variable • Examples • Determining best location for a new store • Studying the effectiveness of advertising dollars in increasing sales volume

  7. Independent variable Independent variable Dependent variable = + Introduction • The variable to be predicted is called the dependent variable • Sometimes called the response variable • The value of this variable depends on the value of the independent variable • Sometimes called the explanatory or predictor variable

  8. Scatter Diagram • Graphing is a helpful way to investigate the relationship between variables • A scatter diagram or scatter plot is often used • The independent variable is normally plotted on the X axis • The dependent variable is normally plotted on the Y axis

  9. Triple A Construction • Triple A Construction renovates old homes • They have found that the dollar volume of renovation work is dependent on the area payroll Table 4.1

  10. 12 – 10 – 8 – 6 – 4 – 2 – 0 – Sales ($100,000) | | | | | | | | 0 1 2 3 4 5 6 7 8 Payroll ($100 million) Triple A Construction Figure 4.1

  11. Simple Linear Regression • Regression models are used to test if there is a relationship between variables (predict sales based on payroll) • There is some random error that cannot be predicted where Y = dependent variable (response) X = independent variable (predictor or explanatory) 0 = intercept (value of Y when X = 0) 1 = slope of the regression line e = random error

  12. where Y = dependent variable (response) X = independent variable (predictor or explanatory) b0 = intercept (value of Y when X = 0) b1 = slope of the regression line ^ Simple Linear Regression • True values for the slope and intercept are not known so they are estimated using sample data

  13. Triple A Construction • Triple A Construction is trying to predict sales based on area payroll Y = Sales X = Area payroll • The line chosen in Figure 4.1 is the one that minimizes the errors Error = (Actual value) – (Predicted value)

  14. Least Squares Regression Errors can be positive or negative so the average error could be zero even though individual errors could be large. Least squares regression minimizes the sum of the squared errors.

  15. Triple A Construction • For the simple linear regression model, the values of the intercept and slope can be calculated using the formulas below

  16. Triple A Construction • Regression calculations Table 4.2

  17. Therefore Triple A Construction • Regression calculations

  18. sales = 2 + 1.25(payroll) If the payroll next year is $600 million Therefore Triple A Construction • Regression calculations

  19. Measuring the Fit of the Regression Model • Regression models can be developed for any variables X and Y • How do we know the model is actually helpful in predicting Y based on X? • We could just take the average error, but the positive and negative errors would cancel each other out • Three measures of variability are • SST – Total variability about the mean • SSE – Variability about the regression line • SSR – Total variability that is explained by the model

  20. Sum of the squares total • Sum of the squared error • Sum of squares due to regression • An important relationship Measuring the Fit of the Regression Model

  21. ^ ^ ^ ^ ^ Measuring the Fit of the Regression Model Table 4.3

  22. Sum of the squares total For Triple A Construction SST = 22.5 SSE = 6.875 SSR = 15.625 • Sum of the squared error • Sum of squares due to regression • An important relationship • SSR – explained variability • SSE – unexplained variability Measuring the Fit of the Regression Model

  23. 12 – 10 – 8 – 6 – 4 – 2 – 0 – ^ Y = 2 + 1.25X ^ Y – Y Y Sales ($100,000) Y – Y Y – Y ^ | | | | | | | | 0 1 2 3 4 5 6 7 8 Payroll ($100 million) Measuring the Fit of the Regression Model Figure 4.2

  24. For Triple A Construction Coefficient of Determination • The proportion of the variability in Y explained by regression equation is called the coefficient of determination • The coefficient of determination is r2 • About 69% of the variability in Y is explained by the equation based on payroll (X)

  25. For Triple A Construction Correlation Coefficient • Thecorrelation coefficientis an expression of the strength of the linear relationship • It will always be between +1 and –1 • The correlation coefficient is r

  26. Y Y * * * * * * * * * * * * * * * * X X (a) Perfect PositiveCorrelation: r = +1 (b) PositiveCorrelation: 0 < r < 1 Y Y * * * * * * * * * * * * * * * * * * X X (c) No Correlation: r = 0 (d) Perfect Negative Correlation: r = –1 Correlation Coefficient Figure 4.3

  27. Using Computer Software for Regression Program 4.1A

  28. Using Computer Software for Regression Program 4.1B

  29. Using Computer Software for Regression Program 4.1C

  30. Using Computer Software for Regression Program 4.1D

  31. Correlation coefficient is called Multiple R in Excel Using Computer Software for Regression Program 4.1D

  32. Assumptions of the Regression Model • If we make certain assumptions about the errors in a regression model, we can perform statistical tests to determine if the model is useful • Errors are independent • Errors are normally distributed • Errors have a mean of zero • Errors have a constant variance • A plot of the residuals (errors) will often highlight any glaring violations of the assumption

  33. Error X Residual Plots • A random plot of residuals Figure 4.4A

  34. Error X Residual Plots • Nonconstant error variance • Errors increase as X increases, violating the constant variance assumption Figure 4.4B

  35. Error X Residual Plots • Nonlinear relationship • Errors consistently increasing and then consistently decreasing indicate that the model is not linear Figure 4.4C

  36. Estimating the Variance • Errors are assumed to have a constant variance ( 2), but we usually don’t know this • It can be estimated using the mean squared error (MSE), s2 where n = number of observations in the sample k = number of independent variables

  37. Estimating the Variance • For Triple A Construction • We can estimate the standard deviation, s • This is also called the standard error of the estimate or the standard deviation of the regression

  38. Testing the Model for Significance • When the sample size is too small, you can get good values for MSE and r2 even if there is no relationship between the variables • Testing the model for significance helps determine if the values are meaningful • We do this by performing a statistical hypothesis test

  39. Testing the Model for Significance • We start with the general linear model • If 1 = 0, the null hypothesis is that there is no relationship between X and Y • The alternate hypothesis is that there is a linear relationship (1≠ 0) • If the null hypothesis can be rejected, we have proven there is a relationship • We use the F statistic for this test

  40. Testing the Model for Significance • The F statistic is based on the MSE and MSR where k = number of independent variables in the model • The F statistic is • This describes an F distribution with degrees of freedom for the numerator = df1 = k degrees of freedom for the denominator = df2 = n – k – 1

  41. Testing the Model for Significance • If there is very little error, the MSE would be small and the F-statistic would be large indicating the model is useful • If the F-statistic is large, the significance level (p-value) will be low, indicating it is unlikely this would have occurred by chance • So when the F-value is large, we can reject the null hypothesis and accept that there is a linear relationship between X and Y and the values of the MSE and r2 are meaningful

  42. Steps in a Hypothesis Test • Specify null and alternative hypotheses Select the level of significance (). Common values are 0.01 and 0.05 Calculate the value of the test statistic using the formula

  43. Reject the null hypothesis if the test statistic is greater than the F-value from the table in Appendix D. Otherwise, do not reject the null hypothesis: • Reject the null hypothesis if the observed significance level, or p-value, is less than the level of significance (). Otherwise, do not reject the null hypothesis: Steps in a Hypothesis Test • Make a decision using one of the following methods

  44. Step 3. Calculate the value of the test statistic Triple A Construction Step 1. H0: 1 = 0 (no linear relationship between X and Y) H1: 1≠ 0 (linear relationship exists between X and Y) Step 2. Select  = 0.05

  45. Step 4. Reject the null hypothesis if the test statistic is greater than the F-value in Appendix D df1 = k = 1 df2 = n – k – 1 = 6 – 1 – 1 = 4 The value of F associated with a 5% level of significance and with degrees of freedom 1 and 4 is found in Appendix D F0.05,1,4 = 7.71 Fcalculated = 9.09 Reject H0 because 9.09 > 7.71 Triple A Construction

  46. 0.05 F = 7.71 9.09 Triple A Construction • We can conclude there is a statistically significant relationship between X and Y • The r2 value of 0.69 means about 69% of the variability in sales (Y) is explained by local payroll (X) Figure 4.5

  47. r2 coefficient of determination • The F-test determines whether or not there is a relationship between the variables. • r2 (coefficient of determination) is the best measure of the strength of the prediction relationship between the X and Y variables. • Values closer to 1 indicate a strong prediction relationship. • Good regression models have a low significance level for the F-test and high r2 value.

  48. Coefficient Hypotheses • Statistical tests of significance can be performed on the coefficients. • The null hypothesis is that the coefficient of X (i.e., the slope of the line) is 0 i.e., X is not useful in predicting Y • P values are the observed significance level and can be used to test the null hypothesis. • Values less than 5% negate the null hypothesis and indicate that X is useful in predicting Y • For a simple linear regression, the test of the regression coefficients gives the same information as the F-test.

  49. Analysis of Variance (ANOVA) Table • When software is used to develop a regression model, an ANOVA table is typically created that shows the observed significance level (p-value) for the calculated F value • This can be compared to the level of significance () to make a decision Table 4.4

  50. ANOVA for Triple A Construction Program 4.1D (partial) P(F > 9.0909) = 0.0394 • Because this probability is less than 0.05, we reject the null hypothesis of no linear relationship and conclude there is a linear relationship between X and Y

More Related