1 / 37

ROHANA BINTI ABDUL HAMID INSTITUT E FOR ENGINEERING MATHEMATICS (IMK) UNIVERSITI MALAYSIA PERLIS

EQT 272 PROBABILITY AND STATISTICS. ROHANA BINTI ABDUL HAMID INSTITUT E FOR ENGINEERING MATHEMATICS (IMK) UNIVERSITI MALAYSIA PERLIS. Free Powerpoint Templates. INTRODUCTION TO LINEAR REGRESSION. CHAPTER 5:. 5.1 INTRODUCTION. 5.2 SCATTER PLOTS. 5.3 LINEAR REGRESSION MODEL.

ivo
Download Presentation

ROHANA BINTI ABDUL HAMID INSTITUT E FOR ENGINEERING MATHEMATICS (IMK) UNIVERSITI MALAYSIA PERLIS

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. EQT 272 PROBABILITY AND STATISTICS ROHANA BINTI ABDUL HAMID INSTITUT E FOR ENGINEERING MATHEMATICS (IMK) UNIVERSITI MALAYSIA PERLIS Free Powerpoint Templates

  2. INTRODUCTION TO LINEAR REGRESSION CHAPTER 5:

  3. 5.1 INTRODUCTION 5.2 SCATTER PLOTS 5.3 LINEAR REGRESSION MODEL 5.4 LEAST SQUARE METHOD 5.5 COEFFICIENT DETERMINATION CHAPTER OUTLINE: 5.6 CORRELATION 5.7 TEST OF SIGNIFICANCE 5.8 ANALYSIS OF VARIANCE (ANOVA)

  4. 5.1 INTRODUCTION TO REGRESSION • Regression – is a statistical procedure for establishing the r/ship between 2 or more variables. • This is done by fitting a linear equation to the observed data. • The regression line is then used by the researcher to see the trend and make prediction of values for the data. • There are 2 types of relationship: • Simple ( 2 variables) • Multiple (more than 2 variables)

  5. THE SIMPLE LINEAR REGRESSION MODEL • is an equation that describes a dependent variable (Y) in terms of an independent variable (X) plus random error where, = intercept of the line with the Y-axis = slope of the line = random error • Random error, is the difference of data point from the deterministic value. • This regression line is estimated from the data collected by fitting a straight line to the data set and getting the equation of the straight line,

  6. Example 5.1: 1) A nutritionist studying weight loss programs might wants to find out if reducing intake of carbohydrate can help a person reduce weight. a) X is the carbohydrate intake (independent variable). b) Y is the weight (dependent variable). 2) An entrepreneur might want to know whether increasing the cost of packaging his new product will have an effect on the sales volume. a) X is cost b) Y is sales volume

  7. 5.2 SCATTER PLOTS • A scatter plot is a graph or ordered pairs (x,y). • The purpose of scatter plot – to describe the nature of the relationships between independent variable, X and dependent variable, Y in visual way. • The independent variable, x is plotted on the horizontal axis and the dependent variable, y is plotted on the vertical axis.

  8. E(y) x SCATTER DIAGRAM • Positive Linear Relationship Regression line Intercept b0 Slope b1 is positive

  9. E(y) x SCATTER DIAGRAM • Negative Linear Relationship Intercept b0 Regression line Slope b1 is negative

  10. E(y) x SCATTER DIAGRAM • No Relationship Intercept b0 Regression line Slope b1 is 0

  11. 5.3 LINEAR REGRESSION MODEL • A linear regression can be develop by freehand plot of the data. Example 5.2: The given table contains values for 2 variables, X and Y. Plot the given data and make a freehand estimated regression line.

  12. 5.4 LEAST SQUARES METHOD • The least squares method is commonly used to determine values for and that ensure a best fit for the estimated regression line to the sample data points • The straight line fitted to the data set is the line:

  13. LEAST SQUARES METHOD Theorem 10.1: • Given the sample data , the coefficients of the least squares line are: • y-Intercept for the Estimated Regression Equation, • and are the mean of x and y respectively.

  14. LEAST SQUARES METHOD ii) Slope for the Estimated Regression Equation, Where,

  15. LEAST SQUARES METHOD • Given any value of the predicted value of the dependent variable , can be found by substituting into the equation

  16. Example 5.3: Students score in history The data below represent scores obtained by ten primary school students before and after they were taken on a tour to the museum (which is supposed to increase their interest in history) • Fit a linear regression model with “before” as the explanatory variable and “after” as the dependent variable. • Predict the score a student would obtain “after” if he scored 60 marks “before”.

  17. 5.5 COEFFICIENT OF DETERMINATION( ) • The coefficient of determination is a measure of the variation of the dependent variable (Y) that is explained by the regression line and the independent variable (X). • The symbol for the coefficient of determination is or . • If =0.90, then =0.81. It means that 81% of the variation in the dependent variable (Y) is accounted for by the variations in the independent variable (X). • The rest of the variation, 0.19 or 19%, is unexplained and called the coefficient of nondetermination. • Formula for the coefficient of nondetermination is

  18. COEFFICIENT OF DETERMINATION( ) SST = SSR + SSE • Relationship Among SST, SSR, SSE where: SST = total sum of squares SSR = sum of squares due to regression SSE = sum of squares due to error • The coefficient of determination is: where: SSR = sum of squares due to regression SST = total sum of squares

  19. Example 5.4 • If =0.919, find the value for and explain the value. Solution : = 0.84. It means that 84% of the variation in the dependent variable (Y) is explained by the variations in the independent variable (X).

  20. 5.6 CORRELATION (r) • Correlation measures the strength of a linear relationship between the two variables. • Also known as Pearson’s product moment coefficient of correlation. • The symbol for the sample coefficient of correlation is , population . • Formula :

  21. Properties of : • Values of close to 1 implies there is a strong positive linear relationship between x and y. • Values of close to -1 implies there is a strong negative linear relationship between x and y. • Values of close to 0implies little or no linear relationship between x and y.

  22. Refer Example 5.3: Students score in history c)Calculate the value of r and interpret its meaning Solution: Thus, there is a strong positive linear relationship between score obtain before (x) and after (y).

  23. 5.7 TEST OF SIGNIFICANCE • To determine whether X provides information in predicting Y, we proceed with testing the hypothesis. • Two test are commonly used: i) ii) t Test F Test

  24. 1) t-Test 1. Determine the hypotheses. ( no linear r/ship) (exist linear r/ship) 2. Compute Critical Value/ level of significance. / p-value 3. Compute the test statistic.

  25. 1) t-Test 4. Determine the Rejection Rule. Reject H0 if : t< -or t> p-value <a 5.Conclusion. There is a significant relationship between variable X and Y.

  26. 2) F-Test 1. Determine the hypotheses. ( no linear r/ship) (exist linear r/ship) 2. Specify the level of significance. Fawith degree of freedom (df) in the numerator (1) and degrees of freedom (df) in the denominator (n-2) 3. Compute the test statistic. F = MSR/MSE 4. Determine the Rejection Rule. Reject H0 if : p-value <a F test >

  27. 2) F-Test 5.Conclusion. There is a significant relationship between variable X and Y.

  28. Refer Example 5.3: Students score in history d) Test to determine if their scores before and after the trip is related. Use a=0.05 Solution: 1. ( no linear r/ship) (exist linear r/ship) 2. 3.

  29. 4. Rejection Rule: 5. Conclusion: Thus, we reject H0. The score before (x) is linear relationship to the score after (y) the trip.

  30. 5.8 ANALYSIS OF VARIANCE (ANOVA) • The value of the test statistic F for an ANOVA test is calculated as: • F=MSR • MSE • To calculate MSR and MSE, first compute the regression sum of squares (SSR) and the error sum of squares (SSE).

  31. ANALYSIS OF VARIANCE (ANOVA) • General form of ANOVA table: • ANOVA Test • 1) Hypothesis: • 2) Select the distribution to use: F-distribution • 3) Calculate the value of the test statistic: F • 4) Determine rejection and non rejection regions: • 5) Make a decision: Reject Ho/ accept H0

  32. Example 5.5 The manufacturer of Cardio Glide exercise equipment wants to study the relationship between the number of months since the glide was purchased and the length of time the equipment was used last week. • Determine the regression equation. • At , test whether there is a linear relationship between the variables

  33. Solution (1): Regression equation:

  34. Solution (2): • Hypothesis: • F-distribution table: • Test Statistic: F = MSR/MSE = 17.303 or using p-value approach: significant value =0.003 • Rejection region: Since F statistic > F table (17.303>11.2586 ), we reject H0 or since p-value (0.003 < 0.01 )we reject H0 5) Thus, there is a linear relationship between the variables (month X and hours Y).

More Related