1 / 15

Chapter 10: Covariance and Correlation

CIS 2033 based on Dekking et al. A Modern Introduction to Probability and Statistics. 2007 Instructor Longin Jan Latecki. Chapter 10: Covariance and Correlation. As an example, take g(x, y) = xy for discrete random variables X and Y with

peeplesl
Download Presentation

Chapter 10: Covariance and Correlation

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. CIS 2033 based onDekking et al. A Modern Introduction to Probability and Statistics. 2007Instructor Longin Jan Latecki Chapter 10: Covariance and Correlation

  2. As an example, take g(x, y) = xy for discrete random variables X and Y with the joint probability distribution given in the table. The expectation of XY is computed as follows:

  3. With the rule above we can compute the expectation of a random variable X with a Bin(n,p) which can be viewed as sum of Ber(p) distributions:

  4. Proof that E[X + Y] = E[X] + E[Y]:

  5. Var(X + Y) is generally not equal to Var(X) + Var(Y)

  6. If Cov(X,Y) > 0 , then X and Y are positively correlated. If Cov(X,Y) < 0, then X and Y are negatively correlated. If Cov(X,Y) =0, then X and Y are uncorrelated. Gustavo Orellana

  7. Now let X and Y be two independent random variables. Then Cov(X, Y ) = E[XY ] − E[X]E[Y ] = 0. Hence, then X and Y are uncorrelated. We proved that if X and Y are two independent random variables, then they are uncorrelated. In general, E[XY] is NOT equal to E[X]E[Y]. INDEPENDENT VERSUS UNCORRELATED. If two random variables X and Y are independent, then X and Y are uncorrelated. The converse is not true as we will see on the next slide.

  8. Then Cov(X, Y ) = E[XY ] − E[X]E[Y ] = 0 and X and Y are uncorrelated,but they are dependent.

  9. The variance of a random variable with a Bin(n,p) distribution:

  10. The covariance changes under a change of units The covariance Cov(X,Y) may not always be suitable to express the dependence between X and Y. For this reason, there is a standardized version of the covariance called the correlation coefficient of X and Y, which remains unaffected by a change of units and, therefore, is dimensionless.

  11. Correlation coefficient is also called Pearson correlation coefficient. (from Wikipedia) Examples of scatter diagrams with different values of correlation coefficient.

  12. (from Wikipedia) Several sets of (x, y) points, with the correlation coefficient of x and y for each set. Note that the correlation reflects the non-linearity and direction of a linear relationship (top row), but not the slope of that relationship (middle), nor many aspects of nonlinear relationships (bottom). N.B.: the figure in the center has a slope of 0 but in that case the correlation coefficient is undefined because the variance of Y is zero.

More Related