1 / 84

Expectation

Expectation. for multivariate distributions. Definition. Let X 1 , X 2 , …, X n denote n jointly distributed random variable with joint density function f ( x 1 , x 2 , …, x n ) then. Example.

badru
Download Presentation

Expectation

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Expectation for multivariate distributions

  2. Definition Let X1, X2, …, Xn denote n jointly distributed random variable with joint density function f(x1, x2, …, xn ) then

  3. Example Let X, Y, Z denote 3jointly distributed random variable with joint density function then Determine E[XYZ].

  4. Solution:

  5. Some Rules for Expectation

  6. Thus you can calculate E[Xi] either from the joint distribution of X1, … , Xn or the marginal distribution of Xi. Proof:

  7. The Linearity property Proof:

  8. (The Multiplicative property)Suppose X1, … , Xq are independent of Xq+1, … , Xk then In the simple case when k = 2 if X and Y are independent

  9. Proof:

  10. Some Rules for Variance

  11. Proof Thus

  12. Note: If X and Y are independent, then

  13. Definition: For any two random variables X and Y then define the correlation coefficient rXY to be: if X and Y are independent

  14. Properties ofthe correlation coefficient rXY The converse is not necessarily true. i.e. rXY= 0 does not imply that X and Y are independent.

  15. More properties ofthe correlation coefficient rXY if there exists a and b such that where rXY = +1 if b > 0 and rXY = -1 if b< 0 Proof: Let Let for all b. Consider choosing b to minimize

  16. Consider choosing b to minimize or Since g(b) ≥ 0, then g(bmin) ≥ 0

  17. Hence g(bmin) ≥ 0 Hence

  18. or Note If and only if This will be true if i.e.

  19. Summary if there exists a and b such that where

  20. Proof Thus

  21. Some Applications (Rules of Expectation & Variance) Let X1, … , Xn be n mutually independent random variables each having mean mand standard deviation s(variance s2). Let Then

  22. Also or Thus Hence the distribution of is centered at mand becomes more and more compact about mas n increases

  23. Tchebychev’s Inequality

  24. Tchebychev’s Inequality Let X denote a random variable with mean m =E(X) and variance Var(X) = E[(X – m)2] = s2 then Note: Is called the standard deviation of X,

  25. Proof:

  26. Tchebychev’s inequality is very conservative • k =1 • k = 2 • k = 3

  27. The Law of Large Numbers

  28. The Law of Large Numbers Let X1, … , Xn be n mutually independent random variables each having mean m. Let Then for any d> 0 (no matter how small)

  29. Proof We will use Tchebychev’s inequality which states for any random variable X. Now

  30. Thus Thus

  31. A Special case Let X1, … , Xn be n mutually independent random variables each having Bernoulli distribution with parameter p. Thus the Law of Large Numbers states

  32. Thus the Law of Large Numbers states that converges to the probability of success p Some people misinterpret this to mean that if the proportion of successes is currently lower that p then the proportion of successes in the future will have to be larger than p to counter this and ensure that the Law of Large numbers holds true. Of course if in the infinite future the proportion of successes is p than this is enough to ensure that the Law of Large numbers holds true.

  33. Some more applications Rules of expectation and Rules of Variance

  34. The mean and variance of a Binomial Random variable We have already computed this by other methods: • Using the probability function p(x). • Using the moment generating function mX(t). Suppose that we have observed n independent repetitions of a Bernoulli trial. Let X1, … , Xn be n mutually independent random variables each having Bernoulli distribution with parameter pand defined by

  35. Now X = X1 + … + Xn has a Binomial distribution with parameters n and p. X is the total number of successes in the n repetitions.

  36. The mean and variance of a Hypergeometric distribution The hypergeometric distribution arises when we sample with replacement n objects from a population of N = a + b objects. The population is divided into to groups (group A and group B). Group A contains a objects while group B contains b objects Let Xdenote the number of objects in the sample of n that come from group A. The probability function of X is:

  37. Let X1, … , Xn be n random variables defined by Then Proof

  38. Therefore and

  39. Thus

  40. Also and We need to also calculate Note:

  41. Thus and Note:

  42. Thus and

  43. Thus with and

  44. Thus

  45. Thus if X has a hypergeometric distribution with parameters a, b and n then

  46. The mean and variance of a Negative Binomial distribution The Negative Binomial distribution arises when we repeat a Bernoulli trial until k successes (S) occur. Then X = the trial on which the kth success occurred. The probability function of X is: Let X1= the number of trial on which the 1st success occurred. and Xi= the number of trials after the (i -1)stsuccess on which the ithsuccess occurred (i ≥ 2)

  47. Then X = X1 + … + Xk and X1, … , Xk are mutually independent Xieach have a geometric distribution with parameter p.

More Related