monte carlo n.
Download
Skip this Video
Loading SlideShow in 5 Seconds..
第九章 Monte Carlo 积分 PowerPoint Presentation
Download Presentation
第九章 Monte Carlo 积分

Loading in 2 Seconds...

play fullscreen
1 / 40

第九章 Monte Carlo 积分 - PowerPoint PPT Presentation


  • 647 Views
  • Uploaded on

第九章 Monte Carlo 积分. 第九章 Monte Carlo 积分. Monte Carlo 法的重要应用领域之一:计算积分和多重积分. 适用于求解:. 被积函数、积分边界复杂,难以用解析方法或一般的数值方法求解; 被积函数的具体形式未知,只知道由模拟返回的函数值。. 本章内容:. 用 Monte Carlo 法求定积分的几种方法: 均匀投点法、期望值估计法、重要抽样法、半解析法、 …. 第九章 Monte Carlo 积分. Goal: Evaluate an integral:. Why use random methods?.

loader
I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.
capcha
Download Presentation

第九章 Monte Carlo 积分


An Image/Link below is provided (as is) to download presentation

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.


- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
    Presentation Transcript
    1. 第九章 Monte Carlo积分

    2. 第九章 Monte Carlo积分 Monte Carlo法的重要应用领域之一:计算积分和多重积分 适用于求解: • 被积函数、积分边界复杂,难以用解析方法或一般的数值方法求解; • 被积函数的具体形式未知,只知道由模拟返回的函数值。 本章内容: 用Monte Carlo法求定积分的几种方法: 均匀投点法、期望值估计法、重要抽样法、半解析法、…

    3. 第九章 Monte Carlo积分 Goal: Evaluate an integral: Why use random methods? Computation by “deterministic quadrature” can become expensive and inaccurate. • grid points add up quickly in high dimensions • bad choices of grid may misrepresent g(x)

    4. 第九章 Monte Carlo积分 • Monte Carlo method can be used to compute integral of any dimension d (d-fold integrals) • Error comparison of d-fold integrals approximating the integral of a function f using quadratic polynomials • Simpson’s rule,… • Monte Carlo method purely statistical, not rely on the dimension ! • Monte Carlo method WINS, when d >> 3

    5. 第九章 Monte Carlo积分 • Hit-or-Miss Method • Sample Mean Method • Variance Reduction Technique • Variance Reduction using Rejection Technique • Importance Sampling Method

    6. Hit-or-Miss Method • Evaluation of a definite integral h X X X X X X O • Probability that a random point reside inside the area O O O O O O a b • N : Total number of points • M : points that reside inside the region

    7. The probability that we are below the curve is So, if we can estimate p, we can estimate I: where is our estimate of p Sample uniformly from the rectangular region Hit-or-Miss Method [a,b]x[0,h]

    8. let Hit-or-Miss Method We can easily estimate p: • throw N “uniform darts” at the rectangle • let M be the number of times you end up under the curve y=g(x)

    9. Hit-or-Miss Method Start Set N : large integer M = 0 h X X X X X Choose a point x in [a,b] X Loop N times O O O O O Choose a point y in [0,h] O O a b if [x,y] reside inside then M = M+1 I = (b-a) h (M/N) End

    10. Hit-or-Miss Method • Error Analysis of the Hit-or-Miss Method • It is important to know how accurate the result of simulations are note that M is binomial(M,p)

    11. 第九章 Monte Carlo积分 • Hit-or-Miss Method • Sample Mean Method • Variance Reduction Technique • Variance Reduction using Rejection Technique • Importance Sampling Method

    12. Sample Mean Method Start Set N : large integer s1 = 0, s2 = 0 xn = (b-a) un + a Loop N times yn = r(xn) s1 = s1 + yn , s2 = s2 + yn2 Estimate mean m’=s1/N Estimate variance V’ = s2/N – m’2 End

    13. Write this as: where X~unif(a,b) Sample Mean Method

    14. Sample Mean Method where X~unif(a,b) So, we will estimate I by estimating E[g(X)] with where X1, X2, …, Xn is a random sample from the uniform(a,b) distribution.

    15. write this as where X~unif(0,3) Example: Sample Mean Method (we know that the answer is e3-1 19.08554)

    16. estimate this with where X1, X2, …, Xn are n independent unif(0,3)’s. Sample Mean Method • write this as where X~unif(0,3)

    17. Simulation Results: true = 19.08554, n=100,000 Sample Mean Method Simulation 1 19.10724 2 19.08260 3 18.97227 4 19.06814 5 19.13261

    18. Don’t ever give an estimate without a confidence interval! Sample Mean Method This estimator is “unbiased”:

    19. Sample Mean Method

    20. Sample Mean Method • an approximation

    21. Then Sample Mean Method • X1, X2, …, Xn iid -> g(X1), g(X2), …, g(Xn) iid • Let Yi=g(Xi) for i=1,2,…,n and we can once again invoke the CLT.

    22. So, a confidence interval for I is roughly given by but since we don’t know , we’ll have to be content with the further approximation: For n “large enough” (n>30), Sample Mean Method

    23. where X~exp(rate=2). By the way… Sample Mean Method No one ever said that you have to use the uniform distribution Example:

    24. Let be the hit-and-miss estimator of I • Let be the sample mean estimator of I Then Comparison of Hit-and-Miss and Sample Mean Monte Carlo Sample Mean Method

    25. Comparison of Hit-and-Miss and Sample Mean Monte Carlo Sample Mean Method Sample mean Monte Carlo is generally preferred over Hit-and-Miss Monte Carlo because: • the estimator from SMMC has lower variance • SMMC does not require a non-negative integrand (or adjustments) • H&M MC requires that you be able to put g(x) in a “box”, so you need to figure out the max value of g(x) over [a,b] and you need to be integrating over a finite integral.

    26. 2.1 Variance Reduction Technique - Introduction

    27. 第九章 Monte Carlo积分 • Hit-or-Miss Method • Sample Mean Method • Variance Reduction Technique • Variance Reduction using Rejection Technique • Importance Sampling Method

    28. Introduction Variance Reduction Technique • Monte Carlo Method and Sampling Distribution • Monte Carlo Method : Take values from random sample • From central limit theorem, • 3s rule • Most probable error • Important characteristics

    29. Introduction Variance Reduction Technique • Reducing error • *100 samples reduces the error order of 10 • Reducing variance  Variance Reduction Technique • The value of variance is closely related to how samples are taken • Unbiased sampling • Biased sampling • More points are taken in important parts of the population

    30. Motivation Variance Reduction Technique • If we are using sample-mean Monte Carlo Method • Variance depends very much on the behavior of r(x) • r(x) varies little  variance is small • r(x) = const  variance=0 • Evaluation of a integral • Near minimum points  contribute less to the summation • Near maximum points  contribute more to the summation • More points are sampled near the peak ”importance sampling strategy”

    31. 第九章 Monte Carlo积分 • 1.2 Hit-or-Miss Method • 1.3 Sample Mean Method • 2.1 Variance Reduction Technique • 2.3 Variance Reduction using Rejection Technique • 2.4 Importance Sampling Method

    32. Variance Reduction Technique • Variance Reduction for Hit-or-Miss method • In the domain [a,b] choose a comparison function w(x) r(x) X X X X X O O O O O O O a b • Points are generated on the area under w(x) function • Random variable that follows distribution w(x)

    33. Variance Reduction Technique Points lying above r(x) is rejected w(x) r(x) X X X X X O O O O O O O a b

    34. Variance Reduction Technique • Error Analysis Hit or Miss method Error reduction

    35. Variance Reduction Technique Start Set N : large integer w(x) r(x) X X N’ = 0 X X X O Generate u1, x= W-1(Au1) O Loop N times O O O O O Generate u2, y=u2 w(x) a b If y<= f(x) accept value N’ = N’+1 Else : reject value I = (b-a) h (N’/N) End

    36. 第九章 Monte Carlo积分 • 1.2 Hit-or-Miss Method • 1.3 Sample Mean Method • 2.1 Variance Reduction Technique • 2.3 Variance Reduction using Rejection Technique • 2.4 Importance Sampling Method

    37. Importance Sampling Method • Basic idea • Put more points near maximum • Put less points near minimum • F(x) : transformation function (or weight function_

    38. Importance Sampling Method if we choose f(x) = cr(x) ,then variance will be small The magnitude of error depends on the choice of f(x)

    39. Importance Sampling Method • Estimate of error

    40. Importance Sampling Method Start Set N : large integer s1 = 0 , s2 = 0 Generate xnaccording to f(x) Loop N times gn = r (xn) / f ( xn) Addgn to s1 Add gn to s2 I ‘ =s1/N , V’=s2/N-I’2 End