1 / 33

Main Effects Screening: A Distributed Continuous Quality Assurance Process for Monitoring Performance Degradation in Evo

Main Effects Screening: A Distributed Continuous Quality Assurance Process for Monitoring Performance Degradation in Evolving Software Systems. Cemal Yilmaz, Arvind S. Krishnaz, Atif Memon, Adam Porter, Douglas C. Schmidt, Aniruddha Gokhale, Balachandran Natarajan

winka
Download Presentation

Main Effects Screening: A Distributed Continuous Quality Assurance Process for Monitoring Performance Degradation in Evo

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Main Effects Screening: A Distributed Continuous Quality Assurance Process for Monitoring Performance Degradation in Evolving Software Systems Cemal Yilmaz, Arvind S. Krishnaz, Atif Memon, Adam Porter, Douglas C. Schmidt, Aniruddha Gokhale, Balachandran Natarajan Presented By: Walaa El-Din M. Moustafa

  2. Objective • As software systems change, developers often run regression tests to detect unintended functional side effects. • QA efforts can be confounded by the enormous configuration space. • Time and resource constraints severely limit the number of configurations that can be examined.

  3. Main Effects Screening • Screening Designs are • Highly economical • Can reveal important low order effects that strongly affect performance.

  4. Low Order Effects • First-, second-, or third-order effects. • nth-order effect is an effect caused by the simultaneous interaction of n factors.

  5. Low Order Effects Example • For certain web server applications: • A 1st-order effect might be that performance slows considerably when logging is turned on and another might be that it also slows when few server threads are used. • A 2ndorder effect involves the interaction of two options, e.g., web server performance may slow down when caching is turned off and the server performs blocking reads.

  6. Factorial Designs • Full factorial design involving k binary factors. Such a design exhaustively tests all combinations. • Fractional factorial designs. These designs use only a carefully selected fraction (such as 1/2 or 1/4) of a full factorial design.

  7. Fractional Factorial Designs • Save money • Do so by giving up the ability to measure some higher-order effects.

  8. Fractional Factorial Design Example • To show how screening designs are computed, we present a hypothetical example of a performance-intensive software system: • 4 binary configuration options, A through D. • No inter-option constraints. • The full configuration space therefore has 24 = 16 configurations.

  9. Fractional Factorial Design Example • 8 configurations. • Will focus only on 1st-order effects. • This design is referred to as a 24 -1 design, where 4 refers to the total number of options we will examine and the -1 (2-1 = 1/2) indicates the fraction of the full factorial over which we will collect data.

  10. Fractional Factorial Design Example • We can extend the design and estimate the effect of option D without going to a 24 full factorial design.

  11. Fractional Factorial Design Example • Design Generator: • Specifies the aliasing patterns used to build the design • For this example, we select the design generator D = ABC.

  12. Fractional Factorial Design Example

  13. Fractional Factorial Design Example • The design we described above is a resolution IV design. • In resolution R designs, no effects involving i factors are aliased with effects involving less than R - i factors.

  14. Fractional Factorial Design Example • Run the experiment • For binary options (with settings - or +), the main effect of option A, ME(A), isME(A) = z(A-) - z(A+)where z(A-) and z(A+) are the mean values of the observed data over all runs where option A is (-) and where option A is (+), respectively. • If desired, 2nd order effects can be calculated in a similar way. The interaction effect of option A and B, INT(A, B) is:= ½ (ME(B|A+) - ME(B|A-)) = ½ (ME(A|B+) - ME(A|B-))

  15. The Experiment • ACE v5.4 + TAO v1.4 + CIAO v0.4. • 14 binary run-time options 214 = 16,384 different configurations. • For each task measure: • Message latency. • Throughput between the client and the server.

  16. The Experiment

  17. The Full Data Set

  18. Evaluating Screening Designs • Can screening designs correctly identify the important options discovered during the analysis of the full data set?

  19. Evaluating Screening Designs • Scr32 is a 214 - 9 IV with design generators F = ABC, G = ABD, H = ACD, I = BCD, J = ABE, K = ACE, L = BCE, M = ADE, N = BDE. • Scr64 is a 214 - 8 IV with design generators G = ABC, H = ABD, I = ABE, J = ACDE, K = ABF, L = ACDF, M = ACEF, N = ADEF. • Scr128 is a 214 - 7 IV with design generators H = ABC, I = ABDE, J = ABDF, K = ACEF, L = ACDG, M = ABEFG, N = BCDEFG.

  20. Evaluating Screening Designs

  21. Evaluating Screening Designs

  22. Evaluating Screening Designs • Screening designs can detect important options at a large fraction of the cost of exhaustive testing. • The smaller the effect, the larger the run size needed to identify it.

  23. Estimating Performance with Screening Suites • The estimates are generated by examining all combinations of the most important options, while defaulting the settings of the unimportant options • Make the estimates based on benchmarking either 4 (all combinations of options B and J) or 32 (all combinations of options B, J, C, I, and F) configurations.

  24. Estimating Performance with Screening Suites

  25. Estimating Performance with Screening Suites • The distributions of the top-5 and top-2 screening suites closely track the overall performance data.

  26. Screening Suites vs. Random Sampling

  27. Screening Suites vs. Random Sampling • These graphs suggest the obvious weakness of random sampling.

  28. Dealing with Evolving Systems • Detect performance degradations in evolving systems quickly.

  29. Dealing with Evolving Systems • Developer records and problem reports indicate that problems were noticed on 12/14/03.

  30. Higher­Order Effects • Resolution VI design • Increased the run size to 2,048 to capture the second-order effects.

  31. Higher­Order Effects

  32. Higher­Order Effects • The important interaction effects involve only options that are already considered important by themselves • The screening design correctly identifies the 5 most important pairwise interactions at 1/8th the cost of exhaustive testing.

  33. Thank you

More Related