1 / 38

Moving Adaptation into Individual Optimizations

Musings on a decade or more of work in adaptive compilation. Industrial. Moving Adaptation into Individual Optimizations. Keith D. Cooper Doerr Professor of Computational Engineering Department of Computer Science Rice University Houston, Texas, USA.

rolandv
Download Presentation

Moving Adaptation into Individual Optimizations

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Musings on a decade or more of work in adaptive compilation Industrial Moving Adaptation into IndividualOptimizations Keith D. Cooper Doerr Professor of Computational Engineering Department of Computer Science Rice University Houston, Texas, USA This talk presents work sponsored under a number of research contracts and grants over a 10-year period, including NSF grants CCR-0205303, CNS-0615180, and CCF-0541165, the DOE-funded Los Alamos Computer Science Institute, and the PACE Project, funded by DARPA through AFRL Contract FA8650-09-C-7915. The opinions and findings in this talk do not necessarily reflect the views of either the United States Government or of Rice University. Slides posted at http://www.cs.rice.edu/~keith/Smart2010 Posted version has a list of references at the end.

  2. Road Map • Background & Context • Classification of Adaptive Schemes • Brief review of adaptive behavior in compilers • Barriers to adoption in “real” compilers • Rough classification of adaptive schemes • Compile time adaptation, runtime parameterization, runtime recompilation • Enticing problems? • Barriers to adoption? SMART 2010 Keynote

  3. Brief History of Adaptive Behavior in Compilers Bergner built on this idea, as did Simpson. Background Bibliography is on slide number 31 Prior to late 1980s, compilers operated on fixed strategies One-size fits all or a couple of command-line flags (-g, -O1, -O2, -O3, …) Those compiler writers may have been a lot smarter than we are Early glimpses of adaptation Bernstein et al.’s “best of three spilling” Motwani et al.s’ “alpha-beta tuning” for allocation & scheduling Granston & Holler’s system to recommend compiler options Adaptation became a minor industry this decade Many submissions, fewer published papers Hard to capture full set of data & experience in a conference paper SMART 2010 Keynote

  4. Brief History of Adaptation in Compilers Background After twenty years, where are we? • Techniques are still (largely) confined to academic sandboxes • Influenced the design of major systems • LLVM, Phoenix built with intent of allowing adaptive choices • Few deployed “real” compilers that use these ideas • Much of the low-hanging fruit has been picked • Optimization choice, sequence ordering, command-line flags, … • Genetic algorithms, genetic programming, heuristic search, greedy algorithms Two major questions for this talk • How do we select problems for future work? • How do we move these techniques into mainstream use? The talk will take a somewhat circuitous path to reach these questions. SMART 2010 Keynote

  5. The Compile / Execute Cycle App Source Static Compiler Source-Level Optimization Application Developer … … Runtime Perf Monitor Target-Level Optimization Tuning JIT Opt. a.out Context End user compiles & runs application Runtime system works to measure & improve performance End user interprets feedback & changes code to improve it Compiler & runtime are opaque boxes to the application developer Compiler features multiple levels of “static” optimization SMART 2010 Keynote

  6. Compile / Execute with Adaptive Techniques App Source Application Developer Static Compiler Source-Level Optimization Adaptive Controller … … Runtime Perf Monitor Target-Level Optimization Tuning JIT Opt. a.out Context • With adaptation, some change is automated • Both compile-time and runtime adaptation • Explicit control & coordination • Feedback occurs in both large and small cycles • Several time scales for adaptation • End user retains the ability to intervene • Compiler structure looks much the same as before SMART 2010 Keynote

  7. Abstracting Away the Details App Source Static Compiler Adaptive Controller Source-Level Optimization … … Runtime Perf Monitor Target-Level Optimization Tuning JIT Opt. a.out Context Ignore the flow of code through the system • Feedback relationships become clear • Many opportunities for feedback & adaptation Good news: more theses are left in this area. SMART 2010 Keynote

  8. Abstracting Away the Details Granston & Holler App Source Static Compiler Adaptive Controller Source-Level Optimization … … Runtime Perf Monitor Target-Level Optimization Tuning JIT Opt. a.out Context Major adaptation cycle • Compile, execute, monitor, analyze, and recompile • Typical adaptations • Change compilation strategies • Change optimizations, sequences, or code shape (e.g., inlining) Behavior evolves slowly over many cycles SMART 2010 Keynote

  9. Abstracting Away the Details Bernstein et al. App Source Static Compiler Adaptive Controller Source-Level Optimization … … Runtime Perf Monitor Target-Level Optimization Tuning JIT Opt. a.out Context Minor adaptation cycles • Repeated multiple times within a single major cycle • Typical adaptations • Modify a heuristic • Modify some threshold • Try several approaches & keep best result Behavior evolves rapidly; may repeat in major cycle SMART 2010 Keynote

  10. Abstracting Away the Details App Source Motwani et al. App Source App Source Static Compiler Adaptive Controller Source-Level Optimization … … Runtime Perf Monitor Target-Level Optimization Tuning JIT Opt. a.out Context Long term learning • Involves multiple runs & multiple applications • Typical adaptations • Refine models of system performance & optimization effects • Recommend (prescribe) optimization strategies & policies • Good subject for a talk; I am the wrong speaker Slow evolution over many compile/execute cycles SMART 2010 Keynote

  11. Defining Internal Adaptation Internal Adaptation Focus on adaptation that changes the behavior of one transformation Ignore, for now, techniques that treat the compiler or its components as black boxes • Optimization choice might pick one of LCM, DVNT, GVN • External or structural adaption • Not a problem for today’s talk • Internal adaptation might parameterize the placement algorithm in LCM • Modify the definition of earliest placement with a tuning parameter • Search for the parameter value that produces the best code Less work has been done on this kind of adaptation SMART 2010 Keynote

  12. Defining Internal Adaptation Let’s examine a number of examples of internal adaptation and try to generalize from them. Internal Adaptation External adaptation changes the way that the compiler runs • Connect the components in different ways • Schielke, Kulkarni derive good optimization sequences for specific code • Manipulate pre-planned options • Triantafyllis chooses from a set of preplanned optimization sequences • Granston & Holler, Cavazos find a set of appropriate compiler flags Internal adaptation changes the behavior of a single pass in the compiler • Evolve a heuristic • Motwani et al. derived new scheduling heuristics • Choose among preplanned options • Bernstein et al. choose best spill heuristic on a procedure-by-procedure basis • Derive internal control parameters • Liu’s adaptive copy coalescing finds a threshold value to throttle coalescing SMART 2010 Keynote

  13. Inline Substitution abcd… call foo(…) efgh … foo{ } becomes abcd … efgh … Examples Inline Substitution • Simple transformation, complex decision problem • Choice at each call site, choices interact • Profitability depends on how the code optimizes • Potential to improve or to degrade performance • Classic examples that produce integer factor speedups • Classic examples that produce significant slowdowns • State of the art is one of two cases • Programmer makes the decision • Compiler applies complex heuristic formula at each call site Textbook example of adaptation within a single optimization Weakness as an example is the slowness of evaluation SMART 2010 Keynote

  14. Inline Substitution App Source App Source Training Production Adaptive Controller App specific CNF expression Inliner Inliner Compiler Run & Measure Compiler Fast Waterman’s solution • Fast inline substitution pass controlled by a heuristic • Feedback-driven system to derive program-specific heuristics • Used a classic compile-run feedback loop to search a huge parameter space • Derive a program-specific heuristic & then use it for subsequent compiles • Consistently beat one-size fits all heuristics SMART 2010 Keynote

  15. Choosing Tile Sizes in the MIPS Compiler Examples The MIPS compiler did a poor job of choosing tile sizes on some codes • Command-line parameter to override its choice • Dasgupta showed that a simple binary search on tile size often beat MIPS • Tile size was used on all loops in the compilation unit (poor parameterization) Similar Setup to Inline Substitution • Need full compile and execute cycle to evaluate each choice • Treated compiler as a black box, so no other choice • Loop-by-loop tile sizes would be better, but would still need the major cycle • Derived compilation-unit specific tile size SMART 2010 Keynote

  16. Lazy Code Motion (negative example) Limit motion Limit replacement profit cost Examples LCM combines code motion & redundancy elimination • Effective technique that improves most programs • Complex transformation, but an easy decision procedure Adaptive LCM? • One downside of LCM is that it can increase register pressure • Introducing spills degrades performance • Might build an adaptive LCM moderated by register pressure • Information on register pressure is not available when LCM runs • Ratio of is too low to justify repeated compilations Many other optimizations are equally unsuited to adaptation • Strength reduction, constant propagation, dead code elimination, … SMART 2010 Keynote

  17. Adaptive Register Coalescing Competing ideas about how to make the decision Examples In allocation & coalescing, k is the number of registers. A node has “high degree” if it has ≥ k neighbors. The Problem • Eliminate all unneeded register-to-register copies (e.g, ri→rj ) • Copies occupy time and code space, also slow down compiler • Simple optimization, simple decision procedure In the context of a Chaitin-style graph-coloring register allocator: • If ri→rj and ri does not interfere with rj, then ri and rj can be coalesced • Combining them can increase register pressure (& lead to more spills) • Popular technique is conservative coalescing • Only combine ri and rj if result has < k neighbors of high degree • Conservative coalescing is safe, but leaves too many unneeded copies Coalescing looks like a good candidate for internal adaptation SMART 2010 Keynote

  18. Adaptive Register Coalescing Examples Liu’s Solution • Perform conservative coalescing, but adaptively vary k • His allocator finds a k such that minimizes copies without adding spill code • Distinct k values for integer & floating-point registers • Range of values for k is limited ( # registers ≤ k ≤ max degree in graph ) • Feedback is quick, accurate, and introspective • The allocator measures its own behavior & uses that metric for feedback Results • Fewer copies, no more spills, relatively inexpensive at compile time • On x86 in LLVM, best value of k is typically 5 or 10 more than # registers SMART 2010 Keynote

  19. Problems Similar to Adaptive Coalescing Examples Adaptive Coloring (for allocation) • On x86, overlapping register classes make graph coloring allocators less effective • Liu made Briggs adaptive by varying k in the coloring phase • Separate k for floating-point and integer (as well as for coalescing) • Introspective evaluation based on estimated spill cost • Preliminary numbers suggest 15 to 20% speedup on x86 in LLVM Software Pipelining (a la Lam) • Uses a simple parameter sweep to find the initiation interval • Bounds the space with lower-bound estimates • Introspective measure of success – modulo scheduler fails or succeeds • Widely used technique SMART 2010 Keynote

  20. Adaptive Reassociation (speculation) + + + + + + b a c a a b c c b Examples The compiler can use commutativity, associativity, and distributivity to reorder expressions so that they optimize in different ways • Expose different sets of expressions • Optimizations need different orders • Choosing a good order for an expression tree depends on the optimizations & the specific performance problems • Daunting number of ways to rearrange an expression (not yes/no decision) Properties & Strategy • Needs major cycle to assess impact on other optimizations • Can rely on introspective evaluation to decrease the cost SMART 2010 Keynote

  21. Adaptive Reassociation (speculation) App Source Static Compiler AdaptiveReassoc’n Other opts Other opts Back End a.out Examples Strategy • Perform reassociation early • For each expression, look @ context & transform • Use a closed-form representation of heuristic & evolve parameters on loop-by-loop basis • Measure code quality in the back end • Operation counts in loops, spills, schedule density Goal • Quick feedback cycle for rapid evolution • Application-specific decision making SMART 2010 Keynote

  22. Summarizing the Examples SMART 2010 Keynote

  23. Summarizing the Examples Major feedback cycle using execution time metric ⇒ slow evolution of solution (1 step per compile) ⇒ clever heavy weight searches to minimize cost SMART 2010 Keynote

  24. Summarizing the Examples Minor feedback cycle using introspective metrics ⇒ rapid evolution of solution (many steps per compile) ⇒ simple lightweight searches, such as parameter sweeps SMART 2010 Keynote

  25. Summarizing the Examples Major feedback cycle using execution time metric ⇒ slow evolution of solution (1 step per compile) ⇒ cost / benefit ratio is too high LCM does well enough on its own. SMART 2010 Keynote

  26. Summarizing the Examples Major feedback cycle using introspective metrics ⇒ choose loop-nest specific solutions ⇒ large expected benefit SMART 2010 Keynote

  27. Looking for Candidates for Internal Adaptation For success, I think that we need to find optimizations that have: • Complex, inter-related decision problems • If each decision is O(1), the problem may be too easy • If context plays a role in the good decision, that’s a good sign • Algorithm that can be parameterized in a clean fashion • Parameter should be tunable, preferably over a small range • We have ignored the issue of parameter choice for too long • Simple, well-defined metrics • Must relate performance to the parameter space • Ideally, should be obtainable with introspection (avoid execution to lower costs) SMART 2010 Keynote

  28. Barriers to Adoption of Adaptive Techniques Industrialization Usability Issues • Where does the system store all those annotations and history? • How does the user specify her objective function? (speed, space, energy, … ?) Mitigating and Explaining the Long Compile Times • These techniques are expensive—especially major cycle examples • Can we incrementalize the search? Work 3 or 4 examples each compile? • Again, where do we store the results & the experience? Implementation complexity • Try making all your optimizations run correctly in any arbitrary order • Good debugging technique for optimizations, bad plan for meeting deadlines • Large & complex systems have issues with reproducibility & publication • Implementation hassle outweighs benefits SMART 2010 Keynote

  29. Parameterization is Crucial Lesson of the VCR STOP Final Points We have ignored the issue of how to parameterize optimizations • Proliferation of command-line flags • Magic number 7 +/- 2 applies • Parameters are chosen to aid compiler writer, not compiler user • Lack of effective control over behavior • What parameter can you set to throttle inline substitution for i-cache size? • Can you specify different tile sizes for different loop nests? • Good parameterization makes adaptive control natural • Same properties that help an adaptive implementation may help the user • Want a clear relationship between parameter and the outcomes This problem may be a “full professor problem”. SMART 2010 Keynote

  30. Introspection Provides a Different Perspective Final Points Execution time is a bad metric for adaptation • Open ended scale with no way to declare victory • Change in execution time may be obscured by unchangeable base number • It is okay to pursue the small improvement while fixing the large problem • Hard to differentiate a long-running code from a poorly compiled code Introspection looks at what is wrong with the code • Optimizer can only fix problems • Optimizers do not invent new algorithms or shrink problem sizes • Introspective measures show the magnitude of the problems • Number of spills, slack cycles in the schedule, … Introspection may provide a better metric for recognizing & declaring victory SMART 2010 Keynote

  31. Introspection in the PACE Compiler App Source App-Aware Partitioner Platform Aware Optimizer IR Target Aware Optimizer IR Native compiler 1 Native compiler 2 X86 Backend Runtime System Runtime opt. Cooper, Mellor-Crummey, Palem, Sarkar, & Torczon The PACE Compiler carries introspection to an extreme Use instrumented version of LLVM to provide feedback to high-level opts Small query language to specify what should be measured LLVM compiles, measures, & reports PACE Runtime provides region-by-region performance metrics to guide optimization, as well Feedback cycles to include both introspection & execution SMART 2010 Keynote

  32. References for Work Mentioned in the Talk These are minimalist citations, but they should provide enough information to find the paper in a good search engine. Bergner et al., Spill Code Minimization Via Interference Region Spilling, PLDI 97 Bernstein et al., Spill Code Minimization Techniques for Optimizing Compilers, PLDI 89 Cavazos et al., “Rapidly Selecting Good Compiler Optimizations using Performance Counters”, CGO 07 Eckhardt et al., “Redundancy Elimination Revisited”, PACT 08, October 2008. (for rematerialization) Granston & Holler, “Automatic Recommendation of Compiler Options,” 4th International Workshop on Feedback and Data-Driven Optimization, 2001. Lam, “Software Pipelining: An Effective Scheduling Technique for VLIW Machines”, PLDI 98 Liu, “Parameterization and Adaptive Search for Graph Coloring Register Allocation”, MS Thesis, Rice University, 2009. Kulkarni et al., “Fast and efficient searches for optimization-phase sequences”, ACM TACO, June 2005. Motwani et al., Combining Register Allocation And Instruction Scheduling, NYU Courant Institute TR 689, July 19955 Schielke et al., “Optimizing for Reduced Code Space with Genetic Algorithms”, LCTES 99 Simpson, “Live Range Splitting in a Graph Coloring Register Allocator”, CC 1998 Stephenson, et al., “Meta-optimization: improving compiler heuristics with machine learning,” PLDI 03 Triantafyllis et al. “Compiler Optimization Space Exploration”, CGO 03 Waterman et al., “An Adaptive Strategy for Inline Substitution”, CC 2008 Waterman et al. “Investigating Adaptive Compilation with the MIPSPro Compiler,” LACSI Symposium, Journal of High Performance Computing Applications 19(4), 2005. SMART 2010 Keynote

  33. Taxonomy of Approaches Background • Ad-hoc Adaptation (Bernstein 89, Triantafyllis 03) • Find a couple of good ideas • Apply them all & keep the best • Static Adaptation (Motwani 95, Schielke 99, Stephenson 03) • Use training & testing approach to derive good general approaches • Systematic Compile-time Search (Grosul 05, Kulkarni 04) • Guided searches in huge spaces that look for good solutions for one application • Knowledge of space & problem can radically reduce costs • Runtime re-optimization (Bala 02, …) • Detect inefficiency at runtime & fix it • Emphasis on payoff/cost ratio • Can be better than compile-time decisions (DasGupta’s unroller) SMART 2010 Keynote

  34. Abstracting Away the Details App Source Static Compiler Adaptive Controller Source-Level Optimization … … Runtime Perf Monitor Target-Level Optimization Tuning JIT Opt. a.out Context • Major adaptation cycle involves compile, run, monitor, analyze, and recompile • Minor adaptation cycles occur within a single major cycle • Long term learning across multiple runs & multiple applications • This talk focuses on the minor cycles (primarily) • Major is well explored • Learning is another talk from another speaker SMART 2010 Keynote

  35. Internal Adaptation Internal Adaptation A good candidate for internal adaptation has the following properties Best decisions require contextual knowledge of application Difference between good & bad decision is large Can distinguish good results from bad results quickly during compilation We want to distinguish carefully between candidates for internal adaptation and optimizations that either require runtime adaptation or require a full-scale feedback-driven optimization cycle. SMART 2010 Keynote

  36. Adaptive Rematerialization (speculation) Examples Sometimes, it is cheaper to recompute than to preserve a value for later use • Recompute x rather than load x or move x from another location • Historically, rematerialization occurred in the register allocator as an alternative to spilling a value & reloading it • Rising memory costs & proliferation of functional units may shift the balance toward recomputation and away from reuse • Adaptive version would apply a parameterized formula to each expression • Formula accounts for operand availability, resources, and costs • Available registers, bandwidth, & issue slots • Complex measurement might boil down to schedule length & density + spills Properties • Needs major cycle to assess impact on effectiveness of other optimizations • Can rely on introspective measures for faster evaluation SMART 2010 Keynote

  37. Adaptive Rematerialization (speculation) App Source App Source App Source Static Compiler AdaptiveRemat. Other opts Other opts Back End a.out Strategy • Perform rematerialization early • Apply formula to each expression & transform • Measure code quality in the back end • Performance-guided search for threshold values Goal • Quick internal feedback cycle that allows rapid evolution of the behavior • Derive, on each compile, a new set of thresholds SMART 2010 Keynote

  38. Blocking or Tiling for Memory Performance Runtime reoptimization seems too heavyweight for this problem. Negative Example If the cache available at runtime is less than the compiler expected, then performance suffers in serious ways • Common problem on multicore systems • Interference from shared I & D cache or from D cache sharing among cores • Effective cache size depends on the activity on other cores This problem cries out for an adaptive solution • Not a weakness in the compile-time algorithm • Lack of knowledge at compile time • Compile-time adaptation cannot provide the right facts Solution: runtime adaptation • Generate parameterized code • Have the runtime tune the parameters SMART 2010 Keynote

More Related