1 / 17

ALICE CPU Benchmarks: Performance Evaluation and Comparison

This article discusses the importance of CPU benchmarks for ALICE simulations, and explores various benchmarking options and their correlation with simulation performance. It also introduces a database of Grid nodes' actual performance and outlines future plans for benchmarking automation and fair accounting of CPU power.

Download Presentation

ALICE CPU Benchmarks: Performance Evaluation and Comparison

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. ALICE CPU benchmarks ALICE CPU benchmarks Costin.Grigoras@cern.ch

  2. ALICE CPU benchmarks Benchmark considerations • ~70% of the Grid time is taken by simulation jobs • A benchmark has to be simple to find and to run • Short execution time relative to the job duration • For automatic benchmarking of nodes • Reflecting the experiment's software performance on the hardware • Simplified method to collect summarize and share the results • No licensing concerns

  3. ALICE CPU benchmarks MC simulation vs benchmarks(old results from Sep 2015) • Reference production: • “pp 13 TeV, new PYTHIA6(Perugia-2011) min.bias, LHC15f anchors” • 200 ev/job, avg(8h) running time, CPU-intensive • Blanket production, 76 sites • Benchmarks: • ROOT's /test/stress (O(30s)) • condor_kflopsfrom ATLAS' repository (if found) (O(15s)) • Each benchmark ran twice after the simulation • To fill in the CVMFS cache and load the libraries in mem • Recording the second iteration only

  4. ALICE CPU benchmarks Results at a glance

  5. ALICE CPU benchmarks Events/s vs KFlops No correlation between Kflops and simulation performance, probably because of small ratio of floating point operations in it.

  6. ALICE CPU benchmarks Events/s vs rootmarks Rootmarks scale ~better with the simulation time in the Grid environment

  7. ALICE CPU benchmarks Is there anything better still? • Running in a controlled environment • ALICE Central Services machines(~50 hosts) • ROOT stress test results don’t look that good

  8. ALICE CPU benchmarks Sysbench ? • Available by default on many Linux variants • ~30s to run • But it doesn’t scale well...

  9. ALICE CPU benchmarks GeekBench • Commercial product • Evaluation license for the 64b version • ~2min to run • Good results in the test environment

  10. ALICE CPU benchmarks GeekBench – correlation with MC jobs

  11. ALICE CPU benchmarks LHCb’s test • Simple python script • ~1 min to run • Used to estimate how many events the job will be able to generate in a fixed amount of time • Very good results on the CS machines

  12. ALICE CPU benchmarks Best correlation so far 5% of the hosts

  13. ALICE CPU benchmarks HepSpec06 from MJF FZK RAL

  14. ALICE CPU benchmarks CPU model performance Trieste NERSC HIP, CERN Torino HLT 3.7x

  15. ALICE CPU benchmarks Site-specific configurationsHT on/off, mem type, #of slots / machine, ... 1.7x HT OFF HT ON OFF ON ON ON ON ON

  16. ALICE CPU benchmarks Summary • Extensive database of Grid nodes’ actual performance • Can be now queried/populated by calling:http://alimonitor.cern.ch/marks/?cpumodel=M&hostname=H(&site=S)(&lhcbmarks=L) • Lots of distinct configurations • ~14K hosts • 114 CPU models (155 combinations with HT on/off) • Can compare any benchmark in production • LHCb’s looks very good

  17. ALICE CPU benchmarks Plans • Use this database to get the slot performance in each pilot job • fast benchmark run + the history of the same (or similar) nodes • Account for the used CPU in this unit • A common version of the fast benchmark that also queries the database for more stable values • Maybe even feed back actual execution performance (events/second/unique job type) • Automatic measurement, also good for cloud resources • Fair accounting of provided CPU power

More Related