1 / 14

Infiniband in EDA (Chip Design)

Infiniband in EDA (Chip Design). Glenn Newell Sr. Staff IT Architect Synopsys. Agenda. Synopsys + Synopsys Computing EDA design flow vs. data size and communication High Performance Linux Clusters at Synopsys Storage is dominant vs. Inter-process communication

eleanorab
Download Presentation

Infiniband in EDA (Chip Design)

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Infiniband in EDA (Chip Design) Glenn Newell Sr. Staff IT Architect Synopsys

  2. Agenda • Synopsys + Synopsys Computing • EDA design flow vs. data size and communication • High Performance Linux Clusters at Synopsys • Storage is dominant vs. Inter-process communication • Performance Increases with Infiniband • Tell the world • Next Steps

  3. Synopsys • “A world leader in semiconductor design software” • Company Founded:1986 • Revenue for FY 2006: $1.096 billion • Employees for FY 2006: ~5,100 • Headquarters: Mountain View, California • Locations: More than 60 sales, support and R&D offices worldwide in North America, Europe, Japan, the Pacific Rim and Israel

  4. Over 60 Offices World Wide Major Data Centers 5 at HQ Hillsboro OR Austin TX Durham NC Nepean Canada Munich Germany Hyderabad India Yerevan Armenia Shanghai China Tokyo Japan Taipei Taiwan 2 Petabytes of NFS Storage ~15000 compute servers 10000 Linux 4000 Solaris 700 HPUX 300 AIX GRID Farms 65 farms composed of 7000 machines 75% SGE 25% LSF Interconnect GigE storage Fast E clients Synopsys IT (2007) • #242 on Nov. ’06 Top400.org • 3.782 TFlops on 1200 Processors

  5. Design Flow vs. Data Size and Communication (2007) • RTL “relatively” small data sets • Physical Layout Data up to 300GB • Inter Process Communication “small” compared to file i/o • Post Optical Proximity Correction 300GB - >1TB • OPC adds complex polygons • Mask machines need flat data (no hierarchy or pattern replication) • Physical world is “messy” (FFT + FDTD)

  6. High Performance Linux ClustersProgression HPLC2 HPLC3

  7. HPLC3 vs. HPLC1 Why IB + Lustre?

  8. Why? – HPLC1 NFS + GigE • Large CPU Count Fractures overtax NFS server CPU • Maximum Read Bandwidth is 90 MB/sec Fracture Explode Fracture Explode

  9. Why? – HPLC3 Lustre + IB 64 CPU Fracture • Lustre splits traffic across resources • 250MB/sec max read bandwidth

  10. Storage + Interconnect Option Comparisons + Myrinet

  11. High Performance Linux ClustersProduction Model • 17x the storage performance of HPLC1 • 6x the storage performance of HPLC3 • IB gateway to Enterprise (6xGigE) means no dual homed hosts • “Fantastic” Performance • More Linear Scaling for Distributed Processing Applications

  12. State of the IB Cluster • Our Customer facing Engineers typically see • 10x improvement for post layout tools over Fast Ethernet • 3x improvement for post OPC tools over GigE NFS, or direct attached storage • So, we are evangelizing IB + parallel file systems (Lustre) with our customers • User Manuals • Papers, Posters, and Presentations at conferences • Presentations and Site visits with customers • Push storage vendors to support IB to the client (vs. 10GigE + TOE) • But…

  13. Estimated EDA Relative CPU Cycles Required 2007  2009 ~10x 65nm  45 nm 2007  2012 ~100x 65nm  22 nm 45 nm

  14. Next Steps • “Inside the box” changing • Multicore • Hardware Acceleration (GPU/co-processor merges with CPU) • Micro Architecture • Applications changing to deal with above and increased data set sizes • Things for IT to explore • Other “new” parallel file systems (e.g. Gluster) • 12X DDR IB • 10GB uplinks • IB Top500 entry? ;-)

More Related