170 likes | 340 Views
High-Performance Computing and Networking. Funding Agencies. Overview of Research Challenges: from Embedded Systems to Supercomputers Alan D. George, Ph.D. Professor of Electrical and Computer Engineering Founder and Director, HCS Research Laboratory. Outline. Research emphases
E N D
High-Performance Computing and Networking Funding Agencies Overview of Research Challenges: from Embedded Systems to Supercomputers Alan D. George, Ph.D. Professor of Electrical and Computer Engineering Founder and Director, HCS Research Laboratory
Outline • Research emphases • Research groups and activities • Reconfigurable computing • Advanced space systems and applications • HPC applications and services • High-performance networking • Advanced simulation and modeling • HPC performance analysis and optimization • Research facilities • Conclusions • Appendix – HPC Initiative at UF
system software architectures networks applications modeling and simulation experimental testbed Research Emphases • Primary Research Areas • high-performance computer architectures • high-performance computer networks • reconfigurable & fault-tolerant computing • parallel and distributed computing • Research Methods • modeling and simulation • experimental testbed research • software design & development • hardware design & development
Reconfigurable Computing • Exciting new technology for HPC • Enabled by new and emerging FPGA technologies (e.g. from Xilinx) • High concurrency potential with RC • Dynamic & partial hardware reconfiguration • Best of ASIC and CPU worlds • Multiparadigm computing for HPC/HPEC • Many research challenges for RC • From hardware structures to middleware, application mapping, & system mgmt. • Offers potentially ideal solutions for many apps in both GP and embedded HPC • Many activities in our RC group • Comprehensive Approach to Reconfigurable Management Architecture (CARMA) • Application mapping from HLL to hardware • RC middleware and API structures • Dynamic and partial reconfigurable structures • Application profiling & analysis as RC target • Computational kernel core libraries • RC/HPC resource monitoring for diagnosis, debug, and performance optimization • RC clusters and grids; network-attached RC • RC applications and benchmarks From Embedded Systems to Supercomputers
Advanced Space Systems & Apps • Broad range of activities • HPEC parallel applications • Multiprocessor architectures • Reconfigurable computing • Fault-tolerant computing • System interconnects • High-speed communications • Testbed experiments and analyses • Virtual and actual prototypes GMTI
HPC Applications & Services • Variety of interests and applications for HPC & HPEC • Computational acoustics and signal processing (ONR) • Computational biomechanics; bioinformatics (w/ MAE for NIH) • Space-based radar (Honeywell); cryptology (DOD) • Exploring additional critical applications for HPC, RC, etc. Number of CPUs Number of CPUs
High-Performance Networking • Research on HPNs for general-purpose HPC • InfiniBand, 10GigE, SCI, Myrinet, Quadrics, etc. • Experimental testbeds • Research on HPNs for embedded HPC systems • RapidIO, WDM LANs • Many critical issues • Protocols and middleware • Multilevel performance traits • Scalability and reliability • Quality of service • From physical to transport
Advanced Simulation & Modeling • FASE Project • Fast and accurate simulation environment; balanced model • HPC application profiling and scripting for speed with fidelity • Architecture, network, subsystem, & system models • HWIL and distributed simulations • Rapid Virtual Prototyping • HPC & HPEC systems • Advanced space systems • Avionics & aerospace networks • Reconfigurable IPv6 routers • Mission assurance systems Processing Raw Data in Heterogeneous Satellite System
Performance Optimization • GAS program models • UPC, SHMEM, etc. • Increasing emphasis at DOD, DOE, etc. • Multilevel optimization • From hardware to apps • Multilevel performance monitoring and profiling • Fusion and integration of metrics for HPC designers • Diverse target platforms • Support leading and emerging HPC systems
Lab Research Facilities • Computational research facilities • Grid of 11 Intel/AMD Linux clusters • New (12th) cluster planned for Su’05 • 480 Pentium-compatible CPUs • Opteron, Xeon, P4, P3, etc. • 308 networked nodes, PCI to PCI-X • 102 GB memory, 5.2 TB storage • Also AlphaServer & Sun clusters • Reconfigurable computing (RC) servers • More details in following slides • Networking research facilities • 10 Gb/s InfiniBand (4X) and 10GigE • 5.3 Gb/s Scalable Coherent Interface • 3.2 Gb/s Quadrics QsNet • 1.28 Gb/s Myrinet • 1.25 Gb/s Cluster LAN (cLAN) • 1.0 Gb/s Gigabit Ethernet
Primary RC Facilities in Lab • Software: • Xilinx ISE • ChipScope • System Generator • Synplify Pro • Handel-C (Celoxica SDK) • Impulse-C • Mitrion-C • Genus-C • Active-HDL RC1000 Dual Xeon Single XCV2000E BenNUEY BenBLUE-II Dual Xeon Triple XC2V6000 Goal: Largest RC Cluster in World Dual Xeon, Dual Opteron, Single Athlon, or Single P3 Single XC2VP30 Content Packet Processor Dual Xeon Dual XC2V1000 (under construction) ~100 nodes Reconfigurable Application- Specific Computer (RASC) Altix350, Single XC2V6000 ADM-XRC Pentium2 Single XCV1000E
Embedded and Custom RC Facilities Total RC Resources (aggregate of all cards) 6,594,624 slices 192 embedded PowerPCs Field-Programmable Object Array (FPOA) (coming soon) UF-developed NARC: Network-Attached RC (ARM AT91RM9200, Ethernet, Xilinx FPGA) Xilinx Development Boards Xilinx FPGAs for Prototyping Twelve HW-AFX-BG560 Two HW-AFX-FF1152 Eight XCV812E Eight XCV1000E Eight XCV2000E Two XC2VP20
CHREC – Proposed New Center • Proposed new center via NSF I/UCRC Program • Center for High-Performance Reconfigurable and Embedded Computing (CHREC) – pronounced “Shreck” • I/UCRC = Industry/University Cooperative Research Center • Focus on both HPC and HPEC sides of RC research • Receiving much interest from variety of potential members in industry (e.g. Honeywell, Boeing, Smiths Aerospace, SGI, Xilinx, Cray), in government (e.g. NSA, AFRL), national labs (e.g. ORNL), academia • Steps Toward Goal • Letter of intent submitted to NSF in Dec’04, approved in Jan’05 • Next step is planning grant proposal for center, due late Sep’05 • Will be requesting strong letters of support and encouragement for this center in July/August (details to follow) from key industry members
Conclusions • Wide range of research expertise in architectures, networks, services, systems, and applications in HPC • Focus on high-performance parallel, distributed, and reconfigurable computing and communications for critical applications • From embedded systems to clusters, supercomputers, and grids • Focal points • Goal of “high performance” in terms of execution time, throughput, latency, quality of service, dependability, etc. • Research challenges in computer engineering in terms of both general-purpose and embedded HPC systems and applications • Close collaboration with emerging application domains of HPC • Both simulative and experimental expertise to achieve distinct and interdependent goals requiring both basic and applied research • World-class experimental and simulative research facilities in an academic setting
High-Performance Computing (HPC) Initiative at UF Applications and Infrastructure Research with Advanced HPC Technologies A. George, Chair, University HPC Committee
HPC Center and Grid Dell Center of Research Excellence (Jun'05) Phase I (#221 on Nov’04 Top500 list of most powerful systems on earth) • Notes: • Phase I is a cluster of 200 dual-Xeon servers with 32TB of storage • Phase II now under development (approx. twice size of Phase I) • All phases and FLR supported by new campus research network of dual 10GigE funded by NSF/MRI grant