1 / 10

Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage. September 2008. Gilad Shainer – Director of Technical Marketing. Company Overview. Silicon-based server and storage interconnect products R&D, Operations in Israel; Business in California

Download Presentation

Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Mellanox InfiniBand InterconnectThe Fabric of Choice for Clustering and Storage September 2008 Gilad Shainer – Director of Technical Marketing

  2. Company Overview • Silicon-based server and storage interconnect products • R&D, Operations in Israel; Business in California • Four generations of products since 1999 • 250+ employees; worldwide sales & support • InfiniBand and Ethernet leadership • Foundation for the world’s most powerful computer • 3.7M 10/20/40Gb/s ports shipped as of Jun08 • Proven execution, high-volume manufacturing & quality • Solid financial position • FY’07 $84.1M, 73% growth from FY’06 • Record Revenue in 2Q’08, $28.2M • 1H’08 $53.4M, 3Q’08 est. $28.5M-$29M • Tier-one, diversified customer base • Includes Cisco, Dawning, Dell, Fujitsu, Fujitsu-Siemens, HP, IBM, NEC, NetApp, QLogic, SGI, Sun, Supermicro, Voltaire $106M raised in IPO Feb07Ticker MLNX

  3. High Throughput - 40Gb/s Low latency - 1us Low CPU overhead Kernel bypass Remote DMA (RDMA) Reliability InfiniBand End-to-End Products Adapter ICs & Cards Adapter ICs & Cards End-to-End Validation Switch ICs Software Cables ADAPTER Cables ADAPTER SWITCH Blade/Rack Servers Storage Switch Maximum Productivity

  4. Virtual Protocol Interconnect … App1 App2 App3 App4 AppX Applications Consolidated Application Programming Interface Networking TCP/IP/UDP Sockets Storage NFS, CIFS, iSCSINFS-RDMA, SRP, iSER, Fibre Channel, Clustered Clustering MPI, DAPL, RDS, Sockets Management SNMP, SMI-SOpenView, Tivoli, BMC, Computer Associates Protocols Networking Clustering Storage Virtualization RDMA Acceleration Engines 10/20/40 InfiniBand 10GigE Data Center Ethernet Any Protocol over Any Convergence Fabric

  5. The Fastest InfiniBand Technology • InfiniBand 40Gb/s QDR in full productions • Multiple sites already utilized InfiniBand QDR performance • ConnectX InfiniBand - 40Gb/s server and storage adapter • 1usec application latency, zero scalable latency impact • InfiniScale IV - 36 InfiniBand 40Gb/s switch device • 3Tb/s switching capability in a single switch device

  6. InfiniBand QDR Switches • 1RU 36-port QSFP, QDR switch • Up to 2.88Tb/s switching capacity • Powered connectors for active cables • Available now • 19U 18 slot chassis, 324-port QDR switch • Up to 25.9Tb/s switching capacity • 18 QSFP ports per switch blade • Available: Q4 2009

  7. InfiniBand Technology Leadership • Industry Standard • Hardware, software, cabling, management • Design for clustering and storage interconnect • Price and Performance • 40Gb/s node-to-node • 120Gb/s switch-to-switch • 1us application latency • Most aggressive roadmap in the industry • Reliable with congestion management • Efficient • RDMA and Transport Offload • Kernel bypass • CPU focuses on application processing • Scalable for Petascale computing & beyond • End-to-end quality of service • Virtualization acceleration • I/O consolidation Including storage The InfiniBand Performance Gap is Increasing 240Gb/s (12X) 120Gb/s 80Gb/s (4X) 60Gb/s 40Gb/s Ethernet 20Gb/s Fibre Channel InfiniBand Delivers the Lowest Latency

  8. InfiniBand 40Gb/s QDR Capabilities • Performance driven architecture • MPI latency 1us, Zero scalable latency • MPI bandwidth 6.5GB/s bi-dir, 3.25GB/s uni-dir • Enhanced communication • Adaptive/static routing, congestion control • Enhanced Scalability • Communication/Computation overlap • Minimizing systems noise effect (DOE funded project) 8-cores 16-cores

  9. HPC Advisory Council • Distinguished HPC alliance (OEMs, IHVs, ISVs, end-users) • Members activities • Qualify and optimize HPC solutions • Early access to new technology, mutual development of future solutions • Outreach • A community effort support center for HPC end-users • End-User Cluster Center • End- user support center • For details – HPC@mellanox.com

  10. Thank YouHPC@mellanox.com 10

More Related