Intro Presentation - Globalpress Connection

Download Report

Transcript Intro Presentation - Globalpress Connection

Chelsio Communications

Ethernet Unified Wire

November 15, 2006

Kianoosh Naghshineh, CEO [email protected]

www.chelsio.com

Pioneering Unified Wire

LAN SAN HPC Unified 10Gb Ethernet Network SAN LAN HPC NAS NAS

Convergence Benefits

Lower Total Cost of Ownership

• Improves CPU efficiency • Minimizes software licenses • Simplifies data center wiring • Leverages staffing skills & tools

Higher Performance and New Apps

• Improves cluster performance • Lowers application latency • Faster backup and recovery • Enables storage applications

Simplified network architecture – reduced operating costs

CHELSIO CONFIDENTIAL

2

Key Market Drivers

Criteria Market Drivers / Enablers

Unit Growth

 3x volume growth in the 10GbE NIC market in 2006 (synergy report)  3x volume growth in 10GbE switch market in 2006 (Dell’Oro)  iSCSI market growing at 50% per quarter

Infrastructure

 High density 10GbE switches available now  OpenFabrics released  can run IB apps unmodified, channel product  Chimney OS released  All NICs become TNICs in channel  iSCSI Target & Initiator ready  can begin to replace FC  Middleware partnerships in place  can run FC apps unmodified

Prices Standards

 XFP over 12 months: prices halved  CX4 switch port at $700/port list price now  10G CX4 RNIC at $995. Expected to be halved by 4Q07  10G CX4 (copper media) introduced and shipping  10G-baseT switches and cards expected by 1Q07

CHELSIO CONFIDENTIAL

3

Current Shipping Products

T210 10GbE Protocol Engine

CX4 copper or XFP SR/LR optics

Full TCP/IP Offload (TOE) Full iSCSI Target solution CX4 copper or XFP fiber

Highest performance and CPU efficiency

N210 10GbE Server Adapter 10Gb Ethernet NIC Linux, Windows, Solaris Turn-key software drivers

Easy-to-use server adapter for 10GbE connectivity

PCI-Express or PCI-X bus interface

T204 4GbE Protocol Engine 4-port Gigabit Ethernet Full TCP/IP Offload (TOE) Full iSCSI Target solution

Best price/performance for servers & IP SANs

 

2 nd generation products shipping in volume since Feb 2005 Over 140 customers in 20 countries using Chelsio adapters

Strong foundation for 3 rd generation ‘unified wire’ solution

Holder of all but one of the current Land Speed Records CHELSIO CONFIDENTIAL

4

Introducing Terminator 3 (T3)

PCI-Express or PCI-X bus interface

T3 ASIC T310 10GbE PCI-Express or PCI-X 2.0

T302 2x1GbE PCI-Express

       

PCIe x8 or PCI-X 2.0

2x10G or 2x1G Ethernet NIC+TOE+iSCSI+RDMA Up to 1M connections ACPI, IPMI, SMBus, I 2 C 0.13um TSMC LV ~11W typical power 37mm FCGBA

      

PCIe x8 or PCI-X 2.0

10GbE CX4 or XFP NIC+TOE+iSCSI+RDMA 32K connections iSCSI Target+Initiator Full-duplex 10Gbbps line-rate performance ½ length form factor

      

PCIe x4 2-port Gigabit Ethernet NIC+TOE+iSCSI+RDMA 32K connections iSCSI Target+Initiator 2 x 1000BASE-T ports Low-profile form factor

5

CHELSIO CONFIDENTIAL

Superior Performance

3 2 1 0 8 7 6 5 4 Chelsio T110 TOE versus 10GbE NIC Network Performance Higher Throughput (Gb/s) Lower CPU Utilization (%) T110 N110 Intel S2io 120% 100% 80% 60% 40% 20% 0% TX RX TX RX Source: Independently verified by VeriTest, Inc.

Test tool: netperf Test configuration: 2 systems connected through 10GbE switch running single TCP channel with 1500-byte Ethernet frames System configuration: AMD Opteron 248 2.2GHz uniprocessor running Linux kernel 2.6.6

    

Performed May 2003 by the world’s leading independent lab, VeriTest, Inc.

TOE delivers ~2X network throughput vs. 10GbE NICs TOE shows ~1/2X CPU utilization vs. 10GbE NICs Net-net: TOE shows ~4X network efficiency vs. NICs No Jumbos, one connection to thousands of connections CHELSIO CONFIDENTIAL

6

Independent HPC Benchmarks Validate Performance 10GbE has higher bandwidth than Infiniband & Myrinet 10GbE has lower latency than Infiniband

CHELSIO CONFIDENTIAL

7000 6000 5000 4000 3000 2000 1000 0 1 Bandwidth vs Message Size 10GigE SDP/IBA (Event) SDP/IBA (Poll) SDP/Myrinet (Event) SDP/Myrinet (Poll) 4 16 64 256 1K 4K Message Size (bytes) 16K 64K 256K Latency vs Message Size 100 90 80 70 60 50 40 30 20 10 0 10GigE TOE SDP/IBA (Event) SDP/IBA (Poll) SDP/Myrinet (Event) SDP/Myrinet (Poll) 1 2 4 8 16 32 64 128 256 512 1K Message Size (bytes) 2K 4K 8K 16K Testing performed by Los Alamos National Laboratory and The Ohio State University 7

RDMA & iSCSI

      

PCI-X 133MHZ 1.8GHZ Dual CPU OPTERON 1GB Memory Mellanox 4X SDR IB PCI-X HCA Chelsio T320X 10GbE PCI-X RNIC Point to point networks Using Open-IB derivative benchmarks 700 600 500 400 300

20 Initiators – MS iSCSI 2.0, 1GbE

 

Dell 2950, 2 x Intel 3GHz EM64T Dual Core Woodcrest, 8GB DDR-II 667MHz memory Linux 2.6.17 SMP x86_64.

T320e (T3b).

iSCSI Target 2.0 release.

Ramdisk. No CRC, TOE mode CHELSIO CONFIDENTIAL 200 100 0 12 10 8 6 4 2 0 1B Read (usec) 1B Write (usec) Chelsio Mellanox 40 30 20 10 0 100 90 80 70 60 50 IO Size

Read Write %CPU Rd %CPU Wr 8

T3 Feature Summary

Feature Hardware Architecture Management QoS Virtualization Filtering Traffic Manager TOE DDP iSCSI iWARP RDMA Description

    

Integrated native PCIe 8x and PCI-X 2.0 266MHz interface Integrated 2 GigE or 2 10GbE ports with CX4/KX4 compliant XAUI Cut-through processing architecture delivers 10Gbps wire rate & ultra low latency Scalable to 100Gbps and extendable to additional protocols (IPSec, SSL, XML, NFS) Flexible in cost, power, performance, programmability

PXE, EFI, Power Management

 

Multiple queues for bandwidth resource provisioning and allocation Multiple channels for simultaneous high bandwidth/low latency operation

Extensive receive and transmit features to support virtualization

Tens of thousands of filter rules in hardware, configurable in software, used for firewalls and virtualization

Simultaneous per-connection and per-class rate control and bandwidth allocation

       

4x performance vs NIC Implements both Full and Partial TCP offload High performance even in demanding network environments (large RTT, packet loss) Zero-copy steering of payload data directly into application space for extremely low CPU utilization No modification to applications Line rate receive on 1 connection 16x performance vs NIC Chelsio offers complete commercial-grade iSCSI software stack for Linux for FREE

 

RDMAC & IETF compliant RNIC-PI, kDAPL and OpenIB software interfaces CHELSIO CONFIDENTIAL

9

Competitive Analysis

 

Chelsio competes primarily with companies that develop Ethernet adapter cards and/or integrated circuits

 Indirect competition from other fabric technologies such as InfiniBand and Fibre Channel

Chelsio delivers a best-in-class, full-featured and seasoned solution Chelsio Broadcom Emulex/Silverback QLogic Intel Neterion NetXen NIC ► ► ► ► ► TOE ► ► ► iSCSI ► ► ► ► RDMA ► ► 10GbE ► ► ► ► GbE ► ► ► ► ► 2-port ► PCI-2X ► PCIe ► ► ► ► ► ► VLIW ► CHELSIO CONFIDENTIAL

10

Why Chelsio

       

Very High Performance

 Very Low Latency  High Transaction Rate  High Bandwidth

Correct architecture

 Data Flow Processor produces a Scalable and Ideal Performance Profile

Value added features

 Integrated Traffic Manager  Filtering 

3 rd

 Virtualization

Generation – Robust, Seasoned

Maximum guarantee of no late arriving QA bugs

Integrated High Performance ULP solutions – run IB or FC apps unmodified

 Commercial grade iSCSI stack integrated with hardware DDP facilities  Fully integrated OpenFabrics RDMA software stack

Broad product offering – future protection

 2x1Gb or 2x10Gb, iSCSI, RDMA, TOE, Chimney, NIC  PCI-Express, PC-X 2.0, PCI-X 1.0

Reliable, predictable supplier (ISO 9001 certified)

 The only vendor with experience shipping Offload products to enterprise OEMs

Availability of software-only products CHELSIO CONFIDENTIAL

11

Thank You

CHELSIO CONFIDENTIAL

12