Upload
danghuong
View
227
Download
5
Embed Size (px)
Citation preview
Anand Rangaswamy | Storage Developer Conference September 2014
“Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet”
© 2014 Mellanox Technologies 2 - Mellanox Confidential -
Mellanox Overview
Leading provider of high-throughput, low-latency server and storage interconnect • FDR 56Gb/s InfiniBand • 10/40/56GbE Ethernet Switches and Adapters
Company headquarters: • Yokneam, Israel; Sunnyvale, California • ~1,432 employees* worldwide
Ticker: MLNX
* As of March 2014
© 2014 Mellanox Technologies 3 - Mellanox Confidential -
Leadership in Strategic Markets with an End to End Portfolio
Host/Fabric Software ICs Switches/Gateways Adapter Cards Cables/Modules
End-to-End - Cloud, Enterprise, and Storage Optimized - InfiniBand and Ethernet Portfolio
This image cannot currently be displayed.
Metro / WAN
DB/Enterprise Cloud Web 2.0 Storage Big Data
© 2014 Mellanox Technologies 4 - Mellanox Confidential -
Mellanox Unique Value Proposition - RDMA
ZERO Copy Remote Data Transfer
Low Latency, High Performance Data Transfers
InfiniBand - 56Gb/s RoCE* – 40Gb/s
Kernel Bypass Protocol Offload
* RDMA over Converged Ethernet
Application Application USER
KERNEL
HARDWARE
Buffer Buffer
© 2014 Mellanox Technologies 5 - Mellanox Confidential -
RDMA – How it Works
RDMA over InfiniBand or Ethernet
KER
NEL
H
AR
DW
AR
E U
SER
RACK 1
OS
NIC Buffer 1
Application
1 Application
2
OS
Buffer 1
NIC Buffer 1
TCP/IP RACK 2
HCA HCA
Buffer 1 Buffer 1
Buffer 1
Buffer 1
Buffer 1
© 2014 Mellanox Technologies 6 - Mellanox Confidential -
Infiniband Acronyms
Reliability / Efficiency
• Link bit encoding – 64/66
• Forward Error Correction
• Low power consumption
Performance / Scalability
• Switch port bandwidth 56Gb/s
• InfiniBand Router and IB-Eth bridges
• First VPI silicon in the World
FDR InfiniBand New Features and Capabilities
2002
10 Gb/s
SDR
2005
20 Gb/s
DDR
2008
40 Gb/s
QDR
2011
56 Gb/s
FDR VPI*
2014
100 Gb/s
EDR
© 2014 Mellanox Technologies 7 - Mellanox Confidential -
5X Performance Improvement Over 10GbE
New class of enterprise file storage with RDMA support Lower latency, higher throughout, lower CPU
overhead Fibre Channel replacement at a lower cost and
higher performance Leverages Windows Server 2012 R2 Mellanox
inbox InfiniBand & Ethernet RDMA drivers Accelerates Microsoft Hyper-V over SMB and
Microsoft SQL Server over SMB solutions
Microsoft Windows Server 2012 R2 SMB Direct
File Client File Server
SMB Client
User
Kernel
Disk RDMA
Adapter
Network w/ RDMA support
RDMA Adapter
SMB Server
Network w/ RDMA support
NTFS SCSI
Application
© 2014 Mellanox Technologies 8 - Mellanox Confidential -
Measuring SMB Direct Performance
Single Server
Fusion IO Fusion IO Fusion IO Fusion-IO
IO Micro Benchmark
SMB Client
SMB Server
Fusion IO Fusion IO Fusion IO Fusion-IO
IO Micro Benchmark
10GbE
10GbE
SMB Client
SMB Server
Fusion IO Fusion IO Fusion IO Fusion-IO
IO Micro Benchmark
FDR IB
FDR IB
SMB Client
SMB Server
Fusion IO Fusion IO Fusion IO Fusion-IO
IO Micro Benchmark
QDR IB
QDR IB
© 2014 Mellanox Technologies 9 - Mellanox Confidential -
Performance Comparison
© 2014 Mellanox Technologies 10 - Mellanox Confidential -
Microsoft Delivers High-End Storage at a Fraction of the Cost
FDR 56Gb/s InfiniBand delivers 5X higher throughput with 50% less CPU overhead vs. 10GbE
Native Throughput Performance over FDR InfiniBand
© 2014 Mellanox Technologies 11 - Mellanox Confidential -
Performance and Resiliency Similar to a Traditional SAN
1.1 Million IOPS for 8KB random reads 700K IOPS for 8KB random writes
100Gbps aggregate read bandwidth
92Gbps aggregate write bandwidth
Source: http://www.mellanox.com/related-docs/applications/Achieving-1.1M-IOPS-over-SMB-Direct.pdf
*
*Cluster Shared Volumes
Scale-Out File Server Cluster using Windows Server 2012 R2 SMB Direct
FDR InfiniBand enables Native Performance
© 2014 Mellanox Technologies 12 - Mellanox Confidential -
Mellanox Testing 8KB I/O Echostreams flash array with 24 SSDs
12
0
100
200
300
400
500
600
700
Local SSDs 10 GE TCP 10 GE RoCE 1xLink
10 GE RoCE 2xlinks
40 GE TCP 40 GE RoCE 1xlink
40 GE RoCE 2xlinks
IB 54Gbps FDR1x link
IB 54Gbps FDR2x links
IB 54Gbps FDR3x links
IOPS (1000's IOs/sec)
IOPS (1000's IOs/sec)
© 2014 Mellanox Technologies 13 - Mellanox Confidential -
Mellanox Testing 32KB I/O Echostreams flash array with 24 SSD
13
0
50
100
150
200
250
300
Local SSDs 10 GE 10 GE RoCE 1xLink
10 GE RoCE 2xlinks
40 GE 40 GE RoCE 1xlink
40 GE RoCE 2xlinks
IB 54Gbps FDR1x link
IB 54Gbps FDR2x links
IOPS (1000's IOs/sec)
IOPS (1000's IOs/sec)
© 2014 Mellanox Technologies 14 - Mellanox Confidential -
Mellanox Testing 512KB I/O Echostreams flash array with 24 SSDs
14
0.0
2.0
4.0
6.0
8.0
10.0
12.0
14.0
16.0
18.0
20.0
Local SSDs 10 GE 10 GE 1x RoCE 10 GE RoCE 2xlinks
40 GE 40 GE RoCE 1xlink
40 GE RoCE 2xlinks
IB 54Gbps FDR1x link
IB 54Gbps FDR2x links
IOPS (1000's IOs/sec)
IOPS (1000's IOs/sec)
© 2014 Mellanox Technologies 15 - Mellanox Confidential -
Hyper-V Live Migration on SMB Direct with Mellanox Interconnect
SMB as a transport for Live Migration of VMs Delivers the power of SMB to provide:
• RDMA (SMB Direct) • Streaming over multiple NICs (SMB Multichannel)
Provides highest bandwidth and lowest latency
0
10
20
30
40
50
60
70
Seco
nds
Live Migration Times* New in
Windows Server
2012 R2
Live migration can stream over
multiple networks for
improved bandwidth
Live Migration can take
advantage of high speed networking
RDMA enables offloading
CPU resources to NIC during live migration
Source: TechED’13 Opening Keynote Session with Satya Nadella
* Lower is better
© 2014 Mellanox Technologies 16 - Mellanox Confidential -
Success Story: Edgenet
Reduced complexity vs. FC storage management • Creates a more manageable storage subsystem
- SMB easier to provision than Fibre Channel devices • The storage server manages all the connections
- Client Servers only need to know one path via SMB
10x Acceleration vs FC • Allows consolidation of many FC arrays over SMB • Live migration between servers at record speeds Mellanox VPI solutions
• Provides user access to the Hyper-V hosted application
Source: http://www.mellanox.com/related-docs/case_studies/CS_Edgenet_Boost_Performance_Lower_Cost.pdf
© 2014 Mellanox Technologies 17 - Mellanox Confidential -
SMB Direct Storage Solutions
18 CONFIDENTIAL | © V i o l i n M e m o r y , I n c . 2 0 1 4
Introducing The Windows Flash Array
19 CONFIDENTIAL | © V i o l i n M e m o r y , I n c . 2 0 1 4
Windows in a Flash!
Transformative Speed in Storage
Consolidation and cost savings
Increases server utilization
SMB Direct for Microsoft applications
Enables file server functions
RDMA as a Storage Superhighway
Encryption Dedupe Live Migration Mirroring Scale-Out
>1 year of Joint Violin - Microsoft Development
20 CONFIDENTIAL | © V i o l i n M e m o r y , I n c . 2 0 1 4
1.2M
1.0M
800K
600K
400K
200K
SQL WRITES*
Beta Site Performance: SQL* and Hyper-V
• SMB Direct reduces CPU load up to 30% • Latency averages 500 microseconds at Load
SQL READS*
HYPER-V WRITES
WFA
AFA
IOPS
41% Increase
54% Increase
2x Increase
30% Decrease
50% Decrease
SQL Writes
SQL Reads
Hyper-V Writes
Application Server CPU Utilization
Latency at Load
WFA
WFA
AFA
AFA
*All Microsoft SQL results based on SQLIO load test with Infiniband connectivity
© 2014 Mellanox Technologies 21 - Mellanox Confidential -
DataOn Cluster-in-a-Box over RDMA
Extremely economical Simplified architecture
Easy to install appliance
CiB interconnects:
• ConnectX-3 40GbE • ConnectX-3 56Gbs InfiniBand
Deployed at:
© 2014 Mellanox Technologies 22 - Mellanox Confidential -
File Server (SMB 3.0)
File Client (SMB 3.0) SQLIO
RDMA NIC
RDMA NIC
RDMA NIC
RDMA NIC
RDMA NIC
RDMA NIC
SAS
SAS HBA
JBOD SSD SSD SSD SSD SSD SSD SSD SSD
SAS
SAS HBA
JBOD SSD SSD SSD SSD SSD SSD SSD SSD
SAS
SAS HBA
JBOD SSD SSD SSD SSD SSD SSD SSD SSD
SAS
SAS HBA
JBOD SSD SSD SSD SSD SSD SSD SSD SSD
SAS
SAS HBA
JBOD SSD SSD SSD SSD SSD SSD SSD SSD
Storage Spaces
SAS
SAS HBA
JBOD SSD SSD SSD SSD SSD SSD SSD SSD
8KB random reads from a mirrored space
(disk) ~600,000 IOPS
8KB random reads from cache (RAM) ~1,000,000 IOPS
32KB random reads from a mirrored space
(disk) ~500,000 IOPS
~16.5 GBytes/sec
EchoStreams: InfiniBand Enables Near Linear Scalability
© 2014 Mellanox Technologies 23 - Mellanox Confidential -
Preconfigured Microsoft Hyper-V cloud IaaS hardware platform • Full integrated converged data center solution
Iron Pod Description • Fault tolerant Mellanox 40/56GbE server and storage network
- Fully redundant network for high availability • ConnectX-3 Pro adapters for RDMA and NVGRE acceleration
- Dual 40/56GbE network adapters
More info about Iron Networks: • http://www.mellanox.com/related-
docs/solutions/Iron_Networks_Microsoft_Fast_Track_Architecture.pdf • http://ironnetworks.com/Microsoft-Converged-Infrastructure/IronPOD-Cloud-
Building-Blocks • http://ironnetworks.com/Microsoft-Converged-Infrastructure/IronPOD-System-500
Iron Networks – IRON POD
© 2014 Mellanox Technologies 24 - Mellanox Confidential -
Questions?
Thank You
© 2014 Mellanox Technologies 26 - Mellanox Confidential -
Accelerating Microsoft SQL 2012 Parallel Data Warehouse V2
Analyze 1 Petabyte of Data in 1 Second
Up to 100X faster performance than legacy data warehouse queries
Up to 50X faster data query, up to 2X the data loading rate
Unlimited storage scalability for future proofing
Accelerated by Mellanox FDR 56Gb/s InfiniBand end-to-end solutions
© 2014 Mellanox Technologies 27 - Mellanox Confidential -
Microsoft SQL Server Appliance – 10X Faster & 50% Lower Capital Cost
Control Node
Mgmt. Node
LZ
Backup Node
Estimated total HW component list price: $1M $ Estimated total HW component list
price: $500K $
Ethernet, InfiniBand & Fibre Channel
• Pure hardware costs are ~50% lower
• Price per raw TB is close to 70% lower due to higher capacity
• 70% more disk I/O bandwidth
InfiniBand
• 128 cores on 8 compute nodes • 2TB of RAM on compute • Up to 168 TB of temp DB • Up to 1PB of user data
• 160 cores on 10 compute nodes • 1.28 TB of RAM on compute • Up to 30 TB of temp DB • Up to 150 TB of user data
Parallel Data Warehouse Appliance V1 Parallel Data Warehouse Appliance V2