Upload
andrey-klyachkin
View
2.870
Download
5
Embed Size (px)
Citation preview
1
© 2009 IBM CorporationSTG Technical Conferences 2009
SMP16Faster Than a Speeding Disk
(Solid State Drives)
Mark OlsonIBM WW Power Systems Product Manager
© 2009 IBM Corporation2
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD Agenda
• Technology / benefit overview
• Implementing Hot/Cold– AIX & IBM i
• Examples/Scenarios
• Technology/Pricing
• Configuring SSD
• Additional information
2
© 2009 IBM Corporation3
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
New Solid State Drives (SSD)
• SSD or “flash drives” offers a new way of running your system to really boost I/O performance
• Ultra high speed I/O can
– Improve your system performance
• Faster response time
• Shorter batch job runtimes
– Save space in your computer room
– Lower your electrical/cooling costs
– Fewer drives improves overall hardware reliability
The size of benefits are dependent upon several factors. Application I/O usage and starting/ending configuration are key.
90% reductionEnergy for I/O
1.65 X more Total transactions
10.8 X more
(SSD do 97% work)
SSD+HHD average drive throughput – I/O per second
86% reductionFewer total drives
3.1 X better ReadsSSD average I/O response time
36 SSD + 80 HDDSSD + HDD
800 HDDAll HDD (Hard Disk Drives)
OLTP / DB Sample* Workload
Compare all HDD vs mixed
HDD+SSD
* This sample workload comparison documented in white paper released late April, 2009 Sample workload selected to stress I/O. Customer workloads, configurations and SSD impact will vary fromthis sample. “Hot” data placed on SSD.
For exa
mple:
© 2009 IBM Corporation4
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Basic Problem --- Disk “Slowing” Down (Relatively)
• Capacity growing ok (35% per year), but Read/Seek -1% & Data Rate only 15% per year
• While processors & memory speed up and add threads and cache
• Net … a growing imbalanced between disk and processor/memory
Seagate 15k RPM/3.5" Drive Specifications
73
450
171
75
3.43.6
2002 2008
Capacity (GB)
Max Sustained
DR (MB/s)
Read Seek (ms)
+35%
+15%
-1%
3
© 2009 IBM Corporation5
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Solid State Drives (SSD) Matching Applications’ Need
• Today’s applications can often benefit with a faster storage option
• SSD high speed can really help get rid of I/O bottlenecks, bridging the gap between memory and disk speeds
– Improve performance
– And save space, energy at the same time
Processors Memory DiskSSD
Very, very, very, very, very fast
Very, very, very fast
Very, very slow comparativelyFast
Access Speed
1,000,000 -8,000,000 ns
~200,000 ns~100 ns< 10’s ns
© 2009 IBM Corporation6
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Solid State Drives (SSD) Matching Applications’ Need
Processors Memory DiskSSD
Very, very, very, very, very fast
Very, very, very fast
Very, very slow comparativelyFast
Access Speed
1,000,000 -8,000,000 ns
~200,000 ns~100 ns< 10’s ns
~ 12.5 hours
~33 minutes
~1 second
DiskSSDmemory
Human Time Context
4
© 2009 IBM Corporation7
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Many Systems Buy “Extra” Disk Arms for Performance
• HDD % utilization of capacity is held low to help ensure higher I/O performance and more consistent response time - arm movement, spinning platter an issue
• Write cache and read cache attempt to buffer the impact
• SSD % utilization of capacity not restricted
Often less than 30-50% full for performance sensitive workloads
Can run much closer to 100% full
© 2009 IBM Corporation8
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Mixed SSD + HDD Can be Great Solution
Cold
Hot
Hot data may be only 10-20% capacity, but represent 80-90% activity
• SSD offers best price performance when focused on “hot” data
• HDD offers best storage cost, so focus it on “cold” data …. sort of a hierarchal approach
• It is typical for data bases to have a large percentage of data which is infrequently used (“cold”) and a small percentage of data which is frequently used (“hot”)
May be able to use larger HDD and/or a larger % capacity used
Can run SSD closer to 100% capacity
5
© 2009 IBM Corporation9
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD
Faster than a spinning disk
Solid State Drive (also called “Flash Drive”)
• High speed – no rotational delay, no arm movement
• Ignore old rules of enough disk arms for performance
• No moving parts improves SSD reliability
• Far lower power/cooling than a 3.5-inch disk drive (HDD)
– Similar power/cooling to 2.5-inch SFF HDD
• Can save floor space and energy by eliminating many HDDs
• High cost on a “per drive” or “per GB” comparison – but for the right application, a very nice fit with good return on investment
© 2009 IBM Corporation10
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD Performance
0
1000
2000
3000
4000
5000
6000
7000
8000
SSD HDD
Power Consumption in Watts Required for 135K IOPS
performance
0
5000
10000
15000
20000
25000
30000
SSD HDD
Random I/O Operations Per Second(Sustained )
6
© 2009 IBM Corporation11
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
IOA Cache Hit SSD 15k RPM HDD
Short Seek
15k RPM HDD
Long Seek
0
1
2
3
4
5
6
7
8
9
0.1 0.33
3.9
8
4KB/op Read Response TimeR
t (m
s)
© 2009 IBM Corporation12
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Sweet Spots for SSD
• Where server performance is really important and I/O dependent
• Where server has a lot of HDD with low capacity utilization (or ought to have them configured this way)
• Where high value (from a performance perspective) can be focused in SSD
– Fairly small portion of total data (“hot data”)
– Specific indexes/tables/files/libraries of the operating system or data base or application
• Best workload characteristics for SSD
– Lots of random reads … and a low percentage of sequential/predictable reads
– Higher percentage reads than writes
• Assuming a disk adapter/controller with enough write cache, SSD writes may or may not be that
much faster than HDD writes
7
© 2009 IBM Corporation13
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Not So Sweet Spots for SSD
• Where server performance with a modest number of HDD is just fine
• Workload characteristics for SSD (and/or cached disk controllers)
– Predictable reads – non-random (especially with read cache controller for HDD)
– High percentage write (especially with write cache controller)
• Where data usage is evenly spread out (no hot data)
• Other examples where SSD won’t make much difference vs HDD
– Boot drives / load source drives (IPLs not disk I/O intensive)
– Journals/journal receivers (heavy writes – if leverage write cache on controller)
– Application binaries
© 2009 IBM Corporation14
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
• Endurance
– Cheaper, “consumer” SSD have a reputation for “wearing out”
– IBM Power SSD have great SSD characteristics - more reliable than HDD
– Over-provisioned – 69 GB available, 128 GB used
• Wear leveling – writes are spread around
• Approximate maximum of 100,000 writes to a cell (SLC)
• ECC – detection and correction
• Bad block relocation if problem detected
• DRAM used to buffer frequently-accessed data and eliminate WRITEs– Like HDD, provides warning to system to call for service if drive starting to wear out
• Performance
– Unlike consumer SSD, Power SSD deliver high performance – consistently, continually – under heavy enterprise workloads 24x7x367
– NAND flash memory can only be electronically erased or reprogrammed in large blocks (as opposed to small blocks, such as a byte of data, for instance).
– The NAND-Flash (the logical “Not And” operation) as used in Power and the DS8000 is working page-based (sector-based) and block-based. A page typically has a size of 512, 2048, 4096, or 8192 bytes. Each page has also a spare area of 64 bytes that is reserved for Error Correction Code, or other internal operations information. Pages are then grouped into blocks. A page is the smallest unit that can be read or written.
– Pages that contain data cannot be directly overwritten with new data, they must be erased first before they can be reused. The page itself cannot be erased because of the grouping into blocks, so the block needs to be erased (erasure procedure takes typically 1.5 to 2.0 ms).
– For each write operation, a copy of the block is needed and is performed at that time. The block that was used before is blocked until it is erased.
Power SSD Technology Insights
8
© 2009 IBM Corporation15
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Placing a charge (electrons) on the floating-gate is called programming or writing, whereas removing the charge from the floating-gate is called erasing.
The current commercial NAND-flash has two varieties of flash cells:
�Single Level Cell Flash (SLC-Flash)
�SLC-Flash technology manipulates the charge on the floating gate of the flash transistor cell to allow representation of two (voltage) states, which translates to a single bit per cell.
�The bit value is a 0 (written state) or a 1 (erased state),
�A Single Level Cell flash typically allows about 100,000 writes per cell. The number of writes to the same cell is limited because each write operation wears out the insulation of the floating gate.
�Multi Level Cell Flash (MLC-Flash)
�MLC-Flash is designed to allow a more precise amount of charge on the floating gate of the transistor to represent four different states, thereby translating to two bits of information per cell and therefore a higher bit density is possible.
�A Multi Level Cell only allows about 10,000 writes per cell and the MLC-Flash.
Although an MLC-Flash can store more information, but...the lifetime of the SLC-Flash is about ten times higher than the MLC-Flash
MLC-Flash is slower in writes per cell than the SLC- Flash.
For those reasons, the Solid State Drives available for Power and the DS8000 use a NAND-Flash with SLC technology.
Data stored in the SLC-flash remains valid for about ten years, without power.
NAND Based Flash
© 2009 IBM Corporation16
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD Agenda
• Technology / benefit overview
• Implementing Hot/Cold – AIX & IBM i
• Examples/Scenarios
• Technology/Pricing
• Configuring SSD
• Additional information
9
© 2009 IBM Corporation17
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Implementing “Hot” and “Cold” Data
IBM i• #1 Best integrated, automated
capability available in the industry today
– “Trace and Balance” function part of IBM i
– Monitors by partition or ASP (Aux Storage Pool) to determine hot/cold
– Upon request, automatically moves hot data to SSD, cold to HDD
– Can re-monitor & rebalance any time
• A few key OS files can automatically be placed on SSD
• Can specify specific data base objects to be placed on SSD
• See white paper released late May for additional insights
• Additional enhancements being developed
AIX• Ability to granularly/flexibly select and
locate hot files on SSD
– Key AIX performance tools are filemon& iostat.
– Database vendors provide hot data analysis tools, example: DB2 Snapshot.
• Migrating hot data
– Migratepv is often useful
– Softek Migration Tool provides a nice suite of functions and can be combined with IBM Services
– See Total Storage Productivity Center
• SSD white paper released late April for additional insights
• New/enhanced tools being developed
© 2009 IBM Corporation18
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
AIX – Choosing Data to Place on SSD
• iostat – identify IOPS (tps) and R/W ratio for PV
– High small-block tps and high R/W ratio suggests good candidate
– iostat –t: if there is no iowait time, SSDs will not improve performance
– iostat –D: Total system storage tps may be substantially higher SDD vs HDDlook for hdisks that do over 200 IOPS (tps) or %tm act at least 99% busy
– Investigate further with lvmstat and # lspv –l <hdisk#>
• lvmstat – identify IOPS (iocnt) and R/W ratio for LVs
– Turn on lvmstat for VG with # lvmstat –e –v <vgname>
# lvmstat -v newvg2
Logical Volume iocnt Kb_read Kb_wrtn Kbps
…
– High iocnt and high R/W ratio LVs are good candidates
– Also reports IOPS on PPs – useful when the LV is relatively large
• filemon – identify IO sizes, sequentiality to PVs, LVstrace for only seconds of timeSee sorted LV utilization in “Most Active Logical Volumes” report. High utilization is good SSD candidate.
• Some applications have tools for identifying hot data
– DB2 snapshot monitoring tool
10
© 2009 IBM Corporation19
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
AIX – Moving Data to SSDs
• Add SSD hdisk to VG with
–# extendvg <vgname> <hdisk#>
• Migrate LV to hdisk dynamically with
–# migratepv –l <lvname> <source hdisk> <target hdisk(s)>
–Repeat with other hdisks the LV resides on
• Or create a new VG for just SSD data
–# mkvg –y ssdvg –s 32 –S <ssdhdisk(s)>
–Offers smaller PP sizes to waste less space
• Stop application and copy the LV to the new VG with
–# cplv –v <sourcevg> -y ssdlv <source LV>
–Or backup/restore/copy data to new file system
© 2009 IBM Corporation20
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
IBM i Load Balancer
• Industry leading automated capability
• Monitors partition/ASP using “trace”
– User turns trace on during a peak time
– User turns trace off after reasonable
sample time
– Negligible performance impact expected
– Tool monitors “reads” to identify hot data
– Looks a 1MB stripes of data
• Upon command, automatically moves hot
data to SSD, cold data to HDD
– Minimal performance impact, done in
background
• Can remonitor and rebalance any time
– Probably a weekly or monthly activity
– Perhaps less often if data not volatile
IBM i intelligent hot/cold placement makes a big difference vs normal IBM striping / scattering of data across all drives.
This example 72 HDD + 16 SSD
Trans/min
Ap
pli
cati
on
Resp
on
se t
ime
72 HDD + 16 SSD No Balance
72 HDD + 16 SSD Data Balanced
Predicting/analyzing what % of data is hot for presale analysis to help size the number of SSD required:
• Use PEX tool/output
• Output from monitor (will need technical person to interpret)
11
© 2009 IBM Corporation21
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD Balancing One of Five Types of ASP Balancing
1. Balance data between busy units and idle units (STRASPBAL TYPE(*USAGE))
2. Make all of the units in the ASP have the same percent full (STRASPBAL TYPE(*CAPACITY))
3. Drain the data from a disk, to prepare unit it to be removed from the configuration (STRASPBAL TYPE(*MOVDTA))
4. (Almost obsolete) move hot data off of a compressed disk, and move cold data to the compressed disk (STRASPBAL
TYPE(*HSM)) Requires specific disk controllers with compression capability – feats #2741/2748/2778. Compression only allowed in user ASPs
5. Move cold data to HDDs and movehot data to SSDs (STRASPBAL TYPE(*HSM))
© 2009 IBM Corporation22
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Using TRCASPBAL to place hot data on SSDs
• Trace ASP balance counts the read operations based on 1MB stripes
– TRCASPBAL SET(*ON) ASP(1) TIMLMT(*NOMAX)
• Start ASP balance moves the data
– STRASPBAL TYPE(*HSM) ASP(1) TIMLMT(*NOMAX)
– Target is 50% of read operations to be on SSD
– Cold data is moved (multiple threads) to HDDs, hot data is moved (single thread) to SSD
HDD3 HDD4 SSDHDD2HDD1
100 500 100 2000 6000
10000 1200 6000 3000 100
300 800 500 900 900
0 4000 300 400 300
200 600 700 1000 100
100
300
900100
100
300
900100 10000
6000
4000
10000
6000
4000
12
© 2009 IBM Corporation23
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
How to Find Hot Tables and Indexes
• Performance Explorer
– BY FAR the best solution
– Perform analysis based on read complete and write complete events
• DB2 maintains statistics about the number of operations on a table or index
– Statistics are zeroed on each IPL
– Statistics only identify candidates (logical operations include both random and sequential operations)
– Available via:
• Display file description (DSPFD)
• Application programming interface (API) QUSRMBRD
• System i Navigator Health Center (V6R1 only)
• SQL catalog queries
© 2009 IBM Corporation24
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD Agenda
• Technology / benefit overview
• Implementing Hot/Cold– AIX & IBM i
• Examples/Scenarios
• Technology/Pricing
• Configuring SSD
• Additional information
13
© 2009 IBM Corporation25
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Example 1: AIX & Mixed SSD+HDD
• Power 550 running AIX
• 800 HDD
– 720 HDD in DS4800s with “hot” data
– 80 HDD in DS3400 with “cold” data
• 116 SSD+HDD
– 36 SSD in #5886 EXP12S Drawers “hot”
– 80 HDD in DS3400 with “cold” data
• Comparisons of 720 HDD to 36 SSD (“hot to hot”) even more impressive
90% reductionEnergy for I/O
1.65 X more Total transactions
10.8 X more
(SSD do 97% work)
SSD+HHD average drive throughput –I/O per second
86% reductionFewer total drives
3.1 X better ReadsSSD average I/O response time
36 SSD + 80 HDDSSD + HDD
800 HDDAll HDD (Hard Disk Drives)
OLTP / DB Sample* Workload
Compare all HDD vs mixed
SSD+HDD
* This sample workload documented in white paper released late April. Sample workload selected to stress I/O. Customer workloads, configurations and SSD impact will vary from this sample.
800 HDD 36 SSD +80 HDD
“Hot”
“Hot”
“Cold” “Cold”
© 2009 IBM Corporation26
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Example 2: #1 in BI for SAP & all SSD
• IBM i with Power 520 and 550 deliver top results for SAP Business Intelligence
workload in April 2009
• SAP BI Mixed Load Standard Application Benchmark
– Using SAP NetWeaver, query activity and load/update activity are executed in parallel
• Re-ran 550 benchmark with SSD to see what would happen ….
• Configuration– IBM i 6.1 with DB2 for i
– SAP NetWeaver 7.0
– 300,000,000 Records
SAP certified results are found at www.sap.com/benchmark
90,634 90,492 Throughput (Query Navigation Steps/per hour)
64 GB64 GBMemory
22 x SSD96 x 15K RPM DiskDisk/SSD
4-core 5.0 GHz4-core 5.0 GHzProcessor
550550
SSD enables the same results with:• 75% fewer drives• Smaller footprint• Energy savings
14
© 2009 IBM Corporation27
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Examples 3/4/5/6:
SSD Cheetah wins race
IBM i & OLTP/DB
• 89% fewer drives• 18% more transactions• 2.5X better user response time• 4.5X better disk response time• 11X more IOPs/device
(2) 1.5GB SAS adapters + 12 SSD
(3) 1.5GB SCSI adapters + 108 SCSI HDD
SCSI HDD to SAS SSD
IBM i & OLTP/DB
• 39% fewer drives• Same number transactions• 2-2.5X better user response time• 1.75X better disk response time• 2.1X more IOPs/device
(4) 1.5GB SAS adapters + 16 SSD + 72 HDD
(4) 1.5GB SAS adapters + 144 HDD
HDD to SSD+HDD
AIX & OLTP/DB
• Same number drives• 42X more transactions• 3X better database response time• 3.5X better disk response time• 42X more IOPs/device
(6) 1.5GB SAS adapters + 36 SSD(hot data only compare)
(6) 1.5GB SAS adapters + 36 15k rpm HDD(hot data only compare)
HDD to SSD
IBM i & OLTP/DB
• Same number drives• 11X more transactions• 10X better user response time• 19X better disk response time• 11X more IOPs/device
1.5GB SAS adapter + 8 SSD
1.5GB SAS adapter + 8 15k rpm HDD
HDD to SSD
OS & sample workloads used
Performance with SSD (Note: These results measured using sample workloads designed to exercise I/O and may differ noticeably from customer workloads.)
Config: With SSD
Config: HDD only
(all 15k rpm)
Example comparisons
3
4
5
6
SSD
© 2009 IBM Corporation28
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Example 7: Batch Window Reduction
• Associated Bank needed to reduce month end batch run time from 4+ hours to under 3 hours
• SSDs cut 1.5 hours from batch run time
– Plus a 16% reduction in # of disk drives
• Leveraged recent IBM i enhancements
– Directed 8 DB2 Objects (table, index, view) to SSD
Batch Performance Runs
0
1
2
3
4
5
Ho
urs
72 HDD 72 HDD + 8 SSD 60 HDD + 4 SSD
SSD run 2
SSD run 1
Base run
2:48460
2:43872
4:22072
Batch Run Time
# of SSDs
# of SAS Disk Drives
40% Reduction
Source: IBM Power Systems Performance and Benchmark Center 5-23-09
15
© 2009 IBM Corporation29
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Scenario: Complete HDD I/O Refresh
• Modernize older disk subsystem, replacing all old HDD with new SSD+HDD• Put “hot” active data on SSD & put “cold” data on spinning disk
– Combination has better price performance than using all SSD– IBM i load balancing function can do this nearly automatically
• Save space, maintenance, power/cooling, improve performance and increase storage capacity
Power 570 w/ older SCSI disk� (360) 35GB disk = 12.6 TB� 152 U rack space (4.1 racks)
Power 570 w/ SSD & newer SAS disk� (32) 69 GB SSD = 2.2 TB� (48) 282 GB disk = 13.5 TB� 36 U rack space (1 rack)
Assumptions:� Amount customer useable space net of parity protection not shown� Ratio of hot to cold data for this scenario = 1 : 6 Will vary by client and application � * Impact/benefit of using a higher % of drive’s capacity not considered� Example uses 4 partitions equal in size and configuration for disk and controllers
4.5 to 1
Drive reductionAnd grow capacity 25%*
© 2009 IBM Corporation30
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Scenario: Existing HDD Refresh - SSD Turbo-charger
Problem: Older HDD not meeting performance needs• Normal tuning efforts have not worked• Prefer not to mix newer HDD with older HDD • Leases/budgets make it difficult to replace existing HDD• Prefer not to invest more money in additional older HDD
Solution: Add SSD and keep HDD. Put “hot” data on SSD & keep “cold” data on HDD• SSD gives performance boost to system …. SWAG 20-40%• SSD provides new technology investment• HDD usable capacity can be larger SWAG 20-30%...as average capacity utilization increased• HDD financial investment protected/leveraged
Power 570 w/ older SCSI disk(360) 35GB disk = 12.6 TB
(12) 69 GB SSD = 756 GPlus xx% more GB usable from HDD
Assume
• older drives 30% data capacity
• 80-20 cold-hot split
Then: 12 SSD covers hot data needsSystem configuration and usage will impact results and applicability
16
© 2009 IBM Corporation31
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Scenario: Existing SAN Refresh - SSD Turbo-charger
Problem: Older HDD in SAN not meeting performance needs
• Normal tuning efforts have not worked
• Leases/budgets make it difficult to replace existing SAN
• Prefer not to invest more money in additional older SAN resource
Solution: Add SSD and keep SAN. Put “hot” data on SSD & keep “cold” data on SAN
• SSD gives performance boost to system …. SWAG 20-40%
• SSD provides new technology investment
• SAN usable capacity can be larger SWAG 10-20%...as average capacity utilization increased
• HDD financial investment protected/leveraged
Same scenario works for SAN
(12) 69 GB SSD = 756 GPlus xx% more GB usable from HDD
Assume
• older drives 30% data capacity
• 80-20 cold-hot split
Then: 12 SSD covers hot data needsSystem configuration and usage will impact results and applicability
Note – operational considerations as some of the data is pulled off SAN assuming using SAN advanced function -- if simple usage, not an issue
© 2009 IBM Corporation32
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD – Internal/External or SAN Scenario
• Flexibility to choose where to implement
– Power BladeCenter JS23/JS43
– Power 520/550/560/570 CEC
– EXP12S SAS Expansion Unit #5886
• On Power 520/550/560/570/575/595
– DS8000 (uses very similar SSD – a Fibre Channel
version instead of SAS)
• All above options can offer the best price/performance if hot data is put on the SSD and cold data on HDD
• As you would expect, tooling to identify/manage the data is somewhat different by OS and if SAN is used or not
• Using non-SAN SSD for hot data and SAN for cold can be done and can provide a performance boost, but if using robust SAN functions, carefully consider operational aspects. For example, if the SAN is doingreplication/backup, this would now need to be coordinated with the SSD to ensure all data was handled.
SSD
HDD
17
© 2009 IBM Corporation33
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD Agenda
• Technology / benefit overview
• Implementing Hot/Cold– AIX & IBM i
• Examples/Scenarios
• Technology/Pricing
• Configuring SSD
• Additional information
© 2009 IBM Corporation34
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Power SSD Offering Technology
• Enterprise Grade Solid State Drive (SSD)
– Built in wear leveling
• Rated capacity: 69.7 GB
– Actually has 128 GB for industrial strength
– Extra 83% capacity for long life of drive
• First SAS SSD in industry
– Higher performance interface
– Higher levels of redundancy/reliability
• SAS Interface ( 3 Gb )
– 2.5 / 3.5 inch inserts/carriers
• Performance Throughput Sustained:
– 220MB/s Read
– 115MB/s Write
• Random transactional operations (IOPS)
– 28,000 IOPS
• Average Access time:
– Typically around 200 microseconds
• Power Consumption: 8.4W max, 5.4W idle
– Same as SFF 15k HDD
– About half 3.5” 15K HDD (16-18W for today’s larger capacity)
2.5 inch ( SFF )
3.5 inch
18
© 2009 IBM Corporation35
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Pricing Comparison
• Price per 69 GB SSD = MUCH higher “per drive” vs HDD
• Great price per performance (I/O Operations Per Second)
• Therefore it makes sense to apply SSD where it offers the best impact for both performance and price/performance
– Use SSD where high performance I/O really matters
– Use HDD where cost per GB storage is important
– Using BOTH SSD and HDD on the same configuration can make a LOT of sense
HDD• Lower $ per GB capacity• Can need lots more I/O
devices for good performance
SSD• Higher $ per GB capacity• Higher I/O performance• Fewer devices = less
energy, floor space
All prices shown are IBM's USA suggested list prices as of April 2009 and are subject to change without notice; reseller prices may vary.
SFF SSD
SFF HDD
© 2009 IBM Corporation36
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD/HDD Maintenance Pricing Insights
• Maintenance price of IBM HDD shown as zero …. not really true – Price buried in maintenance price of CEC or I/O expansion unit with disk bays for
maintenance pricing simplicity.
• IBM SSD is designed to be reliable– Far more reliable comparing chance of 1 SSD failing versus chance of 1 out of
several HDD failing (using scenario of on a few SSD replacing many HDD)– No moving parts – Sophisticated management in SSD of flash memory being used to avoid over using
storage locations (wear leveling)– Huge amount of spare storage in SSD to extend life (83% of SSD actual capacity
reserved)
• But SSD cost results in a maintenance price per SSD, even with its reliability
• Interesting math exercise– Calculate the imputed cost per month of maintenance of HDD – Multiply that cost above by the ratio of (SSD-price)/(HDD price) – Compare it to SSD price– You’ll find that you would have expected a significantly higher SSD
maintenance price ………
19
© 2009 IBM Corporation37
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD Agenda
• Technology / benefit overview
• Implementing Hot/Cold– AIX & IBM i
• Examples/Scenarios
• Technology/Pricing
• Configuring SSD
• Additional information
© 2009 IBM Corporation38
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Configuring SSD - The Basics SFF SSD
SFF HDD• SSD leverages SAS HDD infrastructure
– Located in a SAS drive bay – same as SAS HDD would use
• Subset of SAS HDD location options
– Run by SAS adapter/controller – same as HDD would use
• Subset of SAS adapters
– Contents can be protected by RAID or mirroring same as HDD
• RAID-5, RAID-6, RAID-10, mirroring, hot spare
• Same rule as HDD: optionally protected AIX/Linux, required protection IBM i
– A number of specific configuration rule differences – see details
• POWER6 servers where SSD located in
– BladeCenter JS23/JS24
– Power 520, 550, 560, 570 CEC
– #5886 EXP12S Disk Drawer, #5802/5803* 12X PCIe I/O drawer
• AIX 5.3, IBM i 5.4, Linux SES10, RHEL4.5 or later
* added Oct 2009
20
© 2009 IBM Corporation39
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Pre-requisite Details• AIX 5.3
– with the 5300-07 Technology Level and Service Pack 9– with the 5300-08 Technology Level and Service Pack 7– with the 5300-09 Technology Level and Service Pack 4– with the 5300-10 Technology Level
• AIX 6.1 – with the 6100-00 Technology Level and Service Pack 9– with the 6100-01 Technology Level and Service Pack 5– with the 6100-02 Technology Level and Service Pack 4– with the 6100-03 Technology Level
• IBM i V5R4M5 LIC and V5R4 OS– Cumulative PTF package C9104540– Respin RS545-F LIC, RS540-30 OS (installation not needed, but need CD for system reload purposes)– Database group 22 (SF99507)– The following PTFs should be in temporary, permanent or superseded status on the system:
• MF46591, MF46593, MF46594, MF46595, MF46743*, MF46748, SI35126, SI35365 (* = DELAYED PTF)• IBM i V6R1M0 LIC and OS
– Cumulative PTF package C9111610– Respin RS610-F, RS640-00 OS – (installation not needed, but need CD for system reload purposes)– Database group 10 (SF99601)– The following PTFs should be in temporary, permanent or superseded status on the system:
• MF46390, MF46518, MF46587, MF46588, MF46609, MF46714, MF46771, MF46817, MF47076*, MF47224*, SI35299, SI35379, SI35572 (* = DELAYED PTF)
• SUSE Linux Enterprise Server 10, Service Pack 2 or later• Red Hat Enterprise Linux version 4.7 or later• Red Hat Enterprise Linux version 5.2 or later• System firmware FW 3.4.2
– 520 – EL340_075; 550 – EL340_075; 560 – EM340_075; 570 – EM340_075; 595 – EH340_075 + EB340_078• HMC V7 R3.4.0 Service Pack 2
© 2009 IBM Corporation40
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD Feature Codes
One 69 GB SSD Drive, but five feature codes
– Three packages – BladeCenter, 2.5” SFF and 3.5”
– Two SFF and two 3.5” features help eConfig know if AIX/Linux or IBM i will use the drive
• No physical difference between SSD with AIX/Linux feature vs i feature
• Ignoring eConfig’s confusion, can move SSD from AIX�i or i�AIX partition whenever needed (remembering normal need to “empty” contents and remove from RAID set before moving)
JS23/JS43
520/550 CECor
#5802/5803
560/570 CECor
EXP12S #5886
#8273
SFF#1890 #1909(i)
3.5”
#3586 #3587(i)
SSD Placed in
21
© 2009 IBM Corporation41
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SAS Adapters Supporting SSD on Power 520/550/560/570/595
Three SAS adapter/controller options:– Most powerful: PCI-X 1.5GB Cache RAID Adapter AIX / IBM i / Linux
– Mid choice: PCIe 380MB Cache RAID Adapter AIX / IBM i (Oct 2009) / Linux
– Least costly: imbedded controller* AIX / IBM i / Linux– SAS Adapters not used: PCI-X #5900/5912, PCIe #5901, PCI-X #5902
*175MB Cache RAID #5679 optional on Power 520/550, but often recommended
• PCI-X SAS RAID Adapter CCIN 572F
– Feat code #5904/5906/5908 (all same card, but 3 features indicate
double-wide blind swap cassette)
– 1.5 GB effective write cache
– Read cache disabled using SSD
– Double wide adapter - uses 2 PCI-X slots
• PCIe SAS RAID Adapter CCIN 574E
– Feat Code #5903
– 380 MB write cache
– Used in pairs
• Integrated SAS Controller CCIN 572C
– Recommended cache daughter card FC 5679
– Systems: 520, 550, 560, & 570
pair
© 2009 IBM Corporation42
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SAS Adapters Supporting SSD on Power 520/550/560/570/595
Three SAS adapter/controller options:– Most powerful: PCI-X 1.5GB Cache RAID Adapter AIX / IBM i / Linux
– Mid choice: PCIe 380MB Cache RAID Adapter AIX / IBM i (Oct 2009) / Linux
– Least costly: imbedded controller AIX / IBM i / Linux– SAS Adapters not used: PCI-X #5900/5912, PCIe #5901, PCI-X #5902
520/550CEC
560/570CEC
EXP12S#5886
#5802#5803
Imbedded controller*#5904 PCI-X 1.5GB Adapter
Imbedded controller#5903 PCIe 380MB Adapter#5908 PCI-X 1.5GB Adapter
#5903 PCIe 380MB Adapter#5904/6/8 PCI-X 1.5GB Adapter
#5903 PCIe 380MB Adapter
SFF#1890
#1909(i)
3.5”
#3586 #3587(i)
3.5”
#3586 #3587(i)
SFF#1890
#1990(i)
*175MB Cache RAID #5679 optional, but often recommended
This option announced Oct 2009
22
© 2009 IBM Corporation43
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Configuring SSD - 520/550 CEC Details
520/550CEC
Imbedded controller*#5904 PCI-X 1.5GB Adapter
SFF#1890 #1909(i)
• Use 520/550 CEC SFF SAS bays, NOT 3.5” SAS bays
• Can not mirror SSD against HDD, can not mix in same RAID array
• Using imbedded controller for all 8 SSF bays -- AIX / IBM i / Linux
– * Can augment performance/protection with #5679 175MB cache/RAID enabler
– Can mix HDD and SSD in these 8 bays
• Need enough HDD and enough SSD for each protection scheme used
• Using Split Backplane AIX / Linux Not IBM i
– Use imbedded controller for first four SSF bays
• Can augment performance/protection with #5679 175MB cache/RAID enabler
• Can NOT mix HDD and SSD in these 4 bays– Need enough SSD for protection scheme used
– Use #5904 PCI-X 1.5GB Cache RAID Adapter for second four SSF bays
• Can NOT mix HDD and SSD in these 4 bays– Need enough SSD for protection scheme used
• #5904 located in 520/550 CEC PCI-X slots. – Connect to 520/550 backplane via SAS cables #3679+#3669/3670
• #5904 second/third ports not used – can not attach additional HDD or SSD
Note – #5903 not used. Neither required #3667 SAS cable introduced nor testing done
© 2009 IBM Corporation44
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Configuring SSD - 560/570 CEC Details (1 of 2)
• Use 3.5” SAS bays (no SFF bays offered in 560/570 CEC)• Can not mirror SSD against HDD, can not mix in same RAID array• Using imbedded controller for all 6 3.5” bays -- AIX / IBM i / Linux
– Can mix HDD and SSD in these 6 bays• Need enough HDD and enough SSD for each protection scheme used
• Using PCI adapter(s) to drive all 6 bays AIX / IBM i / Linux(Awkward use of term “split backplane”)
– Use pair #5903 PCIe adapters AIX / IBM i (Oct 2009) / Linux • Can NOT mix HDD and SSD in these 6 bays
– Need enough SSD for protection scheme used
• #5903 adapters located in 560/570 CEC PCIe slots– Connect to 560/570 backplane via SAS cable #3667+#3651
• #5903 second pair ports not used – can not attach additional HDD or SSD
– Use #5908 PCI-X 1.5GB Cache RAID Adapter AIX / IBM i / Linux• Can NOT mix HDD and SSD in these 6 bays
– Need enough SSD for protection scheme used
• #5908 located in 560/570 CEC PCI-X slots. – Connect to 560/570 backplane via SAS cables #3679+#3651
• #5908 second/third ports not used – can not attach additional HDD or SSD
560/570CEC
Imbedded controller#5903 PCIe 380MB Adapter#5908 PCI-X 1.5GB Adapter
3.5”
#3586 #3587(i)
23
© 2009 IBM Corporation45
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Configuring SSD - 560/570 CEC Details (2 of 2)
• Use 3.5” SAS bays (no SFF bays offered in 560/570 CEC)• Can not mirror SSD against HDD, can not mix in same RAID array• Using Classic Split backplane (3+3) AIX / Linux Not IBM i• Using imbedded controller for first three bays AIX / Linux Not IBM i
– Can Not mix HDD and SSD in these 3 bays• Need enough SSD for each protection scheme used
• Using PCI adapter(s) to drive second three bays AIX / Linux Not IBM i– Use pair #5903 PCIe adapters AIX / Linux
• Can NOT mix HDD and SSD in these 3 bays– Need enough SSD for protection scheme used
• #5903 adapters located in 560/570 CEC PCIe slots– Connect to 560/570 backplane via SAS cable #3667+#3650
• #5903 second pair ports not used – can not attach additional HDD or SSD
– Use #5908 PCI-X 1.5GB Cache RAID Adapter AIX / Linux• Can NOT mix HDD and SSD in these 3 bays
– Need enough SSD for protection scheme used
• #5908 located in 560/570 CEC PCI-X slots. – Connect to 560/570 backplane via SAS cables #3679+#3650
• #5908 second/third ports not used – can not attach additional HDD or SSD
560/570CEC
Imbedded controller#5903 PCIe 380MB Adapter#5908 PCI-X 1.5GB Adapter
3.5”
#3586 #3587(i)
© 2009 IBM Corporation46
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Configuring SSD - #5886 EXP12S Details
• EXP12S has twelve 3.5” SAS bays, but max of 8 bays used
– One pair #5903 or one #5904/6/8 adapter runs all bays
– Can not mix HDD and SSD in the EXP12S
– Need enough SSD for protection scheme used
– Can not cascade into second EXP12S via EE SAS cables
• Can not mirror SSD against HDD, can not mix in same RAID array
• Using pair #5903 PCIe adapters AIX / IBM i (Oct 2009) / Linux
– #5903s in PCIe slots in 520/550/560/570/575 CEC or PCIe 12X I/O drawer
– Attached to EXP12S using SAS X cable
– #5903 second pair ports not used – can not attach additional HDD or SSD
• Using #5904/6/8 PCI-X 1.5GB Cache RAID Adapter AIX / IBM i / Linux
– #5904/6/8 in PCI-X slots in 520/550/560/570 CEC or PCI-X 12X I/O drawer or PCI-X RIO/HSL I/O drawer
– Attached to EXP12S using SAS YO cable
– #5904/6/8 second/third ports not used – can not attach additional HDD or SSD
EXP12S#5886
#5903 PCIe 380MB Adapter#5904/6/8 PCI-X 1.5GB Adapter
3.5”
#3586 #3587(i)
24
© 2009 IBM Corporation47
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Configuring SSD - #5802/5803 12X PCIe I/O Drawer Details
• #5802 has 18 SFF SAS bays – bays can be “partitioned”
– Into one set of 18 (AIX/Linux) or 2 sets or 9 (AIX/IBM i/Linux)
• #5803 has 26 SFF SAS bays – bays can be “partitioned”
– Into one set of 26 (AIX/Linux) or 2 sets or 13 (AIX/IBM i/Linux)
• Using pair #5903 PCIe adapters (Oct 2009) AIX / IBM i / Linux
– Max of two pairs of #5903 per #5802/5803
– Max 9 SSD per pair #5903 (SSD located in max of one set of bays)
– Thus max 18 SSD per #5802/5803 with two pairs of #5903
– Rules• Can not mix HDD and SSD in the same set of bays
• Can mix HDD and SSD in the same drawer, but on different SAS adapters
• Need enough SSD for protection scheme used
• Can not mirror SSD against HDD, can not mix in same RAID array
• Using SSD, #5903 second pair ports not used – can not attach additional HDD or SSD
– #5903s located in PCIe slots of #5802/5803 I/O drawer• #5903 SAS ports attached to #5802/5803 with SAS AT cables #3688
#5802#5803
#5903 PCIe 380MB AdapterSFF#1890
#1909(i)pairThis option announced Oct 2009
© 2009 IBM Corporation48
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
AIX SSD Configuration
• Devices are initially configured as pdisks
– # lsdev -Cc pdisk
– pdisk0 Available 02-08-00 Physical SAS Disk Drive
– pdisk1 Available 02-08-00 Physical SAS Disk Drive
• smitty devices -> Disk Array -> IBM SAS Disk Array -> IBM SAS Disk Array Manager
• Create an Array Candidate pdisk and Format to 528 Byte Sectors
• Create a SAS RAID array
– RAID 0, 5, 6 or 10 (2 disk RAID 10 = RAID 1)
– 16, 64 or 256 KB stripe (aka strip) size
• RAID 5 will be popular (because cost of drives & highly reliable)
– An hdisk appears
• # lsdev -Cc disk | grep "SAS RAID"
• hdisk3 Available 02-08-00 SAS RAID 0 Disk Array
• hdisk5 Available 02-08-00 SAS RAID 0 Disk Array
• Choose whether or not to turn on write cache for the adapter
• # sissasraidmgr -K 4 -l <sissas#> -o 1 (disable) | 0 (enable)
• Probably do not create hot spares (because cost of drives & highly reliable)
• Proceed to LVM configuration
25
© 2009 IBM Corporation49
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
AIX SSD Configuration -
© 2009 IBM Corporation50
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
AIX SSD Configuration
• From Disk Array Manager menu -> Diagnostics and Recovery Options -> Change/Show SAS RAID Controller
• Using adapter’s cache with SSD may or may not be best performance
option – is workload/data dependent.
– TPC benchmark
workload & test data
found “Disabled” the
best option
– HA configurations with
SSDs, disabled cache
probably best
26
© 2009 IBM Corporation51
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
DB2 and SSD integration for IBM i – CL enhancements
• IBM i V5R4 and V6R1
– CRTPF, CRTLF, CHGPF, CHGLF, CRTSRCPF, and CHGSRCPF commands
enhanced to indicate preference for placement on SSDs
• V5R4 – example– CRTPF lib1/pf1 SRCFILE(libsrc/dds) UNIT(*255)
– CHGPF lib1tst/pf1 UNIT(*255)
• V6R1 – example– CRTPF lib1/pf1 SRCFILE(libsrc/dds) UNIT(*SSD)
– CHGPF lib1tst/pf1 UNIT(*SSD)
– Delivered via Database Group PTF plus additional PTFs
• V5R4 SF99504
– Version 22 (Recommended minimum level)
• V6R1 SF99601
– Version 10 (Recommended minimum level)
• Capabilities are continuously being added
to DB2. You should stay current to take
advantage of them. See techdocs
entry TD105280.
Notes:
– When using the CRTPF, CRTLF, CHGPF, CHGLF commands, if table or physical file has multiple partitions or members, the media preference applies to all partitions or members.
– An exclusive lock is required to change the PF/LF and is released.
– Movement will be asynchronous (expected delivery 4Q09).
© 2009 IBM Corporation52
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
DB2 and SSD integration for IBM i – SQL enhancements
• IBM i V6R1 SQL support
– UNIT SSD on the object level
• CREATE TABLE
• ALTER TABLE
• CREATE INDEX
– UNIT SSD on the partition level
• CREATE TABLE
• ALTER TABLE ... ADD PARTITION
• ALTER TABLE ... ALTER PARTITION
Example: add partition for current year and
place on SSD
Example: move partition with last year’s data
back to HDDs
27
© 2009 IBM Corporation53
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
SSD Agenda
• Technology / benefit overview
• Implementing Hot/Cold– AIX & IBM i
• Examples/Scenarios
• Technology/Pricing
• Configuring SSD
• Additional information
© 2009 IBM Corporation54
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
New SSD Analyzer Tool for IBM i
SSD ANALYSIS TOOL (ANZSSDDTA)
Type choices, press Enter.
PERFORMANCE MEMBER . . . . . . . *DEFAULT__ Name, *DEFAULTLIBRARY . . . . . . . . . . . __________ Name
Additional Parameters
REPORT TYPE . . . . . . . . . . *SUMMARY *DETAIL, *SUMMARY, *BOTH TIME PERIOD:: START TIME AND DATE:: BEGINNING TIME . . . . . . . . *AVAIL__ Time, *AVAIL BEGINNING DATE . . . . . . . . *BEGIN__ Date, *BEGIN END TIME AND DATE:: ENDING TIME . . . . . . . . . *AVAIL__ Time, *AVAIL ENDING DATE . . . . . . . . . *END____ Date, *END
NUMBER OF RECORDS IN REPORT . . 50__ 0 - 9999
Bottom F3=Exit F4=Prompt F5=Refresh F12=Cancel F13=How to use this display F24=More keys
Available via www.ibm.com/support/techdocs in “Presentations & Tools”. Search using keyword SSD
• Quick, easy, no-charge analysis • Looks at standard performance report output –• Provides “probably yes”, “probably no”, or “maybe
28
© 2009 IBM Corporation55
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
ANZSSDDTA – *SUMMARY output
SSD Data Analysis - Disk Read Wait Summary
Performance member Q224014937 in library @SUE
Time period from 2009-08-12-01.49.40.000000 to 2009-08-13-00.00.00.000000
--------------------------------------------------------------------------------------------------------------------------------
Disk read wait average response was 00.003058. Maybe candidate.
Bottom
F3=Exit F12=Cancel F19=Left F20=Right F24=More keys
© 2009 IBM Corporation56
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
ANZSSDDTA – *DETAIL output
SSD Data Analysis - Jobs Sorted by Disk Read Time Performance member Q224014937 in library @SUE
Time period from 2009-08-12-01.49.40.000000 to 2009-08-13-00.00.00.000000
CPU Disk Read Disk Read Disk Total Wait Total Wait Average Read Wait
Job Name Seconds Seconds Seconds /CPU ------------------------- --------- ----------- ------------ ---------QSPRC00001/QSYS/448980 38.010 4,276.730 .004677 113 DELSMBQPRT/SUEBA02/455198 67.096 3,551.437 .004724 53 QSPLMAINT/QSYS/448961 23.377 2,820.571 .004547 121 PERFNAVDS/SUEBA01/451039 15.865 862.070 .001861 54 WCSMBB01/SUEBAK/456767 144.285 774.387 .002174 5 WCSMBB02/SUEBAK/456856 49.446 589.355 .003625 12 QPADEV000F/SUEBA01/451414 7.612 544.305 .004620 72 SB055J/SUEBA01/453231 690.375 482.659 .002527 1 QCLNSYSPRT/QPGMR/456926 5.232 459.801 .005025 88 DELSMBQQPR/SUEBA02/455585 12.035 431.057 .004763 36
71 minutes of disk read wait time 113 seconds of disk read wait per second of CPU run time
29
© 2009 IBM Corporation57
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Four Power Systems SSD White Papers
• AIX-specific
– “Driving Business Value on Power Systems with Solid State Drives”
– First published April 2009
• IBM i-specific
– “Performance Value of Solid State Drives using IBM i”
– First published May 2009
• More SSD technology specific – AIX/IBM i/Linux appropriate
– “Performance Impacts of Flash SSDs Upon IBM Power Systems”
– First published June 2009
• Above papers in Power Systems web site under “Resources/Literature”
– http://www.ibm.com/common/ssi/apilite?infotype=SA&infosubt=WH&lastdays=1825&hitlimit=200&ctvwcode=US&pubno=POW*USEN&appname=STGE_PO_PO_USEN_WH&additional=summary&contents=keeponlit
• 4th paper recently released for an SAP environment
– http://www.sdn.sap.com/irj/sdn/db4?rid=/library/uuid/90a1637e-065f-2c10-3ab7-bea9375fc88d
© 2009 IBM Corporation58
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
i Oriented References
• IBM - Performance Management on IBM i Resource Libraryhttp://www.ibm.com/systems/i/advantages/perfmgmt/resource.html
• Performance Value of Solid State Drives using IBM ihttp://www.ibm.com/systems/resources/ssd_ibmi.pdf
• Performance Impacts of Flash SSDs Upon IBM Power Systemshttp://www.ibm.com/common/ssi/cgi-
bin/ssialias?infotype=SA&subtype=WH&htmlfid=POW03028USEN&attachment=POW03028
USEN.PDF&appname=STGE_PO_PO_USEN_WH
• Driving Business Value on Power Systems with Solid State Driveshttp://www.ibm.com/common/ssi/cgi-
bin/ssialias?infotype=SA&subtype=WH&htmlfid=POW03025USEN&attachment=POW03025
USEN.PDF&appname=STGE_PO_PO_USEN_WH
• IBM Systems Lab Services and Traininghttp://www.ibm.com/systems/services/labservices
• IBM Power Systems(i) Benchmarking and Proof-of-Concept Centers
http://www.ibm.com/systems/i/support/benchmarkcenters
30
© 2009 IBM Corporation59
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
AIX Oriented References
• Installing and configuring SSDs http://publib.boulder.ibm.com/infocenter/systems/scope/hw/index.jsp?topic=/iphal/iphalssdconfig.htm&resultof=%22%53%53%44%22%20&searchQuery=%53%53%44&searchRank=%31&pageDepth=%30
• Considerations for Solid-State Drives (SSD) http://publib.boulder.ibm.com/infocenter/systems/scope/hw/index.jsp?topic=/arebj/arebjsolidstatedrives.htm&resultof=%22%53%53%44%22%20&searchQuery=%53%53%44&searchRank=%30&pageDepth=%30
• SSD Wikihttps://www.ibm.com/developerworks/wikis/display/WikiPtype/Solid+State+Drives
• SSD movie by Nigel Griffiths https://www.ibm.com/developerworks/wikis/display/WikiPtype/Movies
© 2009 IBM Corporation60
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
THANKS,
Questions?
Evaluations: SMP16
31
© 2009 IBM Corporation61
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
1-8 ms
~100 ns
< 10’s ns
.1’s TB
GB - TB
< MB
CapacityAccessTime
150
10’s M
10’s - 100’s M
4 KB OPS
I/O
Load-store
ProgrammingModel
ProcessorRegisters &
Caches
Processor
DRAM
Rotating Disk
Storage Hierarchy
© 2009 IBM Corporation62
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
ProcessorRegisters &
Caches
Processor
DRAM
SSD
1-8 ms
~200 us
~100 ns
< 10’s ns
.1’s TB
GB - .1’s TB
GB - TB
< MB
CapacityAccessTime
150
10’s – 100’s K
10’s M
10’s - 100’s M
4 KB OPS
I/O
Load-store
ProgrammingModel
Rotating Disk
Storage Hierarchy – New Optimizations
32
© 2009 IBM Corporation63
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
This document was developed for IBM offerings in the United States as of the date of publication. IBM may not make these offerings available in other countries, and the information is subject to change without notice. Consult your local IBM business contact for information on the IBM offerings available in your area.
Information in this document concerning non-IBM products was obtained from the suppliers of these products or other public sources. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products.
IBM may have patents or pending patent applications covering subject matter in this document. The furnishing of this document does not give you any license to these patents. Send license inquires, in writing, to IBM Director of Licensing, IBM Corporation, New Castle Drive, Armonk, NY 10504-1785 USA.
All statements regarding IBM future direction and intent are subject to change or withdrawal without notice, and represent goals and objectives
only.
The information contained in this document has not been submitted to any formal IBM test and is provided "AS IS" with no warranties or guarantees either expressed or implied.
All examples cited or described in this document are presented as illustrations of the manner in which some IBM products can be used and the results that may be achieved. Actual environmental costs and performance characteristics will vary depending on individual client configurations and conditions.
IBM Global Financing offerings are provided through IBM Credit Corporation in the United States and other IBM subsidiaries and divisions worldwide to qualified commercial and government clients. Rates are based on a client's credit rating, financing terms, offering type, equipment type and options, and may vary by country. Other restrictions may apply. Rates and offerings are subject to change, extension or withdrawal without notice.
IBM is not responsible for printing errors in this document that result in pricing or information inaccuracies.
All prices shown are IBM's United States suggested list prices and are subject to change without notice; reseller prices may vary.
IBM hardware products are manufactured from new parts, or new and serviceable used parts. Regardless, our warranty terms apply.
Any performance data contained in this document was determined in a controlled environment. Actual results may vary significantly and are dependent on many factors including system hardware configuration and software design and configuration. Some measurements quoted in this document may have been made on development-level systems. There is no guarantee these measurements will be the same on generally-available systems. Some measurements quoted in this document may have been estimated through extrapolation. Users of this document should verify the applicable data for their specific environment.
Revised September 26, 2006
Special notices
© 2009 IBM Corporation64
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
IBM, the IBM logo, ibm.com AIX, AIX (logo), AIX 6 (logo), AS/400, BladeCenter, Blue Gene, ClusterProven, DB2, ESCON, i5/OS, i5/OS (logo), IBM Business Partner (logo), IntelliStation, LoadLeveler, Lotus, Lotus Notes, Notes, Operating System/400, OS/400, PartnerLink, PartnerWorld, PowerPC, pSeries, Rational, RISC
System/6000, RS/6000, THINK, Tivoli, Tivoli (logo), Tivoli Management Environment, WebSphere, xSeries, z/OS, zSeries, AIX 5L, Chiphopper, Chipkill, Cloudscape, DB2 Universal Database, DS4000, DS6000, DS8000, EnergyScale, Enterprise Workload Manager, General Purpose File System, , GPFS, HACMP, HACMP/6000, HASM, IBM Systems Director Active Energy Manager, iSeries, Micro-Partitioning, POWER, PowerExecutive, PowerVM, PowerVM (logo), PowerHA, Power Architecture, Power
Everywhere, Power Family, POWER Hypervisor, Power Systems, Power Systems (logo), Power Systems Software, Power Systems Software (logo), POWER2, POWER3, POWER4, POWER4+, POWER5, POWER5+, POWER6, POWER6+, System i, System p, System p5, System Storage, System z, Tivoli Enterprise, TME 10,
Workload Partitions Manager and X-Architecture are trademarks or registered trademarks of International Business Machines Corporation in the United States, other countries, or both. If these and other IBM trademarked terms are marked on their first occurrence in this information with a trademark symbol (® or ™), these symbols indicate U.S. registered or common law trademarks owned by IBM at the time this information was published. Such trademarks may also be registered or common law
trademarks in other countries. A current list of IBM trademarks is available on the Web at "Copyright and trademark information" at www.ibm.com/legal/copytrade.shtml
The Power Architecture and Power.org wordmarks and the Power and Power.org logos and related marks are trademarks and service marks licensed by Power.org.UNIX is a registered trademark of The Open Group in the United States, other countries or both.
Linux is a registered trademark of Linus Torvalds in the United States, other countries or both.
Microsoft, Windows and the Windows logo are registered trademarks of Microsoft Corporation in the United States, other countries or both.
Intel, Itanium, Pentium are registered trademarks and Xeon is a trademark of Intel Corporation or its subsidiaries in the United States, other countries or both.
AMD Opteron is a trademark of Advanced Micro Devices, Inc.
Java and all Java-based trademarks and logos are trademarks of Sun Microsystems, Inc. in the United States, other countries or both.
TPC-C and TPC-H are trademarks of the Transaction Performance Processing Council (TPPC).
SPECint, SPECfp, SPECjbb, SPECweb, SPECjAppServer, SPEC OMP, SPECviewperf, SPECapc, SPEChpc, SPECjvm, SPECmail, SPECimap and SPECsfs are trademarks of the Standard Performance Evaluation Corp (SPEC).
NetBench is a registered trademark of Ziff Davis Media in the United States, other countries or both.
AltiVec is a trademark of Freescale Semiconductor, Inc.
Cell Broadband Engine is a trademark of Sony Computer Entertainment Inc.
InfiniBand, InfiniBand Trade Association and the InfiniBand design marks are trademarks and/or service marks of the InfiniBand Trade Association. Other company, product and service names may be trademarks or service marks of others.
Revised April 24, 2008
Special notices (cont.)
33
© 2009 IBM Corporation65
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
The IBM benchmarks results shown herein were derived using particular, well configured, development-level and generally-available computer systems. Buyers should
consult other sources of information to evaluate the performance of systems they are considering buying and should consider conducting application oriented testing. For additional information about the benchmarks, values and systems tested, contact your local IBM office or IBM authorized reseller or access the Web site of the benchmark
consortium or benchmark vendor.
IBM benchmark results can be found in the IBM Power Systems Performance Report at http://www.ibm.com/systems/p/hardware/system_perf.html.
All performance measurements were made with AIX or AIX 5L operating systems unless otherwise indicated to have used Linux. For new and upgraded systems, AIX
Version 4.3, AIX 5L or AIX 6 were used. All other systems used previous versions of AIX. The SPEC CPU2006, SPEC2000, LINPACK, and Technical Computing benchmarks were compiled using IBM's high performance C, C++, and FORTRAN compilers for AIX 5L and Linux. For new and upgraded systems, the latest versions of these compilers were used: XL C Enterprise Edition V7.0 for AIX, XL C/C++ Enterprise Edition V7.0 for AIX, XL FORTRAN Enterprise Edition V9.1 for AIX, XL C/C++
Advanced Edition V7.0 for Linux, and XL FORTRAN Advanced Edition V9.1 for Linux. The SPEC CPU95 (retired in 2000) tests used preprocessors, KAP 3.2 for FORTRAN and KAP/C 1.4.2 from Kuck & Associates and VAST-2 v4.01X8 from Pacific-Sierra Research. The preprocessors were purchased separately from these vendors. Other
software packages like IBM ESSL for AIX, MASS for AIX and Kazushige Goto’s BLAS Library for Linux were also used in some benchmarks.
For a definition/explanation of each benchmark and the full list of detailed results, visit the Web site of the benchmark consortium or benchmark vendor.
TPC http://www.tpc.org
SPEC http://www.spec.org
LINPACK http://www.netlib.org/benchmark/performance.pdf
Pro/E http://www.proe.com
GPC http://www.spec.org/gpc
NotesBench http://www.notesbench.org
VolanoMark http://www.volano.com
STREAM http://www.cs.virginia.edu/stream/
SAP http://www.sap.com/benchmark/
Oracle Applications http://www.oracle.com/apps_benchmark/
PeopleSoft - To get information on PeopleSoft benchmarks, contact PeopleSoft directly
Siebel http://www.siebel.com/crm/performance_benchmark/index.shtm
Baan http://www.ssaglobal.com
Microsoft Exchange http://www.microsoft.com/exchange/evaluation/performance/default.asp
Veritest http://www.veritest.com/clients/reports
Fluent http://www.fluent.com/software/fluent/index.htm
TOP500 Supercomputers http://www.top500.org/
Ideas International http://www.ideasinternational.com/benchmark/bench.html
Storage Performance Council http://www.storageperformance.org/results Revised January 15, 2008
Notes on benchmarks and values
© 2009 IBM Corporation66
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Revised January 15, 2008
Notes on HPC benchmarks and valuesThe IBM benchmarks results shown herein were derived using particular, well configured, development-level and generally-available computer systems. Buyers should
consult other sources of information to evaluate the performance of systems they are considering buying and should consider conducting application oriented testing. For additional information about the benchmarks, values and systems tested, contact your local IBM office or IBM authorized reseller or access the Web site of the benchmark
consortium or benchmark vendor.
IBM benchmark results can be found in the IBM Power Systems Performance Report at http://www.ibm.com/systems/p/hardware/system_perf.html.
All performance measurements were made with AIX or AIX 5L operating systems unless otherwise indicated to have used Linux. For new and upgraded systems, AIX
Version 4.3 or AIX 5L were used. All other systems used previous versions of AIX. The SPEC CPU2000, LINPACK, and Technical Computing benchmarks were compiled using IBM's high performance C, C++, and FORTRAN compilers for AIX 5L and Linux. For new and upgraded systems, the latest versions of these compilers were used: XL C Enterprise Edition V7.0 for AIX, XL C/C++ Enterprise Edition V7.0 for AIX, XL FORTRAN Enterprise Edition V9.1 for AIX, XL C/C++ Advanced Edition V7.0 for Linux, and
XL FORTRAN Advanced Edition V9.1 for Linux. The SPEC CPU95 (retired in 2000) tests used preprocessors, KAP 3.2 for FORTRAN and KAP/C 1.4.2 from Kuck & Associates and VAST-2 v4.01X8 from Pacific-Sierra Research. The preprocessors were purchased separately from these vendors. Other software packages like IBM ESSL
for AIX, MASS for AIX and Kazushige Goto’s BLAS Library for Linux were also used in some benchmarks.
For a definition/explanation of each benchmark and the full list of detailed results, visit the Web site of the benchmark consortium or benchmark vendor.
SPEC http://www.spec.org
LINPACK http://www.netlib.org/benchmark/performance.pdf
Pro/E http://www.proe.com
GPC http://www.spec.org/gpc
STREAM http://www.cs.virginia.edu/stream/
Veritest http://www.veritest.com/clients/reports
Fluent http://www.fluent.com/software/fluent/index.htm
TOP500 Supercomputers http://www.top500.org/
AMBER http://amber.scripps.edu/
FLUENT http://www.fluent.com/software/fluent/fl5bench/index.htm
GAMESS http://www.msg.chem.iastate.edu/gamess
GAUSSIAN http://www.gaussian.com
ABAQUS http://www.abaqus.com/support/sup_tech_notes64.html
select Abaqus v6.4 Performance Data
ANSYS http://www.ansys.com/services/hardware_support/index.htm
select “Hardware Support Database”, then benchmarks.
ECLIPSE http://www.sis.slb.com/content/software/simulation/index.asp?seg=geoquest&
MM5 http://www.mmm.ucar.edu/mm5/
MSC.NASTRAN http://www.mscsoftware.com/support/prod%5Fsupport/nastran/performance/v04_sngl.cfm
STAR-CD www.cd-adapco.com/products/STAR-CD/performance/320/index/html
NAMD http://www.ks.uiuc.edu/Research/namd
HMMER http://hmmer.janelia.org/http://powerdev.osuosl.org/project/hmmerAltivecGen2mod
34
© 2009 IBM Corporation67
STG Technical Conferences 2009
Faster Than a Speeding Disk (Solid State Drives)
Revised April 2, 2007
Notes on performance estimates
rPerf for AIX
rPerf (Relative Performance) is an estimate of commercial processing performance relative to other IBM UNIX systems. It is derived from an IBM analytical model which uses characteristics from IBM internal workloads, TPC and SPEC benchmarks. The rPerf model is not intended to represent any specific public benchmark results and should not be reasonably used in that way. The model simulates some of the system operations such as CPU, cache and memory. However, the model does not simulate disk or network I/O operations.
• rPerf estimates are calculated based on systems with the latest levels of AIX and other pertinent software at the time of systemannouncement. Actual performance will vary based on application and configuration specifics. The IBM eServer pSeries 640 is the baseline reference system and has a value of 1.0. Although rPerf may be used to approximate relative IBM UNIX commercial processing performance, actual system performance may vary and is dependent upon many factors including system hardware configuration and software design and configuration. Note that the rPerf methodology used for the POWER6 systems is identical to that used for the POWER5 systems. Variations in incremental system performance may be observed in commercial workloads due to changes in the underlying system architecture.
All performance estimates are provided "AS IS" and no warranties or guarantees are expressed or implied by IBM. Buyers should consult other sources of information, including system benchmarks, and application sizing guides to evaluate the performance of a system they are considering buying. For additional information about rPerf, contact your local IBM office or IBM authorized reseller.
========================================================================
CPW for IBM i
Commercial Processing Workload (CPW) is a relative measure of performance of processors running the IBM i operating system. Performance in customer environments may vary. The value is based on maximum configurations. More performance information is available in the Performance Capabilities Reference at: www.ibm.com/systems/i/solutions/perfmgmt/resource.html