25
Copyright Gordon Bell Clusters & Grids The CC – GRID? Era Infinite processing, storage, and bandwidth @ zero cost and latency Gordon Bell ([email protected]) Bay Area Research Center Microsoft Corporation

Copyright Gordon Bell Clusters & Grids The CC – GRID? Era Infinite processing, storage, and bandwidth @ zero cost and latency Gordon Bell ([email protected])

  • View
    220

  • Download
    0

Embed Size (px)

Citation preview

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

The CC – GRID? EraInfinite processing, storage,

and bandwidth @ zero cost and latency

Gordon Bell ([email protected])

Bay Area Research Center

Microsoft Corporation

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

deja’ vu ARPAnet: c1969

– To use remote programs & data– Got FTP & mail. Machines & people overloaded.

NREN: c1988– BW => Faster FTP for images, data – Latency => Got http://www…– Tomorrow => Gbit communication BW, latency

<’90 Mainframes, minis, PCs/WSs >’90 very large, dep’t, & personal clusters

VAX: c1979 one computer/scientist Beowulf: c1995 one computer/scientist

1960s batch: opti-use allocate, schedule,$ 2000s GRID: opti-use allocate, schedule, $ (… security, management,

etc.)

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

Some observations Clusters are purchased, managed, and used as a single, one

room facility. Clusters are the “new” computers. They present unique,

interesting, and critical problems… then Grids can exploit them. Clusters & Grids have little to do with one another… Grids use

clusters! Clusters should be a good simulation of tomorrow’s Grid. Distributed PCs: Grids or Clusters? Perhaps some clusterable problems can be solved on a Grid…

but it’s unlikely.– Lack of understanding clusters & variants– Socio-, political, eco- wrt to Grid.

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

Some observations GRID was/is an exciting concept …

– They can/must work within a community, organization, or project. What binds it?

– “Necessity is the mother of invention.” Taxonomy… interesting vs necessity

– Cycle scavenging and object evaluation (e.g. seti@home, QCD)

– File distribution/sharing aka IP theft (e.g. Napster, Gnutella)

– Databases &/or programs and experiments(astronomy, genome, NCAR, CERN)

– Workbenches: web workflow chem, bio…– Single, large problem pipeline… e.g. NASA.– Exchanges… many sites operating together– Transparent web access aka load balancing– Facilities managed PCs operating as cluster!

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

Grids: Why?

The problem or community dictates a Grid

Economics… thief or scavenger Research funding… that’s where

the problems are

In a 5-10 years we can/will have: more powerful personal computers

– processing 10-100x; multiprocessors-on-a-chip– 4x resolution (2K x 2K) displays to impact paper– Large, wall-sized and watch-sized displays– low cost, storage of one terabyte for personal use

adequate networking? PCs now operate at 1 Gbps– ubiquitous access = today’s fast LANs– Competitive wireless networking

One chip, networked platforms e.g. light bulbs, cameras Some well-defined platforms that compete with the PC for mind (time)

and market sharewatch, pocket, body implant, home (media, set-top)

Inevitable, continued cyberization… the challenge… interfacing platforms and people.

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

SNAP … c1995

Scalable Network And Platforms A View of Computing in 2000+

We all missed the impact of WWW!

Gordon Bell Jim GrayNetworkPlatform

How Will Future Computers Be Built?

Thesis: SNAP: Scalable Networks and Platforms• Upsize from desktop to world-scale computer• based on a few standard components

Because: • Moore’s law:

exponential progress• Standardization & Commoditization• Stratification and competition

When: Sooner than you think!• Massive standardization gives massive use • Economic forces are enormous

NetworkPlatform

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

Volume drives simple,cost to standardplatforms

MPPs1-4 processor mP

1-20 processor mP

Distributed workstations

Clustered Computers

price for high speed

interconnect

price

performance

Stand-aloneDesk tops

PCs

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

ComputingSNAPbuilt entirelyfrom PCs Wide & Local

Area Networksfor: terminal,

PC, workstation,& servers

Centralized& departmental

uni- & mP servers(UNIX & NT)

Legacymainframes &

minicomputersservers & terms

Wide-areaglobal

network

Legacymainframe &

minicomputerservers & terminals

Centralized& departmental

servers buit fromPCs

scalable computers

built from PCs

TC=TV+PChome ...

(CATV or ATM or satellite)

???

Portables

A space, time (bandwidth), & generation scalable environment

Person servers (PCs)

Person servers (PCs)

MobileNets

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

SNAP Architecture----------

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

GB plumbing from the baroque:evolving from the 2 dance-hall model

Mp — S — Pc : | :

|——————-- S.fiber ch. — Ms | : |— S.Cluster |— S.WAN —

vs,

MpPcMs — S.Lan/Cluster/Wan — :

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

Modern scalable switches … also hide a supercomputer

Scale from <1 to 120 Tbps 1 Gbps ethernet switches scale to

10s of Gbps, scaling upward SP2 scales from 1.2

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

Interesting “cluster” in a cabinet

366 servers per 44U cabinet– Single processor– 2 - 30 GB/computer (24 TBytes)– 2 - 100 Mbps Ethernets

~10x perf*, power, disk, I/O per cabinet ~3x price/perf Network services… Linux based

*42, 2 processors, 84 Ethernet, 3 TBytes

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

ISTORE Hardware Vision

System-on-a-chip enables computer, memory, without significantly increasing size of disk

5-7 year target:MicroDrive:1.7” x 1.4” x 0.2”

2006: ?1999: 340 MB, 5400 RPM,

5 MB/s, 15 ms seek2006: 9 GB, 50 MB/s ? (1.6X/yr capacity, 1.4X/yr BW)

Integrated IRAM processor2x height

Connected via crossbar switchgrowing like Moore’s law

16 Mbytes; ; 1.6 Gflops; 6.4 Gops10,000+ nodes in one rack! 100/board = 1 TB; 0.16 Tf

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

The Disk Farm? or a System On a Card?

The 500GB disc cardAn array of discsCan be used as 100 discs 1 striped disc 50 FT discs ....etcLOTS of accesses/second of bandwidth

A few disks are replaced by 10s of Gbytes of RAM and a processor to run Apps!!

14"

Telnet & FTP

EMAIL

WWW Audio Video

Voice!Voice!

StandardsStandards

Increase Capacity(circuits & bw)

Lower response time

Create newservice

Increased Demand

The virtuous cycle of bandwidth supply and demand

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

Map of Gray Bell Prize resultsRedmond/Seattle, WA

San Francisco, CA

New York

Arlington, VA

5626 km10 hops

single-thread single-stream tcp/ip single-thread single-stream tcp/ip via 7 hopsvia 7 hops desktop-to-desktop …Win 2K desktop-to-desktop …Win 2K

out of the box performance*out of the box performance*

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

0

50

100

150

200

250

100Mbps Gbps SAN

Transmitreceivercpusender cpu

Time µs toSend 1KB

The Promise of SAN/VIA:10x in 2 years http://www.ViArch.org/

Yesterday: – 10 MBps (100 Mbps Ethernet)

– ~20 MBps tcp/ip saturates 2 cpus

– round-trip latency ~250 µs

Now– Wires are 10x faster

Myrinet, Gbps Ethernet, ServerNet,…

– Fast user-level communication

- tcp/ip ~ 100 MBps 10% cpu- round-trip latency is 15 us

1.6 Gbps demoed on a WAN

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

Labscape1st, 2nd, 3rd, or New Paradigm for science?

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

Labscape

Courtesy of Dr. Thomas Sterling, Caltech

Lessons from Beowulf

An experiment in parallel computing systems Established vision- low cost high end computing Demonstrated effectiveness of PC clusters for some (not all) classes

of applications Provided networking software Provided cluster management tools Conveyed findings to broad community Tutorials and the book Provided design standard to rally community!* Standards beget: books, trained people, software … virtuous cycle**observations

Courtesy, Thomas Sterling, Caltech.

Copyright Gordon Bell Clusters & GridsCopyright Gordon Bell Clusters & Grids

The End

How can GRIDs become a non- ad hoc computer structure?Get yourself an application community!