presented by fault tolerance challenges and solutions al geist computer science research group...

13
Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported by the Department of Energy’s Office of Science Office of Advanced Scientific Computing Research

Upload: christine-chafe

Post on 30-Mar-2015

216 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

Presented by

Fault Tolerance Challengesand Solutions

Al GeistComputer Science Research Group

Computer Science and Mathematics Division

Research supported by the Department of Energy’s Office of ScienceOffice of Advanced Scientific Computing Research

Page 2: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

2 Geist_FT_SC07

Example: ORNL LCFExample: ORNL LCFhardware roadmaphardware roadmap

54 TF(56 cabinets)5294 nodes

10,588 processors

21 TB

119 TF(124 cabinets)11,706 nodes

23,412 processors46 TB

300 TF(84 quad)

71 TB

1 PF (128 new cabinets)

175 TB

Rapid growth in scaledrives fault tolerance need

Jul2006

Nov2006

Dec2007

Nov2008

54 TF

100 TF

1 PF

250 TF

Cray XT3

Cray BakerCray XT4

25 TF

20X scale changein 2.5 years

20X scale changein 2.5 years

Today

Late2006

2009

2007

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

Page 3: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

3 Geist_FT_SC07

ORNL 1 PF Cray “Baker” system, 2009

Today’s applications and their runtime libraries may scale, but they are not prepared for the failure rates of sustained petascale systems

Assuming a linear model and a failure rate 20X what is seen today.

The RAS system automatically configures around faults – up for days.

But every one of these failures kills the application that was using that node!

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

Page 4: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

4 Geist_FT_SC07

Good news: MTTI is betterthan expected for LLNL BG/L

and ORNL XT4(6–7 days, not minutes).

Today’s fault-tolerance paradigm (checkpoint) ceases to be viable on large systems

Mea

n tim

e to

inte

rrup

t (ho

urs)

200

160

120

80

40

0November December January

Time to checkpoint increases with problem size.

Tim

e

2009(estimate)

2006

Crossoverpoint

MTTI decreasesas numberof parts increases.

Page 5: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

5 Geist_FT_SC07

Need to develop new paradigms for applications to handle faults

1. Restart from checkpoint file [large apps today].

2. Restart from diskless checkpoint[Avoids stressing the I/O systemand causing more faults].

3. Recalculate lost data from in-memory RAID.

4. Lossy recalculation of lost data[for iterative methods].

Need todevelop rich methodology

to “run through”

faults.

Need todevelop rich methodology

to “run through”

faults.

No

ch

eckp

oin

tS

om

e st

ate

save

d

5. Recalculate lost data from initialand remaining data.

6. Replicate computation across system.

7. Reassign lost work to another resource.

8. Use natural fault-tolerant algorithms.

Store checkpointin memory.

Page 6: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

6 Geist_FT_SC07

Harness P2P control researchThe file system

can’t let data be corrupted by faults.

I/O nodes must recover andcover failures.

The heterogeneous OS must be able

to tolerate failuresof any of its node type

and instances.

For example:A failed service node shouldn’t take out a bunchof compute nodes.

The schedulersand other system

components must be aware of dynamically

changing system configurations.

So that tasks get assigned around failed components.

Fast recoveryfrom fault

Parallel recovery

from multiplenode failures

Support simultaneous

updates

24/7 system can’t ignore faults

Page 7: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

1. Restart–from checkpointor from beginning.

2. Notify applicationand let it handle the problem.

3. Migrate task to other hardwarebefore failure.

4. Reassign work to spare processor(s).

5. Replicate tasks across machine.

6. Ignore the fault altogether.

6 options for system to handle failures

Need a mechanism for each application (or component)to specify to the system what to do if a fault occurs.

7 Geist_FT_SC07

Page 8: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

8 Geist_FT_SC07

5 recovery modes for MPI applications

1.ABORT: Just do as vendor implementations.

2.BLANK: Leave holes (but make sure collectivesdo the right thing afterward).

3.SHRINK: Reorder processes to makea contiguous communicator (some ranks change).

4.REBUILD: Respawn lost processesand add them to MPI_COMM_WORLD.

5.REBUILD_ALL: Same as REBUILD except that it rebuilds all communicators, and groups and resets all key values, etc.

Harness project’s FT-MPI explored 5 modes of recovery

It may be time to consideran MPI-3 standard that allows applications to recover from faults.

These modes affect the size (extent) and

ordering of the communicators.

Page 9: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

Validation of an answer on such large systems is a growing problem. Simulations are more complex, solutions are being sought in regions never before explored.

• Can’t afford to run every jobthree (or more) times.

• Yearly allocationsare like $5M–$10M grants.

4 ways to fail anyway

9 Geist_FT_SC07

1. Fault may not be detected.

2. Recovery introduces perturbations.

3. Result may depend on which nodes fail.

4. Result looks reasonable, but it is actually wrong.

Page 10: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

1. Detection that something has gone wrong•System: detection in hardware•Framework: detection by runtime environment•Library: detection in math or communication library

2. Notification of the application,runtime, or system components• Interrupt: signal sent to job or system component•Error code returned by application routine

3. Recovery of the application to the fault•By the system•By the application•Neither: natural fault tolerance

Subscriptionnotification

3 steps to fault tolerance

10 Geist_FT_SC07

Page 11: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

11 Geist_FT_SC07

The drive for large-scale simulationsin biology, nanotechnology, medicine,chemistry, materials, etc.

From a fault tolerance perspective:

• Space means that the job ‘state’ to be recovered is huge.

• Time means that many faults will occur during a single run.

2 reasons the problemis only going to get worse

1. Require much larger problems (space):• easily consume the 2 GB per core

in ORNL LCF systems.

2. Require much longer to run (time) • science teams in climate, combustion, and fusion

want to run for a dedicated couple of months.

Page 12: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

12 Geist_FT_SC07

1 holistic solution

Fault Tolerance Backplane

We need coordinated fault awareness, prediction, and recoveryacross the entire HPC system from the application to the hardware.

Middleware

Applications

Operating System

Hardware

CIFTS project

“Prediction and prevention are critical becausethe best fault is the one that never happens.”

Detection

Monitor

Logger

Configuration

Notification

Event manager

Predictionand

prevention

Recovery

Autonomicactions

Recoveryservices

Project under way at ANL, ORNL, LBL, UTK, IU, OSU

Page 13: Presented by Fault Tolerance Challenges and Solutions Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported

13 Geist_FT_SC07

Contact

Al GeistComputer Science Research GroupComputer Science and Mathematics Division (865) [email protected]

13 Geist_FT_SC07