castor project status castor project status cernit-pdp/dm february 2000

18
CASTOR CASTOR Project Project Status Status CERN CERN IT-PDP/DM IT-PDP/DM February 2000 February 2000

Upload: catherine-craig

Post on 18-Dec-2015

214 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

CASTOR Project CASTOR Project StatusStatus

CERNCERN

IT-PDP/DMIT-PDP/DM

February 2000February 2000

Page 2: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

AgendaAgenda CASTOR objectivesCASTOR objectives CASTOR componentsCASTOR components Current statusCurrent status Early testsEarly tests Possible enhancementsPossible enhancements ConclusionConclusion

Page 3: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

CASTORCASTOR

CASTOR stands for “CERN Advanced Storage Manager”CASTOR stands for “CERN Advanced Storage Manager” Evolution of SHIFTEvolution of SHIFT Short term goal: handle NA48 data (25 MB/s) and Short term goal: handle NA48 data (25 MB/s) and

COMPASS data (35 MB/s) in a fully distributed COMPASS data (35 MB/s) in a fully distributed environmentenvironment

Long term goal: prototype for the software to be used to Long term goal: prototype for the software to be used to handle LHC datahandle LHC data

Development started in January 1999Development started in January 1999 CASTOR being put in production at CERNCASTOR being put in production at CERN See: http://wwwinfo.cern.ch/pdp/castorSee: http://wwwinfo.cern.ch/pdp/castor

Page 4: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

CASTOR objectives CASTOR objectives

CASTOR is a disk pool manager coupled with a CASTOR is a disk pool manager coupled with a backend store which provides:backend store which provides:

Indirect access to tapes HSM functionality

Major objectives are:Major objectives are: High performance Good scalability Easy to clone and deploy High modularity to be able to easily replace components and

integrate commercial products

Focussed on HEP requirementsFocussed on HEP requirements Available on most Unix systems and Windows/NTAvailable on most Unix systems and Windows/NT

Page 5: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

CASTOR componentsCASTOR components Client applications use the stager and RFIOClient applications use the stager and RFIO The backend store consists of:The backend store consists of:

RFIOD (Disk Mover) Name server Volume Manager Volume and Drive Queue Manager RTCOPY daemon + RTCPD (Tape Mover) Tpdaemon (PVR)

Main characteristics of the serversMain characteristics of the servers Distributed Critical servers are replicated Use CASTOR Database (Cdb) or commercial databases like Raima

and Oracle

Page 6: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

CASTOR layoutCASTOR layout

STAGER

RFIOD(DISK

MOVER)

TPDAEMON(PVR)

MSGD

DISK POOL

TMS

NAMEserver

VOLUMEmanager

RTCOPY

VDQMserver

RTCPD(TAPE

MOVER)

Page 7: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

Basic Hierarchical Storage Basic Hierarchical Storage Manager (HSM)Manager (HSM)

Automatic tape volume allocationAutomatic tape volume allocation Explicit migration/recall by userExplicit migration/recall by user Automatic migration by disk pool managerAutomatic migration by disk pool manager

Page 8: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

Current statusCurrent status Development completeDevelopment complete New stager with Cdb in production for DELPHINew stager with Cdb in production for DELPHI Mover and HSM being extensively testedMover and HSM being extensively tested

Page 9: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

Early testsEarly tests RTCOPYRTCOPY Name ServerName Server ALICE Data ChallengeALICE Data Challenge

Page 10: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

Hardware configuration for Hardware configuration for RTCOPY tests (1) RTCOPY tests (1)

SUN E450

SCSI disks (striped FS), ~30MB/s

Linux PCs STK Redwood

IBM 3590E

STK 9840

Page 11: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

RTCOPY test results (1)RTCOPY test results (1)

Tapedevice

ReadTape1 stream

ReadTape2 streams

ReadTape3 streams

WriteTape1 stream

WriteTape2 streams

WriteTape3 streams

STKRedwood

13.1 MB/s 21.9 MB/s 24.3 MB/s 11.7 MB/s 23.4 MB/s 28.0 MB/s

STK 9840 11.6 MB/s 22.8 MB/s 24.5 MB/s 11.4 MB/s 22.5 MB/s 32.7 MB/s

IBM3590E

13.5 MB/s 21.2 MB/s 17.4 MB/s 31.9 MB/s

Page 12: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

Hardware configuration for Hardware configuration for RTCOPY tests (2)RTCOPY tests (2)

SUN E450

Linux PC

SCSI disks (striped FS), ~30MB/s

EIDE disks, ~14MB/s

Linux PCs STK Redwood

STK Redwood

STK 9840

Gigabit

Linux PCs

EIDE EIDE

100BaseT

Page 13: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

RTCOPY test results (2)RTCOPY test results (2) A short (1/2 hour) scalability test was run in a A short (1/2 hour) scalability test was run in a

distributed environment:distributed environment: 5 disk servers 3 tape servers 9 drives

120 GB transferred120 GB transferred 70 MB/s aggregate (if mount time overhead included)70 MB/s aggregate (if mount time overhead included) 90 MB/s aggregate (if mount time overhead excluded)90 MB/s aggregate (if mount time overhead excluded) This exceeds COMPASS requirements and is just This exceeds COMPASS requirements and is just

below the ATLAS/CMS requirementsbelow the ATLAS/CMS requirements

Page 14: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

Name server test results (1)Name server test results (1)

Time to create 5000 entries

0

50

100

150

200

250

300

0 50000 100000 150000 200000 250000 300000

Total number of entries

Tim

e (

se

co

nd

s)

Series1

Series2

Series3Cdb

Raima

Oracle

Page 15: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

Name server test results (2)Name server test results (2)

Time to list a directory with 5000 entries

0

1

2

3

4

5

6

7

0 50000 100000 150000 200000 250000 300000

Total number of entries

Tim

e (

se

co

nd

s)

Series1

Series2

Series3

Cdb

Raima

Oracle

Page 16: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

ALICE Data ChallengeALICE Data Challenge

10 * PowerPC 604 200 MHz 32MB 7 * PowerPC 604 200 MHz 32MBHP Kayak

3COM Fast Ethernet Switch 3COM Fast Ethernet Switch

12 * Redwoods

4 * Linux tape servers

12 * Linux disk servers

Gigabit Switch Gigabit Switch

Smart Switch Router

Page 17: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

Possible enhancementsPossible enhancements

RFIO client - name server interfaceRFIO client - name server interface 64 bits support in RFIO (collaboration with IN2P3)64 bits support in RFIO (collaboration with IN2P3) GUI and WEB interface to monitor and administer GUI and WEB interface to monitor and administer

CASTORCASTOR Enhanced HSM functionality:Enhanced HSM functionality:

Transparent migration Intelligent disk space allocation Classes of service Automatic migration between media types Quotas Undelete and Repack functions Import/Export

Page 18: CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000

February 2000CASTOR project status/CHEP2000

ConclusionConclusion 2 man years of design and development2 man years of design and development Easy deployment because of modularity and backward Easy deployment because of modularity and backward

compatibility with SHIFTcompatibility with SHIFT Performance limited only by hardware configurationPerformance limited only by hardware configuration See: http://wwwinfo.cern.ch/pdp/castorSee: http://wwwinfo.cern.ch/pdp/castor