castor project status castor project status cernit-pdp/dm february 2000
TRANSCRIPT
CASTOR Project CASTOR Project StatusStatus
CERNCERN
IT-PDP/DMIT-PDP/DM
February 2000February 2000
February 2000CASTOR project status/CHEP2000
AgendaAgenda CASTOR objectivesCASTOR objectives CASTOR componentsCASTOR components Current statusCurrent status Early testsEarly tests Possible enhancementsPossible enhancements ConclusionConclusion
February 2000CASTOR project status/CHEP2000
CASTORCASTOR
CASTOR stands for “CERN Advanced Storage Manager”CASTOR stands for “CERN Advanced Storage Manager” Evolution of SHIFTEvolution of SHIFT Short term goal: handle NA48 data (25 MB/s) and Short term goal: handle NA48 data (25 MB/s) and
COMPASS data (35 MB/s) in a fully distributed COMPASS data (35 MB/s) in a fully distributed environmentenvironment
Long term goal: prototype for the software to be used to Long term goal: prototype for the software to be used to handle LHC datahandle LHC data
Development started in January 1999Development started in January 1999 CASTOR being put in production at CERNCASTOR being put in production at CERN See: http://wwwinfo.cern.ch/pdp/castorSee: http://wwwinfo.cern.ch/pdp/castor
February 2000CASTOR project status/CHEP2000
CASTOR objectives CASTOR objectives
CASTOR is a disk pool manager coupled with a CASTOR is a disk pool manager coupled with a backend store which provides:backend store which provides:
Indirect access to tapes HSM functionality
Major objectives are:Major objectives are: High performance Good scalability Easy to clone and deploy High modularity to be able to easily replace components and
integrate commercial products
Focussed on HEP requirementsFocussed on HEP requirements Available on most Unix systems and Windows/NTAvailable on most Unix systems and Windows/NT
February 2000CASTOR project status/CHEP2000
CASTOR componentsCASTOR components Client applications use the stager and RFIOClient applications use the stager and RFIO The backend store consists of:The backend store consists of:
RFIOD (Disk Mover) Name server Volume Manager Volume and Drive Queue Manager RTCOPY daemon + RTCPD (Tape Mover) Tpdaemon (PVR)
Main characteristics of the serversMain characteristics of the servers Distributed Critical servers are replicated Use CASTOR Database (Cdb) or commercial databases like Raima
and Oracle
February 2000CASTOR project status/CHEP2000
CASTOR layoutCASTOR layout
STAGER
RFIOD(DISK
MOVER)
TPDAEMON(PVR)
MSGD
DISK POOL
TMS
NAMEserver
VOLUMEmanager
RTCOPY
VDQMserver
RTCPD(TAPE
MOVER)
February 2000CASTOR project status/CHEP2000
Basic Hierarchical Storage Basic Hierarchical Storage Manager (HSM)Manager (HSM)
Automatic tape volume allocationAutomatic tape volume allocation Explicit migration/recall by userExplicit migration/recall by user Automatic migration by disk pool managerAutomatic migration by disk pool manager
February 2000CASTOR project status/CHEP2000
Current statusCurrent status Development completeDevelopment complete New stager with Cdb in production for DELPHINew stager with Cdb in production for DELPHI Mover and HSM being extensively testedMover and HSM being extensively tested
February 2000CASTOR project status/CHEP2000
Early testsEarly tests RTCOPYRTCOPY Name ServerName Server ALICE Data ChallengeALICE Data Challenge
February 2000CASTOR project status/CHEP2000
Hardware configuration for Hardware configuration for RTCOPY tests (1) RTCOPY tests (1)
SUN E450
SCSI disks (striped FS), ~30MB/s
Linux PCs STK Redwood
IBM 3590E
STK 9840
February 2000CASTOR project status/CHEP2000
RTCOPY test results (1)RTCOPY test results (1)
Tapedevice
ReadTape1 stream
ReadTape2 streams
ReadTape3 streams
WriteTape1 stream
WriteTape2 streams
WriteTape3 streams
STKRedwood
13.1 MB/s 21.9 MB/s 24.3 MB/s 11.7 MB/s 23.4 MB/s 28.0 MB/s
STK 9840 11.6 MB/s 22.8 MB/s 24.5 MB/s 11.4 MB/s 22.5 MB/s 32.7 MB/s
IBM3590E
13.5 MB/s 21.2 MB/s 17.4 MB/s 31.9 MB/s
February 2000CASTOR project status/CHEP2000
Hardware configuration for Hardware configuration for RTCOPY tests (2)RTCOPY tests (2)
SUN E450
Linux PC
SCSI disks (striped FS), ~30MB/s
EIDE disks, ~14MB/s
Linux PCs STK Redwood
STK Redwood
STK 9840
Gigabit
Linux PCs
EIDE EIDE
100BaseT
February 2000CASTOR project status/CHEP2000
RTCOPY test results (2)RTCOPY test results (2) A short (1/2 hour) scalability test was run in a A short (1/2 hour) scalability test was run in a
distributed environment:distributed environment: 5 disk servers 3 tape servers 9 drives
120 GB transferred120 GB transferred 70 MB/s aggregate (if mount time overhead included)70 MB/s aggregate (if mount time overhead included) 90 MB/s aggregate (if mount time overhead excluded)90 MB/s aggregate (if mount time overhead excluded) This exceeds COMPASS requirements and is just This exceeds COMPASS requirements and is just
below the ATLAS/CMS requirementsbelow the ATLAS/CMS requirements
February 2000CASTOR project status/CHEP2000
Name server test results (1)Name server test results (1)
Time to create 5000 entries
0
50
100
150
200
250
300
0 50000 100000 150000 200000 250000 300000
Total number of entries
Tim
e (
se
co
nd
s)
Series1
Series2
Series3Cdb
Raima
Oracle
February 2000CASTOR project status/CHEP2000
Name server test results (2)Name server test results (2)
Time to list a directory with 5000 entries
0
1
2
3
4
5
6
7
0 50000 100000 150000 200000 250000 300000
Total number of entries
Tim
e (
se
co
nd
s)
Series1
Series2
Series3
Cdb
Raima
Oracle
February 2000CASTOR project status/CHEP2000
ALICE Data ChallengeALICE Data Challenge
10 * PowerPC 604 200 MHz 32MB 7 * PowerPC 604 200 MHz 32MBHP Kayak
3COM Fast Ethernet Switch 3COM Fast Ethernet Switch
12 * Redwoods
4 * Linux tape servers
12 * Linux disk servers
Gigabit Switch Gigabit Switch
Smart Switch Router
February 2000CASTOR project status/CHEP2000
Possible enhancementsPossible enhancements
RFIO client - name server interfaceRFIO client - name server interface 64 bits support in RFIO (collaboration with IN2P3)64 bits support in RFIO (collaboration with IN2P3) GUI and WEB interface to monitor and administer GUI and WEB interface to monitor and administer
CASTORCASTOR Enhanced HSM functionality:Enhanced HSM functionality:
Transparent migration Intelligent disk space allocation Classes of service Automatic migration between media types Quotas Undelete and Repack functions Import/Export
February 2000CASTOR project status/CHEP2000
ConclusionConclusion 2 man years of design and development2 man years of design and development Easy deployment because of modularity and backward Easy deployment because of modularity and backward
compatibility with SHIFTcompatibility with SHIFT Performance limited only by hardware configurationPerformance limited only by hardware configuration See: http://wwwinfo.cern.ch/pdp/castorSee: http://wwwinfo.cern.ch/pdp/castor