activités cms/grille c.charlot/llr-École polytechnique

22
Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Upload: gavin-larson

Post on 27-Mar-2015

220 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Activités CMS/Grille

C.Charlot/LLR-École Polytechnique

Page 2: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

CMS production use caseNumber of Regional Centers 11

Number of Computing Centers 21

Number of CPU’s ~1000

Number of Production Passes for each Dataset(including analysis group processing done by production)

6-8

Number of Files ~11,000

Data Size (Not including fz files from Simulation)

17TB

File Transfer over the WAN

7TB toward T1

4TB toward T2

Bristol/RAL

Caltech

CERN

FNAL

IC

IN2P3

INFN

Moscow

UCSD

UFL

WISC

Page 3: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Spring2002: cmsim

Wisconsin 2%

IC 4%

IN2P3 7%

CERN 10%

FNAL 12%

Caltech 5%

Bristol 15%

USMOP0.2%

UFL 17%Moscow

2%

UCSD 7%

INFN 19%

6M Events 4.3TB

Page 4: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Spring02: production summary

CMSIM: 1.2 seconds / event for 4 months

High luminosity Digitization:1.4 seconds / event for 2 months

Nbr

of

even

ts 6M

3.5M

April 19 June 7

February 8 May 31

requested

produced

requested

produced1034

Page 5: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Toward ONE Grid

Build a unique CMS-GRID framework (EU+US) EU and US grids not interoperable today. Need for help from the

various Grid projects and middleware experts Work in parallel in EU and US

Main US activities: PPDG, GriPhyN, iVDGL grid projects MOP Virtual Data Toolkit Interactive Analysis: Clarens

Main EU activities: EDG, EDT grid projects Integration of IMPALA with EDG middleware Batch Analysis: user job submission & analysis farm

Page 6: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

EU CMS testbeds participation

DataGrid Testbed Participants: CERN, INFN, CNRS, PPARC, NIKHEF, ESA Curently testbed1, planned testbed2,3 over the project period

DataTAG Participants: CERN, INFN, PPARC, UvA

• France? CMS sites involved

CNAF, CCIN2P3, Bologna, Legnaro/Padova, Pisa, RAL, IC, Moscow, Ecole Polytechnique

Page 7: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

PPDG MOP system

PPDG developed Monte Carlo Distributed Production system Submission of CMS production jobs from a central location on a

remote location, return results Relies on GDMP for replication Globus GRAM CondorG and local queuing systems for job scheduling Impala for job specification DAGMAN for management of dependencies between jobs

(cmsim after cmkin, etc..) Being deployed in US-CMS grid testbed

Page 8: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

GriPhyN/PPDG VDT

US-CMS and GriPhyN/PPDG developing Virtual Data Toolkit Main idea of Virtual Data Materialization

Cache to the user the real processing needed to satisfy request More important to keep information needed to recreate a DataSet

than DataSet itself Introduce new catalogs: Virtual Data Catalog, Materialized

Data Catalog Based on MOP for job submission Plan to use WP1 asap

Page 9: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

CMS EDG Prototype

EU working on integration of CMS production tools (Impala/BOSS) with EDG software

Sites involved: École Polytechnique, Bologna, IC, Padova, Moscow synchronisation avec EDG1.2 devrait permettre d’inclure Lyon

Modify production tools Impala/BOSS to allow remote submission from any site that has UI installed Preinstalled CMS software at the CEs Interface with EDG submission (WP1) Interface with EDG data/file management (WP2)

Page 10: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Production processing“Produce 100000 events dataset mu_MB2mu_pt4” IMPALA

decomposition(Job scripts)

JOBSRC

BOSSDB

IMPALA monitoring(Job scripts)

Production“RefDB”

ProductionInterface

Production manager

coordinatestasks distribution toRegional Centers

Farm storage

RequestSummary

file

RC farm

Regional Center

Data locationthrough

Production DB

Page 11: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Job submission:BOSS submission to EDG scheduler

BOSS on User Interface machine (UI) submits to EDG scheduler

BOSS v3+ BOSS will accept and pass on a JDL file. jobGridExecuter, executable, std in, std out, std err

automatically placed in JDL sandboxes. BOSS job monitoring over the grid.

Monitoring DB open for write access to every machines!

Page 12: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Job submission:IMPALA on User Interface machine

IMPALA uses BOSS on UI machine to submit job to EDG scheduler

Modifications to IMPALA IMPALA creates EDG JDL files Post-process stage: GetOutputJob_edg.sh script Minor changes:

Scripts on UI machine do not set up CMS environment, as only the WNs need the CMS software.

Page 13: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Job submission:IMPALA on User Interface machine

JDL file: CMSIMExecutable = “…/scripts/jm_minbias_000002.sh";InputSandbox ={“…/stationary_env",“…/localFunctions.sh",

“…/jm_minbias_000002.ttl",“…/StageOutFZ.sh",“…/RunCMSIM.sh", “…/LocalEnvironmentSetup.sh",“…/LocalDefaults.rsc",“…/DBAccess.sh", “…/DBAccess.py","in","/tmp/x509up_u33117"};

OutputSandbox ={"impala.jm_minbias_000002.runout", "impala.jm_minbias_000002.runerr","File_location"};

InputData = {"LF:jm_minbias_000002.ntpl"};ReplicaCatalog = "ldap://gatekeeper.phy.bris.ac.uk:2010/rc=Test-Rep-

cat,mds-vo-name=gridpp,o=grid";DataAccessProtocol = "file";

Page 14: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Job submission:IMPALA on User Interface machine

GetOutputJob_edg.sh performs a dg-job-status for all submitted jobs. output from finished jobs is retrieved and moved into

batch/logs. IMPALA tracking is updated

i.e. cleared jobs > “finished” status

aborted jobs > “problem” status

Page 15: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Job Submission:Current Status

Job submission chain: IMPALA > BOSS > EDG-job scheduler is working at several sites.

Jobs only sent to suitable CEs by using RunTimeEnvironment in JDL.

Jobs are sent to the CE which is close to the SE that holds a replica of required input file.

Page 16: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

File management:FlatfilesEDG tools used

Resource Broker creates ‘.BrokerInfo’ file, contains Logical filenames (LFs) Physical filenames (PFNs) Close SEs to the CE

edg-brokerinfo parses ‘.BrokerInfo’ for useful information, e.g. converts Logical Files to Transport filenames TFNs.

GDMP copies files to a SE and registers them in a Replica Catalog. Version 3.0.x has been installed at IN2P3, INFN and IC.

Page 17: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

File management:flatfiles IMPALA modifications

GetReplica() Takes an input list of LFs, queries the .BrokerInfo file to get

the corresponding TFNs (transport filenames) and copies the files to the current directory.

The edg-brokerinfo tool relies on patched RB which produces correct .BrokerInfo file (bug fixed in EDG release 1.2)

Page 18: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

File management: flatfilesIMPALA modifications

StagetoSE() Takes a file and copies it using globus-url-copy to the close

SE named in the .BrokerInfo file. Uses globus-job-run commands to perform remote operations

on the storage element:Find the appropriate GDMP storage directory (by looking at the GDMP.conf file)Check & create output directory if necessary

Information is published to the replica catalog using GDMP and also written to ‘File_location’ which is returned to the user in the output sandbox.

Page 19: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

File management: flatfilesCurrent status

Ntuple file and fz files from CMKIN and CMSIM stages successfully staged out to SEs with information published in the Replica Catalog.

Logical filename of ntuples put in JDL of CMSIM jobs, which is used within the job to retrieve a replica.

Page 20: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

EDG prototype Summary

CMKIN CMSIM stages of production have been run from several sites using EDG software for both job submission and replica management. Consolidate work into single ‘EDG enabled’ version of IMPALA EDG release 1.2 Increase scale of tests

Ensure grid resources have standard installations Interface with Mass Storage System (MSS)

Demonstrate with small production (o(1000) jobs) C’est capital pour CMS …

Page 21: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Conclusions/Remarques CMS n’est pas uniquement européenne

Nombreux projets grilles aux US, nécessité d’inter-opérer• Actuellement l’objet de DataTag/Glue• DataGrid devrait il me semble aller fortement dans cette direction

• Certificats, VO, Dagman?• Autrement, gros risque de divergence

Installation soft EDG actuelle inextricable DataTag travaille sur une distribution simplifiée utilisant

Pacman• CMS prévoit aussi l’utilisation de Pacman pour la distribution des

outils de production• Là aussi, nécessité de converger

On fonde beaucoup d’espoir sur les listes prédéfinies de Yannick, User help desk, etc..

Page 22: Activités CMS/Grille C.Charlot/LLR-École Polytechnique

Conclusions/remarques (II) Testbed(s)

Il y a de nombreux testbeds DataGrid, DataTAG, CMS-US, .. L’évaluation des outils et intégration dans les logiciels actuels

nécessite des séquences d’installations et tests fréquentes et aussi souples que possible

Situation actuelle• DataGrid comme testbed de production

• Validation des releases• stress tests

• Testbed spécifique CMS pour le développement• évaluation des logiciels, accent sur les fonctionnalités• évolution rapide de prototypes EDG+CMS sw