International Workshop on Large Scale of Computing February 8-10, 2006 Variable Energy Cyclotron Centre 1/AF, Bidhan Nagar Kolkata – 700 064, India Tier-2 in India for Alice Susanta Kumar Pal, VECC, Kolkata Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 1 What is GRID computing? Computational Grid is a collection of distributed, heterogeneous resources which can be used as an ensemble to execute large-scale applications. The emphasis is on : Distributed supercomputing High throughput & data intensive applications. Large scale storage High speed network connectivity Why ALICE is interested in GRID? 1 year of Pb-Pb running: 1 PBytes of data 1 year of p-p running : 1 PBytes of data Simulations : 2 PBytes Total Data storage: 4 Pbtyes/year ALICE computing requirements: Simulations, Data reconstruction & analysis will use about 10,000 PCyears. GRID is the solution for ALICE Connect high performance computers from all collaborating countries with a high speed secured network. implementing one virtual environment that is easy for the “end user’’. Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 2 The MONARC Multi-Tier Model (1999) CERN – Tier 0 IN2P3 Tier 1 FNAL RAL Uni n Lab a Tier2 Uni b Department Lab c Desktop MONARC report: http://home.cern.ch/~barone/monarc/RCArchitecture.html Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 3 The opportunity of Grid technology LHC Computing Model Lab m 2001 - evolving CMS Uni x regional group ATLAS Uni a CERN Tier 1 Lab a UK USA Tier3 physics department The opportunity of Tier 1 Uni n Grid technology LHCb The LHC Computing Centre Tier2 France CERN CERN Tier 0 Italy ………. Desktop Lab b Germany Lab c physics group Susanta K Pal ………. Uni y Uni b Tier 0 Centre at CERN IWLSC, Feb 8-10, 2006, Kolkata, India 4 What is a Tier-2? The following is taken from the LCG RTAG 6 on Regional Centre Category and Service Definition. A possible categorization scheme for regional centres could be based on service qualities as follows: Tier1 * CPU cycles (Grid enabled computing elements), advance reservation * Disk storage, resident and temporary (Grid enabled storage elements), advance reservation * Mass storage (Grid enabled storage elements), advance reservation * State-of-the-art network bandwidth, quality of service * Commitment to provide access to primary/master copy of data over lifetime of LHC * Commitment to provide long-term access to specific analysis data * Commitment to resource upgrades as required * 24/7 services and resource support * National support role * Training and user support * Interactive support for particular applications Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 5 Tier2 * CPU cycles (Grid enabled computing elements) * Disk storage, maybe temporary only (Grid enabled storage elements) * May have mass storage * Sufficient network bandwidth for inter-operability * A weaker commitment to provide access to data over LHC lifetime * A weaker commitment to provide long-term access to specific analysis data * A weaker commitment to resource upgrades * Focused user support * 24/7 service but with no guaranteed short-time "crash" response Tier3 * CPU cycles (Grid enabled computing elements) * Local storage (not necessarily Grid enabled) * Focused commitment to data access or resource upgrade * Only local user support * Focused services for agreed and day-by-day analysis activities * Local interactive support Tier4 * Enable Grid access * Provide experiment specific tools Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 6 Calib & Reco 1 Raw Calib ESD AOD Tag RAW data delivered by DAQ undergo Calibration and Reconstruction which produce for each event 3 kinds of objects: 1. 2. 3. ESD object AOD object Tag object Done in Tier-0 site. Further reconstruction and calibration of RAW data will be done at Tier 1 and Tier 2. ESD Reco n AOD The generation, reconstruction, storage and distribution of Monte-Carlo simulated data will be the main task of Tier 1 and Tier 2. Tag DPD (Derived Physics Data) objects will be Processed in Tier 3 and Tier 4. Ana DPD Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 7 Why Tier 2 ? 1. Tier-2 is the lowest level to be accessible by the whole collaboration 2. Each sub-detector of ALICE has to be associated with minimum Tier-2 because of large volume of calibration data 3. PMD and Muon Arm are the important subdetectors of ALICE 4. We are solely responsible for PMD – from conception to commissioning Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 8 ALICE Setup LHC Utilization -- ALICE HMPID Size: 16 x 26 meters Weight: 10,000 tons TOF TRD Cost 120 MCHF TPC PMD ITS Muon Arm PHOS Indian contribution to ALICE : PMD, Muon Arm IWLSC, Feb 8-10, 2006, Kolkata, India 9 Susanta K Pal ALICE Layout : another view Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 10 ALICE PMD (=2.3 – 3.5) Honeycomb counters for ALICE PMD are Modified edition of STAR PMD Two Halves In Vertical When not in use Two planes of honeycomb proportional counters 3 X0 thick lead converter Arranged in two halves in vertical plane Installed at z=360 cm from I.P. PMD in ALICE : Fully Indian Contribution Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 11 ALICE PMD Layout (Data taking position) Converter + Support Plate Super Module 8 Supermodules in 2 planes Total channels Preshower+Veto = 221,184 Unit Module 48 Nos. Total Cooling by air circulation Unit module components Honeycomb (4608 cells) Top, bottom PCBs Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 12 Components of a unit module TOP PCB (details) 4 X 16 cells for one MCM Board 32-pin connector Copper honeycomb Edge frame Bottom PCB Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 13 PMD unit module assembled Unit cell parameters 421 mm Cell cross-section 23 mm2 Cell depth 5 mm Centre-to-centre dist. 5 mm Cell wall thickness 0.4 mm Anode wire (Au-plated tungsten) 20 µm dia Anode-cathode distance (on PCB) 750 µm Unit Module, 4608 cells 421 mm X 260 mm Fabrication of unit modules at Kolkata, Jammu and Bhubaneswar Final assembly at Kolkata Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 14 PMD Parameters Pseudorapidity coverage 2.3 – 3.5 Azimuthal coverage 2p Distance from vertex 350 cm Detector active area 2 m2 Detector weight 1000 kg No. of planes 2 (veto + preshower) Lead plate + SS plate thickness 3 radiation length Detector Gas detector with hexagonal cells Hexagonal cell dimensions Depth: 0.5 mm, cross section: 0.22 cm2 Total number of cells 221184 (=>in each plane: 110592 cells) Detector gas Ar+CO2 (70%+30%) Total gas volume 0.02 m3 (20 Liters) No. of Supermodules per plane 4 No. of unit modules per plane 6 (HV isolation at the unit module level) No. of HV channels 48 Average Occupancy (at full multiplicity) 13% for CPV & 28% for PMD Photon counting efficiency 64% Purity of photon sample 60% Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 15 muon tracking quadrants assembled in Kolkata Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 16 Cell-to-cell gain calibration for PMD Need : To have uniform response throughout the detector In data look for single How to do : isolated cell ADC spectrum 0 0 Test beam studies tell us – hadron response from our detector 0 ADC>0 0 0 0 STAR PMD Isolated cell ADC spectra Single cell is affected Susanta K Pal Pulse height spectra landau in a cell : Mean can be used for calibration IWLSC, Feb 8-10, 2006, Kolkata, India 17 Cell-to-cell gain calibration for PMD ALICE PMD has 221184 cells to be calibrated For this we need isolated cell ADC spectra for each cell Minimum number of entries need for a good Landau distribution ~ 1000 We need at least 1 Million events for calibration So data volume ~200Kx0.25x4BytesX1Million ~ 200GB Although calibration is done once for a period of Running, it may be advisable to check the calibration constants time-to-time. Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 18 MANAS Calibration for Tracking Detectors of MuonSpectrometer With and Without MANAS Calibration – preproduction batch : gain spread 5% Expected Gain Spread in MANAS production : ~ 2.3 % Channel gain calibration may not be essential any more Production Batch Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 19 Physics data Expected Data rate for MuonArm Total number of readout pads in FMS: 1.1 X 106 Trigger rate – 1 kHz Average occupency ~ 2.5% 1 month of Pb-Pb data == 300 TB Pedestal data: Every run ~ 5 MB (σ values of each channel needs to be known for analysis. One pedestal run/hour 40 GB/month Electronic Calibration data The frequency will depend on the observed gain spread (if it is <2.5% then uncalibrated resolution will be satisfactory) Every run ~10 MB (for 2 point calibration) GMS data Comparable or less than pedestal data. Susanta K Pal Estimation from SINP group IWLSC, Feb 8-10, 2006, Kolkata, India 20 Required Computing Resources for Tier 2 By 2006: # CPU(Intel P3) Disk Space (TB) Tier 0: 28800 727 Tier 1 + 2 7200 238 Total 6 Tier 1 centers and for each Tier 1 there will be several (~ 5-6) Tier 2 Centers. Tier 2 centres should have the capacity roughly 30% of Tier1 + 2 capacity. At VECC 2004 2005 2006 2007 2008 Total # CPU (Intel Xeon) 8 8 48 64 64 = 192 Disk space (TB) 0.5 0.5 12 24 12 = 49 Bandwidth: Tier 2 centers rely entirely on associated Tier 1 centers for Reconstruction data storage. For efficient and fast analysis a bandwidth of 1.5 Gb/s is a reasonable value – although smaller values are viable. Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 21 Our Experience with Alice-Grid & PDC’04 The following H/W and S/W infrastructures were used : •8 Node Cluster consisting of dual Xeon CPU & 400GB Disk Space. •PBS Batch System with one Management server and eight Clients under OSCAR cluster management environment •ALICE Environment ( AliEn ) was installed •Data Grid has been registered at cern.ch •AliROOT, GEANT and other production related packages are tested successfully in both ways •Linked with CERN via 2Mbps available Internet link and Participated in PDC’04 Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 22 Our Experience with Alice-Grid & PDC’04 AliEn Architecture in General Core Middleware Low level Susanta K Pal User Application Logger Package Mgr (…) V.O. Packages & Commands Monitor/Gatekeeper Config Mgr Storage Element SOAP/XML Core Modules External Libraries Computing Element User Interface Resource Broker Interfaces API (C/C++/perl/java) Authenticator LDAP File & Metadata Catalogue ADBI Database Proxy DBI DBD RDBMS Core Components & Services FS CLI GUI Web Portal High level IWLSC, Feb 8-10, 2006, Kolkata, India 23 Our Experience with Alice-Grid & PDC’04 Grid Services Workload Management – Reliable Job Submission – Matchmaking – Logging & Bookkeeping Data Management – Replica Management – Metadata Management Resource Management – Gatekeeper (batch Interface) – Storage (disk, tape, etc.) – Database (SQL, …) – Network Susanta K Pal Information System – Service Discovery – Service State Information Security – Authentication – Authorization IWLSC, Feb 8-10, 2006, Kolkata, India 24 Our Experience with Alice-Grid & PDC’04 Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 25 Preparing for next PDC Changed the cluster software from OSCAR to QUATTOR OSCAR Performance degrades with more Nos. of Nodes ( > 32 Nodes ) QUATTOR Perfomence does not degrade with increase of nodes Better performance with more Nos. of Nodes Dedicated Band-Width for Tier2@Kolkata -> 4 Mbps Separate Domain Name for Tier2; ‘tier2-kol.res.in’ Addition of More Nos. of CPUs (48-Xeon) and Storage(7TB) Installation of gLite is in progress Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 26 Preparing for next PDC GAS WM DM RB TQ FPS FTQ PM ACE FC JW JA LRC LJC SE CE SRM CR Grid Access Service Workload Mgmt Data Mgmt Resource Broker Task Queue File Placement Service File Transfer Queue Package Manager AliEn CE (pull) File Catalogue Job Wrapper Job Agent Local Replica Catalogue Local Job Catalogue Storage Element Computing Element Storage Resource Mgr Computing Resource (LSF, PBS,…) Susanta K Pal gLite Architecture API GAS WM DM TQ FTQ gLite PM FC Middleware JW CE SE (JA) L Services L ACE CR (LSF,..) JC IWLSC, Feb 8-10, 2006, Kolkata, India RC SRM 27 Present Status VECC- Cluster : High Availability Quattor CERN Management Node ROUTER Switch Gigabit Network 4Mbps Management Node (Stand-by) Internet Cloud FireWall SINP- Cluster : High Availability Quattor Susanta K Pal Tier-2@Kolkata IWLSC, Feb 8-10, 2006, Kolkata, India 28 Present Status Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 29 Present Status Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 30 Present Status grid.veccal.ernet.in graphs last hour sorted descending Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 31 Tier-II Centre for ALICE (Update on VECC and SINP Activities) New Domain name “tier2-kol.res.in’ has been registered and work is going on CONDOR Batch System is running with one server and eight Clients under QUATTOR cluster management environment AliROOT, GEANT and other production related packages are tested successfully in both ways ALICE Environment ( AliEn ) at present NOT running Data Grid has been registered at cern.ch Linked with CERN via 2Mbps available Internet link. 4MBPS band-width is already installed and commissioned Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 32 Tier-II Centre for ALICE (Update on VECC and SINP Activities) Infrastructure Status: •Fully Equipped GRID Room is Ready (Sufficient to accommodate projected Tier-II H/W) Main Room Main Main Room Room Six Auto Switchable AC units 1. 20ft x 19ft 2. Auto Switchable units SixSix Auto Switchable ACAC units NW to VECC main 30ft x 20ft VECC SINP UPS room 10ft x 19ft UPS room 20ft x 15ft UPS room Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 33 VECC Grid Computing Facility Room Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 34 Proposed Indian ALICE-GRID Project Indian ALICE Collaborators •VECC, SINP - Kolkata •Panjab Univ., Chandigarh •IOP, Bhubaneswar •Jammu Univ., Jammu •Rajasthan Univ., Jaipur •AMU, Aligarh CERN Chandigarh Kolkata Alice-Grid Router Modem Router Modem Jammu Jaipur Router Internet Back-Bone Modem Router Modem Gateway Router +Modem Bhubaneswar Aligarh Modem Router Susanta K Pal Router Modem IWLSC, Feb 8-10, 2006, Kolkata, India 35 Contribution towards LCG-GRID PROJECTS LCG-AliEn-SE Interface using GridFTP 1st-July’03 to 18th Dec’03 Status: The project was successfully completed before schedule Milestones of the Project were as follows : Test Bed Installation for Grid Environment The configuration consists of one central server and two Sites. Certification Authority Server has been installed A simple certification authority has been installed to generate certificates for authentication purpose. Installation of GridFTP Library under AliEn The GridFTP daemon in.ftpd has been used as server and globus-url-copy has been used as client Development of AliEn-SE Interface via GridFTP These newly developed modules along with necessary GridFTP libraries and changes made in existing AliEn Code have been committed to CVS Server at CERN. Susanta K Pal Quality Assurance and Test Environment For AliEn-ARDA* Prototype 23rd Dec’03 to 31st March’04 The Project was successfully completed. Milestones of the Project were: Exploration and Design of Test Scripts using perl. Implementation of Test Scripts for each Individual perl sub-module of AliEn. Individual perl sub-modules of AliEn code were tested for proper functionalities. It Generates a detailed report of the individual tests and maintains a log. Validation of Test-Scripts and Procedures. Testing Modules with perl Harnessing Environment. The Complete Suit was tested at CERN under perl Harnessing Environment for testing AliEn online and generating online consolidated report of the test. Inline Documentation to the extent possible. ARDA Architectural Road-Map towards Distributed Analysis IWLSC, Feb 8-10, 2006, Kolkata, India 36 To Summarize •Infrastructure is ready to accommodate Tier-2 H/W •Upgradation of present AliEn cluster with gLite is in progress •Middleware are installed in the Current Hardware with latest Scientific Linux OS platform •With limited resources, VECC took part in ALICE ProductionData-Challenge(PDC-2004) • Getting ready for Next PDC with upgraded infrastructure •Upgradation of CE and SE is in the process as per requirement •One FTE and 2 Engineers with 30% of time are engaged •Two more FTEs are approved Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India Thank you 37 PMD V0 Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 38 PMD Split Position Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 39 Super Module and Unit module arrangement of ALICE PMD Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 40 Super Module Type B UM : - 12 - FEE Boards in a row - 6 Nos. of rows Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 41 Super Module Type A UM: 24-FEE Boards in a Row 3- Rows Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 42 Contribution towards LCG Project •LCG-AliEn Storage Element Interface (addendum-5) 1st-July’03 to 18th Dec’03 Value: 150 KCHF •Test Suite for perl harnessing with AliEn (addendum-6) 23rd Dec’03 to 31st March’04 Value: 150 KCHF •VECC as a part of the Alice Data-Challenge Team run offline production on the existing infrastructure. •Currently main emphasis is on participation in ALICE Physics Data Challenge 2005-06. Future Projects: • Development of Test Environment for ADRA (Architectural Roadmap for Distributed Analysis) code • Testing ARDA code under Perl Test harness guidelines • Part in EGEE ( Enabling Grid for E-Science Euro) prototype development Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 43 Our Experience with Alice-Grid & PDC’04 Security: Overview User side Getting a Certificate Becoming a member of the VO Server side Authentication /CA Authorization / VO Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 44 Our Experience with Alice-Grid & PDC’04 Security: Authentication Overview Once in every year Once in every year cert-signing grid-cert-request grid-cert-request cert-signing CA ca/url update cert-request certificate cert.pkcs12 convert grid-proxy-init USER host-request host-cert CA-certificate Once in life-time of the VO (only DN) Every night/week crl gridmap Registration proxy-cert SERVICE mkgridmap VO-LDAP Once in every 12/24 Hours host/proxy certificate exchanged Every night/week Susanta K Pal IWLSC, Feb 8-10, 2006, Kolkata, India 45
© Copyright 2026 Paperzz