70 likes | 203 Vues
This document outlines the comprehensive management of the Grid Computing Infrastructure and associated services tailored for scientific data-intensive applications. It discusses the implementation of middleware, application development, and specialized server configurations for projects like ALICE and ATLAS. Key topics include interactive logins, certification authority processes, and data pool management, emphasizing the efficiency and evolution of the Regional Computing and Data Center in Germany. The infrastructure enables high-throughput data operations, ensuring sustainable performance and ongoing upgrades to support advanced research initiatives.
E N D
management Grid-Computing Infrastructure & Services (GIS, H.Marten) 1.2 2.6 FZK tape archive F Z K b a c k b o n e software server interactive login ALICE 1.5.1 1.6 software installation service certification authority compute nodes CPU intensive Data tape archive PBytes G R I D b a c k b o n e 1.3.1 1.8 1.12 1.1 software server interactive login ATLAS licence service firewall 1.4.1 1.7.1 Data server 1 compute nodes I/O intensive . . . . 8x 1.3.2 infrastructure S A N f a b r i c 1.9 1.13 2.1 ALICE Grid services Data server 2 ATLAS .... 1.7.2 Data Pool Volume Management user support . . . 1.4.2 1.14 compute nodes communication intensive ?? 2.2 Data server n training and education software server interactive login other science system monitoring & management 1.4.3 2.3 1.3.9 1.10 compute nodes TESTING; R&D documentation network address translation 1.7 2.4 Data server R&D software server interactive login TEST; R&D 1.11 1.4.4 R & D batch and background; invisible for users private network ?? 1.3.10 2.5 Data Import Data Export Tier-0/1/2 RDCCG – Regional Computing and Data Center Germany WAN 23-nov-2001
RDCCG Evolution (available capacity) 30% rolling upgrade each year after 2007 Networking Evolution 2002 - 2005 1) RDCCG to CERN/FermiLab/SLAC (Permanent point to point): 1 GBit/s – 10 GBit/s => 2 GBit/s could be arranged on a very short timescale 2) RDCCG to general Internet: 34 MBit/s – 100 MBit/s => Current situation, generally less affordable than 1) FTE Evolution 2002 - 2005 Support: 5 - 30 Development: 8 - 10 New office building to accommodate 130 FTE in 2005
Installation and Update of OSThe idea is to set up a dedicated installation server for each rack of CPUsCurrently evaluating the Rocks-Toolkit MonitoringNo use of dedicated monitoring hardware and/or management busTivoli management suite in broad use for already existing operations-> Further study and deployment of tools necessary Management Tools
Certification Authority • Hardware for CA available • Delivery of certificates tested • CP and CPS draft document available • “FZK-Grid CA, Certificate Policy and Certification Practice Statement” • Organization of German virtual organizations • FZK-Grid CA known to Globus and accepted by DataGrid • Policy for acceptance of foreign certificates needs to be discussed Authorization and Registration Procedure • One account manager for each experiment • WWW form to be filled out, signed by experiment account manager and send to FZK by FAX • One experiment specific account per user • One system administrator account per experiment • No super user privileges, sorry
GSI DarmstadtTier-2 will perform e.g. interactive analysis on subsamples (PROOF) Offer support for countries w/o dedicated Tier-1e.g. Russia (MSU), … Tier-2 business needs to be further discussed and developed within Germany Tier-2 Centers LHC Data Challenges • Commitment to fully take part in all data challenges • MoU to integrate non-LHC hardware upon actual needcurrently exercised with ALICE: use BaBar farm for MC challenge-> Synergetic effect until 2004
Department for Grid-Computing and e-Science (GES, M.Kunze): • Middleware and Application Development • Special interest in ROOT core development and support • Meta-Data catalogue • Definition and implementation of interfaces to Globus and .NET • Evolution of object service • Transaction based processing • Interactive analysis (PROOF) • Special interest in database technology • Data mining techniques (Agents and Neural Networks) • Deployment of intelligent server side procedures • Deployment of XML and SOAP interface technology • Involvement into • BaBar MC production and distributed analysis (GanaTools) • LHC test beds (currently ALICE, ATLAS coming up) • CrossGrid • National Grid projects (e.g. DAS-Grid w/ ZIB, UniDo, PC2, FZ Jülich)
SAN SANergy, qfs, gfs, ... Running SAN for Grid Compute Nodes:I/O clients for nfs, cifs, root, scp, gass, ftp, s-ftp, http, https, ... Grid backbone 128 GBit/s I/O server for nfs, cifs, root, scp, gass, ftp, s-ftp, http, https, ... MDC High Throughput Cluster: Direct SAN access (>300 MB/s) Good for interactive Data-Mining: 1 TB in a few minutes tape robotics disk subsystems