1 / 30

Deployment of Storage Resource Broker at CCLRC for E-science Projects Ananta Manandhar

Deployment of Storage Resource Broker at CCLRC for E-science Projects Ananta Manandhar Data Management Group CCLRC Daresbury Laboratory. Overview. SDSC SRB overview Use of SRB in EScience Projects Eminerals/Ematerials Project CERN CMS SRB support UK National Grid Service – SRB Federation

Télécharger la présentation

Deployment of Storage Resource Broker at CCLRC for E-science Projects Ananta Manandhar

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh • Deployment of Storage Resource Broker at CCLRC for E-science Projects • Ananta Manandhar • Data Management Group • CCLRC Daresbury Laboratory

  2. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh Overview • SDSC SRB overview • Use of SRB in EScience Projects • Eminerals/Ematerials Project • CERN CMS SRB support • UK National Grid Service – SRB Federation • Questions

  3. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh Storage Resource Broker • The SDSC Storage Resource Broker is a client server middleware that virtualizes data space by providing a unified view to multiple heterogeneous storage Resources over the network. • It is a software that sits in between users and resources and provides a storage service by managing users, file locations, storage resources and metadata information SRB Space

  4. Hides the physical location of the file by providing a logical view to the users files which is arranged in a UNIX like hierarchical structure The logical directory structure and the physical file locations are maintained in a Database. Also other persistent information such as User accounts, Participating resources and file meta data are stored in it. Scientist Scientist Scientist eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh SRB Logical Structure SRB View Disk Farm Tape Farm Disk Farm DB RDBMS

  5. Oracle RDBMS Oracle Client SRB Server SRB Server SRB Server SRB Server User Storage Driver Storage Driver Storage Driver Storage Space Storage Space Storage Space eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh SRB Physical Structure MES SRB Server SRB Vault SRB Vault SRB Vault

  6. User eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh SRB Communication User File Request MCAT Database • SRB Client sends request for file to SRB server. • SRB Server contacts MCAT Enabled Server (MES). • MES translates query into SQL and sends to database hosting MCAT • Database query returned to MES • Location of file etc returned to SRB Server A. • SRB Server A contacts SRB Server B hosting data file. • Data file transferred to user. d MCAT enabled Server c b e SRB Server SRB Server f a g

  7. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh SRB Access Interfaces • Scommands • Unix file system like interface • Versions available for Unix, Dos and Mac • inQ – • Windows Explorer style interface • Version available for Windows • My SRB • Web Browser interface • Client access API • C API • JARGON – Java API to SRB. • MATRIX – SRB Workflow management system

  8. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh SRB User Interfaces Scommands – primary interface to SRB Bash like commands for interacting with SRB. Versions available for Unix, Windows and Mac Can also be used for scripting for batch processing.

  9. InQ – Windows Explorer Style interface to SRB eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh SRB User Interfaces • Primary highlights includes • Support for drag and drop between Windows mounted filesystems • Provisions for access control enforcements, file replication, metadata entry and metadata query.

  10. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh SRB User Interfaces • My SRB Interface – Web browser interface to SRB • Web based interface to SRB space. • Works through port 80 and hence works around firewall issues.

  11. SRB Admin Tool For managing: Users Domains Resources Collaborating Machines Collaborating Zones (version 3 series) eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh SRB Admin Tool

  12. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh Behind the scenes • Behind the scenes SRB provides many other functionalities in managing files and resources • Supports grouping of multiple physical resources into a logical resource. • Support for direct Client Server parallel file transfers for performance improvements • Support for bulk transfer of multiple small files into SRB server • Supports grouping of multiple files into ‘containers’ which is then manageable for insertion and retrieval from Mass Storage systems. • Fine Grained Access Control • Meta Data Query and File Replication between resources

  13. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh Eminerals Project • e-Minerals:Environment from the Molecular Level • UK e-Science project for investigating the environment from the molecular level • Current collaborating Institutions • Cambridge • Royal Institution • University College London • Bath • Reading • CCLRC Daresbury Laboratory

  14. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh eMinerals SRB Domain Scommands InQ Web Browser DL databases Data Portal MySRB Jinx Elektra SRB Application Server CCLRCFS SRB SRB SRB SRB UCLLake CambsLake cambridgeVault Storage Vaults CCLRC UCL Cambridge

  15. Oracle DB1 Oracle DB2 eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh • eMinerals integration with Data Portal SRB Tools Web Browser Files maybe inserted and retrieved using SRB tools Also files may be accessed via the data portal if user Data Portal is also given access privileges The Data portal may utilize different meta data to locate files HPC Portal Data Portal SRB

  16. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh Use of SRB on Eminerals • Use of SRB with Condor/Globus at Cambridge Condor Client Tools/ DAG Clients SRB Space (Used as a Project file system) Scientist @ Cambridge Condor G interface • Code Repository • Linux • Solaris Globus 2.4 Condor Pool Data Repository Head Node Computing Cluster Scientist @ UCL

  17. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh CMS SRB • CMS is one of the Large Hadron Collider (LHC) particle accelerator experiments at CERN. • For CMS DC04 SRB was used at • RAL - UK • IN2P3 - France • Karlsruhe - Germany • CERN - Switzerland • MCAT hosted at RAL

  18. SRB Disk Server CERN SRB Disk Server1 RAL SRB Disk Server2 RAL SRB MCAT Server Atlas Storage Driver Atlas 1 PB Tape Robot and 5 TB Disk Cache at RAL eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh Disk Server at IN2P3 Tape Servers at IN2P3 Oracle Database Cluster at RAL running SRB MCAT Disk Servers at FZK Oracle Database Cluster at DL running SRB MCAT

  19. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh • File Transfer Agents were implemented at CMS for transferring files from source to the T1 centers. Transfer Management DB CERN T1 Centers Export Buffer T1 RAL MSS RAL Data Generator Export Buffer T1 IN2P3 MSS IN2P3 Export Buffer T1 FZK

  20. CCLRC DL CCLRC RAL Oracle Data Guard Databases Databases eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh MES and Database Oracle Client MCAT Enable Server Application Server Databases are clustering on each site and mirrored between Daresbury and Rutherford Appleton

  21. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh Statistics • Roughly about 260,000 connection were seen per day • It is currently stores about 48 TB of data • 1 million data files • 70 Physical Resources • 18 Countries

  22. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh National Grid Service The National Grid Service is a new initiative at UK to provide computing resources to the e-Science community • Data Nodes • RAL – 20 node data cluster with 20TB RAID • Manchester – 20 node data cluster with 20TB RAID • Computational Nodes • Oxford – 64 node dual Xeon cluster • Leeds – 64 node dual Xeon cluster • Services • Computational service • File management service • Database service

  23. SRB DB eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh SRB Federation Zone A • Enables an SRB Network to recognize the presence of another SRB Network and be able to interact with it. • The overall federation can still be viewed at one single namespace whereby each participating Zone is a child node of the root node. • This keeps the learning curve very low and travelling from one Zone to another is like changing directory. • To go to Zone B: • # Scd /B/home/ananta.Adomain • To copy a file back to Zone A • # Scp foo.txt /A/home/projectX Zone B Trust Relation SRB DB Trust Relation Trust Relation SRB DB Zone C

  24. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh Why Federation • As Federation enables SRB zones to recognize and interact with other zones it opens many possibilities in the types of interaction • Federation Enables Fragmentation of MCAT Database An MCAT fragment coming down due to network or system failure does not bring the entire SRB down. The other sections continue to provide the server leaving only one section temporarily inaccessible. • With federation, meta data can be made to reside closer to the resource by hosting a MCAT server closer to the resource. This enables reducing network latency, thus may enable an overall improvement in performance when an SRB network spans across large geographical distance. • Federation enables local administrative control of resources enabling the notion of Virtual Organizations (VO) and also make it easy for a new VO to join the network • It also enables easy replication and transfer of documents between VO

  25. Oracle Cluster SRB Access Point eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh SRB deployment on NGS DL Access Point Leeds Access Point Leeds Domain Manchester MCAT Trust Relationship Manchester Domain Manchester Zone RAL MCAT Oxford Access Point Oracle Cluster Oxford Domain SRB Access Point 1 SRB Access Point 2 RAL Zone RAL Domain

  26. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh RAL Configuration Oracle 9i RAC (Nodes 1.. 8) Trust Relation Manchester Zone ….. Oracle Client Oxford SRB MCAT Server (SRB v3.1) RAL SRB RAL Zone User / Process

  27. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh Use of GSI on NGS SRB • Verification of the authenticity of the user is left to the certification authority rather than by the use of password stored in the MCAT • MCAT maps the user’s DN to its internal Access control mechanism for providing reliable service. • Benifit: • User can be authenticated to any participating MCAT server without contacting its local MCAT. • A computational process may work on behalf of the user provided that the user has delegated its proxy certificate. • Enables single sign-on for all applications as they all recognize the same authentication framework.

  28. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh User / Process • SRB utilization by Projects on NGS Ematerials Domain at RAL Ematerials Domain at Man /ral/home/ematProj/bin /ral/home/ematProj/data /man/home/ematProj/bin /man/home/ematProj/data Job Execution at Oxford Computing Farm Synchronization of Project Users, Replication of files Job Execution at Leeds Computing Farm Temporary file storage at Oxford SRB Vault Temporary file storage at Leeds SRB Vault Oxford computing Farm Leeds computing Farm

  29. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh Benefits for Projects • Immediate Benefits for projects • Better reliability – One of the site being inaccessible will not bring down the system • Improved performance – Reduction in network latency in WAN environments • Interoperability with other Grid applications deployed on NGS by the use of same authentication mechanism (GSI) • Easy sharing of data within the project and between projects • Provision of storage and computing resource

  30. eScience Grid Environments- 8-12th May 2004 NESC - Edinburgh • Questions ..

More Related