230 likes | 364 Vues
The presentation discusses the Grid-Based Event Processing System (GEPS), designed to facilitate web-based access to grid computing environments for efficient event processing. Developed on the Globus platform, GEPS emphasizes not moving data, ensuring each node manages a subset of data. Key components include a Job Submit Server (JSS) that processes user queries and retrieves results from grid nodes. The framework incorporates various technologies such as PgSQL and ROOT for data handling. Future work will focus on enhancing error handling and load balancing for improved performance.
E N D
Grid-Brick Event Processing Framework in GEPS CHEP 03 – La Jolla, California A. Amorim, P. Trezentos, N. Almeida, H. Fei, L.Pedro, J.Villate, H.Wolters
luis.pedro@cern.ch FCUL - Lisbon Outline • Introduction • Architecture Overview • Data Flow • How do we do it • Action! • Vantages and disadvantages • On-going and future work • Summary
luis.pedro@cern.ch FCUL - Lisbon Introduction • What is GEPS • Grid-based Event Processing System • Developed on top of Globus • Provides web-based access to grid computing environment for event processing • Developed by GridPT working group
luis.pedro@cern.ch FCUL - Lisbon Introduction • Main Idea Do NOT move data! Each node stores and is reponsible by a subset of the whole data…
High speed Disks Data Center luis.pedro@cern.ch FCUL - Lisbon Introduction • The usual way
Meta-data catalogue User terminal User terminal User terminal luis.pedro@cern.ch FCUL - Lisbon Introduction • Our way Job submit server ………..
luis.pedro@cern.ch FCUL - Lisbon Architecture Overview • User submits a query through a interface to the Job Submit Server (JSS). • Job submitted information will be stored in the Metadata Catalog.
luis.pedro@cern.ch FCUL - Lisbon Architecture Overview • The job is submitted to the grid nodes using Globus API functions. • All the nodes query their own information and retrieve a result.
luis.pedro@cern.ch FCUL - Lisbon Architecture Overview • JSS receives result from the Grid nodes and produces a final one • User can download or consult the final result • User can also visualize the state of the job in each Grid node
luis.pedro@cern.ch FCUL - Lisbon Architecture Overview
Meta-data catalogue Brick Grid node luis.pedro@cern.ch FCUL - Lisbon Data-flow JSS B R O K E R User terminal (PHP interface)
luis.pedro@cern.ch FCUL - Lisbon How do we do it: • Technologies used • Globus • PgSQL • LDAP • PHP • ROOT
luis.pedro@cern.ch FCUL - Lisbon How do we do it: • Relevant Features • Globus • Toolkit that provide GRID API functions • PgSQL • Meta-data catalogue implementation • LDAP • Query Grid node information • PHP • Web interface
luis.pedro@cern.ch FCUL - Lisbon Action! • The human interface Main Page Submit a Job GREED info Job status
luis.pedro@cern.ch FCUL - Lisbon Action! • Enabling ROOT Queries • The job is submitted to the grid nodes • All the nodes query their own information with ROOT and retrieve a ROOT file with a TTree • JSS receives the ROOT files and produces a final ROOT file with the result of the query • User can download or consult the final file because it is a TTree.
luis.pedro@cern.ch FCUL - Lisbon Action! • Enabling ROOT Queries (cont) • Stores the information in each node using TTree’s • Filter the information in each node and retrieve a result file that include a TTree • Join all the result files in the Job Submit Server using a TChain and produces a final TTree that is the query result • View the final result file with a TBrowser or with Carrot
luis.pedro@cern.ch FCUL - Lisbon Action! • Enabling ROOT Queries (reading ROOT files)
luis.pedro@cern.ch FCUL - Lisbon Action! • Enabling ROOT Queries (reading ROOT files)
luis.pedro@cern.ch FCUL - Lisbon Action! • Special Features used • ROOT • TObject • TTree • CINT • Filtering data from TTree • ROOT I/O • TChain • Carrot • Browsing ROOT files • Histograming variables
luis.pedro@cern.ch FCUL - Lisbon Vantages and disavantages • Vantages • Commodity Data Storage • Huge Scalability (400 GB/node) • Granularity • Disadvantages • Load balancing • Suitable storage policy • Fault tolerance • Data replication or Backup
luis.pedro@cern.ch FCUL - Lisbon On-going and future work • Error handling and fault-tolerance • Recover mechanisms for each node • Create a redundancy mechanism to recover from a malfunction in the nodes • Develop a storage mechanism to submit more work to the best nodes • Load balancing • Provide to user several interfaces to submit work
luis.pedro@cern.ch FCUL - Lisbon Summary • A different approach is being developed • There is already a real prototype working • Some (good) results have been achieved • A lot of work already done….. • But …. Still a lot of work to do!
luis.pedro@cern.ch FCUL - Lisbon Acknowledgments • Thank you to those who are developing and participating in this project • A. Amorim (Antonio.Amorim@fc.ul.pt) • P. Trezentos (Paulo.Trezentos@iscte.pt) • N. Almeida (Nuno.Almeida@iscte.pt) • H. Fei (hnfi@iscte.pt) • L.Pedro (Luis.Pedro@cern.ch) • J.Villate (villate@fe.up.pt) • H.Wolters (helmut@lip.pt) • Keep the good work Thanks for hearing me!