1 / 23

EDG Testbed Status Moving to Testbed Two

EDG Testbed Status Moving to Testbed Two. Outline. Current production status. Testbed at RAL. Testbed two. Changes from testbed one. LCFG -> LCFGng Software by node type. Status of integration of testbed two. Current Application TB Status.

tacy
Télécharger la présentation

EDG Testbed Status Moving to Testbed Two

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. EDG Testbed StatusMoving to Testbed Two

  2. Outline • Current production status. • Testbed at RAL. • Testbed two. • Changes from testbed one. • LCFG -> LCFGng • Software by node type. • Status of integration of testbed two. 28th April 2003

  3. Current Application TB Status • Recommended testbed is still RH6.2 , edg1.4.9 with LCFG. • Currently eight UK sites contribute to the the total seventeen in datagrid. • Not changed this year since the BD-II was introduced. The RB is again the limiting factor now the information system is reliable. 28th April 2003

  4. Production Testbed at RAL. • 1 CE, 1 SE (350 GB), 10 WNs, 1 UI. • Top edgapp GIIS now at RAL. • 1 CE as a gatekeeper into tier1a system. • In use by Atlas, Babar, LHCb and DZero, 28th April 2003

  5. Running Production Jobs One month to 27th April 28th April 2003

  6. Development Testbed at RAL • CE, SE, WN, MON, RLS, VOMS and LCFGng exist already. • UI exists (gppui06.gridpp.rl.ac.uk) • RB, HLR and PA to be installed this week. • Updates happen twice a day on average. • Improvements now faster than the addition of software. 28th April 2003

  7. Testbed Two • Lots more node types. • Everything is incompatible including schema, globus, gridftp,… • Still a lot of testing to be done. Loose cannons are not yet loose. • GDMP vanishes which is good for integration of software into farms. 28th April 2003

  8. LCFG -> LCFGng • LCFGng is definitely an improvement. • NIS clients can be configured. • DMA can be turned on. • Using autofs is now the default. • Node profile updates happen immediately and reliably. • Reboot for ever does not happen. • PXE support built in from the start.http://gpplcfg.gridpp.rl.ac.uk/install/install.cgi • LCFGng generally is more complete. • WP4s instructions are better and time proof. • LCFGng has a web interface. 28th April 2003

  9. LCFG ->LCFGng(2) • Each node and object reports back its status. • http://gpplcfg.gridpp.rl.ac.uk/status/. • Middleware configuration completed by the developers and is a lot more ‘intelligent’. • EDG profiles are more modular and clearer than before. • Hardware support is still limited, e.G. Raid, SCSI, e1000 all require special case kernels. • Post install notes will be supplied but are smaller, e.g gridmapdir is mounted, pool account lock files created, site GIIS configured to accept registrations from SE. 28th April 2003

  10. Compute Element Node • Now use’s Maui 3.2.6 for scheduling. • Information providers claimed to support this. • MPICH is installed. • Needed for wp1’s support of MPI jobs. • GLOBUS v2.2.4 supplied from VDT 1.1.8. • VDT. Virtual data toolkit from iVDgl. • Gatekeeper, MDS and gridftp server. • R-GMA client • Publishing CE information via GIN. 28th April 2003

  11. Computing Element(2) • GridFTP logs published into R-GMA. • MSA. Monitoring Sensor Agent. • Application software. • DGAS client. • One CE per site is required. 28th April 2003

  12. Worker Node • Application Software. • Globus clients, GridFTP. • RFIO clients. • MSA, Monitoring Sensor Agent. • VOMS, R-GMA, RLS, Reptor, Netcost and SE clients. • At least one per site required. 28th April 2003

  13. Storage Element • Globus, GridFTP and MDS • Information Providers publishing via MDS and GIN, R-GMA. • GridFTP logs published into R-GMA. • Replica Location -> Site Replica mapping. • SE (apache) and SE web service (tomcat). • One per site, possibly one per media, e.g. CASTOR, Atlas Data Store, Disk? • MSA. Monitoring Sensor Agent. 28th April 2003

  14. User Interface • Clients for: • Datagrid job submission. • Globus job submission. • R-GMA • SE • VOMS • RLS, Reptor, Optor • Network cost client. • Access required by all users of datagrid. 28th April 2003

  15. Monitor Box • Two distinct functions. • R-GMA servlets(tomcat) • CE, SE and application producers register here. • The fmonServer collects information from all the MSA. • Sensors such as lm_sensors, load, uptime, network I/O collected on the MON box. • MySQL. • One per site required, unless tomcat is moved elsewhere. 28th April 2003

  16. Replica Location Service Node • Tomcat • MySQL • R-GMA client, publishes service status. • Replica catalogue. • Metadata catalogue. • Replaces the current replica catalogue. • One node per VO until VOMS is integrated. 28th April 2003

  17. Information Catalogue Node • R-GMA registry servlets. • Tomcat • MySQL • One node required per testbed. 28th April 2003

  18. Network Monitor • Iperf – Network bandwidth. • Pinger – RTT time. • UDPmon – UDP throughput. • rTPL – A combination of the above. • Netagent – Network traffic from interface to router. • All published via apache and perl cgi. • One per site. 28th April 2003

  19. VOMS and MyProxy node • VOMS • Apache (mod_perl) • MySQL • One per VO required. • VOMS will not be in TB2. • MyProxy • MyProxy server. • At least one per testbed. 28th April 2003

  20. WP1 Nodes • DGAS (dg accounting service)  3 nodes. • Deployment commences this week. • RB ( Resource Broker) • At least one per testbed. • HLR (Home Location Register Node) • Stores the accounts of users and resources. • One per testbed. (or site?). • PA (Price Authority Node) • Assigns prices to resources. • One per VO. 28th April 2003

  21. Status of Integration • Lots of parallel changes, lots of new software. • Impossible to follow. • Globus job submission is working, with some magic. • Fabric Management and Network Monitoring complete. • Information system , R-GMA was working well but introduction of GLUE has required a re-release. • RLS service has been shown to work. • Integration of the SE and SE with RLS is not completed. 28th April 2003

  22. Status of Integration(2) • Introduction of WP1 software this week and its success is critical to a release date. • Job management interfaces to almost everything though this is the final component. • May is expected to consist of continuous bug fixes once all software has been deployed. 28th April 2003

  23. Testbed Two in the UK • For a UK independent grid the UK will need to support the gridpp VO. • RB, HLR(?), PA, RLS, VOMS. • Once the RB is in place (IC) then it makes sense to move interested sites to testbed two. • What happens next? • LCG1 , Crossgrid and EDG boundaries become blurred. 28th April 2003

More Related