1 / 18

FermiGrid-HA (Delivering Highly Available Grid Services using Virtualization)

This paper discusses the architecture, implementation, and challenges of FermiGrid-HA, a system that ensures highly available grid services using virtualization. It explores the requirements, challenges, and future directions of the system.

wjessee
Télécharger la présentation

FermiGrid-HA (Delivering Highly Available Grid Services using Virtualization)

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. FermiGrid-HA(Delivering Highly Available Grid Services using Virtualization) Keith Chadwick Fermilab chadwick@fnal.gov Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.

  2. Outline • FermiGrid - Architecture & Performance • FermiGrid-HA - Why? • FermiGrid-HA - Requirements & Challenges • FermiGrid-HA - Implementation • Future Work • Conclusions FermiGrid-HA

  3. FermiGrid - Architecture (mid 2007) Step 2 - user issues voms-proxy-init user receives voms signed credentials Step 5 – Gateway requests GUMS Mapping based on VO & Role Step 3 – user submits their grid job via globus-job-run, globus-job-submit, or condor-g Step 4 – Gateway checks against Site Authorization Service Step 6 - Grid job is forwarded to target cluster VOMRS Server Periodic Synchronization VOMS Server Periodic Synchronization GUMS Server Step 1 - user registers with VO Site Wide Gateway Gratia clusters send ClassAds via CEMon to the site wide gateway SAZ Server FERMIGRID SE (dcache SRM) BlueArc Exterior Interior CMS WC1 CMS WC2 CMS WC3 CDF OSG1 CDF OSG2 D0 CAB1 D0 CAB2 GP Farm GP MPI FermiGrid-HA

  4. FermiGrid-HA - Why? • The FermiGrid “core” services (GUMS and/or SAZ) control access to: • Over 2500 systems with more than 12,000 batch slots (today). • Petabytes of storage (via gPlazma / GUMS). • An outage of either GUMS or SAZ can cause 5,000 to 50,000 “jobs” to fail for each hour of downtime. • Manual recovery or intervention for these services can have long recovery times (best case 30 minutes, worst case multiple hours). • Automated service recovery scripts can minimize the downtime (and impact to the Grid operations), but still can have several tens of minutes response time for failures: • How often the scripts run, • Scripts can only deal with failures that have known “signatures”, • Startup time for the service, • A script cannot fix dead hardware. FermiGrid-HA

  5. FermiGrid-HA - Requirements • Requirements: • Critical services hosted on multiple systems (n ≥ 2). • Small number of “dropped” transactions when failover required (ideally 0). • Support the use of service aliases: • VOMS: fermigrid2.fnal.gov -> voms.fnal.gov • GUMS: fermigrid3.fnal.gov -> gums.fnal.gov • SAZ: fermigrid4.fnal.gov -> saz.fnal.gov • Implement “HA” services with services that did not include “HA” in their design. • Without modification of the underlying service. • Desirables: • Active-Active service configuration. • Active-Standby if Active-Active is too difficult to implement. • A design which can be extended to provide redundant services. FermiGrid-HA

  6. FermiGrid-HA - Challenges #1 • Active-Standby: • Easier to implement, • Can result in “lost” transactions to the backend databases, • Lost transactions would then result in potential inconsistencies following a failover or unexpected configuration changes due to the “lost” transactions. • GUMS Pool Account Mappings. • SAZ Whitelist and Blacklist changes. • Active-Active: • Significantly harder to implement (correctly!). • Allows a greater “transparency”. • Reduces the risk of a “lost” transaction, since any transactions which results in a change to the underlying MySQL databases are “immediately” replicated to the other service instance. • Very low likelihood of inconsistencies. • Any service failure is highly correlated in time with the process which performs the change. FermiGrid-HA

  7. FermiGrid-HA - Challenges #2 • DNS: • Initial FermiGrid-HA design called for DNS names each of which would resolve to two (or more) IP numbers. • If a service instance failed, the surviving service instance could restore operations by “migrating” the IP number for the failed instance to the Ethernet interface of the surviving instance. • Unfortunately, the tool used to build the DNS configuration for the Fermilab network did not support DNS names resolving to >1 IP numbers. • Back to the drawing board. • Linux Virtual Server (LVS): • Route all IP connections through a system configured as a Linux virtual server. • Direct routing • Request goes to LVS director, LVS director redirects the packets to the real server, real server replies directly to the client. • Increases complexity, parts and system count: • More chances for things to fail. • LVS director must be implemented as a HA service. • LVS director implemented as an Active-Standby HA service. • LVS director performs “service pings” every six (6) seconds to verify service availability. • Custom script that uses curl for each service. FermiGrid-HA

  8. FermiGrid-HA - Challenges #3 • MySQL databases underlie all of the FermiGrid-HA Services (VOMS, GUMS, SAZ): • Fortunately all of these Grid services employ relatively simple database schema, • Utilize multi-master MySQL replication, • Requires MySQL 5.0 (or greater). • Databases perform circular replication. • Currently have two (2) MySQL databases, • MySQL 5.0 circular replication has been shown to scale up to ten (10). • Failed databases “cut” the circle and the database circle must be “retied”. • Transactions to either MySQL database are replicated to the other database within 1.1 milliseconds (measured), • Tables which include auto incrementing column fields are handled with the following MySQL 5.0 configuration entries: • auto_increment_offset (1, 2, 3, … n) • auto_increment_increment (10, 10, 10, … ) FermiGrid-HA

  9. FermiGrid-HA - Technology • Xen: • SL 5.0 + Xen 3.1.0 (from xensource community version) • 64 bit Xen Domain 0 host, 32 and 64 bit Xen VMs • Paravirtualisation. • Linux Virtual Server (LVS 1.38): • Shipped with Piranha V0.8.4 from Redhat. • Grid Middleware: • Virtual Data Toolkit (VDT 1.8.1) • VOMS V1.7.20, GUMS V1.2.10, SAZ V1.9.2 • MySQL: • MySQL V5 with multi-master database replication. FermiGrid-HA

  10. FermiGrid-HA - Component Design VOMS Active VOMS Active LVS Active LVS Active MySQL Active GUMS Active Replication Heartbeat Heartbeat MySQL Active GUMS Active LVS Standby LVS Standby SAZ Active SAZ Active FermiGrid-HA

  11. FermiGrid-HA - Host Configuration • The fermigrid5&6 Xen hosts are Dell 2950 systems. • Each of the Dell 2950s are configured with: • Two 3.0 GHz core 2 duo processors (total 4 cores). • 16 Gbytes of RAM. • Raid-1 system disks (2 x 147 Gbytes, 10K RPM, SAS). • Raid-1 non-system disks (2 x 147 Gbytes, 10K RPM, SAS). • Dual 1 Gig-E interfaces: • 1 connected to public network, • 1 connected to private network. • System Software Configuration: • Each Domain 0 system is configured with 5 Xen VMs. • Previously we had 4 Xen VMs. • Each Xen VM, dedicated to running a specific service: • LVS Director, VOMS, GUMS, SAZ, MySQL • Previously we were running the LVS director in the Domain-0. FermiGrid-HA

  12. FermiGrid-HA - Actual Component Deployment • Standby fg5x1 LVS • Active fg5x1 LVS Xen VM 0 Xen VM 0 VOMS • Active fg5x1 • Active fg5x1 VOMS Xen VM 1 Xen VM 1 GUMS GUMS Active fg5x2 Active fg5x2 Xen VM 2 Xen VM 2 Active fg5x3 Active fg5x3 SAZ SAZ Xen VM 3 Xen VM 3 MySQL Active fg5x4 MySQL Active fg5x4 Xen VM 4 Xen VM 4 Xen Domain 0 Xen Domain 0 Active fermigrid5 Active fermigrid6 FermiGrid-HA

  13. FermiGrid-HA - Performance • Stress tests of the FermiGrid-HA GUMS deployment: • A stress test demonstrated that this configuration can support ~9.7M mappings/day. • The load on the GUMS VMs during this stress test was ~9.5 and the CPU idle time was 15%. • The load on the backend MySQL database VM during this stress test was under 1 and the CPU idle time was 92%. • Stress tests of the FermiGrid-HA SAZ deployment: • The SAZ stress test demonstrated that this configuration can support ~1.1M authorizations/day. • The load on the SAZ VMs during this stress test was ~12 and the CPU idle time was 0%. • The load on the backend MySQL database VM during this stress test was under 1 and the CPU idle time was 98%. • Stress tests of the combined FermiGrid-HA GUMS and SAZ deployment: • Using a GUMS:SAZ call ratio of ~7:1 • The combined GUMS-SAZ stress test which was performed yesterday (06-Nov-2007) demonstrated that this configuration can support ~6.5 GUMS mappings/day and ~900K authorizations/day. • The load on the SAZ VMs during this stress test was ~12 and the CPU idle time was 0%. FermiGrid-HA

  14. FermiGrid-HA - Production Deployment • FermiGrid-HA was deployed in production on 03-Dec-2007. • In order to allow an adiabatic transition for the OSG and our user community, we ran the regular FermiGrid services and FermiGrid-HA services simultaneously for a three month period (which ended on 29-Feb-2008). • We have already utilized the HA service redundancy on several occasions: • 1 operating system “wedge” of the Domain-0 hypervisor together with a “wedged” Domain-U VM that required a reboot of the hardware to resolve. • multiple software updates. FermiGrid-HA

  15. FermiGrid - Current Architecture Step 2 - user issues voms-proxy-init user receives voms signed credentials Step 5 – Gateway requests GUMS Mapping based on VO & Role Step 3 – user submits their grid job via globus-job-run, globus-job-submit, or condor-g Step 4 – Gateway checks against Site Authorization Service Step 6 - Grid job is forwarded to target cluster VOMRS Server Periodic Synchronization VOMS Server VOMS Server Periodic Synchronization GUMS Server Step 1 - user registers with VO GUMS Server Site Wide Gateway Gratia clusters send ClassAds via CEMon to the site wide gateway SAZ Server FERMIGRID SE (dcache SRM) SAZ Server BlueArc Exterior Interior CMS WC1 CMS WC2 CMS WC3 CDF OSG1 CDF OSG2 CDF OSG3/4 D0 CAB1 D0 CAB2 GP Farm GP MPI FermiGrid-HA

  16. FermiGrid-HA - Future Work • Over the next three to four months we will be deploying “HA” instances of our other services: • Squid, MyProxy (with DRDB), Syslog-Ng, Ganglia, and others. • Redundant side wide gatekeeper: • We have a preliminary “Gatekeeper-HA” design • Based on the “manual” procedure to keep jobs alive during OSG 0.6.0 to 0.8.0 upgrade that Steve Timm described at a previous site administrators meeting. • We expect that this should keep Globus and Condor jobs running. • We also plan to install a test gatekeeper that will be configured to receive Xen VMs as Grid jobs and execute them: • This a test of a possible future dynamic “VOBox” or “Edge Service” capability within FermiGrid. FermiGrid-HA

  17. FermiGrid-HA - Conclusions • Virtualisation benefits: • Significant performance increase, • Significant reliability increase, • Automatic service failover, • Cost savings, • Can be scaled as the load and the reliability needs increase, • Can perform “live” software upgrades and patches without client impact. • Virtualisation drawbacks: • More complex design, • More “moving parts”, • More opportunities for things to fail, • More items that need to be monitored. FermiGrid-HA

  18. Fin • Any Questions? FermiGrid-HA

More Related