1 / 29

Enterprise-Grade Scheduling

Enterprise-Grade OpenStack from a Scheduler Perspective. Enterprise-Grade Scheduling. Gary Kotton - VMware Gilad Zlotkin - Radware. 1. Enterprise Ready Openstack. Migrating existing mission critical and performance critical enterprise applications requires: → High service levels

kobe
Télécharger la présentation

Enterprise-Grade Scheduling

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Enterprise-Grade OpenStackfrom a Scheduler Perspective Enterprise-Grade Scheduling Gary Kotton - VMware Gilad Zlotkin - Radware 1

  2. Enterprise Ready Openstack Migrating existing mission critical and performance critical enterprise applications requires: → High service levels • Availability • Performance • Security → Compliance with existing architectures • Multi-tier • Fault tolerance models 2

  3. Service Level for Applications • Availability • Fault Tolerance (FT) • High Availability (HA) • Disaster Recovery (DR) 3

  4. Service Level for Applications • Availability • Fault Tolerance (FT) • High Availability (HA) • Disaster Recovery (DR) • Performance • Transaction Latency (Sec) • Transaction Load/Bandwidth (TPS) 3

  5. Service Level for Applications • Availability • Fault Tolerance (FT) • High Availability (HA) • Disaster Recovery (DR) • Performance • Transaction Latency (Sec) • Transaction Load/Bandwidth (TPS) • Security • Data Privacy • Data Integrity • Denial of Service 3

  6. Service Level for Applications • Availability • Fault Tolerance (FT) • High Availability (HA) • Disaster Recovery (DR) • Performance • Transaction Latency (Sec) • Transaction Load/Bandwidth (TPS) • Security • Data Privacy • Data Integrity • Denial of Service What all this has to do with the Nova and other Schedulers? 3

  7. Placement Strategies • Availability - anti affinity • Application VMs should be placed in different 'failure domains' (e.g., on different hosts) to ensure application fault tolerance 4

  8. Placement Strategies • Availability - anti affinity • Application VMs should be placed in different 'failure domains' (e.g., on different hosts) to ensure application fault tolerance • Performance • Network proximity • Application VMs should be placed as closely as possible to one another on the network (same 'connectivity domain') to ensure low latency and high performance 4

  9. Placement Strategies • Availability - anti affinity • Application VMs should be placed in different 'failure domains' (e.g., on different hosts) to ensure application fault tolerance • Performance • Network proximity • Application VMs should be placed as closely as possible to one another on the network (same 'connectivity domain') to ensure low latency and high performance • Host Capability • IO-Intensive, Network-Intensive, CPU-Intensive,... • Storage Proximity 4

  10. Placement Strategies • Availability - anti affinity • Application VMs should be placed in different 'failure domains' (e.g., on different hosts) to ensure application fault tolerance • Performance • Network proximity • Application VMs should be placed as closely as possible to one another on the network (same 'connectivity domain') to ensure low latency and high performance • Host Capability • IO-Intensive, Network-Intensive, CPU-Intensive,... • Storage Proximity • Security - Resource Isolation/Exclusivity • Host, Network, ... 4

  11. Group Scheduling • Group together VMs to provide a certain service • Enables scheduling policies per group/sub-group • Provides a multi-VM application designed for fault tolerance and high performance 5

  12. Example 6

  13. Example Bad placement: if a host goes down entire service is down! 6

  14. Placement strategy - anti affinity: achieving fault tolerance Example Bad placement: if a host goes down entire service is down! 6

  15. Group Scheduling in Nova • Expand on “Server Group” support • Topology of resources and relationships between them • Debi Dutta and Yathi Udupi (Cisco) • Mike Spreitzer (IBM) • Gary Kotton (VMware) 7

  16. Anti Affinity in Icehouse • Server groups • Create a server group with a policy • Scheduler hint will ensure that the policy is enforced • Affinity and anti-affinity filters are now default filters • Backward compatible with Havana support • nova boot --hint group=name|uuid--image ws.img --flavor 2 --num 3 Wsi • Completed with the help of: • Russell Bryant (Red Hat) • Xinyuan Huang (Cisco) 8

  17. Beyond Nova Scheduler Providing enterprise grade services: • Availability • Performance • Security Requires: → Hierarchical Scheduling → Cross Scheduling → Rescheduling 9

  18. Storage/Compute Cross Scheduling VMware Storage Policy Based Management (SPBM) → Datastore selection based on  flavor meta data →Compute scheduling on hosts that are connected to the selected datastore 10

  19. Storage/Compute Cross Scheduling Storage Policy Wizard VMware Virtual SAN SPBM virtual disk SPBM Datastore Profile VSAN objects may be (1) mirrored across hosts & (2) striped across disks/hosts to meet VM storage profile policies object manager object 11

  20. Storage/Compute Cross Scheduling Storage/Compute Affinity: Cinder Volume will be on the same host as VM (to optimize performance) 12

  21. Rescheduling VMware vCenter HA(high availability) Automatically orchestrate rescheduling of VMs from failed host on surviving hosts 13

  22. Rescheduling VMware vCenter HA(high availability) Automatically orchestrate rescheduling of VMs from failed host on surviving hosts DRS(Distributed resource scheduling) Evacuate (vMotion) VMs from overloaded host DPM(Distributed Power Management) Evacuate (vMotion) VMs from host to be shutdown 13

  23. Rescheduling Radware’s Neutron LBaaS rescheduling: • Fault-recovery: rescheduling of failed LB instance • Scaling: rolling scale-up (or down) by controlled fail-over X 13

  24. Hierarchical Scheduling Host exclusivity for secure tenant isolation → Select hosts that were allocated to workload of that specific tenant → Compute scheduling on that selected host group 14

  25. Supported by Icehouse Server-Groups: →Anti-affinity →Affinity →Backward Compatible with Havana 15

  26. Future Roadmap Server-Groups: • New filters • Network Proximity • Rack affinity/anti-affinity • Host capabilities • Simultaneous scheduling (Mike Spreitzer - IBM) • Host exclusivity (Phil Day – HP) 16

  27. Future Roadmap Cont. Scheduler-as-a-service project • Gantt (https://wiki.openstack.org/wiki/Gantt) • Initial steps (Sylvain Bauza - Red Hat): • Forklift the Nova scheduler • Discussions of API’s etc. No DB scheduler (Boris Pavlovic – Mirantis) • Ideas on improving scheduler performance 17

  28. Summary High service levels → Scheduling Policies • Availability • Anti-Affinity • Rescheduling • Performance • Proximity • Host Capability • Cross Scheduling & Rescheduling • Security • Resource Exclusivity • Hierachical Scheduling 18

  29. Q&A Thank You Gary Kotton: gkotton@vmware.com Gilad Zlotkin: giladz@radware.com

More Related