1 / 24

Correctness of Gossip-Based Membership under Message Loss

Correctness of Gossip-Based Membership under Message Loss. Maxim Gurevich , Idit Keidar Technion. The Setting. Many nodes – n 10,000s, 100,000s, 1,000,000s, … Come and go Churn Fully connected network Like the Internet Every joining node knows some others (Initial) Connectivity.

tyson
Télécharger la présentation

Correctness of Gossip-Based Membership under Message Loss

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Correctness of Gossip-Based Membership under Message Loss Maxim Gurevich, Idit Keidar Technion

  2. The Setting • Many nodes – n • 10,000s, 100,000s, 1,000,000s, … • Come and go • Churn • Fully connected network • Like the Internet • Every joining node knows some others • (Initial) Connectivity

  3. Membership: Each Node Needs To Know Some Live Nodes • Applications • Gossip partners • Unstructured overlay networks • Gathering statistics • Work best with random node samples • Gossip algorithms converge fast • Overlay networks are robust, good expanders • Statistics are accurate

  4. Membership Protocols • Each node has a view • Set of node ids • Supplied to the application • Used by membership protocol for maintenance • Modeled as a directed graph w y u v

  5. Desirable Properties • Randomness… • Holy grail for samples: IID • Each sample uniformly distributed • Each sample independent of other samples • Avoid spatial dependencies among view entries • Avoid correlations between nodes • Good load balance among nodes

  6. What About Churn? Desirable Properties Cont’d • Views should constantly evolve • Remove failed nodes, add joining ones • Views should evolve to IID from any state • Minimize temporal dependencies • Dependence on the past should decay quickly • Useful for application requiring fresh samples

  7. Do Existing Protocols Measure Up?

  8. Existing Work: Practical Protocols Example: Push protocol • Studied only empirically • Good load balance [Lpbcast, Jelasity et al 07] • Fast decay of temporal dependencies [Jelasity et al 07] • Induces spatial dependence  w z u v

  9. Existing Work: Analysis w z Shuffle protocol • Analyzed theoretically [Allavena et al 05, Mahlmann et al 06] • Uniformity, load balance, spatial independence  • Unrealistic assumptions  • Atomic actions with bi-directional communication • No message loss • No bounds on decay of temporal dependencies  u v

  10. Our Contribution: Bridge This Gap • Formally specify desirable properties outlined above • A practical protocol • Tolerates message loss, churn, failures • No complex bookkeeping for atomic actions • Formally prove the desirable properties • Including under message loss

  11. Send & Forget Membership • The best of push and shuffle • Some view entries may be empty w u v

  12. S&F: Message Loss • Message loss • Or no empty entries in v’s view w w u v u v

  13. S&F: Compensating for Loss • Edges (view entries) disappear due to loss • Need to prevent views from emptying out • Keep the sent ids when too little ids in view w w u v u v

  14. S&F: Advantages over Other Protocols • No bi-directional communication • No complex bookkeeping • Tolerates message loss • Simple • Amenable to formal analysis Easy to implement

  15. Key Contribution: Analysis • Proving all desirable properties • Analytical: degrees distribution w/out loss • Used in setting duplication threshold • Markov 1: degree distribution with loss • Markov 2: Markov Chain of reachable global states • IID samples, Temporal Independence • Hold even under (reasonable) message loss!

  16. Analytic Degree Distribution • Similar (better) to that of a random graph • Validated by a more accurate Markov model

  17. Key Contribution: Analysis • Proving all desirable properties • Analytical: degrees distribution w/out loss • Used in setting duplication threshold • Markov 1: degree distribution with loss • Markov 2: Markov Chain of reachable global states • IID samples, Temporal Independence • Hold even under (reasonable) message loss!

  18. Node Degree Markov Chain • Numerically compute the stationary distribution … outdegree 0 2 4 6 … … State corresponding to isolated node 0 Transitions without loss … 1 … indegree Transitions due to loss 2 … … 3 …

  19. Results • Outdegree is bounded by the protocol • Decreases with increasing loss • Indegree is not bounded • Low variance even under loss • Typical overload at most 2x

  20. Key Contribution: Analysis • Proving all desirable properties • Analytical: degrees distribution w/out loss • Used in setting duplication threshold • Markov 1: degree distribution with loss • Markov 2: Markov Chain of reachable global states • IID samples, Temporal Independence • Hold even under (reasonable) message loss!

  21. Decay of Spatial Dependencies w w • For uniform loss < 15%, dependencies decay faster than they are created • 1 – 2loss rate fraction of view entries are independent • E.g., for loss rate of 3%  more than 90% of entries are independent … u u v v u does not delete the sent ids

  22. Temporal Independence • Dependence on past views decays within O(log n  view size) time • Use “expected conductance” • Ids travel fast enough • Reach random nodes in O(log n) hops • Due to “sufficiently many” independent ids in views - previous slide

  23. Conclusions • Formalized the desired properties of a membership protocol • Send & Forget protocol • Simple for both implementation and analysis • Analysis under message loss • Load balance • Uniformity • Spatial Independence • Temporal Independence

  24. Thank You

More Related