1 / 14

Memory Consistency Models in Wide-area Storage System – Or What Do They Mean?

Memory Consistency Models in Wide-area Storage System – Or What Do They Mean?. CS258 Spring 2002 Mark Whitney and Yitao Duan. Motivations. Global scale computing approaching Wide area storage is becoming a reality The greed for processing power calls the marriage of the two

denali
Télécharger la présentation

Memory Consistency Models in Wide-area Storage System – Or What Do They Mean?

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Memory Consistency Models in Wide-area Storage System – Or What Do They Mean? CS258 Spring 2002 Mark Whitney and Yitao Duan

  2. Motivations • Global scale computing approaching • Wide area storage is becoming a reality • The greed for processing power calls the marriage of the two • Traditional approach to large scale data processing: Hierarchy • What if new algorithms require to touch more data? Scale SMP? • Use OceanStore as testbed

  3. Data and Computation Hungry Applications • Quantum Chromodynamics • Biomolecular Dynamics • Weather Forecasting • Cosmological Dark Matter • Biomolecular Electrostatics • Electric and Magnetic Molecular Properties

  4. ~PBytes/sec ~100 MBytes/sec Offline Processor Farm ~20 TIPS There is a “bunch crossing” every 25 nsecs. There are 100 “triggers” per second Each triggered event is ~1 MByte in size ~100 MBytes/sec Online System Tier 0 CERN Computer Centre ~622 Mbits/sec or Air Freight (deprecated) Tier 1 FermiLab ~4 TIPS France Regional Centre Germany Regional Centre Italy Regional Centre ~622 Mbits/sec Tier 2 Tier2 Centre ~1 TIPS Tier2 Centre ~1 TIPS Caltech ~1 TIPS Tier2 Centre ~1 TIPS Tier2 Centre ~1 TIPS HPSS HPSS HPSS HPSS HPSS ~622 Mbits/sec Institute ~0.25TIPS Institute Institute Institute Physics data cache ~1 MBytes/sec 1 TIPS is approximately 25,000 SpecInt95 equivalents Physicists work on analysis “channels”. Each institute will have ~10 physicists working on one or more channels; data for these channels should be cached by the institute server Pentium II 300 MHz Pentium II 300 MHz Pentium II 300 MHz Pentium II 300 MHz Tier 4 Physicist workstations Data Grid for High Energy Physics - CalTech

  5. Background • What is OceanStore? • A global persistent data store scalable to billions of users • High availability, fault-tolerance, security • Caching to reduce network congestion, guarantee availability and performance • Flexible consistency semantics • Observations and questions • Remarkable resemblance to MP memory system • Replica = cache, client = processor, data object = memory item • OceanStore consistency semantics are typically that of a file system’s. What do they mean to a program?

  6. Running Parallel Applications on OceanStore • Why do we try this • Distributed computing • Grid • World Wide Computing • New programming paradigm? (OceanStore is a new phenomena, will it bring out new applications? Where will computing infrastructure go given the advance of network, storage and parallel processing?

  7. ParaApp ParaApp OS Kernel OS Kernel ParaApp OS Kernel OClient OClient OClient Shared Virtual Memory Space

  8. Running SMP Apps on OceanStore • OceanStore data objects are globally identified • Virtual address in application address space mapped to OceanStore object ID • Shared memory address access turned into OceanStore requests

  9. Consistency Models

  10. Performance Evaluation • OceanStore …(#of inner rings, …) • Nachos++! • MIPS R3000 processor w/FP • Stanford SPLASH-2 benchmark suite • 4 x 4 matrix LU decomposition

  11. Computation Time Number of Cycles

  12. Network Latency Milliseconds

  13. Open Questions • Programming model • Cache policy • Consistency models • Sharing granularity

  14. Conclusion and Future Work • Matrix decomposition runs on OceanStore! • Wide-area distributed synchronizations are expensive (not surprising) • Need better memory model if want to run shared memory applications • Message passing? – Seems to be a better match (use explicit OceanStore APIs) • New programming model?

More Related