1 / 22

A Scalable Distributed Datastore for BioImaging

A Scalable Distributed Datastore for BioImaging. R. Cai, J. Curnutt , E. Gomez, G. Kaymaz, T. Kleffel, K. Schubert, J. Tafas {jcurnutt, egomez, keith, jtafas}@r2labs.org Renaissance Research Labs Department of Computer Science California State University San Bernardino, CA 92407.

orrh
Télécharger la présentation

A Scalable Distributed Datastore for BioImaging

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. A Scalable Distributed Datastore for BioImaging R. Cai, J. Curnutt, E. Gomez, G. Kaymaz, T. Kleffel, K. Schubert, J. Tafas {jcurnutt, egomez, keith, jtafas}@r2labs.org Renaissance Research Labs Department of Computer Science California State University San Bernardino, CA 92407 Supported by NSF ITR #0331697

  2. Background • CSUSB Institute for Applied Supercomputing • Low Latency Communications • UCSB Center for BioImage Informatics • Retinal images • Texture map searches • Distributed consortium (UCB, CMU)

  3. Retina Images 3 day detachment (3d) Normal (n) 1 day detachment followed by 6 day reattached with increased oxygen (1d+6dO2) 3 month detachment (3m) Laser scanning confocal microscope images of the retina

  4. Environment UCSB Image and metadata server Image and metadata server Local Hammer/Nail Cluster BISQUE Image and metadata server LAN Local features search Raven Cluster Lustre WAN analysis internal external CSUSB

  5. Software • Open source • OME • Postgresql 7 • Bisque • Distributed datastore • Clustering • NFS • Lustre • Benchmark: OSDB

  6. Hardware - Raven • 5 year old dual processor • 1.4 GHz Pentium 3 • 256MB RAM • 60GB SCSI • Compaq Proliant DL-360 servers. • Raven has been latency tuned.

  7. Hardware – Hammer/Nail • Hammer headnode • 5 Nail nodes • quad CPUs • 3.2 Ghz Xeon • 4GB RAM • 140GB SCSI • Dell servers • Bandwidth tuned (default) UCSB CSUSB

  8. Outline • Effect of node configuration • Comparison of network file systems • Effects of a wide area network (WAN)

  9. Relative LAN Performance

  10. NFS LAN/WAN Performance

  11. Design Effects? • A few expert users • Metadata searches • Small results to user • Texture searches • Heavy calculation on cluster • Small results to user • Latency tuning

  12. Outline • Effect of node configuration • Comparison of network file systems • Effects of a wide area network (WAN)

  13. NFS / Luster Performance. • NFS • well known standard • Configuration problems with OME • performance comparison of the Lustre file system • Lustre • Journaling • Stripe across multiple computers • Data redundancy and failover

  14. Relative Performance on LAN • NSF/Lustre • Compared to local DB • 1GB LAN • two significant differences

  15. Significant Differences • NSF caching • bulk deletes and bulk modifies • Lustre stripes across computers • increase the bandwidth

  16. Outline • Effect of node configuration • Comparison of network file systems • Effects of a wide area network (WAN)

  17. Effect on Wide Area Network WAN • Compared three connections • Local • Switched, high speed LAN (1 Gb/s) • WAN between UCSB and CSUSB (~50 Mb/s) • NFS only • UCSB didn’t have Lustre installed • Active research prevented reinstalling OS

  18. Local/LAN/WAN Performance

  19. Effect on Wide Area Network WAN • Most significant effect • Not bandwidth intensive operations • Latency intensive operation • Next generation WAN will not solve the problem. • Frequently used data must be kept locally • Database cluster • Daily sync of remote databases

  20. Conclusions • Scientific researchers • Latency tune network • Don’t bandwidth tune • Latency of WAN is too large • replicate data and update. • Bisque/OME NFS issues • Lustre • High bandwidth operations • Stripe Lustre across systems

  21. Future directions: • Agent based texture search engine • Loosely coupled cluster • WAN connection • Unreliable connection • Fault tollerant • Parallelize Jobs • Open source components • Scilab • Convert NSF funded algorithms in Matlab • Simple interface • Superior caching scheme for Lustre

  22. Questions…

More Related