1 / 25

Large Scale Parallel File System and Cluster Management ICT, CAS

Large Scale Parallel File System and Cluster Management ICT, CAS. About ICT, CAS. Institute of Computing Technology, Chinese Academy of Science The first (from 1958) and largest national IT research institute in China The largest graduate school of Computer Science in China

baina
Télécharger la présentation

Large Scale Parallel File System and Cluster Management ICT, CAS

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Large Scale Parallel File System and Cluster ManagementICT, CAS

  2. About ICT, CAS • Institute of Computing Technology, Chinese Academy of Science • The first (from 1958) and largest national IT research institute in China • The largest graduate school of Computer Science in China • Builder of most Chinese systems in HPC TOP 500 • Focusing on computing system architecture: CPU, Compiler, Network, Grid, HPC and Storage

  3. Storage Centre of ICT • Founded in 2001 • Leader: Dr. Xu Lu (from HP Lab) • Storage for scientific computing • BWFS: Parallel cluster file system • Service on Demand system: Storage-based cluster management system. • Storage for business computing • VSDS: Virtual storage research project • Backup / Virtual Computing……

  4. The Storage Bottleneck of Cluster • NFS (Network File System) • Most widely used in clusters to provide shared data access • Simple and easy to use and management • Scalability Problem • Multiple NFS server means multiple name space • Hard to extend in capacity. • The performance do not increase with the capacity • Parallel Access Problem • Poor performance in I/Odensity computing • Weak MS Windows support

  5. What’s BWFS • Parallel network file system • Support multiple storage appliances (8-128) in a single name space (Up to 512 TB) • Separated Data and Meta-Data access to provide parallel accessing between different storage appliance • Global name space between clients with different platforms • Fully compatible with NFS (not 100% POSIX) • Support data sharing between Linux and Windows clients • Support IA32, IA64 and x86_64 hardware platforms

  6. What’s BWFS • Centralized Management • Web based management for the storage appliances and the storage sub-system • Integrated client management with Service on Demand system. • Online extension • Add storage appliances to increase the capacity without stopping the application • The new data will be automatically stripped between all the storage appliances to get a high performance.

  7. Data Access on NFS Meta-Data User Data ` `

  8. Data Access on BWFS User-Data Meta-Data

  9. write large files(20G per node, 1MB record size) 350 300 1SN 250 2SN Aggregate Bandwidth (MB/s) 200 read large files(20G per node, 1MB record size) 150 4SN 350 100 300 NFS 1SN 50 250 Aggregate Bandwidth(MB/s) 2SN 0 200 1 2 4 8 16 Number of client nodes 150 4SN 100 NFS 50 0 1 2 4 8 16 Number of client nodes Bandwidth of BWFS

  10. Paradigm Epos3 (China Petrol, Xinjiang)

  11. Paradigm Disco (China Petrol, Xinjiang)

  12. Management Interface

  13. Service on Demand System • Initially developed as a subsystem of BWFS to provide cluster management • Reduce the management work especially in the system deployments • Increase the availability against the storage components fail • Enable the fast schedule in large server farms with multiple clusters • Boot the system directly from the BWFS storage appliance without the need of local hard disks

  14. Traditional Cluster Deployment System 20mins

  15. Shortcoming 1: Inefficiency in Schedule 20 mins

  16. Shortcoming 2: Inefficiency in Maintains Hard disk errors occupy 30%-50% of all the computer system errors

  17. Shortcoming 3: Inefficiency in Capacity A 5GB system on a 74GB hard disk The disks are getting larger and larger but the system images are keeping small to reduce deployment time

  18. Service on Demand System • Diskless boot OS by TCP/IP • Virtual SCSI disk to support Windows and Linux • Fully compatible with applications • Provide high performance snapshots to support fast cloning of system images • Copy on Write when the system image is modified • Online backup system image with snapshot • Automatic take over on failed clients • Integrated monitor engine to support automatic schedule or adaptive computing (still in researching)

  19. Map to Local Disk Network Service on Demand System Service 2 Service N Service 1

  20. Fast Deployment and Schedule Paradigm Services Paradigm Image Paradigm Snapshot Paradigm Snapshot Paradigm Snapshot Web 系统 Paradigm Snapshot Paradigm Snapshot CGG Services Email 系统 CGG Image CGG Snapshot

  21. Easy to maintain System Image System Snapshot System Snapshot System Snapshot System Snapshot System Snapshot Maintenance

  22. Management UI

  23. Thanks 谢谢!

More Related