1 / 29

The National e-Science Centre and UK e-Science Programme

The National e-Science Centre and UK e-Science Programme. Muffy Calder University of Glasgow http: //www.nesc.ac.uk http://www.dcs.gla.ac.uk/~muffy. An overview. E-Science challenges and opportunities Grid computing challenges and opportunities Activities UK and international

usoa
Télécharger la présentation

The National e-Science Centre and UK e-Science Programme

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. The National e-Science Centreand UK e-Science Programme Muffy Calder University of Glasgow http: //www.nesc.ac.uk http://www.dcs.gla.ac.uk/~muffy

  2. An overview • E-Science • challenges and opportunities • Grid computing • challenges and opportunities • Activities • UK and international • Scotland and National e-Science Centre

  3. e-Science ‘e-Science is about global collaboration in key areas of science, and the next generation of infrastructure that will enable it.’ ‘e-Science will change the dynamic of the way science is undertaken.’ John Taylor Director General of Research Councils Office of Science and Technology

  4. The drivers for e-Science • More data • instrument resolution and laboratory automation • storage capacity and data sources • More computation • computations available, simulations doubling every year • Faster networks • bandwidth • need to schedule • More interplay and collaboration • between scientists, engineers, computer scientists etc. • between computation and data

  5. The drivers for e-Science • Exploration of data and models • in silico discovery • Floods of public data • gene sequence doubling every 9 months • Searches required doubling every 4-5 months In summary Shared data, information and computation by geographically dispersed communities.

  6. HPC Experiment Storage Analysis HPC Scientist Experiment Computing Storage Analysis HPC Current model: ad-hoc client-server

  7. Experiment Analysis MIDLEWARE Computing Storage Scientist Storage Experiment Analysis Computing Storage Computing The vision: computation & information utility

  8. e-Science Examples • Bioinformatics/Functional genomics • Collaborative Engineering • Medical/Healthcare informatics • Earth Observation Systems (flood monitoring) • TeleMicroscopy • Virtual Observatories • Robotic Telescopes • Particle Physics at the LHC • EU DataGrid particle physics, biology & medical imaging, Earth observation • GridPP, ScotGrid • AstroGrid

  9. HumanModels Multi-disciplinary Simulations WingModels • Lift Capabilities • Drag Capabilities • Responsiveness AirframeModels StabilizerModels • Deflection capabilities • Responsiveness Crew Capabilities - accuracy - perception - stamina - re-action times - SOP’s EngineModels • Braking performance • Steering capabilities • Traction • Dampening capabilities • Thrust performance • Reverse Thrust performance • Responsiveness • Fuel Consumption LandingGearModels Whole system simulations are produced by couplingall of the sub-system simulations

  10. Engine Models Wing Models Human Models Multi-disciplinary Simulations US NationalAirSpaceSimulationEnvironment Stabilizer Models GRC 44,000 Wing Runs 50,000 Engine Runs Airframe Models 66,000 Stabilizer Runs ARC LaRC Virtual National Air Space VNAS 22,000 Commercial US Flights a day 22,000 Airframe Impact Runs • FAA Ops Data • Weather Data • Airline Schedule Data • Digital Flight Data • Radar Tracks • Terrain Data • Surface Data Simulation Drivers 48,000 Human Crew Runs 132,000 Landing/ Take-off Gear Runs (Being pulled together under the NASA AvSP Aviation ExtraNet (AEN) Landing Gear Models Many aircraft, flight paths, airport operations, and the environment are combined to get a virtual national airspace

  11. in-flight data global network eg SITA ground station airline DS&S Engine Health Center internet, e-mail, pager data centre maintenance centre Global in-flight engine diagnostics Distributed Aircraft Maintenance Environment: Universities of Leeds, Oxford, Sheffield &York

  12. LHC computing assumes PC = ~ 25 SpecInt95 ~PByte/sec Online System ~100 MByte/sec Offline Farm~20,000 PCs • one bunch crossing per 25 ns • 100 triggers per second • each event is ~1 MByte ~100 MByte/sec Tier 0 CERN Computer Centre >20,000 PCs ~ Gbit/sec or Air Freight HPSS Tier 1 RAL Regional Centre US Regional Centre Italian Regional Centre French Regional Centre HPSS HPSS HPSS HPSS Tier 2 Tier2 Centre ~1000 PCs Tier2 Centre ~1000 PCs Tier2 Centre ~1000 PCs ScotGRID++ ~1000 PCs ~Gbit/sec Tier 3 Institute ~200 PCs physicists work on analysis “channels” each institute has ~10 physicists working on one or more channels data for these channels is cached by the institute server Institute Institute Institute Physics data cache 100 - 1000 Mbit/sec Tier 4 Workstations

  13. Emergency response teams • bring sensors, data, simulations and experts together • wildfire: predict movement of fire & direct fire-fighters • also earthquakes, peacekeeping forces, battlefields,… National Earthquake Simulation Grid Los Alamos National Laboratory: wildfire

  14. Earth observation • ENVISAT • € 3.5 billion • 400 terabytes/year • 700 users • ground deformation prior to a volcano

  15. To achieve e-Science we need…

  16. To achieve e-Science we need…

  17. GRID Grid the vision Computing resources Instruments Complexproblem Data People Solution Knowledge

  18. The Grid • Computing cycles, data storage, bandwidth and facilities viewed as commodities. • like the electricity grid • Software and hardware infrastructure to support model of computation and information utilities on demand. • middleware • An emergent infrastructure • delivering dependable, pervasive and uniform access to a set of globally distributed, dynamic and heterogeneous resources.

  19. The Grid Supporting computations with differing characteristics: • Highthroughput • Unbounded, robust, scalable, use otherwise idle machines • On demand • short-term bounded, hard timing constraints • Data intensive • computed, measured, stored, recalled, e.g. particle physics • Distributed supercomputing • classical CPU and memory intensive • Collaborative • mainly in support of hhi, e.g. access grid

  20. Grid Data • Generated by sensor • From a database • Computed on request • Measured on request

  21. Grid Services • Deliver bandwidth, data, computation cycles • Architecture and basic services for • authentication • authorisation • resource scheduling and coscheduling • monitoring, accounting and payment • protection and security • quality of service guarantees • secondary storage • directories • interoperability • fault-tolerance • reliability

  22. But…has the emperor got any clothes on? • Maybe string vest and pants: • semantic web • machine understandable information on the web • virtual organisations • pervasive computing “ … a billion people interacting with a million e-businesses with a trillion intelligent devices interconnected ” Lou Gerstner, IBM (2000)

  23. National and International Activities • USA • NASA Information Power Grid • DOE Science Grid • NSF National Virtual Observatory etc. • UK e-Science Programme • Japan – Grid Data Farm, ITBL • Netherlands – VLAM, PolderGrid • Germany – UNICORE, Grid proposal • France – Grid funding approved • Italy – INFN Grid • Eire – Grid proposals • Switzerland - Network/Grid proposal • Hungary – DemoGrid, Grid proposal • …

  24. AccessGrid always-on video walls UK e-science centres Edinburgh Glasgow DL Newcastle Belfast Manchester Cambridge Oxford Hinxton RAL Cardiff London Soton

  25. National e-Science Centre • Edinburgh + Glasgow Universities • Physics & Astronomy  2 • Informatics, Computing Science • EPCC • £6M EPSRC/DTI + £2M SHEFC over 3 years • e-Science Institute • visitors, workshops, co-ordination, outreach • middleware development • 50 : 50 industry : academia • UK representation at GGF • coordinate regional centres

  26. Generic Grid Middleware • All e-Science Centres will donate resources to form a UK ‘national’ Grid • All Centres will run same Grid Software - based on Globus*, SRB and Condor • Work with Global Grid Forum (GGF) and major computing companies. *Globus – project to develop open architecture, open source Grid software, Globus toolkit 2.0 now available.

  27. How can you participate? • Informal collaborations • Formal collaborative LINK funded projects • Attend seminars at eSI • Suggest visitors for eSI • Suggest themes and topics for eSI.

  28. eSI forthcoming events Friday, March 15Blue Gene Monday, March 18 IWOX: Introductory Workshop on XML Schema Tuesday, March 19 AWOX: Advanced Workshop on XML: XML Schema, Web Services and Tools Thursday, March 21GOGO: Getting OGSA Going 1 Monday, April 8 Software Management for Grid Projects Monday, April 15 DAI: Database Access and Integration Wednesday, April 17 DBFT Meeting Monday, April 29The Information Grid Sunday, July 21 GGF5/ HPDC11

  29. Watch the web pages…. www.nesc.ac.uk Thank you!

More Related