1 / 38

e-science in the UK

e-science in the UK. Peter Watkins Head of Particle Physics University of Birmingham, UK p.m.watkins@bham.ac.uk. Outline of Talk. UK e-science funding e-science projects e-science infrastructure GridPP Recent developments. RCUK e-Science Funding. Second Phase: 2003 –2006

merrill
Télécharger la présentation

e-science in the UK

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. e-science in the UK Peter Watkins Head of Particle Physics University of Birmingham, UK p.m.watkins@bham.ac.uk

  2. Outline of Talk • UK e-science funding • e-science projects • e-science infrastructure • GridPP • Recent developments

  3. RCUK e-Science Funding Second Phase: 2003 –2006 • Application Projects • £96M • All areas of science and engineering • Core Programme • £16M Research Infrastructure • £10M DTI Technology Fund First Phase: 2001 –2004 • Application Projects • £74M • All areas of science and engineering • Core Programme • £15M Research infrastructure • £20M Collaborative industrial projects • (+£30M Companies)

  4. UK e-Science Budget (2001-2006) Total: £213M Source: Science Budget 2003/4 – 2005/6, DTI(OST)

  5. EPSRC e-science projects(a few examples)

  6. The e-Science Paradigm • The Integrative Biology Project involves the University of Oxford (and others) in the UK and the University of Auckland in New Zealand • Models of electrical behaviour of heart cells developed by Denis Noble’s team in Oxford • Mechanical models of beating heart developed by Peter Hunter’s group in Auckland • Researchers need to be able to easily build a secure ‘Virtual Organisation’ allowing access to each group’s resources • Will enable researchers to do different science

  7. Comb-e-Chem Project Video Simulation Properties Analysis StructuresDatabase Diffractometer X-Raye-Lab Propertiese-Lab Grid Middleware

  8. DAME Project In flight data Global Network eg: SITA Ground Station Airline DS&S Engine Health Center Maintenance Centre Internet, e-mail, pager Data centre

  9. eDiaMoND Project Mammograms have different appearances, depending on image settings and acquisition systems Computer Aided Detection Standard Mammo Format 3D View

  10. KEGG Inter Pro SMART Execute distributed annotation workflow SWISS PROT EMBL NCBI TIGR SNP GO Interactive Editor & Visualisation Discovery Net Project Nucleotide Annotation Workflows Download sequence from Reference Server Save to Distributed AnnotationServer • 1800 clicks • 500 Web access • 200 copy/paste • 3 weeks work in 1 workflow and few second execution

  11. UK Core e-Science programme -

  12. Access Grid Access Grid

  13. National e-science centre NeSC – Edinburgh • Help coordinate and lead UK e-Science • Community building & outreach • Training for UK and EGEE • Undertake R&D projects • Research visitors and events • Engage industry (IBM, Sun, Microsoft, HP, Oracle, …) • Stimulate the uptake of e-Science technology

  14. AbMAN UHI Network Clydenet FaTMAN EaStMAN NNW C&NL MAN Glasgow Edinburgh NorMAN North Wales MAN YHMAN Warrington Leeds EMMAN MidMAN London EastNet Reading TVN Kentish MAN External Links Bristol Portsmouth Northern Ireland LMN External Links South Wales MAN SWERN LeNSE SuperJANET 4 • UK National Network • Managed by UKERNA • 10 Gbit/s core • Regional Distribution • IP Production Service • Multicast enabled • IPv6 rollout • QoS Rollout • MPLS Capable in core

  15. PPARC e-science - two major projects

  16. Powering the Virtual Universehttp://www.astrogrid.ac.uk(Edinburgh, Belfast, Cambridge, Leicester, London, Manchester, RAL) Multi-wavelength showing the jet in M87: from top to bottom – Chandra X-ray, HST optical, Gemini mid-IR, VLA radio. AstroGrid will provide advanced, Grid based, federation and data mining tools to facilitate better and faster scientific output. Picture credits: “NASA / Chandra X-ray Observatory / Herman Marshall (MIT)”, “NASA/HST/Eric Perlman (UMBC), “Gemini Observatory/OSCIR”, “VLA/NSF/Eric Perlman (UMBC)/Fang Zhou, Biretta (STScI)/F Owen (NRA)” p18 Printed: 20/09/2014

  17. The Virtual Observatory • International Virtual Observatory Alliance UK, Australia, EU, China, Canada, Italy, Germany, Japan, Korea, US, Russia, France, India How to integrate manymulti-TB collections ofheterogeneous data distributed globally? Sociological and technological challenges to be met

  18. GridPP – Particle Physics Grid 19 UK Universities, CCLRC (RAL & Daresbury) and CERN Funded by the Particle Physics and Astronomy Research Council (PPARC) GridPP1 - 2001-2004 £17m "From Web to Grid" GridPP2 - 2004-2007 £15m "From Prototype to Production"

  19. The CERN LHC The world’s most powerful particle accelerator - 2007 4 Large Experiments

  20. LHC Experiments Searching for theHiggs Particle and exciting new Physics e.g. ATLAS Starting from this event Looking for this ‘signature’ • > 108 electronic channels • 8x108 proton-proton collisions/sec • 2x10-4 Higgs per sec • 10 Petabytes of data a year • (10 Million GBytes = 14 Million CDs)

  21. ATLAS CAVERN (end 2004)

  22. UK GridPP

  23. GridPP1 Areas Grid Application Development LHC and US Experiments + Lattice QCD UK Tier-1/A Regional Centre Hardware and Manpower Management Travel etc LHC Computing Grid Project (LCG) Applications, Fabrics, Technology and Deployment European DataGrid (EDG) Middleware Development

  24. Application Development AliEn → ARDA BaBar GANGA Lattice QCD SAMGrid CMS

  25. Middleware Development Network Monitoring Configuration Management Grid Data Management Storage Interfaces Information Services Security

  26. International Collaboration • EU DataGrid (EDG) 2001-2004 • Middleware Development Project • LHC Computing Grid (LCG) • Grid Deployment Project for LHC • EU Enabling Grids for e-Science in Europe (EGEE) 2004-2006 • Grid Deployment Project for all disciplines • US and other Grid projects • Interoperability

  27. CERN computer centre Tier 0 Where data comes from Offline farm France RAL,UK USA Germany Italy Tier 1 National centres Online system Tier 2 Regional groups ScotGrid NorthGrid SouthGrid London Tier 3 Institutes Glasgow Edinburgh Durham Detector Tier 4 Workstations

  28. UK Tier-2 Centres ScotGrid Durham, Edinburgh, Glasgow NorthGrid Daresbury, Lancaster, Liverpool, Manchester, Sheffield SouthGrid Birmingham, Bristol, Cambridge, Oxford, RAL PPD, Warwick LondonGrid Brunel, Imperial, QMUL, RHUL, UCL Mostly funded by HEFCE

  29. UK Tier-1/A CentreRutherford Appleton Lab • High quality data services • National and International Role • UK focus for International Grid development • 700 Dual CPU • 80 TB Disk • 60 TB Tape (Capacity 1PB) Grid Operations Centre

  30. Prototype to Production Grid BaBarGrid BaBar EGEE SAMGrid CDF D0 ATLAS EDG LHCb ARDA GANGA LCG ALICE CMS LCG CERN Tier-0 Centre CERN Prototype Tier-0 Centre CERN Computer Centre UK Tier-1/A Centre UK Prototype Tier-1/A Centre RAL Computer Centre 4 UK Tier-2 Centres 19 UK Institutes 4 UK Prototype Tier-2 Centres Separate Experiments, Resources, Multiple Accounts Prototype Grids 'One' Production Grid 2004 2007 2001

  31. GridPP is active part of LCG 20 Sites 2,740 CPUs 67 Tbytes storage

  32. http://www.gridpp.ac.uk

  33. UK Core e-Science: Phase 2 Three major new activities: • National Grid Service and Grid Operation Support Centre • Open Middleware Infrastructure Institute for testing, software engineering and repository for UK middleware • Digital Curation Centre for R&D into long-term data preservation issues

  34. e-Science Institute Globus Alliance HPC(x) Grid Operations Centre Digital Curation Centre Open Middleware Infrastructure Institute The e-Science Centres CeSC (Cambridge)

  35. UK National Grid Service(NGS) • From April 2004, offers free access to two 128 processor compute nodes and two data nodes • Initial software is based on GT2 via VDT and LCG releases plus SRB and OGSA-DAI • Plan to move to Web Services based Grid middleware this summer

  36. 1280 x CPU AIX 512 x CPU Irix HPC(x) 20 x CPU 18TB Disk Linux 64 x CPU 4TB Disk Linux The e-ScienceNational Grid Service 2* 2* CeSC (Cambridge)

  37. VRE, VLE, IE HPCx + HECtoR LHC ISIS TS2 UK e-Infrastructure Users get common access, tools, information, Nationally supported services, through NGS Regional and Campus grids Integrated internationally

  38. Summary • Use common tools and support where possible for GRIDPP and NGS • Strong industrial component in UK e-science • Research Council support with identifiable e-Science funding after 2006 is very important • Integrate e-science infrastructure and posts into the computing services at Universities Acknowledgements Thanks to previous speakers on this topic for their slides and to our hosts for their invitation to this interesting meeting

More Related