1 / 74

Networks and Grids for Global Science

Networks and Grids for Global Science. Harvey B. Newman California Institute of Technology 3 rd International Data Grid Workshop Daegu, Korea August 26, 2004. Challenges for Global HENP Experiments. BaBar/D0 Example - 2004 500+ Physicists 100+ Institutes 35+ Countries. .

tress
Télécharger la présentation

Networks and Grids for Global Science

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Networks and Grids for Global Science Harvey B. Newman California Institute of Technology3rd International Data Grid WorkshopDaegu, Korea August 26, 2004

  2. Challenges for Global HENP Experiments BaBar/D0 Example - 2004 500+ Physicists 100+ Institutes 35+ Countries  LHC Example- 2007 5000+ Physicists 250+ Institutes 60+ Countries  Major Challenges (Shared with Other Fields) • Worldwide Communication and Collaboration • Managing Globally Distributed Computing & Data Resources, for Cooperative Software Development and Analysis

  3. Large Hadron Collider (LHC) CERN, Geneva: 2007 Start • pp s =14 TeV L=1034 cm-2 s-1 • 27 km Tunnel in Switzerland & France CMS TOTEM 5000+ Physicists 250+ Institutes 60+ Countries First Beams: Summer 2007 Physics Runs: from Fall 2007 ALICE : HI LHCb: B-physics Atlas Higgs, SUSY, QG Plasma, CP Violation, … the Unexpected

  4. Tier2 Center Tier2 Center Tier2 Center Tier2 Center Tier2 Center LHC Data Grid Hierarchy:Developed at Caltech CERN/Outside Resource Ratio ~1:2Tier0/( Tier1)/( Tier2) ~1:1:1 ~PByte/sec ~100-1500 MBytes/sec Online System Experiment CERN Center PBs of Disk; Tape Robot Tier 0 +1 Tier 1 10 - 40 Gbps FNAL Center IN2P3 Center INFN Center RAL Center ~10 Gbps Tier 2 ~1-10 Gbps Tier 3 Institute Institute Institute Institute Tens of Petabytes by 2007-8.An Exabyte ~5-7 Years later. Physics data cache 1 to 10 Gbps Tier 4 Workstations Emerging Vision: A Richly Structured, Global Dynamic System

  5. ICFA and Global Networks for Collaborative Science • National and International Networks, with sufficient (rapidly increasing) capacity and seamless end-to-end capability, are essential for • The daily conduct of collaborative work in both experiment and theory • Experiment development & construction on a global scale • Grid systems supporting analysis involving physicists in all world regions • The conception, design and implementation of next generation facilities as “global networks” • “Collaborations on this scale would never have been attempted, if they could not rely on excellent networks”

  6. Challenges of Next Generation Science in the Information Age • Flagship Applications • High Energy & Nuclear Physics, AstroPhysics Sky Surveys: TByte to PByte “block” transfers at 1-10+ Gbps • Fusion Energy: Time Critical Burst-Data Distribution; Distributed Plasma Simulations, Visualization, Analysis • eVLBI: Many real time data streams at 1-10 Gbps • BioInformatics, Clinical Imaging: GByte images on demand • Provide results with rapid turnaround, over networks of varying capability in different world regions • Advanced integrated applications, such as Data Grids, rely on seamless operation of our LANs and WANs • With reliable, quantifiable high performance Petabytes of complex data explored and analyzed by 1000s of globally dispersed scientists, in hundreds of teams

  7. Int’l Networks BW on Major Links for HENP: US-CERN Example • Rate of Progress >> Moore’s Law (US-CERN Example) • 9.6 kbps Analog (1985) • 64-256 kbps Digital (1989 - 1994) [X 7 – 27] • 1.5 Mbps Shared (1990-3; IBM) [X 160] • 2 -4 Mbps (1996-1998) [X 200-400] • 12-20 Mbps (1999-2000) [X 1.2k-2k] • 155-310 Mbps (2001-2) [X 16k – 32k] • 622 Mbps (2002-3) [X 65k] • 2.5 Gbps  (2003-4) [X 250k] • 10 Gbps  (2005) [X 1M] • 4x10 Gbps or 40 Gbps (2007-8) [X 4M] • A factor of ~1M Bandwidth Improvement over 1985-2005 (a factor of ~5k during 1995-2005) • A prime enabler of major HENP programs • HENP has become a leading applications driver, and also a co-developer of global networks

  8. History of Bandwidth Usage – One Large Network; One Large Research Site ESnet Accepted Traffic 1/90 – 1/04Exponential Growth Since ’92;Annual Rate Increased from 1.7 to 2.0X Per Year In the Last 5 Years SLAC Traffic ~400 Mbps; Growth in Steps (ESNet Limit): ~ 10X/4 Years.Projected: ~2 Terabits/s by ~2014

  9. Internet Growth in the World At Large Amsterdam Internet Exchange Point Example 5 MinuteMax 30 Gbps 20 Gbps Average 11.08.04 Some Annual Growth Spurts;Typically In Summer-Fall The Rate of HENP Network Usage Growth (~100% Per Year) is Similar to the World at Large

  10. Internet 2 Land Speed Record (LSR) • Judged on product of transfer speed and distance end-to-end, using standard (TCP/IP) protocols. • Across Production Net: Abilene • IPv6 record: 4.0 Gbps between Geneva and Phoenix (SC2003) • IPv4 Multi-stream record with Windows & Linux: 6.6 Gbps between Caltech and CERN (16 kkm; “Grand Tour d’Abilene”) June 2004 • Exceeded 100 Petabit-m/sec • Single Stream 7.5 Gbps X 16 kkm with Linux Achieved in July • Concentrate now on reliable Terabyte-scale file transfers • Note System Issues: CPU, PCI-XBus, NIC, I/O Controllers, Drivers LSR History – IPv4 single stream Petabitmeter (10^15 bit*meter) • Monitoring of the Abilene traffic in LA: June 2004 Record Network http://www.guinnessworldrecords.com/

  11. HENP Bandwidth Roadmap for Major Links (in Gbps) Continuing Trend: ~1000 Times Bandwidth Growth Per Decade;Keeping Pace with Network BW Usage (ESNet, SURFNet etc.)

  12. Evolving Quantitative Science Requirements for Networks (DOE High Perf. Network Workshop)

  13. National Lambda Rail (NLR) SEA POR SAC BOS NYC CHI OGD DEN SVL CLE WDC PIT FRE KAN RAL NAS STR LAX PHO WAL ATL SDG OLG DAL JAC 15808 Terminal, Regen or OADM site Fiber route Transition beginning now to optical, multi-wavelength Community owned or leased “dark fiber” networks for R&E • NLR • Coming Up Now • Initially 4 10G Wavelengths • Northern Route LA-JAX by 4Q04 • Internet2 HOPI Initiative (w/HEP) • To 40 10G Waves in Future • nl, ca, pl, cz, uk, ko, jp • 18 US States

  14. [Legends ] <10G>  ・Ishikawa Hi-tech Exchange Center (Tatsunokuchi-machi, Ishikawa Prefecture) <100M>  ・Toyama Institute of Information Systems (Toyama)  ・Fukui Prefecture Data Super Highway AP * (Fukui) 20Gbps 10Gbps 1Gbps Optical testbeds Access points <100M>  ・Hokkaido Regional Network Association AP * (Sapporo) Core network nodes <1G>  ・Teleport Okayama (Okayama)  ・Hiroshima University (Higashi Hiroshima) <100M>  ・Tottori University of Environmental Studies (Tottori)  ・Techno Ark Shimane (Matsue)  ・New Media Plaza Yamaguchi (Yamaguchi) <1G>  ・Tohoku University (Sendai)  ・NICT Iwate IT Open Laboratory (Takizawa-mura, Iwate Prefecture) <100M>  ・Hachinohe Institute of Technology (Hachinohe, Aomori Prefecture)  ・Akita Regional IX *(Akita)  ・Keio University Tsuruoka Campus (Tsuruoka, Yamagata Prefecture)  ・Aizu University (Aizu Wakamatsu) <10G>  ・Kyoto University (Kyoto)  ・Osaka University (Ibaraki) <1G>  ・NICT Kansai Advanced Research Center (Kobe) <100M>  ・Lake Biwa Data Highway AP * (Ohtsu)  ・Nara Prefectural Institute of Industrial Technology (Nara)  ・Wakayama University (Wakayama)  ・Hyogo Prefecture Nishiharima Technopolis (Kamigori-cho, Hyogo Prefecture) Sapporo <100M>  ・Niigata University (Niigata)  ・Matsumoto Information Creation Center (Matsumoto, Nagano Prefecture) <10G>  ・Kyushu University (Fukuoka) <100M>  ・NetCom Saga (Saga)  ・Nagasaki University (Nagasaki)  ・Kumamoto Prefectural Office (Kumamoto)  ・Toyonokuni Hyper Network AP *(Oita)  ・Miyazaki University (Miyazaki)  ・Kagoshima University (Kagoshima) <10G>  ・Tokyo University (Bunkyo Ward, Tokyo)  ・NICT Kashima Space Research Center (Kashima, Ibaraki Prefecture) <1G>  ・Yokosuka Telecom Research Park (Yokosuka, Kanagawa Prefecture) <100M>  ・Utsunomiya University (Utsunomiya)  ・Gunma Industrial Technology Center (Maebashi)  ・Reitaku University (Kashiwa, Chiba Prefecture)  ・NICT Honjo Information and Communications Open Laboratory (Honjo, Saitama Prefecture)  ・Yamanashi Prefecture Open R&D Center (Nakakoma-gun, Yamanashi Prefecture) Fukuoka Sendai Kanazawa NICT Kita Kyushu IT Open Laboratory Nagano NICT Koganei Headquarters Osaka Okayama Kochi NICT Tsukuba Research Center Nagoya NICT Keihannna Human Info-Communications Research Center Otemachi USA Okinawa <100M>  ・Kagawa Prefecture Industry Promotion Center (Takamatsu)  ・Tokushima University (Tokushima)  ・Ehime University (Matsuyama)  ・Kochi University of Technology (Tosayamada-cho, Kochi Prefecture) <100M>  ・Nagoya University (Nagoya)  ・University of Shizuoka (Shizuoka)  ・Softopia Japan (Ogaki, Gifu Prefecture)  ・Mie Prefectural College of Nursing (Tsu) *IX:Internet eXchange AP:Access Point JGN2: Japan Gigabit Network (4/04 – 3/08)20 Gbps Backbone, 6 Optical Cross-Connects

  15. GLIF: Global Lambda Integrated Facility “GLIF is a World Scale Lambda based Lab for Application and Middleware development, where Grid applications ride on dynamically configured networks based on optical wavelengths ... Coexisting with more traditional packet-switched network traffic 4th GLIF Workshop:Nottingham UK Sept. 2004 10 Gbps Wavelengths For R&E Network Development Are Prolifering, Across Continents and Oceans

  16. ICFA Standing Committee on Interregional Connectivity (SCIC) • Created by ICFA in July 1998 in Vancouver • CHARGE: Make recommendations to ICFA concerning the connectivity between the Americas, Asia and Europe • As part of the process of developing theserecommendations, the committee should • Monitor traffic • Keep track of technology developments • Periodically review forecasts of future bandwidth needs • Provide early warning of potential problems • Representatives: Major labs, ECFA, ACFA; North American and Latin American Physics Communities • Monitoring, Advanced Technologies, and Digital DivideWorking Groups Formed in 2002

  17. SCIC in 2003-2004http://cern.ch/icfa-scic Three 2004 Reports; Presented to ICFA in February • Main Report: “Networking for HENP” [H. Newman et al.] • Includes Brief Updates on Monitoring, the Digital Divide and Advanced Technologies [*] • A World Network Overview (with 27 Appendices): Status and Plans for the Next Few Years of National & Regional Networks, and Optical Network Initiatives • Monitoring Working Group Report [L. Cottrell] • Digital Divide in Russia [V. Ilyin]August 2004 Update Reports at the SCIC Web Site: See http://icfa-scic.web.cern.ch/ICFA-SCIC/documents.htm • Asia Pacific, Latin America, GLORIAD (US-Ru-Ko-China);Brazil, Korea, ESNet, etc.

  18. SCIC Main Conclusion for 2003Setting the Tone for 2004 • The disparity among regions in HENP could increase even more sharply, as we learn to use advanced networks effectively, and we develop dynamic Grid systems in the “most favored” regions • We must take action, and workto Close the Digital Divide • To make Physicists from All World Regions Full Partners in Their Experiments; and in the Process of Discovery • This is essential for the health of our global experimental collaborations, our plans for future projects, and our field.

  19. Cont’d SCIC Focus on the Digital Divide: Several Perspectives • Work on Policies and/or Pricing: pk, in, br, SE Europe, … • Find Ways to work with vendors, NRENs, and/or Gov’ts • Point to Model Cases: e.g. Poland, Slovakia, Czech Republic • Inter-Regional Projects • GLORIAD, Russia-China-US Optical Ring • Latin America: CHEPREO (US-Brazil); EU CLARA Project • Virtual SILK Highway Project (DESY): FSU satellite links • Workshops and Tutorials/Training Sessions • For Example: Digital Divide and HEPGrid Workshop,UERJ Rio, Feb. 2004; Next DD Workshop in Daegu May 2005 • Help with Modernizing the Infrastructure • Raise Technology Awareness; Help Commission, Develop • Provide Tools for Effective Use: Monitoring, Collaboration • Participate in Standards Development; Open Tools • Advanced TCP stacks; Grid systems

  20. ICFA Report: Networks for HENPGeneral Conclusions (2) • Reliable high End-to-end Performance of networked applications such as large file transfers and Data Grids is required. Achieving this requires: • A coherent approach to End-to-end monitoring extending to all regions that allows physicists throughout the world to extract clear information • Upgrading campus infrastructures.To support Gbps data transfers in most HEP centers. One reason for under-utilization of national and Int’l backbones, is the lack of bandwidth to end-user groups in the campus • Removing local, last mile, and nat’l and int’l bottlenecks end-to-end, whether technical or political in origin.While National and International backbones have reached 2.5 to 10 Gbps speeds in many countries, the bandwidths across borders, the countryside or the city may be much less. This problem is very widespread in our community, with examples stretching from the Asia Pacific to Latin America to the Northeastern U.S. Root causes for this vary, from lack of local infrastructure to unfavorable pricing policies.

  21. ICFA Report Update (8/2004): Main Trends Continue, Some Accelerate • Current generation of 2.5-10 Gbps network backbones and major Int’l links arrived in 2-3 Years [US+Europe+Japan; Now Korea and China] • Capability Grew 4 to 100s of Times; Much Faster than Moore’s Law • Proliferation of 10G links across the Atlantic Now • Direct result of Falling Network Prices: $ 0.5 – 1M Per Year for 10G • Ability to fully use long 10G paths with TCP continues to advance: 7.5 Gbps X 16kkm (August 2004) • Technological progress driving equipment costs in end-systems lower • “Commoditization” of Gbit Ethernet (GbE) ~complete: ($20-50 per port); 10 GbE commoditization underway • Move to owned or leased optical nets (us, ca, nl, sk, po, ko, jp) well underway in several areas of the world • Emergence of “Hybrid” Network Model: GNEW2004; UltraLight, GLIF • While there is progress in some less-advantaged regions, thegap between the technologically “rich” and “poor” is widening

  22. ICFA SCIC Monitoring WG (L. Cottrell)See www.slac.stanford.edu/grp/scs/net/talk03/icfa-aug04.ppt • Now monitoring 650 sites in 115 countries • In last 9 months: • Several sites in Russia (GLORIAD) • Many hosts in Africa (27 of 54 Countries) • Monitoring sites in Pakistan, Brazil C. Asia, Russia, SE Europe, L. America, M. East, China: 4-5 yrs behind India, Africa: 7 yrs behind PingERWorld View Seen from SLAC View from CERNConfirms This View Important for policy makers

  23. AmPath Research Networking in Latin America: August 2004 • The only Countries with research network connectivity now in Latin America: • Argentina, Brazil, Chile, Mexico, Venezuela • AmPathProvided connectivity for some South American countries • New CHEPREO Sao Paolo-Miami Link at 622 Mbps Starting This Month New: CLARA (Funded by EU) • Regional Network Connecting 19 Countries: Argentina Dominican Republic Panama Brasil Ecuador ParaguayBolivia El Salvador PeruChile Guatemala UruguayColombia Honduras VenezuelaCosta Rica Mexico NicaraguaCuba 155 Mbps Backbone with 10-45 Mbps Spurs;4 Mbps Satellite to Cuba; 622 Mbps to Europe Also NSF Proposals To Connect at 2.5G to US

  24. HEPGRID (CMS) in Brazil (Santoro et al.) 622 Mbps Italy France BRAZIL USA Germany HEPGRID-CMS/BRAZIL is a project to build a Grid that • At Regional Level will include CBPF,UFRJ,UFRGS,UFBA, UERJ & UNESP • At International Level will be integrated with CMS Grid based at CERN; focal points include iVGDL/Grid3 and bilateral projects with Caltech Group Brazilian HEPGRID On line systems T0 +T1 2.5 - 10 Gbps CERN T1 UNESP/USP SPRACE-Working T2 T1 UERJ Regional Tier2 Ctr Gigabit T3 T2 UFRGS UERJ: T2T1,100500 Nodes; Plus T2s to 100 Nodes CBPF UERJ UFBA UFRJ Individual Machines T4

  25. PROGRESS in SE Europe (Sk, Pl, Cz, Hu, …) 1660 km of Dark Fiber CWDM Links, up to 112 km. 1 to 4 Gbps (GbE) August 2002:First NREN in Europe to establish Int’l GbE Dark Fiber Link, to AustriaApril 2003 to Czech Republic. Planning 10 Gbps Backbone; dark fiber link to Poland this year.

  26. Dark Fiber in Eastern EuropePoland: PIONIER Network 2650 km Fiber Connects 16 MANs; to 5200 km and 21 MANs by 2005 Support • Computational Grids;Domain-Specific Grids • Digital Libraries • Interactive TV • Add’l Fibers for e-Regional Initiatives

  27. The Advantage of Dark Fiber CESNET Case Study (Czech Republic) Leased 1 x 2,5G Leased fibre with own equipment 1 x 2,5G (EURO/Month) (EURO/Month) about 150km (e.g. Ústí n.L. - Liberec) 7,000 * 5 000 2513 km Leased Fibers (Since 1999) about 300km (e.g. Praha - Brno) 8,000 ** 7 000 * 2 x booster 18dBm ** 2 x booster 27dBm + 2 x preamp + 6 x DCF Leased 4 x 2,5G Leased fibre with own equipment 4 x 2,5G (EURO/Month) (EURO/Month) Case Study ResultWavelength ServiceVs. Fiber Lease: Cost Savings of 50-70% Over 4 Yearsfor Long 2.5G or 10G Links about 150km (e.g. Ústí n.L. - Liberec) 14,000 * 8 000 about 300km (e.g. Praha - Brno) 23,000 ** 11 000 * 2 x booster 24dBm, DWDM 2,5G ** 2 x (booster+In-line+preamp), 6 x DCF, DWDM 2,5G Leased 1 x 10G Leased fibre with own equipment 1 x 10G (EURO/Month) (EURO/Month) about 150km (e.g. Ústí n.L. - Liberec) 14,000 * 5 000 about 300km (e.g. Praha - Brno) 16,000 ** 8 000 * 2 x booster 21dBm, 2 x DCF ** 2 x (booster 21dBm + in-line + preamplifier) + 6 x DCF Leased 4 x 10G Leased fibre with own equipment 4 x 10G (EURO/Month) (EURO/Month) about 150km (e.g. Ústí n.L. - Liberec) 29,000 * 12 000 about 300km (e.g. Praha - Brno) 47,000 ** 14 000 * 2 x booster 24dBm, 2 x DCF, DWDM 10G ** 2 x (booster +In-line+preampr), 6 x DCF, DWDM 10G

  28. Asia Pacific Academic Network Connectivity APAN Status July 2004 RU 200M 20.9G JP Europe 34M 2G KR 1.2G US 155M 310M Connectivity to US from JP, KO, AU is Advancing Rapidly.Progress in the Region, and to Europe is Much Slower 9.1G 622M CN • TW 777M 45M `722M 90M HK 2M 7.5M • IN 45M 155M 1.5M TH PH VN 155M 1.5M 932M (to 21 G) MY • LK 2M 12M SG ID 2.5M Access Point Exchange Point Current Status 2004 (plan) 16M AU Better North/South Linkages within Asia JP-SG link: 155Mbps in 2005 is proposed to NSF by CIREN JP- TH link: 2Mbps  45Mbps in 2004 is being studied. CIREN is studying an extension to India

  29.         

  30. Wireline Dial Up ISDN Broadband 23.4M 45.0M 4.9M 9.8M Internet in China (J.P.Wu APAN July 2004) • Internet users in China: from 6.8 Million to 78 Million within 6 months • Backbone:2.5-10G DWDM+Router • International links:20G • Exchange Points:> 30G(BJ,SH,GZ) • Last Miles • Ethernet,WLAN,ADSL,CTV,CDMA,ISDN,GPRS,Dial-up • IP Addresses: 32M(1A+233B+146C); Need IPv6

  31. China: CERNET Update • 1995, 64K Nationwide backbone connecting 8 cities, 100 Universities • 1998, 2M Nation wide backbone connecting 20 cities, 300 Universities • 2000, Own dark fiber crossing 30+ major cities and 30,000 kilometers • 2001, CERNET DWDM/SDH network finished • 2001, 2.5G/155M Backbone connecting 36 cities, 800 universities • 2003, 1300+ universities and institutes, over 15 Million Users

  32. CERNET2 and Key Technologies • CERNET 2: Next Generation Education and Research Network in China • CERNET 2 Backbone connecting 15-20 GigaPOPs at 2.5G-10Gbps (I2-like Model) • Connecting 200 Universities and 100+ Research Institutes at 1Gbps-10Gbps • Native IPv6 and Lambda Networking • Support/Deployment of: • E2E performance monitoring • Middleware and Advanced Applications • Multicast

  33. APAN-KR : KREONET/KREONet2 II • KREONET2 • Support for Next Gen. Apps: • IPv6, QoS, Multicast; Bandwidth Alloc. Services • StarLight/Abilene Connection • KREONET • 11 Regions, 12 POP Centers • Optical 2.5-10G Backbone; SONET/SDH, POS, ATM • National IX Connection • Int’l Links • US: 2 X 622 Mbps via CA*Net;GbE Link via TransPAC; 155 (to 10G) GLORIAD Link • Japan: 2 Gbps • TEIN to GEANT: 34 Mbps • Singapore (SingAREN): 8 Mbps • SuperSIREN (7 Res. Institutes) • Optical 10-40G Backbone • High Speed Wireless: 1.25 G • Collaborative Environment Support

  34. KR-US/CA Transpacific connection • Participation in Global-scale Lambda Networking • Two STM-4 circuits (1.2G) : KR-CA-US • Global lambda networking : North America, Europe, • Asia Pacific, etc. Global Lambda Networking KREONET/SuperSIReN CA*Net4 StarLight STM-4 * 2 Chicago APII-testbed/KREONet2 PacWave Seattle

  35. New Record: 916 Mbps from CHEP/KNU to Caltech (UDP KOREN-TransPAC-Caltech, 22/06/’04) Date: Tue, 22 Jun 2004 13:47:25 +0900 From: Kihwan Kwon To: Dongchul Son <son@knu.ac.kr> [root@sul Iperf]# ./iperf -c socrates.cacr.caltech.edu -u -b 1000m ------------------------------------------------------------ Client connecting to socrates.cacr.caltech.edu, UDP port 5001 Sending 1470 byte datagrams; UDP buffer size: 64.0 KByte ------------------------------------------------------------ [ 5] local 155.230.20.20 port 33036 connected with 131.215.144.227 [ ID] Interval Transfer Bandwidth [ 5] 0.0-2595.2 sec 277 GBytes 916 Mbits/sec USA TransPAC G/H-Japan KNU/Korea Max. 947.3Mbps

  36. OC3 circuits Moscow-Chicago-Beijing since January 2004 • OC3 circuit Moscow-Beijing July 2004 (completes the ring) • Korea (KISTI) joining US, Russia, China as full partner in GLORIAD • Plans for Central Asian extension, with Kyrgyz Gov’t • Rapid traffic growth with heaviest US use from DOE (FermiLab), NASA, NOAA, NIH and 260+ Univ. (UMD, IU, UCB, UNC, UMN… Many Others) Global Ring Network for Advanced Applications Development • Aug. 8 2004: P.K. Young, Korean IST Advisor to President Announces • Korea Joining GLORIAD • TEIN gradually to 10G, connected to GLORIAD • Asia Pacific Info. Infra- Structure (1G) will be backup net to GLORIAD > 5TBytes now transferred monthly via GLORIAD to US, Russia, China GLORIAD 5-year Proposal Pending (with US NSF) for expansion: 2.5G Moscow-Amsterdam-Chicago-Seattle-Hong Kong-Pusan-Beijing circuits early 2005; 10G ring around northern hemisphere 2007; multiple wavelength service 2009 – providing hybrid circuit-switched (primarily Ethernet) and routed services

  37. AFRICA: NectarNet Initiative W. MatthewsGeorgia Tech Growing Need to connect academic researchers, medicalresearchers & practitioners to many sites in Africa Examples: • CDC & NIH: Global AIDS Project, Dept. of Parasitic Diseases,Nat’l Library of Medicine (Ghana, Nigeria) • Gates $ 50M HIV/AIDS Center in Botswana; Project Coord at Harvard • Africa Monsoon AMMA Project, Dakar Site [cf. East US Hurricanes] • US Geological Survey: Global Spatial Data Infrastructure • Distance Learning: Emory-Ibadan (Nigeria); Research Channel But Africa is Hard: 11M Sq. Miles, 600 M People, 54 Countries • Little Telecommunications Infrastructure Approach:Use SAT-3/WASC Cable (to Portugal), GEANT Across Europe, AMS-NY Link Across Atlantic, Peer with Abilene in NYC • Cable Landings in 8 West African Countries and South Africa • Pragmatic approach to reach end points: VSAT,ADSL,microwave, etc. Note: World Conference on Physics and Sustainable Development,10/31 – 11/2/05 in Durban South Africa; Part of World Year of Physics 2005. Sponsors: UNESCO, ICTP, IUPAP, APS, SAIP

  38. Bandwidth prices in Africa vary dramatically; are in general many times what they could be if universities purchase in volume Sample Bandwidth Costs for African Universities Avg. Unit Cost is 40X US Avg.; Cost is Several Hundred Times, Compared to Leading Countries Sample size of 26 universitiesAverage Cost for VSAT service: Quality, CIR, Rx, Tx not distinguished Roy Steiner Internet2 Workshop

  39. HEP Active in the World Summit on the Information Society 2003-2005 • WSIS I (Geneva 12/03): CERN RSIS Event, SIS Forum, CERN/Caltech Online Stand • Visitors at WSIS I • Kofi Annan, UN Sec’y General • John H. Marburger, Science Adviser to US President • Ion Iliescu, President of Romania, … • Planning Now Underway for WSIS II: Tunis 2005 • GOAL: To Create an “Information Society”. Common Definition Adopted (Tokyo Declaration, January 2003):“… One in which highly developed ICT networks, equitable and ubiquitous access to information, appropriate content in accessible formats and effective communication can help people achieve their potential” • Kofi Annan Challenged the Scientific Community to Help (3/03)

  40. Role of Sciences in Information Society. Palexpo, Geneva 12/2003 • Demos at the CERN/Caltech RSIS Online Stand • Advanced network and Grid-enabled analysis • Monitoring very large scale Grid farms with MonALISA • World Scale multisite multi-protocol videoconference with VRVS (Europe-US-Asia-South America) • Distance diagnosis and surgery using Robots with “haptic” feedback (Geneva-Canada) • Music Grid: live performances with bands at St. John’s, Canada and the Music Conservatory of Geneva on stage VRVS 37k hosts 106 Countries 2-3X Growth/Year

  41. Grid3: An Operational Production Grid, Since October 2003 • 29 sites (U.S., Korea) • to ~3000 CPUs • to ~1200 Concurrent Jobs • Trillium: • PPDGGriPhyNiVDGL www.ivdgl.org/grid2003 • Also LCG and EGEE Project Korea • Prelude to Open Science Grid: www.opensciencegrid.org

  42. HENP Data Grids, and Now Services-Oriented Grids • The original Computational and Data Grid concepts are based on largely stateless, open systems • Analogous to the Web • The classical Grid architecture had a number of implicit assumptions • The ability to locate and schedule suitable resources, within a tolerably short time (i.e. resource richness) • Short transactions with relatively simple failure modes • HENP Grids are Data Intensive & Resource-Constrained • Resource usage governed by local and global policies • Long transactions; some long queues • Grid-Enabled Analysis: 1000s of users compete for resources at dozens of sites: Complex scheduling; management • HENP Stateful, End-to-end Monitored and Tracked Paradigm • Adopted in OGSA [Now WS Resource Framework]

  43. Managing Global Systems: Dynamic Scalable Services Architecture MonALISA: http://monalisa.cacr.caltech.edu 24 X 7 OperationsMultiple Orgs. • Grid2003 • US CMS • CMS-DC04 • ALICE • STAR • VRVS • ABILENE • GEANT • + GLORIAD • “Station Server” Services-engines at sites host many“Dynamic Services” • Scales to thousands of service-Instances • Servers autodiscover and interconnect dynamically to form a robust fabric • Autonomous agents + CLARENS: Web Services Fabric and Portal Architecture

  44. Grid-Enabled Analysis Environment Analysis Client Analysis Client Analysis Client HTTP, SOAP, XML/RPC Grid Services Web Server Scheduler Catalogs Fully- Abstract Planner Metadata Partially- Abstract Planner Virtual Data Applications Data Management Monitoring Replica Fully- Concrete Planner Grid Execution Priority Manager Grid Wide Execution Service CLARENS: Web Services Architecture • Analysis Clients talk standard protocols to the CLARENS “Grid Services Web Server”, with a simple Web service API • The secure Clarens portal hides the complexity of the Grid Services from the client • Key features: Global Scheduler, Catalogs, Monitoring, and Grid-wide Execution service;Clarens servers forma Global Peer to peer Network

  45. UltraLight Collaboration:http://ultralight.caltech.edu • Caltech, UF, FIU, UMich, SLAC,FNAL,CERN, UERJ(Rio), NLR, CENIC, UCAID,Translight, UKLight, Netherlight, UvA, UCLondon, KEK, Taiwan • Cisco, Level(3) • Integrated hybrid experimental network, leveraging Transatlantic R&D network partnerships; packet-switched + dynamic optical paths • 10 GbE across US and the Atlantic: NLR, DataTAG, TransLight, NetherLight, UKLight, etc.; Extensions to Japan, Taiwan, Brazil • End-to-end monitoring; Realtime tracking and optimization; Dynamic bandwidth provisioning • Agent-based services spanning all layers of the system, from the optical cross-connects to the applications.

  46. HEPGRID and Digital Divide Workshop UERJ, Rio de Janeiro, Feb. 16-20 2004 • Theme: Global Collaborations, Grids and Their Relationship to the Digital Divide • For the past three years the SCIC has focused on understanding and seeking the means of reducing or eliminating the Digital Divide. It proposed to ICFA that these issues, as they affect our field of High Energy Physics, be brought to our community for discussion. This led to ICFA’s approval, in July 2003, of the Digital Divide and HEP Grid Workshop.  • Review of R&E Networks; Major Grid Projects • Perspectives on Digital Divide Issues by Major HEP Experiments, Regional Representatives • Focus on Digital Divide Issues in Latin America; Relate to Problems in Other Regions • More Info: http://www.lishep.uerj.br NEWS:Bulletin: ONE TWOWELCOME BULLETIN General InformationRegistrationTravel InformationHotel Registration Participant List How to Get UERJ/HotelComputer Accounts Useful Phone Numbers Program Contact us: SecretariatChairmen • Tutorials • C++ • Grid Technologies • Grid-Enabled Analysis • Networks • Collaborative Systems Sessions & Tutorials Available(w/Video) on the Web SPONSORS CLAF  CNPQ FAPERJ        UERJ

  47. International ICFA Workshop on HEP Networking, Grids and Digital Divide Issues for Global e-Science Dates: May 23-27, 2005 Venue: Daegu, Korea Dongchul Son Center for High Energy Physics Kyungpook National University ICFA, Beijing, China Aug. 2004 Approved by ICFA August 20, 2004

  48. International ICFA Workshop on HEP Networking, Grids and Digital Divide Issues for Global e-Science • Themes • Networking, Grids, and Their Relationship to the Digital Divide for HEP as Global e-Science • Focus on Key Issues of Inter-regional Connectivity • Workshop Goals • Review the current status, progress and barriers to the effective use of the major national, continental and transoceanic networks used by HEP • Review progress, strengthen opportunities for collaboration, and explore the means to deal with key issues in Grid computing and Grid-enabled data analysis, for high energy physics and other fields of data intensive science, now and in the future • Exchange information and ideas, and formulate plans to develop solutions to specific problems related to the Digital Divide in various regions, with a focus on Asia Pacific, as well as Latin America, Russia and Africa • Continue to advance a broad program of work on reducing or eliminating the Digital Divide, and ensuring global collaboration, as related to all of the above aspects.

  49. Networks and Grids for HENP and Global Science • Network backbones and major links used by major experiments in HENP and other fields are advancing rapidly • To the 10 G range in < 3 years; much faster than Moore’s Law • New HENP and DOE Roadmaps: a factor ~1000 improvement per decade • Important advances in Asia-Pacific, notably Korea, Japan, China • We are learning to use long distance 10 Gbps networks effectively • 2004 Developments: to 7.5 Gbps flows with TCP over 16 kkm • A transition to community-owned and operated R&E networks (us, ca, nl, pl, cz, sk, co, jp …); A new generation of “hybird” optical networks is emerging • We Must Work to Close to Digital Divide • To Allow Scientists and Students from All World Regions to Take Part in Discoveries at the Frontiers of Science • Removing Regional, Last Mile, Local Bottlenecks and Compromises in Network Quality are now On the Critical Path • Important Examples on the Road to Progress in Closing the Digital Divide • CLARA, CHEPREO, and Brazil HEP Grid in Latin America • Optical Networking in Central and Southeast Europe • APAN Links in the Asia Pacific • Leadership and Outreach: HEP Groups in Korea, Japan, US and Europe

More Related