1 / 8

‘Rack Status‘

‘Rack Status‘. Jean-Pierre Vanuxem, PH-AIS. Rack Status (1). Layout and Allocation: known to ~ 100% ! 250 racks to be used + 52 spare locations documented in: EDMS ALICE-SER-SPC-021, v.3 Power: known to ~ 100 % ! type: 1-phase / 3-phase and standard power / UPS

ayla
Télécharger la présentation

‘Rack Status‘

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. ‘Rack Status‘ Jean-Pierre Vanuxem, PH-AIS JPV, TB of Nov.15th, 2005

  2. Rack Status (1) • Layout and Allocation: • known to ~ 100% ! • 250 racks to be used + 52 spare locations • documented in: EDMS ALICE-SER-SPC-021, v.3 • Power: • known to ~ 100% ! • type: 1-phase / 3-phase and standard power / UPS • documented in: EDMS ALICE-SER-SPC-022, v.3 • Contents: • known to ~ 80% at sub-rack level • documented under ‘Rack contents’ at: http://aliceinfo.cern.ch/Collaboration/Technical_Coordination/ElectronicsRacks.html • Help from all sub-detectors is still required to complete this task ! JPV, TB of Nov.15th, 2005

  3. Rack Status (2) • Purchase of Infrastructure Components: • Standard: turbine units + air deflectors, heat exchangers, ventilation units, water collectors, brackets, adaptors, fittings, etc. • Non-standard: cooling doors for PC racks (CR1-CR3), internal power distribution boxes (UX25) • Installation: • Cooling: TS-CV • ‘Hazemeyer’ switchboards + ZTBF32 modules: TS-EL • Rack enclosures: ALICE Integration Team (AIT) • Primary power to ‘Hazemeyer’ + control: TS-EL • Secondary power to racks + control: TS-EL + DCS team • Rack infrastructure components: AIT JPV, TB of Nov.15th, 2005

  4. Guidelines to maintain costs low • Recuperate existing LEP equipment • CV (L3): water cooling for racks + air conditioning systems in CR1-CR4 • Mechanics (L3, ALEPH): Rack enclosures + Other Rack components (back doors, brackets, etc.) • Electricity (L3): • UPS units (80 kVA + 60 kVA) • Hazemeyer Switchboards with ZTBF32 modules • Make use of COTS (if possible from CERN stores) • CV: turbines, ventilation units • Mechanics: • rack enclosures  new ‘LHC racks’ 52U/41U, 1000mm deep • VME crates (6U/9U) • Electricity: VME-PS, LVPS, HVPS, Cables, Connectors, • Group Orders with other LHC experiments • cooling doors for PC racks (together with LHCb) • Organize maintenance locally at CERN • define the role of PH-ESS Electronics Pool JPV, TB of Nov.15th, 2005

  5. Installation details (1) • Civil Engineering (TS) • All done, except false floors in C and D areas which are being built • General Cooling and Ventilation (TS-CV) • All done, except in C and D • General Power (TS-EL) • Mostly done in CR1-CR5 + A + B • Rack Placement (AIT) • All racks in CR1-CR5 + A + B are installed • Basic Rack Equipment (AIT) • All 250 racks will need to be equipped with infrastructure components during 1st semester 2006 JPV, TB of Nov.15th, 2005

  6. JPV, TB of Nov.15th, 2005

  7. Installation details (2) • Individual Rack Powering (TS-EL + AIT) • A few racks (5 ?) are already powered for tests in CR1+ CR3 • PX24 (CR1-CR5) : all racks will be powered in April 2006 • UX25 (A + B + C + D): May 2006 • Power Control (TS-EL) • Done by ‘ENS’ at the level of Hazemeyer ZTBF32 modules • PLC of type ‘Premium’ (from Schneider) • DCS as ‘client’ to ENS ‘server’ • Entire system ready in July 2006 JPV, TB of Nov.15th, 2005

  8. And now follows a pictorial trip, around and inside the ALICE racks: ‘ALICE in Rackland’ . . . JPV, TB of Nov.15th, 2005

More Related