70 likes | 190 Vues
This summary outlines the ExScal team's meeting in June 2004 focusing on deployment challenges and automation solutions. Key concerns included the efficiency of battery use, reducing deployment times, ensuring robust operations, and validating correct placements of XSMs and Stargates. Metrics for operational effectiveness and performance were discussed in relation to various intrusion scenarios. The document highlights the importance of dry runs to enhance deployment processes ahead of the December demonstration and outlines experimental proposals to improve localization, synchronization, and routing capabilities.
E N D
ExScal Breakout Session Summary ExScal Team June 2004
Deployment • numbers (>5 tons of batteries, days of recharge time, etc) • emplacement (> 8 hrs for XSM, Stargates, BaseStation) we seek ways to automate -- reduce these numbers • concerns: • how can we make deployment robust? (planning for slack, breaks, mistakes, accidents) • how can we confirm/validate correct deployment? (signals that XSMs are on, are properly emplaced, & Stargates in network); • how can we estimate power reserves, monitor units degrading during deployment? • dry run will improve deployment processes before December demonstration date
Metrics • Operational: deployment time/effort, cleanup, maintenance cost, ability to activate/sleep-control • Effectiveness: PFA metrics, response time, ability to handle different types of intrusion, visualization, coverage • Performance: routing metrics, synchronization, power consumption metrics, sensing accuracy, localization accuracy, time to reprogram Concerns: do operational/effectiveness metrics relate well enough to application/scenarios? Does software design adequately log/report for desired metrics?
Experiments • The main show: grid, intrusion scenarios (persons entering from beach, vehicles entering from beach, 10k run, etc) • The process for adding experiments to follow demo: • Submit “Experimentation/Demonstration Proposal” (no later than 9/15) to DARPA and OSU for approval. Should include: • Purpose and Objectives • Development Schedule • Integration and Test Plan • Experimentation/Demonstration Requirements (e.g., special equipment, support personnel) • DARPA and OSU will coordinate to schedule experiments. • OSU will make available API for invoking ExScal services
Suggestions for Experiments • MIT: goal: observations leading to {localization,synchronization,calibration,…} • MITRE: Measuring acoustics of aircraft • VU: Long-path routing without Stargates • PARC: local, mobile queries by laptop • UCB: RSSI; SNMS; multiple object tracking (grouping, ungrouping of intruders); routing experiments • ARL: Try different sensor patterns (emulate real dispersion) • Degradation testing (saturating bandwidth, probing for weaknesses in architecture, stress testing)
Packaging • XSM package concerns (protection from rain, ability to replace batteries, acoustic & PIR requirements) • Boxing/packaging for deployment 41 sensors in 1ft cube box, 20lbs. stargate batteries, antennas, GPS tools for placement
Timeline Concerns • would like to set up invitations by 8/31, but XSMs not arriving until end August • About 15 XSMv2s in [July 15 – Aug 30] period: can we learn enough? • 1000 node test 10/30/04 --- weather, terrain