1 / 15

Virtual Imaging Peripheral for Enhanced Reality

Virtual Imaging Peripheral for Enhanced Reality. Aaron Garrett, Ryan Hannah, Justin Huffaker , Brendon McCool. Project Overview.

emmy
Télécharger la présentation

Virtual Imaging Peripheral for Enhanced Reality

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Virtual Imaging Peripheral for Enhanced Reality Aaron Garrett, Ryan Hannah, Justin Huffaker, Brendon McCool

  2. Project Overview Our project, code named Virtual Imaging Peripheral for Enhanced Reality or VIPER, is an augmented/virtual reality system. It will track a handheld unit’s location and perspective and use this information to find the location of a camera position in a virtual environment. Through a LCD screen on the handheld unit the user will see the virtual environment at the cameras location as if the handheld unit was a window into the virtual world. As the user moves the handheld unit around a table top sized environment the handheld unit’s actual and virtual perspective changes, allowing for different viewing angles of the virtual space.

  3. Project-Specific Success Criteria • An ability to communicate time stamp data using RF between the base unit and handheld unit. • An ability to display images to the LCD display. • An ability to estimate the angle and position of the handheld unit with respect to an origin point using accelerometer, gyroscope, compass, visual data, and ultrasonic data. • An ability to find angle displacement of the handheld unit’s front face relative to the IR beacon origin using mounted camera. • An ability to find distance from base to handheld unit using ultrasonic emitter and receiver.

  4. Block Diagram

  5. Reliability and Safety Analysis • Definition of Criticality Levels • Components chosen for analysis • FMECA Chart

  6. Criticality Levels • High Criticality: Personal Injury to User • Overheating may cause burn injuries • Low Criticality: No Immediate Threat to User • Loss of accurate video input from camera • Loss of accurate ultrasonic data • Loss of communication between beacon and head unit • Incorrect or loss of display functionality • No power

  7. Components Chosen • Atmel AT91SAM9XE256 • PIC32MX534F064 • Beagleboard (Arm Cortex A8) • 12v reg (AP1509)

  8. Atmel AT91SAM9XE256 λp = (C1 * πT + C2 * πE) * πQ* πL

  9. PIC32MX534F064 λp = (C1 * πT + C2 * πE) * πQ* πL

  10. Beagleboard (Arm Cortex A8) λp = (C1 * πT + C2 * πE) * πQ* πL

  11. 12v reg (AP1509) λp = (C1 * πT + C2 * πE) * πQ* πL

  12. FMECA

  13. FMECA

  14. FMECA

  15. FMECA

More Related