1 / 41

Privacy Protection for Life-log Video

Privacy Protection for Life-log Video. Jayashri Chaudhari November 27, 2007 Department of Electrical and Computer Engineering University of Kentucky, Lexington, KY 40507. Outline. Motivation and Background Proposed Life-Log System Privacy Protection Methodology Face Detection and Blocking

rico
Télécharger la présentation

Privacy Protection for Life-log Video

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Privacy Protection for Life-log Video Jayashri Chaudhari November 27, 2007 Department of Electrical and Computer Engineering University of Kentucky, Lexington, KY 40507

  2. Outline • Motivation and Background • Proposed Life-Log System • Privacy Protection Methodology • Face Detection and Blocking • Voice Segmentation and Distortion • Experimental Results • Segmentation Algorithm Analysis • Audio Distortion Analysis • Conclusions

  3. What is a Life-Log System? “A system that records everything, at every moment and everywhere you go” • Applications include • Law enforcement • Police Questioning • Tourism • Medical Questioning • Journalism • Existing Systems/work • “MyLifeBits Project”: At Microsoft Research • “WearCam” Project: At University of Toronto, Steve Mann • “Cylon Systems”: http::/cylonsystems.com at UK (a portable body worn surveillance system)

  4. Technical Challenges • Security and Privacy • Information management and storage • Information Retrieval • Knowledge Discovery • Human Computer Interface

  5. Technical Challenges • Security and Privacy • Information management and storage • Information Retrieval • Knowledge Discovery • Human Computer Interface

  6. Why Privacy Protection? • Privacy is fundamental right of every citizen • Emerging technologies threaten privacy right • There are no clear and uniform rules and regulations regarding video recording • People are resistant toward technologies like life-log • Without tackling these issues the deployment of such emerging technologies is impossible

  7. Research Contributions • Practical audio-visual privacy protection scheme for life-log systems • Performance measurement (audio) on • Privacy protection • Usability

  8. Proposed Life-log System “A system that protects theaudiovisual privacyof thepersonscaptured by aportable video recording device”

  9. × Usefulness √ Privacy √ Usefulness × Privacy √ Usefulness √ Privacy Privacy Protection Scheme • Design Objectives • Privacy • Hide the identity of the subjects being captured • Privacy verses usefulness: • Recording should convey sufficient information to be useful

  10. Design Objectives • Anonymity or Ambiguity • The scheme should generate ambiguous identity of the recorded subjects. • Every individual will look and sound identical • Reduce correlation attacks • Speed • Protection scheme should work in real time • Interview Scenario • Producer is speaking with a single subject in relative quiet room

  11. Audio Segmentation Audio Distortion audio video Synchronization & Multiplexing storage Face Detection and Blocking Privacy Protection Scheme Overview S P S: Subject (The person who is being recorded) P: Producer (The person who is the user of the system)

  12. Voice Segmentation and distortion Windowed Power, Pk Computation Pk <TS Pk <TP Y Y Statek=Statek-1 orSubjectorProducer Statek=Producer Storage Statek=Subject Pitch Shifting We use the PitchSOLA time-domain pitch shifting method. * “DAFX: Digital Audio Effects” by Udo Zölzer et al.

  13. Input Audio X1(n) X2(n) X2(n) X2(n) N Sa α*Sa Mixing Max correlation Reduce discontinuity in phase and pitch Km Pitch Shifting Algorithm Pitch Shifting (Synchronous Overlap and Add): Steps 1) Time Stretching by a factor of α using window of size N and stepsize Sa Step 2) Re-sampling by a factor of 1/α to change pitch

  14. camera Face detection is based on Viola & Jones 2001. Face Detection Face Tracking Subject Selection Audio segmentation results Selective Blocking Producer talking Subject talking Face Detection and Blocking

  15. Initial Experiments1 • Analysis of Segmentation algorithm • Analysis of Audio distortion algorithm • 1) Accuracy in hiding identity • 2) Usability after distortion 1: Chaudhari J., S.-C. Cheung, and M. V. Venkatesh. Privacy protection for life-log video. In IEEE Signal Processing Society SAFE 2007: Workshop on Signal Processing Applications for Public Security and Forensics, 2007.

  16. S P P S P S P Silence Transitions  S: Subject Speaking P: Producer Speaking Segmentation Experiment • Experimental Data: • Interview Scenario in quiet meeting room • Three interviews recording of about 1 minute and 30 seconds long

  17. Segmentation Results

  18. Comparison With CMU Segmentation Algorithm CMU audio segmentation algorithm1 used as benchmark 1:Matthew A. Seigler, Uday Jain, Bhiksha Raj, and Richard M. Stern. Automatic segmentation, classification and clustering of broadcast news audio. In Proceedings of the Ninth Spoken Language Systems Technology Workshop, Harriman, New York, 1997.

  19. Speaker Identification Experiment • Experimental Data • 11 Test subjects, 2 voice samples from each subject • One voice sample is used as training and the other is used for testing • Public domain speaker recognition software Script1 This script is used for training the speaker recognition software Train Script2 This script is used to test the performance of audio distortion in hiding the identity Test

  20. Speaker Identification Results Distortion 1: (N=2048, Sa=256, α =1.5) Distortion 2: (N=2048, Sa=300, α =1.1) Distortion 3: (N=1024, Sa=128, α =1.5)

  21. Usability Experiments • Experimental Data • 8 subjects, 2 voice samples from each subject • One voice sample is used without distortion and the other is distorted • Manual transcription (5 human tester) Manual Transcription 1.Wav (transcription1) This transcription is of undistorted voice --- stored in one dot wav file. 2.Wav (transcription2) This transcription is of distorted voice sample --- in two dot wav ---. Unrecognized words

  22. Usability after distortion Word Error Rate: Standard measure of word recognition error for speech recognition system WER= (S+D+I) /N S = # substitution D = # deletion I = # insertion N = # words in reference sample Tool used: NIST tool SCLITE

  23. Extended Experiments • Data set • TIMIT (Texas Instruments and Massachusetts Institute of Technology) Speech Corpora • Experimental Setup • Allowable range of alpha (α): 0.2-2.0 • Five alpha values (α=0.5,0.75,1,1.25,1.40) • Increase the scope of experiments • “Subjective Experiments”: Use testers to access privacy and usability • Privacy Experiments (Speaker Identification)

  24. TIMIT Corpora (630 speakers, 10 audio clips per speaker) Our Experiments (30 speakers, 5 audio clips per speaker) Set E (α=1.40) Set D (α=1.25) Set B (α=0.5) Set A (α=1) Set C (α=0.75) Experimental Setup • Total 30 audio clips in each set • Re-divide the audio clips from each sets into five groups (1-5) • Each group consists of 6 audio clips randomly selected from each set • Each group was assigned to three testers and were asked to do 3 tasks

  25. Subjective Experiments • Task 1: Transcribe audio clips in the assigned group. • Purpose: Determine usability of the recording after distortion • Results: • Metric: WER for each transcription by the tester • Average WER for each clip from 3 testers WER for Speaker with the given alpha(α) value

  26. Average WER per speaker for each alpha value (0-30) (0-60) (0-35) (0-35)

  27. 100 22.4 15.3 14.2 14.4 D E B C A Sets Average WER per Set

  28. Statistical Analysis • Z-test calculations • Null Hypothesis: The average WER does not change (from Set A (before distortion) ) after the distortion for a given value of pitch scaling parameter (alpha) • H0: p1 = p2 (Null Hypothesis) Ha: p1 != p2 Z-Test parameters Z-Test Results

  29. Subjective Experiments • Task 2: Identify the number of distinct voices in each subset in the assigned group. • Purpose: Estimate ambiguity created by pitch shifting Results:

  30. Subjective Experiments • Task 3: For each clip from subset of Set A (which is the original un-distorted speech set); identify a clip in other subsets in which the same speaker may be speaking • Purpose: Qualitatively measure the assurance of Privacy Protection achieved by distortion • Results: None of the speakers from set A was identified from other distorted sets. (100% Recognition Error Rate)

  31. Privacy Experiments • Speaker Identification Experiments • ASR tools (LIA_Spk-Det and ALIZE)1 by LIA lab at the University of Avignon • Speaker Verification Tool • GMM-UBM (Gaussian Mixture Model-Universal Background Model) • Single Speaker Independent Background Model • Decision: Likelihood Ratio: 1: Bonastre, J.-F., Wild, F., Alize: a free, open tool for speaker recognition, http://www.lia.univ-avignon.fr/heberges/ALIZE/

  32. Front Processing Feature Extraction (SPRO Tool) Silence Frame Removal (EnergyDetector) Parameter Normalization (NormFeat) 32 coefficients = 16 LFCC + 16 derivative coefficients (SPRO4) Feature Vectors World Modeling Target Speaker Modeling 2 GMM (2048 components) 1: Male 2:Female Warping Initialization Training Adapts a World Model Bayesian Adaptation (MAP) (TrainTarget) (TrainWorld) Speaker Detection (ComputeTest) LIA_RAL Speaker-Det

  33. Experimental Setup • World Model • Number of male speakers     = 325 • Number of female speakers   = 135 • Target Speaker Model • Number of male test clips   =  20  • Number of female test clips =  10 • Two sets of experiments • Same Model: • World Model and Individual Speaker Models: (Training Set: distorted speech with the corresponding alpha) • Cross Model: • World Model and Individual Speaker Models: (Training Set: un-distorted speech)

  34. Privacy Results • Numbers in table is Average rank for the true speakers of the test clips for the corresponding alpha value • Conclusions • Cross Model: Distorted speech, no matter what alpha value is used, is very different from the original speech. • Same Model: Set B and Set C do not provide adequate protection as the rank is still very near the top.

  35. Example Video

  36. Conclusions • Proposed Real time implementation of voice-distortion and face blocking for privacy protection in Life-log video • Analysis of Audio Segmentation • Analysis of Audio Distortion for usability • Analysis of Audio Distortion for privacy protection

  37. Acknowledgment • Prof. Samson Cheung • People at Center of Visualization and Virtual Environment • Prof. Donohue and Prof. Zhang Thank you!

  38. Voice Distortion • Voice Identity • Vocal Track (Formats) : Filters • Vocal Chord (Pitch): Excitation Source • Different ways to distort audio: • Random mixture • Makes the recording useless • Voice Transformation • For example, • More Complex, not suitable for real-time applications • Pitch-shifting • Changes the pitch of voice • Keeps the recording useful • PitchSOLA time-domain pitch shifting method. * “DAFX: Digital Audio Effects” by U. Z. et al. • Simple with less complexity

  39. Cross Model: • World Model and Individual Speaker Models: (Training Set: un-distorted speech) • Same Model • World Model and Individual Speaker Models: (Training Set: distorted speech)

More Related