1 / 36

“Taking Signals to Bits” Sparse Representations for Coding, Sampling and Source Separation

“Taking Signals to Bits” Sparse Representations for Coding, Sampling and Source Separation. Mike Davies University of Edinburgh. with thanks to my family, students & colleagues. What are signals?. A signal is a time (or space) varying quantity that can carry information.

jayme
Télécharger la présentation

“Taking Signals to Bits” Sparse Representations for Coding, Sampling and Source Separation

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. “Taking Signals to Bits”Sparse Representations for Coding, Sampling and Source Separation Mike Davies University of Edinburgh with thanks to my family, students & colleagues

  2. What are signals? A signal is a time (or space) varying quantity that can carry information. The concept is broad, and hard to define precisely. (Wikipedia)

  3. … in medical physics … in telecommunications … in audio Signals are everywhere! … in image/video

  4. Part I: Sparse Representations and Coding

  5. Joseph Fourier What are signals made of? The Frequency viewpoint (Fourier): Signals can be built from the sum of harmonic functions (sine waves) signal Harmonic functions Fourier coefficients

  6. Sampling and the digital revolution • Today we are more familiar with discrete signals (e.g. audio files, digital images). This is thanks to: • Whittaker–Kotelnikov–Shannon Sampling Theorem: • “Exact reconstruction of a continuous-time signal from discrete samples is possible if the signal is bandlimited and the sampling frequency is greater than twice the signal bandwidth.” Sampling below this rate introduces aliasing

  7. frequency time Audio representations Which representation is best: time or frequency?

  8. Frequency (Hz) a Gabor ‘atom’ Time (s) Audio representations Time and Frequency (Gabor) “Theory of Communication,” J. IEE (London) , 1946 “… a new method of analysing signals is presented in which time and frequency play symmetrical parts…”

  9. Gabor and audio coding Time and Frequency (Gabor) “Theory of Communication,” J. IEE (London) , 1946 “…In Part 3, suggestions are discussed for compressed transmission and reproduction of speech or music…” Modern audio coders owe as much to Gabor’s notion of time-frequency analysis as it does to Shannon’s paper of a similar title, two years later, that heralded the birth of information and coding theory. “A Mathematical Theory of Communication,” Bell System Technical Journal, 1948. C. E. Shannon

  10. Image representations … Space and Scale: the wavelet viewpoint: “Daubechies, Ten Lectures on Wavelets,” SIAM 1992 Images can be built of sums of wavelets. These are multi-resolution edge-like (image) functions.

  11. Transform Sparsity What makes a good transform? “TOM” image Wavelet Domain Good representations are efficient - Sparse!

  12. Compressed to 3 bits per pixel Compressed to 0.5 bits per pixel Compressed to 2 bits per pixel Compressed to 1 bits per pixel Compressed to 0.1 bits per pixel Compressed to 2 bits per pixel Tom’s nonzero wavelet coefficients Quantization in wavelet domain Quantization in pixel domain Coding signals of interest What is the difference between quantizing a signal/image in the transform domain rather than the signal domain?

  13. For images (Olshausen and Field, Nature, 1996): For Audio (Abdallah & Plumbley, Proc. ICA 2001): Learning better representations Recent efforts have been targeted at learn better representations for a given set of signals, x(t): That is, learn dictionaries of functions that represent signals of interest with only a small number of significant coefficients, ck.

  14. either good time resolution and poor frequency resolution or good frequency resolution and poor time resolution Build bigger dictionaries Another approach is to try to build bigger dictionaries to provide more flexible descriptions. Consider the following test signal: Heisenberg’s uncertainty principle implies that a Time-Frequency analysis has:

  15. + Combined representation Multi-resolution representations Heisenberg only applies to time-frequency analysis NOT time-frequency synthesis.Consider a TF synthesis representation with a combination of long (40 msec.) atoms and short (5 msec.) atoms. Finding the sparse coefficients is now a nonlinear (and potentially expensive) operation. good frequency representation good time representation New uncertainty principles for sparse representations

  16. Part I Review • Sparse Representations and Coding • How we represent signals is very important; • Sparse representations provide good compression; • Recent efforts have targeted bigger and better representations; • Despite the linear representations nonlinear approximations play an important role.

  17. Part II: Sparse Representations and Source Separation

  18. Observed signals (indexed by i) Mixing process Source signals (indexed by j) Separating mixtures of signals Observed signal Multiple source signals Mathematical model: Aim: extract the sources, sj(t), from the observations, xi(t), often without knowledge of the mixing process (blind).

  19. Blind source separation aims to extract the components associated with each activity. e.g. Observed signals typically contain maternal heartbeat, foetal heartbeat and interference from muscle activity, e.g. breathing. maternal lead 1 foetal (x2 mag.) lead 2 noise (x10 mag.) Time → Biomedical example Foetal ECG signal extraction Using sensors placed on a mother's abdomen, foetal electrical heart activity can be monitored.

  20. Audio example The “Cocktail Party” Problem A microphone array can monitor the acoustic activity in a room. Aim: separate out the individual speakers. x1(t) s1(t) s2(t) x2(t)

  21. x2 s2 x1 s1 Blind source separation Separating the sources without knowledge of the mixing process: Really there are two problems: • Identify the mixing process • This can be solved if the sources are statistically independent (Independent Component Analysis) and non-Gaussian • Separate the sources given the mixing process • Easy if the mixing process is invertible (just invert the mixing matrix). Difficult otherwise, e.g. when the number of source components > number of channels. ICA

  22. Time-Freq. scatter plot Time domain scatter plot Sparse transform More sources than sensors Number of sources > Number of sensors Then the model is said to be overcomplete. e.g. Even if we could determine the mixing process, it is not invertible. We need additional help: 3 speakers 2 microphones sparsity again!

  23. Sparsity and separability Sparse signals can be approximately separated by simple nonlinear operations, e.g. binary thresholding x “ideal” binary mask Black = 0 White = 1 source 1 Black = 1 White = 0 source 2

  24. percussion piano saxophone Blind sparse source separation Given stereo data we can use spatial position of the source to identify the which time-frequency atoms belong to which sources. Here is an example with saxophone, piano and percussion. Left channel only Right channel only Histogram of relative amplitude of TF atoms

  25. Sparse (nonlinear) audio separation Separating Instruments from an extract from John Coltrane. sax piano percussion

  26. Part II Review • Sparse Representations and Source Separation • Representations are important in source Separation • Sparse representations not only provide good basis for compression but also good source separation; • Again nonlinear approximations play an important role.

  27. Part III: Sparse Representations and Sampling (compressed sensing)

  28. This is the philosophy of Compressed Sensing E. Candès, J. Romberg, and T. Tao, “Robust Ucertainty principles: Exact signal reconstruction from highly incomplete frequency information,” IEEE Trans. Information Theory, 2006 D. Donoho, “Compressed sensing,” IEEE Trans. Information Theory, 2006 Compressed sensing Traditionally when compressing a signal (e.g. speech or images) we take lots of samples (sampling theorem) move to a transform domain and then throw most of the coefficients away! Why can’t we just sample signals at the “Information Rate”?

  29. Compressed sensing • The Compressed Sensing principle: • Take a small number of linear observations of a signal (number of observations << number of samples/pixels) • Use nonlinear reconstruction to estimate the signal via a transform domain in which the signal is sparse • Theoretical results • We can achieve an equivalent approximation performance to using the M most significant coefficients for an signal/image (in a sparse domain) by a fixed number of non-adaptive linear observations as long as: • No. of observations ~ M x log N (N, the size of full signal) and • for almost all (random) observations and • can be achieved with practical reconstruction algorithms

  30. 4 3 2 1 Note that 1 is a redundant representation of 2 Wavelet image Invert transform Wavelet image roughly equivalent Sparsifying transform Nonlinear Reconstruction X = original “Tom” sparse “Tom” Observed data Compressed sensing principle

  31. Sub-sampled Fourier Transform ≈ 7 x down sampled (no longer invertible) Haar Wavelet Transform Spatial Fourier Transform Sparse in this domain Logan-Shepp phantom Logan-Shepp phantom …but we wish to sample here We sample in this domain Logan-Shepp phantom Compressed sensing in practice • Compressed Sensing ideas can be applied to reduced sampling in Magnetic Resonance Imaging: • MRI samples lines of spatial frequency • Each line takes time and energy & heats up the patient! • The Logan-Shepp phantom image illustrates this:

  32. ACGP achieves perfect reconstruction at ~ 15% PSNR in dB Best linear reconstruction nonlinear reconstruction (ACGP) Ratio of measurements to signal dimension Compressed sensing in practice Using recently developed algorithms ACGP/GP (Blumensath and Davies, “Gradient Pursuits,” 2007) we can achieve the following results:

  33. Compressed sensing in practice However what we really want to do is… Linear reconstruction (4x under-sampled) Nonlinear reconstruction (4x under-sampled) original (data courtesy of Ian Marshall & Terry Tao, SFC Brain Imaging Centre) …but this reconstruction took us a very long time!

  34. Compressed sensing applications • Compressed Sensing provides a new way of thinking about signal acquisition. • Applications areas include: • Medical imaging • Distributed sensing • Remote sensing • Very fast analogue to digital conversion • (DARPA A2I research program) • Still many unanswered questions… Coding efficiency? Restricted observation domains? Etc.

  35. Summary • Sparse representations provide a powerful mathematical model for many natural signals in signal processing and are a basis for: • good compression; • good source separation and; • efficient sampling • There is an interesting interplay between linear representations and nonlinear approximation • Compressed sensing is only in its infancy…

  36. Final Thought In compressed sensing we have finally learnt that taking signals to bits is easy… its putting the bits back together again that’s difficult.

More Related