1 / 23

Learning with Structured Sparsity

Learning with Structured Sparsity. Authors: Junzhou Huang, Tong Zhang, Dimitris Metaxas. Introduction. Fixed set of p basis vectors where for each j . -->

elgin
Télécharger la présentation

Learning with Structured Sparsity

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Learning with Structured Sparsity Authors: Junzhou Huang, Tong Zhang, Dimitris Metaxas Zhennan Yan

  2. Introduction • Fixed set of p basis vectors where for each j. --> • Given a random observation , which depends on an underlying coefficient vector . • Assume the target coefficient is sparse. • Throughout the paper, assume X is fixed, and randomization is w.r.t. the noise in observation y. Zhennan Yan

  3. Introduction • Define the support of a vector as • So • A natural method for sparse learning is L0 regularization for desired sparsity s: • Here, only consider the least squares loss Zhennan Yan

  4. Introduction • NP-hard! • Standard approach: • Relaxation of L0 to L1 (Lasso) • Greedy algorithms (such as OMP) • In practical applications, often know a structure on β in addition to sparsity. • Group sparsity: variables in the same group tend to be zero or nonzero • Tonal and transient structures: sparse decomposition for audio signals Zhennan Yan

  5. Structured Sparsity • Denote the index set of coefficients • For any sparse subset • Coding complexity of F is defined as:

  6. Structured Sparsity • If a coefficient vector has a small coding complexity, it can be efficiently learned. • Why ? • Number of bits to encode F is cl(F) • Number of bits to encode nonzero coefficients in F is O(|F|)

  7. General Coding Scheme • Block Coding: Consider a small number of base blocks (each element of is a subset of ), every subset can be expressed as union of blocks in . • Define code length on : • Where • So

  8. General Coding Scheme • a structured greedy algorithm that can take advantage of block structures is efficient: • Instead of searching over all subsets of up to a fixed coding complexity s (exponential), we greedily add blocks from one at a time • is supposed to contain only manageable number of base blocks

  9. General Coding Scheme • Standard Sparsity: consisted only of single element sets and each base block has coding length . This uses bits to code each subset of cardinality k. • Group Sparsity: • Graph Sparsity:

  10. General Coding Scheme • Standard Sparsity: • Group Sparsity: Consider , let contain the m groups, and contain p single element blocks. Element in has cl0 of ∞, and element in has cl0 of . only looks for signals consisted of the groups. • The result coding length is: if can be represented as union of g disjoint groups. • Graph Sparsity:

  11. General Coding Scheme • Standard Sparsity: • Group Sparsity: • Graph Sparsity: Generalization of Group Sparsity. Employs a directed graph structure G on . Each element of is a node of G but G may contain additional nodes. • At each node , we define coding length clv(S) on the neighborhood Nv of v, as well as any other single node with clv(u), such that

  12. General Coding Scheme • Example for Graph Sparsity: Image Processing Problem • Each pixel has 4 adjacent pixels, the number of the subsets in its neighborhood is 24 = 16, with a coding length . Encode all other pixels using random jumping with coding length • If connected region F is composed of g sub-regions, then the coding length is • While standard sparse coding length is Zhennan Yan

  13. Algorithms for Structured Sparsity Zhennan Yan

  14. Algorithms for Structured Sparsity • Extend forward greedy algorithms by using block structure, which is only used to limit the search space. Zhennan Yan

  15. Algorithms for Structured Sparsity • Maximize the gain ratio: • Using least squares regression • Where is the projection matrix to the subspaces generated by columns of XF • Select by Zhennan Yan

  16. Experiments-1D • 1D structured sparse signal with values +1~-1, • p = 512, • k =32 • g = 2 • Zero-mean Gaussian noise with standard deviation is a added to the measurements • n = 4k = 128 • Recovery result by Lasso, OMP and structOMP: Zhennan Yan

  17. Experiments-1D Zhennan Yan

  18. Experiments-2D • Generate a 2D structured sparsity image by putting four letters in random locations. • p = H*W = 48*48 • k = 160 • g = 4 • m = 4k = 640 • Strongly sparse signal, Lasso is better than OMP! Zhennan Yan

  19. Experiments-2D Zhennan Yan

  20. Experiments for sample size Zhennan Yan

  21. Experiment on Tree-structured Sparsity • 2D wavelet coefficient • Weakly sparse signal Zhennan Yan

  22. Experiments-Background Subtracted Images Zhennan Yan

  23. Experiments for sample size Zhennan Yan

More Related