1 / 16

Nonlinear Adaptive Distance Metric Learning for Clustering

Nonlinear Adaptive Distance Metric Learning for Clustering. Jianhui Chen, Zheng Zhao, Jieping Ye, Huan Liu KDD, 2007 Reported by Wen-Chung Liao, 2010/01/19. Outlines. Motivation Objective ADAPTIVE DISTANCE METRIC LEARNING: THE LINEAR CASE

cwen
Télécharger la présentation

Nonlinear Adaptive Distance Metric Learning for Clustering

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Nonlinear Adaptive Distance Metric Learning for Clustering Jianhui Chen, Zheng Zhao, Jieping Ye, Huan Liu KDD, 2007 Reported by Wen-Chung Liao, 2010/01/19

  2. Outlines • Motivation • Objective • ADAPTIVE DISTANCE METRIC LEARNING: THE LINEAR CASE • ADAPTIVE DISTANCE METRIC LEARNING: THE NONLINEAR CASE • NAML • Experiment • Conclusions • Comments • In distance metric learning, the goal is to achieve better • compactness (reduced dimensionality) • separability (inter-cluster distance) • on the data, in comparison with usual distance metrics, such as Euclidean distance.

  3. Motivation • Traditionally, dimensionality reduction and clustering are applied in two separate steps. • If distance metric learning (via dimensionality reduction) and clustering can be performed together, the cluster separability in the data can be better maximized in the dimensionality-reduced space. • Many real-world applications may involve data with nonlinear and complex patterns. • Kernel methods

  4. Objectives • Propose NAML for simultaneous distance metric learning and clustering. • NAML • first maps the data to a high-dimensional space through a kernel function; • next applies a linear projection to find a low-dimensional manifold; • and then perform clustering in the low-dimensional space. • The key idea of NAML is to integrate • kernel learning, • dimensionality reduction, • clustering in a joint framework so that the separability of the data is maximized in the low-dimensional space.

  5. ADAPTIVE DISTANCE METRIC LEARNING: THE LINEAR CASE Data set: Data matrix: Linear transformation W: Mahalanobis distance measure: Min Max ≡ Weighted cluster indicator matrix Cluster indicator matrix

  6. ADAPTIVE DISTANCE METRIC LEARNING: THE NONLINEAR CASE Kernel method: nonlinear mapping Hilbert space (feature space) ψK (X) : the data matrix in the feature space Symmetric kernel function K: (inner product) kernel Gram matrix G: For a given kernel function K, the nonlinear adaptive metric learning problem can be formulated as : convex combination of p kernel matrices

  7. ADAPTIVE DISTANCE METRIC LEARNING: THE NONLINEAR CASE 3.1 The computation of L for given Q and G ≡

  8. ADAPTIVE DISTANCE METRIC LEARNING: THE NONLINEAR CASE 3.2 The computation of Q for given L and G ≡ Max

  9. ADAPTIVE DISTANCE METRIC LEARNING: THE NONLINEAR CASE 3.3 The computation of G for given Q and L ≡ Min MOSEK

  10. NAML O(pk3n3)

  11. Experiment • K-means algorithm as the baseline for comparison • three representative unsupervised distance metric learning algorithms: • Principle Component Analysis (PCA), Local Linear Embedding (LLE), and Laplacian Eigenmap (Leigs) Performance Measures ci the obtained cluster indicator yithe true class label C the set of cluster indicators Y be the set of class labels

  12. Experimental Results (10 RBF kernels for NAML) Experiments

  13. Sensitivity Study: the effect of the input kernels and the regularization parameter λ K-means using 10 kernels for NAML NAML provides a way to learn from multiple input kernels and generate a metric, with which an unsupervised learning algorithm, like K-means, is more likely to perform as well as with the best input kernel the quality of the initial kernel is low

  14. A series of different λvalues ranging from 10−8 to 105 Experiments • λ value in the range of [10−4, 102] is helpful in most cases

  15. Conclusion • NAML • The joint kernel learning, metric learning, and clustering can be formulated as a trace maximization problem, which can be solved iteratively in an EM framework. • NAML is effective in learning a good distance metric and improving the clustering performance. • multiple biological data, e.g., amino acid sequences, hydropathy profiles, and gene expression data. • A future work is to study how to combine a set of pre-specified Laplacian matrices to achieve better performance in spectral clustering.

  16. Comments • Advantage • A joint framework for kernel learning, distance metric learning and clustering • Shortage • Applications • clustering

More Related