1 / 1

Correlative Multi-Label Multi-Instance Image Annotation

Overview. Correlative Multi-Label Multi-Instance Image Annotation Xiangyang Xue, Wei Zhang, Jie Zhang, Bin Wu, Jianping Fan, and Yao Lu Fudan University, China & UNC-Charlotte.

berke
Télécharger la présentation

Correlative Multi-Label Multi-Instance Image Annotation

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Overview Correlative Multi-Label Multi-Instance Image Annotation Xiangyang Xue, Wei Zhang, Jie Zhang, Bin Wu, Jianping Fan, and Yao LuFudan University, China & UNC-Charlotte • The global visual features of the entire image and the local features of the regions are extracted to capture coarse and fine patterns, respectively. • The associations between semantic concepts and visual features are mined both at image level and at region level. • Inter-label correlations are captured by a co-occurence matrix of concept pairs. • The cross-level label coherence encodes the consistency between the labels at image level and the labels at region level. • The Model: • are the (nonlinear) functions mapping the input global features of the entire image and the local features of the image region to the kernel spaces, respectively. • denotes the label vector of an image. • denotes the concept label vector of the r-th region in the image. • denotes the set of all concepts related with the l–th concept. • are the parameter vectors to be learned; are the bias parameters. • Part I encodes the associations between image-level labels and global visual features; • Part II models the associations between region-level labels and local visual features; • Part III captures the inter-label correlations dependent on the image features; • Part IV measures the coherence between image-level labels and region-level labels. Learning the model by minimizing the cost function: Based on the design of the proposed model, we can divide the optimization problem into inter-related sub-problems and then learn the model efficiently. Experimental Results Conclusions Results on Corel Dataset Results on MSRC Dataset • Both image-level labels and region-level labels can be obtained in a single framework by capturing the feature-label associations, the inter-label correlations, and the cross-level label coherence. • Structural max-margin technique is used to formulate the proposed model. • By decoupling the annotation task into inter-dependant subproblems, we learn multiple interrelated classifiers jointly. We evaluate our method on MSRC and Corel image datasets in comparisons with other related competitive algorithms: RML[1], RankSVM[1], MLknn[3], and TagProp[4]. The inter-label correlation matrix based on the harmonic mean of empirical conditional probabilities illustrates the interdependency between concepts on the MSRC dataset. The brighter the block is, the stronger the correlation between labels exists. The results of our method in comparison with other related in terms of F score for 10 labels at the image-level from the Corel dataset. The results of our method in comparison with other related competitive algorithms in terms of F score for individual labels at the image-level on the MSRC dataset. References [1] James Petterson and Tiberio Caetano. Reverse multi-label learning. In NIPS, 2010. [2] Andre Elisseeff and Jason Weston. A kernel method for multi-labeled classification. In NIPS, 2002. [3] M-L Zhang and Z-H Zhou. Ml-knn: A lazy learning approach to multi-label learning. Pattern Recognition, 40(7): 2038–2048, 2007 [4]M. Guillaumin, T. Mensink, J. Verbeek, and C. Schmid. Tagprop: Discriminative metric learning in nearest neighbor models for image auto-annotation. In ICCV, 2009. Region-level labeling results of our method for some exemplary images from MSRC. Top: the ground truth; Bottom: our results. Region-level labeling results of our method for some exemplary images from the Corel dataset.

More Related