Scenery image retrieval by meta-feature representation

Chih Fong Tsai*, Wei Chao Lin

*Corresponding author for this work

Research output: Contribution to journalReview articlepeer-review


Purpose - Content-based image retrieval suffers from the semantic gap problem: that images are represented by low-level visual features, which are difficult to directly match to high-level concepts in the user's mind during retrieval. To date, visual feature representation is still limited in its ability to represent semantic image content accurately. This paper seeks to address these issues. Design/methodology/approach - In this paper the authors propose a novel meta-feature feature representation method for scenery image retrieval. In particular some class-specific distances (namely meta-features) between low-level image features are measured. For example the distance between an image and its class centre, and the distances between the image and its nearest and farthest images in the same class, etc. Findings - Three experiments based on 190 concrete, 130 abstract, and 610 categories in the Corel dataset show that the meta-features extracted from both global and local visual features significantly outperform the original visual features in terms of mean average precision. Originality/value - Compared with traditional local and global low-level features, the proposed meta-features have higher discriminative power for distinguishing a large number of conceptual categories for scenery image retrieval. In addition the meta-features can be directly applied to other image descriptors, such as bag-of-words and contextual features.

Original languageEnglish
Pages (from-to)517-533
Number of pages17
JournalOnline Information Review
Issue number4
StatePublished - 2012
Externally publishedYes


  • Class-specific distances
  • Digital images
  • Feature extraction
  • Feature representation
  • Image processing
  • Image retrieval
  • Meta-features


Dive into the research topics of 'Scenery image retrieval by meta-feature representation'. Together they form a unique fingerprint.

Cite this