Scenery image retrieval by meta-feature representation

Chih Fong Tsai*, Wei Chao Lin


研究成果: 期刊稿件文獻綜述同行評審


Purpose - Content-based image retrieval suffers from the semantic gap problem: that images are represented by low-level visual features, which are difficult to directly match to high-level concepts in the user's mind during retrieval. To date, visual feature representation is still limited in its ability to represent semantic image content accurately. This paper seeks to address these issues. Design/methodology/approach - In this paper the authors propose a novel meta-feature feature representation method for scenery image retrieval. In particular some class-specific distances (namely meta-features) between low-level image features are measured. For example the distance between an image and its class centre, and the distances between the image and its nearest and farthest images in the same class, etc. Findings - Three experiments based on 190 concrete, 130 abstract, and 610 categories in the Corel dataset show that the meta-features extracted from both global and local visual features significantly outperform the original visual features in terms of mean average precision. Originality/value - Compared with traditional local and global low-level features, the proposed meta-features have higher discriminative power for distinguishing a large number of conceptual categories for scenery image retrieval. In addition the meta-features can be directly applied to other image descriptors, such as bag-of-words and contextual features.

頁(從 - 到)517-533
期刊Online Information Review
出版狀態已出版 - 2012


深入研究「Scenery image retrieval by meta-feature representation」主題。共同形成了獨特的指紋。