Back to Search Start Over

Texture characterization for joint compression and classification based on human perception in the wavelet domain

Authors :
John Arthur Black
Sethuraman Panchanathan
Gamal Fahmy
Source :
IEEE Transactions on Image Processing. 15:1389-1396
Publication Year :
2006
Publisher :
Institute of Electrical and Electronics Engineers (IEEE), 2006.

Abstract

Today's multimedia applications demand sophisticated compression and classification techniques in order to store, transmit, and retrieve audio-visual information efficiently. Over the last decade, perceptually based image compression methods have been gaining importance. These methods take into account the abilities (and the limitations) of human visual perception (HVP) when performing compression. The upcoming MPEG 7 standard also addresses the need for succinct classification and indexing of visual content for efficient retrieval. However, there has been no research that has attempted to exploit the characteristics of the human visual system to perform both compression and classification jointly. One area of HVP that has unexplored potential for joint compression and classification is spatial frequency perception. Spatial frequency content that is perceived by humans can be characterized in terms of three parameters, which are: 1) magnitude; 2) phase; and 3) orientation. While the magnitude of spatial frequency content has been exploited in several existing image compression techniques, the novel contribution of this paper is its focus on the use of phase coherence for joint compression and classification in the wavelet domain. Specifically, this paper describes a human visual system-based method for measuring the degree to which an image contains coherent (perceptible) phase information, and then exploits that information to provide joint compression and classification. Simulation results that demonstrate the efficiency of this method are presented.

Details

ISSN :
10577149
Volume :
15
Database :
OpenAIRE
Journal :
IEEE Transactions on Image Processing
Accession number :
edsair.doi.dedup.....a5e34a61aaa194ee498359bb54f3928b