Multi-Scale Multi-Level Generative Model in Scene Classification

Wenjie XIE  De XU  Yingjun TANG  Geng CUI  

IEICE TRANSACTIONS on Information and Systems   Vol.E94-D   No.1   pp.167-170
Publication Date: 2011/01/01
Online ISSN: 1745-1361
DOI: 10.1587/transinf.E94.D.167
Print ISSN: 0916-8532
Type of Manuscript: LETTER
Category: Image Recognition, Computer Vision
scene classification,  msml-pLSA,  visual granularity,  

Full Text: PDF(158.3KB)>>
Buy this Article

Previous works show that the probabilistic Latent Semantic Analysis (pLSA) model is one of the best generative models for scene categorization and can obtain an acceptable classification accuracy. However, this method uses a certain number of topics to construct the final image representation. In such a way, it restricts the image description to one level of visual detail and cannot generate a higher accuracy rate. In order to solve this problem, we propose a novel generative model, which is referred to as multi-scale multi-level probabilistic Latent Semantic Analysis model (msml-pLSA). This method consists of two parts: multi-scale part, which extracts visual details from the image of diverse resolutions, and multi-level part, which concentrates multiple levels of topic representation to model scene. The msml-pLSA model allows for the description of fine and coarse local image detail in one framework. The proposed method is evaluated on the well-known scene classification dataset with 15 scene categories, and experimental results show that the proposed msml-pLSA model can improve the classification accuracy compared with the typical classification methods.