Modeling global geometric spatial information for rotation invariant classification of satellite images

PLoS One. 2019 Jul 19;14(7):e0219833. doi: 10.1371/journal.pone.0219833. eCollection 2019.

Abstract

The classification of high-resolution satellite images is an open research problem for computer vision research community. In last few decades, the Bag of Visual Word (BoVW) model has been used for the classification of satellite images. In BoVW model, an orderless histogram of visual words without any spatial information is used as image signature. The performance of BoVW model suffers due to this orderless nature and addition of spatial clues are reported beneficial for scene and geographical classification of images. Most of the image representations that can compute image spatial information as are not invariant to rotations. A rotation invariant image representation is considered as one of the main requirement for satellite image classification. This paper presents a novel approach that computes the spatial clues for the histograms of BoVW model that is robust to the image rotations. The spatial clues are calculated by computing the histograms of orthogonal vectors. This is achieved by calculating the magnitude of orthogonal vectors between Pairs of Identical Visual Words (PIVW) relative to the geometric center of an image. The comparative analysis is performed with recently proposed research to obtain the best spatial feature representation for the satellite imagery. We evaluated the proposed research for image classification using three standard image benchmarks of remote sensing. The results and comparisons conducted to evaluate this research show that the proposed approach performs better in terms of classification accuracy for a variety of datasets based on satellite images.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Algorithms
  • Geographic Information Systems
  • Geographic Mapping
  • Geography*
  • Maps as Topic
  • Models, Theoretical*
  • Satellite Imagery*

Associated data

  • figshare/10.6084/m9.figshare.7006946.v1
  • figshare/10.6084/m9.figshare.8798345.v3
  • figshare/10.6084/m9.figshare.8796980.v1

Grants and funding

This work was supported by Basic Science Research Program (2017R1D1A1B03033526 and 2016R1D1A1B03933860) and Priority Research Centers Program (NRF-2017R1A6A1A03015562) through the National Research Foundation of Korea (NRF) funded by the Ministry of Education.