Enhanced hierarchical model of object recognition based on a novel patch selection method in salient regions

Yan Feng Lu, Tae Koo Kang, Hua Zhen Zhang, Myo Taeg Lim

Research output: Contribution to journalArticle

7 Citations (Scopus)

Abstract

The biologically inspired hierarchical model for object recognition, Hierarchical Model and X (HMAX), has attracted considerable attention in recent years. HMAX is robust (i.e. shift- and scale-invariant), but its use of random-patch-selection makes it sensitive to rotational deformation, which heavily limits its performance in object recognition. The main reason is that numerous randomly chosen patches are often orientation selective, thereby leading to mismatch. To address this issue, the authors propose a novel patch selection method for HMAX called saliency and keypoint-based patch selection (SKPS), which is based on a saliency (attention) mechanism and multi-scale keypoints. In contrast to the conventional random-patch-selection-based HMAX model that involves huge amounts of redundant information in feature extraction, the SKPS-based HMAX model (S-HMAX) extracts a very few features while offering promising distinctiveness. To show the effectiveness of S-HMAX, the authors apply it to object categorisation and conduct experiments on the CalTech101, TU Darmstadt, ImageNet and GRAZ01 databases. The experimental results demonstrate that S-HMAX outperforms conventional HMAX and is very comparable with existing architectures that have a similar framework.

Original languageEnglish
Pages (from-to)663-672
Number of pages10
JournalIET Computer Vision
Volume9
Issue number5
DOIs
Publication statusPublished - 2015 Oct 1

    Fingerprint

ASJC Scopus subject areas

  • Computer Vision and Pattern Recognition
  • Software

Cite this