A deep sparse coding method for fine-grained visual categorization

Author(s):  
Lihua Guo ◽  
Chenggang Guo
2017 ◽  
Vol 61 (1) ◽  
Author(s):  
Lihua Guo ◽  
Chenggang Guo ◽  
Lei Li ◽  
Qinghua Huang ◽  
Yanshan Li ◽  
...  

IEEE Access ◽  
2019 ◽  
Vol 7 ◽  
pp. 158492-158502 ◽  
Author(s):  
Pengfei Wang ◽  
Fugui Qi ◽  
Miao Liu ◽  
Fulai Liang ◽  
Huijun Xue ◽  
...  

2021 ◽  
Vol 13 (4) ◽  
pp. 747
Author(s):  
Yanghua Di ◽  
Zhiguo Jiang ◽  
Haopeng Zhang

Fine-grained visual categorization (FGVC) is an important and challenging problem due to large intra-class differences and small inter-class differences caused by deformation, illumination, angles, etc. Although major advances have been achieved in natural images in the past few years due to the release of popular datasets such as the CUB-200-2011, Stanford Cars and Aircraft datasets, fine-grained ship classification in remote sensing images has been rarely studied because of relative scarcity of publicly available datasets. In this paper, we investigate a large amount of remote sensing image data of sea ships and determine most common 42 categories for fine-grained visual categorization. Based our previous DSCR dataset, a dataset for ship classification in remote sensing images, we collect more remote sensing images containing warships and civilian ships of various scales from Google Earth and other popular remote sensing image datasets including DOTA, HRSC2016, NWPU VHR-10, We call our dataset FGSCR-42, meaning a dataset for Fine-Grained Ship Classification in Remote sensing images with 42 categories. The whole dataset of FGSCR-42 contains 9320 images of most common types of ships. We evaluate popular object classification algorithms and fine-grained visual categorization algorithms to build a benchmark. Our FGSCR-42 dataset is publicly available at our webpages.


Author(s):  
Xianjie Mo ◽  
Tingting Wei ◽  
Hengmin Zhang ◽  
Qiong Huang ◽  
Wei Luo

2021 ◽  
Vol 2021 (29) ◽  
pp. 19-24
Author(s):  
Yi-Tun Lin ◽  
Graham D. Finlayson

In Spectral Reconstruction (SR), we recover hyperspectral images from their RGB counterparts. Most of the recent approaches are based on Deep Neural Networks (DNN), where millions of parameters are trained mainly to extract and utilize the contextual features in large image patches as part of the SR process. On the other hand, the leading Sparse Coding method ‘A+’—which is among the strongest point-based baselines against the DNNs—seeks to divide the RGB space into neighborhoods, where locally a simple linear regression (comprised by roughly 102 parameters) suffices for SR. In this paper, we explore how the performance of Sparse Coding can be further advanced. We point out that in the original A+, the sparse dictionary used for neighborhood separations are optimized for the spectral data but used in the projected RGB space. In turn, we demonstrate that if the local linear mapping is trained for each spectral neighborhood instead of RGB neighborhood (and theoretically if we could recover each spectrum based on where it locates in the spectral space), the Sparse Coding algorithm can actually perform much better than the leading DNN method. In effect, our result defines one potential (and very appealing) upper-bound performance of point-based SR.


Sign in / Sign up

Export Citation Format

Share Document