Mingyi He, Bo Li, Huahui Chen: Al. (11) Proposed A Modified Deep Stacking Network (DSN) For
Mingyi He, Bo Li, Huahui Chen: Al. (11) Proposed A Modified Deep Stacking Network (DSN) For
Mingyi He, Bo Li, Huahui Chen: Al. (11) Proposed A Modified Deep Stacking Network (DSN) For
Table 1. Parameters of convolutional layers
kernel kernel kernel size kernel stride
name number H, W, B Δ(H, W, B)
conv1 16 3,3,11 1,1,3
conv2 1 1,1,1
conv2 2 1,1,3
16 1,1,1
conv2 3 1,1,5
Fig. 2. Illustration of multi-scale 3D convolution block. m1 , conv2 4 1,1,11
m2 and m3 denote the kernel sizes in the 2 spatial and 1 spec- conv3 1 1,1,1
tral dimensions, respectively. conv3 2 1,1,3
16 1,1,1
conv3 3 1,1,5
conv3 4 1,1,11
conv4 16 2,2,3 1,1,1
pooling – 2,2,3 2,2,3
N
1
E=− log(pnk ) (5)
N n=1
classification. Considering the spatial resolution of the data The Indian Pines dataset gathered by Airborne Visible/Infrared
and the target sizes for each classes to be classified, a relative Imaging Spectrometer (AVIRIS) sensor in North-western In-
small kernel size in the 2 spatial dimensions is suit for our diana consists of 145 × 145 pixels with a ground resolution
experiments. of 17 m and 220 spectral reflectance bands in the wavelength
rang 0.4-2.5 μm. We reduce the number of bands to 200
The detailed hyper-parameter setting of this model is pre-
by removing bands covering the region of water absorption.
sented in Table 1. The hyper-parameters are chosen for val-
It includes 16 classes, and we select 8 classes due to some
idation on the training data. In other words, we used 80%
classes having too few labeled samples.
of the training samples to learn weights and the remaining
The University of Pavia dataset acquired by the Reflective
20% to choose the proper hyper-parameters. We used the
Optics System Imaging Spectrometer (ROSIS) sensor during
same model setting for all the three datasets. In other words,
a flight campaign over Pavia University consists of 610 × 340
we don’t deliberately tune the hyper-parameters to pursue a
pixels with ground resolution of 1.3 m and 103 bands.
higher performance.
The Salinas dataset was collected by the 224-band AVIRIS
We train our network with the multinomial logistic loss: sensor over Salinas Valley, California, comprising 512 × 217
pixels with a ground resolution of 3.7 m. We reduce the num-
ber of bands to 204 by removing bands covering the region of
water absorption.
In all the 3 datasets, we randomly select 200 labeled pixels
per class for training and the rest for testing. The input of our
network is the HSI 3D patch in the size of 7 × 7 × Band,
where Band denotes the total number of spectral bands. The
size of the output is the number of the classes. For paper space (a) Ground truth (b) RBF-SVM (c) Hu’s CNN
limit, only the training number and test number for each lass
of the Indian Pines dataset are presented in Table 2.
5. REFERENCES [13] Tran Du, Bourdev Lubomir, Fergus Rob, Torresani
Lorenzo, and Paluri Manohar, “Learning spatio-
[1] Xiuping Jia, Bor-Chen Kou, and Melba Crawford, “Fea- temporal features with 3d convolutional networks,”
ture mining for hyperspectral image classification,” Pro- arxiv.org/abs/1412.0767, 2014.
ceedings of IEEE, vol. 101, no. 3, pp. 676–697, 2013.
[14] Yushi Chen, Hanlu Jiang, Chunyang Li, and Xiuping
[2] Mingyi He, Wenjuan Chang, and Shaohui Mei, “Ad- Jia, “Deep feature extraction and classification of hy-
vance in feature mining from hyperspectral remote sens- perspectral images based on convolutional neural net-
ing data,” Spacecraft Recovery & Remote Sensing, vol. works,” IEEE Transactions on Geoscience & Remote
34, no. 1, pp. 1–12, 2013. Sensing, vol. 54, no. 10, pp. 1–20, 2016.
[3] Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, [15] S. Ji, M. Yang, and K. Yu, “3d convolutional neural
“Gradient-based learning applied to document recogni- networks for human action recognition,” IEEE Transac-
tion,” Proceedings of the IEEE, vol. 86, no. 11, pp. tions on Pattern Analysis & Machine Intelligence, vol.
2278–2324, 1998. 35, no. 1, pp. 221–31, 2013.
[4] Suraj Srinivas and R. Venkatesh Babu, “Deep learning [16] Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky,
in neural networks: An overview,” Computer Science, Ilya Sutskever, and Ruslan Salakhutdinov, “Dropout:
2015. a simple way to prevent neural networks from overfit-
[5] A. Romero, C. Gatta, and G. Camps-Valls, “Unsuper- ting,” Journal of Machine Learning Research, vol. 15,
vised deep feature extraction for remote sensing image no. 1, pp. 1929–1958, 2014.
classification,” IEEE Transactions on Geoscience & Re- [17] Yangqing Jia, Evan Shelhamer, Jeff Donahue, Sergey
mote Sensing, vol. 54, no. 3, pp. 1349–1362, 2015. Karayev, and Jonathan Long, “Caffe: Convolutional ar-
[6] Chen Xing, Li Ma, and Xiaoquan Yang, “Stacked de- chitecture for fast feature embedding,” Eprint Arxiv, pp.
noise autoencoder based feature extraction and classifi- 675–678, 2014.
cation for hyperspectral images,” Journal of Sensors,
[18] John Duchi, Elad Hazan, and Yoram Singer, “Adaptive
vol. 2016, pp. 1–10, 2016.
subgradient methods for online learning and stochastic
[7] Wei Hu, Yangyu Huang, Li Wei, Fan Zhang, and optimization,” Journal of Machine Learning Research,
Hengchao Li, “Deep convolutional neural networks for vol. 12, no. 7, pp. 257–269, 2011.
hyperspectral image classification,” Journal of Sensors,
vol. 2015, no. 2, pp. 1–12, 2015.
[8] Yushi Chen, Xing Zhao, and Xiuping Jia, “Spectralc-
spatial classification of hyperspectral data based on deep
belief network,” IEEE Journal of Selected Topics in Ap-
plied Earth Observations & Remote Sensing, vol. 8, no.
6, pp. 1–12, 2015.
[9] Jun Yue, Wenzhi Zhao, Shanjun Mao, and Hui Liu,
“Spectralcspatial classification of hyperspectral images
using deep convolutional neural networks,” Remote
Sensing Letters, vol. 6, no. 6, pp. 468–477, 2015.
[10] Heming Liang and Qi Li, “Hyperspectral imagery clas-
sification using sparse representations of convolutional
neural network features,” Remote Sensing, vol. 8, no. 2,
2016.
[11] Mingyi He and Xiaohui Li, “Deep stacking network
with coarse features for hyperspectral image classifica-
tion,” in WHISPERS’16, Aug 2016.
[12] S. Mei, J. Ji, Q. Bi, J. Hou, and Q. Du, “Integrating spec-
tral and spatial information into deep convolutional neu-
ral network for hyperspectral classification,” in IGARSS,
July 2016, pp. 5067–5070.