### Description

A big issue with images classification is the limitation of training data. In hyperspectral, the data have high dimensional feature vector, that part can be useful. They propose to compensate this limitation by adding synthetic samples drawn from a Gaussian mixture that is estimated on the feature space.

• GANs in deep-learning.
• Robust Classification: SVM, second part.
• Dimensionality Reduction: PCA

### Evaluation

They use the k-means clustering algorithm to provide reasonable initial values of the estimator.

First reduction (PCA) preserve 99% of the total spectral variance. The result will pass in an extended multi-attribute profile (EMAP). They use four thresholds $$lambda$$: area, diagonal of the bounding box, the standard deviation of the gray values and the moment of inertia. All of those have different values possible.

For the second dimension reduction, non-parametric weighted feature extraction (NWFE) is used to preserve 99% of the feature variance.

For classification, they use a random forest, each experiment is repeated 25 times and the mean average accuracy (AA), overall accuracy (OA) and the Kappa along with their std are reported.

### Compare

• Li et al. 1: For 20, 30 and 40 training samples, OA is slightly higher, AA is comparable and for the Kappas, their approach is in all considerably higher.

• Aptoula et al. 2: Deep-learning classification (CNN), better results for all, but 6 to14 times more training samples used.

• Tao et al. 3: Deep autoencoder, better Kappa, 0.9699 for 50 samples compared to 0.9528 for 40 in their case. Again, the other method higher number of samples.

### Results

• 13 pixels per class: [OA,AA] = [79.52, 84.78]
• 30 pixels per class: [OA,AA] = [76.06, 82.67]
• 40 pixels per class: [OA,AA] = [77.12, 85.12]

learning with stacked sparse autoencoder for hyperspectral imagery classification,” IEEE Geoscience and Remote Sensing Letters, vol. 12, no. 12, pp. 2438–2442, 2015.

1. J. Li, X. Huang, P. Gamba, J. M. Bioucas-Dias, L. Zhang, J. A. Benediktsson, and A. Plaza, “Multiple feature learning for hyperspectral image classification,” IEEE Transactions on Geoscience and Remote Sensing, vol. 53, no. 3, pp. 1592–1606, 2015.

2. E. Aptoula, M. C. Ozdemir, and B. Yanikoglu, “Deep learning with attribute profiles for hyperspectral image classification,” IEEE Geoscience and Remote Sensing Letters, 2016.

3. C. Tao, H. Pan, Y. Li, and Z. Zou, “Unsupervised spectral-spatial feature