Skip to main content

Kernel generalized neighbor discriminant embedding for SAR automatic target recognition


In this paper, we propose a new supervised feature extraction algorithm in synthetic aperture radar automatic target recognition (SAR ATR), called generalized neighbor discriminant embedding (GNDE). Based on manifold learning, GNDE integrates class and neighborhood information to enhance discriminative power of extracted feature. Besides, the kernelized counterpart of this algorithm is also proposed, called kernel-GNDE (KGNDE). The experiment in this paper shows that the proposed algorithms have better recognition performance than PCA and KPCA.

1 Introduction

Synthetic aperture radar (SAR) has been widely used in many fields, such as terrain surveying, marine monitoring, and earth observation, because of its all-time, all-weather, penetrating ability and high resolution. SAR automatic target recognition (ATR) is the essential technology in SAR image interpretation and analysis.

Generally, the procedure of SAR ATR can be divided into four major steps: detection, discrimination, feature extraction, and recognition. The goal of detection is to locate the potential region of interest. In the discrimination phase, the region of interest is processed to remove the false alarms. The feature extraction is one of the crucial steps for SAR ATR, which can reduce the dimensionality of SAR images greatly and improve recognition efficiency. Finally, the extracted features of the target clips are recognized in the last stage of SAR ATR system.

It has been observed that many feather extraction techniques have been proposed. Principal component analysis (PCA) and linear discriminant analysis (LDA) were used for SAR image feature extraction [1, 2] because of their simplicity and effectiveness. Both of them are based on a global linear structure and need to transform a two-dimensional image into a one-dimensional vector. This will cause a large calculation burden since feather extraction is implemented in a very high-dimensional vector space.

In addition, the kernel trick [3, 4] is applied to extending linear feature extraction algorithms to nonlinear ones. These methods transform input space to other higher or even infinite dimensional inner product space, using nonlinear operators, which is performed by a kernel mapping function. Kernel PCA (KPCA) [5] and kernel LDA (KLDA) [6] describe that in detail.

Recently, the manifold learning algorithm-local preserving projection (LPP) is proposed [7]. But it might not be suitable for SAR ATR, because of its minimization problem, which results in discarding larger principle components.

Based on the manifold learning method, we design neighborhood geometry and target function using the average of similar dispersion of dataset, and then, calculate the linear embedding mapping, according to category information. When this method was extended to vector space, we named it as generalized neighbor discriminant embedding (GNDE). In order to reduce calculation burden, a kernel function was employed to replace the high-dimensional vector inner product. This is the kernel GNDE (KGNDE) method mainly discussed in this paper. It was hoped to solve the nonlinear problem better and improve target identification rate in SAR ATR.

The rest of this paper is organized as follows: We introduce the GNDE in section 2, and KGNDE is proposed in section 3. In section 4, we verify GNDE and KGNDE by the MSTAR database. Finally, we conclude the paper in section 5.

2 Generalized neighbor discriminant embedding

Assume that M is the manifold structure embedded in Rm Euclidean space. Given a training set {x i m, i = 1, 2, …, N} M and their homologous labels {y i  [1, 2, …, c], i = 1, 2, …, N}, where N denotes the total number of training samples in training set, and c is the total class number in the training set. In the integrated class and neighborhood information, GNDE aims at finding a linear embedding map Vm × l : x i m → z i  = VTx i l(i = 1, 2, …, N),  lm, so that samples in the same class keep their neighborhood information and samples in different classes apart from each other. The object function of GNDE is as follows:

J V = i j z i z j 2 w ij

W = [w ij ] N × N is the affinity weight matrix [8], which is defined as

w ij = exp t 1 x i x j , y i y j x i x j < ϵ 1 exp t 2 x i x j , y i = y j , x i x j < ϵ 2 0 , otherwise

where t1 and t2 are constants, ϵ1 and ϵ2 define radius of local neighborhood.

Equation 1 shows that maximizing J V makes samples from different classes apart from each other while samples in the same class proximate in the feature space, which is helpful for discrimination.

Referring to (1) and (2), we can infer that

J V = i j V T x i V T x j 2 w ij = trace i j V T x i x j w ij x i x j T V = trace i j V T x i w ij x i T V i j V T x i w ij x j T V = trace V T XD X T V V T XS X T V = trace V T XL X T V

where S = w 11 w 12 w 1 N 0 w 22 0 w N 1 N 0 0 w NN N × N , D ii = i j w ij , X = x 1 , x 2 , , x N m × N , D = diag(D11, D22, , D NN ) N × N,  L = D − SN × N is a Laplacian matrix. We define an object matrix M V

M V = XL X T


J V = trace V T M V V

Impose an additional constraint:

V T V = E l × l

where E l×l is l × l unit matrix. Finally, optimization problem reduces to find:

arg max V trace V T M V V s . t . V T V = E l × l

Therefore, the optimal embedding map V = [v1, v2, …, v l ] is the set of orthogonal eigenvectors of M V corresponding to the l largest eigenvalue.

GNDE is formally stated as follows:

  1. 1)

    Compute affinity weight matrix W according to (2).

  2. 2)

    According to (3) and (4), compute object matrix M V , resolve the maximization problem as (7) and get the optimal embedding map V.

  3. 3)

    Feature extraction: given a testing sample x T , extracted feature is z T  = V Tx i .

3 Kernel generalized neighbor discriminant embedding

The kernel function is widely used to enhance the classification of linear dimensionality reduction methods. GNDE can be further improved by kernel function, which is named KGNDE. Assume that a nonlinear mapping φ : x i m → φ(x i ) H is introduced, where H is a certain high-dimensional feature space.

The main purpose of KGNDE is to find embedding map Φ H × l : x i m →  k(z i ) = ΦTφ(x i ) l(i = 1, 2, …, N),  lm. According to kernel trick property, Φ = [Φ1, Φ2, , Φ l ], where Φ k = p = 1 N α p k φ x p , α p k . The objective function of KGNDE is as follows:

J K = i j k z i k z j 2 w ij

where w ij is defined as (2).

Based on kernel theory, each element of kernel matrix K = [k ij ] N × N is as follows:

k ij = φ x i , φ x j = φ x i T φ x j

Sometimes, we use Gauss or polynomial function instead of (9). Furthermore, we can recompute k(z i ):

k z i = Φ T φ ( x i ) = p α p 1 φ x p T p α p φ x p T φ ( x i ) = α 1 1 α 2 1 α N 1 α 1 2 α 2 2 α N l 1 α 1 l α 2 l α N l k 1 i k 2 i k Ni = A T K i

where A = α 1 , α 2 , , α l N × l , α i = α 1 i , α 2 i α N i T , K i = k 1 i , k 2 i , , k Ni T .

According to (8) and (10), we can get

J K = i j w ij k z i k z j k z i k z j T = trace A T i j w ij K i K i T K i K j T A

Define an object matrix M K ,

M K = i j w ij K i K i T K i K j T

We can infer that

J K = A T M K A

Impose the additional constraint:

A T A = E l × l

Finally, the object function can be written as:

arg max V trace A T M K A s . t . A T A = E l × l

Therefore, the optimal embedding map A = [α1, α2, …, α l ] is the set of orthogonal eigenvectors of M K corresponding to the l largest eigenvalue.

KGNDE is formally stated as follows:

  1. 1)

    Compute the affinity weight matrix W according to (2), compute kernel matrix K according to (9).

  2. 2)

    According to (11) and (12), compute the object matrix M K , resolve the maximization problem as (15) and get the optimal embedding map A.

  3. 3)

    Feature extraction: given a testing sample x T , extracted feature is k(z T ) = A TK • i.

Now, we concern the computational complexity of the proposed algorithms. In most cases, the number of training samples is less than the dimension of the training sample (Nm). Therefore, like most of other feature extraction methods, the computational bottlenecks of GNDE and KGNDE are solving the generalized eigenvalue problems, whose computational complexity are O(m3) and O(N3), respectively.

4 Experiment

We use the Moving and Stationary Target Acquisition and Recognition (MSTAR) dataset to evaluate GNDE and KGNDE. The training dataset contains SAR images at the depression angle 17°, and testing dataset contains SAR images at the depression angle 15°. Both training dataset and testing dataset cover full 0° ~ 360° aspect ranges. Table 1 lists a detailed information about the type and number included in the training and testing datasets [9].

Table 1 Training and testing datasets

4.1 Experiment steps

  1. 1)

    Image pre-processing: Speckle suppression and target segmentation are used for removing speckles and background clutters, respectively. Then we use gray enhancement based on power function to enhance information in the dataset. Finally, we get the dataset {x i m, i = 1, 2, …, N} called DATA, where x i donates each SAR image vectors and its dimensions m = 61 × 61 = 3, 721. The optical images and the corresponding SAR images of the three targets in the MSTAR dataset are shown in Figures 1 and 2. Images of the targets after processing are shown in Figure 3.

Figure 1
figure 1

Optical images for three targets in MSTAR database.

Figure 2
figure 2

Corresponding SAR images of three targets.

  1. 2)

    Feature extraction: Both GNDE and KGNDE are utilized to extract feature of DATA. In order to examine recognition performance of these methods, PCA and KPCA are also used to extract feature. In this paper, both KPCA and KGNDE use polynomial function as the kernel function, as is shown in (16):

    k ij = x i T x j + 1 μ

    where μ is the function parameter. In this paper, we choose μ = 5.

  2. 3)

    Classification: Nearest neighbor classifier (NNC) [10] is utilized to classify extracted feature based on this algorithm.

Figure 3
figure 3

Preprocessed SAR images of three targets.

4.2 Experiment results

Firstly, we compare GNDE with PCA. As is shown in Figure 4, GNDE performs better than PCA. PCA is an unsupervised method based on a linear structure, while GNDE is a supervised method based on manifold structure. Global linear structure is not applicable for high-dimensional dataset, but manifold structure is. Besides, supervised method is conducive to cluster so that classification is easier. Therefore, GNDE is superior to PCA.

Figure 4
figure 4

Recognition performance of PCA and GNDE.

Secondly, KGNDE is compared with KPCA. From Figure 5, we can see that KGNDE performs better than KPCA as well. The kernel trick can handle nonlinear problems in a high-dimensional dataset. However, KGNDE is a supervised method based on not only the kernel trick but also the manifold structure. The manifold structure can fit the real structure of the dataset, and the supervised method is a benefit to classification. So, KGNDE performs better than KPCA.

Figure 5
figure 5

Recognition performance of KPCA and KGNDE.

Finally, as is shown in Table 2, KGNDE performs slightly poorer than GNDE. Recognition performance of kernel trick is closely related to kernel functions; polynomial function may not be suitable for DATA feature extraction.

Table 2 Best recognition performance by various algorithms

5 Conclusions

Feature extraction is the key step in SAR ATR. In this paper, a new feature extraction algorithm and its kernel counterpart are proposed. Based on the manifold structure, both GNDE and KGNDE get linear transformation to achieve low-dimensional embedding of the dataset. Compared with the linear structure, the manifold ways can detect the underlying nonlinear structure, which preserves local information so that manifold ways is more robust. In addition, GNDE and KGNDE are supervised methods. Through these algorithms, the extracted feature can gain better clustering effect than unsupervised methods, which is helpful for classification.


  1. Mishra AK, Mulgrew B: Bistatic SAR ATR Using PCA-based Features. In Proceedings of the SPIE 6234, Automatic Target Recognition XVI. United Kingdom: University of Edinburgh; 18 May 2006, doi:10.1117/12.664117

    Google Scholar 

  2. Mishra AK: Validation of PCA and LDA for SAR ATR. Paper presented at IEEE Region 10 Conference. Guwahati: IIT Guwahati; 19–21 Nov 2008

  3. Gunn SR: Support Vector Machines for Classification and Regression. Technical Report: Analyst, University of Southampton; 2010. doi:10.1039/B918972F

    Google Scholar 

  4. Zhao Q, Principe JC: Support vector machines for SAR automatic target recognition. IEEE Trans. Aerosp. Electron. Syst. 2001, 2: 643-654.

    Article  Google Scholar 

  5. Li Y, Zhang XQ, Bai BD, Zhang YN: Information Compression and Speckle Reduction for Multifrequency Polarimetric SAR Imagery using KPCA. Paper presented at the 2007 International Conference on Machine Learning and Cybernetics. Xi’an: Northwest Polytechnical University; 19–22 Aug 2007

    Google Scholar 

  6. Han P, Wu RB, Wang YH, Wang ZH: An efficient SAR ATR approach. Paper presented at the 2003 IEEE International Conference on the ICASSP. Tian jin: Tian jin University; 6–10 April 2003

    Google Scholar 

  7. He X, Niyogi P: Locality preserving projections. Paper presented at the 16th Processing Conference on the Neural Information Processing Systems. 2003.

    Google Scholar 

  8. Yan S, Xu D, Zhang BY, Zhang HJ, Yang Q, Li S: Graph embedding and extensions: a general framework for dimensionality reduction. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 1: 40-51. doi:10.1109/TPAMI.2007.250598

    Article  Google Scholar 

  9. Ross T, Worrell S, Velten V, Mossing J, Bryant M: Standard SAR ATR evaluation experiments using the MSTAR public release data set. Paper presented at the Processing Conference on the SPIE. 15 September 1998

  10. Cover TM: Estimation by the nearest neighbor rule. IEEE Trans. Inf. Theory 1968, 1: 50-55.

    Article  Google Scholar 

Download references


This research was supported by the National Natural Science Foundation of China (No. 61201272).

Author information

Authors and Affiliations


Corresponding author

Correspondence to Jifang Pei.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ original submitted files for images

Rights and permissions

Open Access  This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made.

The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.

To view a copy of this licence, visit

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Huang, Y., Pei, J., Yang, J. et al. Kernel generalized neighbor discriminant embedding for SAR automatic target recognition. EURASIP J. Adv. Signal Process. 2014, 72 (2014).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: