Abstract
In this paper, an automatic radar waveform recognition system in a high noise environment is proposed. Signal waveform recognition techniques are widely applied in the field of cognitive radio, spectrum management and radar applications, etc. We devise a system to classify the modulating signals widely used in low probability of intercept (LPI) radar detection systems. The radar signals are divided into eight types of classifications, including linear frequency modulation (LFM), BPSK (Barker code modulation), Costas codes and polyphase codes (comprising Frank, P1, P2, P3 and P4). The classifier is Elman neural network (ENN), and it is a supervised classification based on features extracted from the system. Through the techniques of image filtering, image opening operation, skeleton extraction, principal component analysis (PCA), image binarization algorithm and Pseudo–Zernike moments, etc., the features are extracted from the Choi–Williams time-frequency distribution (CWD) image of the received data. In order to reduce the redundant features and simplify calculation, the features selection algorithm based on mutual information between classes and features vectors are applied. The superiority of the proposed classification system is demonstrated by the simulations and analysis. Simulation results show that the overall ratio of successful recognition (RSR) is 94.7% at signal-to-noise ratio (SNR) of −2 dB.
Keywords: LPI radar, time-frequency distribution, digital image processing, waveform recognition
1. Introduction
There are considerable interests in radar systems that "to see and not be seen" can commonly called low probability of intercept (LPI) radar [1]. LPI radar is used more and more widely. Meanwhile, the number of modulating waveforms is increasing quickly for the LPI radar system. However, it is no longer easy, solely relying on human operators, to recognize the LPI radar of interest. That is, automatic recognition of radar waveform is becoming increasingly important for some critical applications in military-like radar identification, threat analysis, electronic warfare (EW), surveillance, etc. Moreover, it can be applied in civilian areas including signal recognition, spectrum management and cognitive radio, etc.
In the past, many scholars have devoted themselves to exploring the automatic recognition system of the radiation sources in the applications. They have proposed several feasible approaches, making the system more intelligent, more robust, and more like a real human operator. These achievements push forward the development in the field of the signal recognition. In [2] and [3] , the statistical characteristics of the radiation signals is estimated by a maximum likelihood algorithm, and the ratio of successful recognition (RSR) is greater than 90% in the case of signal-to-noise ratio (SNR) ≥ 10 dB. However, the classification of the signals are not extensive, as several kinds of LPI radar waveforms such as Frank code are not mentioned. LPI radar waveform usually has lower signal power, which makes it difficult to classify the LPI radar waveforms directly. Time-frequency (T-F) techniques are used to increase the signal processing gain for LPI radar waveforms [4]. With the application of time-frequency techniques, the T-F feature extraction image processing has become a significant branch of radar waveform recognition. In [5] and [6], Wigner–Ville distribution (WVD) and Pseudo–Wigner distribution (PWD) are adopted to distinguish different frequency modulation (FM) signals, including linear frequency modulation (LFM), FM hyperbolic, FM sinusoidal and others. However, the algorithm for distinguishing polyphase codes (such as Frank code) were not mentioned too much. The short-time Fourier transform (STFT) is presented to analyze the differences between four typical radar signals (LFM, Phase shift keying (PSK), Frequency shift keying (FSK) and continuous wave) in [7] and [8]. In such cases, the RSR is more than 90% at SNR ≥ 0 dB. Nevertheless, it is difficult to distinguish polyphase codes. Zilberman and Pace discussed five kinds of radar waveforms (Binary phase shift keying (BPSK), Frequency modulated continuous wave (FMCW), Frank code, P4 and PT1), who used Choi–Williams time-frequency distribution (CWD) image processing to extract signal features, and the RSR is more than 80% at the SNR of 0 dB [9,10]. Lundén addressed the algorithm which can distinguish eight types of radar waveforms based on Wigner–Ville distribution (WVD) and CWD processing with a high RSR [11]. However, the algorithm estimators of subpulse rate and carrier frequency are needed. In [12], Rihaczek distribution (RD) and Hough transform (HT) are presented to derive two new characteristic features that are especially suitable for the LFM and FSK. However, the RSR of the system reached 90%, in the case that SNR is larger than 24 dB. In [13], three kinds of radar waveforms are recognized (LFM, FSK and PSK), based on random projections and sparse classification, with the RSR over 90% at SNR of 0 dB. Normally, the more kinds of radar waveforms the system identifies, the less robust it is. How to classify more types of radar waveforms in a high noise environment is a challenging problem.
In this paper, an automatic radar waveform recognition system is exploited. Eight types of waveforms (LFM, Costas codes, BPSK, Frank code and P1–P4) were applied in LPI radar, which can identify less but necessary features. Some of the features are obtained from signals directly (we call signal features), and others from a T-F distribution 2D image (we call image features). The signal features consist of second order statistics, power spectral density (PSD) and instantaneous properties (including instantaneous frequency and instantaneous phase), etc. These features are helpful to distinguish polyphase codes from others. Image features are extracted from CWD images of detected signals. CWD is a member of Cohen classes that has fewer cross terms than traditional T-F distributions, such as Wigner–Ville distribution. The extraction procedures from the CWD image includes image binarization, image opening operation, skeleton extraction, Pseudo–Zernike moments [14] and PCA, etc. Image features classify different polyphase codes accurately. After feature extraction, a features selection algorithm based on mutual information, is utilized to pick up more utilitarian features from redundant ones. By applying the greedy selection method, the algorithm selects the features one by one. In other words, the selected features vector is the one that maximizes the mutual information in the classes and the current feature vectors. Elman neural network (ENN) is selected as the classifier cell (networks1 and networks2) of the recognition system since it is a feedback network which is suitable for the classifier. The whole structure of the classifier that we proposed consists of two cells, and each cell has a three-layer ENN for different classifying purposes. This structure is suitable for the automatic waveform system, which has been seldom mentioned before. It has higher RSR in the low SNR environments and less training time. Experimental results show that the recognition system has total RSR over 94%, when SNR ≥ −2 dB.
Our major contributions are summarized as follows: (1) modifying the type of network of classifier (using Elman neural network instead of traditional neural network), CWD image feature extraction (using skeleton extraction instead of Wigner–Ville distribution, data driven and peak search) provided in Lundén’s approaches and improving the experimental results effectively (see Section 6); (2) creating two original features, and , which increase the recognition ability; (3) the proposed approaches being classified into eight kinds of waveforms without prior knowledge; and (4) proposing a new classifier structure, with two three-layer Elman neural networks.
The paper is organized as follows. The automatic waveform recognition system and signal model are introduced in Section 2. Section 3 describes the composition of the classifier. Section 4 explains the features that the system required and how to calculate them. After that, a great number of features are calculated. In order to improve the efficiency of the system, the features selection algorithm is given in Section 5. Section 6 shows the simulation results and makes discussions. Finally, Section 7 draws the conclusions.
2. System Overview
As shown in Figure 1, the system is composed of four modules: time-frequency processing, signal feature estimation, image feature extraction and classifiers. In order to obtain the signal power accurately, it is necessary to estimate the adding of white noise () in the samples. Different from majority of the existing systems, the carrier frequency estimation is not required in local systems because the carrier frequency shows the shift of the position in the CWD image. However, the signal structure is not changed at a different frequency, and the advantage of image features is not being sensitive to where the object lies in the image.
The classifier consists of two classifier cells, network1 and network2. It can be identified including LFM, Costas codes, BPSK, as well as Frank code and P1–P4 polyphase codes. According to all of signal features and several image features, four types of classifications are classified in network1. They are LFM, Costas codes, BPSK and polyphase codes (including Frank and P1–P4). After that, the polyphase codes are classified in network2. At the same time, network2 is controlled by network1. As the signal is considered as polyphase codes by network1, network2 will begin to work. The reason to design two subsidiary cell structures is to reduce the number of input features and to improve accuracy of the signal classification. For more details, see Figure 2.
In this paper, the center frequency of the signal bandwidth is considered as carrier frequency. In addition, the signal is disturbed by additive white Gaussian noise, which means that the intercepted discrete time signal model is given by
(1) |
where n is an integer, T is a sampling interval, is the complex of the transmitted signals, and is complex white Gaussian noise with the variance . A is amplitude, for simplicity and without loss of generality, so we assume that A is an invariant constant. is the instantaneous phase of the complex signal. If the input signal is real, the Hilbert transform is used:
(2) |
where is the real signal, and is Hilbert transform. For more details, see [15].
3. Waveform Classifier
The three-layer ENN is used for the signal classification in the paper. There are connections from the hidden layer to these context units fixed with a weight of one [16]. At each time step, the input is propagated in a standard feed-forward fashion, and then the error back propagation (BP) learning rule is applied [17]. The fixed back connections result in context units always maintaining a copy of previous values of hidden units (since they propagate over the connections before the BP learning rule is applied). Thus, the network can maintain a sort of state, allowing it to perform such tasks as sequence-prediction that are beyond the power of a standard multi-layer perceptron [18,19]. In Figure 3, the hidden layer contains 46 neurons, sigmoid activation is selected in each layer, and the number of input and output layers’ neurons are determined by the vector dimension, respectively.
For ENN, a method to fix the number of hidden layer neurons is discussed in [20]. A simple calculation formula is formulated as:
(3) |
where C is the number of classification, and X is the dimension of a feature vector obtained. A slight adjustment of the number of hidden layer neurons needs to be done, until the network achieves the optimal situation. A simple suggestion is to re-enter the training data as the test data, and the correct recognition rate should be greater than 99%. Otherwise, the number of hidden layer neurons should be adjusted. Forty-six neurons are used in this paper.
4. Features Extraction
In this section, the feature extraction is described in details. Feature extraction plays a key role in the automatic recognition system, the algorithm of which determines not only the RSR but also the robustness of the system. The features include signal features and image features. The section is organized as follows. First, the signal features that based on the second order statistics, PSD, and instantaneous properties are illuminated. Then, the Choi–Williams distribution is introduced, and a further eight types of radar waveforms are shown in CWD images, respectively. After that, image preprocessing based on image morphology is addressed. Finally, the image features of the waveforms are estimated and extracted from the CWD image.
Table 1 lists the necessary features that are presented in network1 and network2. Meanwhile, in order to keep the classifiers as concise as possible, other features are considered but not listed. For example, higher order moments (up to the fourth order) and cumulants (up to the fourth order), Pseudo–Zernike moments (up to the eighth order) and other instantaneous properties, etc. However, these features are not found discriminative enough in the system recognition. How to select the final features will be described in Section 5.
Table 1.
Index | Features | Network1 | Network2 |
---|---|---|---|
1 | Moment | ✔ | |
2 | Moment | ✔ | |
3 | Cumulant | ✔ | |
4 | PSD maximum | ✔ | ✔ |
5 | PSD maximum | ✔ | ✔ |
6 | Std. of instantaneous phase | ✔ | ✔ |
7 | Std. of instantaneous freq. | ✔ | |
8 | Num. of objects (10%) | ✔ | ✔ |
9 | Num. of objects (50%) | ✔ | ✔ |
10 | CWD time peak location | ✔ | ✔ |
11 | Std. of object width | ✔ | ✔ |
12 | Maximum of PCA degree | ✔ | |
13 | Std. of | ✔ | |
14 | Statistics test | ✔ | |
15 | Autocorr. of r | ✔ | |
16 | FFT of corr. | ✔ | |
17 | Pseudo–Zernike moment | ✔ | |
18 | Pseudo–Zernike moment | ✔ | |
19 | Pseudo–Zernike moment | ✔ | |
20 | Pseudo–Zernike moment | ✔ | |
21 | Pseudo–Zernike moment | ✔ | |
22 | Pseudo–Zernike moment | ✔ | |
23 | Pseudo–Zernike moment | ✔ |
4.1. Signal Features
4.1.1. Based on Second Order Statistics
The nth order moment of the complex signal is given by
(4) |
where N is the number of samples, and is conjugated components. and are calculated from this formula. The absolute value is ensured that the estimate value is an invariant constant when the signal phase rotated.
The nth order cumulant is given by [21,22]
(5) |
where is given in Equation (4). In the system, is calculated. Additive independent complex second order circular noise does not affect moment features and cumulant features.
4.1.2. Based on Power Spectral Density (PSD)
PSD describes how the signal power distributes in frequency domain. Two features based on PSD are utilized. The signal should be estimated as an invariant scaling before PSD is calculated. Therefore, invariant scaling is estimated as [23]:
(6) |
where is the kth of samples, and is the three-order moment obtained in Equation (4). The variance of the additive noise can be obtained in [24].
The features about PSD are given by
(7) |
where is given in Equation (6). is a nice feature to distinguish between binary phase and Costas codes. In fact, the square of the complex envelope is constant for binary phase modulation signals. The feature makes an outstanding performance in distinguishing binary phase modulation signals from others.
4.1.3. Based on Instantaneous Properties
Instantaneous properties, which include enormous information, work well in distinguishing frequency modulation signal from phase modulation signal. In this paper, instantaneous frequency and instantaneous phase need to be estimated. The standard deviation of the absolute value of instantaneous phase is estimated as [25]. For the sake of simplicity, is utilized, where and are the imaginary and real parts of the signal, respectively:
(8) |
where N is the number of samples, is the instantaneous phase of the complex signal, and the range of is between −π and π. For more details, see [3].
In order to estimate the instantaneous frequency clearly, the method is decomposed into several steps:
calculate ;
compute from ;
set ;
calculate ;
- calculate the normalized centered instantaneous frequency , i.e.;
- the absolute value of the normalized centered instantaneous frequency is given by;
(9)
-
*
corrects the radian phase angles in by adding multiples of when absolute jumps between consecutive element of is greater than or equal to the jump tolerance of π radians.
-
**
There are some spikes in the instantaneous frequency estimation in the vicinity of phase discontinuity of some waveforms. In order to smooth the and to remove the spikes, a median-filter with window size 5 is used.
4.2. Choi–Williams Distribution (CWD)
Choi–Williams distribution is a member in Cohen Classes [26], which can reduce image interference from cross terms effectively:
(10) |
where t and ω are time and frequency axes, and is the kernel function given by
(11) |
The kernel function is a low-pass filter to eliminate cross terms. σ refers to controllable factor. The bigger σ is, the more obvious cross terms are. Meanwhile, is used in this paper to balance the cross terms and resolution. Eight types of waveforms of CWD transformation are shown in Figure 4. A method for fast calculation of CWD is found in [10]. The structure of a fast CWD method is based on the standard fast Fourier transformation (FFT). Therefore, the number of sampling points is recommended to be a small power of two, such as 128, 256, 512, etc. In this paper, 1024 × 1024 points are selected.
4.3. Image Preprocessing
In the following parts, the real parts of CWD results of waveforms is treated as a 2D image. Digital image processing is explored to gain interested features. In this part, CWD image is processed into a binary image with three operations.
First, the length of detected signal, however, is points in most cases. Zero padding is utilized because we select the CWD transformation of 1024 points. Then, the CWD image is resized to to reduce the computation load. Finally, the resized image is converted to a binary image, based on global thresholding algorithm [27]. The operation steps are as following:
- transform the resized image to gray image between , i.e.;
estimate the initial threshold T. It can be obtained from the average of the minimum and maximum from the image ;
divide the image into two pixel groups and after the comparison with the threshold T. includes all pixels in the image that the values , and includes all pixels in the image that the values ;
calculate the average value and of two pixel groups and , respectively;
- update the threshold value;
- repeat (b–e), and calculate , i.e.;
until the is smaller than a predefined convergence value, 0.001 is used in the paper;
- calculate ;
output the final binary image .
After the image binarization, however, there is some isolated noise and processing noise in the binary images. Isolated noise is generated because the signal is transmitted in the noisy environment. In addition, processing noise is generated in the kernel of CWD itself. It is a kind of special straight line, thin but long. The width of the line is less than three pixels, whereas the majority of lines in the CWD image are longer than half of the image length. In order to remove the noise, a morphological opening is applied (erosion followed by dilation). Erosion and dilation are the basic operations in morphological image processing. Morphological techniques probe an image with a small shape or template called a structuring element. The structuring element is positioned at all possible locations in the image. Furthermore, it is compared with the corresponding neighbourhood of pixels. The structuring element is said to fit the image if, for each of its pixels set to 1, the corresponding image pixel is also 1. Similarly, a structuring element is said to hit, or intersect, an image if, at least for one of its pixels set to 1, the corresponding image pixel is also 1. The opening is so called because it can open up a gap between objects connected by a thin bridge of pixels. Any regions that have survived from the erosion are restored to their original size through the dilation. In the paper, the structuring element in the size of 3 × 3 pixels is used. After the opening operation, the groups as small as a minimum 10% of the size of the largest group are removed. The image process is introduced in Figure 5.
4.4. Image Features
The number of objects in the binary image () is a useful feature. For example, Frank code and P3 have two objects, respectively, while LFM and P1 only have one. However, Costas codes have many objects in different location. In order to distinguish different waveforms, two features and are used in the paper. and are the number of the objects that represent larger than 20% and 50% of the size of the largest object, respectively.
A feature is also found in the location of the maximum energy in time domain of the image, i.e.,
(12) |
where is a resized version of the CWD image, and N is the length of the sampling data.
The standard deviation of the width of the signal objects () and the rotate angle of the largest object () are appropriate for polyphase codes discrimination. Namely, the feature is suitable to classify two kinds of waveforms such as “stepped waveform” (including Frank code, P1) and “linear waveform” (including LFM, P3 and P4). In eight types of waveforms, only P2 has a negative slope. Therefore, P2 can be picked out by the parameter from others easily. The features are calculated as follows:
- for each object, ;
- decide the k and mask the other objects away from the binary image;
- calculate the principal components of the binary image;
- rotate∗ the image until the principal component of the object is parallel to the vertical axis, recorded as ;
- calculate the row sum, i.e., , ;
- normalize , i.e., ;
- calculate the standard deviation of , i.e.,
where N is the number of samples.
output the rotation degree of the maximum of objects ;
- output the average of the , i.e.;
(13)
-
*
nearest neighbor interpolation is used in rotation processing.
Next, we select the maximum object with others removed, extracting the skeleton of the maximum object and estimating the linear trend of it. In the estimation, minimizing the square errors method is applied. The linear trend is subtracted from skeleton of the object to acquire the resulting vector . The standard deviation of is given by
(14) |
where M is the length of .
In order to express the randomness of , a statistical characteristics test is proposed. The details of the test are as follows:
calculate the average of ;
calculate as follows , ;
a consecutive sequence of 0’s or 1’s is called a unit, and R is the number of units. Let and denote the statistics number of and , respectively, i.e., ;
- calculate the mean of units, i.e.;
- calculate the variance of units, i.e.;
- calculate the value of test statistic Y, i.e.;
- output the probability feature, i.e.;
(15)
-
*
where is the standard normal cumulative distribution function. The value of is between 0 and 1.
-
**
note that is no longer a probability. It is a measure of the similarity with Gaussian distribution. The standard deviation value is too small for machine precision. Therefore, it is replaced by variance in (f).
When the is closer to 1, the test signal is more similar to Gaussian distribution, and the number of N is more than 50, the distribution of R is similar to the standard normal distribution, i.e., . However, the values of are quite small for machine precision for the P4 codes. Therefore, the value of R is normalized by using the equation . The feature is worked in network2 only.
The other three features are obtained from the autocorrelation of , i.e., , and . Figure 6 indicates the differences of P1 and P4. In order to characterize these differences, the features are introduced.
The ratio of maximum and sidelobe maximum, i.e.,
(16) |
where, is the location of the minimum of the lag value, and is the location of the sidelobe maximum of the lag value in the . The FFT result is selected to characterize the power of oscillation of , and the feature is estimated as:
(17) |
where is normalized by using the maximum value of itself. Namely, the range of is .
The final features are the members of Pseudo–Zernike moments. Pseudo–Zernike moments are invariant to translation, rotation, scaling and mirroring. It is suitable for the problems about pattern recognition [28,29,30]. These invariant features can reduce the amount of data used in training, which makes the recognition easy. The order of the image geometric moments are defined as:
(18) |
where is the binary image. The scaling and translation invariant central geometric moments are given by
(19) |
where and .
The scaling and translation invariant radial geometric moments are given by
(20) |
where and .
The Pseudo–Zernike moments are defined as:
(21) |
where , and
(22) |
The dynamic range can be reduced by calculating the logarithm, i.e., . The members of the Pseudo–Zernike moments, including , , , , , and are selected. The features are used in network2 only because the features in the network1 can distinguish between LFM, Costas codes, BPSK and polyphase codes very well.
5. Features Selection
In this section, the features selection algorithm for final feature vectors is exploited. With the increase of dimension of feature vector, the redundant information should be removed. The algorithm can select excellent features from a large number of them. Thus, the training time is reduced [31]. Based on the mutual information in the features, greedy selection algorithm is selected. It means that a new feature that has the maximum of mutual information is joined to the feature vector (which includes the previously selected features) in each cycle. The procedure is continued until the number of selected features reaches our design. Mutual information is estimated between class C and the feature vectors X:
(23) |
where is the kth class and c is the number of classes. denotes the entropy, N is the total number of training data and is the jth training vector. Probability density function is estimated by using Parzen windows. Entropy of the class is calculated by using the training data, i.e., , where is the total number of training data from class . The is given by
(24) |
where Δ is a covariance matrix of a d-dimensional vector, and h is the window width parameter, (i.e., ), in which, n is the number of samples. For more details, see [31].
In this stage, a total of 23 features for network1 and network2 are selected. The mutual information between each individual feature and class are estimated. In addition, there are 10 and 19 features remaining for network1 and network2, respectively. The mutual information estimation at each step of the features selection algorithm for network1 and network2 are shown in Figure 7. However, features selection would be stopped, when the mutual information less than 10−4 because the estimation of mutual information is not very accurate especially in a high dimension [32]. See Figure 7 and Table 1 for the list of selected features.
6. Simulation and Discussion
In this section, the automatic waveform recognition system is measured by simulation signals. The purpose is to test accurate rate of the identification results, robustness and computational complexity in different conditions.
6.1. Create Simulation Signals
In this part, we create the simulation signals for experiments. In order to expedite the system recognition, the continuous sampling points are intercepted from the observation window. Supposing radar sources are disturbed by additive white Gaussian noise (AWGN) in propagation channel. In addition, SNR is defined as , where and are variances of original signal and noise respectively. denotes a uniform frequency. For example, there is a original frequency , and the sample rate . The uniform result is . For every waveform, there are different parameters that need to be set. For LFM, the initial frequency is a random number between and . Bandwidth ranges from to as well. For BPSK, the random value of the carrier frequency is , and the length of Barker codes selects from . The number of code periods and the number of cycles per phase code are range of and , respectively. For Costas codes, the number of frequency change is a random value between 3 and 6. The fundamental frequency is . For instance, once the Costas codes are selected, the number is 4, namely, a no-repeating sequence is generated. Then, the Costas codes is decided. For Frank code, the carrier frequency is . In addition, and M are range of and , respectively. For P1–P4 polyphase codes, the parameters are similar to Frank code. For more details, see Table 2.
Table 2.
Radar Waveforms | Parameter | Ranges |
---|---|---|
- | Sampling rate | 1 |
LFM | Initial frequency | |
Bandwidth | ||
Number of samples | ||
BPSK | Carrier frequency | |
Barker codes | ||
Number of code periods | ||
Cycles per phase code | ||
Costas codes | Number change | |
Fundamental frequency | ||
N | ||
Frank & P1 code | ||
M | ||
P2 code | ||
M | ||
P3 & P4 code | ||
M |
6.2. Experiment With SNR
The experiment research shows the relationship between the RSR and SNR. One thousand groups are provided for every type of radar waveform, from which 90% of the groups data are used for training and 10% for testing. The SNR would be increased from −8 dB to 14 dB with a step of 2 dB. The system will compare with Lundén’s [11] because his system has similar classifications to us, and also is considered as the satisfactory automatic radar classification system at present. The simulation results are shown in Figure 8.
Figure 8 plots the classification probabilities as a function of the SNR. The probabilities are measured by the testing data of each kind of waveforms, and the overall probabilities are also calculated. As shown in Figure 8, our system performs better on the classification of each kind of waveform than Lundén’s, especially at low SNR. At SNR > 8 dB, RSR of our system approaches 100%. With the decrease of SNR, the successful ratio can still be kept at a high level. However, the performance of Lundén’s is far from satisfactory. At SNR of −2 dB, most waveforms of RSR reach 90%. Moreover, the overall RSR of the waveforms approach 95%. Nevertheless, the overall RSR of Lundén’s is less than 70% in the same conditions. Costas codes and BPSK are not sensitive with SNR, and P4 code has low RSR in the simulations. Table 3 shows the confusion table at SNR of −2 dB, while the overall RSR is 94.7%. As the table shows, the probabilities of error recognition of Frank code and P4 code are higher than others. Some of Frank codes are identified as P3 codes by the system, while P4 codes are identified as LFM. In fact, the two pairs are very similar, which can be seen in Figure 4. When the SNR is low, the probabilities of error recognition is also reasonable.
Table 3.
LFM | BPSK | Costas | Frank | P1 | P2 | P3 | P4 | |
---|---|---|---|---|---|---|---|---|
LFM | 100 | 0 | 0 | 0 | 0 | 0 | 0 | 18 |
BPSK | 0 | 100 | 0 | 0 | 0 | 0 | 0 | 0 |
Costas | 0 | 0 | 100 | 1 | 0 | 0 | 0 | 0 |
Frank | 0 | 0 | 0 | 91 | 3 | 0 | 4 | 0 |
P1 | 0 | 0 | 0 | 0 | 92 | 0 | 0 | 0 |
P2 | 0 | 0 | 0 | 0 | 0 | 100 | 0 | 0 |
P3 | 0 | 0 | 0 | 8 | 0 | 0 | 93 | 0 |
P4 | 0 | 0 | 0 | 0 | 5 | 0 | 3 | 82 |
6.3. Experiment with Robustness
In this part, we explore the robustness of approach through a different number of training samples. There are 200 groups of testing samples for each of the radar waveform, and the number of training samples would be increased from 100 to 900 with a step of 200. To be more adequate, the experiment would be repeated three times with samples from −8 dB, −2 dB and 10 dB.
As shown in Figure 9, with the increase in number of training samples, the successful recognition is also increased. Meanwhile, for all groups of SNR, the change of successful recognition is not obvious with the number over 500. It means that as long as the number is over 500, the classifier will be able to work in the best state. The approach works well in the small number of samples, which has important significance for radar waveform analysis.
6.4. Experiment with Computation
Computational complexity issue is an important indicator to measure the the performance of classification system. We reproduce Lundén’s method [11], and compare it with this paper in the same conditions. All eight kinds of waveforms are tested under three different SNRs: −8 dB, −2 dB and 10 dB, and each test repeats 10 times on average. The testing environment and testing results are demonstrated as Table 4 and Table 5, respectively.
Table 4.
Item | Model/Version |
---|---|
CPU | E5-1620V2 (Intel) |
Memory | 16 GB (DDR3@1600 MHz) |
GPU | NVS315 (Quadro) |
MATLAB | R2012a |
Table 5.
LFM | BPSK | Costas | Frank | |
−8 dB | 55.604/86.324 | 51.332/82.374 | 54.875/84.279 | 56.336/87.022 |
−2 dB | 54.979/86.111 | 51.195/81.560 | 54.009/84.183 | 56.294/86.654 |
10 dB | 54.783/85.899 | 50.867/81.055 | 53.336/83.807 | 55.793/86.131 |
P1 | P2 | P3 | P4 | |
−8 dB | 58.628/88.282 | 56.754/88.360 | 58.830/87.798 | 54.895/85.999 |
−2 dB | 58.422/87.923 | 55.801/88.180 | 58.107/87.353 | 54.428/85.187 |
10 dB | 57.679/87.005 | 55.368/87.448 | 57.505/86.901 | 53.900/84.533 |
As shown in Table 5, the proposed method spends about 55 s and Lundén’s about 85 s, respectively. However, Lundén’s method has several approaches which do not apply in this paper, such as Wigner–Ville distribution, data driven and peak search, etc. These approaches consume extra time. In each waveform, there is the trace of reduction in time, when SNR is increasing. BPSK spends the least time and P3 spends the most time, but in the overall range, the difference is not obvious. The recipe in this paper is stable, and the variety of SNR and waveforms have little effect on the computational complexity of system.
7. Conclusions
In this paper, an automatic radar waveform recognition system is explored. The recognition system can identify eight types of LPI radar waveforms (including LFM, Costas codes, BPSK (Barker modulation), Frank code and P1–P4) in a highly noisy environment. The RSR relies more on the choice of the features extracted from an intercepted radar pulse as well. The features include two categories, one is extracted from waveforms directly (signal features), the other is from time-frequency image (image features). To reduce cross terms, the CWD transformation is adopted in the time-frequency image. The final feature vectors have been selected by using the information theoretic features selection algorithm.
Elman neural network, a typical global network, is applied as the classifier cell of the recognition system. It is a robust network that has a memory function. The structure of the classifier is composed of two cells, network1 and network2. In network1, the waveforms are classified into four types (including LFM, Costas codes, BPSK and polyphase codes), through the use of all signal features and a part of image features. The polyphase codes are further classified into five types (including Frank code, P1–P4) in network2.
New features have been proposed from the time-frequency image. By adopting translation, scaling and rotation features, we can focus more on the structure of the waveform itself. It is important to classify different kinds of polyphase codes and improve the RSR of polyphase codes. The simulation results show when SNR ≥ −2 dB, and the RSR is more than 94%.
Acknowledgments
This work was supported in part by the National Science Foundation of China under Grant 61201410 and 61571149. The authors would like to thank the anonymous reviewers and the associate editor for their valuable comments and suggestions, which have greatly improved the quality of this paper.
Abbreviations
The following abbreviations are used in this manuscript:
LPI | Low Probability of Intercept |
LFM | Linear Frequency Modulation |
FM | Frequency Modulation |
ENN | Elman Neural Network |
BPSK | Binary Phase Shift Keying |
FMCW | Frequency Modulated Continuous Wave |
CWD | Choi–Williams Time-Frequency Distribution |
PSK | Phase Shift Keying |
FSK | Frequency Shift Keying |
PCA | Principal Component Analysis |
RSR | Ratio of Successful Recognition |
SNR | Signal-to-Noise Ratio |
EW | Electronic Warfare |
STFT | Short-Time Fourier Transform |
WVD | Wigner–Ville Distribution |
PWD | Pseudo–Wigner Distribution |
RD | Rihaczek Distribution |
HT | Hough Transform |
AWGN | Additive White Gaussian Noise |
PSD | Power Spectral Density |
Author Contributions
M.Z. and LT.L. conceived and designed the experiments; M.Z. performed the experiments; M.Z. and LT.L. analyzed the data; M.D. contributed analysis tools; M.Z. wrote the paper.
Conflicts of Interest
All co-authors have seen and agree with the contents of the manuscript and declare no conflict of interest. We certify that the submission is original work and is not under review at any other publication. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.
References
- 1.Fuller K. To see and not be seen (radar) IEE Proc. F-Radar Signal Proc. 1990;137:1–9. doi: 10.1049/ip-f-2.1990.0001. [DOI] [Google Scholar]
- 2.Nandi A.K., Azzouz E.E. Algorithms for automatic modulation recognition of communication signals. IEEE Trans. Commun. 1998;46:431–436. doi: 10.1109/26.664294. [DOI] [Google Scholar]
- 3.Nandi A., Azzouz E. Automatic analogue modulation recognition. Signal Proc. 1995;46:211–222. doi: 10.1016/0165-1684(95)00083-P. [DOI] [Google Scholar]
- 4.Cohen L. Time-frequency distributions: A review. Proc. IEEE. 1989;77:941–981. doi: 10.1109/5.30749. [DOI] [Google Scholar]
- 5.Barbarossa S. Analysis of multicomponent LFM signals by a combined Wigner-Hough transform. IEEE Trans. Signal Proc. 1995;43:1511–1515. doi: 10.1109/78.388866. [DOI] [Google Scholar]
- 6.Barbarossa S., Lemoine O. Analysis of nonlinear FM signals by pattern recognition of their time-frequency representation. IEEE Signal Proc. Lett. 1996;3:112–115. doi: 10.1109/97.489064. [DOI] [Google Scholar]
- 7.López-Risueño G., Grajal J., Sanz-Osorio Á. Digital channelized receiver based on time-frequency analysis for signal interception. IEEE Trans. Aerosp. Electron. Syst. 2005;41:879–898. doi: 10.1109/TAES.2005.1541437. [DOI] [Google Scholar]
- 8.Lopez-Risueno G., Grajal J., Yeste-Ojeda O. Atomic decomposition-based radar complex signal interception. IEE Proc. Radar Sonar Navig. 2003;150:323–331. doi: 10.1049/ip-rsn:20030680. [DOI] [Google Scholar]
- 9.Zilberman E.R., Pace P.E. Autonomous time-frequency morphological feature extraction algorithm for LPI radar modulation classification; Proceedings of the 2006 IEEE International Conference on Image Processing; Atlanta, GA, USA. 8–11 October 2006; pp. 2321–2324. [Google Scholar]
- 10.Pace P.E. Detecting and Classifying Low Probability of Intercept Radar. 2nd ed. Artech House; Norwood, MA, USA: 2009. [Google Scholar]
- 11.Lundén J., Koivunen V. Automatic radar waveform recognition. IEEE J. Sel. Top. Signal Proc. 2007;1:124–136. doi: 10.1109/JSTSP.2007.897055. [DOI] [Google Scholar]
- 12.Zeng D., Zeng X., Cheng H., Tang B. Automatic modulation classification of radar signals using the Rihaczek distribution and Hough transform. IET Radar Sonar Navig. 2012;6:322–331. doi: 10.1049/iet-rsn.2011.0338. [DOI] [Google Scholar]
- 13.Ma J., Huang G., Zuo W., Wu X., Gao J. Robust radar waveform recognition algorithm based on random projections and sparse classification. IET Radar Sonar Navig. 2013;8:290–296. doi: 10.1049/iet-rsn.2013.0088. [DOI] [Google Scholar]
- 14.Chen B., Shu H., Zhang H., Chen G., Toumoulin C., Dillenseger J.L., Luo L.M. Quaternion Zernike moments and their invariants for color image analysis and object recognition. Signal Proc. 2012;92:308–318. doi: 10.1016/j.sigpro.2011.07.018. [DOI] [Google Scholar]
- 15.Xu B., Sun L., Xu L., Xu G. Improvement of the Hilbert method via ESPRIT for detecting rotor fault in induction motors at low slip. IEEE Trans. Energy Convers. 2013;28:225–233. doi: 10.1109/TEC.2012.2236557. [DOI] [Google Scholar]
- 16.Johnson A.E., Ghassemi M.M., Nemati S., Niehaus K.E., Clifton D., Clifford G.D. Machine learning and decision support in critical care. Proc. IEEE. 2016;104:444–466. doi: 10.1109/JPROC.2015.2501978. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Bengio Y., Courville A., Vincent P. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013;35:1798–1828. doi: 10.1109/TPAMI.2013.50. [DOI] [PubMed] [Google Scholar]
- 18.Lin W.M., Hong C.M., Chen C.H. Neural-network-based MPPT control of a stand-alone hybrid power generation system. IEEE Trans. Power Electron. 2011;26:3571–3581. doi: 10.1109/TPEL.2011.2161775. [DOI] [Google Scholar]
- 19.Lin C.M., Boldbaatar E.A. Autolanding control using recurrent wavelet Elman neural network. IEEE Trans. Syst. Man Cybern. Syst. 2015;45:1281–1291. doi: 10.1109/TSMC.2015.2389752. [DOI] [Google Scholar]
- 20.Sheela K.G., Deepa S. Review on methods to fix number of hidden neurons in neural networks. Math. Probl. Eng. 2013;2013 doi: 10.1155/2013/425740. [DOI] [Google Scholar]
- 21.Zhu Z., Aslam M.W., Nandi A.K. Genetic algorithm optimized distribution sampling test for M-QAM modulation classification. Signal Proc. 2014;94:264–277. doi: 10.1016/j.sigpro.2013.05.024. [DOI] [Google Scholar]
- 22.Ozen A., Ozturk C. A novel modulation recognition technique based on artificial bee colony algorithm in the presence of multipath fading channels; Proceedings of the 36th IEEE International Conference on Elecommunications and Signal Processing (TSP); Rome, Italy. 2–4 July 2013; pp. 239–243. [Google Scholar]
- 23.Venäläinen J., Terho L., Koivunen V. Modulation classification in fading multipath channel; Proceedings of the 2002 Conference Record of the 36th Asilomar Conference on Signals, Systems and Computers; Pacific Grove, CA, USA. 3–6 November 2002; pp. 1890–1894. [Google Scholar]
- 24.Stoica P., Moses R.L. Spectral Analysis of Signals. Prentice Hall; Upper Saddle River, NJ, USA: 2005. [Google Scholar]
- 25.Wong M.D., Nandi A.K. Automatic digital modulation recognition using artificial neural network and genetic algorithm. Signal Proc. 2004;84:351–365. doi: 10.1016/j.sigpro.2003.10.019. [DOI] [Google Scholar]
- 26.Feng Z., Liang M., Chu F. Recent advances in time–frequency analysis methods for machinery fault diagnosis: A review with application examples. Mech. Syst. Signal Proc. 2013;38:165–205. doi: 10.1016/j.ymssp.2013.01.017. [DOI] [Google Scholar]
- 27.Gonzalez R.C. Digital Image Processing. Pearson Education; Delhi, India: 2009. [Google Scholar]
- 28.Chen Z., Sun S.K. A Zernike moment phase-based descriptor for local image representation and matching. IEEE Trans. Image Proc. 2010;19:205–219. doi: 10.1109/TIP.2009.2032890. [DOI] [PubMed] [Google Scholar]
- 29.Tan C.W., Kumar A. Accurate iris recognition at a distance using stabilized iris encoding and Zernike moments phase features. IEEE Trans. Image Proc. 2014;23:3962–3974. doi: 10.1109/TIP.2014.2337714. [DOI] [PubMed] [Google Scholar]
- 30.Honarvar B., Paramesran R., Lim C.L. Image reconstruction from a complete set of geometric and complex moments. Signal Proc. 2014;98:224–232. doi: 10.1016/j.sigpro.2013.11.037. [DOI] [Google Scholar]
- 31.Kwak N., Choi C.H. Input feature selection by mutual information based on Parzen window. IEEE Trans. Pattern Anal. Mach. Intell. 2002;24:1667–1671. doi: 10.1109/TPAMI.2002.1114861. [DOI] [Google Scholar]
- 32.Van Hulle M.M. Edgeworth approximation of multivariate differential entropy. Neural Comput. 2005;17:1903–1910. doi: 10.1162/0899766054323026. [DOI] [PubMed] [Google Scholar]