,,,
1.College of Civil Aviation,Nanjing University of Aeronautics and Astronautics,Nanjing 211106,P.R.China;
2.The 54th Research Institute of China Electronics Technology Group Corporation,Shijiazhuang 050081,P.R.China;
3.Transport Development Research Center,Zhejiang Scientific Research Institute of Transport,Hangzhou 310023,China
Abstract:It is particular important to identify the pattern of communication signal quickly and accurately at the airport terminal area with the increasing number of radio equipments.A signal modulation pattern recognition method based on compressive sensing and improved residual network is proposed in this work.Firstly,the compressive sensing method is introduced in the signal preprocessing process to discard the redundant components for sampled signals.And the compressed measurement signals are taken as the input of the network.Furthermore,based on a scaled exponential linear units activation function,the residual unit and the residual network are constructed in this work to solve the problem of long training time and indistinguishable sample similar characteristics.Finally,the global residual is introduced into the training network to guarantee the convergence of the network.Simulation results show that the proposed method has higher recognition efficiency and accuracy compared with the state-of-the-art deep learning methods.
Key words:compressed sensing;deep learning;residual network;modulation recognition
The continuous increase of various civil aviation radio stations at the airport terminal area makes the electromagnetic environment more and more complex,and also leads to more serious frequency conflict interference[1].The United States(Next-Gen),Europe(SESAR)and International Civil Aviation Organization(ICAO)Doc9854 have proposed a vision for a new generation of navigation systems,including utilizing the digital communications to reduce the impact of interference in airspace[2].And the modulation recognition method is widely used in the field of digital communications.How to capture and interpret the communication signal effectively plays a major role in ground-to-air communication.
Traditionally,the decision theory is based on hypothesis likelihood ratio and the methods are based on higher-order statistics and feature extraction[3-4].However,the feature extraction is time-consuming and these methods have not been able to meet the demands of high precision and efficiency in the era of data explosion.So that state-of-the-art methods based on deep learning have emerged.The application in the field of communication brings brand-new opportunities and changes.A modulation recognition method based on deep belief network(DBN)was proposed in Ref[.5]to classify the twodimensional images of universal correlation functions of common digital modulation signals.Converting signal into constellation method and recog-nizing by convolution neural network(CNN)were proposed in Ref.[6]which obtained more accurate results than the support vector machine(SVM)and also avoided the manual selection of features.Instead of converting signals into an images,the inphase component and orthogonal component sampling point data proposed in Ref.[7]were directly used as the input of CNN and the stability of the deep learning method was proved at low signal-tonoise ratio(SNR).The reference data set of modulation recognition learning was given in Ref.[8].A simulation data set simulated in a more realistic radio environment was proposed in Ref[.9],which includes 24 kinds of modulation signals.Deep residual network is widely used in the field of computer vision.Through the method of cross-layer connection,the residual network(RN)can avoid the problem of feature graph loss[10-11].RNs for time series radio classification were proposed in Ref.[12]to reduce the training time,but they did not significantly improve the recognition accuracy.
In signal modulation recognition,compressive sensing(CS)is introduced to preprocess the signal to retain the most useful information before inputting the signal into the neural network.The possibility of using measurement signals without reconstruction for detection was provided in Ref.[13].Spectrum detection based on eigenvalues for compressed unreconstructed signals was proposed in Ref.[14],which greatly reduced the computational complexity,but did not affect the probability of signal detection.
Residual neural network can solve the problems that the model training time is too long and the similar features of the sample are hardly to distinguish,which finally lead to a certain enhancement for the recognition accuracy of samples.The image is two-dimensional convolution while the signal is one-dimensional one.The use of neural networks to classify signal modulation is based on the idea of image classification.In this work,the received signal is regarded as a single channel“picture”,and a method based on the improved RN to identify the signal modulation is proposed.Firstly,the received signal is compressed in the preprocessing to extract most of the useful information.Furthermore,the scaled exponential linear units(SELU)activation function is introduced to construct the RN structure.Global residual is introduced to solve the problem of network convergence.Simulation results show that the proposed method has higher recognition efficiency and accuracy.
Combined with the very high frequency(VHF)communications in the airport terminal area,it is assumed that the received signal isr(t),which can be written as
wheres(t)is a modulated signal,crepresents the path loss,andn(t)the white Gaussian noise.The measurement value is obtained by CS as follows
whererisN×1 vector of Nyquist sampling denoting signalr(t),ytheM×1 measurement signal,andΦ∈CM×N(M?N)the measurement matrix.During one cycle,the atom with the largest inner product for the measurement matrix and residual is found and incorporated into the estimation support set.The single atomic selection process can be represented as
whereφiis theith line vector of measurement matrixΦ.
Discrete cosine transformation(DCT)matrix is selected as the sensing matrix in the absence of prior information about the signal.Most of the energy information of the time-domain signal can be compressed into a small number of DCT domains[15],as represented by
wherek∈(0,M-1),i∈(0,N-1),andc(k)is the matrix coefficient.Hence,we arrive at the measurement signal notation of
To prevent over-fitting,three parts data are divided into train set,test set and cross-validation set,y=[y1y2y3]T,wherey1is a training model of training set,y2the validation set,andy3a test set.
The proposed residual network and global skip connection are introduced in this section.
Instead of the original rectified linear unit(RELU),the residual unit based on SELU activation function(R-SELU)is constructed to solve the problem of“neuronal death”at the negative gradient,as shown in Fig.1.
Fig.1 Residual block structure
In the residual unit,the input of the first residual unit is set tox l,then the structure of the residual unit is
whereσis the activation function andF(x l,w l)the function of residual unit.
For any input of unit,the relationship between multiple residual units can be obtained by
Assuming thatJis loss function,then the calculation expression for the error back-propagation is followed by
A 6-layer residual neural network is constructed based on the above residual blocks.The structure diagram is shown in Fig.2.
Fig.2 Structure of ResNet network
The expression of each layer of the above network structure is followed by
whereyis the low dimensional measurements of compressed signals,bthe basis,Wconvolution kernel,the numbernand the sizek×k×c.cis the number of channels.Whenn=1,W1=3×1×1.“*”is convolution operation.
wherey'is the reconstructed measurements of the last convolutional layer.
The activation function used for the full-connection layer is SELU
As can be seen from Eq.(11),when the gradient is greater than 0,the positive half-axis gradient is greater than 1,which is well solved in the RELU activation function.It is not simple to set 0 at the negative half-axis,which solves the case of neuronal death.The introduction of the SELU activation function solves the problem of indistinguishable similar features in the sample.The activation function used by the output layer is
Using training setTto train classifiers is to find appropriate parameters to minimize some loss functions of classifiers.The cross entropy loss function is generally used as
where 1{y i=j}means that wheny=j,the value is 1,otherwise the value is 0.The smaller the value of the loss function,the more accurate the result of the classification training set.
However,Eq.(14)is not a strict convex function,which does not guarantee that it has a unique solution.Increasing the weighted attenuation termpenalizes excessive attenuation value,then,the new loss function is
Using the iterative optimization method to solve Eq.(15),the gradient equation can be obtained as
In the residual block,the relationship between the input and the output can also be written in a simple way as
wherex llandx ll+1are the input and the output of the residual block,respectively.g(·)is the joint operation between the convolution layers andW llrepresents all contained parameters.Sinceσ(·)is nonlinear,g(x l;W l)+x lis not always greater than zero.
Residual block does not accurately learn the residual between the input and the outputx ll+1-x ll,so the overall residualx L-x llis also inaccurate.
Since the input and output of this work are similar,zero padding is used to keep the data dimension consistent.The overall model of the training network is shown in Fig.3.Received signal is compressed using the signal model method to reduce the redundancy and divided into dataset and label.The dataset is based on the constructed residual network and the global residual connection is introduced.Through the full connection layer with SELU as the activation function,the regression output is carried out,the loss function is calculated,and the training network is evaluated.By introducing global skip connection,the residual neural network detects the residuals from the overall network existing between input and output,which leads to a faster convergence rate.
Fig.3 Structure diagram of overall training network
To verify the validity of the proposed method,the open dataset named RML 2018.01a in Ref.[9]was used.74 million signals were extracted,in which 60%data were the training set,20%data were the cross-verification set and the final 20%data were the test set.These extracted data set were divided into 24 parts and preprocessed,including compressing and denoising.The dataset parameters used are listed in Table 1.
Experiments were conducted on Google’s open cloud platform to overcome the inconvenience caused by the hardware limitation.For the sake ofsimplicity,the simulation environment configuration is shown in Table 2.
Table 1 ResNet network layout
Table 2 Simulation environment configuration
After many experiments,the first training time of the whole model is longer by using the setting of hardware conditions.The ResNet network layout is shown in Table 3.
Table 3 Resnet network layout
Dropuout was selected to enhance the generalization ability and robustness of the network.Adam optimizer,an extended algorithm of stochastic gradient descent,was selected in the process of parameter updating,which has excellent performance in practice.Hyper-parameter settings of the training network are shown in Table 4.
Table 4 Hyper-parameter settings of the training network
In this work,the conventional CNN,the ResNet+SELU method used in Ref.[9],the SResNet+RELU method and S-ResNet+SELU method constructed based on the RN proposed in this work are selected for comparative simulations.
Network loss trend of training dataset of four methods is shown in Fig.4(a).Considering the way of batch training,the curve will show ups and downs.At the same iteration number,it can be shown that the RN performs well.It can be shown that the loss function reduction rate of S-ResNet+SELU method is the fastest.At the time of epoch about 13,the loss function begins to converge and finally converges to 1.14.The loss function of SResNet+RELU method finally converges to 1.20,and the loss function of ResNet+SELU method tends to be flat in epoch 30 and finally converges to 1.28.The loss function trend of these three methods confirms the loss function change of the activation function in Fig.4(b).The loss function of CNN shows a“hook back”when epoch is 55,indicating that the network reaches a certain depth and the CNN appeares a gradient disappearance.The convergence trend of using RELU and SELU loss function in the improved network is basically the same,but it is easier to converge than using RELU activation function in the residual function.
Fig.4 Network loss trend of training dataset
The recognition accuracy is the ratio between the correct number of samples predicted and the total number of samples,as
The four models are tested many times,the best results are recorded,and the accuracy is shown in Table 5.
Table 5 Accuracy comparison between the proposed methods and traditional methods on the dataset
Under the condition that the improved residual block uses the SELU,the accuracy of the model using the RELU in the full connection layer is only 1.8%higher.However,SELU converges much faster than RELU,as can be seen from Fig.4(b).
The accuracy in the range of SNR of the four methods is shown in Fig.5.
Fig.5 Accuracy trend of test dataset
When SNR is 10 d B,the three methods based on the RN achieve the best effect.At low SNR,the accuracy of the proposed method has been on top of the other methods.The higher-order modulation of the signal is less easily distinguished at low SNR,however,the recognition rate of the traditional residual module is low when it has similar samples,which explains the poor recognition effect of the ResNet+SELU in the low SNR.The S-ResNet+RESU and the S-ResNet+SELU can still achieve a recognition rate of 50%when SNR is 0,and the SResNet+SELU has reached 90.2%of the recognition accuracy at 8 dB.
The parameters of CNN,ResNet+SELU and the S-ResNet+SELU proposed in this work are compared,as shown in Table 6.
Table 6 Number of parameters
CNN uses 257 099 parameters,while ResNet+SELU uses 238 840 parameters.Furthermore,the number of parameters used in the proposed SResNet+SELU is 139 192.It is found that the number of used parameters is reduced by 99 648.In the process of model training,the method in this paper trains epoch that requires a mean detection time of 175 s,and the average time used by ResNet+SELU is 477 s.In hardware limitations,the detection time used in this paper is longer,and the training of the whole model takes about 2.95 h.
The confusion matrix is compared under different SNRs illustrated in Fig.6.When the three kinds of SNR are-2,6 and 10 dB,the modulation mode can still be identified accurately.It is shown that the modulation modes of 32PSK,AM-DSB-WC,F(xiàn)M and OOK can be accurately identified when SNR is-2 dB.AM-DSB-WC is the modulation mode used in VHF communication system,and 256QAM and QPSK cannot be easily distinguished under low SNR because of their similar characteristics.When SNR is 6 dB,there are 16 modulation modes withthe recognition rate of above 80%.When SNR is 10 dB,the confusion matrix can basically get pure diagonal lines,and most modulation modes can be recognized.However,128QAM and AM-DSB-SC have 20%recognition error rate,128APSK and 128QAM have 20% recognition error rate,256QAM and QPSK modulation mode have about 10%recognition error rate.When SNR are 20 dB and 30 dB,there will still be shadows when the high-order PSK modulation is used.However,it has a good recognition effect for the analog modulation used in the traditional VHF communication,and this work will also provide a reference for the choice of digital modulation in the civil aviation VHF communication system in the future.
Fig.6 Confusion matrix of test set
Aiming at the signal modulation pattern recognition,this work proposes an improved residual network based on SELU activation function combined with CS.It is shown to solve the problem of high computational complexity,complicated steps of the artificial feature extraction and slow convergence speed featured by the conventional residual network.The proposed algorithm firstly discards the redundant components when extracting information from the input signal using the measurement matrix.Then,when training the classification network,a residual unit and residual block structure based on SELU as the activation function is constructed.Furthermore,to solve the inaccuracy of residual learning and difficulty at network convergence,the concept of global skip connection is introduced.According to a series competing simulations,the proposed method is demonstrated to have the short training time,the faster convergence speed and the better recognition accuracy of the modulation mode when comparing with other four popular methods.
In view of the complex electromagnetic environment in the airport terminal area,the prospect offurther research in this work is to design a recognition method for the performance detection of real signal sources based on deep learning spectrum sensing technology.
AcknowledgementsThis work was supported by the National Natural Science Foundation of China(No.71874081),Special Financial Grant from China Postdoctoral Science Foundation(No.2017T 100366)and Open Fund of Hebei Province Key laboratory of Research on data analysis method under dynamic electro-magnetic spectrum situation.
AuthorProf.SHEN Zhiyuan received his B.S.,M.S,and Ph.D.degrees all from Harbin Institute of Technology in 2008,2010 and 2014,respectively.During 2010 to 2012,he was a visiting scholar at Georgia Institute of Technology.His research interests include air traffic management,airport operation and optimization,digital signal processing and deep learning.
Author contributionsProf.SHEN Zhiyuan designed the study,complied the models,conducted the analysis,interpreted the results and wrote the background of the study.Dr.LI Jia contributed to complying the models and data selection for analysis and simulation design.Ms.WANG Qianqian contributed to complying the models,conducting the simulation and discussion,and writing the manuscript.Ms.HU Yingying contributed to adding supplement for revised version and adjusting the paper format.All authors commented on the manuscript draft and approved the submission.
Competing interestsThe authors declare no competing interests.
Transactions of Nanjing University of Aeronautics and Astronautics2021年4期