 Research
 Open access
 Published:
Flight maneuver intelligent recognition based on deep variational autoencoder network
EURASIP Journal on Advances in Signal Processing volumeÂ 2022, ArticleÂ number:Â 21 (2022)
Abstract
The selection and training of aircraft pilots has high standards, long training cycles, high resource consumption, high risk, and high elimination rate. It is the particularly urgent and important requirement for the current talent training strategy of national and military to increase efficiency and speed up all aspects of pilot training, reduce the training cycle and reduce the elimination rate. To this end, this paper uses deep variational autoencoder network and adaptive dynamic time warping algorithms as support to explore the establishment of an integrated evaluation system for flight maneuver recognition and quality evaluation, solve the industry difficulty faced by current flight training data mining applications, and achieve accurate recognition and reliable quality evaluation of flight regimes under the background of high mobility. It will fully explore the benefits of existing airborne flight data for military trainee pilots, support the personalized and accurate training of flight talents, and reduce the rate of talent elimination.
1 Introduction
In the face of the urgent demand for high quality pilots for the launching of aircraft carriers and the commissioning of carrierborne aircraft, how to train high quality and efficient military flying talents is a historical mission in front of military flying academies. At present, there is still a gap in the mining application of big data for the growth of flight cadets and the mining of flight talent cultivation and growth rules based on big data support. Therefore, it is of great practical significance to carry out effective research on intelligent flight maneuver identification and evaluation methods, which are mainly reflected in the following four aspects:
First, launch the pioneering exploration of big data construction and application for the growth of flight cadets. The industry has reached a consensus on the value and significance of big data for the growth of flight cadets. It is urgent to carry out the construction of big data standard system, construction of supporting information conditions, data production and accumulation, data mining, analysis and application.
Second, explore an effective way to improve the quality of flight talent training based on data. With the rapid development of network information system application and communication technology, data collection and accumulation of flight cadets have become increasingly rich [1, 2]; The growth data of flight cadets can accurately reflect their learning motivation, learning attitude and learning effect. Through data mining, and correlation analysis, forming students into multidimensional dynamic curve of growing up, can appear in students learning will reduce, study effect of weakening trend, actively introduce management department, teaching and training department human intervention and guidance to help, to stimulate students learning enthusiasm, supplement short weaknesses; Through the accumulation of a large number of historical flight data and flight statistics, Construct the growth model of flight cadet training, And after a large number of historical data training and test correction, used for scientific early warning of flying cadets, Put forward reasonable opinions and suggestions on the model diversion of students in data, providing an effective way to improve the quality of flight talent training.
Third, solve the urgent need of flight training quality assessment of flight cadets. During flight training, cadets have accumulated a large number of rich flight parameters. For a long time, the training management departments at all levels, in the implementation of the outline of the training requirements, check whether the flight vehicles complete regulations training maneuver, training maneuver whether meet the requirements, etc., in the face of vast amounts of flight parameter data, and to rely on artificial analysis is difficult to achieve, 'Whether to fly, whether the quality of the flight is good' is plagued by departments at all levels of reality problem for a long time. Based on the theoretical method and means of flight parameter data mining, the realization of automatic identification of flight regimes and quantitative evaluation of maneuver quality, automatic intelligent processing of flight parameter data of each flight, and analysis of flight maneuver completion rate and quality can effectively solve the urgent need of management departments at all levels.
Fourth, it is inevitable to give full play to the data effectiveness of flight cadets' growth. Big data is hailed as a fighting force in the new field and new space of military struggle. Education and training are moving towards the era of big data, who can discover the data, who can win the future survival; Who can mine the data, who can win the future development; Who can leverage data and provide personalized services, who can win the competition in the future. With the indepth promotion of the concept of big data and the iterative evolution of data processing technology, how to fullyÂ leverageÂ these real historical data, let data "speak", and use data toÂ support decisions has become the reality of the development of The Times.
2 Related work
It is the premise of pilot control quality monitoring that flight movements can be quickly and accurately identified from the flight parameter data. Applied in the field of flight cadets training, it can realize the evaluation of flight quality of flight cadets, the targeted guidance of weaknesses of flight cadets in flight solitary action, the targeted auxiliary decisionmaking of flight cadets diversion and the early warning of flight cadets in terms of their successful flight so as to improve the efficiency of flight training.
A lot of flight training data resources were accumulated during the daily flight training of military flying cadets, which can be utilized to analyze and assess their flight action and flight quality. But the problems exist widely in the traditional algorithm applied to process such data, such as the not high accuracy in the identification of the flight action, the low resolutionÂ ratio of the quality assessment on a fivepoint scale, the unsatisfactory discrimination of scoring in application etc. Especially in the current situation of the enhancement in military flight training intensity and the better flight mobility, the traditional motion recognition and quality evaluation algorithm is almost invalid in the application, resulting in the difficulty in carrying out the objective data evaluation of the current military flight quality.
Flight motion recognition is essentially a pattern recognition problem. Pilot control quality monitoring was carried out earlier in civil aviation aircraft. Because civil aviation aircraft were in flat flight state for a long time. pilot control quality monitoring was mainly reflected in the identification of takeoff, climb, descent and landing stages, which is easy to identify and of little value and significance. Military aircraft, especially fighter aircraft, are difficult to be identified because of their strong flight maneuverability, complex and changeable movements, and are prone to misidentification and missing identification. Mature application has been used in foreign military aircraft based on radar observation and data mining to analyze the aircraft flight movements and movement trend in order to realize the prediction of the other sideâ€™s situational awareness on the battlefield and win the battlefield situational countermeasures initiative. The research on flight motion recognition methods of domestic military aircraft began in the beginning of this century, mainly by Air Force Engineering University and Naval Aviation University etc. Aiming at flight motion recognition tasks, the flight motion recognition methods widely used in engineering mainly rely on expert systems and the fast identification of the flight action was achieved by the established knowledge base of motion recognition according to the changing characteristics of flight parameters extracted from the flying data by the specialistsâ€™ experience, the developed inferenceÂ engine applying advanced computer language and the forward accurate reasoning strategy adopted. In 2004, Xie Chuan et al. used support vector machine to divide the flight action recognition task into two stages: action data screening and action segment classification [3]. In 2005, Xie Chuan et al. proposed a method to extract the characteristics of flight parameters using rough set theory to improve the accuracy of model classification [4]. Ni Shihong et al. constructed the knowledge base of flight parameter discrimination rules for typical flight actions based on the idea of expert system [5]. In 2006, Zhang Ruifeng et al. proposed an algorithm using Cardinals cubic spline curve to generate aircraft flight paths [6] as the basis of flight motion evaluation. Yin Wenjun et al. proposed a flight action recognition method based on genetic algorithm [7]. In 2011, Su Chen et al. and Gao Gao et al. used artificial immune system algorithm [8] and improved quantum genetic algorithm [9] respectively to automatically extract flight action discrimination rules without relying on expert knowledge. In 2018, Wang Yuwei et al. proposed an extraction method of flight action recognition rules based on whale optimization algorithm [10]. In addition, expert knowledge was used to decompose complex actions into five basic actions and establish a flight action knowledge base to realize flight action recognition [11]. The disadvantage of this method is that it is difficult to judge the completeness and accuracy of knowledge base and knowledge multilayer nested relationship will appear for the representation of complex maneuvering. In extreme cases, for some tactical maneuvers, it is difficult to extract the characteristics of flight parameters, and the knowledge of motion recognition cannot be expressed.
In recent years, scholars have proposed that the problem of motion recognition can be transformed into the similarity query of flight time series and standard time series. The earliest research paper on time series similarity was proposed by Agrawal et al. in 1993 [12, 13], and subsequently became one of the hot issues with many valuable studies, such as discrete Fourier Transform method (DFT) [12, 13], discrete wavelet transform method (DWT) [14], Singular value decomposition (SVD) [15], piecewise Approximation (PAA) [16], dynamic time warping (DTW) [17, 18], etc. The relevant research began earlier that Discrete Fourier transform and Rtree search are used to realize the similarity query of time series. The discrete wavelet transform method uses the scaling and translation wavelet to replace the fixed window for calculation and analysis. However, the information loss of this algorithm is serious, so it is not suitable for nonstationary sequence. The piecewise approximation method segmented the time series by equal width window, and the time series in each window was represented by the window average value. A piecewise linear representation of the time series was obtained, and then the similarity query was carried out. Its advantage is that it has strong data compression ability and can keep the main shape of time series well. Its disadvantages are that it cannot handle sequences of arbitrary length and does not support distance measurements with weights. Singular value decomposition method regards variables in time series as random variables, the observed values at each time point as sample points and the correlation coefficient matrix as the basis of feature extraction and constructs a pattern matching model according to the coordinate transformation principle in linear space. Its advantage is that it has strong data compression ability, and the precise details of data processing can be retained. Its disadvantages are that it cannot handle sequences of arbitrary length and does not support distance measurements with weights. Singular value decomposition (SVD) can also realize the motion analysis of multiple flight parameters, and its transformation is global, but the time and space complexity of the algorithm are relatively high. Dynamic time warping is widely used in speech recognition, and its core idea is to obtain the minimum path through dynamic time warping technology, which has the advantage of realizing the optimal alignment and matching, supporting the time axis bending of sequence and the similarity measurement of time series of different lengths, but the calculation is relatively complicated. In 2015, Li Hongli et al. use multiple dynamic time neat (Multivariant Dynamic Time Warping MDTW) algorithm to calculate the similarity between the time sequence and standard action sequence to determine its action categories [19]. In 2017, Shen Yichao et al. used hierarchical clustering based on multivariate dynamic time regularization to screen node features, and then constructed a Bayesian network to predict the probability of time series belonging to different categories [20]. Zhou Chao et al. used the improved multivariate dynamic time warping algorithm to achieve the hierarchical preclassification and subclassification of flight movements [21]. In 2019, Shen Yichao et al. proposed a dynamic timewarping path illcondition matching algorithm to remove invalid fragments in time series and improve the precision of action starting and ending points [22]. In 2018, Xiao Yanping et al. conducted a comprehensive analysis of the lateraldirectional flight quality evaluation indexes of aircraft from the perspective of system dynamics model [23]. In 2019, Wang Yuwei et al. constructed a fivelevel comprehensive evaluation structure system for the flight quality, and used the comprehensive weighting method to conduct weighted evaluation on the score of basic movements [24]. In 2020, Xu Gang et al. proposed an automatic evaluation method of training effectiveness of combat simulation [25]. In addition, some studies used lowcost simulated flight data to replace costly actual flight [26]. However, influenced by many factors, such as the aircraft performance, the pilot operation habits, the flight environment et. complex maneuver are with strong randomness and fuzziness. Hence, the above algorithms are, generally, falling into disuse because the recognition features is not diverse enough, they adopt the hard division for the precise threshold and the complex maneuver randomness and fuzziness fail to be fully illustrated for the knowledge expression and reasoning process.
3 Flight maneuver recognition and quality assessment system
This paper solves the identification of the maneuver name, flight time starting point and quality assessment of three specific issues, build flight regimes intelligent identification and quality evaluation system. Based on flight parameters of flight training subsets according to get the plane threedimensional trajectory and attitude data, in order to carry out the maneuver criterion standard establishment, flight regimes intelligent identification, adaptive flight quality assessment.
3.1 Establishment of flight maneuver criteria
Flight parameter data in flight training data comprehensively record flight training subjects and content, flight maneuver data of aircraft and aircraft control data, etc., with abundant, authoritative, objective and true data accumulation. Flight maneuver is the basic element of flight training unit and the key index of flight assessment, and a single flight training includes several flight regimes that need to be assessed. Aimed at training model, combining with the aircraft maneuver characteristics and flight instructor expert experience, based on the maneuver of the aircraft flight parameters data sets, a comprehensive combing to identify flight regimes of classification methods and standards requirements, key flight maneuver set, put forward the classification standard, the index requirements of flight regimes, lay the foundation for flight gesture recognition and evaluation.
3.2 Intelligent identification of flight maneuver
In the mining analysis and application of flight training data, the traditional algorithms generally have some problems, such as low recognition rate of flight maneuver, low resolution of quality evaluation and low differentiation of evaluation. Deep learning methods based on deep neural network have been widely used in feature extraction of big data. In order to give full play to the feature extraction ability of neural network and make the deep neural network pay attention to the timing characteristics of collected flight maneuver signals and adaptively extract features from them, based on deep residual convolution, a deep variational autoencoder network is constructed to realize flight maneuver recognition, which is expected to fill the gap of the lack of effective mining and analysis methods for flight training data.
3.3 Adaptive evaluation of flight quality
On the basis of efficient completion of automatic identification of flight regimes, according to the flight maneuver and flight sorties two levels, combined with the proposed maneuver classification criterion index system and deep variational autoencoder network, a multiview fusion maneuver scoring algorithm based on variational autoencoder feature compression network and adaptive dynamic time warping was implemented.
4 Flight training maneuver library
4.1 4.1 Criterion standard
Flight parameters derived from different aircraft components, points belong to different hardware management system, causes the tend to have different sampling frequency and scope of said, after data preprocessing is needed to input variables of the model, such as neural networks as the depth at the same time also ensures network in training can give equal importance to different parameters. Parameter types include Boolean switching variables, Continuous variables, Discrete variables and Angle variables, so data preprocessing includes resampling, normalization and slicing. Firstly, the sampling frequency of fast variable parameters and slow variable parameters is unified by downsampling and oversampling techniques. Then, according to the physical characteristics of each parameter, the upper and lower bounds of its theoretical amplitude are determined, and the input parameters of the model are normalized to the same numerical range by means of amplitude scaling, which is usually [0,1] or [âˆ’â€‰1,1]. This process can significantly improve the learning efficiency and prediction accuracy of the model. In the slicing process, the entire flight sorties with overlap are divided into multiple segments to ensure that the input samples of the segmentation network have the same time sequence length. Finally, the output sequences obtained are spliced and combined to obtain the output results of the whole flight sorties. When only a few training samples are available, data enhancement is essential to make the network as invariant and robust as possible.
4.2 The threedimensional trajectory reproduction
Due to the increasing intensity of flight training and stronger maneuverability of flight, the corresponding flight regimes cannot be accurately identified only by analyzing flight parameters However, 3d trajectory reconstruction can reproduce the real flight trajectory in flight sortals by observing the realtime changes of flight trajectory, supplemented by the variation trend of main flight parameters, flight regimes and their starting and ending moments can be accurately, realtime and intuitively identified.
The threedimensional position of an aircraft is determined by and only by longitude, latitude, and altitude. Therefore, the flight parameters of longitude, latitude and barometric height in flight sorties are extracted as x, y and z coordinates respectively to draw scatter plots in threedimensional space. In order to obtain a smooth 3D trajectory, the moving average method is used to smooth the 3D coordinates of the aircraft at each time point. By adding new and old data one by one in order to calculate the average value, the moving average method eliminates accidental variation factors and achieves the effect of smoothing the curve.
After obtaining the overall threedimensional trajectory of the aircraft, we expect to observe the trend of the flight trajectory with time. Therefore, the way of sliding window is adopted to draw the flight track of the aircraft in the window [t, tâ€‰+â€‰100] at time t. With the change of time t, the window continues to move along the time axis, and the flight path of the aircraft changes at the same time, realizing the realtime change of the flight path, which is more conducive to the identification of flight regimes. In order to realize intuitive and accurate judgment of flight regimes by combining dynamic 3D trajectory reconstruction and realtime flight parameter data, flight history tracing software is built, as shown in Fig.Â 1. The interface is mainly divided into timeofflight area, dynamic 3D trajectory area, realtime flight parameters area and interaction area. The time of flight area is located at the top of the interface, showing the running time of the flight track. The dynamic 3D trajectory area is located at the left end of the interface, where the dynamic threedimensional trajectory of flight at different moments is displayed as time changes. By dragging the mouse position, you can show the 3D flight track from different angles for further maneuver analysis and judgment. The realtime flight parameter area is located on the right side of the interface, which dynamically displays six major flight parameters highly correlated with flight regimes at that time. By analyzing the realtime flight parameters, the flight regimes can be more accurately mined. The interaction area is located at the bottom of the interface and contains two control buttons "Stop" and "Start", which are used to control the pause and start of the flight history backtracking algorithm. The flight history tracing interface can be used to analyze and trace any historical flight sorties, and then build flight maneuver database, which lays a foundation for the establishment of classification model.
4.3 Flight regimes to be identified
As shown in Table 1, there are a total of 21 categories of flight regimes to be identified, which are divided into three categories: takeoff and landing routes, basic regimes and stunt regimes.
5 Depth variational autoencoder maneuver recognition model
In the mining of flight training data, the traditional algorithms rely on expert experience, the recognition rate of flight maneuver is not high, and the model generalization ability is poor. Deep learning algorithm based on convolutional neural network improves the accuracy and generalization ability of recognition model in a datadriven way. However, the traditional classification and recognition model is aimed at a single category of samples, and it still needs the help of expert knowledge to classify specific maneuver samples, so it cannot process the flight data containing multiple types of flight regimes. In this paper, pointbypoint prediction was made based on the time sequence characteristics of the flight parameters in multiple time series, and a pointbypoint maneuver recognition network based on the depth variational autoencoder network was constructed to perform intelligent segmentation of flight regimes in the complete sorties.
5.1 Variational autoencoder
variational autoencoder is a kind of considering the generated model derived from the hidden variable approximation, it includes coding device network \(q_{\phi } (zx)\) and decoder network \(p_{\theta } (xz)\), as shown in Fig.Â 2. Where, x is the visible variable of input sample features, and z is a series of unobserved hidden layer variables. The deep variational autoencoder network uses variational inference theory and stochastic gradient descent principle to train the network. The encoder compresses the raw data into a lowdimensional space so that the visible variable x maps to the continuous hidden variable z. The decoder uses hidden layer variables to generate data and uses hidden variable z to reconstruct the original data x. Depth variation since the coding network distribution with depth of the neural network structure encoder \(q_{\phi } (zx)\) and decoder network \(p_{\theta } (xz)\), thus to extract the characteristics with abundant information of hidden layer.
According to prior assumptions, the implicit variable z is constrained to be normally distributed \(p_{\theta } (z)\sim N(\mu ,\sigma )\), that the model learns the distribution of input data. But the marginal likelihood function of distribution is difficult to solve. According to the variational theory, the lower boundary of the marginal likelihood function is taken as the objective function. Given the approximate posterior \(q_{\phi } (zx)\), variational since coding network loss function is shown by
where \({\mathbb{E}}_{{q{(}zx{)}}}\) is the expect under the posterior distribution \(q_{\phi } (zx)\), \(D_{{{\text{KL}}}}\) is the KL divergence between the approximate posterior \(q_{\phi } (zx)\) and the prior distribution \(p_{\theta } (z)\) of the potential variable z, which in used to measure the gap between approximate posterior and prior distributions. The greater the KL divergence value, the greater the difference between the two distributions.
5.2 Deep variational autoencoder maneuver recognition network
For the multivariate time series in the flight process, the depth variational multivariate variational autoencoder network is constructed, as shown in Fig.Â 3. It can be divided into two parts: variational autoencoder feature extraction and multiple time series segmentation. For input of multivariate time series \(x \in R^{l \times n}\), where l is the length of the time series, n is the number of flying parameters entered.
Firstly, the hidden layer features of flying parameter are extracted by deep variational autoencoder network \(\overline{x} \in R^{l \times z}\), where z is the dimension of hidden layer features, let the features extracted can be standardized without redundancy, which is conducive to further flight maneuver sequence segmentation. Then, convolution and downsampling are carried out to encode the hidden layer features as highlevel features with temporal relations. Finally, deconvolution and upsampling are performed on the highlevel features to output the sequence prediction category probability vector \(\hat{y} \in R^{l \times 1}\). The classification error is computed by the sequence prediction category probability vector \(\hat{y} \in R^{l \times 1}\) and the onehot encoding vector of sequence real class \(y \in R^{l \times 1}\) together. Combined with classification error and reconstruction error, the model parameters are updated iteratively by gradient descent method to determine the final network. The specific network structure and parameters are shown in Table 2.
5.3 Loss function and model training
The loss function of the model is divided into two parts: reconstruction error and classification error. For variational autocoding networks, we make the minimum meansquare error between the decoder output \(\hat{x} \in R^{l \times n}\) and the original input \(x \in R^{l \times n}\) as small as possible to ensure that no information is lost in the extracted hidden variables. At the same time, in order to make the approximate posterior of hidden variables close to the real prior, the KL divergence between them should be minimum. Therefore, the reconstruction error can be expressed as
where \(\mu\) and \(\sigma\) are the mean vector and variance vector output by encoder g respectively.
In order to improve the accuracy of classification error, the predicted category should be consistent with the label as much as possible. The cross entropy loss function is adopted as
where \(p_{ij}\) represents the probability of class J maneuver predicted by the model at moment i, and \(y_{ij}\) represents the real flight maneuver performed at moment i. Therefore, the total loss function of time series segmentation is the sum of reconstruction error and classification error, as shown by
Other model training Settings are as follows. All network parameters are initialized by kaiming initialization. The number of training samples per batch was 512. The total loss of time series segmentation is taken as the objective function, and the gradient of the objective function about the weight is calculated by the back propagation strategy, and the network weight is updated in the opposite direction of the gradient. The random gradient descent (Adam) optimizer was selected, and the initial learning rate was 0.1, the exponential decay strategy was used to attenuate the learning rate, and the decay rate was selected as 0.95. The model parameters were saved after 100 rounds of training.
5.4 Flight data preprocessing
Collect Kâ€‰=â€‰21 flight parameters (such as heading Angle, pitch Angle, barometric height, etc.) within multiple flight sorties. It is spliced into a multiple time series of Lâ€‰Ã—â€‰K, where L is the total time of the series. In order to speed up the convergence process of the model and improve the calculation speed of the network, the overall statistical indicators of each flight parameter are calculated respectively and the overall normalization process is carried out so that the value of each parameter is about 1. Finally, it is divided into training set and test set in the ratio of 6:4.
Then, to generate training samples, resampling the training set is performed. In order to ensure consistent model input size, a sliding window with a fixed size L_{t} was used to traverse the whole training set and divide the time series into several subsamples. At the same time, in order to collect a sufficient number of samples to meet the requirements of model training, the whole time series is traversed with a step size of I_{t}. So, for the time series with points L, the sample number N obtained by resampling is shown by
5.5 Model testing process
In the test phase, the sequence to be predicted was truncated into several subsequences whose length met the network input length Lt, and each subsequence was input into the maneuver recognition network to obtain several prediction subsequences. Then the prediction subsequence is spliced to obtain the pointbypoint prediction label of the whole flight sortie. Finally, the boundary judgment is carried out to extract each flight maneuver segment of flight sorties as the output result. The test process is shown in Fig.Â 4.
The identification accuracy of network on flight maneuver types and maneuver start and end times has a significant impact on the identification results of the whole flight sortie. Therefore, this paper adopts three types of evaluation indicators to evaluate the model, which are described as follows.
The IntersectionoverUnion refers to the ratio of intersection and union between the predicted result and the real tag, and measures the similarity between the predicted sequence and the real tag sequence, as shown by
where I_{j} is an indicator function used to indicate whether there is an element point belonging to maneuver j in the sample sequence. p_{j} indicates that the maximum probability of output prediction is the number of points of maneuver j, and y_{j} is the number of points that are truly labeled as maneuver j. I(Â·) and U(Â·) represent the intersection and union of p_{j} and y_{j} respectively, and C represents the total number of categories. Overall accuracy refers to the proportion of correctly predicted points outputted by the network to all sequence points, as shown in Formula (9).
where P indicates the maneuver corresponding to the maximum probability predicted by the output.
The F1 score also focuses on the overall prediction accuracy of the model and the balance between different categories. It is the harmonic mean of accuracy rate and recall rate, and its value ranges from 0 to 1, as shown in Eq.Â (10). The closer the value is to 1, the stronger the model performance is.
The evaluation results of overall accuracy, intersection ratio and F1 score are shown in Table 3. Among them, the training results of all indicators are close to 100%, and the overall accuracy of the model is 92.76% when the model is run in the test set, indicating that the model can identify flight regimes well. The intersection ratio of the test is 0.7354, indicating that the recognition of the start and end moments of flight regimes is basically correct, but there is still room for improvement. The recognition accuracy of some categories is not high, which is related to the quality and data size of tags. More highquality flight data and tags need to be added to train better classification models.
6 Flight quality assessment algorithms
Due to the high maneuverability and complex and changeable regimes of military aircraft, the flight quality evaluation of military aircraft lacks reasonable measures. The existing quality evaluation largely depends on the expert experience of flight instructors, so it is not highly differentiated in practical application. At the same time, the highly nonlinear mutation between multiple flight regimes interferes with the quality evaluation of the whole flight envelope. In this paper, a flight maneuver quality evaluation algorithm based on deep variational autoencoder network is proposed to eliminate redundancy of contextual maneuver information in the whole flight scene, and a scoring rule combining global feature and local feature is constructed to realize the comprehensive evaluation of flight maneuver quality from multiple perspectives.
6.1 Standard maneuver base library is established
In order to realize the maneuver quality assessment of different flight categories, it is necessary to select the corresponding standard maneuver datum and establish the standard maneuver benchmark library, which can be used as the reference datum of flight regimes and guide the threshold setting of subsequent scoring rules. The establishment process of the standard maneuver reference library is as follows: Combine the flight training outline of the designated aircraft type with the experience and knowledge of the flight experts, the relevant standard flight maneuver items are selected from the rich historical flight data, and the typical standard maneuver database is established. The relevant standards include flight parameter standards and threedimensional trajectory standards of various flight regimes. The criteria for flight parameters include entry speed, exit speed, takeoff Angle, time, altitude, intake pressure and other flight parameters. The 3D trajectory standard is used to judge whether the 3D trajectory of flight maneuver conforms to the maneuver description and whether its smoothness meets the requirements.
According to the above standards, some standard data of each type of flight regimes are extracted from the flight history data, and the start and end times of the regimes are recorded, which are included in the standard regimes reference base.
6.2 Variational autoencoder feature compression network
The flight environment is complex when an aircraft enters a maneuver, and the flight maneuver parameters monitored by airborne sensors contain variables such as spaceâ€“time coordinates and state parameters. The variables are of various types and vary in different ranges, and there are huge differences among them. In addition, when the aircraft is performing the maneuver, the collected signals show the same or opposite trend of change, and there is highly jumbled information among the feature dimensions. Direct use of original signals to evaluate flight quality will cause serious inductive bias and cannot be applied to the actual flight environment. Its specific performance is as follows:

1.
The value range of different dimensions varies greatly, which makes the evaluation model focus on the dimension with a larger value range. For example, the altitude of air pressure ranges from 0 to 2500Â m, and the elevation angle ranges from 0 to 180Â°. When measured against standard maneuver, the distance depends mainly on barometric height and ignores the influence of pitch angle.

2.
The original value distribution is easy to sample to extreme values, which will lead to low or high ratings of highly or lowly quality regimes.

3.
Information redundancy or correlation between dimensions is high, and the influence of some flight parameters on maneuver quality is ignored.
In order to reduce the impact of the above problems on flight quality assessment, a depth variational autoencoder network is constructed to constrain the distribution of hidden layer features, remove redundant information among features, standardize features of each dimension, make distance measurement calculation more reasonable, and improve the accuracy of scoring.
According to the variational autoencoder principle described in Sect.Â 3.1, a variational autoencoder feature compression network is constructed, as shown in Fig.Â 5. The network is divided into two stages: model training and feature compression During model training, the convolution layer is used to form the encoder of the model, and the multiple time signal input x is accepted and encoded to obtain the hidden variable z (z is constrained by KL divergence), which contains the characteristic information of flight regimes for subsequent quality assessment tasks. Then, the decoder composed of deconvolution uses hidden variables to reconstruct the signal, and takes the mean square error between the reconstructed signal and the original signal as the loss function, and trains the network parameters of the encoder by minimizing the reconstruction error. The specific parameter Settings of encoder and decoder are shown in Table 4.
After the network training is completed, flight parameters are input to the encoder with fixed weights to obtain the hidden layer variables with normal distribution, so as to realize the standardization of features and eliminate redundancy, which is conducive to improving the accuracy of subsequent distance measurement.
6.3 Adaptive dynamic time warping
After variational autoencoder feature compression, the next step is to measure the distance between the compression features of the flight maneuver to be evaluated and the standard maneuver to measure the flight maneuver quality. Common Euclidean distance, Mahalanobis distance and cosine pair distance isometric measures can calculate the similarity between two time series. However, influenced by a variety of external factors during flight, the duration of similar regimes varies greatly and is difficult to predict, which cannot meet the requirements of the common distance measurement method to measure the object scale consistency. Adaptive dynamic time programming calculates distances by means of dynamic programming to find the best matching relationship between sequences. It allows sequence points to be mismatched after selfreplication, can measure nonisometric sequences and is robust to noise. Therefore, this paper uses adaptive dynamic time planning algorithm to measure the similarity between the evaluated maneuver and the standard maneuver.
Suppose that for two timeofflight sequences Xâ€‰=â€‰â€‰<â€‰x_{1}, x_{2}, â€¦, x_{m}â€‰>â€‰, Yâ€‰=â€‰â€‰<â€‰y_{1}, y_{2}, â€¦, y_{n}â€‰>â€‰, and the maneuver duration is m and n respectively, then the path after dynamic state interval planning is Wâ€‰=â€‰â€‰<â€‰w_{1}, w_{2}, â€¦,w_{k}, â€¦, w_{K}â€‰>â€‰, and the calculation process of similarity between X and Y is as follows:

1.
Construct matrix D with size nâ€‰Ã—â€‰m, and element dijâ€‰=â€‰dist(xi, yj) of row I and column, where dist is the distance calculation function and Euclidean distance is usually used by
$${\text{dist}}(x_{i} ,y_{i} ) = \sqrt {\sum\limits_{j = 1}^{n} {(x_{ij}  y_{ij} )^{2} } }$$(11) 
2.
Dynamic path planning is used to search the shortest path from d11 to dnm from matrix D, and the search process must meet the following constraints:

1.
Boundary conditions:w_{1}â€‰=â€‰(1, 1), w_{K}â€‰=â€‰(m, n). The speed of any flight may vary, but the start and end times match. The selected path must start at the lower left corner and end at the upper right corner.

2.
Continuity: if w_{k1}â€‰=â€‰(a_{k1}, b_{k1}), then the next point of the path w_{k}â€‰=â€‰(a_{k}, b_{k}) need to meet (a_{k}a_{k1})â€‰<â€‰â€‰=â€‰1 and (b_{k}b_{k1})â€‰<â€‰â€‰=â€‰1. That is, you can't match across a point, you can only align with your adjacent points. This ensures that each coordinate in X and Y appears in W.

3.
Monotonicity: if w_{k1}â€‰=â€‰(a_{k1}, b_{k1}), then the next point of the path w_{k}â€‰=â€‰(a_{k}, b_{k}) need to meet (a_{k}a_{k1})â€‰>â€‰â€‰=â€‰0 and (b_{k}b_{k1})â€‰>â€‰â€‰=â€‰0. This restricts the W path to be monotonous over time to ensure that matching points do not intersect.
Combined with monotonicity and continuity constraints, the path of any grid point can only be selected in three directions. If the path W has passed the point (i, j), the next grid point can only be the following three cases: (iâ€‰+â€‰1, j), (i, jâ€‰+â€‰1) or (iâ€‰+â€‰1, jâ€‰+â€‰1), as shown in Fig.Â 6.

1.

3.
Search the shortest path from d11 to dnm in matrix D as the similarity of X and Y sequences. Therefore, the total dynamic time planning distance is shown by
$$D_{dtw} (X,Y) = d_{i,j} (x_{i} ,y_{j} ) + \min \left\{ \begin{gathered} D_{dtw} (X_{i  1} ,Y_{j} ) \hfill \\ D_{dtw} (X_{i} ,Y_{j  1} ) \hfill \\ D_{dtw} (X_{i  1} ,Y_{j  1} ) \hfill \\ \end{gathered} \right.$$(12)where iâ€‰=â€‰1, 2, â€¦,m; jâ€‰=â€‰1, 2, â€¦, n; D_{dtw}(X,Y) represents the distance between sequence X and Y, x_{i} and y_{j} represent the points in sequence X and Y respectively, d_{i,j}(x_{i}, y_{j}) represents the Euclidean distance between the points x_{i} and y_{j}.
Computing all the paths in the entire matrix results in enormous time complexity. In order to achieve fast and accurate flight maneuver evaluation, a fixed window is selected to limit the maximum distance of sequence offset. In addition, in order to suppress the phenomenon that abnormal differences at both ends of the sequence lead to too low dynamic time warping distance, an offset is selected to represent the maximum offset of the abnormal points at both ends of the sequence. Finally, for the flight maneuver sequence to be evaluated, the dynamic time planning distance is calculated with each standard flight maneuver of the corresponding category in the standard flight reference base for subsequent flight maneuver quality evaluation.
6.4 Globalâ€“local information multiview fusion scoring rule
When evaluating flight maneuver quality, a reasonable scoring algorithm should take into account both the completion of all indexes and the completion of single specific indexes during flight. If all the indicators of flight regimes are above the standard, the weighted sum is carried out according to the quality of each indicator to obtain the overall score. However, if a single specific index does not meet the requirements, the final score shall not exceed the fixed score. This paper considers both global information and local information, and forms a set of multiview fusion scoring rules based on globalâ€“local information to achieve multiview comprehensive scoring.
For a period of stay evaluation maneuver \(x \in R^{l \times n}\) and k standard regimes of the same class {s_{1}, s_{2}, â€¦, s_{k}}, \(s_{i} \in R^{{l_{i} \times n}}\) for iâ€‰=â€‰1,2,â€¦k, where l and l_{i} represent the duration of flight regimes to be evaluated and each standard flight maneuver respectively, n is the total number of flying parameter. The normalized hidden layer features \(\overline{x} \in R^{l \times z}\) å’Œ \(\{ \overline{{s_{1} }} ,\overline{{s_{2} }} , \ldots ,\overline{{s_{k} }} \}\), \(\overline{{s_{i} }} \in R^{{l_{i} \times z}}\) for iâ€‰=â€‰1,2,â€¦k are obtained by inputting them into variational autoencoder feature compression network, Where z is the independent feature dimension of the longitude variational autoencoder feature compression network after compression. Then the adaptive dynamic time planning algorithm is used to calculate the dynamic time planning distance of each standard maneuver and the maneuver to be evaluated in each dimension d_{z}â€‰=â€‰{d_{1,z}, d_{2,z},â€¦d_{k,z}}, where \(d_{i,z} = {\text{adist}}(\overline{{s_{i} }} (;,z),\overline{x}(;,z))\). Take the minimum value of k distances after calculation as the similarity measure between the maneuver to be evaluated and the standard maneuver benchmark in this dimension D_{z}â€‰=â€‰min(d_{1,z}, d_{2,z},â€¦d_{k,z}).
According to the similarity measure between the maneuver to be recognized and the standard maneuver, the globalâ€“local information fusion multiview comprehensive score is carried out. Two threshold vectors A and B were selected to divide the overall score into three ranges of 100 points, 60â€“100 points and 60 points. Considering global information, if the distance between the maneuver to be recognized and the standard maneuver datum in each characteristic dimension is less than the lower threshold a, the flying maneuver is regarded as 100 points. Considering local information, if there is a distance metric dimensions is greater than the threshold is lower, then the maneuver is less than 100 points, if on the basis of the characteristic dimensions of distance is less than b, upper and lower bounds of the threshold is lower bound by each dimension scores more than threshold of the sum of allowance to decide, and assume that score with various dimensions exceed the threshold is lower total surplus to a linear relationship. The larger the total margin, the lower the score; The smaller the total margin, the higher the score; When the total margin is 0, the score reaches 100 points. If the distance measure on a feature dimension is greater than the upper bound of the threshold, the maneuver is 60 points. The scoring rules are shown in Eq.Â (13).
After simplification, the final flight maneuver quality score can be expressed as Eq.Â (14):
As described above, the quality score is closely related to the choice of upper and lower thresholds. If the threshold value is too high, the algorithm tends to output higher scores, resulting in inflated scores; If the threshold value is too low, the algorithm is too strict and the overall score of flight maneuver quality is low. Therefore, reasonable selection of upper bound b and lower bound a of threshold is crucial to scoring results.
Firstly, considering the sparsity of the number of 100 scores, the lower bound of threshold b is determined by using the standard maneuver reference library. The adaptive dynamic time warping algorithm is used to calculate the similarity measure of each standard maneuver in the standard maneuver base and select their minimum value as the lower threshold. Since the flight parameters of standard regimes are very close to each other, the minimum value between them is taken as the lower bound of threshold value, which not only ensures that the 100point regimes meet the requirements of standard regimes, but also has a high sparseness. Secondly, the upper bound of threshold a divides the boundaries of two intervals of 60â€‰~â€‰100 and 60Â min. It is assumed that all kinds of flight regimes obey the Gaussian distribution, and the standard flight regimes are distributed near the mean of the Gaussian distribution. The maximum similarity measure between standard regimes in the standard maneuver base is taken as variance Ïƒ of gaussian distribution. According to the 3Ïƒ principle, the probability of flight maneuver at a distance of more than 3Ïƒ is almost zero. Therefore, 3 times of the maximum similarity measure between each standard maneuver is selected as the upper bound of threshold a to ensure the sparsity of 60min flying maneuver. To sum up, the lower bound of threshold b and the upper bound of threshold a are given by
Combined with the above, the overall process of multi view fusion scoring algorithm based on global local information is shown in Table 5.
7 Simulation results
Because the depth convolution autoencoder imposes an independent and identically distributed a priori on the hidden layer features, the features after dimension reduction by depth convolution autoencoder often approximately obey the standard normal distribution of independent and identically distributed, which can eliminate the scale difference between dimensions and is beneficial to the distance calculation between features.
When different dimensionality reduction dimensions are selected, it will affect the training effect of the network. When the dimension is too low, the feature information is missing and the original signal cannot be reconstructed effectively; When the dimension is too high, redundancy inevitably occurs between dimensions, breaking the independence between different features and affecting the performance of the scoring model. Therefore, we tested the overall distribution of scores obtained under different feature dimensions to achieve super parameter optimization. By comparing the score distribution results under five groups of feature dimensions [5, 6, 8, 10, 12], when the feature dimension is greater than or equal to 10, the scores of some regimes obviously tend to be low segmentation, indicating that the autoencoder model will produce some invalid dimensions at this time; When the feature dimension is less than or equal to 6, the score of some regimes obviously tends to be high segmentation, indicating that the network is difficult to converge and the feature discrimination is insufficient. Therefore, the feature dimension selected by the final scoring model is 8, as shown in Fig.Â 7.
8 Conclusions
In this paper, a scheme of deep learning for accurate recognition of high maneuvering flight regimes in military flight is proposed for the first time. Make full use of the neural network, which has the characteristics of high invariance to the flight parameter data of multivariate time series, such as scaling, translation, tilt and other forms of deformation. Combined with the spatial mapping ability of variational autoencoder, the flight maneuver segmentation and accurate evaluation are realized. At the same time, it effectively solves the practical problem of scientific evaluation of training quality in flight training. Based on the accurate identification and quality evaluation of flight regimes, the concept of building a basic database of flight quality for flight cadets is proposed, and the indepth application level analysis and mining of flight training big data is expected to solve the specific engineering practical problems of the management department on specific air regimes, such as "whether to fly, how to fly quality, how to express and locate Flight points", It can fill the gap that the current military training lacks effective mining and analysis means for flight training data.
By establishing a comprehensive evaluation method system for flight quality of flight cadets, we can give play to the advantages of big data and analyze the overall law of flight training from multiple dimensions in followup research and practice. From the individual dimension of flight cadets, comprehensively study and judge the technical maturity curve formed by the factors such as technical maturity, stability and isolated regimes of a single flight Cadet during flight training, so as to provide auxiliary data support for the personalized and accurate training of flight cadets; From the group dimension of flight cadets, mine and analyze the flight training quality law of characteristic groups, summarize and refine the advantageous flight training regimes and group training difficulties, form auxiliary data analysis and decisionmaking, and support the reform of flight training; From the dimension of flight instructors, systematically mine and analyze the overall flight quality, advantageous regimes and isolated regimes of flight instructors, summarize experience and avoid weaknesses through data, and support the overall improvement of teaching ability and level of flight instructors.
Availability of data and materials
The datasets used and analysed during the current study are available from the corresponding author on reasonable request.
References
X. Liu, X.B. Zhai, W. Lu, C. Wu, QoSguarantee resource allocation for multibeam satellite industrial internet of things with NOMA. IEEE Tran. Ind. Inf. 17(3), 2052â€“2061 (2019). https://doi.org/10.1109/TII.2019.2951728
X. Liu, X. Zhang, Rate and energy efficiency improvements for 5Gbased IoT with simultaneous transfer. IEEE Internet Things J. 6, 5971â€“5980 (2019). https://doi.org/10.1109/JIOT.2018.2863267
C. Xie, S.H. Ni, Z.L. Zhang, Y.H. Wang, Recognition method of acrobatic maneuver based on state matching and support vector machines. J. Project. Rockets Missiles Guidance 3, 240â€“242 (2004). https://doi.org/10.15892/j.cnki.djzdxb.2004.s3.015
C. Xie, S.H. Ni, Z.L. Zhang, Pattern attributes extraction of flight data based on rough set. Comput. Eng. 12, 169â€“171 (2005). https://doi.org/10.3969/j.issn.10003428.2005.12.062
S.H. Ni, Z.K. Shi, C. Xie, Y.H. Wang, Establishment of avion inflight maneuver action recognizing knowledge base. Comput. Simul. (2005). https://doi.org/10.3969/j.issn.10069348.2005.04.007
R.F. Zhang, S.H. Ni, The algorithm of flight trajectory creatation based on Cardinals cubic splines. Calc. Technol. Autom. S2, 109â€“111 (2006). https://doi.org/10.3969/j.issn.16734599.2005.01.009
W.J. Yin, S.H. Ni, A method of recognizing flight maneuver based on genetic algorithm. Comput. Dev. Appl. (2006). https://doi.org/10.3969/j.issn.10035850.2006.04.008
C. Su, S.H. Ni, Y.H. Wang, Method of rule acquirement of flight state based on improved AIS. Comput. Eng. Appl. 47(3), 237â€“239 (2011). https://doi.org/10.3778/j.issn.10028331.2011.03.069
Y. Gao, S.H. Ni, Y.H. Wang, P. Zhang, A flight state rule extraction method based on improved quantum genetic algorithm. Electroopt. Control 18(01), 28â€“31 (2011). https://doi.org/10.3969/j.issn.1671637X.2011.01.007
Y.W. Wang, Y. Gao, A rule extraction method for flight action recognition based on whale optimization algorithm. J. Naval Aeronaut. Astronaut. Univ. 33(05), 447451+498 (2018). https://doi.org/10.7682/j.issn.16731522.2018.05.005
Y.W. Wang, Y. Gao, Research on complex motion recognition method based on basic flying motion. Ship Electron. Eng. 38(10), 74â€“76 (2018). https://doi.org/10.3969/j.issn.16729730.2018.10.018
R. Agrawal, C. Faloutsos, A. Swami, Efficient Similarity Search in Sequence Databases (Springer, Berlin, 1993)
R.A.K. Lin, H.S.S.K. Shim, Fast similarity search in the presence of noise. Scaling. and translation in time. Series Databases. In: Proceedings of the 21 St Int'1 Conference on Very Large Databases, Zurich. Switzerland, pp. 490â€“501(1995)
Y. Cai, R. Ng, Indexing spatiotemporal trajectories with Chebyshev polynomials. In: Proceedings of the 2004 ACM SIGMOD International Conference on Management of Data. ACM, pp. 599â€“610 (2004)
F. Korn, H.V. Jagadish, C. Faloutsos, Efficiently supporting ad hoc queries in large datasets of time sequence. ACM SIGMOD Rec. 26(2), 289â€“300 (1997). https://doi.org/10.1145/253262.253332
T. Pavlidis, S.L. Horowitz, Segmentation of plane curves. IEEE Trans. Comput. 23(8), 860â€“870 (1974). https://doi.org/10.1109/TC.1974.224041
D.J. Berndt, J. Clifford, Using dynamic time warping to find patterns in time series. Working Notes of the Knowledge Discovery in Databases Workshop. Seatle, pp. 359â€“370 (1994)
A. Olemskoi, S. Kokhan, Effective temperature of selfsimilar time series: analytical and numerical developments. Physica A 360(1), 37â€“58 (2006). https://doi.org/10.5488/CMP.8.4.761
H.L. Li, Z. Shan, H.R. Guo, Flight action recognition algorithm based on MDTW. Comput. Eng. Appl. 51(09), 267â€“270 (2015). https://doi.org/10.3778/j.issn.10028331.13070018
Y.C. Shen, S.H. Ni, P. Zhang, Flight action recognition method based on Bayesian network. Comput. Eng. Appl. 53(24), 161167+218 (2017). https://doi.org/10.3778/j.issn.10028331.17070156
C. Zhou, R. Fan, G. Zhang, Z.Y. Huang, A flight action recognition based on multivariate time series fusion. J. Air Force Eng. Univ. (Natural Science Edition) 18(04), 34â€“39 (2017). https://doi.org/10.3969/j.issn.10093516.2017.04.007
Y.C. Shen, S.H. Ni, P. Zhang, Similarity query method for flight data time series subsequences. J. Air Force Eng. Univ. (Natural Science Edition) 20(02), 7â€“12 (2019)
Y.P. Xiao, Q.Q. Fu, A review of the research on lateral direction flight quality assessment methods. J. Civ. Aviat. 2(05), 42â€“45 (2018)
Y.W. Wang, Y. Gao, Comprehensive evaluation method of UAV flight quality based on comprehensive weighting method. Ordnance Ind. Autom. 38(05), 14+10 (2019). https://doi.org/10.7690/bgzdh.2019.05.001
G. Xu, J.M. Liu, An automatic evaluation method for military simulation training result. Fire Control Command Control 45(05), 162â€“169 (2020). https://doi.org/10.3969/j.issn.10020640.2020.05.030
S. Xu, Y. Wang, L.R. Bai, Overview of military helicopter training effectiveness based on simulation. Helicopter Tech. 03, 58â€“62 (2020)
Funding
This work was supported by the Aviation Science Fund under Grant 2020Z014066001, and the 13th special grant of the China Postdoctoral Science Fund under Grant 2020T130181.
Author information
Authors and Affiliations
Contributions
All authors have contributed toward this work as well as in compilation of this manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Tian, W., Zhang, H., Li, H. et al. Flight maneuver intelligent recognition based on deep variational autoencoder network. EURASIP J. Adv. Signal Process. 2022, 21 (2022). https://doi.org/10.1186/s1363402200850x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s1363402200850x