Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
As described in the background art, most of the existing radar target identification methods belong to conventional target identification methods, and targets to be identified are treated equally, that is, the importance of the targets is considered to be the same. In this case, it is an object to improve the correct recognition rate. However, the conventional target identification method has a problem that the different importance (different threat degrees) of the target and the possible risks of misclassification of the target are not considered.
The invention adopts a cost-sensitive target identification method, and because the importance (namely threat degree) of each target to be identified is different, the loss possibly brought by the error classification of different targets is also different, so the cost-sensitive target identification method is more consistent with the actual situation. Cost sensitive target identification seeks to reduce the overall target misidentification cost such that the risk or loss due to misidentification is minimized.
The method simultaneously considers the test cost and the misclassification cost to select and determine the optimal feature subset for classification, constructs the feature vector of the sample to be classified, trains the extreme learning machine, and realizes the radar target identification with high performance and minimum misclassification cost by utilizing the trained extreme learning machine.
In order to make the aforementioned objects, features and advantages of the present invention comprehensible, embodiments accompanied with figures are described in further detail below.
Example 1:
as shown in fig. 1, the present invention provides a radar target identification method inembodiment 1, where S1-S6 correspond to the steps in the method, and the method includes:
s1, preprocessing radar target characteristic information collected by different sensors and then fusing to obtain characteristic vectors for characteristic classification;
s2, determining an objective function for feature selection; the target function considers the misclassification cost and the testing cost at the same time;
s3, minimizing the objective function by using an HO optimization method, and obtaining an optimal feature subset according to an optimization result;
s4, generating a target data set according to the optimal feature subset;
s5, training the classifier by using the target data set; if the training error is within a preset range, obtaining a trained classifier;
and S6, directly utilizing the trained classifier to identify the target to be identified in the unknown category.
The following is a detailed description of the above steps:
object recognition and cost-sensitive classification
The target identification of the invention belongs to cost-sensitive target identification, namely the costs generated by error classification are different, and the essential problem is cost-sensitive classification.
In the invention, in order to realize optimization in the process of feature selection, a newly proposed optimization method, namely hunting optimization, is adopted; considering that the cost sensitive target identification has test cost and misclassification cost, a new feature selection target function is constructed by considering two types of cost simultaneously during feature selection, so that a new cost sensitive feature selection method is provided; and (4) using the preferred features to form feature vectors, and using an extreme learning machine as a classifier to realize the classification of the cost-sensitive target feature data.
HO optimization algorithm
The HO optimization algorithm stems from human heuristics for hunting activities. In human hunting activities, a hunter and several hunting dogs form a group, and they work in concert to search for and catch game prey. An extension based on such a single group is that several groups cooperate together to capture the most valuable (or as many as possible) game within a certain hunting horizon, similar to finding a globally optimal solution defined by a specific objective function in an optimization problem. Based on the analysis of the hunting activities, a new optimization model called Hunting Optimization (HO) was proposed.
First, assuming that a single hunting group consists of one hunter and a certain number of hunting dogs, the visual field distance of the hunter determines the hunting range (hunting field) of the group, a conceptual diagram of a hunting group with three hunting dogs is shown in fig. 2.
Each panel mainly contained the following behaviors:
(1) and (6) searching. Each hunter randomly releases the dog to which he belongs to his hunting field for searching, as shown in figure 2. If one of the game dogs found a location with more game parts, the game player moves to the location where the game dog is located and begins searching again around the new location. Also, as the game dog becomes more fatigued, the hunter will reduce his field of view to shrink the game. When the hunting dogs were exhausted, the hunting field became small and the group fell into local optima, at which time the exhausted hunting dogs were replaced with new ones to revive the group.
(2) Follow and gather. In addition to searching its local hunting field, each team always exchanges information with other teams in the population to follow and search the locations where the whole population finds the most animals. That is, each team moves not only toward its local optimum point, but also a distance toward the globally optimum (most game). By this following action, the entire population is gradually gathered together, which enables the entire population to converge to the same location in the end.
(3) Competition. In the hunting process, the group with the fewest prey has a high probability of being in the wrong hunting direction. Thus, this group would be eliminated and its game dogs would be assigned to other game players. This strategy can balance global and local searches. At the beginning of the search, there were more hunters and fewer dogs per hunter, thus a better global search; in the later stages of the search, fewer hunters remain, and each hunter has more hunters and therefore has better local search.
(4) And (4) recombining. When two subgroups are close together, they may repeatedly search the same area, which is very inefficient. Thus, their hunting grounds will be redistributed. This strategy ensures that each team resource is fully utilized as much as possible.
Through the above behaviors, each group searches the hunting ground of the group and simultaneously cooperates with other groups to search the area with the most prey. The above process is also an optimization process, and the search space represents a feasible solution for hunting field, hunter and hounds, and the fitness value of a specific objective function represents the number or value of the prey.
For a d-dimensional minimization optimization problem that only considers boundary constraints, it can be expressed as:
where f (x) is an objective function, x ═ x
1,x
2,…x
d]Is an arbitrary feasible solution, S ═ x
i|l
i≤x
i≤u
iI-1, 2, … d is a non-empty finite set, l
i,u
iIs i
thThe upper and lower bounds of the dimension. The objective of solving equation (1) is to find a global optimal solution x
*So that f (x)
*)≤f(x),
First, the definition parameters are defined as follows: n is a radical ofmax,NminMaximum minimum number of hunters, Qmax,QminMaximum minimum number of hunting dogs; r ismax,rminA maximum minimum field-of-view attenuation factor; x is the number ofiHunters or hunting dogs; psi0An original search space; psii xiThe search space of (2); a τ congestion factor; d*A local optimal solution; n is a radical ofiteNumber of hunters at the second iteration; qiteNumber of hounds at the ite iteration; v. ofi-hunter's visual field; v. ofrThe elimination rate; gBest global optimal solution.
The hunting optimization HO algorithm process is as follows:
andStep 1, initializing.
Initializing N
max,N
min,Q
max,Q
min,r
max,r
min∈(0,1),ψ
0={ψ
i|l
i≤ψ
i≤u
iI ═ 1,2, … d } and τ. Initializing all hunters
i=1,2,…,N
0Wherein
j-1, 2, …, d, rand is a uniformly distributed random number with a value between 0 and 1, and N
0=N
max. For all x
iInitializing search field of view with original search range
Namely, it is
j is 1,2, …, d. And calculating the fitness of all hunters, and recording the optimal solution gBest.
AndStep 2, eliminating.
To balance global and local searches, the number of hunters is gradually reduced and the number of hunting dogs is gradually increased during the search. That is, in each cycle, a portion of hunters with the worst fitness values will be eliminated and their dogs will be assigned to hunters with better fitness values. To ensure that there are enough hunters in the beginning search phase to ensure a global search and enough dogs in the end phase to ensure a local search, a cosine function is used to control the culling ratio:
vr0.5 cos (ite pi/maxIte) +0.5 (formula 2)
Adjusting N based on (equation 2)iteAnd Qite:
I.e., in the second ite cycle, Nite-Nite-1The hunters are eliminated and the number of hunting dogs of each hunter is from Qite-1Is changed into Qite。
Step3. recombination.
If a hunter xiVery close to other hunters or xiBecomes very small, xiShould be recombined. Since all hunters are randomly distributed throughout the search space, the probability that two hunters are in the same position is small; on the other hand, x is continuously deeper as the search progresses, since all subgroups dynamically move towards gBest, which attenuates the field of view while keeping the position unchanged until a new optimal solution is found (see Step 4-5)iAnd gBest gradually overlap. In order to simplify the HO,checking only xiAnd gBest, i.e. if xiAnd gBest is less than a predetermined congestion factor, xiWill be re-randomized to a new position with its field of view initialized to the original search range:
furthermore, if viVery small, xiWill fall into local optima and cannot jump out, at this point, v will beiResetting as the product of the original search range and the attenuation factor of the current search stage:
recombination is highly necessary in the search process, which ensures that the search potential of each subgroup is exploited as much as possible, reducing invalid or inefficient searches.
And Step 4, searching.
Each hunter x remainingiAttenuating its field distance by:
vi=vi*r,i=1,2,…Nite(formula 6)
Then each hunter passes viAnd current position xiUpdating the hunting field:
as can be seen from (equation 7), xiIs xiA central hypercube. Since r < 1, the distance of the field of view becomes smaller and smaller in each cycle according to (equation 6). This results in the search space in (equation 7) becoming more and more compact. This allows each hunter to focus on searching for local areas around it.
After updating the search space, xiRandom use of QiteIndividual hunting dog search psiiQ in (1)iteA position di,i=1,2,…,QiteThen find out the one with the best fitness, and mark as d*. If f (d)*) < f (gBest), gBest and xiQuilt d*And (6) replacing, directly turning to Step6. If f (d)*) F (gBest) and f (d)*)<f(xi) Hunter will move to d*I.e. xiQuilt d*And (6) replacing. Otherwise, xiRemain unchanged.
Step 5, follow.
After the search process is complete, each hunter moves a small step distance toward gBest by:
xi=xi+(gBest-xi) (1-r) (equation 8)
Through this process, each hunter perceives and learns the search results of the other hunters. This process can be considered as the process of information exchange between different hunters. In the current cycle, each hunter always moves from the local optimal solution of the hunter to the global optimal solution while searching, so that the whole population is guaranteed to be always converged to an optimal solution, which is important for the global convergence.
The above search and follow-up process can be summarized as:
xi=xi+(d*-xi)+(gBest-d*) (1-r) (equation 9)
From (equation 6) and (equation 9), it can be seen that r is used to control the decay rate of the field of view and the speed of movement towards the optimal solution. To balance the global and local search, the variation of r is controlled by a nonlinear decay function:
rite=rmax-(rmax-rmin)*(ite/maxite)2(formula 10)
For viA larger r is advantageous for finding more promising regions by global search; a smaller r will cause the panel to focus on searching the local search space around the hunter. For the step size of the move in (equation 9), a larger r favors the subgroupSearching a local search space where the local search space is located; conversely, a smaller r allows the population to converge quickly to the globally optimal solution. Thus, (equation 10) can balance not only the global and local searches, but also the convergence speed and premature convergence.
And (5) terminating.
If the termination condition is met, the algorithm is ended, and the global optimal solution is used as final output; otherwise, go toStep 2.
HO-based cost sensitive feature selection method
In the target identification, the test cost and the misclassification cost are simultaneously considered, namely the total cost comprises two parts: test costs and misclassification costs. The misclassification cost refers to the loss of samples due to misclassification. The cost of testing refers to how much time, money, etc., is spent obtaining the characteristics of a certain sample.
In the invention, cost sensitive feature selection is carried out, and the aim is to obtain a feature subset with the minimum total cost by balancing test cost and misclassification cost. I.e. to determine a subset of features that have a smaller average test cost and that achieve better classification performance.
In the present invention, it is assumed that all test costs are independent of each other. For a classification problem with d-dimensional features, if Fe is the original feature set, the test cost can be expressed as a d-dimensional vector tc ═ tc
1,tc
2,…tc
D]Wherein tc
kIs the cost of the kth feature. For arbitrary feature subsets
The total test cost equals the sum of all individual feature test costs in B:
for an m-class classification problem, the misclassification cost information is usually represented by a cost matrix:
wherein c isijIndicating the cost of identifying class i as class j. The cost matrix should satisfy the following conditions:
(1) the cost of misclassification should always be greater than the cost of correct classification. Namely to
Is provided with c
ij>c
ii. Only under this condition is cost sensitive learning meaningful.
(2) To pair
Cannot always have c
ij≥c
ikI is 1, …, m. If this condition is violated, class j is dominant and the sample cannot be classified into that class. In this case the presence of class j is meaningless.
For a data set I { (x) containing N samples
i,y
i)|x
i∈R
d,y
iE {1,2, …, m }, i ═ 1,2, …, N }, the total misclassification cost equals the sum of the classification costs of all samples on feature set B:
wherein
Is a sample x
iThe prediction category of (a) is determined,
denotes a number y
iClass is identified as
The cost of the class.
The process of selecting the cost sensitive features is an optimization problem, the method simultaneously considers the misclassification cost Tc (B) and the test cost Fc (B) to construct an optimized objective function, adopts a newly proposed optimization method HO to carry out optimization, and determines the selected features according to the HO optimization result.
The HO-based cost sensitive feature selection method comprises the following process steps:
step 1. define the objective function.
An objective function that considers both the misclassification cost and the testing cost is defined as follows:
f (b) ═ fc (b) + λ tc (b) (formula 15)
Where λ is the balance factor. When λ is small, the size of f (b) is determined by fc (b), and the algorithm is primarily focused on reducing the overall test cost. When λ is large, the size of f (b) is determined by tc (b), and the algorithm is mainly focused on reducing the total misclassification cost.
AndStep 2, optimizing the constructed objective function by using HO, and minimizing the value of the objective function.
Step3. real code of HO is converted into binary code using a step function.
Since HO uses a real number encoding method for an individual, it cannot be directly applied to feature selection (binary optimization problem). To take advantage of HO's advantages in real number optimization problems, the original search space is limited to [ -1,1], while each real coded individual is converted to a binary vector using a step function:
in the binary vector, a 1 indicates that the corresponding feature is selected, otherwise the corresponding feature is not selected.
For example, before feature selection, the feature dimension is 5, and after HO algorithm optimization, one individual is obtained as x ═ 0.2, -0.3,0.4,0.1, -0.5, and after step function processing, a corresponding binary vector is obtained as sf ═ 1,0,1,1,0, and this vector represents that the 1 st, 3 rd, and 4 th features are selected to constitute a feature subset.
In the process of selecting the characteristics, the innovation point of the method is mainly embodied in two stages ofStep 1 andStep 2.
In thestage 1, the misclassification cost and the test cost are simultaneously considered when the objective function of the optimization problem is constructed, the problems that only the misclassification cost or only the test cost is considered in the traditional method are solved, the characteristics of the problems are better met, and a better result can be obtained.
In the 2 nd stage, a newly proposed optimization model HO is adopted for optimization, so that a better result can be obtained more quickly.
Data classification method based on cost sensitive feature selection
And meanwhile, constructing a target function of the feature selection optimization problem by considering the misclassification cost and the test cost, optimizing by adopting HO, and obtaining the selected features according to the optimization result.
On the basis of an original data set, a target data set can be generated according to an optimization result of feature selection, namely the selected features, and the data set is divided into a training data set and a testing data set for training a classifier and evaluating the performance of the classifier.
In the present invention, an Extreme Learning Machine (ELM) is used as a classifier for target recognition. The basic network structure of the ELM is a 3-layer structure as shown in fig. 3, and is composed of an input layer, a hidden layer, and an output layer, and nodes between the two layers are all connected.
The basic idea of ELM is to initialize and fix the parameters of the hidden layer randomly without iterative adjustment, then transform the input data from the input space to the high-dimensional hidden space by ELM feature mapping, and solve the output parameters by ELM learning algorithm.
Supposing that N training samples are arranged, and classifying the training samples into m categories;
the training sample set is
Wherein xi=[xi1,xi2,...,xid]TIs the ith training sample, d is the dimensionality of the data, ti=[ti1,ti2,...,tim]TTarget output for the ith sample. The ELM network outputs are:
wherein (w)i,bi) The ith implicit node parameter, W ═ W1,w2,...,wL) Is weighted from the hidden layer, b ═ b1,b2,…,bL) Is the hidden layer bias. g (-) is an activation function, such as Sigmoid, Tanh, etc.
As shown in fig. 4, to implement cost-sensitive classification and target identification, the steps of the present invention are as follows:
A. training phase
Step 1, acquiring characteristic information of a target, fusing the characteristic information from different sensors, and carrying out normalization preprocessing on characteristic values of all dimensions to form a characteristic vector for characteristic classification.
And 2, determining the testing cost, the misclassification cost and the balance factor.
And 3, selecting cost sensitive features, and determining an optimal feature subset for classification. The specific process is divided into the following three stages:
3.1. and constructing an objective function according to the specific cost sensitive characteristic selection problem. The objective function considers both misclassification costs and testing costs.
3.2. The constructed objective function is optimized by using a Hunting Optimization (HO) method.
3.3. The selected features are determined based on the results of the HO optimization. The real code of HO needs to be converted into binary code using a step function to determine the features.
And 4, generating a target data set according to the selected characteristics, wherein the target data set comprises a training data set and a testing data set, the training data set is used for training the classifier, and the testing data set is used for evaluating the performance of the classifier.
And 5, training a target recognition classifier. An extreme learning machine is used as a classifier, and the classifier is trained by using a training data set. The training algorithm of the extreme learning machine is as follows:
inputting: training sample
Number of hidden layer nodes L, activation function g (·).
And (3) outputting: the hidden layer outputs weights β.
5.1 randomly generating hidden layer weight values and bias (W, b);
5.2 calculating hidden layer output H according to the following formula;
5.3 calculating the hidden layer output weight beta according to the following formula.
B. Identification phase
Step 1, acquiring characteristic information of a target, fusing the characteristic information from different sensors, and carrying out normalization preprocessing on characteristic values of all dimensions to form a characteristic vector for characteristic classification.
And 2, according to the result of the cost sensitive feature selection, forming a feature vector by the selected features.
And 3, identifying the unknown class target by using the trained ELM, namely calculating the output of the network by using the formula 17, thereby obtaining the class of the unknown target.
By means of the feature selection considering the misclassification cost and the test cost at the same time, the problem that only the misclassification cost or only the test cost is considered in the traditional method is solved, the characteristics of the problem are better met, and a better result can be obtained; in the optimization of feature selection, the optimization is carried out by adopting the provided optimization method HO, so that a better result can be obtained more quickly.
In conclusion, for cost-sensitive classification and target identification, the technical method can realize classification and target identification with high performance and minimum misclassification cost. Example 2:
as shown in fig. 5, the present invention further provides, inembodiment 2, a radar target recognition system implemented based on the radar target recognition method described inembodiment 1, where the system includes:
the system comprises afeature fusion module 1, a cost sensitivefeature selection module 2, a training and testdata generation module 3, a classifier training module 4 and aclassification module 5.
Thefeature fusion module 1 is used for preprocessing radar target feature information collected by different sensors and then fusing the preprocessed radar target feature information to obtain feature vectors for feature classification;
the cost sensitivefeature selection module 2 is used for determining an objective function for feature selection; the target function considers the misclassification cost and the testing cost at the same time; minimizing a target function by using an HO optimization method, and obtaining an optimal characteristic subset according to an optimization result;
the training and testingdata generation module 3 is used for generating a target data set according to the optimal feature subset;
the classifier training module 4 is used for training a classifier by using a target data set; if the training error is within a preset range, obtaining a trained classifier;
theclassification module 5 is used for directly utilizing the trained classifier to identify the target to be identified in the unknown class.
Because the system utilizes the radar target identification method as described inembodiment 1 of the present invention, after the system modules in the embodiment construct and train classifiers meeting the requirements, the trained classifiers can be directly utilized to identify targets of unknown types.
In summary, the radar target identification method and the radar target identification system provided by the invention can effectively solve the problem of high risk of radar target identification error in the background art, and can realize target identification with high performance, high precision and minimum error cost on the basis of simultaneously considering test cost and error cost.
The principle and the implementation of the present invention are explained in the present text by applying specific examples, and the above description of the examples is only used to help understanding the method and the core idea of the present invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, the specific embodiments and the application range may be changed. In view of the above, the present disclosure should not be construed as limiting the invention.