Movatterモバイル変換


[0]ホーム

URL:


CN107154024A - Dimension self-adaption method for tracking target based on depth characteristic core correlation filter - Google Patents

Dimension self-adaption method for tracking target based on depth characteristic core correlation filter
Download PDF

Info

Publication number
CN107154024A
CN107154024ACN201710355456.3ACN201710355456ACN107154024ACN 107154024 ACN107154024 ACN 107154024ACN 201710355456 ACN201710355456 ACN 201710355456ACN 107154024 ACN107154024 ACN 107154024A
Authority
CN
China
Prior art keywords
mrow
target
msup
scale
msub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710355456.3A
Other languages
Chinese (zh)
Inventor
刘忠耿
练智超
濮柯佳
李杨
张伟
李敏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Science and Technology
Original Assignee
Nanjing University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Science and TechnologyfiledCriticalNanjing University of Science and Technology
Priority to CN201710355456.3ApriorityCriticalpatent/CN107154024A/en
Publication of CN107154024ApublicationCriticalpatent/CN107154024A/en
Pendinglegal-statusCriticalCurrent

Links

Classifications

Landscapes

Abstract

The invention discloses a kind of dimension self-adaption method for tracking target based on depth characteristic core correlation filter.This method comprises the following steps:The convolutional neural networks of pre-training completion are input an image into, depth convolution feature is extracted;Target tracking, using the model of training, estimates position and the yardstick of target;According to the target location of current detection and yardstick, core correlation filter is trained;Using the model update method of adaptive high confidence level, core correlation filter is updated.The present invention is extracted depth convolution feature, using adaptive scale method of estimation and the model modification strategy of adaptive high confidence level, improve the robustness of target following of the target in complex scene and cosmetic variation, target scale change can efficiently and accurately be handled, in addition, due to the model modification strategy using adaptive high confidence level, model following drift has been reduced as far as.

Description

Translated fromChinese
基于深度特征核相关滤波器的尺度自适应目标跟踪方法Scale Adaptive Target Tracking Method Based on Deep Feature Kernel Correlation Filter

技术领域technical field

本发明涉及计算机视觉技术领域,特别是一种基于深度特征核相关滤波器的尺度自适应目标跟踪方法。The invention relates to the technical field of computer vision, in particular to a scale-adaptive target tracking method based on a deep feature kernel correlation filter.

背景技术Background technique

近年来,随着大规模标注数据集的出现,以及计算机计算能力的提升,深度学习方法尤其是卷积神经网络成功应用于图像分类、目标检测、目标识别以及语义分割等计算机视觉领域,这主要归功于卷积神经网络强大的目标表示能力。与传统的图像特征不同,深度卷积特征从大量的数千个类别的图像数据学习得到,所以高层的卷积特征表示目标的语义特征,适用于图像分类问题。由于高层的卷积特征分辨率很低,所以并不适合确定目标的位置,而且由于训练数据的缺失,在跟踪开始的前几帧训练一个深度模型困难重重。In recent years, with the emergence of large-scale labeled data sets and the improvement of computer computing power, deep learning methods, especially convolutional neural networks, have been successfully applied to computer vision fields such as image classification, target detection, target recognition, and semantic segmentation. Thanks to the powerful object representation ability of convolutional neural network. Different from traditional image features, deep convolutional features are learned from a large number of thousands of categories of image data, so high-level convolutional features represent the semantic features of the target and are suitable for image classification problems. Due to the low resolution of high-level convolutional features, it is not suitable for determining the position of the target, and due to the lack of training data, it is difficult to train a deep model in the first few frames of tracking.

最近,基于相关滤波器的判别式跟踪方法由于跟踪效果高效而准确,引起很多研究者的兴趣。基于相关滤波器的跟踪方法通过将输入特征回归为目标高斯分布在线训练一个相关滤波器,并在后续的跟踪过程中寻找相关滤波器输出相应图谱的峰值确定目标的位置。相关滤波器在运算中巧妙应用快速傅里叶变换,降低了计算复杂度,使得跟踪速度大幅度提升。但是,核相关滤波器算法使用传统的梯度方向直方图特征,当目标外观表示发生变化时,跟踪算法容易漂移;另外,该算法不能估计目标的尺度变化,而且采用简单的线性插值更新模型,这样当目标追踪发生错误的时候,由于更新机制的最终导致跟踪算法漂移。Recently, the discriminative tracking method based on correlation filter has aroused the interest of many researchers because of its efficient and accurate tracking effect. The tracking method based on the correlation filter trains a correlation filter online by regressing the input features to the Gaussian distribution of the target, and finds the peak of the corresponding spectrum output by the correlation filter in the subsequent tracking process to determine the position of the target. The correlation filter cleverly applies the fast Fourier transform in the operation, which reduces the computational complexity and greatly improves the tracking speed. However, the kernel correlation filter algorithm uses the traditional gradient direction histogram feature, and the tracking algorithm is prone to drift when the appearance representation of the target changes; in addition, the algorithm cannot estimate the scale change of the target, and uses simple linear interpolation to update the model, so When an error occurs in target tracking, the tracking algorithm drifts due to the update mechanism.

发明内容Contents of the invention

本发明的目的在于提供一种基于深度特征核相关滤波器的尺度自适应目标跟踪方法,从而提高目标在复杂场景和外观变化中的目标跟踪的鲁棒性,高效准确地处理目标尺度变化,尽可能地减少因错误检测而导致的模型跟踪漂移。The purpose of the present invention is to provide a scale-adaptive target tracking method based on the depth feature kernel correlation filter, thereby improving the robustness of target tracking in complex scenes and appearance changes, and efficiently and accurately processing target scale changes, as much as possible Potentially reduce model tracking drift due to false detections.

实现本发明目的的技术解决方案为:一种基于深度特征核相关滤波器的尺度自适应目标跟踪方法,包括以下几个步骤:The technical solution to realize the object of the present invention is: a scale-adaptive target tracking method based on depth feature kernel correlation filter, comprising the following steps:

步骤1,输入目标的初始位置p0和尺度s0,设置窗口大小为目标初始包围盒的2.0倍;Step 1, input the initial position p0 and scale s0 of the target, and set the window size to 2.0 times the initial bounding box of the target;

步骤2,根据第t-1帧的目标位置pt-1,获得目标区域xt-1,尺寸为窗口大小;Step 2, according to the target position pt-1 in the t-1th frame, obtain the target area xt-1 , whose size is the window size;

步骤3,提取目标区域xt-1的深度卷积特征,并进行快速傅立叶变换,得到特征图谱其中^表示离散傅立叶变换;Step 3, extract the deep convolution feature of the target area xt-1 , and perform fast Fourier transform to obtain the feature map Wherein ^ represents the discrete Fourier transform;

步骤4,根据特征图谱计算核自相关Step 4, according to the feature map Computing Kernel Autocorrelation

步骤5,训练位置和尺度相关滤波器;Step 5, training location and scale correlation filters;

步骤6,根据第t-1帧目标的位置pt-1,获得目标在第t帧的候选区域zt,尺寸为窗口大小;Step 6, according to the position pt-1 of the target in the t-1th frame, obtain the candidate area zt of the target in the t-th frame, and the size is the window size;

步骤7,提取候选区域zt的深度卷积特征,并进行快速傅立叶变换,得到特征图谱其中^表示离散傅立叶变换;Step 7, extract the depth convolution feature of the candidate area zt , and perform fast Fourier transform to obtain the feature map Wherein ^ represents the discrete Fourier transform;

步骤8,根据目标前一帧的特征图谱计算核互相关Step 8, according to the feature map of the previous frame of the target Computing Kernel Cross-Correlation

步骤9,分别检测位置滤波器和尺度滤波器的输出图谱中最大值对应的位置,来确定目标在当前帧的位置pt和尺度stStep 9, detecting the position corresponding to the maximum value in the output spectrum of the position filter and the scale filter respectively, to determine the position pt and scalest of the target in the current frame;

步骤10,采用自适应的模型更新策略更新核相关滤波器。In step 10, an adaptive model updating strategy is used to update the kernel correlation filter.

进一步地,步骤3和步骤7所述深度卷积特征的提取方法,具体如下:Further, the extraction method of the depth convolution feature described in step 3 and step 7 is as follows:

(3.1)预处理,将窗口区域I缩放到卷积神经网络规定的输入尺寸224×224;(3.1) Preprocessing, scaling the window area I to the input size 224×224 specified by the convolutional neural network;

(3.2)特征提取,提取卷积神经网络第3、4和5个卷积层的特征图谱;(3.2) Feature extraction, extracting the feature maps of the 3rd, 4th and 5th convolutional layers of the convolutional neural network;

(3.3)双线性插值,将提取的3层卷积特征上采样到同等尺寸大小。(3.3) Bilinear interpolation, upsampling the extracted 3-layer convolution features to the same size.

进一步地,步骤4所述计算核自相关和步骤8所述计算核互相关具体方法如下:Further, the calculation kernel autocorrelation described in step 4 Cross-correlate with the calculation kernel described in step 8 The specific method is as follows:

(4.1)采用高斯核,公式如下:(4.1) Using the Gaussian kernel, the formula is as follows:

其中,k(x,x′)表示两个特征图谱x和x′计算的高斯核,exp(.)表示e指数函数,σ是高斯函数的标准差,σ取值为0.5,||.||2表示向量或矩阵的2范式;Among them, k(x, x′) represents the Gaussian kernel calculated by two feature maps x and x′, exp(.) represents the e exponential function, σ is the standard deviation of the Gaussian function, and the value of σ is 0.5, ||.| |2 means the 2 normal form of the vector or matrix;

(4.2)计算核相关,公式如下:(4.2) Calculate the kernel correlation, the formula is as follows:

其中,kxx′表示特征图谱x和x′的核相关,exp(.)表示e指数函数,σ是高斯函数的标准差,σ取值为0.5,||.||2表示向量或矩阵的2范式,表示离散傅立叶变换的逆变换,*表示复共轭,^表示离散傅立叶变换,⊙表示两个矩阵对应元素相乘。Among them, kxx' represents the kernel correlation of feature maps x and x', exp(.) represents the e exponential function, σ is the standard deviation of the Gaussian function, and the value of σ is 0.5, and ||.||2 represents the vector or matrix 2 paradigms, Represents the inverse transform of the discrete Fourier transform, * represents the complex conjugate, ^ represents the discrete Fourier transform, and ⊙ represents the multiplication of the corresponding elements of the two matrices.

进一步地,步骤5所述的训练位置和尺度相关滤波器,具体如下:Further, the training location and scale correlation filters described in step 5 are as follows:

根据步骤3所提取的深度卷积特征,对每层特征图谱分别训练1个核相关滤波器,采用如下公式训练模型:According to the deep convolution features extracted in step 3, a kernel correlation filter is trained for each layer of feature maps, and the following formula is used to train the model:

其中,表示根据第l层深度卷积特征图谱求得的相关滤波器模型,表示特征图谱的核自相关,^表示离散傅立叶变换,λ是正则化参数,防止训练的模型过拟合,λ取值为0.001。in, Indicates that the feature map is convolved according to the depth of the l layer The obtained correlation filter model, Represents a feature map Kernel autocorrelation of , ^ means discrete Fourier transform, λ is a regularization parameter to prevent overfitting of the trained model, and the value of λ is 0.001.

进一步地,步骤9所述分别检测位置滤波器和尺度滤波器的输出图谱中最大值对应的位置,来确定目标在当前帧的位置pt和尺度st,具体如下:Further, in step 9, respectively detect the position corresponding to the maximum value in the output spectrum of the position filter and the scale filter to determine the position pt and scalest of the target in the current frame, as follows:

(9.1)从第t帧图像,以位置pt-1和窗口大小选取候选区域zt,trans(9.1) From the tth frame image, select the candidate area zt,trans with the position pt-1 and the window size;

(9.2)提取候选区域zt,trans的3层深度卷积特征图谱(9.2) Extract the 3-layer deep convolution feature map of the candidate area zt, trans

(9.3)对第l层特征图谱计算位置滤波器相关输出置信度图谱ft,trans(l)(9.3) Calculate the position filter correlation output confidence map ft,trans(l) for the l-th layer feature map:

其中,ft(l)表示第l层特征图谱的位置滤波器的输出相应图谱,表示特征图谱的核互相关,是前一帧训练得到的并且更新过的位置滤波器,表示离散傅立叶变换的逆变换,^表示离散傅立叶变换,表示两个矩阵对应元素相乘;Among them, ft(l) represents the output corresponding map of the position filter of the l-th layer feature map, Represents a feature map with The nuclear cross-correlation of is the position filter trained and updated in the previous frame, Represents the inverse transform of the discrete Fourier transform, ^ represents the discrete Fourier transform, Indicates that the corresponding elements of the two matrices are multiplied;

(9.4)从输出相应图谱ft(3)开始由粗到细估计目标位置,第l层的目标位置pt(l)为输出相应图谱ft,trans(l)最大值对应的位置;(9.4) Estimating the target position from coarse to fine from the output corresponding map ft(3) , the target position pt(l) of the first layer is the position corresponding to the maximum value of the output corresponding map ft,trans(l) ;

(9.5)从第t帧图像中,以位置pt和尺度st-1提取尺度估计的候选区域zt,sacle,构建尺度金字塔;(9.5) From the t-th frame image, extract the candidate region zt,sacle for scale estimation with the position pt and scale st-1 , and construct a scale pyramid;

(9.6)提取候选区域的梯度方向直方图特征计算尺度滤波器相关输出置信度图谱ft,sacle(9.6) Extract the gradient direction histogram feature of the candidate area Calculate scale filter correlation output confidence map ft,sacle ;

(9.7)第t帧检测到的目标尺度st为输出相应图谱ft,sacle最大值对应的尺度。(9.7) The target scale st detected in the tth frame is the scale corresponding to the maximum value of the output corresponding map ft,sacle .

进一步地,步骤10所述采用自适应的模型更新策略更新核相关滤波器,具体如下:Further, as described in step 10, an adaptive model update strategy is used to update the kernel correlation filter, specifically as follows:

(10.1)在完成估计目标位置和尺度之后,计算两种跟踪结果置信度度量准则,其一是相关相应输出图谱的峰值:(10.1) After estimating the target position and scale, calculate two tracking result confidence metrics, one of which is the peak value of the corresponding output spectrum:

fmax=maxffmax = maxf

其中,f为核相关滤波器与候选区域相关输出相应图谱,fmax是该图谱的峰值;Wherein, f is the corresponding map output by the kernel correlation filter and the candidate region, and fmax is the peak value of the map;

另一种是平均峰值相关能量APCE:The other is the average peak correlation energy APCE:

其中,fmax和fmin分别是输出图谱f的最大值与最小值,mean(.)表示求均值函数,fi,j表示输出图谱f的第i行第j列数值;Among them, fmax and fmin are the maximum value and minimum value of the output spectrum f respectively, mean(.) represents the mean function, and fi, j represent the value of the i-th row and the j-th column of the output spectrum f;

(10.2)如果fmax和APCE均大于fmax和APCE历史平均值,则模型进行更新,否则,不进行更新;对每一深度卷积层,使用线性插值方法更新,公式如下:(10.2) If fmax and APCE are both greater than fmax and the historical average value of APCE, the model is updated, otherwise, no update is performed; for each deep convolutional layer, the linear interpolation method is used to update, the formula is as follows:

其中,分别表示第l层的前一帧的特征图谱和相关滤波器,η为学习率,η越大模型更新越快,η取值为0.02。in, with respectively represent the feature map and correlation filter of the previous frame of the first layer, η is the learning rate, the larger η is, the faster the model is updated, and the value of η is 0.02.

本发明与现有技术相比,其显著优点为:(1)使用深度卷积特征,深度卷积特征是从大量的数千类别的图像数据中学习得到,具有强大的目标表示能力,使得算法在目标外观发生变化时以及外部因素如光照变化时具有较强的鲁棒性;(2)采用自适应尺度估计方法,该方法与位置估计类似,训练一个单独的尺度滤波器,巧妙使用快速傅里叶变换,实现高效,尺度估计准确,可以结合到任何判别式跟踪算法框架中;(3)采用自适应的高置信度的模型更新策略,当在目标追踪阶段发生错误时,目标检测的置信度较低,此时模型不应更新,这种策略有效地降低了跟踪算法漂移的危险。Compared with the prior art, the present invention has the following significant advantages: (1) using deep convolution features, which are learned from a large number of thousands of categories of image data, and have powerful target representation capabilities, making the algorithm It has strong robustness when the appearance of the target changes and external factors such as illumination changes; (2) adopts the adaptive scale estimation method, which is similar to the position estimation method, trains a separate scale filter, and cleverly uses the fast Fu Lie transform, which achieves high efficiency and accurate scale estimation, can be combined into any discriminative tracking algorithm framework; (3) adopts an adaptive high-confidence model update strategy, when an error occurs in the target tracking phase, the confidence of target detection At this time, the model should not be updated, and this strategy effectively reduces the risk of tracking algorithm drift.

附图说明Description of drawings

图1为本发明基于深度特征核相关滤波器的尺度自适应目标跟踪方法的流程图。FIG. 1 is a flow chart of the scale-adaptive target tracking method based on the deep feature kernel correlation filter of the present invention.

图2为提取深度卷积特征示意图。Figure 2 is a schematic diagram of extracting deep convolution features.

图3为目标位置和尺度估计示意图,其中(a)是由粗到细的目标位置估计示意图,(b)是自适应的目标尺度估计示意图。Fig. 3 is a schematic diagram of target position and scale estimation, where (a) is a schematic diagram of target position estimation from coarse to fine, and (b) is a schematic diagram of adaptive target scale estimation.

图4为自适应的高置信度的模型更新示意图。Fig. 4 is a schematic diagram of an adaptive high-confidence model update.

图5为本发明在标准视觉跟踪数据集上评测结果图,其中(a)是OTB50数据集的准确度绘图,(b)是OTB50数据集的正确率绘图,(c)是OTB100数据集的准确度绘图,(d)是OTB100数据集的正确率绘图。Fig. 5 is the evaluation result diagram of the present invention on the standard visual tracking data set, wherein (a) is the accuracy drawing of the OTB50 data set, (b) is the accuracy drawing of the OTB50 data set, (c) is the accurate drawing of the OTB100 data set Degree drawing, (d) is the correct rate drawing of the OTB100 dataset.

图6为本发明实际视频目标跟踪结果图,其中(a)是OTB100数据集上Human测试视频结果图,(b)是OTB100数据集上Walking测试视频结果图,(c)是OTB50数据集上Tiger测试视频结果图,(d)是OTB50数据集上Dog测试视频结果图。Fig. 6 is the actual video target tracking result diagram of the present invention, wherein (a) is the Human test video result diagram on the OTB100 dataset, (b) is the Walking test video result diagram on the OTB100 dataset, and (c) is the Tiger test video diagram on the OTB50 dataset Test video result graph, (d) is the Dog test video result graph on the OTB50 dataset.

具体实施方式detailed description

为使本发明的结构特征以及所达成的功效有进一步的了解与认识,用以较佳的实施例及附图配合详细的说明,说明如下:In order to further understand and understand the structural features of the present invention and the achieved effects, a detailed description is provided in conjunction with preferred embodiments and accompanying drawings, as follows:

结合图1,本发明基于深度特征核相关滤波器的尺度自适应目标跟踪方法,包括以下几个步骤:In conjunction with Fig. 1, the scale adaptive target tracking method based on the depth feature kernel correlation filter of the present invention includes the following steps:

步骤1,输入目标的初始位置p0和尺度s0,设置窗口大小为目标初始包围盒的2.0倍;Step 1, input the initial position p0 and scale s0 of the target, and set the window size to 2.0 times the initial bounding box of the target;

步骤2,根据第t-1帧的目标位置pt-1,获得目标区域xt-1,尺寸为窗口大小;Step 2, according to the target position pt-1 in the t-1th frame, obtain the target area xt-1 , whose size is the window size;

步骤3,提取目标区域xt-1的深度卷积特征,并进行快速傅立叶变换,得到特征图谱其中^表示离散傅立叶变换;Step 3, extract the deep convolution feature of the target area xt-1 , and perform fast Fourier transform to obtain the feature map Wherein ^ represents the discrete Fourier transform;

步骤4,根据特征图谱计算核自相关Step 4, according to the feature map Computing Kernel Autocorrelation

步骤5,训练位置和尺度相关滤波器;Step 5, training location and scale correlation filters;

步骤6,根据第t-1帧目标的位置pt-1,获得目标在第t帧的候选区域zt,尺寸为窗口大小;Step 6, according to the position pt-1 of the target in the t-1th frame, obtain the candidate area zt of the target in the t-th frame, and the size is the window size;

步骤7,提取候选区域zt的深度卷积特征,并进行快速傅立叶变换,得到特征图谱其中^表示离散傅立叶变换;Step 7, extract the depth convolution feature of the candidate area zt , and perform fast Fourier transform to obtain the feature map Wherein ^ represents the discrete Fourier transform;

步骤8,根据目标前一帧的特征图谱计算核互相关Step 8, according to the feature map of the previous frame of the target Computing Kernel Cross-Correlation

步骤9,分别检测位置滤波器和尺度滤波器的输出图谱中最大值对应的位置,来确定目标在当前帧的位置pt和尺度stStep 9, detecting the position corresponding to the maximum value in the output spectrum of the position filter and the scale filter respectively, to determine the position pt and scalest of the target in the current frame;

步骤10,采用自适应的模型更新策略更新核相关滤波器。In step 10, an adaptive model updating strategy is used to update the kernel correlation filter.

作为一种具体示例,步骤3和步骤7所述深度卷积特征的提取方法,具体如下:As a specific example, the extraction method of the depth convolution feature described in step 3 and step 7 is as follows:

(3.1)预处理,将窗口区域I缩放到卷积神经网络规定的输入尺寸224×224;(3.1) Preprocessing, scaling the window area I to the input size 224×224 specified by the convolutional neural network;

(3.2)特征提取,提取卷积神经网络第3、4和5个卷积层的特征图谱;(3.2) Feature extraction, extracting the feature maps of the 3rd, 4th and 5th convolutional layers of the convolutional neural network;

(3.3)双线性插值,将提取的3层卷积特征上采样到同等尺寸大小。(3.3) Bilinear interpolation, upsampling the extracted 3-layer convolution features to the same size.

作为一种具体示例,步骤4所述计算核自相关和步骤8所述计算核互相关具体方法如下:As a specific example, the calculation kernel autocorrelation described in step 4 Cross-correlate with the calculation kernel described in step 8 The specific method is as follows:

(4.1)采用高斯核,公式如下:(4.1) Using the Gaussian kernel, the formula is as follows:

其中,k(x,x′)表示两个特征图谱x和x′计算的高斯核,exp(.)表示e指数函数,σ是高斯函数的标准差,σ取值为0.5,||.||2表示向量或矩阵的2范式;Among them, k(x, x′) represents the Gaussian kernel calculated by two feature maps x and x′, exp(.) represents the e exponential function, σ is the standard deviation of the Gaussian function, and the value of σ is 0.5, ||.| |2 means the 2 normal form of the vector or matrix;

(4.2)计算核相关,公式如下:(4.2) Calculate the kernel correlation, the formula is as follows:

其中,kxx′表示特征图谱x和x′的核相关,exp(.)表示e指数函数,σ是高斯函数的标准差,σ取值为0.5,||.||2表示向量或矩阵的2范式,表示离散傅立叶变换的逆变换,*表示复共轭,^表示离散傅立叶变换,表示两个矩阵对应元素相乘。Among them, kxx' represents the kernel correlation of feature maps x and x', exp(.) represents the e exponential function, σ is the standard deviation of the Gaussian function, and the value of σ is 0.5, and ||.||2 represents the vector or matrix 2 paradigms, Represents the inverse transform of the discrete Fourier transform, * represents the complex conjugate, ^ represents the discrete Fourier transform, Represents the multiplication of corresponding elements of two matrices.

作为一种具体示例,步骤5所述的训练位置和尺度相关滤波器,具体如下:As a specific example, the training position and scale correlation filter described in step 5 is as follows:

根据步骤3所提取的深度卷积特征,对每层特征图谱分别训练1个核相关滤波器,采用如下公式训练模型:According to the deep convolution features extracted in step 3, a kernel correlation filter is trained for each layer of feature maps, and the following formula is used to train the model:

其中,表示根据第l层深度卷积特征图谱求得的相关滤波器模型,表示特征图谱的核自相关,^表示离散傅立叶变换,λ是正则化参数,防止训练的模型过拟合,λ取值为0.001。in, Indicates that the feature map is convolved according to the depth of the l layer The obtained correlation filter model, Represents a feature map Kernel autocorrelation of , ^ means discrete Fourier transform, λ is a regularization parameter to prevent overfitting of the trained model, and the value of λ is 0.001.

作为一种具体示例,步骤9所述分别检测位置滤波器和尺度滤波器的输出图谱中最大值对应的位置,来确定目标在当前帧的位置pt和尺度st,具体如下:As a specific example, in step 9, respectively detect the position corresponding to the maximum value in the output spectrum of the position filter and the scale filter to determine the position pt and scalest of the target in the current frame, as follows:

(9.1)从第t帧图像,以位置pt-1和窗口大小选取候选区域zt,trans(9.1) From the tth frame image, select the candidate area zt,trans with the position pt-1 and the window size;

(9.2)提取候选区域zt,trans的3层深度卷积特征图谱(9.2) Extract the 3-layer deep convolution feature map of the candidate area zt, trans

(9.3)对第l层特征图谱计算位置滤波器相关输出置信度图谱ft,trans(l)(9.3) Calculate the position filter correlation output confidence map ft,trans(l) for the l-th layer feature map:

其中,ft(l)表示第l层特征图谱的位置滤波器的输出相应图谱,表示特征图谱的核互相关,是前一帧训练得到的并且更新过的位置滤波器,表示离散傅立叶变换的逆变换,^表示离散傅立叶变换,表示两个矩阵对应元素相乘;Among them, ft(l) represents the output corresponding map of the position filter of the l-th layer feature map, Represents a feature map with The nuclear cross-correlation of is the position filter trained and updated in the previous frame, Represents the inverse transform of the discrete Fourier transform, ^ represents the discrete Fourier transform, Indicates that the corresponding elements of the two matrices are multiplied;

(9.4)从输出相应图谱ft(3)开始由粗到细估计目标位置,第l层的目标位置pt(l)为输出相应图谱ft,trans(l)最大值对应的位置;(9.4) Estimating the target position from coarse to fine from the output corresponding map ft(3) , the target position pt(l) of the first layer is the position corresponding to the maximum value of the output corresponding map ft,trans(l) ;

(9.5)从第t帧图像中,以位置pt和尺度st-1提取尺度估计的候选区域zt,sacle,构建尺度金字塔;(9.5) From the t-th frame image, extract the candidate region zt,sacle for scale estimation with the position pt and scale st-1 , and construct a scale pyramid;

(9.6)提取候选区域的梯度方向直方图特征计算尺度滤波器相关输出置信度图谱ft,sacle(9.6) Extract the gradient direction histogram feature of the candidate area Calculate scale filter correlation output confidence map ft,sacle ;

(9.7)第t帧检测到的目标尺度st为输出相应图谱ft,sacle最大值对应的尺度。(9.7) The target scale st detected in the tth frame is the scale corresponding to the maximum value of the output corresponding map ft,sacle .

作为一种具体示例,步骤10所述采用自适应的模型更新策略更新核相关滤波器,具体如下:As a specific example, in step 10, an adaptive model update strategy is used to update the kernel correlation filter, specifically as follows:

(10.1)在完成估计目标位置和尺度之后,计算两种跟踪结果置信度度量准则,其一是相关相应输出图谱的峰值:(10.1) After estimating the target position and scale, calculate two tracking result confidence metrics, one of which is the peak value of the corresponding output spectrum:

fmax=maxffmax = maxf

其中,f为核相关滤波器与候选区域相关输出相应图谱,fmax是该图谱的峰值;Wherein, f is the corresponding map output by the kernel correlation filter and the candidate region, and fmax is the peak value of the map;

另一种是平均峰值相关能量(average peak-to-correlation energy,APCE):The other is the average peak-to-correlation energy (APCE):

其中,fmax和fmin分别是输出图谱f的最大值与最小值,mean(.)表示求均值函数,fi,j表示输出图谱f的第i行第j列数值;Among them, fmax and fmin are the maximum value and minimum value of the output spectrum f respectively, mean(.) represents the mean function, and fi, j represent the value of the i-th row and the j-th column of the output spectrum f;

(10.2)如果fmax和APCE均大于fmax和APCE历史平均值,则模型进行更新,否则,不进行更新;对每一深度卷积层,使用线性插值方法更新,公式如下:(10.2) If fmax and APCE are both greater than fmax and the historical average value of APCE, the model is updated, otherwise, no update is performed; for each deep convolutional layer, the linear interpolation method is used to update, the formula is as follows:

其中,分别表示第l层的前一帧的特征图谱和相关滤波器,η为学习率,η越大模型更新越快,η取值为0.02。in, with respectively represent the feature map and correlation filter of the previous frame of the first layer, η is the learning rate, the larger η is, the faster the model is updated, and the value of η is 0.02.

本发明解决了由目标形变、光照变化、目标旋转和尺度变化等外观变化以及目标遮挡等导致跟踪失败的缺陷。借助深度卷积特征的强大目标表示能力,提高了目标在复杂场景和外观变化中的目标跟踪的鲁棒性;另外,本发明能够高效准确地处理目标尺度变化,最后,由于采用自适应的高置信度的模型更新策略,减少了因错误检测而导致的模型跟踪漂移。The invention solves the defect of tracking failure caused by object deformation, illumination change, object rotation, scale change and other appearance changes, as well as object occlusion. With the help of the powerful target representation ability of deep convolution features, the robustness of target tracking in complex scenes and appearance changes is improved; in addition, the present invention can efficiently and accurately deal with target scale changes. Finally, due to the use of adaptive high Confidence-based model update strategy that reduces model tracking drift due to false detections.

下面结合具体实施例对本发明做进一步详细说明。The present invention will be described in further detail below in conjunction with specific embodiments.

实施例1Example 1

本发明基于深度特征核相关滤波器的尺度自适应目标跟踪方法,该方法主要分为四步,第一步为提取深度卷积特征;第二步训练核相关滤波器;第三步估计目标在当前帧的位置和尺度;第四步,采用自适应的高置信度的模型更新策略。The present invention is based on the scale adaptive target tracking method of the deep feature kernel correlation filter, the method is mainly divided into four steps, the first step is to extract the depth convolution feature; the second step is to train the kernel correlation filter; the third step is to estimate the target at The position and scale of the current frame; the fourth step is to adopt an adaptive high-confidence model update strategy.

步骤1,输入目标的初始位置p0和尺度s0,设置窗口大小为目标初始包围盒的2.0倍;Step 1, input the initial position p0 and scale s0 of the target, and set the window size to 2.0 times the initial bounding box of the target;

步骤2,根据第t-1帧的目标位置pt-1,获得目标区域xt-1,尺寸为窗口大小;Step 2, according to the target position pt-1 of the t-1th frame, obtain the target area xt-1 , whose size is the window size;

步骤3,提取目标区域xt-1的深度卷积特征,快速傅立叶变换,得到特征图谱其中^表示离散傅立叶变换;Step 3, extract the deep convolution feature of the target area xt-1 , fast Fourier transform, and obtain the feature map Wherein ^ represents the discrete Fourier transform;

步骤4,根据特征图谱计算核自相关Step 4, according to the feature map Computing Kernel Autocorrelation

步骤5,训练位置和尺度相关滤波器;Step 5, training location and scale correlation filters;

步骤6,根据第t-1帧目标的位置pt-1,获得目标在第t帧的候选区域zt,尺寸为窗口大小;Step 6, according to the position pt-1 of the target in the t-1th frame, obtain the candidate area zt of the target in the t-th frame, and the size is the window size;

步骤7,提取候选区域zt的深度卷积特征,并进行快速傅立叶变换,得到特征图谱其中^表示离散傅立叶变换;Step 7, extract the depth convolution feature of the candidate area zt , and perform fast Fourier transform to obtain the feature map Wherein ^ represents the discrete Fourier transform;

步骤8,根据目标前一帧的特征图谱计算核互相关Step 8, according to the feature map of the previous frame of the target Computing Kernel Cross-Correlation

步骤9,分别检测位置滤波器和尺度滤波器的输出相应图谱的最大值对应的位置来确定目标在当前帧的位置pt和尺度stStep 9, respectively detecting the position corresponding to the maximum value of the output corresponding map of the position filter and the scale filter to determine the position pt and scalest of the target in the current frame;

步骤10,采用自适应的高置信度的模型更新策略更新核相关滤波器。In step 10, an adaptive high-confidence model update strategy is used to update the kernel correlation filter.

如图2所示,给出了基于深度特征核相关滤波器的尺度自适应目标跟踪方法的提取深度卷积特征的示意图。本发明基于深度特征核相关滤波器的尺度自适应目标跟踪方法,其特征在于提取深度卷积特征,深度卷积特征是从大量的数千类别的图像数据中学习得到,具有强大的目标表示能力,使得算法在目标外观发生变化时以及外部因素如光照变化时具有较强的鲁棒性,具体步骤如下:As shown in Figure 2, a schematic diagram of extracting deep convolutional features of the scale-adaptive target tracking method based on deep feature kernel correlation filters is given. The present invention is based on the scale-adaptive target tracking method of the deep feature kernel correlation filter, which is characterized in that the deep convolution feature is extracted, and the deep convolution feature is learned from a large number of thousands of categories of image data, and has a strong target representation ability , which makes the algorithm more robust when the appearance of the target changes and external factors such as illumination changes. The specific steps are as follows:

(3.1)预处理,将窗口区域I缩放到卷积神经网络规定的输入尺寸224×224;(3.1) Preprocessing, scaling the window area I to the input size 224×224 specified by the convolutional neural network;

(3.2)特征提取,提取卷积神经网络第3、4和5个卷积层的特征图谱;(3.2) Feature extraction, extracting the feature maps of the 3rd, 4th and 5th convolutional layers of the convolutional neural network;

(3.3)双线性插值,将提取的3层卷积特征上采样到同等尺寸大小。(3.3) Bilinear interpolation, upsampling the extracted 3-layer convolution features to the same size.

如图3所示,给出了基于深度特征核相关滤波器的尺度自适应目标跟踪方法的目标位置和尺度估计示意图,其中图3(a)是由粗到细的目标位置估计示意图,图3(b)是自适应的目标尺度估计示意图。本发明基于深度特征核相关滤波器的尺度自适应目标跟踪方法,其特征在于由粗到细的分层位置估计方法和自适应的目标尺度估计方法。传统的核相关滤波方法目标的模型尺寸固定,无法处理目标尺度的变化,因此容易导致跟踪失败。本发明提出了一种自适应尺度估计方法,它主要思想是训练一个独立的尺度滤波器,通过尺度滤波器相关响应最大时对应的尺度来估计,这种方法巧妙应用快速傅立叶变换,简单而高效,可以集成到传统的判别式目标跟踪方法中,具体步骤如下:As shown in Figure 3, a schematic diagram of the target position and scale estimation of the scale-adaptive target tracking method based on the deep feature kernel correlation filter is given, where Figure 3(a) is a schematic diagram of target position estimation from coarse to fine, and Figure 3 (b) is a schematic diagram of adaptive target scale estimation. The scale-adaptive target tracking method based on the deep feature kernel correlation filter of the present invention is characterized by a coarse-to-fine layered position estimation method and an adaptive target scale estimation method. The model size of the target in the traditional kernel correlation filtering method is fixed, which cannot deal with the change of the target scale, so it is easy to cause tracking failure. The present invention proposes an adaptive scale estimation method, the main idea of which is to train an independent scale filter and estimate the scale corresponding to the maximum correlation response of the scale filter. This method cleverly uses fast Fourier transform, which is simple and efficient , which can be integrated into the traditional discriminative target tracking method, the specific steps are as follows:

(9.1)从第t帧图像,以位置pt-1和窗口大小选取候选区域zt,trans(9.1) From the tth frame image, select the candidate area zt,trans with the position pt-1 and the window size;

(9.2)提取候选区域zt,trans的3层深度卷积特征图谱(9.2) Extract the 3-layer deep convolution feature map of the candidate area zt, trans

(9.3)对第l层特征图谱计算位置滤波器相关输出置信度图谱ft,trans(l)(9.3) Calculate the position filter correlation output confidence map ft,trans(l) for the l-th layer feature map:

其中,ft(l)表示第l层特征图谱的位置滤波器的输出相应图谱,表示特征图谱的核互相关,是前一帧训练得到的并且更新过的位置滤波器,表示离散傅立叶变换的逆变换,^表示离散傅立叶变换,表示两个矩阵对应元素相乘。Among them, ft(l) represents the output corresponding map of the position filter of the l-th layer feature map, Represents a feature map with The nuclear cross-correlation of is the position filter trained and updated in the previous frame, Represents the inverse transform of the discrete Fourier transform, ^ represents the discrete Fourier transform, Represents the multiplication of corresponding elements of two matrices.

(9.4)从输出相应图谱ft(3)开始由粗到细估计目标位置,第l层的目标位置pt(l)为输出相应图谱ft,trans(l)最大值对应的位置;(9.4) Estimating the target position from coarse to fine from the output corresponding map ft(3) , the target position pt(l) of the first layer is the position corresponding to the maximum value of the output corresponding map ft,trans(l) ;

(9.5)从第t帧图像中,以位置pt和尺度st-1提取尺度估计的候选区域zt,sacle,构建尺度金字塔;(9.5) From the t-th frame image, extract the candidate region zt,sacle for scale estimation with the position pt and scale st-1 , and construct a scale pyramid;

(9.6)提取候选区域的梯度方向直方图特征计算尺度滤波器相关输出置信度图谱ft,sacle(9.6) Extract the gradient direction histogram feature of the candidate area Calculate scale filter correlation output confidence map ft,sacle ;

(9.7)第t帧检测到的目标尺度st为输出相应图谱ft,sacle最大值对应的尺度。(9.7) The target scale st detected in the tth frame is the scale corresponding to the maximum value of the output corresponding map ft,sacle .

如图4所示,给出了基于深度特征核相关滤波器的尺度自适应目标跟踪方法的自适应的高置信度的模型更新示意图。本发明基于深度特征核相关滤波器的尺度自适应目标跟踪方法,其特征在于自适应的高置信度的模型更新策略。传统的核相关滤波模型更新方法采用简单线性插值方法,没有进行跟踪错误检测,当跟踪时目标检测出现错误,模型更新会污染核相关滤波器,导致跟踪算法漂移。本发明提出了两种检测置信度准则,当确定目标跟踪正确时,则模型更新,否则,模型不更新,具体步骤如下:As shown in Figure 4, a schematic diagram of the adaptive high-confidence model update of the scale-adaptive target tracking method based on the deep feature kernel correlation filter is given. The scale adaptive target tracking method based on the depth feature kernel correlation filter of the present invention is characterized by an adaptive high-confidence model update strategy. The traditional nuclear correlation filter model update method uses a simple linear interpolation method without tracking error detection. When the target detection error occurs during tracking, the model update will pollute the nuclear correlation filter and cause the tracking algorithm to drift. The present invention proposes two detection confidence criteria. When it is determined that the target is tracked correctly, the model is updated; otherwise, the model is not updated. The specific steps are as follows:

(10.1)在完成估计目标位置和尺度之后,计算两种跟踪结果置信度度量准则,其一是相关相应输出图谱的峰值:(10.1) After estimating the target position and scale, calculate two tracking result confidence metrics, one of which is the peak value of the corresponding output spectrum:

fmax=maxffmax = maxf

其中,f为核相关滤波器与候选区域相关输出相应图谱,fmax是该图谱的峰值。另一种是平均峰值相关能量(average peak-to-correlation energy,APCE):Among them, f is the corresponding map output by the kernel correlation filter and the candidate region, and fmax is the peak value of the map. The other is the average peak-to-correlation energy (APCE):

其中,fmax和fmin是输出图谱f的最大值与最小值,mean(.)表示求均值函数。fi,j表示输出图谱f的第i行第j列数值。Among them, fmax and fmin are the maximum and minimum values of the output spectrum f, and mean(.) represents the mean function. fi, j represents the value of row i and column j of the output spectrum f.

(10.2)如果fmax和APCE均大于fmax和APCE历史平均值,则模型进行更新,否则,不进行更新。对每一深度卷积层,使用线性插值方法更新,公式如下:(10.2) If both fmax and APCE are greater than fmax and APCE historical average, the model is updated, otherwise, no update is performed. For each deep convolutional layer, it is updated using the linear interpolation method, the formula is as follows:

其中,分别表示第l层的前一帧的特征图谱和相关滤波器,η为学习率,η越大模型更新地越快,在发明中η取值为0.02。in, with Respectively represent the feature map and the correlation filter of the previous frame of the first layer, η is the learning rate, the larger the η, the faster the model is updated, and the value of η is 0.02 in the invention.

如图5所示,展示了本发明在标准视觉追踪数据集OTB50和OTB100上评测结果图,其中(a)是OTB50数据集的准确度绘图,(b)是OTB50数据集的正确率绘图,(c)是OTB100数据集的准确度绘图,(d)是OTB100数据集的正确率绘图。OTB50数据集有50个视频序列,总共拥有29000帧,而OTB100数据集拥有100个视频序列,总共拥有58897帧,它们每帧都有目标的标记。评测指标主要有两种:准确度和成功率。在准确度绘图(a)和(c)中,准确度定义为算法检测位置与目标标定位置之间的距离不超过20像素的帧数占总评测帧数的百分比;在成功率绘图(b)和(d)中,重叠率指的是算法检测目标包围盒与目标标定包围盒两者之间重叠部分面积(交运算)占总面积(并运算)的百分比超过50%的帧数占总评测帧数的百分比。从评测结果可以看出,与经典的目标跟踪算法相比,本发明(fSCF2)在目标追踪任务中表现出色。As shown in Figure 5, it shows the evaluation results of the present invention on the standard visual tracking data sets OTB50 and OTB100, wherein (a) is the accuracy drawing of the OTB50 data set, (b) is the correct rate drawing of the OTB50 data set, ( c) is the accuracy plot of the OTB100 dataset, and (d) is the accuracy plot of the OTB100 dataset. The OTB50 dataset has 50 video sequences with a total of 29,000 frames, while the OTB100 dataset has 100 video sequences with a total of 58,897 frames, each of which has a target label. There are two main evaluation indicators: accuracy and success rate. In the accuracy plots (a) and (c), the accuracy is defined as the percentage of the number of frames whose distance between the algorithm detection position and the target calibration position does not exceed 20 pixels to the total number of evaluation frames; in the success rate plot (b) In (d) and (d), the overlap rate refers to the percentage of the overlapping area (intersection operation) between the algorithm detection target bounding box and the target calibration bounding box to the total area (combination operation) and the number of frames exceeding 50% of the total evaluation Percentage of frames. It can be seen from the evaluation results that, compared with the classic target tracking algorithms, the present invention (fSCF2) performs well in the target tracking task.

如图6所示,展示了本发明与近年来一些出色的算法在实际视频中目标追踪结果比较图,其中(a)是OTB100数据集上Human测试视频结果图,(b)是OTB100数据集上Walking测试视频结果图,(c)是OTB50数据集上Tiger测试视频结果图,(d)是OTB50数据集上Dog测试视频结果图。总体来看,与经典的目标跟踪算法相比,本发明(fSCF2)追踪效果最好,由于使用深度卷积特征,具有很强的目标表示能力,自适应尺度估计机制以及自适应的高置信度的模型更新策略使本发明能够在目标发生遮挡,尺度变化,目标变形以及目标快速运动等不利因素条件下仍然能够准确追踪目标。As shown in Figure 6, it shows the comparison between the present invention and some excellent algorithms in recent years in the actual video target tracking results, where (a) is the Human test video result graph on the OTB100 dataset, (b) is the OTB100 dataset The results of the Walking test video, (c) is the result of the Tiger test video on the OTB50 dataset, and (d) is the result of the Dog test video on the OTB50 dataset. Overall, compared with the classic target tracking algorithm, the present invention (fSCF2) has the best tracking effect. Due to the use of deep convolution features, it has strong target representation ability, adaptive scale estimation mechanism and adaptive high confidence The model updating strategy enables the present invention to track the target accurately even under adverse conditions such as target occlusion, scale change, target deformation and fast movement of the target.

Claims (6)

CN201710355456.3A2017-05-192017-05-19Dimension self-adaption method for tracking target based on depth characteristic core correlation filterPendingCN107154024A (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN201710355456.3ACN107154024A (en)2017-05-192017-05-19Dimension self-adaption method for tracking target based on depth characteristic core correlation filter

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN201710355456.3ACN107154024A (en)2017-05-192017-05-19Dimension self-adaption method for tracking target based on depth characteristic core correlation filter

Publications (1)

Publication NumberPublication Date
CN107154024Atrue CN107154024A (en)2017-09-12

Family

ID=59794201

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN201710355456.3APendingCN107154024A (en)2017-05-192017-05-19Dimension self-adaption method for tracking target based on depth characteristic core correlation filter

Country Status (1)

CountryLink
CN (1)CN107154024A (en)

Cited By (56)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN107644217A (en)*2017-09-292018-01-30中国科学技术大学Method for tracking target based on convolutional neural networks and correlation filter
CN107730536A (en)*2017-09-152018-02-23北京飞搜科技有限公司A kind of high speed correlation filtering object tracking method based on depth characteristic
CN108053424A (en)*2017-12-152018-05-18深圳云天励飞技术有限公司Method for tracking target, device, electronic equipment and storage medium
CN108182388A (en)*2017-12-142018-06-19哈尔滨工业大学(威海)A kind of motion target tracking method based on image
CN108280808A (en)*2017-12-152018-07-13西安电子科技大学The method for tracking target of correlation filter is exported based on structuring
CN108345885A (en)*2018-01-182018-07-31浙江大华技术股份有限公司A kind of method and device of target occlusion detection
CN108346159A (en)*2018-01-282018-07-31北京工业大学A kind of visual target tracking method based on tracking-study-detection
CN108550126A (en)*2018-04-182018-09-18长沙理工大学A kind of adaptive correlation filter method for tracking target and system
CN108573499A (en)*2018-03-162018-09-25东华大学 A Visual Object Tracking Method Based on Scale Adaptation and Occlusion Detection
CN108665481A (en)*2018-03-272018-10-16西安电子科技大学 Adaptive anti-occlusion infrared target tracking method based on multi-layer deep feature fusion
CN108734151A (en)*2018-06-142018-11-02厦门大学Robust long-range method for tracking target based on correlation filtering and the twin network of depth
CN108830878A (en)*2018-04-132018-11-16上海大学A kind of method for tracking target based on FPN neural network
CN108846345A (en)*2018-06-062018-11-20安徽大学Moving object scale estimation method in monitoring scene
CN109035290A (en)*2018-07-162018-12-18南京信息工程大学A kind of track algorithm updating accretion learning based on high confidence level
CN109035300A (en)*2018-07-052018-12-18桂林电子科技大学A kind of method for tracking target based on depth characteristic Yu average peak correlation energy
CN109064491A (en)*2018-04-122018-12-21江苏省基础地理信息中心A kind of nuclear phase pass filter tracking method of adaptive piecemeal
CN109146928A (en)*2017-12-292019-01-04西安电子科技大学A kind of method for tracking target that Grads threshold judgment models update
CN109146917A (en)*2017-12-292019-01-04西安电子科技大学A kind of method for tracking target of elasticity more new strategy
CN109166106A (en)*2018-08-022019-01-08山东大学A kind of target detection aligning method and apparatus based on sliding window
CN109255304A (en)*2018-08-172019-01-22西安电子科技大学Method for tracking target based on distribution field feature
CN109410251A (en)*2018-11-192019-03-01南京邮电大学Method for tracking target based on dense connection convolutional network
CN109410247A (en)*2018-10-162019-03-01中国石油大学(华东)A kind of video tracking algorithm of multi-template and adaptive features select
CN109410246A (en)*2018-09-252019-03-01深圳市中科视讯智能系统技术有限公司The method and device of vision tracking based on correlation filtering
CN109461172A (en)*2018-10-252019-03-12南京理工大学Manually with the united correlation filtering video adaptive tracking method of depth characteristic
CN109584271A (en)*2018-11-152019-04-05西北工业大学High speed correlation filtering tracking based on high confidence level more new strategy
CN109741366A (en)*2018-11-272019-05-10昆明理工大学 A Correlation Filtering Target Tracking Method Fusion Multi-layer Convolution Features
CN109785360A (en)*2018-12-182019-05-21南京理工大学A kind of adaptive method for real time tracking excavated based on online sample
CN109801311A (en)*2019-01-312019-05-24长安大学A kind of visual target tracking method based on depth residual error network characterization
CN109859244A (en)*2019-01-222019-06-07西安微电子技术研究所A kind of visual tracking method based on convolution sparseness filtering
CN109858326A (en)*2018-12-112019-06-07中国科学院自动化研究所Based on classification semantic Weakly supervised online visual tracking method and system
CN109858455A (en)*2019-02-182019-06-07南京航空航天大学A kind of piecemeal detection scale adaptive tracking method for circular target
CN109858454A (en)*2019-02-152019-06-07东北大学 A Dual-Model Adaptive Kernel Correlation Filtering Tracking Method
CN109886996A (en)*2019-01-152019-06-14东华大学 A visual tracking optimization method
CN109934098A (en)*2019-01-242019-06-25西北工业大学 A camera intelligent system with privacy protection and its realization method
CN110033006A (en)*2019-04-042019-07-19中设设计集团股份有限公司Vehicle detecting and tracking method based on color characteristic Nonlinear Dimension Reduction
CN110197126A (en)*2019-05-062019-09-03深圳岚锋创视网络科技有限公司A kind of target tracking method, device and portable terminal
CN110211149A (en)*2018-12-252019-09-06湖州云通科技有限公司A kind of dimension self-adaption nuclear phase pass filter tracking method based on context-aware
CN110414439A (en)*2019-07-302019-11-05武汉理工大学 Anti-Occlusion Pedestrian Tracking Method Based on Multi-Peak Detection
CN110544267A (en)*2019-07-242019-12-06中国地质大学(武汉) A Correlation Filtering Tracking Method Based on Adaptive Feature Selection
CN110555870A (en)*2019-09-092019-12-10北京理工大学DCF tracking confidence evaluation and classifier updating method based on neural network
CN110633595A (en)*2018-06-212019-12-31北京京东尚科信息技术有限公司Target detection method and device by utilizing bilinear interpolation
CN110689559A (en)*2019-09-302020-01-14长安大学 A Visual Object Tracking Method Based on Dense Convolutional Network Features
CN110807473A (en)*2019-10-122020-02-18浙江大华技术股份有限公司Target detection method, device and computer storage medium
CN110889863A (en)*2019-09-032020-03-17河南理工大学 A target tracking method based on target-aware correlation filtering
CN111161323A (en)*2019-12-312020-05-15北京理工大学重庆创新中心 A method and system for target tracking in complex scenes based on correlation filtering
CN111192288A (en)*2018-11-142020-05-22天津大学青岛海洋技术研究院Target tracking algorithm based on deformation sample generation network
CN111221770A (en)*2019-12-312020-06-02中国船舶重工集团公司第七一七研究所Kernel correlation filtering target tracking method and system
CN111428740A (en)*2020-02-282020-07-17深圳壹账通智能科技有限公司Detection method and device for network-shot photo, computer equipment and storage medium
CN111476819A (en)*2020-03-192020-07-31重庆邮电大学 A long-term target tracking method based on multi-correlation filtering model
CN111507999A (en)*2019-01-302020-08-07北京四维图新科技股份有限公司 A target tracking method and device based on FDSST algorithm
CN111696132A (en)*2020-05-152020-09-22深圳市优必选科技股份有限公司Target tracking method and device, computer readable storage medium and robot
CN112053386A (en)*2020-08-312020-12-08西安电子科技大学Target tracking method based on depth convolution characteristic self-adaptive integration
CN112200833A (en)*2020-09-172021-01-08天津城建大学Relevant filtering video tracking algorithm based on residual error network and short-term visual memory
CN112560695A (en)*2020-12-172021-03-26中国海洋大学Underwater target tracking method, system, storage medium, equipment, terminal and application
CN113379804A (en)*2021-07-122021-09-10闽南师范大学Unmanned aerial vehicle target tracking method, terminal equipment and storage medium
CN114708300A (en)*2022-03-022022-07-05北京理工大学Anti-blocking self-adaptive target tracking method and system

Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2015163830A1 (en)*2014-04-222015-10-29Aselsan Elektronik Sanayi Ve Ticaret Anonim SirketiTarget localization and size estimation via multiple model learning in visual tracking
CN106570486A (en)*2016-11-092017-04-19华南理工大学Kernel correlation filtering target tracking method based on feature fusion and Bayesian classification

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2015163830A1 (en)*2014-04-222015-10-29Aselsan Elektronik Sanayi Ve Ticaret Anonim SirketiTarget localization and size estimation via multiple model learning in visual tracking
CN106570486A (en)*2016-11-092017-04-19华南理工大学Kernel correlation filtering target tracking method based on feature fusion and Bayesian classification

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
张雷: "复杂场景下实时目标跟踪算法及实现技术研究", 《中国博士学位论文全文数据库 信息科技辑》*

Cited By (92)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN107730536A (en)*2017-09-152018-02-23北京飞搜科技有限公司A kind of high speed correlation filtering object tracking method based on depth characteristic
CN107730536B (en)*2017-09-152020-05-12苏州飞搜科技有限公司High-speed correlation filtering object tracking method based on depth features
CN107644217A (en)*2017-09-292018-01-30中国科学技术大学Method for tracking target based on convolutional neural networks and correlation filter
CN107644217B (en)*2017-09-292020-06-26中国科学技术大学Target tracking method based on convolutional neural network and related filter
CN108182388A (en)*2017-12-142018-06-19哈尔滨工业大学(威海)A kind of motion target tracking method based on image
CN108053424B (en)*2017-12-152020-06-16深圳云天励飞技术有限公司Target tracking method and device, electronic equipment and storage medium
CN108053424A (en)*2017-12-152018-05-18深圳云天励飞技术有限公司Method for tracking target, device, electronic equipment and storage medium
CN108280808A (en)*2017-12-152018-07-13西安电子科技大学The method for tracking target of correlation filter is exported based on structuring
CN109146928B (en)*2017-12-292021-09-24西安电子科技大学 A Target Tracking Method Based on Gradient Threshold Judging Model Update
CN109146917A (en)*2017-12-292019-01-04西安电子科技大学A kind of method for tracking target of elasticity more new strategy
CN109146928A (en)*2017-12-292019-01-04西安电子科技大学A kind of method for tracking target that Grads threshold judgment models update
CN109146917B (en)*2017-12-292020-07-28西安电子科技大学 A Target Tracking Method for Elastic Update Policy
CN108345885A (en)*2018-01-182018-07-31浙江大华技术股份有限公司A kind of method and device of target occlusion detection
CN108346159A (en)*2018-01-282018-07-31北京工业大学A kind of visual target tracking method based on tracking-study-detection
CN108346159B (en)*2018-01-282021-10-15北京工业大学 A Tracking-Learning-Detection-Based Visual Object Tracking Method
CN108573499A (en)*2018-03-162018-09-25东华大学 A Visual Object Tracking Method Based on Scale Adaptation and Occlusion Detection
CN108573499B (en)*2018-03-162021-04-02东华大学 A Visual Object Tracking Method Based on Scale Adaptive and Occlusion Detection
CN108665481A (en)*2018-03-272018-10-16西安电子科技大学 Adaptive anti-occlusion infrared target tracking method based on multi-layer deep feature fusion
CN108665481B (en)*2018-03-272022-05-31西安电子科技大学 Adaptive anti-occlusion infrared target tracking method based on multi-layer deep feature fusion
CN109064491A (en)*2018-04-122018-12-21江苏省基础地理信息中心A kind of nuclear phase pass filter tracking method of adaptive piecemeal
CN108830878A (en)*2018-04-132018-11-16上海大学A kind of method for tracking target based on FPN neural network
CN108830878B (en)*2018-04-132021-02-23上海大学 A target tracking method based on FPN neural network
CN108550126A (en)*2018-04-182018-09-18长沙理工大学A kind of adaptive correlation filter method for tracking target and system
CN108846345B (en)*2018-06-062021-09-17安徽大学Moving object scale estimation method in monitoring scene
CN108846345A (en)*2018-06-062018-11-20安徽大学Moving object scale estimation method in monitoring scene
CN108734151B (en)*2018-06-142020-04-14厦门大学 Robust long-range target tracking method based on correlation filtering and deep Siamese network
CN108734151A (en)*2018-06-142018-11-02厦门大学Robust long-range method for tracking target based on correlation filtering and the twin network of depth
CN110633595B (en)*2018-06-212022-12-02北京京东尚科信息技术有限公司Target detection method and device by utilizing bilinear interpolation
CN110633595A (en)*2018-06-212019-12-31北京京东尚科信息技术有限公司Target detection method and device by utilizing bilinear interpolation
CN109035300A (en)*2018-07-052018-12-18桂林电子科技大学A kind of method for tracking target based on depth characteristic Yu average peak correlation energy
CN109035300B (en)*2018-07-052021-03-26桂林电子科技大学Target tracking method based on depth feature and average peak correlation energy
CN109035290A (en)*2018-07-162018-12-18南京信息工程大学A kind of track algorithm updating accretion learning based on high confidence level
CN109166106A (en)*2018-08-022019-01-08山东大学A kind of target detection aligning method and apparatus based on sliding window
CN109255304A (en)*2018-08-172019-01-22西安电子科技大学Method for tracking target based on distribution field feature
CN109255304B (en)*2018-08-172021-07-27西安电子科技大学 Target Tracking Method Based on Distributed Field Features
CN109410246A (en)*2018-09-252019-03-01深圳市中科视讯智能系统技术有限公司The method and device of vision tracking based on correlation filtering
CN109410246B (en)*2018-09-252021-06-11杭州视语智能视觉系统技术有限公司Visual tracking method and device based on correlation filtering
CN109410247A (en)*2018-10-162019-03-01中国石油大学(华东)A kind of video tracking algorithm of multi-template and adaptive features select
CN109461172A (en)*2018-10-252019-03-12南京理工大学Manually with the united correlation filtering video adaptive tracking method of depth characteristic
CN111192288B (en)*2018-11-142023-08-04天津大学青岛海洋技术研究院Target tracking algorithm based on deformation sample generation network
CN111192288A (en)*2018-11-142020-05-22天津大学青岛海洋技术研究院Target tracking algorithm based on deformation sample generation network
CN109584271A (en)*2018-11-152019-04-05西北工业大学High speed correlation filtering tracking based on high confidence level more new strategy
CN109410251A (en)*2018-11-192019-03-01南京邮电大学Method for tracking target based on dense connection convolutional network
CN109741366A (en)*2018-11-272019-05-10昆明理工大学 A Correlation Filtering Target Tracking Method Fusion Multi-layer Convolution Features
CN109858326A (en)*2018-12-112019-06-07中国科学院自动化研究所Based on classification semantic Weakly supervised online visual tracking method and system
CN109785360A (en)*2018-12-182019-05-21南京理工大学A kind of adaptive method for real time tracking excavated based on online sample
CN110211149A (en)*2018-12-252019-09-06湖州云通科技有限公司A kind of dimension self-adaption nuclear phase pass filter tracking method based on context-aware
CN110211149B (en)*2018-12-252022-08-12湖州云通科技有限公司 A Scale-adaptive Kernel Correlation Filter Tracking Method Based on Background Perception
CN109886996B (en)*2019-01-152023-06-06东华大学 A Visual Tracking Optimization Method
CN109886996A (en)*2019-01-152019-06-14东华大学 A visual tracking optimization method
CN109859244A (en)*2019-01-222019-06-07西安微电子技术研究所A kind of visual tracking method based on convolution sparseness filtering
CN109859244B (en)*2019-01-222022-07-08西安微电子技术研究所Visual tracking method based on convolution sparse filtering
CN109934098A (en)*2019-01-242019-06-25西北工业大学 A camera intelligent system with privacy protection and its realization method
CN111507999B (en)*2019-01-302023-07-18北京四维图新科技股份有限公司 A target tracking method and device based on FDSST algorithm
CN111507999A (en)*2019-01-302020-08-07北京四维图新科技股份有限公司 A target tracking method and device based on FDSST algorithm
CN109801311A (en)*2019-01-312019-05-24长安大学A kind of visual target tracking method based on depth residual error network characterization
CN109801311B (en)*2019-01-312021-07-16长安大学 A Visual Object Tracking Method Based on Deep Residual Network Features
CN109858454A (en)*2019-02-152019-06-07东北大学 A Dual-Model Adaptive Kernel Correlation Filtering Tracking Method
CN109858454B (en)*2019-02-152023-04-07东北大学Adaptive kernel correlation filtering tracking method based on dual models
CN109858455A (en)*2019-02-182019-06-07南京航空航天大学A kind of piecemeal detection scale adaptive tracking method for circular target
CN110033006A (en)*2019-04-042019-07-19中设设计集团股份有限公司Vehicle detecting and tracking method based on color characteristic Nonlinear Dimension Reduction
CN110197126A (en)*2019-05-062019-09-03深圳岚锋创视网络科技有限公司A kind of target tracking method, device and portable terminal
CN110544267B (en)*2019-07-242022-03-15中国地质大学(武汉) A Correlation Filter Tracking Method for Adaptive Feature Selection
CN110544267A (en)*2019-07-242019-12-06中国地质大学(武汉) A Correlation Filtering Tracking Method Based on Adaptive Feature Selection
CN110414439B (en)*2019-07-302022-03-15武汉理工大学 Anti-occlusion pedestrian tracking method based on multi-peak detection
CN110414439A (en)*2019-07-302019-11-05武汉理工大学 Anti-Occlusion Pedestrian Tracking Method Based on Multi-Peak Detection
CN110889863A (en)*2019-09-032020-03-17河南理工大学 A target tracking method based on target-aware correlation filtering
CN110889863B (en)*2019-09-032023-03-24河南理工大学Target tracking method based on target perception correlation filtering
CN110555870B (en)*2019-09-092021-07-27北京理工大学 A Neural Network-based DCF Tracking Confidence Evaluation and Classifier Update Method
CN110555870A (en)*2019-09-092019-12-10北京理工大学DCF tracking confidence evaluation and classifier updating method based on neural network
CN110689559A (en)*2019-09-302020-01-14长安大学 A Visual Object Tracking Method Based on Dense Convolutional Network Features
CN110689559B (en)*2019-09-302022-08-12长安大学 A Visual Object Tracking Method Based on Dense Convolutional Network Features
CN110807473A (en)*2019-10-122020-02-18浙江大华技术股份有限公司Target detection method, device and computer storage medium
CN110807473B (en)*2019-10-122023-01-03浙江大华技术股份有限公司Target detection method, device and computer storage medium
CN111161323B (en)*2019-12-312023-11-28北京理工大学重庆创新中心Complex scene target tracking method and system based on correlation filtering
CN111221770A (en)*2019-12-312020-06-02中国船舶重工集团公司第七一七研究所Kernel correlation filtering target tracking method and system
CN111161323A (en)*2019-12-312020-05-15北京理工大学重庆创新中心 A method and system for target tracking in complex scenes based on correlation filtering
WO2021169625A1 (en)*2020-02-282021-09-02深圳壹账通智能科技有限公司Method and apparatus for detecting reproduced network photograph, computer device, and storage medium
CN111428740A (en)*2020-02-282020-07-17深圳壹账通智能科技有限公司Detection method and device for network-shot photo, computer equipment and storage medium
CN111476819A (en)*2020-03-192020-07-31重庆邮电大学 A long-term target tracking method based on multi-correlation filtering model
WO2021227519A1 (en)*2020-05-152021-11-18深圳市优必选科技股份有限公司Target tracking method and apparatus, and computer-readable storage medium and robot
CN111696132B (en)*2020-05-152023-12-29深圳市优必选科技股份有限公司Target tracking method, device, computer readable storage medium and robot
CN111696132A (en)*2020-05-152020-09-22深圳市优必选科技股份有限公司Target tracking method and device, computer readable storage medium and robot
CN112053386B (en)*2020-08-312023-04-18西安电子科技大学Target tracking method based on depth convolution characteristic self-adaptive integration
CN112053386A (en)*2020-08-312020-12-08西安电子科技大学Target tracking method based on depth convolution characteristic self-adaptive integration
CN112200833A (en)*2020-09-172021-01-08天津城建大学Relevant filtering video tracking algorithm based on residual error network and short-term visual memory
CN112560695B (en)*2020-12-172023-03-24中国海洋大学Underwater target tracking method, system, storage medium, equipment, terminal and application
CN112560695A (en)*2020-12-172021-03-26中国海洋大学Underwater target tracking method, system, storage medium, equipment, terminal and application
CN113379804A (en)*2021-07-122021-09-10闽南师范大学Unmanned aerial vehicle target tracking method, terminal equipment and storage medium
CN113379804B (en)*2021-07-122023-05-09闽南师范大学 A UAV target tracking method, terminal equipment and storage medium
CN114708300A (en)*2022-03-022022-07-05北京理工大学Anti-blocking self-adaptive target tracking method and system
CN114708300B (en)*2022-03-022024-07-23北京理工大学 An adaptive target tracking method and system capable of resisting occlusion

Similar Documents

PublicationPublication DateTitle
CN107154024A (en)Dimension self-adaption method for tracking target based on depth characteristic core correlation filter
CN108734151B (en) Robust long-range target tracking method based on correlation filtering and deep Siamese network
CN107316316A (en)The method for tracking target that filtering technique is closed with nuclear phase is adaptively merged based on multiple features
CN104574445B (en)A kind of method for tracking target
CN105741316B (en)Robust method for tracking target based on deep learning and multiple dimensioned correlation filtering
CN108665481B (en) Adaptive anti-occlusion infrared target tracking method based on multi-layer deep feature fusion
CN111080675B (en)Target tracking method based on space-time constraint correlation filtering
CN112364931B (en)Few-sample target detection method and network system based on meta-feature and weight adjustment
CN104268594B (en)A kind of video accident detection method and device
CN106952288B (en)Based on convolution feature and global search detect it is long when block robust tracking method
CN107680119A (en)A kind of track algorithm based on space-time context fusion multiple features and scale filter
CN109461172A (en)Manually with the united correlation filtering video adaptive tracking method of depth characteristic
CN110175649B (en) A fast multi-scale estimation object tracking method for re-detection
CN108062531A (en)A kind of video object detection method that convolutional neural networks are returned based on cascade
CN111462191B (en) A deep learning-based method for unsupervised optical flow estimation with non-local filters
CN107369166A (en)A kind of method for tracking target and system based on multiresolution neutral net
CN104573731A (en)Rapid target detection method based on convolutional neural network
CN103295242A (en)Multi-feature united sparse represented target tracking method
CN107424177A (en)Positioning amendment long-range track algorithm based on serial correlation wave filter
CN109410251B (en)Target tracking method based on dense connection convolution network
CN110245587B (en) A target detection method for optical remote sensing images based on Bayesian transfer learning
CN110276784B (en)Correlation filtering moving target tracking method based on memory mechanism and convolution characteristics
CN110287760A (en) A method for occlusion detection of facial facial features based on deep learning
CN112419317A (en)Visual loopback detection method based on self-coding network
CN110555864B (en)Self-adaptive target tracking method based on PSPCE

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
RJ01Rejection of invention patent application after publication

Application publication date:20170912

RJ01Rejection of invention patent application after publication

[8]ページ先頭

©2009-2025 Movatter.jp