Movatterモバイル変換


[0]ホーム

URL:


CN113284166A - High-speed face tracking method and system based on deep cascade neural network - Google Patents

High-speed face tracking method and system based on deep cascade neural network
Download PDF

Info

Publication number
CN113284166A
CN113284166ACN202110403624.8ACN202110403624ACN113284166ACN 113284166 ACN113284166 ACN 113284166ACN 202110403624 ACN202110403624 ACN 202110403624ACN 113284166 ACN113284166 ACN 113284166A
Authority
CN
China
Prior art keywords
frame
face
neural network
face tracking
input
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110403624.8A
Other languages
Chinese (zh)
Inventor
不公告发明人
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Warmnut Beijing Technology Development Co ltd
Original Assignee
Warmnut Beijing Technology Development Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Warmnut Beijing Technology Development Co ltdfiledCriticalWarmnut Beijing Technology Development Co ltd
Priority to CN202110403624.8ApriorityCriticalpatent/CN113284166A/en
Publication of CN113284166ApublicationCriticalpatent/CN113284166A/en
Pendinglegal-statusCriticalCurrent

Links

Images

Classifications

Landscapes

Abstract

Translated fromChinese

本发明涉及一种基于深度级联神经网络的高速人脸跟踪方法及系统,其方法包括:S1:建立多层级联神经网络的人脸跟踪模型;S2:对人脸跟踪模型进行训练,得到训练好的人脸跟踪模型;S3:输入人脸视频帧到训练好的人脸跟踪模型,若该帧为第一帧或校准帧,通过完整的多层级联神经网络进行检测,得到人脸框位置;若该帧是后续帧,则将该帧的上一帧的输出的人脸候选框,输入多层级联神经网络的最后一层进行检测,得到人脸框位置。本发明提供的方法,充分利用人脸跟踪任务中帧与帧之间在图像特征与语义上有较大关联,因此利用上一帧中高层高分辨率级联神经网络输出的高精度人脸框作为先验信息,作为对下一帧的跟踪流程的基础,进而大大加速人脸跟踪过程。

Figure 202110403624

The invention relates to a high-speed face tracking method and system based on a deep cascaded neural network. The method includes: S1: establishing a face tracking model of a multi-level cascaded neural network; S2: training the face tracking model to obtain the trained A good face tracking model; S3: Input the face video frame to the trained face tracking model. If the frame is the first frame or calibration frame, it is detected through a complete multi-level cascaded neural network to obtain the position of the face frame. ; If the frame is a subsequent frame, input the output face candidate frame of the previous frame of the frame into the last layer of the multi-layer cascaded neural network for detection to obtain the face frame position. The method provided by the present invention makes full use of the large correlation between image features and semantics between frames in the face tracking task, so the high-precision face frame output by the high-level high-resolution cascaded neural network in the previous frame is used. As a priori information, it serves as the basis for the tracking process of the next frame, thereby greatly accelerating the face tracking process.

Figure 202110403624

Description

High-speed face tracking method and system based on deep cascade neural network
Technical Field
The invention relates to the field of image enhancement and computer vision, in particular to a high-speed face tracking method and system based on a deep cascade neural network.
Background
In the use scene of face tracking, the inferred speed of face tracking is directly reflected as the frame number of the processed and output video stream, thereby directly influencing the user experience. The traditional face tracking algorithm carries out matching of manual feature characterization through a traditional machine learning algorithm to determine the corresponding relation between an image area and a face image to be detected, but the method cannot fully utilize deep information of the face image, so that the tracking effect is poor; moreover, the depth tracking algorithm suffers from the complexity of calculation, and cannot achieve the high-speed real-time tracking effect in a production environment with limited calculation capacity.
Disclosure of Invention
In order to solve the technical problem, the invention provides a high-speed face tracking method and system based on a deep cascade neural network.
The technical solution of the invention is as follows: a high-speed face tracking method based on a deep cascade neural network comprises the following steps:
step S1: establishing a face tracking model comprising a multilayer cascade neural network;
step S2: training the face tracking model to obtain a trained face tracking model;
step S3: inputting a face video frame to the trained face tracking model, and if the face video frame is a first frame or a calibration frame, detecting through the complete multilayer cascade neural network to obtain a face frame position; and if the face video frame is a subsequent frame, inputting a face candidate frame output by the previous frame of the frame as input into the last layer of the multilayer cascade neural network for detection to obtain the position of the face frame. Compared with the prior art, the invention has the following advantages:
the high-speed face tracking method based on the deep cascade neural network provided by the invention fully utilizes the fact that the image characteristics and semantics of frames in a face tracking task are greatly related, so that a high-precision face frame output by a high-layer high-resolution cascade neural network in the previous frame is used as prior information and is used as the basis of a tracking flow of the next frame, and the face tracking process is greatly accelerated. The method provided by the invention can achieve the tracking speed of vision in real time (more than 24 frames per second), thereby realizing the face tracking task on a mobile terminal platform with limited calculation capacity.
Drawings
FIG. 1 is a flowchart of a high-speed face tracking method based on a deep cascade neural network in an embodiment of the present invention;
fig. 2 is a step S3 in the high-speed face tracking method based on the deep cascade neural network in the embodiment of the present invention: inputting a face video frame to the trained face tracking model, and if the face video frame is a first frame or a calibration frame, detecting through the complete multilayer cascade neural network to obtain a face frame position; if the face video frame is a subsequent frame, taking a face candidate frame output from the previous frame of the frame as input, and inputting the face candidate frame into the last layer of the multilayer cascade neural network for detection to obtain a flow chart of the position of the face frame;
fig. 3 is a block diagram of a high-speed face tracking system based on a deep cascade neural network in the embodiment of the present invention.
Detailed description of the preferred embodiments
The invention provides a high-speed face tracking method and a high-speed face tracking system based on a deep cascade neural network.
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings.
Example one
As shown in fig. 1, a high-speed face tracking method based on a deep cascade neural network provided in an embodiment of the present invention includes the following steps:
step S1: establishing a face tracking model comprising a multilayer cascade neural network;
step S2: training a face tracking model to obtain a trained face tracking model;
step S3: inputting a face video frame to a trained face tracking model, and if the face video frame is a first frame or a calibration frame, detecting through a complete multilayer cascade neural network to obtain a face frame position; and if the face video frame is a subsequent frame, taking the face candidate frame output by the previous frame of the frame as input, and inputting the face candidate frame into the last layer of the multilayer cascade neural network for detection to obtain the position of the face frame.
In one embodiment, the step S1: establishing a face tracking model comprising a multilayer cascade neural network, which specifically comprises the following steps:
establishing a multilayer cascade neural network, wherein different levels of the multilayer cascade neural network respectively receive corresponding images with different resolution scales and output gradually accurate positions of the face candidate frames; and each layer of neural network adjusts the resolution of the face candidate frame output by the previous layer of neural network and then takes the face candidate frame as the input of the layer.
The face tracking model constructed by the invention comprises a cascade neural network of a plurality of layers of independent networks. The networks are mutually independent, and are more and more distant from the input image and closer to the output result from a lower layer to a higher layer network; the lower layer network receives the low-resolution input image and outputs a tracking result with lower precision, and the higher layer network receives the high-resolution input network and outputs a high-precision tracking result.
In one embodiment, the step S2: training a face tracking model to obtain the trained face tracking model, and specifically comprising:
and performing end-to-end training by using training data, wherein the training data is screened according to the score of the face candidate box after each iteration process is finished, and meanwhile, the proportion of positive and negative samples is adjusted.
The invention adopts a training mode of a face detection network to train the face tracking model. End-to-end training is performed by using a large amount of training data generated offline. After each iteration process, the training data is screened according to the score of the face candidate box, meanwhile, the proportion of positive and negative samples is properly adjusted according to the actual situation, particularly the proportion of the difficult samples is increased, and the positive samples with lower scores and the negative samples with higher scores are trained.
As shown in fig. 2, in one embodiment, the step S3: inputting a face video frame to the trained face tracking model, and if the face video frame is a first frame or a calibration frame, detecting through the complete multilayer cascade neural network to obtain a face frame position; if the face video frame is a subsequent frame, taking a face candidate frame output from the previous frame of the frame as an input, and inputting the face candidate frame into the last layer of the multilayer cascade neural network for detection to obtain the position of the face frame, specifically comprising the following steps:
step S31: judging the input human face video frame, and if the frame is the first frame or the calibration frame, jumping to the step S32; if not, jumping to step S34;
when the face tracking is actually carried out, a face video frame is input to the trained face tracking model. Firstly, judging an input human face video frame, and if the frame is a first frame or a calibration frame, jumping to the step S32; if not, go to step S34.
Step S32: the frame is detected through a first layer of a multilayer cascade neural network, and is screened according to a non-maximum suppression algorithm to obtain a face candidate frame;
when the input frame is the first frame or the calibration frame, the frame is detected by the complete multilayer cascade neural network. Firstly, detecting through a first layer of a multilayer cascade neural network, screening according to a non-maximum value inhibition algorithm, screening face frames with more coincidence and lower scores, namely selecting the rest face frames with the coincidence proportion lower than a threshold value as an output result to obtain a face candidate frame, and cutting a face image with a corresponding resolution scale according to the result to be used as the input of a subsequent high-resolution cascade neural network; if the frame is not the first frame or the calibration frame, then the process directly jumps to step S34.
Step S33: inputting the face candidate frame into a next layer of cascade neural network, and screening according to a non-maximum suppression algorithm to obtain the face candidate frame;
and (8) inputting the screened face candidate frame in the step (S32) into the next layer of cascade neural network for face judgment, further screening according to a non-maximum suppression algorithm to obtain a face candidate frame, and cutting a face image with a corresponding resolution scale according to the result to be used as the input of the subsequent high-resolution cascade neural network.
Step S34: repeating the step S33 until the last layer of the cascade neural network outputs the position of the face frame;
and step S33 is repeated until the last layer of the cascade neural network is reached, similarly, screening is carried out according to a non-maximum value inhibition algorithm, and finally, the face candidate frame with the coincidence proportion lower than the threshold value is selected as a result, and the face frame position is output.
Step S35: and inputting the output obtained by the last frame of the subsequent frame in the last layer of the cascade neural network into the last layer of the cascade neural network to obtain a face candidate frame, screening according to the non-maximum suppression algorithm, and outputting the position of the face frame.
And for the subsequent frames which are not the first frame or the calibration frame, taking the output of the last frame in the last layer of the cascade neural network as the input, inputting the output into the last layer of the cascade neural network to obtain a face candidate frame, screening according to the non-maximum inhibition algorithm, and outputting the position of the face frame.
The high-speed face tracking method based on the deep cascade neural network provided by the invention fully utilizes the fact that the image characteristics and semantics of frames in a face tracking task are greatly related, so that a high-precision face frame output by a high-layer high-resolution cascade neural network in the previous frame is used as prior information and is used as the basis of a tracking flow of the next frame, and the face tracking process is greatly accelerated. The method provided by the invention can achieve the tracking speed of vision in real time (more than 24 frames per second), thereby realizing the face tracking task on a mobile terminal platform with limited calculation capacity.
Example two
As shown in fig. 3, an embodiment of the present invention provides a high-speed face tracking system based on a deep cascade neural network, including the following modules:
the model construction module is used for establishing a face tracking model comprising a multilayer cascade neural network;
the model training module is used for training the face tracking model to obtain a trained face tracking model;
the face tracking module is used for inputting a face video frame to the trained face tracking model, and if the face video frame is a first frame or a calibration frame, the face video frame is detected through a complete multilayer cascade neural network to obtain the position of a face frame; and if the face video frame is a subsequent frame, taking the face candidate frame output by the previous frame of the frame as input, and inputting the face candidate frame into the last layer of the multilayer cascade neural network for detection to obtain the position of the face frame.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
The above examples are provided only for the purpose of describing the present invention, and are not intended to limit the scope of the present invention. The scope of the invention is defined by the appended claims. Various equivalent substitutions and modifications can be made without departing from the spirit and principles of the invention, and are intended to be within the scope of the invention.

Claims (5)

Translated fromChinese
1.一种基于深度级联神经网络的高速人脸跟踪方法,其特征在于,包括:1. a high-speed face tracking method based on deep concatenated neural network, is characterized in that, comprises:步骤S1:建立包括多层级联神经网络的人脸跟踪模型;Step S1: establish a face tracking model including a multi-layer cascaded neural network;步骤S2:对所述人脸跟踪模型进行训练,得到训练好的人脸跟踪模型;Step S2: training the face tracking model to obtain a trained face tracking model;步骤S3:输入人脸视频帧到所述训练好的人脸跟踪模型,如果所述人脸视频帧为第一帧或校准帧,则通过完整的所述多层级联神经网络进行检测,得到人脸框位置;如果所述人脸视频帧是后续帧,则将该帧的上一帧的输出的人脸候选框作为输入,输入所述多层级联神经网络的最后一层进行检测,得到人脸框位置。Step S3: input the face video frame to the trained face tracking model, if the face video frame is the first frame or the calibration frame, then detect through the complete described multi-level cascaded neural network to obtain the human face. face frame position; if the face video frame is a subsequent frame, the face candidate frame of the output of the previous frame of the frame is used as an input, and the last layer of the multi-layer cascaded neural network is input to detect, and the human face is obtained. face frame position.2.根据权利要求1所述的基于深度级联神经网络的高速人脸跟踪方法,其特征在于,所述步骤S1:建立包括多层级联神经网络的人脸跟踪模型,具体包括:2. the high-speed face tracking method based on deep cascaded neural network according to claim 1, is characterized in that, described step S1: establish the face tracking model that comprises multi-level cascaded neural network, specifically comprises:建立多层级联神经网络,其中,所述多层级联神经网络的不同层级分别接收对应的不同分辨率尺度的图像,并输出逐渐精确的所述人脸候选框位置;每层神经网络将上一层神经网络输出的所述人脸候选框,调整其分辨率后,作为该层的所述输入。A multi-level cascaded neural network is established, wherein different levels of the multi-level cascaded neural network receive corresponding images of different resolution scales respectively, and output the gradually accurate position of the face candidate frame; The face candidate frame output by the layer neural network, after adjusting its resolution, is used as the input of the layer.3.根据权利要求1所述的基于深度级联神经网络的高速人脸跟踪方法,其特征在于,所述步骤S2:对所述人脸跟踪模型进行训练,得到训练好的人脸跟踪模型,具体包括:3. the high-speed face tracking method based on deep cascading neural network according to claim 1, is characterized in that, described step S2: described face tracking model is trained, obtains trained face tracking model, Specifically include:使用训练数据进行端对端训练,其中,所述训练数据在每次迭代过程结束后根据所述人脸候选框的得分值进行筛选,同时,调整正负样本的比例。End-to-end training is performed using training data, wherein the training data is screened according to the score value of the face candidate frame after each iteration process, and at the same time, the ratio of positive and negative samples is adjusted.4.根据权利要求1所述的基于深度级联神经网络的高速人脸跟踪方法,其特征在于,所述步骤S3:输入人脸视频帧到所述训练好的人脸跟踪模型,如果所述人脸视频帧为第一帧或校准帧,则通过完整的所述多层级联神经网络进行检测,得到人脸框位置;如果所述人脸视频帧是后续帧,则将该帧的上一帧的输出的人脸候选框作为输入,输入所述多层级联神经网络的最后一层进行检测,得到人脸框位置,具体包括下述步骤:4. the high-speed face tracking method based on deep cascading neural network according to claim 1, is characterized in that, described step S3: input face video frame to described trained face tracking model, if described The face video frame is the first frame or the calibration frame, then the complete multi-layer cascaded neural network is used for detection to obtain the position of the face frame; if the face video frame is a subsequent frame, the previous frame of the frame is detected. The face candidate frame of the output of the frame is used as input, and the last layer of the multi-layer cascaded neural network is input to detect, and the position of the face frame is obtained, which specifically includes the following steps:步骤S31:判断输入的所述人脸视频帧,如果该帧为第一帧或校准帧,则该帧通过所述多层级联神经网络的第一层进行检测,并根据非极大值抑制算法进行筛选,得到人脸候选框;如果否,则跳至步骤S34;Step S31: Judging the input face video frame, if the frame is the first frame or the calibration frame, then the frame is detected by the first layer of the multi-layer cascaded neural network, and the non-maximum value suppression algorithm is used to detect the frame. Perform screening to obtain a face candidate frame; if not, skip to step S34;步骤S32:将所述人脸候选框输入下一层所述级联神经网络,并根据所述非极大值抑制算法进行筛选,得到所述人脸候选框;Step S32: Input the face candidate frame into the next layer of the cascaded neural network, and perform screening according to the non-maximum suppression algorithm to obtain the face candidate frame;步骤S33:重复步骤S32,直至所述级联神经网络的最后一层,输出人脸框位置;Step S33: Repeat step S32 until the last layer of the cascaded neural network, and output the position of the face frame;步骤S34:将所述后续帧的上一帧在所述级联神经网络的最后一层得到的输出,输入所述级联神经网络的最后一层,得到所述人脸候选框,并根据所述非极大值抑制算法进行筛选,输出所述人脸框位置。Step S34: Input the output of the last frame of the subsequent frame obtained in the last layer of the cascaded neural network into the last layer of the cascaded neural network to obtain the face candidate frame, and according to the The non-maximum value suppression algorithm is used for screening, and the position of the face frame is output.5.一种基于深度级联神经网络的高速人脸跟踪系统,其特征在于,包括下述模块:5. a high-speed face tracking system based on deep concatenated neural network, is characterized in that, comprises following module:模型构建模块,用于建立包括多层级联神经网络的人脸跟踪模型;A model building module for building a face tracking model including a multi-level cascaded neural network;模型训练模块,用于对所述人脸跟踪模型进行训练,得到训练好的人脸跟踪模型;a model training module for training the face tracking model to obtain a trained face tracking model;人脸跟踪模块,用于输入人脸视频帧到所述训练好的人脸跟踪模型,如果所述人脸视频帧为第一帧或校准帧,则通过完整的所述多层级联神经网络进行检测,得到人脸框位置;如果所述人脸视频帧是后续帧,则将该帧的上一帧的输出的人脸候选框作为输入,输入所述多层级联神经网络的最后一层进行检测,得到人脸框位置。The face tracking module is used to input the face video frame to the trained face tracking model, if the face video frame is the first frame or the calibration frame, then carry out through the complete described multi-level cascaded neural network Detect to obtain the position of the face frame; if the face video frame is a subsequent frame, then the face candidate frame of the output of the previous frame of the frame is used as input, and the last layer of the multi-layer cascaded neural network is input to carry out Detect to get the face frame position.
CN202110403624.8A2021-04-152021-04-15High-speed face tracking method and system based on deep cascade neural networkPendingCN113284166A (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202110403624.8ACN113284166A (en)2021-04-152021-04-15High-speed face tracking method and system based on deep cascade neural network

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202110403624.8ACN113284166A (en)2021-04-152021-04-15High-speed face tracking method and system based on deep cascade neural network

Publications (1)

Publication NumberPublication Date
CN113284166Atrue CN113284166A (en)2021-08-20

Family

ID=77276844

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202110403624.8APendingCN113284166A (en)2021-04-152021-04-15High-speed face tracking method and system based on deep cascade neural network

Country Status (1)

CountryLink
CN (1)CN113284166A (en)

Citations (13)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN104331903A (en)*2014-10-252015-02-04北京航空航天大学Moving-target progressive locating and tracking method based on water strider food location
CN104966095A (en)*2015-06-032015-10-07深圳一电科技有限公司Image target detection method and apparatus
CN106250850A (en)*2016-07-292016-12-21深圳市优必选科技有限公司 Face detection and tracking method and device, robot head rotation control method and system
US20170339409A1 (en)*2016-05-232017-11-23Intel CorporationFast and robust human skin tone region detection for improved video coding
CN108171196A (en)*2018-01-092018-06-15北京智芯原动科技有限公司A kind of method for detecting human face and device
CN108334885A (en)*2018-02-052018-07-27湖南航升卫星科技有限公司A kind of video satellite image space object detection method
WO2018187622A1 (en)*2017-04-052018-10-11Lyrical Labs Holdings, LlcVideo processing and encoding
CN108875506A (en)*2017-11-172018-11-23北京旷视科技有限公司Face shape point-tracking method, device and system and storage medium
CN109919096A (en)*2019-03-082019-06-21广州二元科技有限公司A kind of method of video real-time face detection
CN110991287A (en)*2019-11-232020-04-10深圳市恩钛控股有限公司Real-time video stream face detection tracking method and detection tracking system
CN111259742A (en)*2020-01-092020-06-09南京理工大学Abnormal crowd detection method based on deep learning
CN111915639A (en)*2020-08-062020-11-10广州市百果园信息技术有限公司 Target detection and tracking method, device, electronic device and storage medium
CN112465872A (en)*2020-12-102021-03-09南昌航空大学Image sequence optical flow estimation method based on learnable occlusion mask and secondary deformation optimization

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN104331903A (en)*2014-10-252015-02-04北京航空航天大学Moving-target progressive locating and tracking method based on water strider food location
CN104966095A (en)*2015-06-032015-10-07深圳一电科技有限公司Image target detection method and apparatus
US20170339409A1 (en)*2016-05-232017-11-23Intel CorporationFast and robust human skin tone region detection for improved video coding
CN106250850A (en)*2016-07-292016-12-21深圳市优必选科技有限公司 Face detection and tracking method and device, robot head rotation control method and system
WO2018187622A1 (en)*2017-04-052018-10-11Lyrical Labs Holdings, LlcVideo processing and encoding
CN108875506A (en)*2017-11-172018-11-23北京旷视科技有限公司Face shape point-tracking method, device and system and storage medium
CN108171196A (en)*2018-01-092018-06-15北京智芯原动科技有限公司A kind of method for detecting human face and device
CN108334885A (en)*2018-02-052018-07-27湖南航升卫星科技有限公司A kind of video satellite image space object detection method
CN109919096A (en)*2019-03-082019-06-21广州二元科技有限公司A kind of method of video real-time face detection
CN110991287A (en)*2019-11-232020-04-10深圳市恩钛控股有限公司Real-time video stream face detection tracking method and detection tracking system
CN111259742A (en)*2020-01-092020-06-09南京理工大学Abnormal crowd detection method based on deep learning
CN111915639A (en)*2020-08-062020-11-10广州市百果园信息技术有限公司 Target detection and tracking method, device, electronic device and storage medium
CN112465872A (en)*2020-12-102021-03-09南昌航空大学Image sequence optical flow estimation method based on learnable occlusion mask and secondary deformation optimization

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
ZHICHAO LIAN等: "A Real Time Face Tracking System based on Multiple Information Fusion", 《 MULTIMEDIA TOOLS AND APPLICATIONS》, vol. 79, 5 May 2020 (2020-05-05), pages 16751 - 16769, XP037180489, DOI: 10.1007/s11042-020-08889-4*
康一梅等: "目标尺度自适应的Mean Shift跟踪算法", 《兵工学报》, vol. 32, no. 2, 15 February 2011 (2011-02-15), pages 210 - 216*
濮柯佳: "基于多特征融合的多目标跟踪算法研究及应用", 《中国优秀硕士学位论文全文数据库_信息科技辑》, 15 June 2020 (2020-06-15), pages 138 - 994*

Similar Documents

PublicationPublication DateTitle
Prakash et al.It GAN do better: GAN-based detection of objects on images with varying quality
CN110619319A (en)Improved MTCNN model-based face detection method and system
CN111105375B (en)Image generation method, model training method and device thereof, and electronic equipment
CN111695421B (en)Image recognition method and device and electronic equipment
CN108648253B (en)Method and device for generating dynamic picture
WO2023050650A1 (en)Animation video generation method and apparatus, and device and storage medium
EP3712825A1 (en)Model prediction method and device
KR20160134019A (en)Method for training top-down selective attention in artificial neural networks
CN112861659A (en)Image model training method and device, electronic equipment and storage medium
CN110503957A (en)A kind of audio recognition method and device based on image denoising
CN117315387A (en)Industrial defect image generation method
CN114038045B (en)Cross-modal face recognition model construction method and device and electronic equipment
CN114519880B (en)Active speaker recognition method based on cross-modal self-supervision learning
US20220366262A1 (en)Method and apparatus for training neural network model
CN109005402A (en)A kind of appraisal procedure and device of video
CN112132797A (en)Short video quality screening method
CN114882234A (en)Construction method of multi-scale lightweight dense connected target detection network
KR20220136884A (en) Point Cloud Completion Network Creation and Point Cloud Data Processing
CN114529484B (en)Deep learning sample enhancement method for direct current component change in imaging
CN113361568B (en) Target recognition method, device and electronic system
CN113906443B (en) Point cloud data completion and point cloud data processing
CN113284166A (en)High-speed face tracking method and system based on deep cascade neural network
CN111144497B (en) Image saliency prediction method under multi-task deep network based on aesthetic analysis
CN115631192B (en)Control method, device, equipment and medium for valve pressure tester
CN110490165B (en)Dynamic gesture tracking method based on convolutional neural network

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
WD01Invention patent application deemed withdrawn after publication
WD01Invention patent application deemed withdrawn after publication

Application publication date:20210820


[8]ページ先頭

©2009-2025 Movatter.jp