Movatterモバイル変換


[0]ホーム

URL:


CN109409307A - A kind of Online Video behavioral value system and method based on space-time contextual analysis - Google Patents

A kind of Online Video behavioral value system and method based on space-time contextual analysis
Download PDF

Info

Publication number
CN109409307A
CN109409307ACN201811298487.0ACN201811298487ACN109409307ACN 109409307 ACN109409307 ACN 109409307ACN 201811298487 ACN201811298487 ACN 201811298487ACN 109409307 ACN109409307 ACN 109409307A
Authority
CN
China
Prior art keywords
behavior
video
behavioral
frame
online
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811298487.0A
Other languages
Chinese (zh)
Other versions
CN109409307B (en
Inventor
李楠楠
张世雄
张子尧
李革
安欣赏
张伟民
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Longgang Intelligent Audiovisual Research Institute
Original Assignee
Shenzhen Longgang Intelligent Audiovisual Research Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Longgang Intelligent Audiovisual Research InstitutefiledCriticalShenzhen Longgang Intelligent Audiovisual Research Institute
Priority to CN201811298487.0ApriorityCriticalpatent/CN109409307B/en
Publication of CN109409307ApublicationCriticalpatent/CN109409307A/en
Application grantedgrantedCritical
Publication of CN109409307BpublicationCriticalpatent/CN109409307B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Classifications

Landscapes

Abstract

A kind of Online Video behavioral value method based on space-time contextual analysis, this method realizes the on-line checking of generation behavior in input video using deep learning frame and in conjunction with space-time contextual analysis technology, and the detection is carried out in time-domain and spatial domain joint.The present invention includes two parts: the link between behavioral value and video clip in video clip.Algorithm utilizes coding-decoded model in video clip, in conjunction with present frame and space-time dynamic information, generates candidate actions region;Candidate actions Regional links at behavioral chain, which gives more sustained attention specified action object for link between video clip, occurs from it until terminating, while predicting the classification of behavior on-line manner.

Description

A kind of Online Video behavioral value system and method based on space-time contextual analysis
Technical field
The present invention relates to video behavior analysis technical fields, and in particular to a kind of based on the online of space-time contextual analysisVideo behavior detection system and its method, this method are realized using deep learning frame and in conjunction with space-time contextual analysis technologyThe on-line checking of behavior occurs in input video, and the detection is carried out in time-domain and spatial domain joint.
Background technique
Video behavioral value not only will correctly classify to the behavior occurred in given video, also carry out to behaviorPositioning in time-domain and spatial domain is the key that video human behavior understands a step in research.Briefly, presently, there areMethod generallys use the program of two steps processing to solve this problem: being generated with the motion detector by re -training singleThe motion detection of frame is as a result, include the object frame by recurrence and corresponding classification of motion score;By continuing in entire videoThe motion detection result that single frames is perhaps tracked in connection in time forms final space-time action chain usually these connections or trackingBe carried out under some constraint conditions, such as: it is required that the motion detection frame overlapping region of consecutive frame is big as far as possible.It is thisThe limitation of processing is mainly reflected in two aspects: 1) it carries out the row of single frames just with current image or motion informationTo detect, and have ignored the continuity of action behavior in time;2) join algorithm usually uses a kind of offline and batch processing sideFormula carries out, namely: action chain is extended to always video trailer since video, then with other time-domain pruning algorithms comeEliminate false testing result.In the present invention, above-mentioned two problems are solved by following approach: 1) combines present frame and space-timeInformation above carries out motion detection down;2) by the way of on-line checking, the generation of consummatory behavior chain and behavior in single treatmentClassification prediction.
2017, Zhu et al. (Zhu H., Vial R., and Lu S.2017. " A Spatio-temporalConvolutional Regression Network for Video Action Proposal”,IEEEInternational Conference on Computer Vision, pp.5814-5822) propose a kind of Recurrent networks mouldType is for generating action behavior proposal, and the model is with Conv LSTM (Convolutional Long Short-TermMemory it is constructed based on), has merged space-time dynamic information and current frame information carries out motion detection.This method lacksIt is trapped in in a bit of video, usually only the video frame of position rearward can use space-time dynamic information and assist current inspectionIt surveys.
Summary of the invention
The Online Video behavioral value system based on space-time contextual analysis that the object of the present invention is to provide a kind of, so that rightIt can use video sequence contextual information, while continually entering with video frame when the action behavior of present frame is detectedBehavioral chain can be incrementally generated, and is dynamically classified to video behavior.
The Online Video behavioral value method based on space-time contextual analysis that it is a further object of the present invention to provide a kind of.
Method proposed by the present invention and have the main improvement of two o'clock compared with the conventional method: 1) method of the invention is based onConvGRU (Convolutional Gated Recurrent Unit), compared to ConvLSTM, it is a kind of following for lightweightRing memory models, the parameter with much less, reduce the risk of over-fitting on small sample quantity collection;2) their model isSingle forward model, therefore when video frame only positioned at input video sequence rear end does behavioral value, could be using fusionSpace-time dynamic information, and method proposed by the present invention is a kind of coding-decoded model, the spatio-temporal context information of video sequence existsIt can be all used in each frame when decoding.
The principle of the present invention is: 1) using depth convolutional neural networks extract single frame video feature, it will continuous several frame viewsFrequency feature is input to ConvGRU building coding-decoded video sequence descriptive model, the behavior space-time context in forward direction transmittingInformation is encoded, and is decoded to each frame in the rear space-time dynamic information encoded into transmitting, is completed in conjunction with current frame information dynamicIt detects;2) maintain a dynamic behavior classification candidate pool that possibility is gradually reduced with the continuous growth of input video sequenceBehavior class scope, while the behavioral chain being currently generated dynamically is trimmed, comprising: increase, terminate and time-domain is repairedIt cuts.
Technical solution provided by the invention is as follows:
Temporal-spatial field behavioral value method proposed by the present invention includes two parts: the behavioral value in video clipLink between video clip.Algorithm utilizes coding-decoded model in video clip, believes in conjunction with present frame and space-time dynamicBreath generates candidate actions region;Candidate actions Regional links at behavioral chain, which gives more sustained attention specified for link between video clipAction object, occur while predicting the classification of behavior on-line manner until terminating from it.
A kind of Online Video behavioral value system based on space-time contextual analysis, including video behavior space-time context letterIt ceases converged network and motion frame links online and sorting algorithm;Wherein: the video behavior spatio-temporal context information converged network,For behavior spatio-temporal context information in current frame information and a video clip to be merged;The motion frame links onlineAnd sorting algorithm forms one completely for the motion frame for corresponding to same moving target to be chained up on-line mannerBehavioral chain, and classify to its behavior classification.
The video behavior spatio-temporal context information converged network specifically includes: single frames feature extraction network, for extractingPresent frame RGB image and light stream picture depth expressing feature in video clip;Video clip spatio-temporal context information converged network,Coding-decoder module of the building based on ConvGRU model for extracting video clip space-time context expressing feature, and be allowed toVideo present frame feature is merged, and fusion feature is obtained;Behavioral value network, for carrying out single frames behavior on fusion featureDetection, the position for obtaining behavior classification score and behavior occurring position, and generate motion frame.
The motion frame link online and sorting algorithm specifically includes: building behavior classification candidate pool, for maintain forThe behavior classification that given video currently most possibly occurred specify number;Behavior classification candidate pool more new algorithm, for rowIt gives a mark for classification, behavior class scope belonging to current video possibility is gradually reduced, realizes the online Fast Classification of behavioral chain;The online growth algorithm of behavioral chain is realized for the corresponding behavior candidate region of video clip to be linked in already present behavioral chainBehavioral chain increases online;Or behavior candidate region is determined as to new behavioral chain.
A kind of Online Video behavioral value method based on space-time contextual analysis, including following several steps:
Step 1: light stream image being calculated to present frame, extracts the depth expressing feature of RGB image and light stream image;
Step 2: building coding-decoding network extracts video behavior spatio-temporal context information, and carries out with current frame informationFusion, obtains fusion feature;
Step 3: to fusion feature carry out classification and position return, generate motion frame, with Viterbi algorithm to motion frame intoRow link, obtains behavior candidate region;
Step 4: building behavior classification candidate pool updates the behavior classification being likely to occur;
Step 5: behavior candidate region being linked in already present behavioral chain on-line manner or generates new behaviorChain;
Step 6: the testing result of RGB image branch and light stream image branch being merged, final detection knot is obtainedFruit.
Compared with prior art, the beneficial effects of the present invention are:
Piece of video is utilized when carrying out behavioral value to video single-frame images using technical solution provided by the inventionBehavior spatio-temporal context information in section, improves the accuracy rate of behavioral value;Online inspection can be carried out to video behavior simultaneouslyIt surveys, compared to the previous offline method based on batch processing mode, improves the timeliness of video behavioral value, can be applied to pairThe relatively high occasion of requirement of real-time, such as intelligent robot, man-machine interactive system.Compared with video behavior detection technique phaseThan on currently a popular open test collection, technology provided by the invention is achieved using less Candidate SubmissionBetter detection effect.
With reference to the accompanying drawing, by embodiment, the present invention is further described:
Detailed description of the invention
Fig. 1 is flow chart of the invention.
Fig. 2 is video unit motion information encoding-decoder En-Decoder model framework figure.
Fig. 3 is the video behavior single frame detection model framework figure based on space-time contextual analysis.
Fig. 4 is video behavioral chain set TdOnline dynamic updates operational flowchart.
In attached drawing:
1-single-frame images expressing feature p 'i, 2-ConVGRU units, 3-fusion expressing feature pd, 4-video unit institutesInclude image sequence, 5-feature extraction networks, 6-dimensionality reduction networks, 7-RPN networks, 8-Detection Network, 9-Score, 12-movement proposals, the cutting of 13-time-domains, 14-are proposed in behavior classification results, 10-position adjustment amounts, 11-movementsCalculating behavior score, 15-building behavior candidate pools, 16-building candidate collection Pt, 17-regeneration behavior chains, 18-increase it is newBehavioral chain.
Specific embodiment
A kind of Online Video behavioral value method based on space-time contextual analysis of the invention, embodiment mode is by such asLower step carries out:
1) one section of video sequence to be detected is divided evenly that (8 frames are a segment, adjacent segment for several video clipsBetween have a frame overlapping);
2) Optic flow information, original RGB image and optical-flow light stream image difference are extracted to each video clipTwo independent calculating branches of composition in model are input to, are illustrated by taking one of RGB branch as an example below, optical-Flow branch situation is identical;
3) every frame image is separately input to a depth convolutional network trained in advance (as behavior point in video clipClass) carry out motion feature extraction;
4) motion feature extracted is input in coding-decoding network that one is made of ConvGRU and extracts video lineIt for spatio-temporal context information, and is merged with the motion information of present frame, every frame exports fusion feature;
5) fusion feature is by an access one behavior sorter network and position Recurrent networks, the behavior occurred to every frame intoRow classification, while the position that behavior occurs positions, and generates behavior frame;
6) it is linked, is constituted by behavior classification with the behavior frame that Viterbi algorithm detects frame every in video clipSeveral behavior candidate regions, following steps (7-9) circulation executes, until input video sequence terminates;
If 7) multiple that current video segment ordinal number is 10, process performing chain time-domain Pruning Algorithm calculates behavioral chain phaseFor the score of each behavior classification, regeneration behavior classification candidate pool, making it only includes the maximum several behavior classifications of score;
8) to certain current behavioral chain, behavior candidate region overlaps area (overlapping area refers to that behavioral chain is last if it existsOverlapping area between one behavior frame and the behavior frame of candidate region first) be greater than specified threshold, then the maximum row of scoreBehavior chain is linked to for candidate region;Behavior candidate region overlaps area greater than specified threshold if it does not exist, then terminatesBehavior chain.This step operation executes respectively for different classes of in behavior classification candidate pool;
9) behavior candidate region is not if it exists and any behavioral chain constitutes link, then using behavior candidate region as oneNew behavioral chain;
10) the behavioral chain testing result of RGB branch and Optical-flow branch is merged, obtains final detectionAs a result.
Fig. 1 is flow chart of the invention, wherein S1-S8 is corresponding in turn in specific embodiment step 1)-8).It is a kind ofOnline Video behavioral value method based on space-time contextual analysis, concrete operations process are now described below:
1) video is evenly dividing into segment S1: one section of input video is given, several video clips are evenly divided into,Every section includes 8 frame images.Each section is used as independent video unit, extracts behavior candidate region;
2) RGB image or light stream S2 are extracted: for each video unit, extracting wherein every frame optical-flow light streamImage is described as motion information.Original RGB image and light stream image is separately input in model as two independent pointZhi Jinhang is calculated.It is illustrated based on RGB image branch below, optical-flow light stream branch is same;
3) single frames expressing feature p ' is extractediS3: the behavioral value network frame of single-frame images is as shown in Fig. 3.It is shown in Fig. 3The image 4 that a video unit is included is shown.Every frame image extracts expressing feature with feature extraction network 5, is denoted as pi.FeatureIt extracts network and is based on VGG-16 model (Simonyan K.and Zisserman A.2014.Very DeepConvolutional Networks for Large Scale Image Recognition.ArXiv(2014).https://Doi.org/arXiv:1409.1556) fine tuning training obtains, and takes its conv5 layers of feature.A dimensionality reduction network 6 is constructed, by pi'sFeature number of layers drops to 128 from 512, is denoted as p 'i, prevent overall network model over-fitting.Dimensionality reduction network is by 128 layers of convolutional layerThe convolution module of composition;
4) fusion expressing feature p is extracteddS4: by the expressing feature p ' of every frame imageiIt is input to ConvGRU network 2, carrys out structureIt is motion encoded to build video unit temporal-spatial fusion.Motion information codec En-Decoder model structure is as shown in Fig. 2, Fig. 2Show present frame input feature vector p 'i1 and fusion expressing feature pd2.En-Decoder model action time range is entire videoUnit, including forward coding and backward decoding process.Feature p 'iForward coding and backward decoding process are simultaneously participated in, it is specific defeatedIt is as shown in Figure 2 to enter mode.Feature p ' of the forward coding to single framesiIt is accumulated, is obtained to video unit motion sequence at any timeCharacterization;Backward decoding motion sequence characterization propagate backward to each frame in video unit and with feature p 'iIt is merged,The feature p of present frame and spatio-temporal context information is mergedd
5) the output S5 of Detection Networka and RPN network is calculated: feature pdIt is input to RPN network 7, is passed throughOperation obtains movement and proposes score 11, is denoted as sr, and movement proposal 12, it is denoted as pr.RPN network is one 2 layers of 3*3 convolution netNetwork, in pdUpper sliding calculates movement at each position and proposes score sr, area of the score value greater than a designated value (such as 0.5)Domain is considered as that p is proposed in movementr.Detection Network8 receives pdAnd prAs input, behavior classification results 9, note are exportedFor scAnd position adjustment amount 10, it is denoted as δr.Detection Network is by 2 layers of full articulamentum structure comprising 1024 hidden unitsAt behavior classification results scInclude the classification score to each class behavior and background classes, position adjustment amount δrEach class behavior is givenCorresponding 3 position deviations (center, width and height) out.By prAnd δrIt can calculate by modified behavior candidate frame bt
6) computer video unit behavior candidate region P S6: note btCorresponding RPN movement is proposed to be scored at sr(bt),Using Viterbi algorithm b on different images frame in same video unittIt is chained up, behavior candidate region p is obtained, such as formula(1) shown in:
TpIt is video unit persistence length, is taken as 8 here;For btAnd bt-1Between friendship and compare IntersectionOver Union(IoU);For harmonic coefficient, it is taken as 0.5;
7) video behavioral chain set T is calculateddS7: it as video constantly inputs, obtains corresponding to each video unitBehavior candidate region p obtains the increased video behavioral chain set T of dynamic by following rule (a)-(f)d.Fig. 4 is video behaviorChain set TdOnline dynamic updates operational flowchart, forExecuting rule (a)-(f), main thought are as follows: maintain oneThe behavior classification candidate pool that a dynamic updates, differentiates according to the video continually entered, and it is candidate gradually to reduce behavior classificationNumber;Determine that newly generated candidate region p is linked to original video behavioral chain or work according to the link method of settingFor new behavioral chain.As shown in figure 4, comprising the concrete steps that:
(a) time-domain cuts 13.If current collection TdElement number > upper limit Nd, update terminates.Otherwise, it utilizesViterbi algorithm carries out time-domain cutting to T, as shown in formula (2):
TlIt include b by behavioral chain TtNumber;lt∈ { 0, c } is btGeneric, 0 represents backgroundClass, c represent behavior classification c;If lt=c, thenFor btCorresponding Detection Network classification c classification score sc(bt), if lt=0,It is defined as 1-sc(bt);If lt=lt-1, ω (lt, lt-1)=0, otherwise, ω (lt, lt-1)=0.7;λω=0.4;It is cut by time-domain, background block included in T will be reduced;
(b) it calculates behavior score 14: for T, calculating its score s (T) relative to each behavior classification, s (T) is defined asBelong to the average value of all p score s (p) of T;Similarly, the score s (p) of p is defined as belonging to all b of ptScore sc(bt) averageValue;
(c) it constructs behavior candidate pool 15: constructing a behavior classification candidate pool in descending order according to s (T), specificallyAre as follows: when i) starting, retain all categories;When ii) handling the 10th video unit, retain preceding 5 behavior classifications;Iii) processing theWhen 20 video units, retain preceding 3 behavior classifications;Iv) handle the 30th video unit and after, only retain and rank the firstBehavior classification.If the current candidate pool behavior classification upper limit is Np, for each of candidate pool behavior classification j≤Np, execute ruleThen (d)-(e):
(d) candidate collection P is constructedt16: new generation behavior candidate region p is set, if the IoU between T and p is greater than specified thresholdThen set P is added in p by (such as 0.5)t(initial ptFor sky).The last one behavior that IoU between T and p is defined as belonging to T is candidateIoU between frame and first behavior candidate frame of p;
(e) regeneration behavior chain 17: if PtIt is not sky, then the (p ' ∈ P of p ' corresponding to score maximum s (p ')t) it is linked to T,P ' is added to behind T, forms new behavioral chain T ', updates T=T ';
(f) increase new behavioral chain 18: p if it existsnew(pnew∈Pt) be not linked in any one T, then pnewAsSet T is added in one new behavioral chaind
8) RGB and Optical-flow testing result merges S8: the behavioral chain of RGB branch and Optical-flow branchTesting result is merged, and final testing result is obtained.The method of fusion is: setting TrgbFor a behavioral chain of RGB branch,ToptFor a behavioral chain of Optical-flow branch, if TrgbWith ToptBetween IoU be greater than a specified threshold (such as 0.7), thenTake max (s (Trgb), s (Topt)) corresponding to that behavioral chain, delete it is other one;Otherwise, retain this two behavioral chains.
Take mAP (mean Average Precision) be evaluation criterion when, method proposed by the present invention is in J-HMDB-21Behavioral value best at present is achieved on data set as a result, as shown in table 1 compared with current other methods:
mAP0.50.5:0.95
Gurkirt et al.[1]72.041.6
ACT[2]72.243.2
Peng and Schmid[3]73.1-
Harkirat et al.[4]67.336.1
The present invention75.944.8
Table 1. is compared with other methods, and '-' indicates not refer to, the higher the better for result value
The method compared in table 1 is listed below:
[1]S.G.,S.S.,and C.F.,“Online real time multiple spatiotemporalaction localisation and prediction on a single platform,”ArXiv,2016.
[2]V.Kalogeiton,P.Weinzaepfel,V.Ferrari,and C.Schmid,“Action tubeletdetector for spatio-temporal action localization,”in IEEE InternationalConference on Computer Vision,2017,pp.4415–4423.
[3]P.X.and S.C.,“Multi-region two-stream r-cnn for action detection,”European Conference on Computer Vision,pp.744–759,2016.
[4]B.H.,S.M.,S.G.,S.S.,C.F.,and T.P.H.,“Incremental tube constructionfor human action detection,”ArXiv,2017.
It should be noted that the purpose for publicizing and implementing example is to help to further understand the present invention, but the skill of this fieldArt personnel, which are understood that, not to be departed from the present invention and spirit and scope of the appended claims, and various substitutions and modifications are allIt is possible.Therefore, the present invention should not be limited to embodiment disclosure of that, and the scope of protection of present invention is with claimSubject to the range that book defines.

Claims (8)

CN201811298487.0A2018-11-022018-11-02Online video behavior detection method based on space-time context analysisActiveCN109409307B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN201811298487.0ACN109409307B (en)2018-11-022018-11-02Online video behavior detection method based on space-time context analysis

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN201811298487.0ACN109409307B (en)2018-11-022018-11-02Online video behavior detection method based on space-time context analysis

Publications (2)

Publication NumberPublication Date
CN109409307Atrue CN109409307A (en)2019-03-01
CN109409307B CN109409307B (en)2022-04-01

Family

ID=65471476

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN201811298487.0AActiveCN109409307B (en)2018-11-022018-11-02Online video behavior detection method based on space-time context analysis

Country Status (1)

CountryLink
CN (1)CN109409307B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN110059761A (en)*2019-04-252019-07-26成都睿沿科技有限公司A kind of human body behavior prediction method and device
CN110348345A (en)*2019-06-282019-10-18西安交通大学A kind of Weakly supervised timing operating position fixing method based on continuity of movement
CN110472230A (en)*2019-07-112019-11-19平安科技(深圳)有限公司The recognition methods of Chinese text and device
CN111414876A (en)*2020-03-262020-07-14西安交通大学Violent behavior identification method based on time sequence guide space attention
CN111523421A (en)*2020-04-142020-08-11上海交通大学 Multi-person behavior detection method and system based on deep learning fusion of various interactive information
CN113569824A (en)*2021-09-262021-10-29腾讯科技(深圳)有限公司Model processing method, related device, storage medium and computer program product
CN113837091A (en)*2021-09-242021-12-24上海商汤智能科技有限公司Identification method, identification device, electronic equipment and computer-readable storage medium
CN115424207A (en)*2022-09-052022-12-02南京星云软件科技有限公司Self-adaptive monitoring system and method
CN120071606A (en)*2025-01-032025-05-30华南理工大学Traffic flow prediction method, system and medium based on space-time large language model

Citations (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN1466104A (en)*2002-07-032004-01-07中国科学院计算技术研究所 Speech-driven face animation method based on the combination of statistics and rules
US20080310734A1 (en)*2007-06-182008-12-18The Regents Of The University Of CaliforniaHigh speed video action recognition and localization
CN107330920A (en)*2017-06-282017-11-07华中科技大学A kind of monitor video multi-target tracking method based on deep learning
CN108573246A (en)*2018-05-082018-09-25北京工业大学 A temporal sequence action recognition method based on deep learning

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN1466104A (en)*2002-07-032004-01-07中国科学院计算技术研究所 Speech-driven face animation method based on the combination of statistics and rules
US20080310734A1 (en)*2007-06-182008-12-18The Regents Of The University Of CaliforniaHigh speed video action recognition and localization
CN107330920A (en)*2017-06-282017-11-07华中科技大学A kind of monitor video multi-target tracking method based on deep learning
CN108573246A (en)*2018-05-082018-09-25北京工业大学 A temporal sequence action recognition method based on deep learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
BO HUANG ET AL: "《Convolutional Gated Recurrent Units Fusion for Video Action Recognition》", 《ICONIP 2017: NEURAL INFORMATION PROCESSING》*
郭会文: "《基于多载体的人群异常事件检测方法研究》", 《中国博士学位论文全文数据库 信息科技辑》*

Cited By (12)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN110059761A (en)*2019-04-252019-07-26成都睿沿科技有限公司A kind of human body behavior prediction method and device
CN110348345A (en)*2019-06-282019-10-18西安交通大学A kind of Weakly supervised timing operating position fixing method based on continuity of movement
CN110348345B (en)*2019-06-282021-08-13西安交通大学 A Weakly Supervised Temporal Action Localization Method Based on Action Coherence
CN110472230A (en)*2019-07-112019-11-19平安科技(深圳)有限公司The recognition methods of Chinese text and device
CN110472230B (en)*2019-07-112023-09-05平安科技(深圳)有限公司Chinese text recognition method and device
CN111414876A (en)*2020-03-262020-07-14西安交通大学Violent behavior identification method based on time sequence guide space attention
CN111523421A (en)*2020-04-142020-08-11上海交通大学 Multi-person behavior detection method and system based on deep learning fusion of various interactive information
CN111523421B (en)*2020-04-142023-05-19上海交通大学Multi-person behavior detection method and system based on deep learning fusion of various interaction information
CN113837091A (en)*2021-09-242021-12-24上海商汤智能科技有限公司Identification method, identification device, electronic equipment and computer-readable storage medium
CN113569824A (en)*2021-09-262021-10-29腾讯科技(深圳)有限公司Model processing method, related device, storage medium and computer program product
CN115424207A (en)*2022-09-052022-12-02南京星云软件科技有限公司Self-adaptive monitoring system and method
CN120071606A (en)*2025-01-032025-05-30华南理工大学Traffic flow prediction method, system and medium based on space-time large language model

Also Published As

Publication numberPublication date
CN109409307B (en)2022-04-01

Similar Documents

PublicationPublication DateTitle
CN109409307A (en)A kind of Online Video behavioral value system and method based on space-time contextual analysis
CN109801256B (en)Image aesthetic quality assessment method based on region of interest and global features
CN111462282B (en)Scene graph generation method
CN108596958B (en) A Target Tracking Method Based on Difficult Positive Sample Generation
CN106650806B (en) A Collaborative Deep Network Model Method for Pedestrian Detection
CN114898293B (en) A multimodal trajectory prediction method for pedestrian groups crossing the street for autonomous vehicles
CN107066973B (en)Video content description method using space-time attention model
CN110443173A (en)A kind of instance of video dividing method and system based on inter-frame relation
CN111709410B (en)Behavior identification method for strong dynamic video
CN110164476A (en)A kind of speech-emotion recognition method of the BLSTM based on multi output Fusion Features
CN116342953A (en) Bimodal object detection model and method based on residual shrinkage attention network
CN113239801B (en)Cross-domain action recognition method based on multi-scale feature learning and multi-level domain alignment
CN107529650A (en)Network model construction and closed loop detection method, corresponding device and computer equipment
CN114693942A (en)Multimode fault understanding and auxiliary labeling method for intelligent operation and maintenance of instruments and meters
CN110110648A (en)Method is nominated in view-based access control model perception and the movement of artificial intelligence
CN109753853A (en)One kind being completed at the same time pedestrian detection and pedestrian knows method for distinguishing again
CN110032952B (en)Road boundary point detection method based on deep learning
CN114626598B (en) A multimodal trajectory prediction method based on semantic environment modeling
CN108629370A (en)A kind of classification and identification algorithm and device based on depth confidence network
CN110110686A (en)Based on the human motion recognition methods for losing double-current convolutional neural networks more
CN109376677A (en)A kind of video behavior detection method merged based on shape-movement double fluid information
CN114219763A (en) Infrared image detection method for abnormal hot spots of power distribution equipment based on Faster RCNN algorithm
CN115546252A (en) A target tracking method and system based on cross-correlation matching enhanced Siamese network
CN116883650A (en)Image-level weak supervision semantic segmentation method based on attention and local stitching
CN116977367A (en)Campus multi-target tracking method based on transform and Kalman filtering

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp