Movatterモバイル変換


[0]ホーム

URL:


CN113129306A - Occlusion object segmentation solving method based on deep learning - Google Patents

Occlusion object segmentation solving method based on deep learning
Download PDF

Info

Publication number
CN113129306A
CN113129306ACN202110504652.9ACN202110504652ACN113129306ACN 113129306 ACN113129306 ACN 113129306ACN 202110504652 ACN202110504652 ACN 202110504652ACN 113129306 ACN113129306 ACN 113129306A
Authority
CN
China
Prior art keywords
image
overall appearance
parameters
area
deep learning
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110504652.9A
Other languages
Chinese (zh)
Other versions
CN113129306B (en
Inventor
邹倩颖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu College of University of Electronic Science and Technology of China
Original Assignee
Chengdu College of University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu College of University of Electronic Science and Technology of ChinafiledCriticalChengdu College of University of Electronic Science and Technology of China
Priority to CN202110504652.9ApriorityCriticalpatent/CN113129306B/en
Publication of CN113129306ApublicationCriticalpatent/CN113129306A/en
Application grantedgrantedCritical
Publication of CN113129306BpublicationCriticalpatent/CN113129306B/en
Expired - Fee Relatedlegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Images

Classifications

Landscapes

Abstract

The invention relates to the technical field of image analysis, and aims to provide a method for segmenting and solving a sheltered object based on deep learning, which comprises the following steps of 1: when the object is shielded, acquiring an image of the object which is not shielded, performing initialization processing on the image to extract an area parameter and a frame parameter, and executing the step 2; step 2: the area parameters and the frame parameters are used as input and sent to an image deep learning model, the model outputs corresponding frame overall appearance parameters and area overall appearance parameters, the parameters are continuously screened according to the requirements, and the step 3 is executed; and step 3: the method comprises the steps of screening standard frame overall appearance parameters and standard area overall appearance parameters from a plurality of frame overall appearance parameters and area overall appearance parameters, and sending the standard frame overall appearance parameters and the standard area overall appearance parameters to an image construction model to obtain the object overall appearance.

Description

Occlusion object segmentation solving method based on deep learning
Technical Field
The invention relates to the technical field of image processing, in particular to a method for segmenting and solving a sheltered object based on deep learning.
Background
Given a scene, human beings can perform scene understanding well, for example, human beings can not only recognize objects in the scene well, but also perceive relationships between the objects, including occlusion relationships. Occlusion often occurs in two-dimensional scenes, and the occlusion relationship reflects the depth order relationship between objects, i.e., near objects occlude distant objects. The human being can easily judge the relationship between the shielding objects and can identify the shielded objects at the same time because the human eyes acquire a large amount of prior knowledge along with the accumulation of experience of long-term observation of surrounding images.
Scene understanding is an extremely important fundamental task in the field of computer vision, with the aim of making computers understand scenes like humans. The current studies of scholars on scene understanding are mainly divided into two types: one is based on a neural network model and one is based on a probabilistic graphical model. With the wide application of deep learning in recent years, particularly after the Convolutional Neural Network (CNN) is applied in the image field with great success, various subtasks for scene understanding, such as scene recognition, target detection, scene segmentation and the like, have achieved breakthrough progress. However, the scene understanding based on the neural network model has less research on the occlusion objects, only the objects themselves are concerned (for example, the recognition is only to classify the objects in the picture, and the segmentation is only to classify the pixels), and the relationship between the objects is not considered, so that the occlusion relationship cannot be judged. Secondly, the CNN network generally needs a large amount of data supervision information to support, and needs to see samples shielded at various angles in order to identify the shielded object; in addition, the cognitive process of the neural network is a process based on forward transmission and backward transmission of the Convolutional Neural Network (CNN), and there is no feedback mechanism similar to the human brain, and the essential differences between the two are: the feedforward network is a bottom-up process, while reasoning and feedback based on knowledge and experience is a top-down process. The probabilistic graphical model has certain advantages in the aspects of logical reasoning and context information relationship, some researches can carry out depth order reasoning by using some probabilistic graphical models such as Bayesian reasoning and Markov, but the probabilistic graphical model is only a mathematical model, so that the accuracy is lower compared with a neural network model, different models are required to be established according to different scenes, the universality is poor, and the probabilistic graphical model cannot be used for modeling according to more complex scenes.
Disclosure of Invention
The invention aims to provide a method for segmenting and solving a sheltered object based on deep learning, which comprises the steps of obtaining the area of a part of the object which is not sheltered, obtaining a standard frame global parameter and a standard area global parameter through an image deep learning model, and sending the standard frame global parameter and the standard area global parameter to an image construction model to obtain the object global;
the technical scheme adopted by the invention is as follows: a method for segmenting and solving an occluded object based on deep learning comprises the following steps:
step 1: when the object is shielded, acquiring an image of the object which is not shielded, performing initialization processing on the image to extract an area parameter and a frame parameter, and executing the step 2;
step 2: the area parameters and the frame parameters are used as input and sent to an image deep learning model, the model outputs corresponding frame overall appearance parameters and area overall appearance parameters, the parameters are continuously screened according to the requirements, and the step 3 is executed;
and step 3: and screening out standard frame overall appearance parameters and standard area overall appearance parameters from the plurality of frame overall appearance parameters and area overall appearance parameters, and sending the standard frame overall appearance parameters and the standard area overall appearance parameters to an image construction model to obtain the object overall appearance.
Preferably, in the step 1, the initialization processing includes dividing the obtained image of the object that is not occluded into a plurality of cells, and counting the number and area of the cells, where the cells are rectangles, and the areas of the cells close to the edge of the object and the occlusion boundary gradually decrease until the image is fully covered.
Preferably, the acquired image is a planar image of the object, and a single object acquires planar images in at most three directions.
Preferably, in the step 2, the building of the image depth learning model includes the following steps:
step 41: establishing a neural network structure comprising a full connection layer, a convolution layer, an activation function and a pooling layer, establishing forward propagation of image parameters through the full connection layer, and executing step 42;
step 42: extracting image features through the convolutional layer, outputting the feature matrix channel number which is the same as the number of the convolutional kernels through training the convolutional kernels and the deviation features, and executing the step 43;
step 43: in the activation function, the calculation process is linear, a nonlinear factor is introduced, and a matrix size calculation formula after convolution is obtained, wherein W is (W-F +2P)/S +1
Step 44: and only changing W and H of the characteristic matrix without changing the number of depth channels, and obtaining a learning model formula by setting the probability sum of all the processed nodes to be 1.
Preferably, in step 42, a channel of each convolution kernel is the same as a channel of the input feature layer, each channel of the convolution kernel is convolved with the input layer of the corresponding channel, and then the channels are summed to obtain a feature matrix which is used as an output of the deep learning image and is used as a channel of the next layer of input features to be propagated forward.
Preferably, the standard frame complete picture and the standard area complete picture are average values output by the image depth learning model through multiple times of training.
Preferably, the error calculation of the image depth learning model adopts a cross Encopy Loss of Loss cross entropy algorithm.
Preferably, in step 43, Dropout random neuron inactivation operation is used in the first two layers of the fully-connected layer through the relu activation function to reduce fitting.
Compared with the prior art, the invention has the beneficial effects that:
1. the full-view of the object can be quickly obtained through the only area of the object image, and the method is quick and high in accuracy.
Drawings
Fig. 1 is a schematic diagram of a method for segmentation solution of an occluded object based on deep learning.
Detailed Description
The technical solutions in the embodiments of the present invention are clearly and completely described below with reference to fig. 1 of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments. All other implementations made by those of ordinary skill in the art based on the embodiments of the present invention are obtained without inventive efforts.
Example 1:
a method for segmenting and solving an occluded object based on deep learning comprises the following steps:
step 1: when the object is shielded, acquiring an image of the object which is not shielded, performing initialization processing on the image to extract an area parameter and a frame parameter, and executing the step 2;
step 2: the area parameters and the frame parameters are used as input and sent to an image deep learning model, the model outputs corresponding frame overall appearance parameters and area overall appearance parameters, the parameters are continuously screened according to the requirements, and the step 3 is executed;
and step 3: and screening out standard frame overall appearance parameters and standard area overall appearance parameters from the plurality of frame overall appearance parameters and area overall appearance parameters, and sending the standard frame overall appearance parameters and the standard area overall appearance parameters to an image construction model to obtain the object overall appearance.
It should be noted that, in step 1, the initialization processing includes dividing the obtained image of the object that is not occluded into a plurality of cells, and counting the number and area of the cells, where the cells are rectangles, and the area of the cells near the edge of the object and the occlusion boundary gradually decreases until the image is fully covered.
It should be noted that the acquired image is a planar image of an object, and a single object acquires planar images in at most three directions.
It should be noted that, in the step 2, the building of the image depth learning model includes the following steps:
step 41: establishing a neural network structure comprising a full connection layer, a convolution layer, an activation function and a pooling layer, establishing forward propagation of image parameters through the full connection layer, and executing step 42;
step 42: extracting image features through convolutional layers, training convolutional kernels and deviation features, wherein a channel of each convolutional kernel is the same as a channel of an input feature layer (for example, a 5 x 5 RGB image has 3 input layers, and each convolutional kernel also has three channels corresponding to the three input layers respectively), each channel of the convolutional kernel is convolved with the input layer of the corresponding channel respectively, then summing is carried out to obtain a feature matrix which is used as the output of a deep learning image, the number of characteristic matrix channels which are used as the forward propagation output of one channel of the next layer of input features is the same as the number of convolutional kernels, and executing step 43;
step 43: in the activation function, the calculation process is linear, a nonlinear factor is introduced, and the disappearance of the gradient is easy to occur when the sigmoid-layer number is deep; relu-when a very large gradient passes through the back propagation process, the updated weight distribution center may be smaller than zero, the derivative is always 0, the back propagation cannot update the weight, and the neuron is in an inactive state. Obtaining the calculation formula of the matrix size after convolution, wherein W is (W-F +2P)/S +1
Step 44: only changing W and H of the feature matrix, not changing the number of depth channels, enabling the probability sum of all nodes after processing to be 1, and carrying out Max boosting downsampling to aim at carrying out sparse processing on the feature map and reducing the operation amount of data. AveragePooling downsampling layer, characteristics: # No training parameters # only changes W and H of the feature matrix, does not change the depth (number of channels) # the same 5SoftMax layers as the general poolsize and stride.
It should be noted that, in step 42, the channel of each convolution kernel is the same as the channel of the input feature layer, and each channel of the convolution kernel is convolved with the input layer of the corresponding channel, and then summed to obtain a feature matrix as the output of the deep learning image, which is used as a channel of the next layer of input features to propagate forward.
It should be noted that the standard frame full-face and the standard area full-face are average values output by the image depth learning model through multiple training.
It is worth to be noted that the error calculation of the image depth learning model adopts a cross entropy Loss algorithm of cross entropy Loss, wherein the cross entropy Loss algorithm of cross entropy Loss is as follows: in the process of solving the maximum value of the likelihood function of the sample, the minimum value of a function is deduced, and the function is made to be a cross entropy function. (the goal of maximum likelihood estimation is to extrapolate back to the parameter values most likely to lead to such a result using known sample results.) the logistic regression-two classification reflects the degree of deviation of the two distributions (the distributions of groudtruth and output), the optimization goal being to approximate the two distributions.
For the multi-class problem (the output belongs to only one class), crossentry _ Loss is the function 1.
For the two-class problem (the output may fall into multiple classes), crossentry _ Loss is function 2.
It is worth noting that in said step 43, using Dropout random neuron inactivation operation in the first two layers of the fully-connected layer by relu activation function, fitting is reduced, and the LRN layer: a competition mechanism is created for the activity of local neurons, so that the response value becomes relatively larger, and other neurons with smaller feedback are inhibited, and the generalization capability of the model is enhanced. LRN (local Response normalization) local Response normalization.
In summary, by acquiring the area of the part of the object which is not shielded, the standard frame global parameters and the standard area global parameters are obtained through the image depth learning model, and the standard frame global parameters and the standard area global parameters are sent to the image construction model to obtain the object global.

Claims (8)

CN202110504652.9A2021-05-102021-05-10Occlusion object segmentation solving method based on deep learningExpired - Fee RelatedCN113129306B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202110504652.9ACN113129306B (en)2021-05-102021-05-10Occlusion object segmentation solving method based on deep learning

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202110504652.9ACN113129306B (en)2021-05-102021-05-10Occlusion object segmentation solving method based on deep learning

Publications (2)

Publication NumberPublication Date
CN113129306Atrue CN113129306A (en)2021-07-16
CN113129306B CN113129306B (en)2022-12-02

Family

ID=76781274

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202110504652.9AExpired - Fee RelatedCN113129306B (en)2021-05-102021-05-10Occlusion object segmentation solving method based on deep learning

Country Status (1)

CountryLink
CN (1)CN113129306B (en)

Citations (25)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20060122480A1 (en)*2004-11-222006-06-08Jiebo LuoSegmenting occluded anatomical structures in medical images
CN101853395A (en)*2010-05-272010-10-06南昌航空大学 A Method for Recognizing Partially Covered 3D Objects from Single Graphics and Images
US20100284609A1 (en)*2008-02-052010-11-11CENTRE DE RECHERCHE INDUSTRIELLE DU QUéBECApparatus and method for measuring size distribution of granular matter
CN103839279A (en)*2014-03-182014-06-04湖州师范学院Adhesion object segmentation method based on VIBE in object detection
CN106485215A (en)*2016-09-292017-03-08西交利物浦大学Face occlusion detection method based on depth convolutional neural networks
CN107403200A (en)*2017-08-102017-11-28北京亚鸿世纪科技发展有限公司Improve the multiple imperfect picture sorting technique of image segmentation algorithm combination deep learning
CN107563303A (en)*2017-08-092018-01-09中国科学院大学A kind of robustness Ship Target Detection method based on deep learning
CN107622503A (en)*2017-08-102018-01-23上海电力学院 A Hierarchical Segmentation Method for Recovering Image Occlusion Boundaries
US10067509B1 (en)*2017-03-102018-09-04TuSimpleSystem and method for occluding contour detection
CN108629782A (en)*2018-04-282018-10-09合肥工业大学The road target depth estimation method propagated based on ground clue
CN108764186A (en)*2018-06-012018-11-06合肥工业大学Personage based on rotation deep learning blocks profile testing method
CN109190458A (en)*2018-07-202019-01-11华南理工大学A kind of person of low position's head inspecting method based on deep learning
US20190050667A1 (en)*2017-03-102019-02-14TuSimpleSystem and method for occluding contour detection
CN109360206A (en)*2018-09-082019-02-19华中农业大学 Segmentation method of field rice ears based on deep learning
CN109558810A (en)*2018-11-122019-04-02北京工业大学Divided based on position and merges target person recognition methods
CN109858455A (en)*2019-02-182019-06-07南京航空航天大学A kind of piecemeal detection scale adaptive tracking method for circular target
CN110414430A (en)*2019-07-292019-11-05郑州信大先进技术研究院A kind of pedestrian recognition methods and device again based on the fusion of more ratios
CN110415181A (en)*2019-06-122019-11-05勤耕仁现代农业科技发展(淮安)有限责任公司Flue-cured tobacco RGB image intelligent recognition and grade determination method under a kind of open environment
CN111310624A (en)*2020-02-052020-06-19腾讯科技(深圳)有限公司Occlusion recognition method and device, computer equipment and storage medium
CN111325764A (en)*2020-02-112020-06-23广西师范大学Fruit image contour recognition method
CN111428556A (en)*2020-02-172020-07-17浙江树人学院(浙江树人大学)Traffic sign recognition method based on capsule neural network
CN111489373A (en)*2020-04-072020-08-04北京工业大学Occlusion object segmentation method based on deep learning
JP2020188432A (en)*2019-05-172020-11-19株式会社国際電気通信基礎技術研究所 Shooting direction setting device, shooting direction setting program, shooting direction setting method, intrusion detection device, intrusion detection program, intrusion detection method
CN112418043A (en)*2020-11-162021-02-26安徽农业大学 Method, device, robot, equipment and storage medium for determining corn weed shelter
CN112634313A (en)*2021-01-082021-04-09云从科技集团股份有限公司Target occlusion assessment method, system, medium and device

Patent Citations (25)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20060122480A1 (en)*2004-11-222006-06-08Jiebo LuoSegmenting occluded anatomical structures in medical images
US20100284609A1 (en)*2008-02-052010-11-11CENTRE DE RECHERCHE INDUSTRIELLE DU QUéBECApparatus and method for measuring size distribution of granular matter
CN101853395A (en)*2010-05-272010-10-06南昌航空大学 A Method for Recognizing Partially Covered 3D Objects from Single Graphics and Images
CN103839279A (en)*2014-03-182014-06-04湖州师范学院Adhesion object segmentation method based on VIBE in object detection
CN106485215A (en)*2016-09-292017-03-08西交利物浦大学Face occlusion detection method based on depth convolutional neural networks
US10067509B1 (en)*2017-03-102018-09-04TuSimpleSystem and method for occluding contour detection
US20190050667A1 (en)*2017-03-102019-02-14TuSimpleSystem and method for occluding contour detection
CN107563303A (en)*2017-08-092018-01-09中国科学院大学A kind of robustness Ship Target Detection method based on deep learning
CN107403200A (en)*2017-08-102017-11-28北京亚鸿世纪科技发展有限公司Improve the multiple imperfect picture sorting technique of image segmentation algorithm combination deep learning
CN107622503A (en)*2017-08-102018-01-23上海电力学院 A Hierarchical Segmentation Method for Recovering Image Occlusion Boundaries
CN108629782A (en)*2018-04-282018-10-09合肥工业大学The road target depth estimation method propagated based on ground clue
CN108764186A (en)*2018-06-012018-11-06合肥工业大学Personage based on rotation deep learning blocks profile testing method
CN109190458A (en)*2018-07-202019-01-11华南理工大学A kind of person of low position's head inspecting method based on deep learning
CN109360206A (en)*2018-09-082019-02-19华中农业大学 Segmentation method of field rice ears based on deep learning
CN109558810A (en)*2018-11-122019-04-02北京工业大学Divided based on position and merges target person recognition methods
CN109858455A (en)*2019-02-182019-06-07南京航空航天大学A kind of piecemeal detection scale adaptive tracking method for circular target
JP2020188432A (en)*2019-05-172020-11-19株式会社国際電気通信基礎技術研究所 Shooting direction setting device, shooting direction setting program, shooting direction setting method, intrusion detection device, intrusion detection program, intrusion detection method
CN110415181A (en)*2019-06-122019-11-05勤耕仁现代农业科技发展(淮安)有限责任公司Flue-cured tobacco RGB image intelligent recognition and grade determination method under a kind of open environment
CN110414430A (en)*2019-07-292019-11-05郑州信大先进技术研究院A kind of pedestrian recognition methods and device again based on the fusion of more ratios
CN111310624A (en)*2020-02-052020-06-19腾讯科技(深圳)有限公司Occlusion recognition method and device, computer equipment and storage medium
CN111325764A (en)*2020-02-112020-06-23广西师范大学Fruit image contour recognition method
CN111428556A (en)*2020-02-172020-07-17浙江树人学院(浙江树人大学)Traffic sign recognition method based on capsule neural network
CN111489373A (en)*2020-04-072020-08-04北京工业大学Occlusion object segmentation method based on deep learning
CN112418043A (en)*2020-11-162021-02-26安徽农业大学 Method, device, robot, equipment and storage medium for determining corn weed shelter
CN112634313A (en)*2021-01-082021-04-09云从科技集团股份有限公司Target occlusion assessment method, system, medium and device

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
JIAHONG WU 等: "AINNOSEG:PANORAMIC SEGMENTATION WITH HIGH PERFOMANCE", 《HTTPS://ARXIV.ORG/ABS/2007.10591》*
唐旭晟等: "基于局部边缘特征的快速目标检测", 《计算机辅助设计与图形学学报》*
张栩华等: "基于车辆轮廓凹陷区域的分割算法", 《电子设计工程》*
王立春等: "基于无人机航拍图像的公路标线检测算法", 《计算机技术与发展》*

Also Published As

Publication numberPublication date
CN113129306B (en)2022-12-02

Similar Documents

PublicationPublication DateTitle
CN110210551B (en) A Visual Object Tracking Method Based on Adaptive Subject Sensitivity
CN112069896B (en) A video target tracking method based on Siamese network fusion of multi-template features
US10936913B2 (en)Automatic filter pruning technique for convolutional neural networks
CN110070107B (en)Object recognition method and device
US20190228268A1 (en)Method and system for cell image segmentation using multi-stage convolutional neural networks
CN109949255B (en)Image reconstruction method and device
CN111476219A (en) Image object detection method in smart home environment
WO2020244261A1 (en)Scene recognition system for high-resolution remote sensing image, and model generation method
CN114694261B (en) A video 3D human pose estimation method and system based on multi-level supervised graph convolution
CN107862275A (en)Human bodys' response model and its construction method and Human bodys' response method
CN106295694B (en)Face recognition method for iterative re-constrained group sparse representation classification
CN107609460A (en)A kind of Human bodys' response method for merging space-time dual-network stream and attention mechanism
CN113536970B (en) A video classification model training method and related device
US10643092B2 (en)Segmenting irregular shapes in images using deep region growing with an image pyramid
CN111898432A (en) A pedestrian detection system and method based on improved YOLOv3 algorithm
CN112861718A (en)Lightweight feature fusion crowd counting method and system
Irfan et al.Comparison Of Sgd, Rmsprop, And Adam Optimation In Animal Classification Using Cnns
Jeziorek et al.Memory-efficient graph convolutional networks for object classification and detection with event cameras
CN116309632A (en) 3D Liver Semantic Segmentation Method Based on Multi-scale Cascade Feature Attention Strategy
CN111489373B (en)Occlusion object segmentation method based on deep learning
US10776923B2 (en)Segmenting irregular shapes in images using deep region growing
Jeong et al.Congestion-aware Bayesian loss for crowd counting
Dorgham et al.Grasshopper Optimization Algorithm and Neural Network Classifier for Detection and Classification of Barley Leaf Diseases
CN113129306B (en)Occlusion object segmentation solving method based on deep learning
CN112189217B (en) Segmenting irregular shapes in images using deep region growing

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant
CF01Termination of patent right due to non-payment of annual fee
CF01Termination of patent right due to non-payment of annual fee

Granted publication date:20221202


[8]ページ先頭

©2009-2025 Movatter.jp