Movatterモバイル変換


[0]ホーム

URL:


CN112837320A - A Semantic Segmentation Method of Remote Sensing Image Based on Parallel Atrous Convolution - Google Patents

A Semantic Segmentation Method of Remote Sensing Image Based on Parallel Atrous Convolution
Download PDF

Info

Publication number
CN112837320A
CN112837320ACN202110129416.3ACN202110129416ACN112837320ACN 112837320 ACN112837320 ACN 112837320ACN 202110129416 ACN202110129416 ACN 202110129416ACN 112837320 ACN112837320 ACN 112837320A
Authority
CN
China
Prior art keywords
remote sensing
convolution
sensing image
network
semantic segmentation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110129416.3A
Other languages
Chinese (zh)
Other versions
CN112837320B (en
Inventor
张东映
唐振超
罗蔚然
洪志明
梁忠壮
刘震
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huazhong University of Science and Technology
Original Assignee
Wuhan Shanlai Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan Shanlai Technology Co LtdfiledCriticalWuhan Shanlai Technology Co Ltd
Priority to CN202110129416.3ApriorityCriticalpatent/CN112837320B/en
Publication of CN112837320ApublicationCriticalpatent/CN112837320A/en
Application grantedgrantedCritical
Publication of CN112837320BpublicationCriticalpatent/CN112837320B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Images

Classifications

Landscapes

Abstract

The invention discloses a remote sensing image semantic segmentation method based on parallel void convolution, which relates to the technical field of remote sensing images and comprises the following steps: the method comprises the steps of obtaining a high-resolution remote sensing image in advance, slicing the high-resolution remote sensing image, normalizing and standardizing the high-resolution remote sensing image, and obtaining a source high-resolution remote sensing image; initializing a low-layer network of a feature extraction network resnet101 based on a pretrained resnet101 parameter on ImageNet, constructing a parallel cavity convolution network, and extracting shallow features of a source high-resolution remote sensing image; inputting the shallow layer characteristics into a parallel cavity convolution network to obtain multi-scale information, and fusing the multi-scale information; and fusing the fused features and the shallow features again, repairing image-level information by using a full-connection conditional random field, and acquiring a semantic segmentation result. The invention enlarges the experience field of convolution under the condition of not increasing extra parameters, and compared with the standard convolution which achieves the same experience field, the parallel cavity convolution method can save video memory.

Description

Remote sensing image semantic segmentation method based on parallel hole convolution
Technical Field
The invention relates to the technical field of remote sensing images, in particular to a remote sensing image semantic segmentation method based on parallel void convolution.
Background
With the maturity of satellite remote sensing technology, the promotion of commercialization and the encouragement and promotion of governments around the world, satellite remote sensing is rapidly developed and applied in more and more fields. The semantic segmentation of the remote sensing image is an important link of the satellite remote sensing application. The semantic segmentation of the remote sensing image is widely applied to the field of pattern recognition such as city planning, road planning, ground object target extraction and even automatic driving. The semantic segmentation precision is improved, and the method has important significance in the problem of processing the remote sensing image.
The ground feature information in the remote sensing image is complex and various, and in order to improve the semantic segmentation precision of the remote sensing image, a great deal of research has been carried out by related scholars, and a plurality of algorithms are proposed. The idea of the algorithms mainly comprises (a) adopting a full convolution network to carry out semantic segmentation on a remote sensing image, and (b) fusing feature information with symmetrical size on the basis of the full convolution network, and recording indexes in an anti-pooling process to make up for loss of position information. The methods are all based on standard convolution, and the receptive field of the standard convolution has limitation, so that the expansion of the receptive field has important research and application values for the semantic segmentation of the remote sensing image.
An effective solution to the problems in the related art has not been proposed yet.
Disclosure of Invention
Aiming at the problems in the related art, the invention provides a remote sensing image semantic segmentation method based on parallel void convolution, so as to overcome the technical problems in the prior related art.
The technical scheme of the invention is realized as follows:
a remote sensing image semantic segmentation method based on parallel void convolution comprises the following steps:
the method comprises the steps of obtaining a high-resolution remote sensing image in advance, slicing the high-resolution remote sensing image, normalizing and standardizing the high-resolution remote sensing image, and obtaining a source high-resolution remote sensing image;
initializing a low-layer network of a feature extraction network resnet101 based on a pretrained resnet101 parameter on ImageNet, constructing a parallel cavity convolution network, and extracting shallow features of a source high-resolution remote sensing image;
inputting shallow layer characteristics into a parallel cavity convolution network to obtain multi-scale information, and fusing the multi-scale information, wherein different expansion rates are set to capture the multi-scale information;
and fusing the fused features and the shallow features again, repairing image-level information by using a full-connection conditional random field, and acquiring a semantic segmentation result.
Further, the slicing the high-resolution remote sensing image comprises slicing the high-resolution remote sensing image to 512 pixels in length and width.
Further, the method also comprises the following steps:
and extracting RGB three channels from the sliced high-resolution remote sensing image.
Further, the parallel hole convolution network includes the following steps:
starting from a standard ordinary convolution, a discrete function F is scaled:
Figure BDA0002924965170000021
and is provided with
Figure BDA0002924965170000022
And k is as follows: omegar→ i is the discrete convolution kernel, developed with p as the centerThe convolution calculation process of (a) is as follows:
Figure BDA0002924965170000023
by extending the standard convolution, let l be the dilation rate of the hole convolution, then the hole convolution is:
Figure BDA0002924965170000024
and (3) parallelly carrying out cavity convolution on the shallow layer features with different expansion rates to obtain multi-scale features, and fusing the multi-scale features by using a splicing mode to form a parallel cavity convolution network layer.
Further, the expansion ratios are set to 2, 3, 4, and 5, respectively.
Further, the fully connected conditional random field inpainting image-level information comprises the following steps:
the energy function used by the fully connected conditional random field is:
Figure BDA0002924965170000025
the elementary potential energy function of the method is used for describing the influence of an observed object and a label:
θi(xi)=-log P(xi)
wherein, the pixel points i, P (x) of the imagei) For the probability of classification of the network on a pixel, a binary potential energy function describes the correlation between observed objects:
Figure BDA0002924965170000026
when x isi≠yjWhen u (x)i,yj) 1, otherwise, u (x)i,yj) 0, another km(fi,fj) As fiAnd fjGaussian kernel of between fiIs the color information corresponding to the pixel i, i.e. the feature vector, wmIs the weight used by the Gaussian kernel;
in the process of minimizing the energy function, pixels which are unreasonably classified in the image can be corrected, and a repaired semantic segmentation result is obtained.
The invention has the beneficial effects that:
the invention relates to a remote sensing image semantic segmentation method based on parallel cavity convolution, which comprises the steps of obtaining a high-resolution remote sensing image in advance, slicing the high-resolution remote sensing image, normalizing and standardizing to obtain a source high-resolution remote sensing image, initializing the characteristic based on the pretrained resnet101 parameter on ImageNet to extract the lower layer network of the network resnet101, constructing a parallel cavity convolution network, extracting shallow features of the source high-resolution remote sensing image, inputting the shallow features into a parallel cavity convolution network to obtain multi-scale information, and the multi-scale information is fused, the fused features and the shallow features are fused again, the full-connection conditional random field is used for restoring the image-level information, the semantic segmentation result is obtained, and the semantic segmentation is realized without adding extra parameters, the field of experience of convolution is enlarged, and compared with standard convolution which achieves the same field of experience, the parallel hole convolution method can save video memory; the parallel computing structure is adopted, so that the nodes in the neural network computing graph are conveniently distributed on distributed hardware, and the computing speed is improved; the multi-scale information is beneficial to capturing detailed objects and large objects by a network, small target objects are not easy to miss, and the semantic segmentation precision is improved; in addition, the void convolution can widely perceive the adjacent objects of the target object, the pixel-level classification can be more effectively carried out by means of the adjacent information, and the pixel-level classification effect is better than that of the standard convolution.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings needed in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings without creative efforts.
FIG. 1 is a schematic diagram of hole convolution sampling with different expansion rates of a remote sensing image semantic segmentation method based on parallel hole convolution according to an embodiment of the invention;
FIG. 2 is a schematic diagram of a multi-scale parallel void convolution network of a remote sensing image semantic segmentation method based on parallel void convolution according to an embodiment of the invention;
FIG. 3 shows a parallel void convolution semantic segmentation result of the remote sensing image semantic segmentation method based on parallel void convolution according to the embodiment of the invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments that can be derived by one of ordinary skill in the art from the embodiments given herein are intended to be within the scope of the present invention.
According to the embodiment of the invention, a remote sensing image semantic segmentation method based on parallel void convolution is provided.
As shown in fig. 1 to fig. 3, the method for segmenting the remote sensing image semantic based on the parallel void convolution according to the embodiment of the present invention includes the following steps:
the method comprises the steps of obtaining a high-resolution remote sensing image in advance, slicing the high-resolution remote sensing image, normalizing and standardizing the high-resolution remote sensing image, and obtaining a source high-resolution remote sensing image;
initializing a low-layer network of a feature extraction network resnet101 based on a pretrained resnet101 parameter on ImageNet, constructing a parallel cavity convolution network, and extracting shallow features of a source high-resolution remote sensing image;
inputting shallow layer characteristics into a parallel cavity convolution network to obtain multi-scale information, and fusing the multi-scale information, wherein different expansion rates are set to capture the multi-scale information;
and fusing the fused features and the shallow features again, repairing image-level information by using a full-connection conditional random field, and acquiring a semantic segmentation result.
And slicing the high-resolution remote sensing image, wherein slicing the high-resolution remote sensing image to 512 pixels in length and width.
Wherein, still include the following step:
and extracting RGB three channels from the sliced high-resolution remote sensing image.
The parallel hole convolution network comprises the following steps:
starting from a standard ordinary convolution, a discrete function F is scaled:
Figure BDA0002924965170000041
and is provided with
Figure BDA0002924965170000042
And k is as follows: omegar→ i is the discrete convolution kernel, the convolution computation process with p as the center is:
Figure BDA0002924965170000051
by extending the standard convolution, let l be the dilation rate of the hole convolution, then the hole convolution is:
Figure BDA0002924965170000052
and (3) parallelly carrying out cavity convolution on the shallow layer features with different expansion rates to obtain multi-scale features, and fusing the multi-scale features by using a splicing mode to form a parallel cavity convolution network layer.
Wherein the expansion ratios are set to 2, 3, 4 and 5, respectively.
Wherein the fully connected conditional random field inpainting image-level information comprises the steps of:
the energy function used by the fully connected conditional random field is:
Figure BDA0002924965170000053
the elementary potential energy function of the method is used for describing the influence of an observed object and a label:
θi(xi)=-log P(xi)
wherein, the pixel points i, P (x) of the imagei) For the probability of classification of the network on a pixel, a binary potential energy function describes the correlation between observed objects:
Figure BDA0002924965170000054
when x isi≠yjWhen u (x)i,yj) 1, otherwise, u (x)i,yj) 0, another km(fi,fj) As fiAnd fjGaussian kernel of between fiIs the color information corresponding to the pixel i, i.e. the feature vector, wmIs the weight used by the gaussian kernel.
By means of the technical scheme, the high-resolution remote sensing image is sliced to obtain the pre-trained resnet101, the low-level migration is used as a feature extraction network, and the shallow features of the sliced image are extracted; constructing parallel cavity convolution, setting the expansion rate of a convolution kernel from 2 to 5, inputting shallow layer characteristics into a parallel cavity convolution network, and splicing information under different scales; the output features of the cavity convolution network and the shallow features are fused again, resolution is restored through up-sampling, and a segmentation result is restored through a conditional random field; and merging the segmentation results of the slices, and repairing unreasonable prediction results by simply filling holes and removing small connected domains.
Specifically, in one embodiment, the method comprises the following steps:
s1: firstly, preprocessing a high-resolution remote sensing image, wherein the resolution of the high-resolution remote sensing image is too high, the memory and the video memory of a common computer are not easy to bear the calculation of the whole image, and the image is sliced to 512 pixels in length and width according to the commonly used 512 resolution of a mainstream semantic segmentation model;
s2: in order to be compatible with the conventional deep convolutional neural network, RGB three channels need to be extracted from the sliced remote sensing image. Conventional data enhancement was performed: random horizontal turning, random vertical turning and color dithering. When data enhancement is carried out, the same processing is carried out on the annotation image along with the RGB image;
s3: and carrying out proper scaling or normalization on the RGB three-channel tensor. Suppose a data set has a total of m RGB images that can be divided into 3 channels of tensors [ x ]1,x2,x3]Normalizing the tensor to obtain [ y [)1,y2,y3]The tensor normalization formula for each channel is:
Figure BDA0002924965170000061
s4: then, normalizing according to the mean value mu and the standard deviation sigma of each channel to obtain tensor [ z [ [ z ]1,z2,z3]The normalized calculation formula is:
Figure BDA0002924965170000062
s5: initializing a network based on resnet101, intercepting layer1 to layer4, wherein the cavity convolution expansion rate of layer4 is 2, and the cavity convolution expansion rates of layer1 to layer3 are all 1, which is equivalent to standard common convolution;
s6: performing cavity space pyramid convolution on the features output by the resnet101, performing parallel convolution by using different expansion rates, wherein the space pyramid convolution does not perform global pooling, but replaces global pooling branches with standard convolution instead, and the purpose is to obtain semantic information more deeply and improve classification accuracy;
s7: the low-level features generated by layer1 in resnet101 are fused with the space pyramid convolution result after linear interpolation by using a skip level structure, the low-level features can bring partial position information for the high-level features, and because global pooling is cancelled in the space pyramid convolution layer, the network lacks the position information of image level features, the rough segmentation result output by the network needs to be post-processed based on a conditional random field;
s8: calculating loss by using cross entropy, adding weight to each class of objects during cross entropy calculation due to unbalanced object distribution of the remote sensing image, reversely propagating and calculating gradient in a calculation graph by using the loss function, and updating network parameters;
s9: adadelta is adopted as an optimization method of model training, and the initial learning rate is set to be 1e-1
Adadelta can achieve a faster effect in the early and middle training period. The main network for extracting the features of the model is resnet101, although the specific ground object information of the remote sensing image cannot be directly detected by the resnet101 pre-trained on ImageNet, the low-level information such as edges, angles, colors and the like can be effectively sensed, so that the feature extraction layer of the network can be initialized by using the parameters of the resnet101 pre-trained on ImageNet, and the network can obtain a good initial solution; performing random initialization of other layer parameters of the network, wherein the random initialization obeys Gaussian distribution;
s10: the model can be converged after traversing the whole data set for 256 times, the batch processing size is set to be 8, and the co-iteration number of model training is 5e4
S11: the high-resolution remote sensing image cannot be segmented at one time, so that the high-resolution remote sensing image is required to be sliced first and then semantically segmented one by one, and unreasonable prediction results are restored by simply filling holes and removing small connected domains when all slices are spliced.
Further, as shown in fig. 1, (a)/(b)/(c) in fig. 1 indicates characteristic sampling of the hole convolutions having different expansion rates, and the expansion rates are 1, 2, and 3, respectively, and it can be seen from fig. 1 that the receptive field increases as the expansion rate of the hole convolution increases. The hole convolution can be sparse sampled on the features by setting the expansion rate, and can be performed by using any expansion rate, so that the method is favorable for clearly controlling the receptive field and acquiring the context information in the intensive calculation task. The setting of the hole convolution expansion rate does not affect the structure of the original network parameters, which is friendly to the migration learning, so that fine adjustment can be still performed based on the parameters of the original network after the expansion rate is set.
In addition, according to steps S1 to S4, the GID high resolution remote sensing image is sliced to have a resolution of 512, normalized and normalized, and the statistical data set is used to obtain 3 channel mean values of RGB required for normalization: 0.3515224,0.38427463,0.35403764. The standard deviation is: 0.19264674,0.18325084,0.17028946.
In addition, as shown in fig. 2, according to steps S5 and S6, the feature is initialized by using the prestrained resnet101 parameter on ImageNet to extract the lower network of the network resnet101, the lower network can effectively detect the position information of edges, corners, and the like, and construct a parallel cavity convolution network, and the expansion rates are set to 2, 3, 4, and 5, respectively. The template parameters of the convolution kernel are all tensors consisting of (3, 3) size. The shallow layer features are input into a parallel cavity convolution network to obtain multi-scale information, the multi-scale information is fused in a splicing mode, and the calculation process of the feature input cavity convolution is shown in the attached figure 2.
In addition, according to step S7, the fused features and the shallow features are fused again to compensate for the position detail information, and the resolution is restored by upsampling, so that the obtained semantic segmentation result is rough, and the full-connection conditional random field is used to repair the image-level information, that is, the image-level features are repaired to improve the semantic segmentation result.
In addition, according to steps S8 to S10, the data set is traversed, forward calculation is performed, the loss function is updated after each batch processing is finished, the loss function is firstly counted according to the pixel classes to obtain the proportion of each class, and the proportion is fused to the cross entropy loss to serve as the coefficient of each class loss. And starting from the node of the loss function, reversely calculating in the calculation graph, acquiring the gradient and updating the model parameters. The optimizer used for model updating is Adadelta, and Adadelta can achieve a quicker effect in the early and middle training period.
In addition, as shown in fig. 3, after the training is finished, parameters of the semantic segmentation network are obtained, and the parameters are loaded into the network with the corresponding structure during reasoning. Semantic segmentation is performed on each slice of the high-resolution remote sensing image, and (a)/(b)/(c) in fig. 3 respectively represents the original slice image, the real label corresponding to the slice, and the semantic segmentation result of the parallel cavity convolution. Different terrain objects are represented using different pixels. The figure 3 shows that the remote sensing image semantic segmentation method based on the parallel cavity convolution achieves good effect, and the segmentation result is close to the real label.
In addition, according to step S11, the semantic segmentation results of each slice are sorted and merged, and an unreasonable prediction result is repaired by simply filling holes and removing small connected domains when each slice is spliced.
In summary, according to the technical solution of the present invention, by acquiring the high-resolution remote sensing image in advance, slicing the high-resolution remote sensing image, normalizing and standardizing to obtain a source high-resolution remote sensing image, initializing the characteristic based on the pretrained resnet101 parameter on ImageNet to extract the lower layer network of the network resnet101, constructing a parallel cavity convolution network, extracting shallow features of the source high-resolution remote sensing image, inputting the shallow features into a parallel cavity convolution network to obtain multi-scale information, and the multi-scale information is fused, the fused features and the shallow features are fused again, the full-connection conditional random field is used for restoring the image-level information, the semantic segmentation result is obtained, and the semantic segmentation is realized without adding extra parameters, the field of experience of convolution is enlarged, and compared with standard convolution which achieves the same field of experience, the parallel hole convolution method can save video memory; the parallel computing structure is adopted, so that the nodes in the neural network computing graph are conveniently distributed on distributed hardware, and the computing speed is improved; the multi-scale information is beneficial to capturing detailed objects and large objects by a network, small target objects are not easy to miss, and the semantic segmentation precision is improved; in addition, the void convolution can widely perceive the adjacent objects of the target object, the pixel-level classification can be more effectively carried out by means of the adjacent information, and the pixel-level classification effect is better than that of the standard convolution.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that fall within the spirit and principle of the present invention are intended to be included therein.

Claims (6)

Translated fromChinese
1.一种基于并行空洞卷积的遥感影像语义分割方法,其特征在于,包括以下步骤:1. a remote sensing image semantic segmentation method based on parallel hole convolution, is characterized in that, comprises the following steps:预先获取高分辨率遥感影像,并对高分辨率遥感影像进行切片,进行归一化和标准化,获取源高分辨率遥感影像;Acquire high-resolution remote sensing images in advance, and slice, normalize and standardize high-resolution remote sensing images to obtain source high-resolution remote sensing images;基于ImageNet上预训练的resnet101参数初始化特征提取网络resnet101的低层网络,构建并行空洞卷积网络,并提取源高分辨率遥感影像的浅层特征;Initialize the low-level network of the feature extraction network resnet101 based on the resnet101 parameters pre-trained on ImageNet, construct a parallel atrous convolutional network, and extract the shallow features of the source high-resolution remote sensing images;将浅层特征输入至并行空洞卷积网络获取多尺度信息,并将多尺度信息融合,其中包括设置不同的膨胀率捕获多尺度信息;Input the shallow features into the parallel atrous convolutional network to obtain multi-scale information, and fuse the multi-scale information, including setting different expansion rates to capture multi-scale information;将融合后的特征与浅层特征再次融合,并使用全连接条件随机场修复图像级信息,获取语义分割结果。The fused features are re-fused with shallow features, and image-level information is repaired using a fully-connected conditional random field to obtain semantic segmentation results.2.根据权利要求1所述的基于并行空洞卷积的遥感影像语义分割方法,其特征在于,所述对高分辨率遥感影像进行切片,包括将高分辨率遥感影像切片至长和宽为512像素。2. the remote sensing image semantic segmentation method based on parallel hole convolution according to claim 1, is characterized in that, the described high-resolution remote sensing image is sliced, including slices high-resolution remote sensing image to a length and a width of 512 pixel.3.根据权利要求2所述的基于并行空洞卷积的遥感影像语义分割方法,其特征在于,还包括以下步骤:3. the remote sensing image semantic segmentation method based on parallel hole convolution according to claim 2, is characterized in that, also comprises the following steps:基于切片后的高分辨率遥感影像中提取RGB三通道。Extract RGB three channels from the sliced high-resolution remote sensing images.4.根据权利要求1所述的基于并行空洞卷积的遥感影像语义分割方法,其特征在于,所述并行空洞卷积网络,包括以下步骤:4. The remote sensing image semantic segmentation method based on parallel hole convolution according to claim 1, wherein the parallel hole convolution network comprises the following steps:从标准的普通卷积开始,标定有离散的函数F:
Figure FDA0002924965160000013
和有
Figure FDA0002924965160000015
其k:
Figure FDA0002924965160000014
为离散卷积核,以p为中心展开的卷积计算过程为:Starting from the standard ordinary convolution, a discrete function F is scaled:
Figure FDA0002924965160000013
And there
Figure FDA0002924965160000015
Its k:
Figure FDA0002924965160000014
is a discrete convolution kernel, and the calculation process of convolution centered on p is:
Figure FDA0002924965160000011
Figure FDA0002924965160000011
通过扩充标准卷积,使l为空洞卷积的膨胀率,则空洞卷积为:By expanding the standard convolution, let l be the expansion rate of the atrous convolution, then the atrous convolution is:
Figure FDA0002924965160000012
Figure FDA0002924965160000012
将浅层特征并行通过不同膨胀率的空洞卷积得到多尺度的特征,使用拼接方式融合多尺度特征,即构成并行空洞卷积网络层。Multi-scale features are obtained by parallel convolution of shallow layers with different expansion rates, and multi-scale features are fused by splicing, that is, a parallel hole convolution network layer is formed.5.根据权利要求4所述的基于并行空洞卷积的遥感影像语义分割方法,其特征在于,所述膨胀率分别设置为2、3、4和5。5 . The method for semantic segmentation of remote sensing images based on parallel hole convolution according to claim 4 , wherein the expansion ratios are set to 2, 3, 4, and 5, respectively. 6 .6.根据权利要求1所述的基于并行空洞卷积的遥感影像语义分割方法,其特征在于,所述全连接条件随机场修复图像级信息,包括以下步骤:6. The remote sensing image semantic segmentation method based on parallel hole convolution according to claim 1, wherein the fully connected conditional random field repairs image-level information, comprising the following steps:全连接条件随机场使用的能量函数为:The energy function used by the fully connected conditional random field is:
Figure FDA0002924965160000021
Figure FDA0002924965160000021
其一元势能函数用于描述观测对象和标注的影响:Its unary potential energy function is used to describe the influence of observation objects and labels:θi(xi)=-log P(xi)θi (xi )=-log P(xi )其中,图像的像素点i,P(xi)为网络在像素上分类的概率,二元势能函数描述观测对象之间的相关性:Among them, the pixel point i of the image, P(xi ) is the probability that the network classifies the pixel on the pixel, and the binary potential energy function describes the correlation between the observed objects:
Figure FDA0002924965160000022
Figure FDA0002924965160000022
当xi≠yj时,u(xi,yj)=1,否则,u(xi,yj)=0,另外km(fi,fj)作为fi与fj之间的高斯核,fi是像素i对应的颜色信息,也就是特征向量,wm是该高斯核使用的权重。When xi ≠yj , u(xi , yj )=1, otherwise,u (xi , yj )=0, andkm (fi , fj ) as the distance between fi and fj The Gaussian kernel of , fi is the color information corresponding to pixel i, that is, the feature vector, and wm is the weight used by the Gaussian kernel.
CN202110129416.3A2021-01-292021-01-29Remote sensing image semantic segmentation method based on parallel hole convolutionActiveCN112837320B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202110129416.3ACN112837320B (en)2021-01-292021-01-29Remote sensing image semantic segmentation method based on parallel hole convolution

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202110129416.3ACN112837320B (en)2021-01-292021-01-29Remote sensing image semantic segmentation method based on parallel hole convolution

Publications (2)

Publication NumberPublication Date
CN112837320Atrue CN112837320A (en)2021-05-25
CN112837320B CN112837320B (en)2023-10-27

Family

ID=75931168

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202110129416.3AActiveCN112837320B (en)2021-01-292021-01-29Remote sensing image semantic segmentation method based on parallel hole convolution

Country Status (1)

CountryLink
CN (1)CN112837320B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN113486840A (en)*2021-07-212021-10-08武昌理工学院Building rapid extraction method based on composite network correction
CN113780297A (en)*2021-09-152021-12-10北京百度网讯科技有限公司Image processing method, device, equipment and storage medium
CN114067221A (en)*2022-01-142022-02-18成都数联云算科技有限公司Remote sensing image woodland extraction method, system, device and medium
CN116563816A (en)*2023-04-042023-08-08江西通慧科技集团股份有限公司Road identification method and system based on intelligent lamp post

Citations (12)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN108062756A (en)*2018-01-292018-05-22重庆理工大学Image, semantic dividing method based on the full convolutional network of depth and condition random field
CN108877832A (en)*2018-05-292018-11-23东华大学A kind of audio sound quality also original system based on GAN
CN109285162A (en)*2018-08-302019-01-29杭州电子科技大学 A Semantic Image Segmentation Method Based on Local Area Conditional Random Field Model
CN109461157A (en)*2018-10-192019-03-12苏州大学Image, semantic dividing method based on multi-stage characteristics fusion and Gauss conditions random field
CN109741383A (en)*2018-12-262019-05-10西安电子科技大学 Image depth estimation system and method based on atrous convolution and semi-supervised learning
CN110070022A (en)*2019-04-162019-07-30西北工业大学A kind of natural scene material identification method based on image
CN110232394A (en)*2018-03-062019-09-13华南理工大学A kind of multi-scale image semantic segmentation method
CN110781775A (en)*2019-10-102020-02-11武汉大学Remote sensing image water body information accurate segmentation method supported by multi-scale features
CN111259828A (en)*2020-01-202020-06-09河海大学High-resolution remote sensing image multi-feature-based identification method
US20200234447A1 (en)*2019-01-222020-07-23Kabushiki Kaisha ToshibaComputer vision system and method
CN111539959A (en)*2020-07-132020-08-14浙江省肿瘤医院(浙江省癌症中心) Ultrasound image processing method of thyroid nodule based on cross-layer sparse hole convolution
CN112183360A (en)*2020-09-292021-01-05上海交通大学Lightweight semantic segmentation method for high-resolution remote sensing image

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN108062756A (en)*2018-01-292018-05-22重庆理工大学Image, semantic dividing method based on the full convolutional network of depth and condition random field
CN110232394A (en)*2018-03-062019-09-13华南理工大学A kind of multi-scale image semantic segmentation method
CN108877832A (en)*2018-05-292018-11-23东华大学A kind of audio sound quality also original system based on GAN
CN109285162A (en)*2018-08-302019-01-29杭州电子科技大学 A Semantic Image Segmentation Method Based on Local Area Conditional Random Field Model
CN109461157A (en)*2018-10-192019-03-12苏州大学Image, semantic dividing method based on multi-stage characteristics fusion and Gauss conditions random field
CN109741383A (en)*2018-12-262019-05-10西安电子科技大学 Image depth estimation system and method based on atrous convolution and semi-supervised learning
US20200234447A1 (en)*2019-01-222020-07-23Kabushiki Kaisha ToshibaComputer vision system and method
CN110070022A (en)*2019-04-162019-07-30西北工业大学A kind of natural scene material identification method based on image
CN110781775A (en)*2019-10-102020-02-11武汉大学Remote sensing image water body information accurate segmentation method supported by multi-scale features
CN111259828A (en)*2020-01-202020-06-09河海大学High-resolution remote sensing image multi-feature-based identification method
CN111539959A (en)*2020-07-132020-08-14浙江省肿瘤医院(浙江省癌症中心) Ultrasound image processing method of thyroid nodule based on cross-layer sparse hole convolution
CN112183360A (en)*2020-09-292021-01-05上海交通大学Lightweight semantic segmentation method for high-resolution remote sensing image

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
MINGYANG ZHONG等: "Multi-Receptive Atrous Convolutional Network for Semantic Segmentation", 《2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORK》*
李平: "基于深度学习的命名实体识别方法研究", 《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》*

Cited By (7)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN113486840A (en)*2021-07-212021-10-08武昌理工学院Building rapid extraction method based on composite network correction
CN113780297A (en)*2021-09-152021-12-10北京百度网讯科技有限公司Image processing method, device, equipment and storage medium
CN113780297B (en)*2021-09-152024-03-12北京百度网讯科技有限公司Image processing method, device, equipment and storage medium
CN114067221A (en)*2022-01-142022-02-18成都数联云算科技有限公司Remote sensing image woodland extraction method, system, device and medium
CN114067221B (en)*2022-01-142022-04-15成都数联云算科技有限公司Remote sensing image woodland extraction method, system, device and medium
CN116563816A (en)*2023-04-042023-08-08江西通慧科技集团股份有限公司Road identification method and system based on intelligent lamp post
CN116563816B (en)*2023-04-042025-09-09江西通慧科技集团股份有限公司Road identification method and system based on intelligent lamp post

Also Published As

Publication numberPublication date
CN112837320B (en)2023-10-27

Similar Documents

PublicationPublication DateTitle
CN108154192B (en)High-resolution SAR terrain classification method based on multi-scale convolution and feature fusion
CN110443818B (en) A Graffiti-based Weakly Supervised Semantic Segmentation Method and System
Henry et al.Road segmentation in SAR satellite images with deep fully convolutional neural networks
CN111191583B (en)Space target recognition system and method based on convolutional neural network
CN108647742B (en) A fast target detection method based on lightweight neural network
CN107424159B (en)Image semantic segmentation method based on super-pixel edge and full convolution network
CN109741341B (en)Image segmentation method based on super-pixel and long-and-short-term memory network
CN112837320A (en) A Semantic Segmentation Method of Remote Sensing Image Based on Parallel Atrous Convolution
CN107092870B (en)A kind of high resolution image Semantic features extraction method
CN110728658A (en) A deep learning-based method for detecting weak targets in high-resolution remote sensing images
CN112488025B (en)Double-temporal remote sensing image semantic change detection method based on multi-modal feature fusion
CN108537102B (en)High-resolution SAR image classification method based on sparse features and conditional random field
CN107480726A (en)A kind of Scene Semantics dividing method based on full convolution and shot and long term mnemon
CN114048822A (en) An Image Attention Mechanism Feature Fusion Segmentation Method
CN113065594A (en)Road network extraction method and device based on Beidou data and remote sensing image fusion
CN111126127B (en)High-resolution remote sensing image classification method guided by multi-level spatial context characteristics
CN114863348A (en)Video target segmentation method based on self-supervision
CN112101364B (en)Semantic segmentation method based on parameter importance increment learning
CN106991411B (en)Remote Sensing Target based on depth shape priori refines extracting method
CN114972753A (en) A lightweight semantic segmentation method and system based on contextual information aggregation and assisted learning
CN116645592B (en) A crack detection method and storage medium based on image processing
Bagwari et al.A comprehensive review on segmentation techniques for satellite images
KR102026280B1 (en)Method and system for scene text detection using deep learning
CN116486273B (en) A Method for Extracting Water Body Information from Small Sample Remote Sensing Images
CN111260660A (en) A meta-learning-based transfer method for semantic segmentation of 3D point clouds

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
TA01Transfer of patent application right

Effective date of registration:20230615

Address after:430074 Hubei Province, Wuhan city Hongshan District Luoyu Road No. 1037

Applicant after:HUAZHONG University OF SCIENCE AND TECHNOLOGY

Address before:Room 305-65, 2-3 / F, 5-15 / F, R & D building / unit 1, modern service industry base, Science Park, Huazhong University of science and technology, 13-1, daxueyuan Road, Donghu New Technology Development Zone, Wuhan City, Hubei Province, 430000

Applicant before:Wuhan shanlai Technology Co.,Ltd.

TA01Transfer of patent application right
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp