Movatterモバイル変換


[0]ホーム

URL:


CN113673274A - Road boundary detection method, road boundary detection device, computer equipment and storage medium - Google Patents

Road boundary detection method, road boundary detection device, computer equipment and storage medium
Download PDF

Info

Publication number
CN113673274A
CN113673274ACN202010400797.XACN202010400797ACN113673274ACN 113673274 ACN113673274 ACN 113673274ACN 202010400797 ACN202010400797 ACN 202010400797ACN 113673274 ACN113673274 ACN 113673274A
Authority
CN
China
Prior art keywords
image
point cloud
point
gray value
peak
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010400797.XA
Other languages
Chinese (zh)
Inventor
罗哲
肖振宇
李琛
周旋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Changsha Intelligent Driving Research Institute Co Ltd
Original Assignee
Changsha Intelligent Driving Research Institute Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Changsha Intelligent Driving Research Institute Co LtdfiledCriticalChangsha Intelligent Driving Research Institute Co Ltd
Priority to CN202010400797.XApriorityCriticalpatent/CN113673274A/en
Priority to PCT/CN2021/088583prioritypatent/WO2021227797A1/en
Publication of CN113673274ApublicationCriticalpatent/CN113673274A/en
Pendinglegal-statusCriticalCurrent

Links

Images

Classifications

Landscapes

Abstract

Translated fromChinese

本申请涉及一种道路边界检测方法、装置、计算机设备和存储介质。所述方法包括:获取探测设备针对目标道路采集的点云数据,点云数据包括各点云的点云位置和反射强度;将点云数据投影至二维图像中,获得第一图像,第一图像中包含各点云对应的图像点,各点云对应的图像点的图像灰度值根据该点云的点云位置和反射强度确定;根据各图像灰度值,确定第一图像的目标分割阈值;根据目标分割阈值,对第一图像进行分割处理,获得第二图像,基于第二图像确定道路边界。采用本方法能够提高边界检测准确性。

Figure 202010400797

The present application relates to a road boundary detection method, device, computer equipment and storage medium. The method includes: acquiring point cloud data collected by a detection device for a target road, where the point cloud data includes the point cloud position and reflection intensity of each point cloud; projecting the point cloud data into a two-dimensional image to obtain a first image, the first The image contains image points corresponding to each point cloud, and the image gray value of the image point corresponding to each point cloud is determined according to the point cloud position and reflection intensity of the point cloud; according to the gray value of each image, the target segmentation of the first image is determined Threshold; according to the target segmentation threshold, the first image is segmented to obtain the second image, and the road boundary is determined based on the second image. Using this method can improve the accuracy of boundary detection.

Figure 202010400797

Description

Road boundary detection method, road boundary detection device, computer equipment and storage medium
Technical Field
The application relates to the technical field of intelligent driving, in particular to a road boundary detection method, a road boundary detection device, computer equipment and a storage medium.
Background
At present, road cleaning work in cities mainly depends on a large number of sanitation workers to carry out manual cleaning, and manual cleaning is low in efficiency and high in labor intensity, so that the development trend of road cleaning work is realized by adopting cleaning equipment with high intelligent degree to replace manual cleaning in order to improve the working efficiency and reduce the manual labor intensity and labor cost.
The intelligent electric sweeper adopts electric energy as drive, cannot pollute the environment again in the driving process, has smaller volume compared with a common sweeper, can sweep places which cannot be removed by a garden, a street lane and a plurality of large sweepers, and can detect and track places with boundaries such as road teeth, guardrails, flower beds and the like by utilizing the vehicle-mounted sensor, thereby realizing automatic welting sweeping.
However, when the current intelligent electric sweeper detects the road boundary, the boundary detection is not accurate enough, so that the accurate edge-attaching sweeping cannot be realized.
Disclosure of Invention
In view of the above, it is necessary to provide a road boundary detection method, apparatus, computer device and storage medium capable of improving the accuracy of boundary detection in view of the above technical problems.
A road boundary detection method, the method comprising:
acquiring point cloud data acquired by a detection device aiming at a target road, wherein the point cloud data comprises point cloud positions and reflection intensities of point clouds;
projecting the point cloud data into a two-dimensional image to obtain a first image, wherein the first image comprises image points corresponding to the point clouds, and the image gray value of the image point corresponding to the point clouds is determined according to the point cloud positions and the reflection intensity of the point clouds;
determining a target segmentation threshold of the first image according to each image gray value;
and according to the target segmentation threshold, performing segmentation processing on the first image to obtain a second image, and determining a road boundary based on the second image.
A roadway boundary detection apparatus, the apparatus comprising:
the acquisition module is used for acquiring point cloud data acquired by the detection equipment aiming at a target road, wherein the point cloud data comprises point cloud positions and reflection intensities of point clouds;
the projection module is used for projecting the point cloud data into a two-dimensional image to obtain a first image, wherein the first image comprises image points corresponding to the point clouds, and the image gray value of the image point corresponding to the point clouds is determined according to the point cloud positions and the reflection intensity of the point clouds;
the determining module is used for determining a target segmentation threshold of the first image according to each image gray value;
and the processing module is used for carrying out segmentation processing on the first image according to the target segmentation threshold value to obtain a second image, and determining a road boundary based on the second image.
A computer device comprising a memory and a processor, the memory storing a computer program, the processor implementing the following steps when executing the computer program:
acquiring point cloud data acquired by a detection device aiming at a target road, wherein the point cloud data comprises point cloud positions and reflection intensities of point clouds;
projecting the point cloud data into a two-dimensional image to obtain a first image, wherein the first image comprises image points corresponding to the point clouds, and the image gray value of the image point corresponding to the point clouds is determined according to the point cloud positions and the reflection intensity of the point clouds;
determining a target segmentation threshold of the first image according to each image gray value;
and according to the target segmentation threshold, performing segmentation processing on the first image to obtain a second image, and determining a road boundary based on the second image.
A computer-readable storage medium, on which a computer program is stored which, when executed by a processor, carries out the steps of:
acquiring point cloud data acquired by a detection device aiming at a target road, wherein the point cloud data comprises point cloud positions and reflection intensities of point clouds;
projecting the point cloud data into a two-dimensional image to obtain a first image, wherein the first image comprises image points corresponding to the point clouds, and the image gray value of the image point corresponding to the point clouds is determined according to the point cloud positions and the reflection intensity of the point clouds;
determining a target segmentation threshold of the first image according to each image gray value;
and according to the target segmentation threshold, performing segmentation processing on the first image to obtain a second image, and determining a road boundary based on the second image.
According to the road boundary detection method, the road boundary detection device, the computer equipment and the storage medium, point cloud data collected by the detection equipment aiming at a target road are obtained, wherein the point cloud data comprise point cloud positions and reflection intensities of point clouds; projecting the point cloud data into a two-dimensional image to obtain a first image, wherein the first image comprises image points corresponding to each point cloud, and the image gray value of the image point corresponding to each point cloud is determined according to the point cloud position and the reflection intensity of the point cloud; determining a target segmentation threshold of the first image according to the gray value of each image; and according to the target segmentation threshold, performing segmentation processing on the first image to obtain a second image, and determining a road boundary based on the second image. Therefore, the position information and the reflection intensity information of the point cloud are converted into image gray values to be reflected in the first image, the obtained image gray values can be used for accurately distinguishing image points corresponding to the road surface point cloud and the road boundary point cloud, therefore, the first image is divided according to the target division threshold determined by the image gray values, the divided second image can accurately restore the road boundary information, and the boundary detection effect is improved.
Drawings
Fig. 1 is a schematic flowchart of a road boundary detection method according to an embodiment.
FIG. 2 is a schematic diagram of a first image in one embodiment.
FIG. 3 is a diagram of a second image in one embodiment.
FIG. 4 is a flowchart illustrating the step of determining the target segmentation threshold for the first image according to the gray-level value of each image in one embodiment.
FIG. 5 is a diagram illustrating an exemplary gray scale value distribution of an image.
FIG. 6 is a flowchart illustrating the step of determining a road boundary based on the second image in one embodiment.
FIG. 7 is a schematic diagram of a road boundary curve in one embodiment.
Fig. 8 is a flowchart illustrating a road boundary detection method according to another embodiment.
Fig. 9 is a block diagram showing the structure of a road boundary detection device according to an embodiment.
FIG. 10 is a diagram showing an internal structure of a computer device according to an embodiment.
FIG. 11 is a diagram illustrating an internal structure of a computer device in one embodiment.
Detailed Description
In order to make the objects, technical solutions and advantages of the present application more apparent, the present application is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the present application and are not intended to limit the present application.
The road boundary detection method can be applied to an intelligent vehicle driving system, the intelligent vehicle driving system comprises an industrial personal computer and detection equipment (such as a laser radar), the detection equipment can be installed on a vehicle, corresponding point cloud data are collected along with the movement of the vehicle on a target road, the industrial personal computer obtains the point cloud data collected by the detection equipment, the point cloud data are processed, a road boundary is determined, and the vehicle can be further controlled to move along the road boundary.
In one embodiment, as shown in fig. 1, a road boundary detection method is provided, which is described by taking an example of the method applied to an industrial personal computer, and includes the following steps S102 to S108.
S102, point cloud data collected by the detection equipment for the target road are obtained, and the point cloud data comprise point cloud positions and reflection intensities of the point clouds.
The detection device can adopt a laser radar, the working principle of the laser radar is to transmit a detection signal (laser beam) to a target, then the received signal reflected from the target is compared with the transmitted signal, and relevant information of the target, such as parameters of distance, direction, height, reflection intensity and the like, can be obtained after appropriate processing, so that the target can be detected, tracked and identified. Specifically, the laser radar is installed on a vehicle running on a target road, and point cloud data on the target road is collected along with the movement of the vehicle, wherein the point cloud data comprises a point cloud position and a reflection intensity of each point cloud, and the point cloud position can be a coordinate position of the point cloud under a laser radar coordinate system.
In one embodiment, the lidar is mounted on the top of the vehicle and is inclined downward at an angle (e.g., 15 degrees) toward the front, the horizontal scanning range of the lidar is approximately 100 degrees, the vertical scanning angle of the lidar is approximately 40 degrees, and the scanning range can cover most of the area right in front of the vehicle. And after the laser radar is installed, calibrating external parameters of the laser radar to obtain calibration parameters of the laser radar reaching the vehicle body. The industrial personal computer performs coordinate conversion on original point cloud data acquired by the laser radar after obtaining the original point cloud data, specifically, the point cloud data under a laser radar coordinate system is converted into an automobile body coordinate system through a laser radar which is calibrated in advance, and a conversion formula can be as follows:
Figure BDA0002489383350000051
wherein, PLRepresenting point cloud coordinate points (forward x-axis, leftward y-axis and upward z-axis) in a laser radar coordinate system, PCAnd point cloud coordinate points (the x axis is right in front of the vehicle, the y axis is right in left of the vehicle, the z axis is right above the vehicle) under the vehicle body coordinate system are represented, R is a 3 x 3 rotation matrix, and T is a translation vector. R and T can be calibrated by combining actual conditions.
And S104, projecting the point cloud data into a two-dimensional image to obtain a first image, wherein the first image comprises image points corresponding to the point clouds, and the image gray value of the image point corresponding to each point cloud is determined according to the point cloud position and the reflection intensity of the point cloud.
The point cloud data are projected into the two-dimensional image, so that all information contained in the point cloud is converted into the image, and the point cloud data can be processed more conveniently in the image. Each point cloud corresponds to one image point in the image, and the position information and the reflection intensity information of each point cloud are converted into the position information and the image gray value of the corresponding image point.
For example, in a 1000 × 1000 single-channel grayscale image, the coordinate origin is at the upper left of the image, the w coordinate axis represents the width of the image, the h coordinate axis represents the height of the image, and assuming that the image coordinate point (w is 500, h is 1000) represents the coordinate point of the vehicle body coordinate origin in the image, the conversion formula of the coordinates of the point cloud in the image coordinate system and the coordinates thereof in the vehicle body coordinate system is as follows:
w′=500-y*100;h′=1000-x*100
wherein, w 'and h' represent the coordinate value of the point cloud under the image coordinate system, x and y represent the longitudinal value and the transverse value of the point cloud under the vehicle body coordinate system respectively, the unit is meter (m), and the unit of the image coordinate value is centimeter (cm). Therefore, the position information of the point cloud in the x-axis direction and the y-axis direction under the vehicle body coordinate system is converted into the coordinate values of the corresponding image points in the image. The position information of the point cloud in the z-axis direction under the vehicle body coordinate system represents the height of the point cloud, and the height and the reflection intensity of the point cloud are respectively mapped to a pixel gray value space (0-255), so that the height and the reflection intensity of the point cloud are converted into image gray values of corresponding image points in the image. Referring to fig. 2, a schematic diagram of a first image in an embodiment is shown, where the first image includes image points corresponding to road surface point clouds (hereinafter referred to as road surface image points) and image points corresponding to road boundary point clouds (hereinafter referred to as boundary image points).
And S106, determining a target segmentation threshold of the first image according to the gray value of each image.
The target segmentation threshold is used for segmenting road surface image points and boundary image points in the first image. The point cloud position and the reflection intensity of the road surface point cloud and the boundary point cloud are different, so that the image gray values of the corresponding image points are also different, and an image gray value capable of distinguishing the road surface image point from the boundary image point is determined from the image gray values in the first image and is used as a target segmentation threshold.
And S108, segmenting the first image according to the target segmentation threshold value to obtain a second image, and determining a road boundary based on the second image.
Image points in the first image can be divided into two types according to the target segmentation threshold, wherein one type corresponds to road surface image points, and the other type corresponds to boundary image points. Specifically, the first image may be binarized, and the second image may be obtained by setting the grayscale value of the divided road surface image point to 255 (white) and the grayscale value of the divided boundary image point to 0 (black). Referring to FIG. 3, a schematic diagram of a second image in an embodiment is shown, in which road surface image points are removed and boundary image points are retained, as compared to the first image shown in FIG. 2.
In the road boundary detection method, the position information and the reflection intensity information of the point cloud are converted into the image gray value to be reflected in the first image, and the obtained image gray value can be used for accurately distinguishing the road surface point cloud and the image points corresponding to the road boundary point cloud.
In one embodiment, the point cloud position includes a point cloud height, and the determining of the image gray value of the image point corresponding to each point cloud includes: and determining the image gray value of the image point corresponding to each point cloud according to the point cloud height of each point cloud, the corresponding first conversion factor and first distribution weight, the reflection intensity of each point cloud, the corresponding second conversion factor and second distribution weight.
The first conversion factor and the second conversion factor are respectively used for converting the point cloud height and the reflection intensity into gray values, and the converted values are between 0 and 255. The first distribution weight value and the second distribution weight value respectively represent the weight occupied by the gray value after the point cloud height and the reflection intensity are converted, and the value obtained by adding the first distribution weight value and the second distribution weight value is 1. And determining the final image gray value of the image point corresponding to the point cloud together by the gray value and the weight of the converted point cloud height and reflection intensity.
Specifically, the step of determining the image gray value of the image point corresponding to each point cloud according to the point cloud height of each point cloud, the first conversion factor and the first distribution weight corresponding to each point cloud, and the reflection intensity of each point cloud, the second conversion factor and the second distribution weight corresponding to each point cloud may include the following steps: multiplying the point cloud height and the reflection intensity of each point cloud by the first conversion factor and the second conversion factor respectively to obtain a first conversion gray value and a second conversion gray value of each point cloud; multiplying the first conversion gray value and the second conversion gray value of each point cloud by the first distribution weight value and the second distribution weight value respectively to obtain a first weighted gray value and a second weighted gray value of each point cloud; and adding the first weighted gray value and the second weighted gray value of each point cloud to determine the image gray value of the image point corresponding to each point cloud.
The calculation formula of the image gray value of the image point corresponding to each point cloud is as follows:
P=m*j*z+n*k*i
wherein z and i represent the point cloud height and the reflection intensity, respectively, j and k represent the first conversion factor and the second conversion factor, respectively, j x z and k x i represent the first conversion gray value (between 0 and 255) and the second conversion gray value (between 0 and 255), respectively, m and n represent the first distribution weight value and the second distribution weight value, respectively, m + n is 1, P represents the image gray value, and P is 0 ≦ P ≦ 255. j. k, m and n can be calibrated by combining the actual conditions.
In the embodiment, the point cloud height and the reflection intensity information are converted into the gray values through the conversion factors corresponding to the point cloud height and the reflection intensity, and the image gray values of the image points corresponding to the point cloud are determined together based on the gray values obtained after the point cloud height and the reflection intensity are converted and the corresponding distribution weights, so that the determined image gray values can be used for accurately distinguishing the road surface image points and the boundary image points, and the boundary detection effect is improved.
In one embodiment, before the step of determining the target segmentation threshold of the first image according to the gray-scale value of each image, the method further comprises: according to the position of each image point in the first image, carrying out blocking processing on the first image to obtain at least two blocked images; the step of determining a target segmentation threshold for the first image based on the respective image gray-scale values comprises: for each block image, determining a block segmentation threshold of the block image according to the image gray value of each image point in the block image, wherein the target segmentation threshold comprises each block segmentation threshold; according to the target segmentation threshold, the step of carrying out segmentation processing on the first image to obtain a second image comprises the following steps: and according to the block segmentation threshold value of each block image, respectively carrying out segmentation processing on the corresponding block image to obtain a second image.
When a vehicle runs on an uneven road surface, the vehicle may bump up and down or left and right, so that the height of the point cloud fluctuates greatly, the point cloud fluctuation close to the vehicle body is small, the point cloud fluctuation far away from the vehicle body is large, and a single segmentation threshold value is difficult to find to accurately segment the point clouds of a near part and a far part. Based on this, before determining the target segmentation threshold of the first image, the first image is segmented, and specifically, the first image may be segmented into an upper portion and a lower portion, that is, two segmented images, where the lower portion represents a point cloud image in front of the vehicle and closer to the vehicle body (0-5 m), and the upper portion represents a point cloud image in front of the vehicle and farther from the vehicle body (5-10 m). After the block images are obtained, for each block image, the following processing is performed: and determining a block segmentation threshold of the block image according to the image gray value of each image point in the block image, and then performing segmentation processing on the block image according to the block segmentation threshold of the block image. After each block image is subjected to segmentation processing, the images subjected to segmentation processing of each block image jointly form a second image.
The image blocking process is not limited to the vertical blocking, and other blocking processes may be performed according to actual circumstances, for example, when the road surface is inclined left and right, left and right blocking may be considered.
In this embodiment, by partitioning the first image, determining the segmentation threshold of each partitioned image, and performing segmentation processing on each partitioned image, the influence of point cloud fluctuation caused by uneven road surface or vehicle fluctuation on the segmentation threshold can be reduced, and the segmentation accuracy of road surface image points and boundary image points can be improved.
In an embodiment, as shown in fig. 4, the step of determining the target segmentation threshold of the first image according to the gray-level value of each image may specifically include the following steps S1062 to S1066.
S1062, obtaining a gray value distribution graph according to the gray value of each image and the number of the image points corresponding to the gray value, wherein the first coordinate of the gray value distribution graph represents the gray value of the image, and the second coordinate represents the number of the image points.
Specifically, histogram statistics may be performed on image gray values of all image points in the first image to obtain a gray value distribution graph, where the range of the image gray values is 0 to 255, that is, the image gray values include 256 values. For example, the first coordinate may be divided into 26 parts, each of the first 25 parts includes 10 gray-scale values (0-9, 10-19, …, 240-249), and the 26 th part includes 6 gray-scale values (250-255). Referring to fig. 5, a schematic diagram of an image gray-level value distribution in an embodiment is shown, in which an abscissa represents an image gray-level value and an ordinate represents a second coordinate represents a number of image points corresponding to the image gray-level value.
S1064, detecting wave crests in the gray value distribution diagram in the second coordinate direction, and determining a first wave crest and a second wave crest which are maximum in the second coordinate value in each wave crest.
The gray value distribution map may have a plurality of peaks in the second coordinate (i.e. ordinate) direction, and it can be understood that the image points are more likely to form peaks in the ordinate direction when the image gray value is closer. If more than two wave crests exist in the gray value distribution diagram in the longitudinal coordinate direction, longitudinal coordinate values of the wave crests are obtained, and the two wave crests with the largest longitudinal coordinate values (namely the two wave crests corresponding to the two longitudinal coordinates arranged in the first two are sorted from large to small) are selected to serve as the first wave crest and the second wave crest respectively. As shown in fig. 5, there are two peaks in the ordinate direction, and the higher peak is taken as the first peak, and the lower peak is taken as the second peak.
S1066, based on whether other peaks exist between the first peak and the second peak, selecting a corresponding segmentation threshold determination mode to determine a target segmentation threshold of the first image.
When no other peak exists between the first peak and the second peak, the road boundary is considered to be obvious, and the difference between the image gray values of the boundary image point and the road surface image point is large. When other wave crests exist between the first wave crest and the second wave crest, the road boundary is not obvious, and the image gray values of the boundary image points and the road surface image points which are lower in height from the road surface are closer and are not easy to distinguish due to the fact that the image gray values are highly related to the point cloud. In these two different cases, different segmentation threshold determination methods are used to determine the target segmentation threshold for the first image.
In this embodiment, the target segmentation threshold of the first image is determined by detecting the peak in the gray value distribution map and selecting a corresponding segmentation threshold determination mode based on whether other peaks exist between the first peak and the second peak, so that the segmentation effect of the road surface image point and the boundary image point can be improved, and the road boundary information can be restored more accurately.
In one embodiment, when no other peak exists between the first peak and the second peak, the maximum inter-class variance value is determined according to the gray value of each image, and is used as the target segmentation threshold of the first image.
Specifically, the maximum inter-class variance method (otsu) may be used to calculate the object segmentation threshold of the first image, where for the first image, the object segmentation threshold of the foreground (i.e., the boundary, which may refer to all object boundaries except the road surface) and the background (i.e., the road surface) is denoted as T, and the ratio of the foreground image points to the whole image points is denoted as w0The mean gray value of u0(ii) a The proportion of background image points to the whole image point is w1The mean gray value of u1(ii) a The total average gray value of the image is recorded as mu, between-class squareThe difference is recorded as g; the total number of image points is denoted as W x H, and the number of image points with image gray-scale values smaller than a threshold value T is denoted as N0The number of image points with the gray value of the image greater than the threshold value T is recorded as N1Then, there are: w is a0=N0/M*N,w1=N0/M*N,w0+w1=1,N0+N1=M*N,u=u0*w0+u1*w1,g=w0*(u-u0)2+w1*(u-u1)2=w0*w1*(u0-u1)2And traversing the variance value between classes of each grade, and taking the variance value between the maximum classes as a target segmentation threshold T.
In this embodiment, the target segmentation threshold of the first image is determined by the maximum inter-class variance value, and since the inter-class variance value represents the deviation degree between the foreground gray value and the background gray value and the average gray value, and the larger the deviation degree is, the better the segmentation effect is, the maximum inter-class variance value is selected as the target segmentation threshold, and the segmentation of the road surface image point and the boundary image point can be well realized in most scenes.
In one embodiment, when at least one other peak exists between the first peak and the second peak, a peak adjacent to the first peak is selected from the at least one other peak as a third peak; the second coordinate value of the first peak is greater than or equal to the second coordinate value of the second peak; and determining a target segmentation threshold of the first image according to the image gray value corresponding to the minimum second coordinate value between the first peak and the third peak.
And sequencing the second coordinate values of all the wave crests from large to small, wherein the first wave crest is the wave crest arranged corresponding to the first coordinate value, the second wave crest is the wave crest arranged corresponding to the second coordinate value, and the third wave crest is positioned between the first wave crest and the second wave crest and is adjacent to the first wave crest. And determining an image gray value corresponding to the minimum second coordinate value between the first peak and the third peak as a target segmentation threshold of the first image. For example, when the gray-level distribution graph is a histogram, the image gray-level value corresponding to the minimum second coordinate value between the first peak and the third peak is a gray-level value interval including more than one gray-level value, and the average gray-level value in the interval can be obtained by adding all the gray-level values in the interval and dividing the sum by the number of the gray-level values in the interval, and is used as the target segmentation threshold.
In this embodiment, the target segmentation threshold of the first image is determined by the image gray value corresponding to the minimum second coordinate value between the first peak and the third peak, and since the image gray value is related to the point cloud height, the image gray values of the boundary image point and the road image point which are lower in height from the road surface are closer, the distribution of the road image points is the largest, the third peak is adjacent to the first peak corresponding to the first peak, and the third peak corresponds to the boundary image point which is lower in height from the road surface, so that the image gray value corresponding to the minimum second coordinate value between the first peak and the third peak is selected as the target segmentation threshold of the first image, the segmentation effect of the road image point and the boundary image point which are lower in height from the road surface can be improved, and the road boundary information can be more accurately restored.
In one embodiment, as shown in fig. 6, the step of determining the road boundary based on the second image may specifically include the following steps S602 to S608.
S602, extracting boundary contour image points on the object side from the second image.
In addition to the image points at the road edge, the boundary image points in the second image may also include image points at the boundary of guard rails, flower beds and the like, and for the application of the intelligent sweeper, the boundary image points are more concerned with the contour of the road edge, so that the sweeper can clean along the road edge, and therefore, other image points which do not belong to the road edge can be filtered, and only the contour image points at the road edge are reserved. Further, when the sweeper travels along the boundary, the sweeper travels either towards the left boundary or towards the right boundary, and the driving habit is determined as driving towards the right, so that the boundary contour image points on the left side in the second image can be filtered, and the boundary contour image points on the right side (i.e. the target map) are reserved.
Specifically, before extracting the boundary contour image points of the second image, the second image is preprocessed, and the preprocessing may include removing discrete points and erosion expansion, because the image points projected into the image are discrete when the resolution of the lidar is low, and the preprocessing may not only filter accidental false detection points of the lidar, but also ensure the integrity of the boundary. After the second image is preprocessed, all contours in the second image can be stored in an array sequence by using an image processing contour searching method, in a boundary contour on the right side, an image point between a first image point traversed from left to right of the image and an image point closest to the vehicle body is considered as a required boundary contour image point, and an image point on the left side of the image in the boundary contour image points is selected as a final boundary contour image point.
S604, when the boundary contour formed based on the image points of the boundary contour is discontinuous, the positions of two end points of the boundary contour at each disconnected position are obtained, and the connection curve of the boundary contour at each disconnected position is determined according to the positions of the end points and the tangent lines at the end points.
The boundary contour image points extracted from the second image may have partial missing boundaries, which may cause discontinuity of the boundary contour formed based on the boundary contour image points, and if a plurality of boundary contours are included, different boundary contours may not be continuous. Based on this, an Hermite (Hermite) interpolation method can be used to interpolate between the two endpoints at each break of the boundary contour, according to Hermite: the principle that two end point coordinates of a known curve and a tangent line at an end point can determine a curve is adopted, the coordinates of each end point are obtained, the tangent line at each end point is calculated according to the coordinates of each end point and the coordinates of adjacent points, and then a connection curve at each disconnected point is obtained.
And S606, interpolating between two endpoints at each break of the boundary contour according to each connection curve to obtain interpolated boundary contour image points.
After the connection curves of the boundary contour at each disconnection position are obtained, Hermite interpolation is carried out between two endpoints at each disconnection position, the partially missing boundary contour can be supplemented, and all the boundary contours can be connected, so that the obtained image points of the interpolated boundary contour can embody the boundary contour more completely.
And S608, performing curve fitting on the interpolated boundary contour image points to obtain a road boundary curve, and determining a road boundary based on the road boundary curve.
Specifically, the B-spline curve is adopted to fit the interpolated boundary contour image points, and compared with other fitting modes (such as a least square method fitting mode), the road boundary curve closer to the real road boundary can be obtained through B-spline curve fitting, so that the high-precision requirement of the sweeper on welt sweeping is met. Referring to fig. 7, a schematic diagram of a road boundary curve in one embodiment is shown, and the final road boundary curve is a curve indicated by an arrow. After the final road boundary curve is obtained, the coordinates of each boundary point on the road boundary curve are converted from the image coordinate system to the vehicle body coordinate system according to the conversion formula in the foregoing, so that the sweeper is controlled to carry out welting sweeping along the road boundary curve.
In this embodiment, by performing extraction, interpolation, and fitting processing on the road boundary contour image points included in the second image, the obtained road boundary curve can reflect the road boundary information more completely and truly, and the boundary detection effect is improved.
In one embodiment, after obtaining the road boundary curve, the method further comprises the following steps: and filtering the road boundary curve to obtain a filtered road boundary curve, and determining a road boundary based on the filtered road boundary curve.
Road boundary detection is applied to when intelligent motor sweeper, in actual measuring environment, may meet various emergency, for example the motor sweeper appears very big shake in the in-process of traveling, may cause road boundary curve to appear very big fluctuation this moment, and based on this, through filtering road boundary curve, can reduce road boundary curve's volatility.
Specifically, a Kalman Filter (KF) method may be used to filter the road boundary curve, and considering that the sweeper moves at a relatively low speed at a uniform speed during operation, a uniform motion model is suitable for use, so the kalman filter design process is as follows:
the formula of the kalman filter prediction section includes:
x′=Fx+f (1)
P′=FPFT+Q (2)
where x represents the state vector, F represents the state transition matrix, FTRepresenting state transition matrix transposition, f representing external influence, x 'representing the updated state vector, P representing the state covariance matrix, Q representing the process noise matrix, and P' representing the state covariance matrix after state update.
Setting the state vector to x ═ x, y, vx,vy]TWhere x and y represent the coordinates of the boundary point in the image coordinate system, v, respectivelyxAnd vyEach of the velocities representing the boundary points is a uniform motion model, and f may be set to 0,
Figure BDA0002489383350000131
the point cloud returned by the laser radar is corresponding to the boundary point and is obtained by direct measurement of the laser radar, and the speed of the boundary point cannot be measured, so that the position information of the boundary point can be accurately acquired, the uncertainty is low, and the uncertainty is high for the speed information of the boundary point. Can be provided with
Figure BDA0002489383350000132
Q has an influence on the entire system, but it is difficult to determine how much the influence is on the system, so that setting Q as an identity matrix, then
Figure BDA0002489383350000133
The formula of the kalman filter measurement section includes:
y=Z-Hx′ (3)
S=HP′HT+R (4)
K=P′HTS-1 (5)
x=x′+Ky (6)
P=(I-KH)P′ (7)
wherein the formulas (3), (6) and (7) are observation formulasEquations (4) and (5) are used to find the Kalman gain K, the observed value of the boundary point
Figure BDA0002489383350000134
H represents a measurement matrix, which is mainly used for converting a state vector space into a measurement space and is obtained according to Z ═ Hx
Figure BDA0002489383350000135
R represents a measurement noise matrix, the value represents the difference between the measured value and the true value, and when the laser radar ranging precision is 2cm, the difference can be set
Figure BDA0002489383350000136
S represents a temporary variable of the simplified formula and I represents the identity matrix with the state vector. Through all the known variables, updating the state vector x and the state covariance matrix P is realized by using the formulas (6) and (7), and the prediction and measurement are iterated continuously, so that a predicted state vector close to a true value can be obtained.
In this embodiment, by filtering the road boundary curve, the volatility of the boundary curve can be reduced, and the boundary detection effect can be improved. It is to be understood that the filtering method is not limited to the Kalman Filter (KF) described above, and for example, the road boundary curve may be filtered by using a filtering method such as an extended kalman filter (EFK) or an Unscented Kalman Filter (UKF).
In one embodiment, as shown in fig. 8, a road boundary detection method is provided, which is described by taking an example of the method applied to an industrial personal computer, and includes the following steps S801 to S816.
S801, point cloud data collected by the laser radar aiming at a target road are obtained, wherein the point cloud data comprise point cloud positions and reflection intensities of all point clouds.
And S802, performing coordinate conversion on the point cloud data, and converting the point cloud data from a laser radar coordinate system to a vehicle body coordinate system.
And S803, projecting the point cloud data converted into the vehicle body coordinate system into a two-dimensional image to obtain a first image, wherein the first image comprises image points corresponding to the point clouds, so that the point cloud coordinates are converted into the image coordinate system from the vehicle body coordinate system, and the image gray values of the image points corresponding to the point clouds are determined according to the point cloud height and the reflection intensity of the point clouds.
S804, according to the position of each image point in the first image, the first image is subjected to blocking processing, and at least two blocking images are obtained.
And S805, for each block image, obtaining a gray value distribution graph according to the gray value of each image in the block image and the number of the image points corresponding to the gray value, wherein the first coordinate of the gray value distribution graph represents the gray value of the image, and the second coordinate represents the number of the image points.
S806, detecting wave peaks in the gray value distribution diagram in the second coordinate direction, and determining a first wave peak and a second wave peak with the largest second coordinate value in each wave peak.
S807, judging whether other wave crests exist between the first wave crest and the second wave crest, if so, entering step S808, otherwise, entering step S809.
S808, selecting a wave peak adjacent to the first wave peak as a third wave peak from other existing wave peaks; and determining a target segmentation threshold of the block image according to the image gray value corresponding to the minimum second coordinate value between the first peak and the third peak.
And S809, determining the maximum inter-class variance value according to the gray value of each image, and using the maximum inter-class variance value as a block segmentation threshold value of the block image.
And S810, respectively carrying out segmentation processing on the corresponding segmented images according to the segmented segmentation threshold of each segmented image to obtain second images.
S811, boundary outline image points on the object side are extracted from the second image.
S812, when the boundary contour formed based on the image points of the boundary contour is discontinuous, acquiring the positions of two end points of the boundary contour at each disconnected position, and determining a connection curve of the boundary contour at each disconnected position according to the positions of the end points and tangent lines at the end points.
And S813, interpolating between two endpoints at each break of the boundary contour according to each connection curve to obtain interpolated boundary contour image points.
And S814, performing curve fitting on the interpolated boundary contour image points to obtain a road boundary curve.
And S815, filtering the road boundary curve to obtain the filtered road boundary curve.
And S816, converting the coordinates of each boundary point on the filtered road boundary curve from the image coordinate system to the vehicle body coordinate system, and determining the road boundary.
For specific description of steps S801 to S816, reference may be made to the foregoing embodiments, which are not described herein again. In the embodiment, the point cloud data is projected to the two-dimensional image, so that the height and the reflection intensity of the point cloud are converted into the gray value of the image, and important information is provided for extracting the road boundary; by carrying out blocking processing on the images and calculating the segmentation threshold value of each image, the influence of point cloud fluctuation on the segmentation threshold value caused by the conditions of uneven road surface or vehicle fluctuation and the like can be reduced; and the target segmentation threshold of the image is determined by detecting the wave crest in the gray value distribution diagram and selecting a corresponding segmentation threshold determination mode based on whether other wave crests exist between the first wave crest and the second wave crest, so that the segmentation effect of the road surface image points and the boundary image points can be improved, and the road boundary information can be restored more accurately.
It should be understood that although the steps in the flowcharts of fig. 1, 4, 6, 8 are shown in order as indicated by the arrows, the steps are not necessarily performed in order as indicated by the arrows. The steps are not performed in the exact order shown and described, and may be performed in other orders, unless explicitly stated otherwise. Moreover, at least some of the steps in fig. 1, 4, 6, and 8 may include multiple steps or multiple stages, which are not necessarily performed at the same time, but may be performed at different times, which are not necessarily performed in sequence, but may be performed alternately or alternately with other steps or at least some of the other steps.
In one embodiment, as shown in fig. 9, there is provided a road boundary detecting apparatus 900, including: an obtainingmodule 910, a projectingmodule 920, a determiningmodule 930, and aprocessing module 940, wherein:
an obtainingmodule 910, configured to obtain point cloud data collected by a detection device for a target road, where the point cloud data includes a point cloud position and a reflection intensity of each point cloud.
Theprojection module 920 is configured to project the point cloud data to a two-dimensional image to obtain a first image, where the first image includes image points corresponding to each point cloud, and an image gray value of an image point corresponding to each point cloud is determined according to a point cloud position and a reflection intensity of the point cloud.
A determiningmodule 930, configured to determine a target segmentation threshold of the first image according to the gray-level value of each image.
And aprocessing module 940, configured to perform segmentation processing on the first image according to the target segmentation threshold to obtain a second image, and determine a road boundary based on the second image.
In one embodiment, the point cloud position includes a point cloud height, and theprojection module 920 further includes a gray value determining unit, configured to determine an image gray value of the image point corresponding to each point cloud according to the point cloud height of each point cloud and its corresponding first conversion factor and first distribution weight, and the reflection intensity of each point cloud and its corresponding second conversion factor and second distribution weight.
In one embodiment, the gray value determining unit is specifically configured to: multiplying the point cloud height and the reflection intensity of each point cloud by the first conversion factor and the second conversion factor respectively to obtain a first conversion gray value and a second conversion gray value of each point cloud; multiplying the first conversion gray value and the second conversion gray value of each point cloud by the first distribution weight value and the second distribution weight value respectively to obtain a first weighted gray value and a second weighted gray value of each point cloud; and adding the first weighted gray value and the second weighted gray value of each point cloud to determine the image gray value of the image point corresponding to each point cloud.
In an embodiment, the determiningmodule 930 further comprises an image blocking unit, configured to perform a blocking process on the first image according to the position of each image point in the first image, to obtain at least two blocked images. The determiningmodule 930 is further configured to determine, for each block image, a block segmentation threshold of the block image according to an image gray value of each image point in the block image; the target segmentation threshold includes a respective block segmentation threshold. Theprocessing module 940 is further configured to perform segmentation processing on the corresponding segmented images according to the segmentation threshold of each segmented image to obtain a second image.
In one embodiment, the determiningmodule 930 includes: the device comprises a gray value distribution acquisition unit, a wave crest detection unit and a division threshold value determination unit. Wherein:
and the gray value distribution acquisition unit is used for acquiring a gray value distribution graph according to the gray value of each image and the number of the image points corresponding to the gray value, wherein the first coordinate of the gray value distribution graph represents the gray value of the image, and the second coordinate represents the number of the image points.
And the peak detection unit is used for detecting peaks in the gray value distribution diagram in the second coordinate direction and determining a first peak and a second peak with the largest second coordinate value in each peak.
And the segmentation threshold determining unit is used for selecting a corresponding segmentation threshold determining mode to determine the target segmentation threshold of the first image based on whether other peaks exist between the first peak and the second peak.
In an embodiment, the segmentation threshold determination unit is specifically configured to determine a maximum inter-class variance value according to gray values of the respective images as the target segmentation threshold of the first image when no other peak exists between the first peak and the second peak.
In an embodiment, the segmentation threshold determination unit is specifically configured to select, when at least one other peak exists between the first peak and the second peak, a peak adjacent to the first peak as a third peak from the at least one other peak; the second coordinate value of the first peak is greater than or equal to the second coordinate value of the second peak; and determining a target segmentation threshold of the first image according to the image gray value corresponding to the minimum second coordinate value between the first peak and the third peak.
In one embodiment, theprocessing module 940 further includes: the device comprises an extraction unit, a connection curve determination unit, an interpolation unit and a fitting unit. Wherein:
and the extraction unit is used for extracting the boundary contour image point of the target side from the second image.
And the connection curve determining unit is used for acquiring the positions of two end points of the boundary contour at each disconnected position when the boundary contour formed based on the image points of the boundary contour is discontinuous, and determining the connection curve of the boundary contour at each disconnected position according to the positions of the end points and the tangent lines at the end points.
And the interpolation unit is used for interpolating between two endpoints at each disconnection position of the boundary contour according to each connection curve to obtain an interpolated boundary contour image point.
And the fitting unit is used for performing curve fitting on the interpolated boundary contour image points to obtain a road boundary curve and determining a road boundary based on the road boundary curve.
In one embodiment, theprocessing module 940 further includes a filtering unit, configured to filter the road boundary curve to obtain a filtered road boundary curve, and determine the road boundary based on the filtered road boundary curve.
For the specific definition of the road boundary detection device, reference may be made to the above definition of the road boundary detection method, which is not described herein again. The various modules in the road boundary detection apparatus described above may be implemented in whole or in part by software, hardware, and combinations thereof. The modules can be embedded in a hardware form or independent from a processor in the computer device, and can also be stored in a memory in the computer device in a software form, so that the processor can call and execute operations corresponding to the modules.
In one embodiment, a computer device is provided, which may be a server, and its internal structure diagram may be as shown in fig. 10. The computer device includes a processor, a memory, and a network interface connected by a system bus. Wherein the processor of the computer device is configured to provide computing and control capabilities. The memory of the computer device comprises a nonvolatile storage medium and an internal memory. The non-volatile storage medium stores an operating system, a computer program, and a database. The internal memory provides an environment for the operation of an operating system and computer programs in the non-volatile storage medium. The network interface of the computer device is used for communicating with an external terminal through a network connection. The computer program is executed by a processor to implement a road boundary detection method.
In one embodiment, a computer device is provided, which may be a terminal, and its internal structure diagram may be as shown in fig. 11. The computer device includes a processor, a memory, a communication interface, a display screen, and an input device connected by a system bus. Wherein the processor of the computer device is configured to provide computing and control capabilities. The memory of the computer device comprises a nonvolatile storage medium and an internal memory. The non-volatile storage medium stores an operating system and a computer program. The internal memory provides an environment for the operation of an operating system and computer programs in the non-volatile storage medium. The communication interface of the computer device is used for carrying out wired or wireless communication with an external terminal, and the wireless communication can be realized through WIFI, an operator network, NFC (near field communication) or other technologies. The computer program is executed by a processor to implement a road boundary detection method. The display screen of the computer equipment can be a liquid crystal display screen or an electronic ink display screen, and the input device of the computer equipment can be a touch layer covered on the display screen, a key, a track ball or a touch pad arranged on the shell of the computer equipment, an external keyboard, a touch pad or a mouse and the like.
Those skilled in the art will appreciate that the configurations shown in fig. 10 or 11 are merely block diagrams of some configurations relevant to the present disclosure, and do not constitute a limitation on the computing devices to which the present disclosure may be applied, and that a particular computing device may include more or less components than those shown, or may combine certain components, or have a different arrangement of components.
In one embodiment, a computer device is provided, comprising a memory and a processor, the memory having stored therein a computer program, the processor implementing the steps of the above-described method embodiments when executing the computer program.
In an embodiment, a computer-readable storage medium is provided, on which a computer program is stored, which computer program, when being executed by a processor, carries out the steps of the respective method embodiment as described above.
It should be understood that the terms "first", "second", etc. in the above-described embodiments are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implying any number of technical features indicated.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by hardware instructions of a computer program, which can be stored in a non-volatile computer-readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. Any reference to memory, storage, database or other medium used in the embodiments provided herein can include at least one of non-volatile and volatile memory. Non-volatile Memory may include Read-Only Memory (ROM), magnetic tape, floppy disk, flash Memory, optical storage, or the like. Volatile Memory can include Random Access Memory (RAM) or external cache Memory. By way of illustration and not limitation, RAM can take many forms, such as Static Random Access Memory (SRAM) or Dynamic Random Access Memory (DRAM), among others.
The technical features of the above embodiments can be arbitrarily combined, and for the sake of brevity, all possible combinations of the technical features in the above embodiments are not described, but should be considered as the scope of the present specification as long as there is no contradiction between the combinations of the technical features.
The above-mentioned embodiments only express several embodiments of the present application, and the description thereof is more specific and detailed, but not construed as limiting the scope of the invention. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the concept of the present application, which falls within the scope of protection of the present application. Therefore, the protection scope of the present patent shall be subject to the appended claims.

Claims (11)

Translated fromChinese
1.一种道路边界检测方法,所述方法包括:1. A road boundary detection method, the method comprising:获取探测设备针对目标道路采集的点云数据,所述点云数据包括各点云的点云位置和反射强度;acquiring point cloud data collected by the detection device for the target road, where the point cloud data includes the point cloud position and reflection intensity of each point cloud;将所述点云数据投影至二维图像中,获得第一图像,所述第一图像中包含各所述点云对应的图像点,各所述点云对应的图像点的图像灰度值根据该点云的所述点云位置和所述反射强度确定;Project the point cloud data into a two-dimensional image to obtain a first image, the first image includes image points corresponding to each of the point clouds, and the image gray value of the image points corresponding to each of the point clouds is based on The point cloud position and the reflection intensity of the point cloud are determined;根据各所述图像灰度值,确定所述第一图像的目标分割阈值;determining the target segmentation threshold of the first image according to the grayscale values of each of the images;根据所述目标分割阈值,对所述第一图像进行分割处理,获得第二图像,基于所述第二图像确定道路边界。According to the target segmentation threshold, the first image is segmented to obtain a second image, and a road boundary is determined based on the second image.2.根据权利要求1所述的方法,其特征在于,所述点云位置包括点云高度,各所述点云对应的图像点的图像灰度值的确定方式包括:2. The method according to claim 1, wherein the point cloud position comprises a point cloud height, and the method for determining the image gray value of the image point corresponding to each of the point clouds comprises:根据各所述点云的所述点云高度及其对应的第一转换因子和第一分配权值、以及各所述点云的所述反射强度及其对应的第二转换因子和第二分配权值,确定各所述点云对应的图像点的图像灰度值。According to the point cloud height of each of the point clouds and its corresponding first conversion factor and first distribution weight, as well as the reflection intensity of each of the point clouds and its corresponding second conversion factor and second distribution The weight value is used to determine the image gray value of the image point corresponding to each of the point clouds.3.根据权利要求2所述的方法,其特征在于,根据各所述点云的所述点云高度及其对应的第一转换因子和第一分配权值、以及各所述点云的所述反射强度及其对应的第二转换因子和第二分配权值,确定各所述点云对应的图像点的图像灰度值,包括:3. The method according to claim 2, characterized in that, according to the height of the point cloud of each of the point clouds, the corresponding first conversion factor and the first distribution weight, and the height of each of the point clouds. The reflection intensity and its corresponding second conversion factor and second distribution weight are determined, and the image gray value of the image point corresponding to each of the point clouds is determined, including:将各所述点云的所述点云高度和所述反射强度,分别与所述第一转换因子和所述第二转换因子相乘,获得各所述点云的第一转换灰度值和第二转换灰度值;Multiply the point cloud height and the reflection intensity of each of the point clouds by the first conversion factor and the second conversion factor, respectively, to obtain the first converted gray value of each of the point clouds and the second converted gray value;将各所述点云的所述第一转换灰度值和所述第二转换灰度值,分别与所述第一分配权值和所述第二分配权值相乘,获得各所述点云的第一加权灰度值和第二加权灰度值;Multiplying the first converted gray value and the second converted gray value of each of the point clouds by the first distribution weight and the second distribution weight, respectively, to obtain each of the points a first weighted gray value and a second weighted gray value of the cloud;将各所述点云的所述第一加权灰度值与所述第二加权灰度值相加,确定各所述点云对应的图像点的图像灰度值。The first weighted gray value and the second weighted gray value of each of the point clouds are added to determine the image gray value of the image point corresponding to each of the point clouds.4.根据权利要求1所述的方法,其特征在于,在根据各所述图像灰度值,确定所述第一图像的目标分割阈值之前,还包括:根据所述第一图像中各图像点的位置,对所述第一图像进行分块处理,获得至少两个分块图像;4 . The method according to claim 1 , wherein before determining the target segmentation threshold of the first image according to the gray value of each image, the method further comprises: according to each image point in the first image. 5 . position, perform block processing on the first image to obtain at least two block images;根据各所述图像灰度值,确定所述第一图像的目标分割阈值,包括:对于每个所述分块图像,根据所述分块图像中各图像点的图像灰度值,确定所述分块图像的分块分割阈值;所述目标分割阈值包括各所述分块分割阈值;Determining the target segmentation threshold of the first image according to each of the image grayscale values includes: for each segmented image, determining the The segmentation threshold of the segmented image; the target segmentation threshold includes each of the segmentation thresholds;根据所述目标分割阈值,对所述第一图像进行分割处理,获得第二图像,包括:根据各所述分块图像的分块分割阈值,分别对对应的所述分块图像进行分割处理,获得第二图像。Performing segmentation processing on the first image according to the target segmentation threshold to obtain a second image includes: performing segmentation processing on the corresponding segmented images respectively according to the segmented segmentation thresholds of the segmented images, Obtain a second image.5.根据权利要求1所述的方法,其特征在于,根据各所述图像灰度值,确定所述第一图像的目标分割阈值,包括:5. The method according to claim 1, wherein determining the target segmentation threshold of the first image according to the grayscale values of each of the images, comprising:根据各所述图像灰度值及其对应的图像点数量,获得灰度值分布图,所述灰度值分布图的第一坐标表示图像灰度值,第二坐标表示图像点数量;According to each of the image gray values and the corresponding number of image points, a gray value distribution map is obtained, wherein the first coordinate of the gray value distribution map represents the image gray value, and the second coordinate represents the number of image points;检测所述灰度值分布图中在第二坐标方向上的波峰,确定各所述波峰中第二坐标值最大的第一波峰和第二波峰;Detecting the peaks in the second coordinate direction in the gray value distribution map, and determining the first peak and the second peak with the largest second coordinate value in each of the peaks;基于所述第一波峰和所述第二波峰之间是否存在其它波峰,选用对应的分割阈值确定方式,确定所述第一图像的目标分割阈值。Based on whether other peaks exist between the first peak and the second peak, a corresponding segmentation threshold determination method is selected to determine the target segmentation threshold of the first image.6.根据权利要求5所述的方法,其特征在于,基于所述第一波峰和所述第二波峰之间是否存在其它波峰,选用对应的分割阈值确定方式,确定所述第一图像的目标分割阈值,包括:6 . The method according to claim 5 , wherein, based on whether there are other peaks between the first peak and the second peak, a corresponding segmentation threshold determination method is selected to determine the target of the first image. 7 . segmentation thresholds, including:当所述第一波峰和所述第二波峰之间不存在其它波峰时,根据各所述图像灰度值确定最大类间方差值,作为所述第一图像的目标分割阈值。When there is no other peak between the first peak and the second peak, the maximum inter-class variance value is determined according to the gray value of each image as the target segmentation threshold of the first image.7.根据权利要求5所述的方法,其特征在于,基于所述第一波峰和所述第二波峰之间是否存在其它波峰,选用对应的分割阈值确定方式,确定所述第一图像的目标分割阈值,包括:7 . The method according to claim 5 , wherein, based on whether there are other peaks between the first peak and the second peak, a corresponding segmentation threshold determination method is selected to determine the target of the first image. 8 . segmentation thresholds, including:当所述第一波峰和所述第二波峰之间存在至少一个其它波峰时,从所述至少一个其它波峰中,选取与所述第一波峰相邻的波峰作为第三波峰;所述第一波峰的第二坐标值大于或等于所述第二波峰的第二坐标值;When there is at least one other peak between the first peak and the second peak, from the at least one other peak, a peak adjacent to the first peak is selected as a third peak; the first The second coordinate value of the wave crest is greater than or equal to the second coordinate value of the second wave crest;根据所述第一波峰与所述第三波峰之间的最小第二坐标值对应的图像灰度值,确定所述第一图像的目标分割阈值。The target segmentation threshold of the first image is determined according to the image gray value corresponding to the smallest second coordinate value between the first wave peak and the third wave peak.8.根据权利要求1至7任意一项所述的方法,其特征在于,基于所述第二图像确定道路边界,包括:8. The method according to any one of claims 1 to 7, wherein determining a road boundary based on the second image comprises:从所述第二图像中提取目标侧的边界轮廓图像点;extracting boundary contour image points on the target side from the second image;当基于所述边界轮廓图像点形成的边界轮廓不连续时,获取所述边界轮廓在各断开处的两个端点的位置,根据各所述端点的位置以及各所述端点处的切线,确定所述边界轮廓在各断开处的连接曲线;When the boundary contour formed based on the boundary contour image points is discontinuous, obtain the positions of the two end points of the boundary contour at each disconnection, and determine according to the position of each end point and the tangent at each end point. the connecting curve of the boundary contour at each disconnection;根据各所述连接曲线,在所述边界轮廓的各断开处的两个端点之间进行插值,获得插值后的边界轮廓图像点;According to each of the connection curves, interpolation is performed between the two endpoints of each disconnection of the boundary contour to obtain an interpolated boundary contour image point;对所述插值后的边界轮廓图像点进行曲线拟合,获得道路边界曲线,基于所述道路边界曲线确定道路边界。Curve fitting is performed on the interpolated boundary contour image points to obtain a road boundary curve, and a road boundary is determined based on the road boundary curve.9.一种道路边界检测装置,其特征在于,所述装置包括:9. A road boundary detection device, wherein the device comprises:获取模块,用于获取探测设备针对目标道路采集的点云数据,所述点云数据包括各点云的点云位置和反射强度;an acquisition module, configured to acquire point cloud data collected by the detection device for the target road, where the point cloud data includes the point cloud position and reflection intensity of each point cloud;投影模块,用于将所述点云数据投影至二维图像中,获得第一图像,所述第一图像中包含各所述点云对应的图像点,各所述点云对应的图像点的图像灰度值根据该点云的所述点云位置和所述反射强度确定;The projection module is used for projecting the point cloud data into a two-dimensional image to obtain a first image, wherein the first image includes image points corresponding to each of the point clouds, and the image points corresponding to each of the point clouds are The image gray value is determined according to the point cloud position and the reflection intensity of the point cloud;确定模块,用于根据各所述图像灰度值,确定所述第一图像的目标分割阈值;a determination module, configured to determine the target segmentation threshold of the first image according to the grayscale values of each of the images;处理模块,用于根据所述目标分割阈值,对所述第一图像进行分割处理,获得第二图像,基于所述第二图像确定道路边界。The processing module is configured to perform segmentation processing on the first image according to the target segmentation threshold to obtain a second image, and determine a road boundary based on the second image.10.一种计算机设备,包括存储器和处理器,所述存储器存储有计算机程序,其特征在于,所述处理器执行所述计算机程序时实现权利要求1至8中任一项所述方法的步骤。10. A computer device, comprising a memory and a processor, wherein the memory stores a computer program, wherein the processor implements the steps of the method according to any one of claims 1 to 8 when the processor executes the computer program .11.一种计算机可读存储介质,其上存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现权利要求1至8中任一项所述的方法的步骤。11. A computer-readable storage medium on which a computer program is stored, wherein when the computer program is executed by a processor, the steps of the method according to any one of claims 1 to 8 are implemented.
CN202010400797.XA2020-05-132020-05-13Road boundary detection method, road boundary detection device, computer equipment and storage mediumPendingCN113673274A (en)

Priority Applications (2)

Application NumberPriority DateFiling DateTitle
CN202010400797.XACN113673274A (en)2020-05-132020-05-13Road boundary detection method, road boundary detection device, computer equipment and storage medium
PCT/CN2021/088583WO2021227797A1 (en)2020-05-132021-04-21Road boundary detection method and apparatus, computer device and storage medium

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202010400797.XACN113673274A (en)2020-05-132020-05-13Road boundary detection method, road boundary detection device, computer equipment and storage medium

Publications (1)

Publication NumberPublication Date
CN113673274Atrue CN113673274A (en)2021-11-19

Family

ID=78526357

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202010400797.XAPendingCN113673274A (en)2020-05-132020-05-13Road boundary detection method, road boundary detection device, computer equipment and storage medium

Country Status (2)

CountryLink
CN (1)CN113673274A (en)
WO (1)WO2021227797A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN114155258A (en)*2021-12-012022-03-08苏州思卡信息系统有限公司Detection method for highway construction enclosed area
CN114581464A (en)*2022-03-072022-06-03广州极飞科技股份有限公司 Boundary detection method, apparatus, electronic device, and computer-readable storage medium
CN114723900A (en)*2022-02-222022-07-08高德软件有限公司Contour drawing method of ground element, electronic device and computer program product
CN114821501A (en)*2022-04-292022-07-29北京智行者科技有限公司 A road boundary classification method and device, computer equipment and mobile tool
CN114880334A (en)*2022-03-162022-08-09高德软件有限公司 Map data update method and electronic device
CN114926751A (en)*2022-04-082022-08-19广州市华南自然资源科学技术研究院Farmland road extraction threshold value selection method and device based on data mining

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN114519782B (en)*2021-12-272025-09-26中智行(苏州)科技有限公司 A roadside extraction method and device
CN115436936A (en)*2022-08-312022-12-06南京慧尔视智能科技有限公司 Radar map determination method, device, equipment and medium for target detection
CN115131761B (en)*2022-08-312022-12-06北京百度网讯科技有限公司 Recognition method, drawing method and device of road boundary
CN117368879B (en)*2023-12-042024-03-19北京海兰信数据科技股份有限公司Radar diagram generation method and device, terminal equipment and readable storage medium
CN117764992B (en)*2024-02-222024-04-30山东乔泰管业科技有限公司Plastic pipe quality detection method based on image processing
CN119359964B (en)*2024-12-232025-09-16思看科技(杭州)股份有限公司Three-dimensional reconstruction method and laser scanner

Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN105741298A (en)*2016-02-032016-07-06浙江工商大学Image segmenting method for defect detection of thin film labels
WO2020083024A1 (en)*2018-10-242020-04-30腾讯科技(深圳)有限公司Obstacle identification method and device, storage medium, and electronic device

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN106067003B (en)*2016-05-272020-05-19山东科技大学 A method for automatic extraction of road vector marking lines in vehicle-mounted laser scanning point cloud
CN110163047B (en)*2018-07-052023-04-07腾讯大地通途(北京)科技有限公司Method and device for detecting lane line
CN109034047B (en)*2018-07-202021-01-22京东方科技集团股份有限公司Lane line detection method and device
CN109766878B (en)*2019-04-112019-06-28深兰人工智能芯片研究院(江苏)有限公司A kind of method and apparatus of lane detection
CN110502973B (en)*2019-07-052023-02-07同济大学Automatic extraction and identification method for road marking based on vehicle-mounted laser point cloud
CN110866449A (en)*2019-10-212020-03-06北京京东尚科信息技术有限公司Method and device for identifying target object in road

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN105741298A (en)*2016-02-032016-07-06浙江工商大学Image segmenting method for defect detection of thin film labels
WO2020083024A1 (en)*2018-10-242020-04-30腾讯科技(深圳)有限公司Obstacle identification method and device, storage medium, and electronic device

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
张达;李霖;李游;: "基于车载激光扫描的城市道路提取方法", 测绘通报, no. 07, 25 July 2016 (2016-07-25), pages 1 - 3*
王淑燕;陈晓勇;余广旺;: "基于点云片段法提取道路边界线", 东华理工大学学报(自然科学版), no. 01, 31 March 2017 (2017-03-31)*
陈兆学;姚春荣;: "基于OTSU多阈值分割算法的激光线扫点云数据表达及精简方法研究", 光学技术, no. 05, 15 September 2016 (2016-09-15)*

Cited By (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN114155258A (en)*2021-12-012022-03-08苏州思卡信息系统有限公司Detection method for highway construction enclosed area
CN114723900A (en)*2022-02-222022-07-08高德软件有限公司Contour drawing method of ground element, electronic device and computer program product
CN114581464A (en)*2022-03-072022-06-03广州极飞科技股份有限公司 Boundary detection method, apparatus, electronic device, and computer-readable storage medium
CN114880334A (en)*2022-03-162022-08-09高德软件有限公司 Map data update method and electronic device
CN114926751A (en)*2022-04-082022-08-19广州市华南自然资源科学技术研究院Farmland road extraction threshold value selection method and device based on data mining
CN114821501A (en)*2022-04-292022-07-29北京智行者科技有限公司 A road boundary classification method and device, computer equipment and mobile tool

Also Published As

Publication numberPublication date
WO2021227797A1 (en)2021-11-18

Similar Documents

PublicationPublication DateTitle
CN113673274A (en)Road boundary detection method, road boundary detection device, computer equipment and storage medium
Wang et al.Computational methods of acquisition and processing of 3D point cloud data for construction applications
US20230111722A1 (en)Curb detection by analysis of reflection images
CN107330925B (en) A Multiple Obstacle Detection and Tracking Method Based on LiDAR Depth Image
Labayrade et al.In-vehicle obstacles detection and characterization by stereovision
CN109635816B (en)Lane line generation method, apparatus, device, and storage medium
US20050015201A1 (en)Method and apparatus for detecting obstacles
CN111104933A (en)Map processing method, mobile robot, and computer-readable storage medium
Wedel et al.Realtime depth estimation and obstacle detection from monocular video
CN111491845B (en) Method for radar classification of road surfaces
CN104950313A (en)Road-surface abstraction and road gradient recognition method
WO2020080088A1 (en)Information processing device
Wang et al.Automated pavement distress survey: a review and a new direction
CN113379776A (en)Road boundary detection method
Konrad et al.Localization in digital maps for road course estimation using grid maps
CN114740867B (en)Intelligent obstacle avoidance method and device based on binocular vision, robot and medium
CN113240735A (en)Slope displacement activity monitoring method
CN114842166A (en) Negative obstacle detection method, system, medium and device applied to structured road
Janda et al.Road boundary detection for run-off road prevention based on the fusion of video and radar
CN114663855A (en) A road surface water and roughness detection method for unstructured roads
CN118898825A (en) Road environment state perception method, equipment, medium, program product and vehicle
CN118068366B (en)Intelligent wheelchair self-adaptive travelable area detection method, system and equipment
Oniga et al.Curb detection based on elevation maps from dense stereo
WO2020135325A1 (en)Mobile device positioning method, device and system, and mobile device
Mandici et al.Aggregate road surface based environment representation using digital elevation maps

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
CB02Change of applicant information
CB02Change of applicant information

Country or region after:China

Address after:Building A3 and A4, Hunan Inspection and testing characteristic industrial park, No. 336, bachelor Road, Yuelu District, Changsha City, Hunan Province

Applicant after:Xidi Intelligent Driving Technology Co.,Ltd.

Address before:Building A3 and A4, Hunan Inspection and testing characteristic industrial park, No. 336, bachelor Road, Yuelu District, Changsha City, Hunan Province

Applicant before:CHANGSHA INTELLIGENT DRIVING RESEARCH INSTITUTE Co.,Ltd.

Country or region before:China


[8]ページ先頭

©2009-2025 Movatter.jp