Movatterモバイル変換


[0]ホーム

URL:


US11440537B2 - Apparatus and method for estimating position in automated valet parking system - Google Patents

Apparatus and method for estimating position in automated valet parking system
Download PDF

Info

Publication number
US11440537B2
US11440537B2US16/810,669US202016810669AUS11440537B2US 11440537 B2US11440537 B2US 11440537B2US 202016810669 AUS202016810669 AUS 202016810669AUS 11440537 B2US11440537 B2US 11440537B2
Authority
US
United States
Prior art keywords
vehicle
map
lane
controller
processor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US16/810,669
Other versions
US20200298836A1 (en
Inventor
Dong Wook Kim
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hyundai Mobis Co Ltd
Original Assignee
Hyundai Mobis Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hyundai Mobis Co LtdfiledCriticalHyundai Mobis Co Ltd
Assigned to HYUNDAI MOBIS CO., LTD.reassignmentHYUNDAI MOBIS CO., LTD.ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: KIM, DONG WOOK
Publication of US20200298836A1publicationCriticalpatent/US20200298836A1/en
Priority to US17/881,854priorityCriticalpatent/US12326337B2/en
Application grantedgrantedCritical
Publication of US11440537B2publicationCriticalpatent/US11440537B2/en
Activelegal-statusCriticalCurrent
Adjusted expirationlegal-statusCritical

Links

Images

Classifications

Definitions

Landscapes

Abstract

An apparatus for estimating a position in an automated valet parking system includes a front camera processor processing a front image of a vehicle, a surround view monitor (SVM) processor recognizing a short-distance lane and stop line by processing a surround view image of the vehicle, a map data unit storing a high definition map, and a controller downloading a map including an area set as a parking zone from the map data unit when the entry of the vehicle to a parking lot is identified and correcting a position measurement value of the vehicle by performing map matching based on results of the recognition and processing of the front camera processor and SVM processor and the parking lot map of the map data unit when an automated valet parking start position is recognized based on the recognized short-distance lane and stop line.

Description

CROSS-REFERENCE TO RELATED APPLICATION
The present application claims priority to Korean Patent Application No. 10-2019-0031092 filed on Mar. 19, 2019 in the Korean Intellectual Property Office, which is incorporated herein by reference in its entirety.
BACKGROUND1. Technical Field
Embodiments of the present disclosure relate to an apparatus and method for estimating a position in an automated valet parking system, and more particularly, to an apparatus and method for estimating a position in an automated valet parking system, which can estimate an initial position in an automatic valet parking (AVP) system using a surround view monitor (SVM).
2. Related Art
In general, an autonomous vehicle refers to a vehicle which autonomously determines a driving path by recognizing a surrounding environment using a function for detecting and processing external information upon driving and independently travels using its own power.
Positioning methods applied to autonomous vehicles include a satellite positioning method based on a global navigation satellite system (GNSS), such as a global positioning system (GPS), a differential GPS (DGPS) or network-real time kinematic (RTK), vehicle behavior-based dead reckoning for correcting satellite positioning using vehicle sensors and an inertial measurement unit (IMU) (e.g., a vehicle speed, a steering angle, and a wheel odometer/yaw rate/acceleration), and a map-matching method of relatively estimating the position of a vehicle by comparing a precise map for autonomous driving with data from various sensors (e.g., a camera, a stereo camera, an SVM camera, and a radar).
Recently, automated valet parking (AVP) has been developed for more convenient parking. An autonomous vehicle on which an AVP system has been mounted can autonomously travel without a driver, search for a parking space, and perform parking or exit from a parking lot. Furthermore, even a function for performing parking by extending a target parking space to a surrounding parking lot in a traffic congestion area has been developed.
Accordingly, a positioning method for estimating a position becomes important. However, a conventional satellite positioning method has problems in that the method is very expensive because it requires a high definition GPS, a high definition radar, and a high resolution camera, that the method has low processing speed and accuracy because it is configured with a complicated algorithm, and that the method cannot constantly maintain its performance because it is influenced by characteristics of a road and characteristics of surrounding geographic features.
The related art of the disclosure is disclosed in U.S. Patent Application Publication No. 2018-0023961 (Jan. 25, 2018) entitled “SYSTEMS AND METHODS FOR ALIGNING CROWDSOURCED SPARSE MAP DATA.”
SUMMARY
Various embodiments are directed to the provision of an apparatus and method for estimating a position in an automated valet parking system, which can estimate an initial position of an automatic valet parking (AVP) system without expensive equipment using a surround view monitor (SVM).
In an embodiment, an apparatus for estimating a position in an automated valet parking system includes a front camera processor configured to process a front image of a vehicle, a surround view monitor (SVM) processor configured to recognize a short-distance lane and a stop line by processing a surround view image of the vehicle, a map data unit configured to store a high definition map, and a controller configured to download a map including an area set as a parking zone from the map data unit when the entry of the vehicle to a parking lot is identified and to correct a position measurement value of the vehicle by performing map matching based on the results of the recognition and processing of the front camera processor and the SVM processor and the parking lot map of the map data unit when an automated valet parking (AVP) start position is recognized based on the short-distance lane and stop line recognized by the SVM processor.
In an embodiment, the controller is configured to predict a behavior of the vehicle through dead reckoning when the AVP start position is recognized and to estimate an AVP initial position of the vehicle by fusing the position measurement value of the vehicle corrected through the map matching and the predicted behavior of the vehicle.
In an embodiment, the controller includes a vehicle behavior prediction unit configured to predict a behavior of the vehicle through dead reckoning based on GPS information received from a GPS receiver and a vehicle steering wheel angle, yaw rate and wheel speed received from a vehicle sensor unit.
In an embodiment, the controller includes a map-matching unit configured to perform the map matching based on at least one of lane fusion data in which a long-distance lane recognized by the front camera processor and the short-distance lane and stop line recognized by the SVM processor have been fused, parking lot map data from the map data unit, and vehicle behavior data for each time predicted through dead reckoning.
In an embodiment, the map-matching unit is configured to compute a position and rotation correction quantity in which a distance error between sensor data and map data is minimized using iterative closest point (ICP) logic.
In an embodiment, the controller includes a position fusion unit configured to fuse a vehicle pose output as the results of the map matching and GPS information of a vehicle position predicted through dead reckoning.
In an embodiment, the controller includes a fail-safe diagnosis unit configured to receive the vehicle position and flags output by the position fusion unit and to perform fail-safe diagnosis. The fail-safe diagnosis unit is configured to perform the fail-safe diagnosis using a distribution chart configured with estimated positioning results in which positioning results at past timing have been projected on current timing and positioning results input at current timing.
In an embodiment, the vehicle pose includes one or more of longitude, latitude, heading, covariance, warning/fail/safe, flags and a lane offset.
In an embodiment, a method of estimating a position in an automated valet parking system includes downloading, by a controller, a map including an area set as a parking zone from a map data unit for storing a high definition map when the entry of a vehicle to a parking lot is identified, recognizing, by the controller, an automated valet parking (AVP) start position based on a short-distance lane and stop line recognized by a surround view monitor (SVM) processor, and correcting, by the controller, a position measurement value of the vehicle by performing map matching based on the results of the recognition and processing of a front camera processor and the SVM processor and the parking lot map of the map data unit.
In an embodiment, the method further includes predicting, by the controller, a behavior of the vehicle through dead reckoning when the AVP start position is recognized and estimating, by the controller, an AVP initial position of the vehicle by fusing the position measurement value of the vehicle corrected through the map matching and the predicted behavior of the vehicle.
In an embodiment, in the predicting of the behavior of the vehicle, the controller predicts the behavior of the vehicle through dead reckoning based on GPS information received from a GPS receiver and a vehicle steering wheel angle, yaw rate and wheel speed received from a vehicle sensor unit.
In an embodiment, in the correcting of the position measurement value, the controller performs the map matching based on at least one of lane fusion data in which a long-distance lane recognized by the front camera processor and the short-distance lane and stop line recognized by the SVM processor have been fused, parking lot map data from the map data unit, and vehicle behavior data for each time predicted through dead reckoning.
In an embodiment, in the correcting of the position measurement value, the controller computes a position and rotation correction quantity in which a distance error between sensor data and map data is minimized using iterative closest point (ICP) logic.
In an embodiment, in the estimating of the AVP initial position, the controller fuses a vehicle pose output as results of the map matching and GPS information of a vehicle position predicted through dead reckoning.
In an embodiment, the method further includes receiving, by the controller, the vehicle position and flags output as the result of the position fusion and performing fail-safe diagnosis. In the performing of the fail-safe diagnosis, the controller performs the fail-safe diagnosis using a distribution chart configured with estimated positioning results in which positioning results at past timing have been projected on current timing and positioning results input at current timing.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram illustrating an apparatus for estimating a position in an automated valet parking system according to an embodiment of the disclosure.
FIG. 2 is a diagram more specifically describing the apparatus for estimating a position in an automated valet parking system according to an embodiment of the disclosure.
FIG. 3 is a flowchart for describing a method of estimating a position in an automated valet parking system according to an embodiment of the disclosure.
FIG. 4 is an exemplary diagram of an apparatus and method for estimating a position in an automated valet parking system according to an embodiment of the disclosure.
FIG. 5 is a diagram for describing map-matching logic for the apparatus and method for estimating a position in an automated valet parking system according to an embodiment of the disclosure.
DETAILED DESCRIPTION
Hereinafter, an apparatus and method for estimating a position in an automated valet parking system according to embodiments of the present disclosure are described with reference to the accompanying drawings. For clarity and convenience in this description, thicknesses of lines, sizes of constituent elements, and the like are illustrated in an exaggerated manner in the drawings.
Furthermore, terms to be described hereunder have been defined by taking into consideration functions in the present disclosure, and may be different depending on a user, an operator's intention or practice. Accordingly, each term should be defined based on contents over the entire specification.
Furthermore, an implementation described in this specification may be realized as a method or process, apparatus, software program, data stream or signal, for example. Although the disclosure has been discussed only in the context of a single form of an implementation (e.g., discussed as only a method), an implementation having a discussed characteristic may also be realized in another form (e.g., apparatus or program). The apparatus may be implemented as proper hardware, software or firmware. The method may be implemented in an apparatus, such as a processor commonly referring to a processing device, including a computer, a microprocessor, an integrated circuit or a programmable logic device, for example. The processor includes a communication device, such as a computer, a cell phone, a mobile phone/personal digital assistant (“PDA”) and another device, which facilitates the communication of information between end-users.
FIG. 1 is a block diagram illustrating an apparatus for estimating a position in an automated valet parking system according to an embodiment of the disclosure.FIG. 2 is a diagram more specifically describing the apparatus for estimating a position in an automated valet parking system according to an embodiment of the disclosure.FIG. 5 is a diagram for describing map-matching logic for the apparatus and method for estimating a position in an automated valet parking system according to an embodiment of the disclosure. The apparatus for estimating a position in an automated valet parking system is described below with reference toFIGS. 1, 2 and 5.
As illustrated inFIG. 1, the apparatus for estimating a position in an automated valet parking system according to an embodiment of the disclosure includes afront camera processor10, a surround view monitor (SVM)processor20, amap data unit30, aGPS receiver40, avehicle sensor unit50, acontroller60 and anoutput unit70.
As publicly known in the art, some of exemplary embodiments may be illustrated in the accompanying drawings from the viewpoint of function blocks, units, section, and/or modules. Those skilled in the art will understood that such blocks, units and/or modules are physically implemented by electronic (or optical) circuits such as logic circuits, discrete components, processors, hard wired circuits, memory devices and wiring connections. When the blocks, units and or modules are implemented by processors or other similar hardware, the blocks, units and modules may be programmed and controlled through software (for example, codes) in order to perform various functions discussed in this specification. Furthermore, each of the blocks, units and/or modules may be implemented by dedicated hardware or a combination of dedicated hardware for performing some functions and a processor for performing another function (for example, one or more programmed processors and related circuits). In some exemplary embodiments, each of the blocks, units and/or modules may be physically divided into two or more blocks, units and or modules which are interactive and discrete, without departing from the scope of the disclosure. Furthermore, blocks, units and/or modules in some exemplary embodiments may be physically coupled as a more complex block, unit and/or module without departing from the scope of the disclosure.
First, the present embodiment is for estimating an initial position in an automated valet parking (AVP) system using an SVM, which facilitates parking by allowing surrounding spaces of a vehicle to be seen within the vehicle through cameras attached to the front, back and sides of the vehicle. That is, the present embodiment relates to a vehicle positioning device, and can measure the position of a vehicle using images captured by cameras without expensive equipment, a stop line, etc. and improve the accuracy of map matching.
Thefront camera processor10 may receive a front image of a vehicle from a front camera of the vehicle, and may recognize a long-distance lane and a traffic sign by processing the front image of the vehicle.
Furthermore, thefront camera processor10 may include inside lane recognition means for recognizing an inside lane in the front image, lane tracking means for tracking a lane having the same characteristic as the recognized lane, and reliability computation means.
The inside lane recognition means may recognize a lane having a solid line or dotted-line form with a specific color (e.g., white, or yellow) in a front image.
The lane tracking means may track a lane having the same characteristics within a pre-designated margin although components (e.g., color, thickness, and form) of the recognized lane do not partially maintain the same characteristics (e.g., the same line color, the same line thickness, and the same line interval) by considering a flow (or direction) of the recognized lane.
Furthermore, the reliability computation means may compute the ratio (i.e., lane component matching ratio) in which the components (e.g., color, thickness, and form) of the tracked lane are identical with pre-designated reference values for each component. As the computed lane component matching ratio is closer to 100%, this means high reliability. In contrast, as the computed lane component matching ratio is closer to 0%, this means low reliability. Furthermore, the reliability computation means may predict a current lane (i.e., predicted lane) using the results of the recognition of a previous lane and motion information of a vehicle, and may compute reliability in such a way to compare the predicted lane with a current lane recognized in the front image and increase a reliability count (or reliability score) when a difference between the predicted lane and the current lane is a preset threshold or less. When the reliability count is more than the preset threshold, the reliability computation means may determine that corresponding lane recognition is valid (i.e., the corresponding lane is a valid lane).
TheSVM processor20 may recognize a short-distance lane and a stop line by processing a surround view image of a vehicle.
Furthermore, theSVM processor20 functions to recognize a lane in a surround view image (or surround view composition image). The surround view image means an image obtained by composing, in a top view or surround view form, surrounding images (e.g., front, side, and back images) of a vehicle captured by one or more cameras. Accordingly, theSVM processor20 may recognize a lane (i.e., short-distance lane) in an area close to a vehicle.
In this case, the cameras are disposed on the front, back, left and right sides of the vehicle. In order to increase a degree of completion of the top view or surround view image and prevent a photographing blind spot from occurring, additional cameras may also be disposed on the upper sides at the front and back of the vehicle, that is, relatively higher positions than the positions of the cameras disposed on the front, back, left and right sides.
Furthermore, like thefront camera processor10, theSVM processor20 may include inside lane recognition means, lane tracking means and reliability computation means.
That is, the inside lane recognition means may recognize an inside lane in a surround view image, and may recognize a lane having a solid line or a dotted-line form with a specific color (e.g., white or yellow) in the surround view image. In the present embodiment, in particular, the inside lane recognition means may recognize a stop line.
The lane tracking means may track a lane having the same characteristics within a pre-designated margin although components (e.g., color, thickness, and form) of the recognized lane do not partially maintain the same characteristics (e.g., the same line color, the same line thickness, and the same line interval) by considering a flow (or direction) of the recognized lane.
Furthermore, the reliability computation means may compute the ratio (i.e., lane component matching ratio) in which the components (e.g., color, thickness, and form) of the tracked lane are identical with pre-designated reference values for each component. As the computed lane component matching ratio is closer to 100%, this means high reliability. In contrast, as the computed lane component matching ratio is closer to 0%, this means low reliability. Furthermore, the reliability computation means may predict a current lane (i.e., predicted lane) using the results of the recognition of a previous lane and motion information of a vehicle, and may compute reliability in such a way to compare the predicted lane with a current lane recognized in the surround view image and increase a reliability count (or reliability score) when a difference between the predicted lane and the current lane is a preset threshold or less. When the reliability count is more than the preset threshold, the reliability computation means may determine that corresponding lane recognition is valid (i.e., the corresponding lane is a valid lane).
The short-distance lane may mean a lane in an area, which may be recognized in a surround view image. The long-distance lane may mean a lane in a long-distance area, which may be recognized in a front image.
Themap data unit30 stores a high definition map in which information on roads and surrounding terrains has been constructed with high precision, and provides the high definition map in response to a request from thecontroller60. In the present embodiment, in particular, themap data unit30 may store a high definition (HD) map for a parking lot (i.e., area set as a parking zone).
TheGPS receiver40 receives GPS signals from satellites and provides the GPS signals to thecontroller60 so that the position of a vehicle can be set based on a current position.
Thevehicle sensor unit50 means various sensors within a vehicle. In the present embodiment, in particular, thevehicle sensor unit50 may include a vehicle steering wheel angle sensor, yaw rate sensor and wheel speed sensor for vehicle behavior prediction.
Thecontroller60 identifies that a vehicle enters an area set as a parking lot or parking zone, and downloads a map for the corresponding area. That is, when the entry of a vehicle to a parking lot is identified, thecontroller60 may download, from themap data unit30, a map including an area set as the parking zone.
Furthermore, thecontroller60 may recognize an AVP start position based on a short-distance lane and stop line recognized by theSVM processor20.
At this time, thecontroller60 may generate a single fused lane (i.e., a single lane not divided into a short-distance lane and a long-distance lane) by fusing a lane recognized by theSVM processor20 and a lane recognized by thefront camera processor10.
That is, thecontroller60 may fuse lanes through a lane error comparison, may determine a valid lane, and may generate a fused lane.
Thecontroller60 computes (or determines) a position error (e.g., an interval between the ends of lanes in a vehicle reference coordinate system and an angle of each lane) by comparing the lane (i.e., short-distance lane) recognized by theSVM processor20 and the lane (i.e., long-distance lane) recognized by thefront camera processor10. In this case, the vehicle reference coordinate system means a coordinate system indicative of a traverse coordinates X, longitudinal coordinates Y and vehicle movement direction θ corresponding to a moved distance and direction of a vehicle with respect to the center of the vehicle.
When the position error is within a preset permission range as a result of the comparison between the two lanes (i.e., the long-distance lane and the short-distance lane), thecontroller60 generates a single fused lane (i.e., a single lane not divided into the short-distance lane and the long-distance lane) by fusing the two lanes (i.e., the long-distance lane and the short-distance lane). Furthermore, when the position error is out of the preset permission range as a result of the comparison between the two lanes (i.e., the long-distance lane and the short-distance lane), thecontroller60 does not fuse the two lanes (i.e., the long-distance lane and the short-distance lane) and determines a lane having relatively high reliability as a valid lane.
Accordingly, thecontroller60 may determine that the two lanes (i.e., the long-distance lane and the short-distance lane) are not valid lanes when the reliability of each lane is smaller than the preset threshold. Thecontroller60 may generate a single fused lane (i.e., a single lane not divided into the short-distance lane and the long-distance lane) by fusing the two lanes when the reliability of each of the two lanes is the preset threshold or more and the position error between the two lanes is within the preset permission range. Thecontroller60 may determine a lane having relatively higher reliability, among the two lanes, as a valid lane when the position error between the two lanes is out of the preset permission range.
Furthermore, thecontroller60 includes a vehiclebehavior prediction unit62, a map-matchingunit64, aposition fusion unit66 and a fail-safe diagnosis unit68. Thecontroller60 may predict a behavior of a vehicle through dead reckoning, may correct a position measurement value of the vehicle through map matching based on the results of the recognition and processing of thefront camera processor10 and theSVM processor20 and the parking lot map of themap data unit30, and may finally estimate an AVP initial position of the vehicle by fusing the predicted behavior of the vehicle and the corrected position measurement value of the vehicle.
The vehiclebehavior prediction unit62 may predict a behavior of a vehicle through dead reckoning based on GPS information received from theGPS receiver40 and a vehicle steering wheel angle, yaw rate and wheel speed received from thevehicle sensor unit50.
Furthermore, the map-matchingunit64 may perform map matching based on at least one of lane fusion data in which a long-distance lane recognized by thefront camera processor10 and a short-distance lane and stop line recognized by theSVM processor20 have been fused, parking lot map data from themap data unit30, and vehicle behavior data for each time predicted through dead reckoning.
At this time, the map-matchingunit64 may compute a position and rotation correction quantity in which a distance error between sensor data and map data is minimized using iterative closest point (ICP) logic. The ICP logic is a method of registering current data with the existing data set, and is a method of finding an association based on the closest points of data, moving and rotating current data based on the association, and adding the current data to the existing data set.
For example, a position T and a rotation (R) correction quantity may be computed with reference to the following equation andFIG. 5.
minmizeE=i=1Nei=i=1Nwi((R·li+T-mpair,i)·ηpair,i)2
Furthermore, theposition fusion unit66 may fuse a vehicle pose, output as the results of the map matching, and the GPS information of the vehicle position predicted through dead reckoning.
In this case, theposition fusion unit66 may be implemented like the method of fusing lanes recognized by theSVM processor20 and thefront camera processor10, but may fuse position measurement values using another method.
Thecontroller60 includes the fail-safe diagnosis unit68 for receiving a vehicle position and flags output by theposition fusion unit66 and performing fail-safe diagnosis. The fail-safe diagnosis unit68 may perform fail-safe diagnosis using a distribution chart configured with estimated positioning results in which positioning results at the past timing have been projected on current timing and positioning results input at current timing.
In the present embodiment, the vehicle pose may include one or more of longitude, latitude, heading, covariance, warning/fail/safe, flags and a lane offset.
That is, in the present embodiment, theoutput unit70 may output the results of the diagnosis of the fail-safe diagnosis unit68. In this case, theoutput unit70 may output the results of fail-safe diagnosis of vehicle pose information.
In the present embodiment, an autonomous driving system can perform sensor fusion positioning based on map matching, and can perform fail-safe diagnosis for improving the reliability of the system and enabling the computation (or calculation or estimation) of robust and stable positioning information in a process of performing sensor fusion positioning. Furthermore, in the present embodiment, the fail-safe diagnosis does not require additional hardware because it is analytic redundancy-based fault diagnosis, but the disclosure is not limited thereto.
Referring toFIG. 2, the present embodiment may basically include a performance core for performing a process of fusing a position measurement value, corrected through map matching, and the predicted results of a vehicle behavior and a safety core for performing fail-safe diagnosis on a vehicle position fused in the performance core.
In the performance core, thefront camera processor10 and theSVM processor20 may perform sensor value processing, and themap data unit30 may download and manage a map. Furthermore, in the safety core, theGPS receiver40 may perform GPS signal processing.
Furthermore, in thecontroller60, the map-matchingunit64 and theposition fusion unit66 may be included in the performance core, and the vehiclebehavior prediction unit62 and the fail-safe diagnosis unit68 may be included in the safety core, but the disclosure is not limited thereto.
In other words, the performance core receives lanes and a stop line from an SVM, and receives lanes and a traffic sign from a front camera. Furthermore, the performance core may process recognition data, that is, sensor values received from the SVM and the front camera. In other words, thecontroller60 may fuse recognition data from the SVM and the front camera, may download the HD map of a parking zone, and may perform map matching. In this case, thecontroller60 may perform map matching using GPS signals, vehicle trajectory information predicted through dead reckoning, and GPS information. Furthermore, thecontroller60 may fuse a vehicle pose (or position) corrected through the map matching and a position value based on the GPS information, and may finally estimate an initial position in the automated valet parking system by performing fail-safe diagnosis on the fused results.
FIG. 3 is a flowchart for describing a method of estimating a position in an automated valet parking system according to an embodiment of the disclosure.FIG. 4 is an exemplary diagram of an apparatus and method for estimating a position in an automated valet parking system according to an embodiment of the disclosure. The method of estimating a position in an automated valet parking system is described below with reference toFIGS. 3 and 4.
As illustrated inFIG. 3, in the method of estimating a position in an automated valet parking system according to an embodiment of the disclosure, first, thecontroller60 identifies that a vehicle enters a parking lot (S10).
In this case, thecontroller60 may identify that the vehicle enters the parking lot by receiving a vehicle position from theGPS receiver40, but the disclose is not limited to such a method.
Furthermore, when the entry of the vehicle to the parking lot is identified, thecontroller60 downloads a map including an area set as the parking zone from themap data unit30 in which a high definition (HD) map is stored (S20).
Furthermore, if it is identified that the vehicle has parked at an AVP start position (S30), thecontroller60 determines whether an AVP start area has been recognized through the SVM processor20 (S40).
That is, when the vehicle parks at an AVP start position as illustrated inFIG. 4(a), thecontroller60 may recognize the AVP start position based on a short-distance lane and stop line recognized by theSVM processor20, as illustrated inFIG. 4(b).
If an AVP start area is not recognized through theSVM processor20, thecontroller60 may return to step S30 and perform AVP start position parking. The AVP start position parking may be performed by a user. Thecontroller60 may identify that the vehicle has parked at the AVP start position by receiving GPS information or a signal, indicating that the vehicle has parked, from AVP infrastructure installed in the parking lot.
Furthermore, thecontroller60 sets a position as an initial value of the AVP start position (S50).
That is, as illustrated inFIG. 4(c), thecontroller60 may set the position as the initial value of the AVP start position based on the AVP start area recognized through theSVM processor20.
Furthermore, as illustrated inFIG. 4(d), thecontroller60 corrects the AVP start position (S60).
At this time, thecontroller60 may predict a behavior of the vehicle through dead reckoning, and may correct a position measurement value of the vehicle through map matching based on the results of the recognition and processing of thefront camera processor10 and theSVM processor20 for recognizing a long-distance lane and a traffic sign and the parking lot map of themap data unit30.
Furthermore, thecontroller60 may finally estimate an AVP initial position of the vehicle by fusing the corrected position measurement value of the vehicle and the predicted behavior of the vehicle.
In this case, thecontroller60 may predict the behavior of the vehicle through dead reckoning based on the GPS information received from theGPS receiver40 and a vehicle steering wheel angle, yaw rate and wheel speed received from thevehicle sensor unit50. Furthermore, thecontroller60 may perform map matching based on at least one of lane fusion data in which a long-distance lane recognized by thefront camera processor10 and a short-distance lane and stop line recognized by theSVM processor20 have been fused, parking lot map data from themap data unit30, and vehicle behavior data for each time predicted through dead reckoning.
In the present embodiment, thecontroller60 may compute a position and rotation correction quantity in which a distance error between sensor data and map data is minimized using iterative closest point (ICP) logic.
Furthermore, thecontroller60 may fuse a vehicle pose, output as the results of the map matching, and the GPS information of the vehicle position predicted through dead reckoning.
Finally, thecontroller60 may receive a vehicle position and flags output as the results of the position fusion, and may perform fail-safe diagnosis (S70).
In this case, thecontroller60 may perform the fail-safe diagnosis using a distribution chart configured with estimated positioning results in which positioning results at the past timing have been projected on current timing and positioning results input at current timing. In this case, the vehicle pose may include one or more of longitude, latitude, heading, covariance, warning/fail/safe, flags and a lane offset.
As described above, the apparatus and method for estimating a position in an automated valet parking system according to embodiments of the disclosure can perform map matching without expensive equipment and estimate an initial position regardless of the inside and outside by estimating the initial position in an automated valet parking (AVP) system using a surround view monitor (SVM) and can improve map matching accuracy through increased cognition distance accuracy (i.e., correction accuracy) by performing measurement in the proximity of geographic features.
Although preferred embodiments of the present disclosure have been disclosed for illustrative purposes, those skilled in the art will appreciate that various modifications, additions and substitutions are possible, without departing from the scope of the present disclosure as defined in the accompanying claims.
Thus, the true technical scope of the present disclosure should be defined by the following claims.

Claims (11)

What is claimed is:
1. An apparatus for estimating a position in an automated valet parking system, the apparatus comprising:
a front camera configured to capture a front image of a vehicle;
a front camera processor configured to process the front image of the vehicle received from the front camera;
a surround view monitor (SVM) processor configured to recognize a short-distance lane and a stop line by processing a surround view image of the vehicle;
a map data unit configured to store a high definition map; and
a controller configured to:
download a map comprising an area set as a parking zone from the map data unit when an entry of the vehicle to a parking lot is identified, and
correct a position measurement value of the vehicle by performing map matching based on results of a recognition and processing of the front camera processor and the SVM processor and the parking lot map of the map data unit, when an automated valet parking (AVP) start position is recognized based on the short-distance lane and stop line recognized by the SVM processor and when it is determined that the vehicle has parked at the AVP start position by receiving GPS information received from a GPS receiver or a signal, indicating that the vehicle has parked, from AVP infrastructure installed in the parking lot,
wherein the controller comprises a vehicle behavior prediction unit configured to predict a behavior of the vehicle through dead reckoning based on the GPS information received from the GPS receiver and a vehicle steering wheel angle, yaw rate and wheel speed received from a vehicle sensor unit, and
wherein the controller comprises a map-matching unit configured to perform the map matching based on lane fusion data in which a long-distance lane recognized by the front camera processor and the short-distance lane and stop line recognized by the SVM processor have been fused, parking lot map data from the map data unit, and vehicle behavior data for each time predicted through dead reckoning.
2. The apparatus ofclaim 1, wherein the controller is configured to:
predict a behavior of the vehicle through dead reckoning when the AVP start position is recognized, and
estimate an AVP initial position of the vehicle by fusing the position measurement value of the vehicle corrected through the map matching and the predicted behavior of the vehicle.
3. The apparatus ofclaim 1, wherein the map-matching unit is configured to compute a position and rotation correction quantity in which a distance error between sensor data and map data is minimized using iterative closest point (ICP) logic.
4. The apparatus ofclaim 1, wherein the controller comprises a position fusion unit configured to fuse a vehicle pose output as results of the map matching and GPS information of a vehicle position predicted through dead reckoning.
5. The apparatus ofclaim 4, wherein:
the controller comprises a fail-safe diagnosis unit configured to receive the vehicle position and flags output by the position fusion unit and to perform fail-safe diagnosis, and
the fail-safe diagnosis unit is configured to perform the fail-safe diagnosis using a distribution chart configured with estimated positioning results in which positioning results at past timing have been projected on current timing and positioning results input at current timing.
6. The apparatus ofclaim 4, wherein the vehicle pose comprises one or more of longitude, latitude, heading, covariance, warning/fail/safe, flags and a lane offset.
7. A method of estimating a position in an automated valet parking system, the method comprising:
capturing, by a front camera, a front image of a vehicle;
processing, by a front camera processor, the front image of the vehicle received from the front camera
downloading, by a controller, a map comprising an area set as a parking zone from a map data unit for storing a high definition map when an entry of the vehicle to a parking lot is identified;
recognizing, by the controller, an automated valet parking (AVP) start position based on a short-distance lane and stop line recognized by a surround view monitor (SVM) processor, when it is determined that the vehicle has parked at the AVP start position by receiving GPS information or a signal, indicating that the vehicle has parked, from AVP infrastructure installed in the parking lot; and
correcting, by the controller, a position measurement value of the vehicle by performing map matching based on results of a recognition and processing of the front camera processor and the SVM processor and the parking lot map of the map data unit,
wherein in the predicting of the behavior of the vehicle, the controller predicts the behavior of the vehicle through dead reckoning based on the GPS information received from the GPS receiver and a vehicle steering wheel angle, yaw rate and wheel speed received from a vehicle sensor unit, and
wherein in the correcting of the position measurement value, the controller performs the map matching based on lane fusion data in which a long-distance lane recognized by the front camera processor and the short-distance lane and stop line recognized by the SVM processor have been fused, parking lot map data from the map data unit, and vehicle behavior data for each time predicted through dead reckoning.
8. The method ofclaim 7, further comprising:
predicting, by the controller, a behavior of the vehicle through dead reckoning when the AVP start position is recognized, and
estimating, by the controller, an AVP initial position of the vehicle by fusing the position measurement value of the vehicle corrected through the map matching and the predicted behavior of the vehicle.
9. The method ofclaim 7, wherein in the correcting of the position measurement value, the controller computes a position and rotation correction quantity in which a distance error between sensor data and map data is minimized using iterative closest point (ICP) logic.
10. The method ofclaim 7, wherein in the estimating of the AVP initial position, the controller fuses a vehicle pose output as results of the map matching and GPS information of a vehicle position predicted through dead reckoning.
11. The method ofclaim 10, further comprising receiving, by the controller, the vehicle position and flags output as the result of the position fusion and performing fail-safe diagnosis,
wherein in the performing of the fail-safe diagnosis, the controller performs the fail-safe diagnosis using a distribution chart configured with estimated positioning results in which positioning
results at past timing have been projected on current timing and positioning results input at current timing.
US16/810,6692019-03-192020-03-05Apparatus and method for estimating position in automated valet parking systemActive2040-04-23US11440537B2 (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
US17/881,854US12326337B2 (en)2019-03-192022-08-05Apparatus and method for estimating position in automated valet parking system

Applications Claiming Priority (2)

Application NumberPriority DateFiling DateTitle
KR10-2019-00310922019-03-19
KR1020190031092AKR102696094B1 (en)2019-03-192019-03-19Appratus and method for estimating the position of an automated valet parking system

Related Child Applications (1)

Application NumberTitlePriority DateFiling Date
US17/881,854ContinuationUS12326337B2 (en)2019-03-192022-08-05Apparatus and method for estimating position in automated valet parking system

Publications (2)

Publication NumberPublication Date
US20200298836A1 US20200298836A1 (en)2020-09-24
US11440537B2true US11440537B2 (en)2022-09-13

Family

ID=72333901

Family Applications (2)

Application NumberTitlePriority DateFiling Date
US16/810,669Active2040-04-23US11440537B2 (en)2019-03-192020-03-05Apparatus and method for estimating position in automated valet parking system
US17/881,854Active2040-06-07US12326337B2 (en)2019-03-192022-08-05Apparatus and method for estimating position in automated valet parking system

Family Applications After (1)

Application NumberTitlePriority DateFiling Date
US17/881,854Active2040-06-07US12326337B2 (en)2019-03-192022-08-05Apparatus and method for estimating position in automated valet parking system

Country Status (4)

CountryLink
US (2)US11440537B2 (en)
KR (2)KR102696094B1 (en)
CN (1)CN111721285B (en)
DE (1)DE102020105639B4 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US12392619B2 (en)2020-05-222025-08-19Profound Positioning Inc.Vehicle localization system and method

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
JP7157876B2 (en)*2019-05-082022-10-20日立Astemo株式会社 Vehicle position detection device and parameter set creation device for vehicle position detection
US11781875B2 (en)*2019-08-212023-10-10Toyota Motor Engineering & Manufacturing North America, Inc.Apparatus and method for forming and analyzing connected roads
EP4194298A4 (en)*2020-08-052024-01-31Hitachi Astemo, Ltd. VEHICLE CONTROL DEVICE, VEHICLE CONTROL METHOD AND VEHICLE CONTROL SYSTEM
CN112319464B (en)*2020-11-092021-10-15恒大新能源汽车投资控股集团有限公司 Automatic parking method, device, equipment and storage medium
US11827244B2 (en)*2020-12-042023-11-28Ford Global Technologies, LlcEnhanced vehicle operation
CN112835359B (en)*2020-12-242022-11-04合众新能源汽车有限公司 An AVP control method and device based on visual SLAM technology
US20220205788A1 (en)*2020-12-282022-06-30Continental Automotive Systems, Inc.Ground vehicle monocular visual-inertial odometry via locally flat constraints
JP7294356B2 (en)*2021-02-192023-06-20トヨタ自動車株式会社 Vehicle control method, vehicle control system, and vehicle control program
JP7250833B2 (en)*2021-03-092023-04-03本田技研工業株式会社 OBJECT RECOGNITION DEVICE, OBJECT RECOGNITION METHOD, AND PROGRAM
CN113112847B (en)*2021-04-122025-09-30蔚来汽车科技(安徽)有限公司 Vehicle positioning method and system for fixed parking scene
CN113173158A (en)*2021-04-262021-07-27安徽域驰智能科技有限公司Vehicle positioning method based on look-around SLAM and vehicle kinematics
DE102021117744B4 (en)*2021-07-092025-02-27Cariad Se Self-localization of a vehicle based on an initial pose
US20230051155A1 (en)*2021-08-132023-02-16Here Global B.V.System and method for generating linear feature data associated with road lanes
CN113946956B (en)*2021-10-152025-07-08北京经纬恒润科技股份有限公司Method and device for simulating parking of passengers
DE102022112331A1 (en)2022-05-172023-11-23Valeo Schalter Und Sensoren Gmbh METHOD FOR OPERATING A PARKING ASSISTANCE SYSTEM, COMPUTER PROGRAM PRODUCT, PARKING ASSISTANCE SYSTEM AND A VEHICLE
EP4336144A1 (en)*2022-09-062024-03-13Aptiv Technologies LimitedMethod and system for determining a path of a vehicle and method for updating information contained in a digital map

Citations (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20160097862A1 (en)*2014-10-062016-04-07Hyundai Mobis Co., Ltd.System and method for complex navigation using dead reckoning and gps
US20170039439A1 (en)*2015-08-032017-02-09Hyundai Mobis Co., Ltd.Parking space recognition apparatus and method of controlling the same
US20180023961A1 (en)2016-07-212018-01-25Mobileye Vision Technologies Ltd.Systems and methods for aligning crowdsourced sparse map data
CN109326136A (en)*2017-07-312019-02-12中兴通讯股份有限公司Parking navigation method, equipment and computer readable storage medium
US20190367012A1 (en)*2018-05-292019-12-05Hitachi Automotive Systems, Ltd.Road marker detection method
US20200249695A1 (en)*2019-02-052020-08-06Visteon Global Technologies, Inc.Method for localizing a vehicle

Family Cites Families (18)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20100121518A1 (en)*2008-11-112010-05-13Timothy Arthur TiernanMap enhanced positioning sensor system
JP2013088208A (en)*2011-10-142013-05-13Furuno Electric Co LtdRoad map feedback server for tightly coupled gps and dead-reckoning vehicle navigation
EP3261074A1 (en)*2016-06-202017-12-27Volvo Car CorporationMethod for autonomous vehicle parking
US10248124B2 (en)*2016-07-212019-04-02Mobileye Vision Technologies, Inc.Localizing vehicle navigation using lane measurements
US20190271550A1 (en)*2016-07-212019-09-05Intelligent Technologies International, Inc.System and Method for Creating, Updating, and Using Maps Generated by Probe Vehicles
KR102420597B1 (en)*2016-11-172022-07-13현대모비스 주식회사Autonomous driving system fail-safe utility and method thereof
KR102660497B1 (en)*2016-12-142024-04-24현대모비스 주식회사System for positioning location information of car
KR102273355B1 (en)*2017-06-202021-07-06현대모비스 주식회사Apparatus for correcting vehicle driving information and method thereof
CN108482366A (en)*2018-03-232018-09-04重庆长安汽车股份有限公司Valet parking system and method based on Vehicular automatic driving
WO2019222358A1 (en)*2018-05-152019-11-21Mobileye Vision Technologies Ltd.Systems and methods for autonomous vehicle navigation
JP7119720B2 (en)*2018-07-302022-08-17株式会社デンソー Driving support device
US20200130676A1 (en)*2018-10-252020-04-30Magna Electronics Inc.Autonomous vehicle parking system
US20200132473A1 (en)*2018-10-262020-04-30Ford Global Technologies, LlcSystems and methods for determining vehicle location in parking structures
JP2022028092A (en)*2018-12-202022-02-15ソニーグループ株式会社Vehicle controller, vehicle control method, program, and vehicle
JP7120036B2 (en)*2019-01-162022-08-17トヨタ自動車株式会社 Automatic parking management device
JP7151526B2 (en)*2019-02-082022-10-12トヨタ自動車株式会社 Automatic parking management device
JP7155047B2 (en)*2019-03-072022-10-18本田技研工業株式会社 VEHICLE CONTROL DEVICE, VEHICLE CONTROL METHOD, AND PROGRAM
JP7123836B2 (en)*2019-03-082022-08-23本田技研工業株式会社 Vehicle control system, vehicle control method, and program

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20160097862A1 (en)*2014-10-062016-04-07Hyundai Mobis Co., Ltd.System and method for complex navigation using dead reckoning and gps
US20170039439A1 (en)*2015-08-032017-02-09Hyundai Mobis Co., Ltd.Parking space recognition apparatus and method of controlling the same
US20180023961A1 (en)2016-07-212018-01-25Mobileye Vision Technologies Ltd.Systems and methods for aligning crowdsourced sparse map data
CN109326136A (en)*2017-07-312019-02-12中兴通讯股份有限公司Parking navigation method, equipment and computer readable storage medium
US20190367012A1 (en)*2018-05-292019-12-05Hitachi Automotive Systems, Ltd.Road marker detection method
US20200249695A1 (en)*2019-02-052020-08-06Visteon Global Technologies, Inc.Method for localizing a vehicle

Cited By (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US12392619B2 (en)2020-05-222025-08-19Profound Positioning Inc.Vehicle localization system and method

Also Published As

Publication numberPublication date
DE102020105639B4 (en)2025-06-05
KR20220033477A (en)2022-03-16
DE102020105639A1 (en)2020-09-24
US12326337B2 (en)2025-06-10
US20220388500A1 (en)2022-12-08
KR102696094B1 (en)2024-08-20
CN111721285A (en)2020-09-29
CN111721285B (en)2023-09-26
US20200298836A1 (en)2020-09-24
KR20200119920A (en)2020-10-21

Similar Documents

PublicationPublication DateTitle
US11440537B2 (en)Apparatus and method for estimating position in automated valet parking system
CN114728657B (en) Vehicle control method and vehicle control device
JP6451857B2 (en) Method for controlling travel control device and travel control device
KR102441073B1 (en)Apparatus for compensating sensing value of gyroscope sensor, system having the same and method thereof
CN112797998A (en)Vehicle lane level positioning method, corresponding program carrier, product, device and vehicle
US20200110183A1 (en)Method of determining location of vehicle, apparatus for determining location, and system for controlling driving
JP6943127B2 (en) Position correction method, vehicle control method and position correction device
JP6790951B2 (en) Map information learning method and map information learning device
JP7234840B2 (en) position estimator
CN113375679B (en) Lane-level positioning method, device, system and related equipment
US11754403B2 (en)Self-position correction method and self-position correction device
US12270661B2 (en)Lane marking localization and fusion
US12221135B2 (en)Vehicle controller, method, and computer program for controlling vehicle
US12157466B2 (en)Vehicle controller, and method and computer program for controlling vehicle
CN116448128A (en)Travel route creation device, travel route creation method, and non-transitory computer-readable medium
RU2781373C1 (en)Method for correcting one's location and a device for correcting one's location
US12420841B2 (en)Vehicle controller, vehicle control method, and vehicle control computer program for vehicle control
US20240217545A1 (en)Vehicle controller, method, and computer program for vehicle control
JP7313325B2 (en) Self-localization device
JP7632419B2 (en) Vehicle control device, vehicle control method, and vehicle control computer program
JP7538740B2 (en) Driving assistance method and driving assistance device
KR20230053054A (en)System and method for determining location of a vehicle using historical information of roads and lanes
CN117985047A (en)Method, device, equipment, medium and product for controlling vehicle to run in tunnel

Legal Events

DateCodeTitleDescription
ASAssignment

Owner name:HYUNDAI MOBIS CO., LTD., KOREA, REPUBLIC OF

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KIM, DONG WOOK;REEL/FRAME:052032/0961

Effective date:20200213

FEPPFee payment procedure

Free format text:ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPPInformation on status: patent application and granting procedure in general

Free format text:DOCKETED NEW CASE - READY FOR EXAMINATION

STPPInformation on status: patent application and granting procedure in general

Free format text:NON FINAL ACTION MAILED

STPPInformation on status: patent application and granting procedure in general

Free format text:RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPPInformation on status: patent application and granting procedure in general

Free format text:FINAL REJECTION MAILED

STPPInformation on status: patent application and granting procedure in general

Free format text:DOCKETED NEW CASE - READY FOR EXAMINATION

STPPInformation on status: patent application and granting procedure in general

Free format text:NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPPInformation on status: patent application and granting procedure in general

Free format text:PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED

STCFInformation on status: patent grant

Free format text:PATENTED CASE


[8]ページ先頭

©2009-2025 Movatter.jp