Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
Training a gaze detection model requires a large number of gaze data samples as a training basis. The existing sight line data acquisition technology is limited by the mutual matching between an acquirer and an acquirer, the acquirer commands the acquirer to observe sight line points required to be acquired, then an acquisition tool is used for shooting images of the sight line points observed by the acquirer, and meanwhile labels of the acquired sight line points are recorded, wherein the labels are generally simple, such as watching an area A and not watching the area A. In the scheme, once the acquirer is not supervised and the acquirer is not matched with the command of the acquirer, the acquirer is difficult to find, and the error rate of the obtained sight line data sample is high.
In order to solve the above problems, embodiments of the present invention provide a method, an apparatus, a system, and a computer-readable storage medium for line-of-sight acquisition, which may be implemented by using corresponding software, hardware, and a combination of software and hardware. The following describes embodiments of the present invention in detail.
First, agaze acquisition system 100 for implementing a gaze acquisition method, apparatus, and method according to an embodiment of the present invention will be described with reference to fig. 1.
Thegaze acquisition system 100 may include a controller 120, a memory 110 coupled to the controller 120, adisplay device 130, asensing device 140, an image acquisition device 150, and a gaze acquisition apparatus, and these components of the memory 110, the controller 120, thedisplay device 130, thesensing device 140, the image acquisition device 150, and the gaze acquisition apparatus may be interconnected by a bus system and/or other form of connection mechanism (not shown). It should be noted that the components and configuration of thegaze acquisition system 100 shown in fig. 1 are exemplary only, and not limiting, and that thegaze acquisition system 100 may have other components and configurations as desired. The memory 110, thedisplay device 130, thesensing device 140, and the image capturing device 150 may be integrated into a module, or may be separately configured as a module.
The gaze acquisition device comprises at least one software functional module which may be stored in the memory 110 in the form of software or firmware or may be fixed in an Operating System (OS) of thegaze acquisition system 100. The controller 120 is configured to execute executable modules stored in the memory 110, such as software functional modules or computer programs included in the gaze acquisition apparatus.
The memory 110 may include one or more computer program products that may include various forms of computer-readable storage media, such as volatile memory and/or non-volatile memory. The volatile memory may include, for example, Random Access Memory (RAM), cache memory (cache), and/or the like. The non-volatile memory may include, for example, Read Only Memory (ROM), hard disk, flash memory, etc. On which one or more computer program instructions may be stored that may be executed by the controller 120 to implement the functions desired to be implemented in the embodiments of the invention described below. Various applications and various data, such as various data used and/or generated by the applications, may also be stored in the computer-readable storage medium. Of course, the device coordinates of the image capturing device 150 with respect to thedisplay device 130 and the target picture generation data may be stored in the memory 110 in advance.
The controller 120 may be an integrated circuit chip having signal processing capabilities. The controller 120 may be a general-purpose Processor, and includes a Central Processing Unit (CPU), a Network Processor (NP), and the like; but may also be a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic device, discrete hardware components. The controller 120 may implement or perform the methods, steps, and logic blocks disclosed in embodiments of the present invention.
Thedisplay device 130 provides an interactive interface (e.g., a user operation interface) between thegaze acquisition system 100 and the tester or is used for displaying image data to the tester for reference, and in this embodiment, thedisplay device 130 may be a tablet with a display function, a mobile terminal, or the like.
Thesensing device 140 is an input device, and may be a capacitive touch screen or a resistive touch screen supporting single-point and multi-point touch operations. Supporting single-point and multi-point touch operations means that the touch display screen can sense touch operations simultaneously generated from one or more positions on the touch display screen, and the sensed touch operations are sent to the controller 120 for calculation and processing. Of course, in an alternative embodiment, thesensing device 140 may also be a mouse for sensing the position of the user click.
The image capturing device 150 is used for capturing a picture of a face of a tester, and may be a camera, or the like.
The following will describe a method of line-of-sight acquisition for optimization purposes in line-of-sight data acquisition:
referring to fig. 2, fig. 2 is a flowchart of a gaze acquisition method according to an embodiment of the present invention, which is described from the perspective of the controller 120 of thegaze acquisition system 100. The memory 110 of thegaze acquisition system 100 stores therein target picture generation data and device coordinates of the plurality of image acquisition devices 150 with respect to thedisplay device 130 in advance.
The flow shown in fig. 2 will be described in detail below, and the method includes:
step S110: when the controller acquires a picture replacing instruction, a current target picture comprising a sight point is determined based on the target picture generation data, the display equipment is controlled to display the current target picture, and the image acquisition equipment is controlled to acquire a face picture of a tester.
Since the sight line acquisition data requires the image data of the pupils of the eyes of the testee, the face picture may be a face picture including the eyes of the testee acquired by each image acquisition device 150 connected to the controller 120.
As an optional implementation manner, the target picture generation data may include a plurality of pictures, each of which includes a preset sight point, and the controller 120 may determine, when the picture replacement instruction is obtained, a target picture from the plurality of pictures as the current target picture.
In this embodiment, optionally, the controller 120 may randomly select one picture from the plurality of pictures to be determined as the current target picture when the command for replacing a picture is obtained.
In this embodiment, optionally, the multiple pictures may be stored in the memory 110 in advance according to a time sequence, and the controller 120 may sequentially select one picture from the multiple pictures according to the sequence of the storage times of the multiple pictures when the picture replacement instruction is obtained, and determine the selected picture as the current target picture. Alternatively, a plurality of pictures may be stored in a queue in the memory 110, and since the reading rule of the queue is first in first out, the picture stored in the queue first is read by the controller 120 and determined as the current target picture.
The controller 120 controls thedisplay device 130 to display the current target picture after determining the current target picture. Each target picture may include a sight line point, and the memory 110 stores coordinates of the sight line point included in each target picture. When displaying the current target picture, thedisplay device 130 displays the sight points included in the current target picture together.
The sight point included in the target picture may be a specific shape displayed in the target picture, such as a cross as shown in fig. 3, or may be any other irregular shape. Of course, the shapes of the sight line points in each target picture may also be different, and in this case, in order to make the sight line points stand out and be easily found by the testee, the background of the target picture may be set to be a pure color, and the color of the sight line point may be set to be another color having a larger contrast with the background color of the target picture.
As another optional implementation, the target picture generation data may include: the device comprises a picture, a sight point identifier and a plurality of sight point coordinates, wherein the picture is used as a background picture. The controller 120 may determine a target sight point coordinate from the plurality of sight point coordinates when the image replacement instruction is obtained, and then display the sight point identifier at the target sight point coordinate in the background image, so as to obtain the current target image.
In this embodiment, optionally, the plurality of sight point coordinates may be generated by a function stored in the memory 110 and stored in the memory 110, or may be generated by a function stored outside the memory 110 and transmitted to the memory 110 for storage. Of course, in this embodiment, regular coordinates generated by a function, such as { (1, 1), (1, 2), (1, 3), (1, 4), (2, 1), (2, 2), (2, 3), (2, 4) } and the like, may be between the plurality of gaze point coordinates, at which time the gaze point identifiers appear on the picture in a varying order from left to right, from top to bottom; the coordinates of the plurality of sight points may be random coordinates randomly generated by a random function.
As another optional implementation, the target picture generation data may include: the method comprises the steps of one picture, one sight point mark and a random number generation function, wherein the picture is used as a background picture. The controller 120 may control the random number generation function to randomly generate a target sight point coordinate when the image replacement instruction is obtained, and then display the sight point identifier at the target sight point coordinate in the background image, so as to obtain the current target image.
In this embodiment, optionally, in order to increase the diversity of the target gaze point coordinates, the target picture generation data further includes a plurality of existing gaze point coordinates, for example { (1, 1), (1, 2), (1, 3), (1, 4), (2, 1), (2, 2), (2, 3), (2, 4) } and the like. The controller 120 controls the random number generating function to randomly generate a coordinate, superimposes the coordinate onto one of the existing multiple sight point coordinates to generate a target sight point coordinate, and displays the sight point identifier at the target sight point coordinate in the background picture to obtain the current target picture. It is to be noted that one of the plurality of sight point coordinates may be selected at random from the plurality of sight point coordinates, or may be selected sequentially in the order of storing the plurality of sight point coordinates.
Optionally, the above-mentioned picture replacing instruction may be automatically generated by the controller 120 every preset time (for example, 2 seconds), or may be generated by the controller 120 when acquiring the gaze point acquisition signal sent by thesensing device 140. Of course, the image replacement command may also be generated by combining the above two situations, for example, when the controller 120 does not acquire the sight point acquisition signal sent by thesensing device 140 within the preset time, the image replacement command may be automatically generated.
Since the sight line acquisition data required to be obtained by the embodiment of the present invention is pupil data of the sight line point included in the current target picture seen by the eyes of the tester when thedisplay device 130 displays the current target picture, if the controller 120 controls the image acquisition device 150 to acquire the face picture of the tester when the tester looks at the sight line point included in the current target picture a and gives feedback, the face picture is most likely to be a face picture when the tester looks at the sight line point included in the current target picture B presented at the next time point, and the accuracy of the sight line acquisition data is reduced. In order to avoid the above problem and ensure the accuracy of the sight line acquisition data as much as possible, the controller 120 needs to control the image acquisition device 150 to acquire a face image of the tester while displaying the current target image.
Step S120: and the controller caches the coordinates of the sight points corresponding to the face picture acquired by the image acquisition equipment and the current target picture in the display equipment in a one-to-one correspondence manner to form alternative picture information.
Regardless of the manner in which the current target picture is determined, the controller 120 may obtain the coordinates of the gaze point corresponding to the current target picture in thedisplay device 130. Of course, it is worth noting that one vertex of the display interface of thedisplay device 130 may be defined as the origin of coordinates to determine the coordinates of the gaze point within the display device.
The number of the image acquisition devices 150 may be one, and after the image acquisition devices 150 acquire the face picture, the coordinates of the sight points included in the face picture and the current target picture in thedisplay device 130 may be cached in a one-to-one correspondence manner, so as to form alternative picture information. As time goes by, the image capturing device 150 may capture a plurality of face pictures, each of which corresponds to a different current target picture, and finally form a plurality of candidate picture information to be stored in the memory 110.
Optionally, in order to speed up the acquisition of the gaze acquisition data, the number of image acquisition devices 150 may also be multiple, and each image acquisition device 150 has different device coordinates with respect to thedisplay device 130.
In this embodiment, for the same current target picture, when the controller 120 controls the image capturing device 150 to capture a face picture of the tester, a plurality of face pictures can be obtained according to the difference of the image capturing device 150. Similarly, when the candidate picture information is formed subsequently, for the coordinates of the same sight point in thedisplay device 130, the controller 120 may obtain multiple sets of candidate picture information corresponding to the image capturing devices 150 one to one according to the differences of the image capturing devices 150.
Since the face image acquired by the image acquisition device 150 may not meet the requirement of forming the sight line data, optionally, a preset condition of the face orientation state may be stored in the memory 110 in advance. The preset condition of the face orientation state may be: in the face picture corresponding to the image acquisition device 150, the distance between the face and thedisplay device 130 is within a preset distance range; or in the face picture corresponding to the image acquisition device 150, the orientation of the face is within a preset angle range; or in the face picture corresponding to the image acquisition device 150, human eyes are not shielded; any two or all three of the above three conditions may be satisfied simultaneously.
Before the controller 120 caches the coordinates of the sight points corresponding to the face picture acquired by one image acquisition device 150 and the current target picture in the display device in a one-to-one correspondence manner to form alternative picture information, the controller 120 may determine whether the face picture corresponding to the image acquisition device 150 meets the preset condition of the face orientation state; when the condition is met, the controller 120 caches the coordinates of the sight point in thedisplay device 130 corresponding to the face picture acquired by the image acquisition device 150 and the current target picture in a one-to-one correspondence manner, so as to form alternative picture information. If not, the controller 120 may re-determine the next current target picture and re-capture the face picture of the tester via the image capture device 150.
Of course, optionally, the controller 120 may further control thedisplay device 130 to issue a capture error prompt before the next current target picture is determined again, so that the tester can adjust his posture. The acquisition error prompt may be prompt content for correcting the posture of the tester displayed according to a specific situation that the preset condition of the face orientation state is not satisfied, for example: "too far away, please get close to the display device".
The following description will be made with respect to the controller 120 re-determining the next current target picture.
As an alternative embodiment, when the controller 120 re-determines the next current target picture, the next current target picture may be the same as the last current target picture, i.e. the controller 120 may directly re-display the last current target picture.
As another alternative, when the controller 120 re-determines the next current target picture, the next current target picture may be different from the last current target picture.
In this embodiment, if the target picture generation data includes a plurality of pictures, the controller 120 determines a picture as the next current target picture after excluding the current target picture from the plurality of pictures.
In this embodiment, if the target picture generation data includes a picture, a sight point identifier, and a random number generation function, the controller 120 may obtain the next target sight point coordinate generated by the random number generation function, so as to determine a target picture as the next current target picture.
In this embodiment, if the target picture generation data includes a picture, a sight point identifier, and a plurality of sight point coordinates, the controller 120 determines a new target sight point coordinate after excluding the target sight point coordinate corresponding to the current target picture from the plurality of sight point coordinates, thereby determining a target picture as the next current target picture.
In this embodiment, if the target picture generation data includes a picture, a sight point identifier, a random number generation function, and a plurality of sight point coordinates, the controller 120 may superimpose the random coordinates onto the target sight point coordinates corresponding to the current target picture after the random number generation function generates a random coordinate, so as to form a new target sight point coordinate, thereby determining a target picture as the next current target picture.
Step S130: and when the controller acquires a sight point acquisition signal sent by the induction equipment, determining target picture information from at least one piece of alternative picture information corresponding to the image acquisition equipment.
The sight-line point acquisition signal may be a signal generated when thesensing device 140 detects that the tester touches the sight-line point with a finger or a stylus, or may be a signal generated when the tester clicks the sight-line point with a mouse.
When the controller 120 acquires the sight point acquisition signal, determining a target picture information from at least one candidate picture information acquired by the image acquisition device 150 may include:
the controller 120 may obtain the generation time of the sight-line point acquisition signal, and then subtract a preset value from the generation time to obtain a comparison time. The preset value represents the time interval for triggering the gaze point to acquire the signal after the tester sees the gaze point, and can be set to 300 milliseconds generally. Then, the controller 120 selects one candidate picture information from at least one candidate picture information acquired by the image acquisition device 150 as the target picture information, wherein the storage time of the target picture information is closest to the comparison time.
Optionally, when the controller 120 detects that the mouse moves in thedisplay device 130 or detects that the finger or the stylus slides on thesensing device 140, the controller 120 may further collect coordinates of an operation position of the mouse, the finger or the stylus in thedisplay device 130, and store the coordinates as reference coordinates in the memory 110.
Of course, it is worth pointing out that, if the controller 120 detects that the finger or the stylus pen slides on thesensing device 140, at this time, the controller may further detect a staying time of the finger or the stylus pen at the same position of thesensing device 140, if the staying time does not exceed the preset staying time, the operating position coordinate at this time is saved as a reference coordinate, if the staying time exceeds the preset staying time, the finger or the stylus pen is pressed for a long time on thesensing device 140, and at this time, the sight point acquisition signal is triggered.
In this embodiment, when the controller 120 caches the coordinates of the sight point corresponding to the face picture acquired by the image acquisition device 150 and the current target picture in thedisplay device 130 in a one-to-one correspondence manner to form the candidate picture information, the controller 120 may also cache the face picture acquired by the image acquisition device 150, the reference coordinates corresponding to the face picture acquired by the image acquisition device 150, and the coordinates of the sight point corresponding to the current target picture in thedisplay device 130 in a one-to-one correspondence manner to form the candidate picture information.
It should be noted that there may be no reference coordinate corresponding to the time point when a certain face picture is captured, because the stylus or mouse or the finger of the captured person is not in a sliding state. At this time, the controller 120 may use, as the reference coordinate of the face picture, a reference coordinate corresponding to a time point at which the reference coordinate is stored, which is closest to the time point at which the face picture is acquired, among the already stored reference coordinates. Accordingly, when the controller 120 determines a target picture information from the at least one candidate picture information corresponding to the image capturing device 150, the controller 120 may select one candidate picture information from the at least one candidate picture information captured by the image capturing device 150 as the target picture information. The distance between the reference coordinate corresponding to the target picture information and the coordinate of the sight point included in the target picture information in the display device 150 is greater than a preset value (that is, the reference coordinate is out of the sight point coordinate range), and the storage time of the target picture information is closest to the comparison time, so that a picture that the testee does not move yet but is about to move to the sight point can be found, and the state of the human eyes is most concentrated on the sight point.
Because the coordinate of the point in contact with thesensing device 140 is deviated from the coordinate of the sight point corresponding to the current target picture when the tester clicks the sight point by using a finger, a stylus, or a mouse to trigger generation of the sight point acquisition signal, optionally, before determining one target picture information from at least one candidate picture information corresponding to the image acquisition device, the controller 120 determines whether an error between the coordinate triggering the sight point acquisition signal and the coordinate of the sight point corresponding to the current target picture in the display device exceeds a preset error threshold; if not, the controller 120 determines a target picture information from at least one candidate picture information corresponding to the image capturing device 150; when yes, the controller 120 re-determines the next current target picture.
The case where the controller 120 re-determines the next current target picture here is the same as the case where the controller 120 re-determines the next current target picture described in step S120 above, and is not described here again to avoid repetition.
Step S140: and the controller stores the target picture information corresponding to the image acquisition equipment and the equipment coordinates into the memory to form sight line acquisition data.
Since the device coordinates of the image capturing device 150 with respect to thedisplay device 130 are stored in the memory 110 in advance, after the target picture information is determined, the controller 120 stores the device coordinates of the image capturing device 150 corresponding to the target picture information in the memory 110, and forms the line-of-sight capturing data.
Wherein one line of sight acquisition data is uniquely determined by the device coordinates, the face picture and the coordinates of the line of sight point within thedisplay device 130.
According to the sight line acquisition method applied to the embodiment of the invention, a picture replacement instruction is obtained through a controller, a current target picture comprising a sight line point is determined based on pre-stored target picture generation data, the display equipment is controlled to display the current target picture, and the image acquisition equipment is controlled to acquire a face picture of a tester; and then, caching the coordinates of the sight point corresponding to the face picture acquired by the image acquisition equipment and the current target picture in the display equipment in a one-to-one correspondence manner to form alternative picture information, subsequently determining one piece of target picture information from at least one piece of alternative picture information corresponding to the image acquisition equipment when the controller acquires a sight point acquisition signal sent by the induction equipment, and storing the target picture information corresponding to the image acquisition equipment and the equipment coordinates in the memory to form sight acquisition data. When the sight line point is clicked, the portrait before the clicking moment is acquired, and the acquired sight line acquisition data is ensured to be correct as much as possible, so that the accuracy of the sight line data sample is improved; in addition, the whole process can be completed by only one tester, so that the condition that the action of a director is inconsistent with that of the tester is avoided, or the tester does not listen to the command condition of the director, or the condition that the director needs to supervise is avoided.
Referring to fig. 4, in response to the view line acquiring method provided in fig. 2, an embodiment of the invention further provides a viewline acquiring apparatus 400 applied to the controller 120 in the viewline acquiring system 100, where the viewline acquiring apparatus 400 may include: aselection display module 410, acaching module 420, adetermination module 430, and a savingmodule 440.
Theselection display module 410 is used for determining a current target picture comprising a sight point based on target picture generation data when a picture replacement instruction is obtained, controlling the display device to display the current target picture, and controlling the image acquisition device to acquire a portrait picture of a tester;
acache module 420, configured to cache coordinates of a sight point in the display device, where the sight point corresponds to the face picture acquired by the image acquisition device and the current target picture, in a one-to-one correspondence manner, so as to form alternative picture information;
a determiningmodule 430, configured to determine, when a gaze point acquisition signal sent by the sensing device is acquired, target picture information from at least one piece of candidate picture information corresponding to the image acquisition device, where the gaze point acquisition signal is a signal generated by the sensing device when it is detected that the tester clicks the gaze point;
and a savingmodule 440, configured to save the target picture information and the device coordinates corresponding to the image capturing device to the memory to form the gaze capturing data.
Optionally, the target picture generation data includes a picture, a sight point identifier, and a plurality of sight point coordinates, and theselection display module 410 is configured to determine a target sight point coordinate from the plurality of sight point coordinates when the picture replacement instruction is obtained; and displaying the sight point identification at the position of the target sight point coordinate on the picture to obtain the current target picture.
Optionally, the plurality of sight point coordinates are generated by a function stored in the memory, or generated by a function stored outside the memory and then sent to the memory.
Optionally, the target picture generation data includes a picture, a sight point identifier, and a random number generation function, and theselection display module 410 is configured to control the random number generation function to randomly generate a target sight point coordinate when the picture replacement instruction is acquired; and displaying the sight point identification at the position of the target sight point coordinate on the picture to obtain the current target picture.
Optionally, the target picture generation data further includes a plurality of sight point coordinates, and theselection display module 410 is configured to control the random number generation function to randomly generate one coordinate; and superposing the coordinate to one of the plurality of sight point coordinates to generate a target sight point coordinate.
Optionally, the target picture generation data includes a plurality of pictures, each of the pictures includes a preset sight point, and theselection display module 410 is configured to determine a target picture from the plurality of pictures as the current target picture when the picture change instruction is acquired.
Optionally, theselection display module 410 is configured to randomly select one picture from the multiple target pictures to determine that the picture is the current target picture when the picture replacement instruction is obtained.
Optionally, the multiple target pictures are stored in the memory in advance according to a time sequence, and theselection display module 410 is configured to, when the picture replacement instruction is obtained, sequentially select one picture from the multiple target pictures according to the sequence of the storage times of the multiple target pictures and determine the selected picture as the current target picture.
Optionally, the image replacement instruction is automatically generated at preset time intervals or generated by the device when the sight point acquisition signal is acquired.
Optionally, the memory stores preset conditions of the face orientation state in advance, and the device further includes a determining module and a re-determining module. And the judging module is used for judging whether the face picture corresponding to the image acquisition equipment meets the preset condition of the face azimuth state. If the result of the determination is yes, thecaching module 420 caches the coordinates of the sight point in the display device, where the sight point corresponds to the face picture acquired by the image acquisition device and the current target picture, in a one-to-one correspondence manner, so as to form alternative picture information, otherwise, the re-determining module is configured to re-determine the next current target picture.
Optionally, the re-determining module may be configured to re-determine the next current target picture after the display device issues the acquisition error prompt.
Optionally, the preset condition of the face orientation state may be: in the face picture corresponding to the image acquisition equipment, the distance between the face and the display equipment is within a preset distance range, and/or in the face picture corresponding to the image acquisition equipment, the orientation of the face is within a preset angle range, and/or the eyes are not shielded.
The judging module is further configured to judge whether an error between the coordinate triggering the sight point acquisition signal and the coordinate of the sight point corresponding to the current target picture in the display device exceeds a preset error threshold. If the determination result is negative, the determiningmodule 430 determines a target picture information from at least one candidate picture information corresponding to the image capturing device; and when the judgment result is yes, the re-determining module is used for re-determining the next current target picture.
Optionally, the next current target picture and the last current target picture may be the same or different.
Optionally, the determiningmodule 430 is configured to acquire, by the controller, a generation time of the gaze point acquisition signal; the controller subtracts a preset value from the generation time to obtain comparison time; and the controller selects one piece of alternative picture information from at least one piece of alternative picture information acquired by the image acquisition equipment as the target picture information, wherein the storage time of the target picture information is closest to the comparison time.
Optionally, the sight line point acquisition signal is a signal generated when a finger or a touch pen touches the sight line point or a signal generated when a mouse clicks the sight line point.
Optionally, when the controller 120 detects that the mouse moves in thedisplay device 130 or detects that the finger or the stylus slides on thesensing device 140, the controller 120 may further collect coordinates of an operation position of the mouse, the finger or the stylus in thedisplay device 130, and store the coordinates as reference coordinates in the memory 110.
Of course, it is worth pointing out that, if the controller 120 detects that the finger or the stylus pen slides on thesensing device 140, at this time, the controller 120 may further detect a staying time of the finger or the stylus pen at the same position of thesensing device 140, if the staying time does not exceed the preset staying time, the coordinate at this time is saved as the reference coordinate, if the staying time exceeds the preset staying time, the finger or the stylus pen is pressed for a long time on thesensing device 140, and at this time, the sight line point acquisition signal is triggered.
In this embodiment, when the controller 120 caches the coordinates of the sight point corresponding to the face picture acquired by the image acquisition device 150 and the current target picture in thedisplay device 130 in a one-to-one correspondence manner to form the candidate picture information, the controller 120 may also cache the face picture acquired by the image acquisition device 150, the reference coordinates corresponding to the face picture acquired by the image acquisition device 150, and the coordinates of the sight point corresponding to the current target picture in thedisplay device 130 in a one-to-one correspondence manner to form the candidate picture information.
Accordingly, when a target picture information is determined from at least one candidate picture information corresponding to the image capturing device 150, the controller 120 may select one candidate picture information from the at least one candidate picture information captured by the image capturing device 150 as the target picture information. The distance between the reference coordinate corresponding to the target picture information and the coordinate of the sight point included in the target picture information in the display device 150 is greater than a preset value (that is, the reference coordinate is out of the sight point coordinate range), and the storage time of the target picture information is closest to the comparison time, so that a picture that the testee does not move yet but is about to move to the sight point can be found, and the state of the human eyes is most concentrated on the sight point.
The device provided in this embodiment has the same implementation principle and technical effect as those of the foregoing embodiments, and for the sake of brief description, reference may be made to the corresponding contents in fig. 1 to 3 in the foregoing method embodiments without reference to the device embodiment.
In addition, an embodiment of the present invention further provides a sight line acquisition system, including: the system comprises a controller, a display device coupled with the controller, a sensing device, a memory and an image acquisition device, wherein the memory is pre-stored with device coordinates of the image acquisition device relative to the display device and target picture generation data, and the memory is further stored with a computer program, and when the computer program is executed by the controller, the sight line acquisition system is enabled to execute the method of any one of the embodiments in the first embodiment. The structure of the sight line acquisition system can be as shown in fig. 1.
For the specific implementation process of the sight line acquisition system, reference is made to the foregoing embodiments, and details are not repeated here.
In addition, an embodiment of the present invention further provides a computer-readable storage medium, in which a computer program is stored, and when the computer program runs on a computer, the computer is caused to execute the gaze acquisition method provided in any one of the embodiments of the present invention.
In addition, an embodiment of the present invention further provides a computer program, where the computer program may be stored in a cloud or a local storage medium, and when the computer program runs on a computer, the computer is enabled to execute the gaze acquisition method provided in any embodiment of the present invention.
In summary, according to the sight line acquisition method, the sight line acquisition device, the sight line acquisition system and the computer-readable storage medium provided by the embodiments of the present invention, when the image replacement instruction is acquired by the controller, a current target image including a sight line point is determined based on pre-stored target image generation data, and the display device is controlled to display the current target image and the image acquisition device is controlled to acquire a face image of a tester; and then the controller caches the coordinates of the sight point corresponding to the face picture acquired by the image acquisition equipment and the current target picture in the display equipment in a one-to-one correspondence manner to form alternative picture information, subsequently when the controller acquires a sight point acquisition signal sent by the induction equipment, the controller determines one target picture information from at least one alternative picture information corresponding to the image acquisition equipment, and stores the target picture information corresponding to the image acquisition equipment and the equipment coordinates corresponding to the image acquisition equipment in the memory to form sight acquisition data. When the sight line point is clicked, the portrait before the clicking moment is acquired, and the acquired sight line acquisition data is ensured to be correct as much as possible, so that the accuracy of the sight line data sample is improved; in addition, the whole process can be completed by only one tester, so that the condition that the action of a director is inconsistent with that of the tester is avoided, or the tester does not listen to the command condition of the director, or the condition that the director needs to supervise is avoided.
In the embodiments provided in the present application, it should be understood that the disclosed apparatus and method can be implemented in other ways. The apparatus embodiments described above are merely illustrative, and for example, the flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of apparatus, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
In addition, the functional modules in the embodiments of the present invention may be integrated together to form an independent part, or each module may exist separately, or two or more modules may be integrated to form an independent part.
The functions, if implemented in the form of software functional modules and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present invention may be embodied in the form of a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present invention. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
It is noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
The above description is only for the specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present invention, and all the changes or substitutions should be covered within the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.