Embodiment
Be described below in detail embodiments of the invention, the example of described embodiment is shown in the drawings, and wherein same or similar label represents same or similar element or has element that is identical or similar functions from start to finish.Being exemplary below by the embodiment be described with reference to the drawings, only for explaining the present invention, and can not limitation of the present invention being interpreted as.
In describing the invention, it will be appreciated that, term " longitudinal direction ", " transverse direction ", " on ", D score, "front", "rear", "left", "right", " vertically ", " level ", " top ", " end " " interior ", the orientation of the instruction such as " outward " or position relationship be based on orientation shown in the drawings or position relationship, only the present invention for convenience of description and simplified characterization, instead of indicate or imply that the device of indication or element must have specific orientation, with specific azimuth configuration and operation, therefore can not be interpreted as limitation of the present invention.
In describing the invention, it should be noted that, unless otherwise prescribed and limit, term " installation ", " being connected ", " connection " should be interpreted broadly, such as, can be mechanical connection or electrical connection, also can be the connection of two element internals, can be directly be connected, also indirectly can be connected by intermediary, for the ordinary skill in the art, the concrete meaning of described term can be understood as the case may be.
In order to solve the intelligent difference of the anthropomorphic robot existed in correlation technique and can not very well and the mankind carry out mutual problem, the present invention is based on artificial intelligence and achieve intelligent height and the anthropomorphic robot's control method experienced with human interaction, system and anthropomorphic robot, wherein, artificial intelligence (ArtificialIntelligence, be called for short: AI), be research, develop the theory of intelligence for simulating, extending and expand people, method, one of application system new technological sciences.Artificial intelligence is a branch of computer science, the essence of intelligence is understood in attempt, and produce a kind of intelligent machine can made a response in the mode that human intelligence is similar newly, the research in this field comprises robot, speech recognition, image recognition, natural language processing and expert system etc.
Artificial intelligence is the simulation to the consciousness of people, the information process of thinking.Artificial intelligence is not the intelligence of people, but can think deeply as people, may exceed the intelligence of people yet.Artificial intelligence comprises science very widely, be made up of different fields, as machine learning, computer vision etc., generally speaking, a main target of artificial intelligence study is the complex work enabling machine be competent at some usually to need human intelligence just can complete.
Describe according to the anthropomorphic robot's control method based on artificial intelligence of the embodiment of the present invention, control system and anthropomorphic robot below in conjunction with accompanying drawing.
Fig. 1 is according to an embodiment of the invention based on the process flow diagram of anthropomorphic robot's control method of artificial intelligence.As shown in Figure 1, the method comprises the steps:
S101: the voice signal and/or the picture signal that receive user's input.
Particularly, the voice signal of user's input can be that user is inputted by microphone; Above-mentioned picture signal can be obtained by camera collection.
S102: the intention determining described user according to voice signal and/or picture signal.That is: analyzing and processing can be carried out to voice signal and/or picture signal by artificial intelligence, thus determine the intention of user.It should be noted that, can carry out the intention that analyzing and processing determines user by artificial intelligence to any one in voice signal and picture signal, also can be the intention by determining user to both combinations.
Particularly, as shown in Figure 2, the intention of user can be determined according to following two kinds of modes, specifically comprise:
1, speech recognition is carried out to voice signal, and one or more in natural language understanding, semantic analysis, sentiment analysis are carried out to recognition result, to determine the intention of user.Specifically, process for voice signal needs to carry out speech recognition, natural language understanding, semantic analysis, mechanical translation, sentiment analysis etc., by above-mentioned process, anthropomorphic robot is carrying out in mutual process with user, when user says in short time, the implication of the voice that user inputs can be learnt.
2, speech recognition is carried out to voice signal, and one or more in natural language understanding, semantic analysis, sentiment analysis are carried out to recognition result, and combining image signal determines the intention of user.Specifically, when anthropomorphic robot learns the implication of the voice that user inputs, the intention of definitely user can be carried out further combined with the picture signal of user.Such as: when the voice of user's input are " shaking hands ", and identify user by the picture signal of user and give the action of stretching out the right hand, then now can specify user be intended to shake hands with anthropomorphic robot.
In addition, the mode combined by both not only can the intention of definitely user, can also wherein one when not identifying, and is determined the intention of user by another.Such as: when the action of user is ambiguous or when not providing clear and definite action, " can be shaken hands " intention determining user by the voice identified.Equally, when voice signal is bad, when can not identify voice signal, then can determine that user goes out the intention of user, the intention of namely shaking hands according to " user stretches out the action of the right hand " of the user recognized in picture signal.
As shown in Figure 3, the intention of user can also be determined according to other two kinds of modes, specifically comprise:
3, image recognition is carried out to picture signal, determine the user in picture signal, and determine the limb action of user according to the action difference of the user in picture signal between multiple image, and determine the intention of user according to the limb action of user.Specifically, (as 2 seconds) gather the vision signal (i.e. multiple continuous print image) of user in a short period of time, then the limb action of user is determined according to the action difference of user in multiple continuous print picture signal, also for " user stretches out the action of the right hand ", then determine that the limb action of user is " action that user stretches out the right hand ", thus determine being intended to " shaking hands " of user.
4, image recognition is carried out to picture signal, determine the user in picture signal, and determine the limb action of user according to the action difference of the user in picture signal between multiple image, and determine the intention of user according to the limb action of user and/or voice signal.Specifically, after being determined that by picture signal the limb action of user is " action that user stretches out the right hand ", in conjunction with voice signal, the implication identifying voice signal is " shaking hands ", then both combine can the intention of definitely user.
In addition, the mode combined by both not only can the intention of definitely user, can also wherein one when not identifying, and is determined the intention of user by another.Such as: when the action of user is ambiguous or when not providing clear and definite action, " can be shaken hands " intention determining user by the voice identified.Equally, when voice signal is bad, when can not identify voice signal, then can determine that user goes out the intention of user, the intention of namely shaking hands according to " user stretches out the action of the right hand " of the user recognized in picture signal.
S103: the intention of user is processed, and result is fed back to user with the multi-modal way of output, the multi-modal way of output comprise in the action way of output of anthropomorphic robot, image or video frequency output mode and the audio frequency way of output one or more.
In one embodiment of the invention, result is fed back to user with the multi-modal way of output to comprise: control anthropomorphic robot and perform the action corresponding with the intention of user; And/or; Show the expression relevant to the intention of described user; And/or carry out the image demonstration relevant to the intention of described user or video display; And/or play the audio frequency relevant to the intention of described user.
Such as: when anthropomorphic robot determines being intended to " shaking hands " of user, then can control the one that anthropomorphic robot performs action, the display of stretching out its right hand " smiling face's image " corresponding with the scene of the close friend such as " shaking hands ", plays the music corresponding with the friendly scene such as " shaking hands " etc. intelligently, or the multiple combination in above-mentioned various feedback mode, as stretch out the right hand action while display " smiling face's image " to show friendship.Thus promote the man-machine interaction experience of anthropomorphic robot.
As shown in Figure 4, in one embodiment of the invention, before the intention determining user according to voice signal and/or picture signal, also comprise:
S401: detect user and whether call out anthropomorphic robot.Such as: set a name to anthropomorphic robot in advance, as " little Bai ", then, as user's calling " little Bai ", anthropomorphic robot carries out speech recognition by the mode of artificial intelligence to " little Bai " that user calls out, thus make anthropomorphic robot understand its implication, just know whether as calling out oneself.
S402: if then activate anthropomorphic robot, and carry out auditory localization to determine the position of user according to the calling of user, and control anthropomorphic robot moves in face of user.
Specifically, if anthropomorphic robot determines that user calls out oneself really, then activate, each functional module being such as anthropomorphic robot by battery etc. powers on, and makes anthropomorphic robot be in state of activation.In one embodiment of the invention, in order to promote man-machine interaction experience, when anthropomorphic robot determines that user is when calling out it, the activation expression of anthropomorphic robot can be shown, activate expression can be predefined, just can as long as can reflect that robot is activated intuitively.Thus, promote man-machine interaction experience further.
In addition, auditory localization is carried out to determine the position of user according to the calling of user.The sound around anthropomorphic robot is detected by ManyEars auditory localization technology, particularly, sound-source signal can be gathered according to microphone array, effective sound signal detection can be carried out to sound-source signal afterwards, and by ManyEars technology, the multi-acoustical detected be carried out being separated to obtain multiple independently sound source.Wherein, the term " at least one " of the present embodiment can be regarded as one or more.Further, by the auditory localization computing in ManyEars technology, certain sound source above-mentioned is positioned, thus determine the position of user.
After the position determining user, just can automatically control anthropomorphic robot and move in face of user.Further, also comprise: detect anthropomorphic robot and whether move in face of user; If not, then control anthropomorphic robot further to carry out moving until reach in face of user.Thus closed loop is formed to the motion control of anthropomorphic robot, ensure that the motion of anthropomorphic robot is more accurate.
In one embodiment of the invention, the method be also included in control anthropomorphic robot move to user in face of time, by the camera steering surface of anthropomorphic robot to the direction of user, to take pictures to user, and carry out recognition of face according to the picture signal of user, to determine the identity information of user.Like this, can prevent disabled user from using anthropomorphic robot on the one hand, on the other hand, also anthropomorphic machine people can be made to provide personalized service for user according to the identity information of user, promote the experience of anthropomorphic robot.Such as: computer can set up multiple account for different users, each user can enter system by the account of oneself, personal settings can be carried out under the account of oneself, anthropomorphic machine people provides personalized service to be also similar for user, anthropomorphic robot, by after the identity of determining user, can be just that different users provides personalized service according to the identity of user.
As a concrete example, suppose to judge that user calls out as anthropomorphic robot by artificial intelligence, anthropomorphic robot collects the position of user by microphone array, then move to intelligently in face of user, and anthropomorphic robot's head oscillation is aimed at user, carry out by camera laggard row recognition of face determination user identity of taking pictures, thus provide desirable interactive means for user or provide personalized service for user.At this moment when user reaches, robot collects picture signal corresponding to This move, then utilize artificial intelligence after carrying out specific aim analysis to user behavior, determine the action of shaking hands, just can initiatively make a stretch of the arm suitable position, display " smiling face's image " corresponding with the scene of the close friend such as " shaking hands " and play the one of the music corresponding with the friendly scene such as " shaking hands " etc., or the multiple combination in above-mentioned various feedback mode, as stretch out the right hand action while display " smiling face's image " to show friendship.Thus promote the man-machine interaction experience of anthropomorphic robot.
Make a more detailed description below in conjunction with the method for concrete example to the embodiment of the present invention.Such as: an indoor environment, the user of anthropomorphic robot calls its name, and anthropomorphic robot captures this sound by the microphone array be arranged on health, and is navigated to the position of user by location algorithm.Anthropomorphic robot is automatic activation after capturing call, power to each ingredient of anthropomorphic robot, expression (namely activating expression) after the display screen display of the head of anthropomorphic robot activates, then anthropomorphic robot is made to move in face of user by the wheel on the chassis of anthropomorphic robot in the mode of two-wheel differential motion, by head rotation to the direction towards user, the camera of head can be taken pictures and identification to user, go out the most appropriate mode mutual with user by existing database retrieval again or provide personalized service for it.Collected by microphone user voice signal or by camera collection to the picture signal of user after, the intention of user is learnt through the process of artificial intelligence and analysis, then just the audio frequency relevant to the intention of user can be exported by audio amplifier, also the picture relevant to the intention of user or video be can export by main display, even can expression and action and user be passed through mutual.Such as say " we have a dance " user, anthropomorphic robot is captured after this voice signal by after its implication of analysis and understanding of artificial intelligence by microphone, the music can liked by audio amplifier broadcasting user, then position and the action of user is captured by camera, move to suitable position and head and arm etc. are swung thereupon, reach the implementation effect of user's " we have a dance " this voice signal, thus the constructional hardware of anthropomorphic robot is organically combined with artificial intelligence, reach the interaction effect good with user.
According to the anthropomorphic robot's control method based on artificial intelligence of the embodiment of the present invention, the voice signal of collection user that can be real-time and/or picture signal, after the analysis of artificial intelligence, independently control anthropomorphic robot carry out corresponding action or the display image relevant to user view or the broadcasting audio frequency etc. relevant with user view, enrich with the interactive means of user, can more intuitively effectively show the achievement of artificial intelligence.In addition, the motion of anthropomorphic robot is the feedback system realization of view-based access control model and the sense of hearing completely, there is anthropomorphic paleocinetic consciousness, be easy to user operation, embody the intelligent of anthropomorphic robot more all sidedly simultaneously, promote the experience of user, in publicity, demonstration and service field, there is very affine effective effect.
Fig. 5 is according to an embodiment of the invention based on the structured flowchart of the Control system of humanoid robot of artificial intelligence.As shown in Figure 5, this control system 500, comprising: receiver module 510, artificial intelligence module 520, control module 530 and feedback module 540.
Specifically, receiver module 510 is for receiving voice signal and/or the picture signal of user's input.The voice signal of user's input can be that user is inputted by microphone; Above-mentioned picture signal can be obtained by camera collection.
Artificial intelligence module 520 is for determining the intention of user according to voice signal and/or picture signal.
Such as: analyzing and processing can be carried out to voice signal and/or picture signal by artificial intelligence, thus determine the intention of user.It should be noted that, can carry out the intention that analyzing and processing determines user by artificial intelligence to any one in voice signal and picture signal, also can be the intention by determining user to both combinations.
Particularly, as shown in Figure 2, the intention of user can be determined according to following two kinds of modes, specifically comprise:
1, speech recognition is carried out to voice signal, and one or more in natural language understanding, semantic analysis, sentiment analysis are carried out to recognition result, to determine the intention of user.Specifically, process for voice signal needs to carry out speech recognition, natural language understanding, semantic analysis, mechanical translation, sentiment analysis etc., by above-mentioned process, anthropomorphic robot is carrying out in mutual process with user, when user says in short time, the implication of the voice that user inputs can be learnt.
2, speech recognition is carried out to voice signal, and one or more in natural language understanding, semantic analysis, sentiment analysis are carried out to recognition result, and combining image signal determines the intention of user.Specifically, when anthropomorphic robot learns the implication of the voice that user inputs, the intention of definitely user can be carried out further combined with the picture signal of user.Such as: when the voice of user's input are " shaking hands ", and identify user by the picture signal of user and give the action of stretching out the right hand, then now can specify user be intended to shake hands with anthropomorphic robot.
In addition, the mode combined by both not only can the intention of definitely user, can also wherein one when not identifying, and is determined the intention of user by another.Such as: when the action of user is ambiguous or when not providing clear and definite action, " can be shaken hands " intention determining user by the voice identified.Equally, when voice signal is bad, when can not identify voice signal, then can determine that user goes out the intention of user, the intention of namely shaking hands according to " user stretches out the action of the right hand " of the user recognized in picture signal.
As shown in Figure 3, the intention of user can also be determined according to other two kinds of modes, specifically comprise:
3, image recognition is carried out to picture signal, determine the user in picture signal, and determine the limb action of user according to the action difference of the user in picture signal between multiple image, and determine the intention of user according to the limb action of user.Specifically, (as 2 seconds) gather the vision signal (i.e. multiple continuous print image) of user in a short period of time, then the limb action of user is determined according to the action difference of user in multiple continuous print picture signal, also for " user stretches out the action of the right hand ", then determine that the limb action of user is " action that user stretches out the right hand ", thus determine being intended to " shaking hands " of user.
4, image recognition is carried out to picture signal, determine the user in picture signal, and determine the limb action of user according to the action difference of the user in picture signal between multiple image, and determine the intention of user according to the limb action of user and/or voice signal.Specifically, after being determined that by picture signal the limb action of user is " action that user stretches out the right hand ", in conjunction with voice signal, the implication identifying voice signal is " shaking hands ", then both combine can the intention of definitely user.
In addition, the mode combined by both not only can the intention of definitely user, can also wherein one when not identifying, and is determined the intention of user by another.Such as: when the action of user is ambiguous or when not providing clear and definite action, " can be shaken hands " intention determining user by the voice identified.Equally, when voice signal is bad, when can not identify voice signal, then can determine that user goes out the intention of user, the intention of namely shaking hands according to " user stretches out the action of the right hand " of the user recognized in picture signal.
Control module 530 is for processing the intention of user.
Feedback module 540 for the result of processing module is fed back to described user with the multi-modal way of output, the multi-modal way of output comprise in the action way of output of anthropomorphic robot, image or video frequency output mode and the audio frequency way of output one or more.
In one embodiment of the invention, result is fed back to user with the multi-modal way of output to comprise: control anthropomorphic robot and perform the action corresponding with the intention of user; And/or; Show the expression relevant to the intention of user; And/or carry out the image demonstration relevant to the intention of user or video display; And/or play the audio frequency relevant to the intention of user.
Such as: when anthropomorphic robot determines being intended to " shaking hands " of user, then can control the one that anthropomorphic robot performs action, the display of stretching out its right hand " smiling face's image " corresponding with the scene of the close friend such as " shaking hands ", plays the music corresponding with the friendly scene such as " shaking hands " etc. intelligently, or the multiple combination in above-mentioned various feedback mode, as stretch out the right hand action while display " smiling face's image " to show friendship.Thus promote the man-machine interaction experience of anthropomorphic robot.
As shown in Figure 4, in one embodiment of the invention, artificial intelligence module 520 before the intention determining user according to voice signal and/or picture signal, also for:
S401: detect user and whether call out anthropomorphic robot.Such as: set a name to anthropomorphic robot in advance, as " little Bai ", then, as user's calling " little Bai ", anthropomorphic robot carries out speech recognition by the mode of artificial intelligence to " little Bai " that user calls out, thus make anthropomorphic robot understand its implication, just know whether as calling out oneself.
S402: if then activate anthropomorphic robot, and carry out auditory localization to determine the position of user according to the calling of user, and control anthropomorphic robot moves in face of user.
Specifically, if anthropomorphic robot determines that user calls out oneself really, then activate, each functional module being such as anthropomorphic robot by battery etc. powers on, and makes anthropomorphic robot be in state of activation.In one embodiment of the invention, in order to promote man-machine interaction experience, when anthropomorphic robot determines that user is when calling out it, the activation expression of anthropomorphic robot can be shown, activate expression can be predefined, just can as long as can reflect that robot is activated intuitively.Thus, promote man-machine interaction experience further.
In addition, auditory localization is carried out to determine the position of user according to the calling of user.The sound around anthropomorphic robot is detected by ManyEars auditory localization technology, particularly, sound-source signal can be gathered according to microphone array, effective sound signal detection can be carried out to sound-source signal afterwards, and by ManyEars technology, the multi-acoustical detected be carried out being separated to obtain multiple independently sound source.Wherein, the term " at least one " of the present embodiment can be regarded as one or more.Further, by the auditory localization computing in ManyEars technology, certain sound source above-mentioned is positioned, thus determine the position of user.
After the position determining user, just can automatically control anthropomorphic robot and move in face of user.Further, whether control module 530 also moves in face of user for detecting anthropomorphic robot; If not, then control anthropomorphic robot further to carry out moving until reach in face of user.Thus closed loop is formed to the motion control of anthropomorphic robot, ensure that the motion of anthropomorphic robot is more accurately reliable.
In one embodiment of the invention, control module 530 also for control anthropomorphic robot move to user in face of time, by the camera steering surface of anthropomorphic robot to the direction of user, to take pictures to user, and carry out recognition of face according to the picture signal of user, to determine the identity information of user.Like this, can prevent disabled user from using anthropomorphic robot on the one hand, on the other hand, also anthropomorphic machine people can be made to provide personalized service for user according to the identity information of user, promote the experience of anthropomorphic robot.Such as: computer can set up multiple account for different users, each user can enter system by the account of oneself, personal settings can be carried out under the account of oneself, anthropomorphic machine people provides personalized service to be also similar for user, anthropomorphic robot, by after the identity of determining user, can be just that different users provides personalized service according to the identity of user.
As a concrete example, suppose to judge that user calls out as anthropomorphic robot by artificial intelligence, anthropomorphic robot collects the position of user by microphone array, then move to intelligently in face of user, and anthropomorphic robot's head oscillation is aimed at user, carry out by camera laggard row recognition of face determination user identity of taking pictures, thus provide desirable interactive means for user or provide personalized service for user.At this moment when user reaches, robot collects picture signal corresponding to This move, then utilize artificial intelligence after carrying out specific aim analysis to user behavior, determine the action of shaking hands, just can initiatively make a stretch of the arm suitable position, display " smiling face's image " corresponding with the scene of the close friend such as " shaking hands " and play the one of the music corresponding with the friendly scene such as " shaking hands " etc., or the multiple combination in above-mentioned various feedback mode, as stretch out the right hand action while display " smiling face's image " to show friendship.Thus promote the man-machine interaction experience of anthropomorphic robot.
Make a more detailed description below in conjunction with the system of concrete example to the embodiment of the present invention.Such as: an indoor environment, the user of anthropomorphic robot calls its name, and anthropomorphic robot captures this sound by the microphone array be arranged on health, and is navigated to the position of user by location algorithm.Anthropomorphic robot is automatic activation after capturing call, power to each ingredient of anthropomorphic robot, expression (namely activating expression) after the display screen display of the head of anthropomorphic robot activates, then anthropomorphic robot is made to move in face of user by the wheel on the chassis of anthropomorphic robot in the mode of two-wheel differential motion, by head rotation to the direction towards user, the camera of head can be taken pictures and identification to user, go out the most appropriate mode mutual with user by existing database retrieval again or provide personalized service for it.Collected by microphone user voice signal or by camera collection to the picture signal of user after, the intention of user is learnt through the process of artificial intelligence and analysis, then just the audio frequency relevant to the intention of user can be exported by audio amplifier, also the picture relevant to the intention of user or video be can export by main display, even can expression and action and user be passed through mutual.Such as say " we have a dance " user, anthropomorphic robot is captured after this voice signal by after its implication of analysis and understanding of artificial intelligence by microphone, the music can liked by audio amplifier broadcasting user, then position and the action of user is captured by camera, move to suitable position and head and arm etc. are swung thereupon, reach the implementation effect of user's " we have a dance " this voice signal, thus the constructional hardware of anthropomorphic robot is organically combined with artificial intelligence, reach the interaction effect good with user.
According to the Control system of humanoid robot based on artificial intelligence of the embodiment of the present invention, the voice signal of collection user that can be real-time and/or picture signal, after the analysis of artificial intelligence, independently control anthropomorphic robot carry out corresponding action or the display image relevant to user view or the broadcasting audio frequency etc. relevant with user view, enrich with the interactive means of user, can more intuitively effectively show the achievement of artificial intelligence.In addition, the motion of anthropomorphic robot is the feedback system realization of view-based access control model and the sense of hearing completely, there is anthropomorphic paleocinetic consciousness, be easy to user operation, embody the intelligent of anthropomorphic robot more all sidedly simultaneously, promote the experience of user, in publicity, demonstration and service field, there is very affine effective effect.
It should be noted that, the specific implementation of the specific implementation of the Control system of humanoid robot based on artificial intelligence of the embodiment of the present invention and the anthropomorphic robot's control method based on artificial intelligence of the embodiment of the present invention is similar, specifically refer to the description of method part, in order to reduce redundancy, do not repeat herein.
Further, the invention discloses a kind of anthropomorphic robot, comprising: the Control system of humanoid robot based on artificial intelligence described in any one embodiment above-mentioned.This anthropomorphic robot can be real-time the voice signal of collection user and/or picture signal, after the analysis of artificial intelligence, independently control anthropomorphic robot carry out corresponding action, enrich with the interactive means of user.In addition, the motion of anthropomorphic robot is the feedback system realization of view-based access control model and the sense of hearing completely, has anthropomorphic paleocinetic consciousness, is easy to user operation, embody the intelligent of anthropomorphic robot more all sidedly simultaneously, promote the experience of user.
As shown in Figure 6, as a concrete example, anthropomorphic robot mainly comprises the requirement in following four in structural design: the ability to express of have the locomitivity of omnidirectional, have the multivariant manner of presentation, have vision and the mode such as the sense of hearing and expression, voice, video.
The composition structural drawing of anthropomorphic robot as shown in Figure 6, is respectively head assembly, skeleton shell, control system, energy resource system and chassis assembly, sets forth product structure design and the functional realiey of five parts below respectively.
One, head assembly
Head assembly comprise head skeleton and head install expression display screen, audio amplifier and the first-class mechanism that makes a video recording, one can carry out video acquisition function, as the input element of visual feedback, gather the information such as the position of user, action and expression, by analysis, feed back to output again.Carry the equipment of audio amplifier, expression display screen on the other hand, the output of audio frequency and expression can be carried out.
Two, skeleton shell
The skeleton structure of anthropomorphic robot is equivalent to the bone of the mankind, is the main load part of robot and the mounting structure providing each appliance arrangement of robot.Shell mechanism is equivalent to the skin of the mankind, can provide the outward appearance that one, robot is more anthropomorphic, thus makes user more natural when exchanging with robot.Simultaneously at the correct position of shell, point two-layer uniform eight microphone arrays, as the input element of audio feedback, collect the sound of user, through the comparative analysis of microphone array, through calculating the relative position of user, thus provide foundation for the outgoing position of motion.
Three, control system
Control system is the core of robot architecture, comprises mainboard, drive plate and topworks three chief components.Action command, receiving after artificial intelligence analysis sends instruction, can be converted into control signal, make steering wheel carry out the output of speed and position, realize the motion of robot by drive plate by mainboard on the one hand; Other interactive instructions can be converted on the other hand and control the signal such as audio amplifier and screen, thus realize the mutual of the aspects such as robot audio frequency, video and expression.
Four, energy resource system
Energy resource system provides energy source and power for whole robot, comprises 36V40AH high-capacity lithium battery, power transformation module, power switch, suddenly stops air switch and route bus etc., ensures that each position of robot can realize the function responded.
Five, chassis assembly
Chassis assembly comprises the parts such as high-power brushless motor, chassis skeleton, driving wheel and engaged wheel.Energy resource system drives driving wheel to move by brushless electric machine after powering, and adopts the motion planning mode of two-wheel differential, makes robot can omnidirectional moving.Play a supporting role with engaged wheel simultaneously, keep the centre of gravity place of robot to drop between driving wheel and engaged wheel, keep the whole machine balancing of robot.
Such as: an indoor environment, the user of anthropomorphic robot calls its name, and anthropomorphic robot captures this sound by the microphone array be arranged on health, and is navigated to the position of user by location algorithm.Anthropomorphic robot is automatic activation after capturing call, power to each ingredient of anthropomorphic robot, expression (namely activating expression) after the display screen display of the head of anthropomorphic robot activates, then anthropomorphic robot is made to move in face of user by the wheel on the chassis of anthropomorphic robot in the mode of two-wheel differential motion, by head rotation to the direction towards user, the camera of head can be taken pictures and identification to user, go out the most appropriate mode mutual with user by existing database retrieval again or provide personalized service for it.Collected by microphone user voice signal or by camera collection to the picture signal of user after, the intention of user is learnt through the process of artificial intelligence and analysis, then just the audio frequency relevant to the intention of user can be exported by audio amplifier, also the picture relevant to the intention of user or video be can export by main display, even can expression and action and user be passed through mutual.Such as say " we have a dance " user, anthropomorphic robot is captured after this voice signal by after its implication of analysis and understanding of artificial intelligence by microphone, the music can liked by audio amplifier broadcasting user, then position and the action of user is captured by camera, move to suitable position and head and arm etc. are swung thereupon, reach the implementation effect of user's " we have a dance " this voice signal, thus the constructional hardware of anthropomorphic robot is organically combined with artificial intelligence, reach the interaction effect good with user.
According to the anthropomorphic robot of the embodiment of the present invention, the voice signal of collection user that can be real-time and/or picture signal, after the analysis of artificial intelligence, independently control anthropomorphic robot carry out corresponding action or the display image relevant to user view or the broadcasting audio frequency etc. relevant with user view, enrich with the interactive means of user, can more intuitively effectively show the achievement of artificial intelligence.In addition, the motion of anthropomorphic robot is the feedback system realization of view-based access control model and the sense of hearing completely, there is anthropomorphic paleocinetic consciousness, be easy to user operation, embody the intelligent of anthropomorphic robot more all sidedly simultaneously, promote the experience of user, in publicity, demonstration and service field, there is very affine effective effect.
In addition, according to the anthropomorphic robot of the embodiment of the present invention other form and effect be all known for the ordinary skill in the art, in order to reduce redundancy, do not repeat herein.
In describing the invention, it will be appreciated that, term " " center ", " longitudinal direction ", " transverse direction ", " length ", " width ", " thickness ", " on ", D score, " front ", " afterwards ", " left side ", " right side ", " vertically ", " level ", " top ", " end " " interior ", " outward ", " clockwise ", " counterclockwise ", " axis ", " radial direction ", orientation or the position relationship of the instruction such as " circumference " are based on orientation shown in the drawings or position relationship, only the present invention for convenience of description and simplified characterization, instead of indicate or imply that the device of indication or element must have specific orientation, with specific azimuth configuration and operation, therefore limitation of the present invention can not be interpreted as.
In addition, term " first ", " second " only for describing object, and can not be interpreted as instruction or hint relative importance or imply the quantity indicating indicated technical characteristic.Thus, be limited with " first ", the feature of " second " can express or impliedly comprise at least one this feature.In describing the invention, the implication of " multiple " is at least two, such as two, three etc., unless otherwise expressly limited specifically.
In the description of this instructions, specific features, structure, material or feature that the description of reference term " embodiment ", " some embodiments ", " example ", " concrete example " or " some examples " etc. means to describe in conjunction with this embodiment or example are contained at least one embodiment of the present invention or example.In this manual, to the schematic representation of above-mentioned term not must for be identical embodiment or example.And the specific features of description, structure, material or feature can combine in one or more embodiment in office or example in an appropriate manner.In addition, when not conflicting, the feature of the different embodiment described in this instructions or example and different embodiment or example can carry out combining and combining by those skilled in the art.
Describe and can be understood in process flow diagram or in this any process otherwise described or method, represent and comprise one or more for realizing the module of the code of the executable instruction of the step of specific logical function or process, fragment or part, and the scope of the preferred embodiment of the present invention comprises other realization, wherein can not according to order that is shown or that discuss, comprise according to involved function by the mode while of basic or by contrary order, carry out n-back test, this should understand by embodiments of the invention person of ordinary skill in the field.
In flow charts represent or in this logic otherwise described and/or step, such as, the sequencing list of the executable instruction for realizing logic function can be considered to, may be embodied in any computer-readable medium, for instruction execution system, device or equipment (as computer based system, comprise the system of processor or other can from instruction execution system, device or equipment instruction fetch and perform the system of instruction) use, or to use in conjunction with these instruction execution systems, device or equipment.With regard to this instructions, " computer-readable medium " can be anyly can to comprise, store, communicate, propagate or transmission procedure for instruction execution system, device or equipment or the device that uses in conjunction with these instruction execution systems, device or equipment.The example more specifically (non-exhaustive list) of computer-readable medium comprises following: the electrical connection section (electronic installation) with one or more wiring, portable computer diskette box (magnetic device), random access memory (RAM), ROM (read-only memory) (ROM), erasablely edit ROM (read-only memory) (EPROM or flash memory), fiber device, and portable optic disk ROM (read-only memory) (CDROM).In addition, computer-readable medium can be even paper or other suitable media that can print described program thereon, because can such as by carrying out optical scanning to paper or other media, then carry out editing, decipher or carry out process with other suitable methods if desired and electronically obtain described program, be then stored in computer memory.
Should be appreciated that each several part of the present invention can realize with hardware, software, firmware or their combination.In the above-described embodiment, multiple step or method can with to store in memory and the software performed by suitable instruction execution system or firmware realize.Such as, if realized with hardware, the same in another embodiment, can realize by any one in following technology well known in the art or their combination: the discrete logic with the logic gates for realizing logic function to data-signal, there is the special IC of suitable combinational logic gate circuit, programmable gate array (PGA), field programmable gate array (FPGA) etc.
Those skilled in the art are appreciated that realizing all or part of step that above-described embodiment method carries is that the hardware that can carry out instruction relevant by program completes, described program can be stored in a kind of computer-readable recording medium, this program perform time, step comprising embodiment of the method one or a combination set of.
In addition, each functional unit in each embodiment of the present invention can be integrated in a processing module, also can be that the independent physics of unit exists, also can be integrated in a module by two or more unit.Above-mentioned integrated module both can adopt the form of hardware to realize, and the form of software function module also can be adopted to realize.If described integrated module using the form of software function module realize and as independently production marketing or use time, also can be stored in a computer read/write memory medium.
The above-mentioned storage medium mentioned can be ROM (read-only memory), disk or CD etc.Although illustrate and describe embodiments of the invention above, be understandable that, above-described embodiment is exemplary, can not be interpreted as limitation of the present invention, and those of ordinary skill in the art can change above-described embodiment within the scope of the invention, revises, replace and modification.