Embodiment
Although mobile terminal apparatus now can provide voice system, link up with mobile terminal apparatus to allow the user send voice, the user still must start by mobile terminal apparatus itself when starting this voice system.Therefore can't touch immediately mobile terminal apparatus the user, but the situation that voice system is opened often can't satisfy user's demand immediately.For this reason, the present invention proposes device and the corresponding method thereof that a kind of assistant voice system opens, and allows user opening voice system more easily.In order to make content of the present invention more clear, below the example that really can implement according to this as the present invention especially exemplified by embodiment.
Fig. 1 is the calcspar of the speech control system that illustrates according to one embodiment of the invention.Please refer to Fig. 1, speech control system 100 comprisesauxiliary actuating apparatus 110,mobile terminal apparatus 120 and server 130.In the present embodiment,auxiliary actuating apparatus 110 can start the voice system ofmobile terminal apparatus 120 by wireless signal transmission, makesmobile terminal apparatus 120 link up according to voice signal andserver 130.
Specifically,auxiliary actuating apparatus 110 comprises the firstwireless transport module 112 andtrigger module 114, and whereintrigger module 114 is coupled to the first wireless transport module 112.The firstwireless transport module 112 is for example to support wireless compatible authentication (Wireless delity, Wi-Fi), global intercommunication microwave access (Worldwide Interoperability for Microwave Access, WiMAX), bluetooth (Bluetooth), ultra broadband (ultra-wideband, UWB) or radio-frequency (RF) identification (Radio-frequencyidentification, the device of communication protocol such as RFID), it can send wireless signal transmission, to correspond to each other and to set up wireless link with another wireless transportmodule.Trigger module 114 is such as being button, button etc.In the present embodiment, after the user presses thesetrigger module 114 generation one triggering signals, the firstwireless transport module 112 receives this triggering signal and starts, this moment, the firstwireless transport module 112 can send wireless signal transmission, and transmitted this wireless signal transmission tomobile terminal apparatus 120 by the first wireless transport module 112.In one embodiment, above-mentionedauxiliary actuating apparatus 110 can be a bluetooth earphone.
Although it should be noted that the earphone/of some hand-free also has the design that startsmobile terminal apparatus 120 some function at present, in another embodiment of the present invention, auxiliary actuatingapparatus 110 can be different from above-mentioned earphone/.Above-mentioned earphone/by with the line of mobile terminal apparatus, listen/converse to replace the earphone/ onmobile terminal apparatus 120, start-up performance is additional design, but the application'sauxiliary actuating apparatus 110 " only " is used for opening the voice system ofmobile terminal apparatus 120, do not have the function of listening/conversing, therefore inner circuit design can be simplified, cost is also lower.In other words, for above-mentioned hands-free headsets/microphone, auxiliary actuatingapparatus 110 is other devices, and namely the user may possess earphone/and the application's of hand-freeauxiliary actuating apparatus 110 simultaneously.
In addition, the body of above-mentioned auxiliary actuating apparatus 110 can be the articles for use that the user can reach conveniently, ornaments such as ring, wrist-watch, earrings, necklace, glasses, be various Portable article, or installation component, for example for being disposed at the driving accessory on steering wheel, be not limited to above-mentioned.That is to say, auxiliary actuating apparatus 110 is the device of " life-stylize ", by the setting of built-in system, allows the user can touch easily trigger module 114, with the opening voice system.For instance, when the body of auxiliary actuating apparatus 110 was ring, user's moveable finger trigger module 114 of pressing ring easily was triggered it.On the other hand, the body when auxiliary actuating apparatus 110 is that when being disposed at the device of driving accessory, the user also can trigger the trigger module 114 of driving accessory device during the road easily.In addition, compared to the discomfort of wearing earphone/and listening/converse, use the application's auxiliary actuating apparatus 110 voice system in mobile terminal apparatus 120 can be opened, even and then open sound amplification function (then will describe in detail), make the user need not wear earphone/, still can directly listen/converse by mobile terminal apparatus 120.In addition, for the user, the article of auxiliary actuating apparatus 110 for originally wearing or use of these " life-stylizes " therefore do not have in the use the uncomfortable or problem of discomfort, namely do not need the adaptation of taking time.For instance, when the user cooks in the kitchen, in the time of need to dialing the mobile phone that is positioned over the parlor, suppose its wear have ring, the auxiliary actuating apparatus of the present invention 110 of necklace or wrist-watch body, just can touch ring, necklace or wrist-watch with the opening voice system with inquiry friend recipe details.Also can reach above-mentioned purpose although partly have at present the earphone/of start-up performance, but in the process of at every turn cooking, be not all to need to call to consult the friend at every turn, therefore for the user, wear at any time earphone/and cook, can say suitable inconvenience in order to controlling mobile terminal apparatus at any time.
In other embodiments,auxiliary actuating apparatus 110 also may be configured withwireless charging battery 116, in order to drive the first wireless transport module 112.Furthermore,wireless charging battery 116 comprisesbattery unit 1162 andwireless charging module 1164, and whereinwireless charging module 1164 is coupled to battery unit 1162.At this,wireless charging module 1164 can receive the energy of supplying from a wireless power supply (not illustrating), and is that electric power comesbattery unit 1162 chargings with this power conversion.Thus, the firstwireless transport module 112 ofauxiliary actuating apparatus 110 can charge bywireless charging battery 116 expediently.
On the other hand, mobileterminal apparatus 120 is for example mobile phone (Cellphone), personal digital assistant (Personal Digital Assistant, PDA) mobile phone, smart mobile phone (Smart phone), or palmtop computer (Pocket PC), Tablet PC (Tablet PC) or mobile computer of communication software etc. are installed.Mobileterminal apparatus 120 can be any portable (Portable) mobile device that possesses communication function, does not limit its scope at this.In addition, mobileterminal apparatus 120 can use Android operating system, microsoft operating system, Android operating system, (SuSE) Linux OS etc., is not limited to above-mentioned.
Mobileterminal apparatus 120 comprises the secondwireless transport module 122, the secondwireless transport module 122 can be complementary with the firstwireless transport module 112 ofauxiliary actuating apparatus 110, and adopt corresponding wireless communication protocol (communication protocols such as wireless compatible authentication, global intercommunication microwave access, bluetooth, ultra-wideband communication protocol or radio-frequency (RF) identification), use with the firstwireless transport module 112 and set up wireless link.It should be noted that " first " described hereinwireless transport module 112, " second "wireless transport module 122 are disposed at different devices in order to wireless transport module to be described, is not to limit the present invention.
In other embodiments, mobileterminal apparatus 120 also comprisesvoice system 121, thisvoice system 121 is coupled to the secondwireless transport module 122, therefore after the user triggers thetrigger module 114 ofauxiliary actuating apparatus 110, can wirelessly startvoice system 121 by the firstwireless transport module 112 and the second wireless transport module 122.In one embodiment, thisvoice system 121 can comprisespeech sample module 124, voicesynthetic module 126 and voice output interface 127.Speech sample module 124 is in order to receiving the voice signal from the user, and thisspeech sample module 124 is such as the device that is the audio reception such as microphone (Microphone).Voicesynthetic module 126 can be inquired about a speech database for speech synthesis, and this speech database for speech synthesis be for example record word with and the information of corresponding voice, make voicesynthetic module 126 can find out voice corresponding to the specific character message, so that message language is carried out phonetic synthesis.Afterwards, voicesynthetic module 126 can with synthetic voice by 127 outputs of voice output interface, be used to play and give the user.Above-mentionedvoice output interface 127 is such as being loudspeaker or earphone etc.
In addition, mobileterminal apparatus 120 also may be configured with communication module 128.Communication module 128 is for example can transmit and the element that receives wireless signal, as radio-frequency (RF) transceiver.Furthermore,communication module 128 can allow the user answer or call or use by mobileterminal apparatus 120 other services that telecommunication operator provides.In the present embodiment,communication module 128 can receive response message fromserver 130 by the Internet, and set up conversation line between mobileterminal apparatus 120 and at least one electronic installation according to this response message, wherein said electronic installation for example is another mobile terminal apparatus (not illustrating).
Server 130 is such as being the webserver or cloud server etc., and it has speech understanding module 132.In the present embodiment,speech understanding module 132 comprisesvoice identification module 1322 andspeech processing module 1324, and whereinspeech processing module 1324 is coupled to voice identification module 1322.At this,voice identification module 1322 can receive the voice signal that transmits fromspeech sample module 124, voice signal is converted to a plurality of segmentations semantic (such as vocabulary or words and expressions etc.).1324 of speech processing module can parse according to these segmentations semantemes mean (such as intention, time, place etc.) of the semantic representatives of these segmentations, and then judge the meaning represented in above-mentioned voice signal.In addition,speech processing module 1324 also can produce corresponding response message according to the result of resolving.In the present embodiment,speech understanding module 132 can be come implementation by the hardware circuit that or several gates combine, and can be also to come implementation with computer program code.It is worth mentioning that, in another embodiment,speech understanding module 132 is configurable in mobileterminal apparatus 220,speech control system 200 as shown in Figure 2.
The method that the above-mentioned speech control system 100 plain language sounds of below namely arranging in pairs or groups are controlled.Fig. 3 is the flow chart of the speech control method that illustrates according to one embodiment of the invention.Please be simultaneously with reference to Fig. 1 and Fig. 3, in step 302, auxiliary actuating apparatus 110 sends wireless signal transmission to mobile terminal apparatus 120.Detailed explanation is that when the first wireless transport module 112 of auxiliary actuating apparatus 110 was triggered because receiving a triggering signal, this auxiliary actuating apparatus 110 can send wireless signal transmission to mobile terminal apparatus 120.Particularly, when the trigger module 114 in auxiliary actuating apparatus 110 is pressed by the user, this moment, trigger module 114 meetings be triggered because of triggering signal, and make the first wireless transport module 112 send wireless signal transmission to the second wireless transport module 122 of mobile terminal apparatus 120, use making the first wireless transport module 112 link by wireless communication protocol and the second wireless transport module 122.Above-mentioned auxiliary actuating apparatus 110 only is used for opening the voice system of mobile terminal apparatus 120, does not have the function of listening/conversing, therefore inner circuit design can be simplified, cost is also lower.In other words, for the additional hands-free headsets/microphone of general mobile terminal apparatus 120, auxiliary actuating apparatus 110 is another devices, and namely the user may possess earphone/and the application's of hand-free auxiliary actuating apparatus 110 simultaneously.
It is worth mentioning that, the body of above-mentionedauxiliary actuating apparatus 110 can be the articles for use that the user can reach conveniently, various Portable article such as ring, wrist-watch, earrings, necklace, glasses, or installation component, for example for being disposed at the driving accessory on steering wheel, be not limited to above-mentioned.That is to say,auxiliary actuating apparatus 110 is the device of " life-stylize ", by the setting of built-in system, allows the user can touch easily triggermodule 114, with opening voice system 121.Therefore, use the application'sauxiliary actuating apparatus 110voice system 121 in mobileterminal apparatus 120 can be opened, even and then open sound amplification function (then will describe in detail), make the user need not wear earphone/, still can directly listen/converse by mobile terminal apparatus 120.In addition, for the user, the article ofauxiliary actuating apparatus 110 for originally wearing or use of these " life-stylizes " are not therefore have in the use the uncomfortable or problem of discomfort.
In addition, the firstwireless transport module 112 and the secondwireless transport module 122 all can be in sleep pattern or mode of operation.Wherein, it is closed condition that sleep pattern refers to wireless transport module, that is wireless transport module can not receive/the detecting wireless signal transmission, and can't link with other wireless transport module.It is opening that mode of operation refers to wireless transport module, that is wireless transport module detecting wireless signal transmission constantly, or sends at any time wireless signal transmission, and can link with other wireless transport module.At this, whentrigger module 114 is triggered, if the firstwireless transport module 112 is in sleep pattern,trigger module 114 can wake the firstwireless transport module 112 up, make the firstwireless transport module 112 enter mode of operation, and make the firstwireless transport module 112 send wireless signal transmission to the secondwireless transport module 122, and allow the firstwireless transport module 112 link by the secondwireless transport module 122 of wireless communication protocol and mobileterminal apparatus 120.
On the other hand, continue to maintain mode of operation and consume too much electric power for fear of the firstwireless transport module 112, in Preset Time after the firstwireless transport module 112 enters mode of operation (being for example 5 minutes), iftrigger module 114 is not triggered again, the firstwireless transport module 112 can enter sleep pattern from mode of operation, and stops linking with the secondwireless transport module 120 of mobileterminal apparatus 120.
Afterwards, in step 304, the secondwireless transport module 122 of mobileterminal apparatus 120 can receive wireless signal transmission, to start voice system 121.Then, at step S306, when the secondwireless transport module 122 detected wireless signal transmission, mobileterminal apparatus 120 can startvoice system 121, and 121sampling modules 124 of voice system can begin received speech signal, for example " temperature several years today ", " phone Lao Wang.", " ask enquiring telephone number." etc.
At step S308,speech sample module 124 can be sent to above-mentioned voice signal thespeech understanding module 132 inserver 130, to resolve voice signal and to produce response message by speech understanding module 132.Furthermore,voice identification module 1322 inspeech understanding module 132 can receive the voice signal fromspeech sample module 124, and it is semantic that voice signal is divided into a plurality of segmentations,speech processing module 1324 can be carried out speech understanding to above-mentioned segmentation semanteme, to produce in order to respond the response message of voice signal.
In another embodiment of the present invention, mobileterminal apparatus 120 more can receive the response message thatspeech processing module 1324 produces, and perhaps carries out by interior invoice output interface 127 output response messages the operation that response message is assigned according to this.At step S310, the voicesynthetic module 126 of mobileterminal apparatus 120 can receive the response message thatspeech understanding module 132 produces, and carry out phonetic synthesis according to the content in response message (such as vocabulary or words and expressions etc.), and produces voice answer-back.And at step S312,voice output interface 127 can receive and export this voice answer-back.
For example, when the user pressestrigger module 114 inauxiliary actuating apparatus 110, the first 112 of wireless transport modules can send wireless signal transmission to the secondwireless transport module 122, make mobileterminal apparatus 120 start thespeech sample module 124 of voice system 121.At this, suppose that the voice signal from the user is an inquiry sentence, for example " temperature several years today ",speech sample module 124 just can receive and thespeech understanding module 132 that this voice signal is sent inserver 130 is resolved, andspeech understanding module 132 can send back mobileterminal apparatus 120 with resolving the response message that produces.Suppose that the content in response message thatspeech understanding module 132 produces is " 30 ℃ ", voicesynthetic module 126 can synthesize voice answer-back with the message of these " 30 ℃ ", andvoice output interface 127 can should be reported these voice to the user.
In another embodiment, suppose that the voice signal from the user is an imperative sentence, for example " phone Lao Wang.", can pick out this imperative sentence inspeech understanding module 132 and be " dialing to the request of Lao Wang ".In addition,speech understanding module 132 can produce new response message again, and for example " whether PLSCONFM sets aside Lao Wang ", and the response message that this is new is sent to mobile terminal apparatus 120.At this, voicesynthetic module 126 can synthesize voice answer-back by the response message that this is new, and reports in the user by voice output interface 127.Further say, when the user reply sure answer for "Yes" and so on the time, similarly,speech sample module 124 can receive and transmit this voice signal toserver 130, to allowspeech understanding module 132 resolve.Afterspeech understanding module 132 is resolved and finished, just can record a dialing command information at response message, and be sent to mobile terminal apparatus 120.At this moment, the contact information that 128 of communication modules can record according to call database inquires the telephone number of " Lao Wang ", and setting up the conversation line between mobileterminal apparatus 120 and another electronic installation, that is " Lao Wang " given in dialing.
In other embodiments, except above-mentioned speech control system 100, also can utilizespeech control system 200 or other similar systems, carry out above-mentioned method of operation, not be limited with the above embodiments.
In sum, in the speech control system and method for the present embodiment, auxiliary actuating apparatus can wirelessly be opened the phonetic function of mobile terminal apparatus.And, the body of this auxiliary actuating apparatus can be the user conveniently can and the articles for use of " life-stylize ", ornaments such as ring, wrist-watch, earrings, necklace, glasses, be various Portable article, or installation component, for example for being disposed at the driving accessory on steering wheel, be not limited to above-mentioned.Thus, compared to the discomfort of wearing in addition at present hands-free headsets/microphone, will be more convenient with the voice system that the application'sauxiliary actuating apparatus 110 is opened in mobileterminal apparatus 120.
It should be noted that above-mentionedserver 130 with speech understanding module may be the webserver or cloud server, and cloud server may relate to the problem of user's the right of privacy.For example, the user need upload complete address list to cloud server, just can complete as calling, send out the operation relevant to address list such as news in brief.Even cloud server adopt to be encrypted line, and instant biography do not preserve, and the load that still is difficult to eliminate the user is excellent.Accordingly, below provide the method for another kind of speech control and corresponding voice interactive system thereof, mobile terminal apparatus can in the situation that do not upload complete address list, be carried out the interactive voice service with cloud server.In order to make content of the present invention more clear, below the example that really can implement according to this as the present invention especially exemplified by embodiment.
Fig. 4 is the calcspar according to the voice interactive system of one embodiment of the invention.Please refer to Fig. 4, voiceinteractive system 400 can comprisecloud server 410 and mobileterminal apparatus 420, butcloud server 410 and mobileterminal apparatus 420 interconnecting lines.Voiceinteractive system 400 is to carry out the interactive voice service by cloud server 410.That is, come processed voice identification by thecloud server 410 with powerful operational capability, reduce by this data of mobileterminal apparatus 420 and process load, also can promote accuracy and the recognition speed of speech recognition.
In mobileterminal apparatus 420, comprise processingunit 422,communication module 424,voice system 426, memory cell 428.In one embodiment, mobileterminal apparatus 420 also disposes a display unit 430.Wherein, processingunit 422 is coupled tocommunication module 424,voice system 426,memory cell 428 and display unit 430.More store anaddress list 429 inmemory cell 428.
Above-mentionedprocessing unit 422 is for possessing the hardware (such as chipset, processor etc.) of operational capability, in order to control the overall operation of mobile terminal apparatus 420.Processing unit 422 is for example CPU (Central Processing Unit, CPU), or other programmable microprocessors (Microprocessor), digital signal processor (Digital Signal Processor, DSP), Programmable Logic Controller, Application Specific Integrated Circuit (Application Specific Integrated Circuits, ASIC), programmable logic device (Programmable Logic Device, PLD) or other similar devices.
Above-mentionedcommunication module 424 is for example network card, and it can be to link up via wire transmission or wireless transmission and cloud server 410.And above-mentionedvoice system 426 comprises the radio reception devices such as microphone at least, so that sound is converted to electronic signal.Saidmemory cells 428 is for example random access memory (RandomAccess Memory, RAM), read-only memory (Read-Only Memory, ROM), flash memory (Flash memory) or disk storage device (Magnetic disk storage device) etc.Above-mentioneddisplay unit 430 is for example liquid crystal display (Liquid Crystal Display, LCD) or the Touch Screen (touch screen) with touch-control module etc.
On the other hand, thecloud server 410 entity main frame for having powerful operational capability, a super virtual machine that perhaps can be comprised of a group entity main frame uses to carry out large-scale task.At this,cloud server 410 comprises processingunit 412 and communication module 414.At this, thecommunication module 414 ofcloud server 410 is coupled to its processing unit 412.Communication module 414 is in order to link up with thecommunication module 424 of mobile terminal apparatus 420.Communication module 414 is for example network card, and it can be to link up via wire transmission or wireless transmission and mobileterminal apparatus 420.
In addition, processing unit incloud server 410 412 is for having more powerful operational capability, for example by the CPU of multi-core or formed the CPU array by a plurality of CPU.Theprocessing unit 412 ofcloud server 410 for example comprisesspeech understanding module 132 as shown in Figure 1 atleast.Processing unit 412 can be resolved the voice signal that receives from mobileterminal apparatus 420 by the speech understanding module.And the result thatcloud server 410 will be resolved bycommunication module 414 is sent to mobileterminal apparatus 420, makes mobileterminal apparatus 420 be able to carry out corresponding action according to result.
The above-mentioned Fig. 4 that below namely arranges in pairs or groups is illustrated in the exchange of speech flow process of voice interactive system.
Fig. 5 is the schematic diagram according to the voice communication flow process that is used for voice interactive system of one embodiment of the invention.Please simultaneously with reference to Fig. 4 and Fig. 5, in step S501, in mobileterminal apparatus 420, receive the first voice signal byvoice system 426, and in step S503, bycommunication module 424, the first voice signal is sent to cloud server 410.At this, mobileterminal apparatus 420 is such as being by elements such as the microphones invoice system 426 and receive the first voice signal from the user.For instance, suppose that mobileterminal apparatus 420 is mobile phone, the user says " phoning Lao Wang " facing to mobile phone,voice system 426 is receiving this voice signal " phone Lao Wang " after, can be bycommunication module 424 with this voice signal " phone Lao Wang " be sent to cloud server 410.In one embodiment, above-mentionedvoice system 426 can start by Fig. 1 ~ auxiliary actuating apparatus shown in Figure 3.
Then, in step S505, beyond the clouds inserver 410, processingunit 412 utilizes the speech understanding module to resolve the first voice signal, and, in step S507, the communication target thatprocessing unit 412 will be obtained by the first voice signal is sent to mobileterminal apparatus 420 by communication module 414.Content take the first voice signal " is phoned Lao Wang " as example, and theprocessing unit 412 ofcloud server 410 can utilize the speech understanding module to resolve the first voice signal, obtains by this communication instruction and communication target.Namely, the speech understanding module can parse the first voice signal and comprise " making a phone call " and " Lao Wang ", accordingly, theprocessing unit 412 ofcloud server 410 just can be judged communication instruction and be the dialing instruction, and communication target is " Lao Wang ", and is sent to mobileterminal apparatus 420 bycommunication module 414.
Then, in step S509, in mobileterminal apparatus 420, theprocessing unit 422 of mobileterminal apparatus 420 is according to theaddress list 429 in communication targetsearch memory cell 428, and acquisition meets the selective listing of communication target.For example, theprocessing unit 422 of mobileterminal apparatus 420 finds many and has the contact information of " king ", thereby produce selective listing, and be shown indisplay unit 430 in the process of searching address list, selects for the user.
For instance, shown in table 1, search the contact information that meets communication target " Lao Wang " in address list under selective listing for example.In this example, suppose the contact information that finds 4 to meet, and with the coordinator's title in contact information, namely " Wang Congming ", " king five ", " Wang Anshi " and " Wang Wei ", write in selective listing.
Table 1
And if the user speaks facing to mobileterminal apparatus 420, as shown in step S511, mobileterminal apparatus 420 can receive the second voice signal by voice system 426.And when mobileterminal apparatus 420 received the second voice signal, in step S513, mobileterminal apparatus 420 can be sent tocloud server 410 bycommunication module 424 simultaneously with the second voice signal and selective listing.Such as: user after watching selective listing and say " the 1st " or contents such as " Wang Congming " facing to mobileterminal apparatus 420, and when forming the second voice signal, mobileterminal apparatus 420 just can be sent to the second voicesignal cloud server 410 together with selective listing.
In addition, the user also can arbitrarily say other guide, that is to say, no matter the content that the user says why, as long as mobileterminal apparatus 420 receives the second voice signal, just can be sent tocloud server 410 with the second voice signal and selective listing simultaneously.
It is worth mentioning that, in this application, the address list with " complete " is not uploaded tocloud server 410, and only will meet communication target with the form of " selective listing ", is uploaded tocloud server 410 to carry out speech signal analysis for the second time.In other words, only have " part " contact data can be uploaded.In one embodiment, mobileterminal apparatus 420 is uploaded in the selective listing ofcloud server 410 can include only coordinator's title, and does not comprise telephone number or other information.The content of the selective listing of uploading can be set according to user's demand.
In addition, it should be noted that, in this application, the second voice signal and selective listing are sent tocloud server 410 simultaneously, need gradation to resolve each voice signal and each list compared to the communication means that need not upload at present address list, namely a step only comprises an information, and the application's voice switching method is more quick.
Then, inserver 410, processingunit 412 can utilize the speech understanding module to resolve the second voice signal, as shown in step S515 beyond the clouds.For example, utilize the speech understanding module parses to go out the included content of the second voice signal and be " the 3rd ", theprocessing unit 412 ofcloud server 410 just can further remove to compare the 3rd contact information in the selective listing that mobileterminal apparatus 420 receives.Take table 1 as example, the 3rd contact information is " Wang Anshi ".
It should be noted that, the design of thespeech understanding module 132 by as shown in Figure 1, the user does not need the complete content of selective listing of telling as the second voice signal, as " the 1st Wang Congming ", only need tell the content of part selective listing, as the second voice signal, and the selective listing of arranging in pairs or groups simultaneously is uploaded to thespeech understanding module 132 of cloud server, can parse select target as " the 1st " or " Wang Congming ".In other words, the selective listing content comprises a plurality of project information, and each project information has the content (as: name, telephone number etc.) of numbering and corresponding this numbering at least, and the second voice signal comes from partial content or the numbering of corresponding this numbering.
Afterwards, in step S517,cloud server 410 is sent to mobileterminal apparatus 420 by itscommunication module 414 with communication instruction and select target.And in other embodiments,cloud server 410 also can namely first transmit communication instruction to mobileterminal apparatus 420 storages after step S505 has resolved the first voice signal, transmits afterwards select target again, does not limit the delivery time point of communication instruction at this.
After mobile terminal apparatus 420 received communication instruction and select target, in step S519, mobile terminal apparatus 420 was by 422 pairs of select targets of its processing unit, the communication operation that the executive communication instruction is corresponding.Above-mentioned communication instruction is such as needing to use the instruction of this address list content for dialing instruction or citation instruction etc., and communication instruction is to be obtained based on the first voice signal by cloud server 410.For example, the content of supposing the first voice signal is " phoning Lao Wang ", and cloud server 410 is judged communication instruction and is the dialing instruction by " making a phone call ".Again for example, suppose that the content of the first voice signal is " passing news in brief to Lao Wang ", cloud server 410 is judged communication instruction for summoning instruction by " biography news in brief ".In addition, above-mentioned select target is based on the second voice signal and selective listing and obtain by cloud server 410.Take the selective listing shown in above-mentioned table 1 as example, the content of supposing the second voice signal is " the 3rd ", and cloud server 410 just can be judged select target and is " Wang Anshi ".For example, call to select target, or start a citation interface, to transmit news in brief to select target.
It should be noted that mobileterminal apparatus 420 can include only coordinator's title in the selective listing that above-mentioned steps S509 obtains, and do not comprise telephone number or other information.Therefore, when mobileterminal apparatus 420 whencloud server 410 receives communication instruction and select target, theprocessing unit 422 of mobileterminal apparatus 420 can take out the telephone number of corresponding selection target in address list, and comes communication operation corresponding to executive communication instruction according to telephone number.
In addition, in other embodiments, mobileterminal apparatus 420 also can comprise coordinator's title and telephone number simultaneously in the selective listing that above-mentioned steps S509 obtains, perhaps also can comprise other information.Therefore, in step S515, theprocessing unit 412 ofcloud server 410 just can be based on the second voice signal and selective listing, and obtains the telephone number of select target, and in step S517, communication instruction and telephone number are sent to mobile terminal apparatus 420.Accordingly, in step S519, mobileterminal apparatus 420 comes communication operation corresponding to executive communication instruction according to telephone number.
In sum, the application's utilization is uploaded simultaneously mode to the cloud server with powerful operational capability of selective listing that the first voice produce, select target that the second voice signal produces and is carried out speech understanding program, and this selective listing only comprises the address list of part.Therefore, the application's speech control system can be possessed higher treatment efficiency and better fail safe simultaneously.
On the other hand, although it should be noted that above-mentioned auxiliary actuating apparatus has solved the user and can't touch immediately mobile terminal apparatus, need to use the problem of voice system, make the user can pass through the speech understanding technology, allow user and mobile terminal apparatus carry out question and answer.Yet, for the situation that needs public address system to open, need be still to start expansion by mobile terminal apparatus itself at present, therefore can't satisfy user's demand immediately.For this reason, the present invention proposes a kind of method and corresponding device thereof of opening public address system, allows the user can open more easily public address system.In order to make content of the present invention more clear, below the example that really can implement according to this as the present invention especially exemplified by embodiment.
On the other hand, although it should be noted that above-mentioned auxiliary actuating apparatus has solved the user and can't touch immediately mobile terminal apparatus, need to use the voice system problem, make the user can pass through the speech understanding technology, allow user and mobile terminal apparatus carry out question and answer.Yet, for the situation that needs sound amplification function to open, still need at present to start sound amplification function by mobile terminal apparatus itself, when the user can't touch mobile terminal apparatus immediately, but in the time of need making sound amplification function, the design that needs at present to start by mobile terminal apparatus itself will cause user's inconvenience.For this reason, the present invention proposes a kind of method and corresponding device thereof of opening sound amplification function, allows the user can open more easily sound amplification function.In order to make content of the present invention more clear, below the example that really can implement according to this as the present invention especially exemplified by embodiment.
Fig. 6 is the system schematic according to the mobile terminal apparatus of one embodiment of the invention.Please refer to Fig. 6, in the present embodiment, mobile terminal apparatus 600 comprises voice system, input unit 620, pulls and connects unit 630, receiver 640, public address equipment 650 and processing unit 660.In another embodiment of the present invention, mobile terminal apparatus 600 also can comprise earphone 670.Mobile terminal apparatus 600 can be mobile phone or other similar electronic installations, and it is similar to the mobile terminal apparatus 120 of Fig. 1, and its detailed content can with reference to aforementioned content, not repeat them here.Processing unit 660 couples speech sample module 610, input unit 620, pulls and connects unit 630, receiver 640, public address equipment 650, earphone 670.Voice system comprises speech sample module 610, and this speech sample module 610 is converted to input speech signal SAI with sound, and above-mentioned speech sample module 610 can be microphone or similar electronic component.In other words, speech sample module 610 can be considered the part of voice system, and this described voice system is similar to the voice system 121 of Fig. 1, and its detailed content can with reference to aforementioned content, not repeat them here.Input unit 620 corresponding users' operation provides input operation signal SIO, and input unit 620 can be keyboard, contact panel or similar electronic component.Pull and connect unit 630 and pull and connect function in order to be controlled by processing unit 660 execution.Receiver 640, public address equipment 650, earphone 670 are converted to sound in order to the output voice signal SAO that processing unit 660 is provided, therefore can be considered the voice output interface.Above-mentioned public address equipment 650 is such as being loud speaker etc.Above-mentioned earphone 670 can be wired earphone and wireless headset at least one of them.
As from the foregoing, the physical button that the unlatching of phonetic function can be by pressing mobile communications device, control screen or utilize auxiliary actuating apparatus of the present invention.In the situation that the hypothesis phonetic function is opened, when the user talks facing to mobile terminal apparatus 600, sound can be converted to input speech signal SAI by speech sample module 610, processing unit 660 can be according to input speech signal SAI, when carrying out content matching for information such as the coordinator's title in address list or telephone numbers, when the information in address list conforms to input speech signal SAI, 660 of processing units can open pull and connect unit 630 pull and connect function and public address equipment 650, so that after connecting, the user can with coordinator's conversation.Detailed explanation is that processing unit 660 can be converted to input speech signal SAI one input word string, and will input the information such as a plurality of coordinator's titles in word string and address list, a plurality of telephone numbers relatively.When the input word string met one of them of the information such as these coordinator's titles, these telephone numbers, processing unit 660 was opened the function of pulling and connecting of pulling and connecting unit 630.On the contrary, when the input word string did not meet these coordinator's titles and these telephone numbers, processing unit 660 was not opened the function of pulling and connecting of pulling and connecting unit 630.
In other words, in the present embodiment, when the content matching inprocessing unit 660 confirmation input speech signal SAI and address list, processingunit 660 can provide enabling signal, in order to automatically open the conversation sound amplification function of mobile terminal apparatus 100.Specifically, processingunit 660 can supply enabling signal topublic address equipment 650 by automatic lifting, and input speech signal SAI is converted to conversation transmits data DTC, and transmit conversation transmission data DTC to coordinator (another mobile terminal apparatus does not illustrate) by pulling and connecting unit 630.Simultaneously, processingunit 660 can receive conversation receive data DRC by pulling and connectingunit 630, and provide output audio signal SAO topublic address equipment 650 according to conversation receive data DRC, so that output audio signal SAO is converted to sound, and in the mode that amplifies with voice output.
It is worth mentioning that, in the mode of present startup sound amplification function, be still and adopt the mode that starts by mobile terminal apparatus itself to carry out, can't touch immediately mobile terminal apparatus but work as the user, in the time of but need using sound amplification function, present design will cause user's inconvenience.So in the present embodiment, in the situation that voice system opens, sound amplification function is further opened in the action that can pull and connect by voice, is user-friendly for conversation.
In another embodiment, whenpublic address equipment 650 andearphone 670 all with the situation of mobileterminal apparatus 600 lines under (being thatpublic address equipment 650 all couples processing unit with earphone 670), if provide toprocessing unit 660 and be input speech signal SAI, processingunit 660 can be according to user's setting, makingearphone 670 conversations is the first preferential talking mode (preset value), andpublic address equipment 650 is the second preferential talking mode.Perhaps,public address equipment 650 is made as the first preferential talking mode (preset value),earphone 670 conversations are made as the second preferential talking mode.
In addition, in another embodiment, when the user provides input operation signal SIO byinput unit 620, the expression user can't not touch the problem of mobile terminal apparatus immediately, therefore carry out the address book data coupling according to input operation signal SIO atprocessing unit 660 after, by processingunit 660, pull and connectunit 630 and output audio signal SAO can be sent to the voice output interfaces such aspublic address equipment 650,receiver 640 orearphone 670, it is looked closely the output interface (preset value) that the user presets and decides.
For instance, when the user says " phoning Lao Wang " facing to mobile terminal apparatus, after speech sample module 610 receives these sound at this moment, it is changed into input speech signal SAI, and this input speech signal SAI is by the parsing of speech understanding module, (for example: Lao Wang), and and then (for example: Wang Anshi) obtain select target obtain communication instruction (for example: make a phone call) and communication target.Owing to being the communication instruction of resolving from " voice ", therefore processing unit 660 automatic liftings are opened public address equipment 650 for enabling signal, in order to the follow-up conversation that amplifies.That is to say, when pull and connect the unit complete pull and connect after, the user can utilize public address equipment directly to talk with Lao Wang.Perhaps, in another example, when the user says facing to mobile terminal apparatus " answer the call ", after speech sample module 610 receives these sound at this moment, it is changed into input speech signal SAI, and this input speech signal SAI obtains communication instruction (as: answering the call) by the parsing of speech understanding module.Owing to being the communication instruction of resolving from " voice ", therefore processing unit 660 automatic liftings are opened public address equipment 650 for enabling signal, can utilize public address equipment directly and the Lao Wang dialogue in order to the user.Embodiment about configuration mode and the correlative detail of above-mentioned speech understanding module has been described in the front does not repeat them here.In addition, about communication target and last resulting select target, its execution mode can be taked aforementioned method or other the similar methods of utilizing cloud server, does not repeat them here.Certainly, as mentioned above, in public address equipment 650 and earphone 670 and the situation of depositing, processing unit 660 can be according to user's setting, and making earphone 670 conversations is the first preferential talking mode, and public address equipment 650 is the second preferential talking mode.
In another example, if the user is by thedisplay unit 430 of similar Fig. 4, when utilizing button or touch-control to select " Wang Anshi " in address list, owing to being when providing input operation signal SIO byinput unit 620, processingunit 660 can carry out the address book data coupling according to input operation signal SIO, and by processingunit 660, pull and connectunit 630 and user's setting, output audio signal SAO is sent to the voice output interfaces such aspublic address equipment 650,receiver 640 orearphone 670, makes the user to talk with Wang Anshi.
According to above-mentioned, can converge the whole automatic starting method that goes out a kind of sound amplification function of conversing of a mobile terminal apparatus.Fig. 7 is the flow chart according to the automatic starting method of the conversation sound amplification function of the mobile terminal apparatus of one embodiment of the invention.Please simultaneously with reference to Fig. 7, in the present embodiment, judge whether theprocessing unit 660 of mobileterminal apparatus 600 pulls and connects function (step S710) with unlatching.In other words, from the input speech signal SAI of the input operation signal SIO ofinput unit 620 orspeech sample module 610 may not with pull and connect relevantly, it might be the operation of carrying out other.Such as: enable the computer function in mobile terminal apparatus or utilize voice system inquiry weather etc.When processingunit 660 according to input-signal judging with unlatching pull and connectunit 630 pull and connect function the time, that is input signal and to pull and connect action relevant, step S710 judgment result is that "Yes", execution in step S720; Otherwise, when processingunit 660 will can not be pulled and connected function according to input-signal judging, that is step S710 judgment result is that "No", finish the automatic starting method of this conversation sound amplification function.
Then, in step S720, judge whether processing unit 660 receives the input speech signal SAI that pulls and connects function in order to unlatching.When processing unit 660 receive from speech sample module 610 pull and connect the input speech signal SAI of function in order to unlatching the time, that is step S720 judgment result is that "Yes", can Check processing unit 660 whether be connected (step S730) with earphone 670.When processing unit 660 is connected with earphone 670, that is step S730 judgment result is that "Yes", processing unit 660 automatic liftings for enabling signals starting earphone, and output audio signal SAO is to earphone 670(step S740); Otherwise, when processing unit 660 is not connected with earphone 670, that is step S730 judgment result is that "No", processing unit 660 automatic liftings supply enabling signal to start public address equipment 650, and output voice signal SAO is to the public address equipment 650 of mobile terminal apparatus 600, to open the conversation sound amplification function (step S750) of mobile terminal apparatus 600.It is worth mentioning that, when processing unit 660 receives when pulling and connecting the input speech signal of function in order to unlatching, above-mentioned step 730 ~ step 750 is in the situation that the user is set as preferential voice output interface (suppose all lines of public address equipment 650 and earphone 670) with earphone 670 carries out.In other embodiments, the user also can be set as public address equipment 650 preferential voice output interface.Certainly, when earphone 670 and public address equipment 650 only have one of them line, can set equipment on line as preferential voice output interface.Above-mentioned implementation step is to know the operator can do corresponding change according to its demand.
On the other hand, when processing unit 660 do not receive from speech sample module 610 pull and connect the input speech signal SAI of function in order to unlatching the time, that is step S720 judgment result is that "No", can follow Check processing unit 660 and whether be connected (step S760) with earphone 670.Specifically, processing unit 660 does not receive the input speech signal SAI from speech sample module 610, but processing unit is pulled and connected function with unlatching again, and expression processing unit 660 receives the input operation signal SIO from input unit 620, and this input operation signal SIO and pull and connect move relevant.When processing unit 660 is connected with earphone 670, that is step S760 judgment result is that "Yes", processing unit 660 can automatic liftings for enabling signals starting earphone 670, and output voice signal SAO is to earphone 670(step S740).Otherwise, when processing unit 660 is not connected with earphone 670, that is step S760 judgment result is that "No", processing unit 660 provides output voice signal SAO one of them (step S770) to public address equipment and receiver according to preset values.Wherein, the order of above-mentioned steps is the use as explanation, and the embodiment of the present invention is not as limit.It is worth mentioning that, when step 760 is judged as " be ", will provide output audio signal SAO to earphone 670, above-mentioned condition is set as earphone 670 situation of preferential voice output interface (supposing all lines of receiver 640, public address equipment 650, earphone 670) for the user.In other embodiments, the user also can be set as preferential voice output interface with receiver 640 or public address equipment 650.Certainly, at receiver 640, public address equipment 650, when earphone 670 equipment only have one of them line, can set equipment on line as preferential voice output interface.Above-mentioned implementation step is to know the operator can do corresponding change according to its demand.
In sum, the automatic starting method of the mobile terminal apparatus of the embodiment of the present invention and conversation sound amplification function thereof, when the processing unit reception is pulled and connected the input speech signal of function in order to unlatching, pull and connect function except opening, more can automatically open sound amplification function, will export voice signal to public address equipment.Thus, when the user can't touch mobile terminal apparatus immediately, but sound amplification function need be made the time, can start sound amplification function by voice system, to improve the ease of use of mobile terminal.
Although the present invention with embodiment openly as above; so it is not to limit the present invention; those skilled in the art when doing a little change and retouching, are as the criterion therefore protection scope of the present invention ought be looked the appended claims confining spectrum without departing from the spirit and scope of the present invention.