Movatterモバイル変換


[0]ホーム

URL:


US10395651B2 - Device and method for activating with voice input - Google Patents

Device and method for activating with voice input
Download PDF

Info

Publication number
US10395651B2
US10395651B2US13/780,010US201313780010AUS10395651B2US 10395651 B2US10395651 B2US 10395651B2US 201313780010 AUS201313780010 AUS 201313780010AUS 10395651 B2US10395651 B2US 10395651B2
Authority
US
United States
Prior art keywords
circuitry
input voice
voice instruction
information processing
processing apparatus
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US13/780,010
Other versions
US20140244269A1 (en
Inventor
Kenji TOKUTAKE
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sony Corp
Original Assignee
Sony Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US13/780,010priorityCriticalpatent/US10395651B2/en
Application filed by Sony CorpfiledCriticalSony Corp
Priority to EP13166981.4Aprioritypatent/EP2772907A1/en
Priority to EP21193046.6Aprioritypatent/EP3989043A1/en
Priority to EP17205856.2Aprioritypatent/EP3324404A1/en
Priority to EP18166758.5Aprioritypatent/EP3379530A1/en
Publication of US20140244269A1publicationCriticalpatent/US20140244269A1/en
Assigned to SONY MOBILE COMMUNICATIONS ABreassignmentSONY MOBILE COMMUNICATIONS ABASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: Tokutake, Kenji
Assigned to SONY CORPORATIONreassignmentSONY CORPORATIONASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: SONY MOBILE COMMUNICATIONS AB
Priority to US16/505,737prioritypatent/US10825457B2/en
Application grantedgrantedCritical
Publication of US10395651B2publicationCriticalpatent/US10395651B2/en
Priority to US17/026,323prioritypatent/US11580976B2/en
Activelegal-statusCriticalCurrent
Adjusted expirationlegal-statusCritical

Links

Images

Classifications

Definitions

Landscapes

Abstract

An information processing apparatus that detects a voice command via a microphone in order to activate the device and execute certain applications. The apparatus comprises a digital signal processor (DSP) and a host controller which are responsible for processing the voice commands. The DSP recognizes and processes voice commands intermittently while the host processor is in a sleep state, thereby reducing the overall power consumption of the apparatus. Further, when the DSP is configured to recognize voice commands intended only to activate the device, a memory having a sufficiently lower storage capacity suffices.

Description

BACKGROUND1. Field of the Disclosure
This disclosure relates to a method of efficiently operating a device through voice instructions.
2. Description of the Related Art
Mobile devices, such as smart phones and tablets are examples of devices that are adapted to take as input a user's voice (in the form of voice commands) and perform certain instructions.
Traditionally, when processing a voice command that is intended to execute a certain application, the processing circuitry of the device (responsible for processing and analyzing the voice input), continuously monitors the input such as a microphone or the like, for voice commands inputted from the user. In doing so, the control circuit of such a device which executes an application essentially waits for the processing circuit to complete its tasks of analyzing the voice input.
In such scenarios, the electric power consumption of the device increases dramatically. Accordingly, there is a requirement to lower the power consumption in order to utilize the device in an efficient manner.
SUMMARY
Devices and methods for operating the devices via voice commands in order to lower the power consumption of the device are discussed herein.
According to one exemplary embodiment, the disclosure is directed to an information processing apparatus comprising: a microphone configured to detect input voice instructions; a memory configured to store a plurality of keywords, each keyword being associated with a predetermined function of the information processing apparatus; first circuitry configured to compare an input voice instruction with the plurality of keywords stored in the memory; and activate second circuitry, which is configured to execute an application based on the comparison.
According to another exemplary embodiment, the disclosure is directed to a method performed by an information processing apparatus, the method comprising: receiving input voice instructions from a microphone; storing a plurality of keywords in a memory, each keyword associated with a predetermined function of the information processing apparatus; comparing by a first circuitry the input voice instruction with the plurality of keywords; activating a second circuitry, by the first circuitry based on the comparing; and executing an application by the second circuitry corresponding to the keyword associated with the application.
According to another exemplary embodiment, the disclosure is directed to a non-transitory computer-readable medium including computer program instructions, which when executed by an information processing apparatus, cause the information processing apparatus to perform a process comprising: receiving input voice instructions from a microphone; storing a plurality of keywords in a memory, each keyword associated with a predetermined function of the information processing apparatus; comparing by a first circuitry the input voice instruction with the plurality of keywords; activating a second circuitry, by the first circuitry based on the comparing; and executing an application by the second circuitry corresponding to the keyword associated with the application.
The foregoing general description of the illustrative implementations and the following detailed description thereof are merely exemplary aspects of the teachings of this disclosure, and are not restrictive.
BRIEF DESCRIPTION OF THE DRAWINGS
A more complete appreciation of the invention and many of the attendant advantages thereof will be readily obtained as the same becomes better understood by reference to the following detailed description when considered in connection with the accompanying drawings, wherein:
FIG. 1 illustrates schematically an exemplary scenario outlining the sleep and active states of a digital signal processor (DSP) and a host processor;
FIG. 2 illustrates schematically an exemplary mobile phone terminal device;
FIG. 3 illustrates schematically the structure of a DSP speech processing unit;
FIG. 4 illustrates an exemplary non-limiting example of the processing states in the DSP;
FIG. 5 illustrates the processes conducted by a host processor;
FIG. 6 illustrates an exemplary flowchart depicting a method to wake up a device (i.e., brought to an active state) via a voice recognition process according to one aspect of the present disclosure;
FIG. 7 illustrates the audio (speech) analysis performed by the DSP;
FIG. 8 illustrates an exemplary flowchart depicting a method to wake up a device via a voice recognition process according to a second aspect of the present disclosure;
FIG. 9 depicts a non-limiting example of an apparatus (a mobile phone) that transitions to an active state via voice commands according to the present disclosure;
FIG. 10 illustrates an exemplary flowchart depicting the steps taken while performing a voice pattern analysis;
FIGS. 11A and 11B illustrate an exemplary flowchart depicting a method to wake up a device by a voice recognition process according to another aspect of the present disclosure; and
FIG. 12 illustrates an exemplary flowchart depicting a voice recognition process implemented in the method ofFIG. 8 to wake up a device.
FIGS. 13A-13C depict another non-limiting example illustrating the execution of an application using the voice pattern analysis ofFIG. 10.
DETAILED DESCRIPTION
Referring now to the drawings, wherein like reference numerals designate identical or corresponding parts throughout the several views.
FIG. 1 illustrates a non-limiting exemplary scenario outlining the sleep and active states of a digital signal processor (DSP) and a host processor that are comprised within a device which is configured to take as input, voice commands (instructions) from a user in order to bring it to an active state. As shown inFIG. 1, the DSP2, comprises avoice pattern memory3, which stores user input data i.e., keywords which are used to activate the device. The keywords can be short words such as the name of user or simply commonly used words such as “hello”. These keywords are stored in the memory by a registration process which is explained in detail later with reference toFIGS. 6 and 8.
The DSP2, takes as input voice commands from a user which are inputted via amicrophone1. The analog data (user's voice) is modulated using a pulse code modulation (PCM) technique or the like and fed in to the DSP for further processing. The details of this voice processing and analysis are explained in detail with reference toFIG. 3. Note that DSP is an active state while it receives information from the user. At this instant in time, the host processor is kept in a “sleep” state (i.e., a suspended state). The host processor is responsible for executing certain applications based on the keyword input from a user. Low power consumption is obtained by maintaining the host processor in a sleep state, while the DSP processes the voice commands input by a user.
On receiving the input from the microphone, the DSP compares the input keyword with those stored in thevoice pattern memory3. If a match is found, the DSP notifies the host processor about a successful match via an interrupt command (discussed later with reference toFIG. 8) and brings the host processor to an active state. This process is referred to as “waking up” the host processor. The host processor on being transitioned to the active state executes a certain application (denoted by AP1-AP4) based on the recognized speech.
In one aspect of the present disclosure, a method of processing the speech data, wherein the host processor processes the data and the DSP is responsible only for the analysis of the keyword is described. In such a setting, note that the keywords are used to wake up the device. The specific execution of an application proceeds later with voice commands inputted by a user. In another aspect of the present disclosure, a method wherein the DSP is responsible for all the speech processing is described. Moreover, both embodiments outline as to how lower power consumption is obtained by specifically turning the DSP and the host processor into sleep modes.
FIG. 2 illustrates a schematic block diagram of an exemplary mobilephone terminal device100. As shown inFIG. 2, the mobilephone terminal device100 may include anantenna101 and a wirelesscommunication processing section102. The wirelesscommunication processing section102 may communicate wirelessly via radio signals, or the like, with other mobile devices via a base station. Further, a data signal, such as a voice transmission from another user, may be received byantenna101 and sent to the wirelesscommunication processing section102 for further processing. In the case of an incoming voice transmission, the voice data signal may be sent from the wirelesscommunication processing section102 to avoice processing section103. Incoming voice data received by thevoice processing section103 via the wirelesscommunication processing section102 may be output as sound via aspeaker104.
Conversely, an outgoing voice signal may be supplied by a user to thevoice processing section103 via amicrophone105. The voice signal received viamicrophone105 and processed by thevoice processing section103 may be sent to wirelesscommunication processing section102 for transmission by theantenna101. Thevoice processing section103 comprises a digital signal processor (DSP)103awhich digitizes the incoming analog signal and processes the audio input to detect for keywords. Keywords enable the operation ofdevice100, when it is configured to operate under the instructions of specific voice commands. These keywords are preset in the device with the aid of a voice registration unit and stored in thevoice pattern library103b. The detailed description of this processes is explained later with reference toFIGS. 5 and 6.
Asecond antenna106 may be supplied for use with a short distance wirelesscommunication processing section107. The short distance wirelesscommunication processing section107 may communicate wirelessly with other devices over a network, such as the Internet, a local area network (LAN), or a wide area network (WAN). Thesecond antenna106 may, e.g., by a Wi-Fi transceiver.
Asensor section108 may be provided for the mobilephone terminal device100. Thesensor section108 may be a motion sensor that detects a motion of an object in the proximity of the mobilephone terminal device100. The motion may correspond to a user moving an instruction object, such as a finger or stylus, in the proximity of the mobilephone terminal device100 for the purpose of selecting data displayed ondisplay120.
The mobilephone terminal device100 may includedisplay120. Thedisplay120 may be, for example a liquid crystal display (LCD) panel, an organic electroluminescent (OLED) display panel, a plasma display panel, or the like. Thedisplay120 may display text, an image, a web page, a video, or the like. For example, when the mobilephone terminal device100 connects with the Internet, thedisplay120 may display text and/or image data which is transmitted from a web server in Hyper Text Markup Language (HTML) format and displayed via a web browser. Thedisplay120 may additionally display data stored in amemory150.
Atouch panel section130 can detect a touch operation on the surface of thedisplay120. For example thetouch panel130 can detect a touch operation performed by an instruction object such as a finger or stylus. Touch operations may correspond to user inputs such as a selection of an icon or a character string displayed on thedisplay120. Thetouch panel section130 may be an electrostatic capacitance type device, a resistive type touch panel device, or other such type devices for detecting a touch on a display panel.
Thetouch panel section130 may perform processing related to touch operation classification. For example, thetouch panel section130 may assign a predetermined function to be performed when a “tap” touch operation is detected. Similarly, the touch panel section may analyze a touch operation in which the instruction object makes continuous contact with thedisplay120 while moving the instruction object around the display120 (e.g., a “swipe” operation). Thetouch panel section130 may output a signal based on a classification of the touch operation performed. The signal may for example include information indicating the touch operation classification, the location on thedisplay120 where the touch operation was performed, and the operation to be performed based on the touch operation.
Data which is detected and processed by thetouch panel130 can be transmitted to ahost controller110. The host controller/processor110 may include one or more processor units and can control each element of the mobilephone terminal device100 based on data detected by thetouch panel130, or by inputs received fromoperation key140. Theoperation key140 may receive inputs, e.g., from external control buttons included with the mobilephone terminal device100. The external control buttons may for example control the volume, the power, or a hold operation for the mobilephone terminal device100.
Thehost controller110 may further execute instructions stored in thememory150. The controller may further comprise of aDSP driver111, which is configured to communicate with theDSP103a. Specifically, the driver may actuate the DSP during a voice registering phase, or theDSP103amay initiate communication with the driver upon the successful detection of a voice command. Thedriver111 may further activate the host processor to execute a certain application based on the received voice commands. The specific details pertaining to thedriver111 and the host controller are explained later with reference toFIG. 5. To this end, thememory150 may be a non-transitory computer readable medium having instructions stored therein for controlling the mobilephone terminal device100. Further, thecontroller110 may include one or more processors for executing the instructions stored on thememory150.
The mobilephone terminal device100 can include a control line CL and a data line DL as internal bus lines for communication. The control line CL can be used to transmit control data from thecontroller110. The data line DL may be used for the transmission of voice data, display data, or the like, throughout the various elements of the mobilephone terminal device100.
FIG. 3 illustrates a block diagram depicting the processing of a voice command input by a user via themicrophone105. The output of the microphone is first amplified before further processing by anamplifier103c. After modulating the input voice (speech) by using a modulation scheme like the pulse code modulation or the like, the signal is input to an analog todigital converter103d. This converter obtains digital samples of the input analog voice signal and stores them in amemory buffer103e.
TheDSP processor103aperforms processing on these digitized voice samples and checks for a match with certain user input keywords which are stored in avoice pattern memory103b. Note that these keywords are small words/phrases such as “hello”, the user's name and are used to wake up the device. Specific details pertaining to how the keywords are stored in the memory and the process of initiating an application execution will be explained later with reference toFIGS. 6 and 8. Note that the DSP may perform further processing on the voice samples such as noise removal, voice compression etc, to further clean and/or make compact the input voice samples for better processing.
The DSP communicates with the host processor via a control line (CL) and a data line (DL). Specific instructions and control signals are exchanged between the processors by using the CL. Data pertaining to the inputted voice command via the microphone is exchanged with the host processor via the DL. This exchange of information between the DSP and the host processor is explained in detail with reference toFIGS. 5 and 7.
FIG. 4 depicts a non-limiting example of the processing states involved in the DSP and illustrates the active and sleep cycles of the DSP.FIG. 4A depicts a standard analog audio signal that is input toapparatus100 via themicrophone105. InFIG. 4B, a modulation scheme such as pulse code modulation (PCM) is applied to the analog signal, which is further processed in an A/D converter to obtain digital samples. As stated previously, these digital signals are stored in thebuffer memory103e. The digital data is read from the buffer memory and transferred to theDSP103 for processing at every fixed time-period. Note, that as shown inFIG. 4C, the DSP is in an active state only when the data (at every fixed time period) is read from the memory and transferred to the processor. At all other times the DSP remains in a sleep mode, thus enabling the decrease of power consumption.
While being in active state, the DSP reads the data from thebuffer memory103eand compares (for a match) the input data to the set of keywords that are stored in thevoice pattern memory103b. On finding a successful match, the DSP notifies the host processor of the matched keyword and brings it to an active state. Note that the host processor is in the sleep state while the DSP performs the comparison process. Further, it must be appreciated that the DSP is in an active state only when data is read from thebuffer103efor further processing. At all other times when data is not fed from the buffer, the DSP remains in a sleep state. Thus it must be appreciated that by alternating from a sleep to active state, the DSP operates in an intermittent manner thus providing a valuable savings in terms of operating power. Further, note that by keeping the host processor in a sleep state (while the DSP performs it's processing), and activating it only when a keyword match is successfully detected by the DSP, a further savings in the power consumption of theapparatus100 is obtained.
FIG. 5 illustrates the processing steps performed by a host processor upon successful keyword match detected by the DSP. As shown inFIG. 5, theDSP103 includes avoice analysis library103xand avoice pattern memory103b. The voice command received from themicrophone105 is input to the voice analysis library wherein the input voice command is compared with the keywords stored in thevoice pattern memory103bfor a match. Upon a successful match, the DSP instructs (via control signals) theDSP driver111 comprised within the host processor to transition to an active state.
Once in active state, the DSP driver initiates the keyword recognition process by instructing thekeyword analysis application113 to perform a search for the appropriate application that is to be executed. Note that the instruction to perform the application search is sent from aframework unit112, to theapplication analysis unit113 via an application interface. On receiving instructions from theDSP driver111, the keyword analysis application searches for the application within the application set AP1-AP4, for the keyword associated with the application. On finding the correct match the appropriate application is executed.
Note that in the above process, theDSP103 performs only a keyword match of the input voice command with the keywords registered by a user and stored in the voice pattern memory. Until a successful match is obtained the host processor remains in a sleep mode. On finding a successful match, the DSP activates the host controller (processor) via the DSP driver. The host controller on receiving the keyword from the DSP, runs a search through the application database to search for the application associated with the keyword. Thus the process of application recognition is performed by the host controller.
FIG. 6 is a flowchart of the process for waking up the device and outlines the steps undertaken to recognize an input voice command according to one aspect of the present disclosure. Note that in this example thehost processor110 performs the voice registration process and the input processing of the audio signal. Further we assume for the sake of simplicity that the application intended to be executed by the voice command is one of making a telephone call.
In step S11, a user inputs a voice command that is assigned to start an application. This command is registered in the voice pattern registration application. Note that commands such as “hello” or a user's name can be assigned to indicate a start of the application execution. In step S12, the voice pattern registration application registers the start of the specific application (with respect to the input keyword) in the application launcher.
In step S13, the voice pattern registration application notifies theDSP driver111 about the assigned voice pattern, which in turn, in step S14 registers the assigned voice pattern in theDSP103. The DSP stores information pertaining to the assigned voice keyword in itsvoice pattern memory103b. This stored information is used later to obtain a match, when a user inputs a voice command via the microphone115.
In step S15, the host controller initiates a suspension (sleep) state for all units other than theDSP103a. In step S16, the control unit of the host processor brings the DSP driver111 (comprised within the host processor) to a sleep state. Note (as shown in step S17) that the DSP driver initiates a start of comparison of voice pattern to theDSP103a. Once the DSP driver has initiated this start of comparison process to the DSP, thehost controller101 is completely in a suspension state (step S18).
In step S19, the DSP101atakes as input, voice commands from the microphone115 and compares the inputted commands to the keywords stored in thevoice pattern memory103b. Note that this process is performed intermittently as shown inFIG. 4. Specifically, the DSP undergoes cycles of sleep and active states depending on the time it receives input from the buffer memory.
Step S20 depicts the case wherein the inputted voice command has resulted in a match with one of the keywords stored in the voice pattern memory. Note that keywords intended to start the apparatus are small words/phrases such a “hello” or a users name. On detecting a match, in step S21 the DSP triggers the host processor via an interrupt signal. In doing so, the host processor which was in a sleep/suspended state is transitioned to an active state. The host processor further triggers theDSP driver111 to an active state (step S22). In step S23, the DSP driver initiates the application launcher which in turn signals the DSP driver to be in a wait state (step S24). Specifically, the application launcher upon transitioning to an active state signals the DSP driver to wait for further voice commands from the DSP. Note that at this point of time no application is yet executed, rather, with the use of a keyword the DSP and the host processor are activated to receive and process further instructions as to which specific application is to be executed.
Steps S25 to S35 illustrate the process from the time an input voice instruction is fed to theDSP103 to the time the host processor executes the specific application. In step S25, the user inputs a voice instruction such as “make a telephone call” via the microphone115. This instruction is transmitted from the DSP to the DSP driver in step S26, which in turn invokes the application launcher and notifies it about the specific voice instruction, as shown in Step S27.
In step S28, the application launcher performs an audio analysis as described inFIG. 5. Based on the content of the data (voice) obtained, theaudio analysis application113 searches for the appropriate application to execute. In the example depicted inFIG. 6, we consider an application of making a telephone call. Thus, the application launcher initiates the call application in step S29. In step S30, the application launcher notifies the DSP driver to wait for further instructions from the user. For example, the application launcher may request the DSP driver to receive the specific number to be called from the DSP.
On receiving this information in step S31, the DSP notifies the driver of the number to call (step S32) which in turn passes the information to the application launcher (shown in step S33). Finally, the application launcher notifies the call application of the number to call in step S34, wherein-after the specific application (call application in this example) executes the associated task as shown in Step S35.
FIG. 7 illustrates according to another aspect of the present disclosure, a method of waking up the device and outlines the steps undertaken by the host processor when the DSP conducts the processes pertaining to voice analysis. Note that in the first aspect of the disclosure (FIGS. 5 and 6), the DSP performed voice analysis only to activate the host processor. This was accomplished with the use of a keyword such as “hello” which would initiate (i.e., bring to active state) the host processor.
In the illustration ofFIG. 7, the DSP's voice pattern memory memorizes not only keywords associated to start the apparatus but also maintains a list of voice instructions required to execute an application. Thevoice analysis unit103x, receives input instructions from themicrophone105, and compares the instructions with those stored in thevoice pattern memory103b.
Upon finding a match, the DSP notifies the DSP driver111 (of the host processor) of the match, through an interrupt command. Note that the interrupt command varies depending upon the application type to be executed. The driver transmits the information received from the DSP to aframework unit112, which through an application interface (API) executes the appropriate application.
FIG. 8 represents a flowchart of the voice recognition process for the example described inFIG. 7. For sake of illustration, we consider the same application, that of making a phone call to explain the steps of the process ofFIG. 8.
In step S41, a user inputs a voice command that is assigned to start an application. This command is registered in the voice pattern registration application. Note that commands such as “hello” or a user's name can be assigned to indicate a wake up of the processors to further execute a start of the application. In step S42, the voice pattern registration application notifies theDSP driver111 of the pattern registered to activate theapparatus100.
In step S43, the DSP driver notifies the DSP of the voice pattern. Note, that in this case, thememory103bassociated with the DSP stores the voice pattern for activating the host controller as well as the instructions that initiates the execution of an application. In step S44, the host controller initiates a suspension (sleep) state for all units other than theDSP103a. In step S45, the control unit of the host processor brings the DSP driver111 (comprised within the host processor) to a sleep state. In step S46, the DSP driver signals theDSP103 to start the process of comparing an input voice pattern with those stored in the memory of the DSP. At this time instant, the DSP conducts a voice analysis by the method described inFIG. 4. Specifically, the DSP operates in an intermittent fashion wherein it undergoes a series of periodic sleep and active states. Note that while the DSP performs processing of the input voice instructions (step S48), the host controller is in a suspension state as shown in step S47.
In step S49, once the DSP has confirmed a match between the voice instruction inputted from themicrophone105 and the voice pattern stored in thememory103b, it sends an interrupt signal to the controller, thereby indicating the controller to resume an active state (step S50). The controller once activated, instructs the DSP driver in step S51, to transition the active state.
Once the DSP driver is in the active state it selects the application to start and invokes the corresponding application as shown in Step S52. Note that in this case, we consider a simple application to make a telephone call. Accordingly the DSP driver, in step S53 initiates a telephone call operation. In step S54, the application (for example the call application) instructs the DSP driver to wait for further input from theDSP103. This input pertains to the number required by the call application in order to make the telephone call. This number is inputted as shown in step S55 and is notified to the DSP driver in step S56. The number is further transmitted to the call application (step S57) which finally makes the telephone call in step S58. Note that in both aspects presented thus far, the DSP and the host processor are put in a sleep mode when not processing information thereby providing a means to lower the power consumption of the apparatus.
FIG. 9 illustrates how an apparatus (e.g. a mobile phone terminal) can be configured to be activated by voice instructions to perform a certain application. We consider an example of transferring data (images, text or the like) from theapparatus100 to adisplay terminal200. Note that the example depicts the two devices connected by a wired mechanism. However, the means of connecting this device is in no way limiting the example. For example, theapparatus100 and thedisplay terminal200 may well be connected by a wireless mechanism to achieve the end objective,
Initially, both theapparatus100 and thedisplay terminal200 are in a suspended state. On receiving instructions from a user theapparatus100 makes a transition to the active state. Specifically, theapparatus100 recognizes the voice pattern (keyword) inputted by a user, to bring it in an active state. On transitioning to the active state, theapparatus100 instructs the display to turn in to the active state.
Once the devices are activated, thedisplay terminal200displays data121 of the mobile phone terminal on itsdisplay screen201. Hence, in this manner the display of information from a mobile phone to a display terminal is accomplished by the use of voice instructions implemented in a manner presented in the present disclosure.
FIG. 10 illustrates an exemplary flowchart depicting the steps taken while performing a voice pattern analysis (i.e., voice recognition process). The voice pattern analysis process includes two steps: a voice verification step and a voice pattern matching step.
In the voice verification step, the voice of a registered (i.e., authorized) user is first verified. Specifically, an input voice is matched with a registered voice by verifying, in detail, the characteristics of the input voice waveform. Further, the keyword associated with the voice is also verified. In doing so, a security mechanism is created wherein unauthorized users are not granted access to unauthorized devices. For example, if an authorized user has registered the word ‘hello’ to activate a mobile phone, only the keyword (‘hello’) spoken by the authorized user will activate the mobile phone.
Once the device is activated by the authorized user, the voice pattern matching step processes further voice instructions. The processing in this step simply verifies if an appropriate keyword associated with an application is input. For example, if a voice instruction “make telephone call”, is associated with the application of making a telephone call by the mobile phone, in the voice pattern matching step the input keyword is verified. Note that access in this step is granted to all users. Specifically, once the device is activated by the authorized user, any user can execute an application on the device, as long as the correct instruction is input to the device.
FIG. 10 illustrates the steps performed by the voice pattern analysis process. First, the voice verification process is executed at specific time instants that correspond to the instants of time the DSP of the device is active (as shown inFIG. 4). In step S91, a query is made to check if the voice verification is successful. If the response to the query is affirmative, the process proceeds to step S92, wherein the voice pattern matching process is performed. If the response to the query in step S91 is negative (i.e., the voice verification is erroneous) the process simply loops back to step S91.
In step S92 a query is made to check if the voice pattern matching is successful. If the response to this query is negative, the process loops back to step S92, wherein the device awaits the input of the appropriate voice instruction. If the response in step S92 is successful, the process proceeds to step S93 wherein the application associated with the instruction is executed.
In step S94, a query is made to check if a security time has elapsed from the onset of a successful voice verification process. If the response to the query is affirmative, the process loops back to step S91, wherein the device needs to be activated again by the authorized user. However, if the response to the query is negative, the process loops back to step S92, wherein the DSP of the device awaits further instructions from a user to execute an appropriate application.
FIGS. 11A and 11B illustrate an exemplary flowchart depicting a method to wake up a device by a voice recognition process according to another aspect of the present disclosure. Note that the flowchart ofFIG. 11B depict the steps performed after the steps of the flowchart depicted inFIG. 11A are executed. Further, note that the flowcharts ofFIGS. 11A-11B are similar to the flowchart ofFIG. 6, and differ in the voice pattern analysis process. In what follows, we explain the steps of the flowcharts ofFIGS. 11A-11B that differ from the steps ofFIG. 6.
After the voice pattern registration application registers the start of a specific application (with respect to an input keyword) in the application launcher (step S12), the voice pattern registration application notifies the DSP driver, in step S13, the voice of the authorized user.
In Step S14, the authorized voice is registered inDSP103, by the DSP driver. Note, that in doing so, an unauthorized user is not granted permission to activate the device. Further, the flowcharts ofFIGS. 11A-11C, differ in the manner in which the voice analysis is performed. As shown in step S21, a voice input via the microphone is first verified by processing the input voice waveform and matching the voice to that of the authorized user. Further, in step S22, the keyword input to activate the device is analyzed.
Upon a successful match, the device is activated, thus enabling it to accept further voice instructions (step S23) from any user to execute an application. For the sake of simplicity, we consider the same application (as inFIG. 6) of making a telephone call. The further steps that execute the telephone call application are similar to those ofFIG. 6. Similarly,FIG. 12 depicts according to another embodiment of the present disclosure a flowchart depicting the steps taken to wake up a device by using a voice recognition process. Specifically, the flowcharts ofFIG. 12 is similar to the flowchart ofFIG. 8, and differs in the voice pattern analysis process, which is similar to the voice pattern analysis process as described inFIGS. 11A and 11B.
FIGS. 13A-13C depict a non-limiting example illustrating the execution of an application using the voice pattern analysis process ofFIG. 10.
Initially, as shown inFIG. 13A, thedisplay terminal200 is in an inactive state (OFF). On receiving instructions from an authorized user, themobile terminal100 that is connected to thedisplay terminal200, is brought to an active state. Note, that only the voice of the authorized user can enable the mobile terminal to transition to an active state (ON state). A predetermined instruction such as ‘Hello my phone’ can be used by the user to activate themobile terminal100, as shown inFIG. 13B.
Upon activating the mobile terminal, an application can be executed by any user.FIG. 13 C depicts an example wherein a user intends to display pictures from themobile terminal100 to thedisplay terminal200. A pre-determined instruction such as ‘Show pictures on TV’ can be used to indicate the execution of transferring pictures to the display terminal. Note that the processing involved in this stage, matches only the instructions spoken by the user to the keyword associated with the application. Specifically, the voice waveform is not processed in detail to match a registered user's voice, thus enabling any user to execute an application, provided that the devices have been activated by a authorized user.
Obviously, numerous modifications and variations of the present invention are possible in light of the above teachings. It is therefore to be understood that within the scope of the appended claims, the invention may be practiced otherwise than as specifically described herein. For example, the DSP as presented inFIG. 3 may be configured in another manner, wherein the input data is modulated by mechanisms other PCM. Further, the application considered in this disclosure was that of making a telephone call. Similarly, other applications which can be voice activated can be implemented in a manner as disclosed inFIGS. 6 and 8. Additionally, devices other than the mobilephone terminal device100 as shown inFIG. 9 may be used to perform the features discussed in present disclosure. For example, aspects of the present disclosure may be executed on a smart phone, a tablet, a general purpose computer, a laptop, an electronic reading device, or other such display terminals.
The above disclosure also encompasses the embodiments noted below.
(1) An information processing apparatus comprising: a microphone configured to detect input voice instructions; a memory configured to store a plurality of keywords, each keyword being associated with a predetermined function of the information processing apparatus; first circuitry configured to compare an input voice instruction with the plurality of keywords stored in the memory; and activate second circuitry, which is configured to execute an application based on the comparison.
(2) The information processing apparatus of (1), wherein the second circuitry is in a suspended state when the first circuitry performs the comparison.
(3) The information processing apparatus of (1), wherein the first circuitry is continuously in an active state when the second circuitry is in an active state.
(4) The information processing apparatus of (1), wherein the first circuitry activates the second circuitry based on a successful match of the input voice instruction with a keyword.
(5) The information processing apparatus of (1), wherein the voice instructions are modulated by pulse code modulation before being stored in the memory.
(6) The information processing apparatus of (5), wherein the modulated voice instructions are input to the memory at a fixed time period.
(7) The information processing apparatus of (6), wherein the first circuitry is in an active state for a fraction of the fixed time period.
(8) The information processing apparatus of (1), wherein the input voice instructions include the keyword and further instructions to execute the application.
(9) The information processing apparatus of (1), wherein the first circuitry activates the second circuitry by an interrupt signal.
(10) The information processing apparatus of (1), wherein the first circuitry and the second circuitry exchange data through a plurality of data lines.
(11) The information processing apparatus of (1), wherein the first circuitry and the second circuitry exchange control information through a plurality of control lines.
(12) The information processing apparatus of (1), wherein the second circuitry receives information from the first circuitry regarding a successful match of the voice instruction and the keyword and searches for the corresponding application associated with the keyword.
(13) The information processing apparatus of (12), wherein the second circuitry executes the application upon a successful keyword search.
(14) The information processing apparatus of (1), wherein the first circuitry includes a driving unit configured to activate the second circuitry.
(15) The information processing apparatus of (1), wherein the first circuitry is configured to match an input voice with a registered voice of an authorized user.
(16) The information processing apparatus of (15), wherein the first circuitry is further configured to match an input keyword to a predetermined keyword.
(17) The information processing apparatus of (16), wherein the second circuitry is configured to execute the application by any user.
(18) An information processing method performed by an information processing apparatus, the method comprising: receiving input voice instructions from a microphone; storing a plurality of keywords in a memory, each keyword associated with a predetermined function of the information processing apparatus; comparing by a first circuitry the input voice instruction with the plurality of keywords; activating a second circuitry, by the first circuitry based on the comparing; and executing an application by the second circuitry corresponding to the keyword associated with the application.
(19) A non-transitory computer-readable medium including computer program instructions, which when executed by an information processing apparatus, cause the information processing apparatus to perform a process, the process comprising: receiving input voice instructions from a microphone; storing a plurality of keywords in a memory, each keyword associated with a predetermined function of the information processing apparatus; comparing by a first circuitry the input voice instruction with the plurality of keywords; activating a second circuitry, by the first circuitry based on the comparing; and executing an application by the second circuitry corresponding to the keyword associated with the application.

Claims (13)

The invention claimed is:
1. An information processing apparatus comprising:
a mobile terminal device;
a microphone, disposed in the mobile terminal device, configured to receive input voice instructions;
first circuitry, disposed in the mobile terminal device, configured to determine a first input voice instruction as an activation keyword spoken by a registered user and to output an activation signal when the first input voice instruction is determined as the activation keyword and as having been spoken by the registered user, the first circuitry not outputting the activation signal when the first input voice instruction is determined as the activation keyword and as having been spoken by an un-registered user, the first circuitry being configured to operate in an active state; and
second circuitry, disposed in the mobile terminal device, configured to be in a sleep state while the first circuitry is in the active state until the second circuitry receives the activation signal, to enter the active state in response to the activation signal triggered by the spoken activation keyword of the registered user, and to execute an application based on a second input voice instruction that is received by the microphone subsequent to the first voice input instruction after entering the active state, the second input voice instruction being spoken by the un-registered user, wherein the un-registered user is a different person than the registered user.
2. The information processing apparatus ofclaim 1, wherein
the second circuitry is in the sleep state when the first circuitry performs the determination.
3. The information processing apparatus ofclaim 1, wherein
the first circuitry is continuously in the active state when the second circuitry is in the active state.
4. The information processing apparatus ofclaim 1, wherein
the voice instructions are modulated by pulse code modulation.
5. The information processing apparatus ofclaim 1, wherein
the first circuitry activates the second circuitry by an interrupt signal.
6. The information processing apparatus ofclaim 1, wherein
the first circuitry and the second circuitry exchange data through a plurality of data lines.
7. The information processing apparatus ofclaim 1, wherein
the first circuitry and the second circuitry exchange control information through a plurality of control lines.
8. The information processing apparatus ofclaim 1, wherein
the second circuitry receives information from the first circuitry regarding a successful match of the second input voice instruction and a corresponding one of a plurality of application keywords and searches for the corresponding application associated with the corresponding one of the application keywords.
9. The information processing apparatus ofclaim 8, wherein the second circuitry executes the application upon a successful keyword search.
10. The information processing apparatus ofclaim 1, wherein
the first circuitry includes a driving unit configured to activate the second circuitry.
11. The information processing apparatus ofclaim 1, further comprising:
a memory circuit configured to store the activation keyword and a plurality of application keywords, each of the plurality of application keywords being associated with a predetermined function of the information processing apparatus; and
a buffer memory circuit configured to operate independently of the active state and the sleep state of the first circuitry, to store the first input voice instruction while the first circuitry operates in the sleep state, and to supply the first input voice instruction stored while the first circuitry operates in the sleep state to the first circuitry when the first circuitry is in the active state, wherein
the second circuitry is configured to compare the second input voice instruction with the plurality of application keywords, determine when the application is finished executing, and, when the application is finished executing, to determine whether a security time requiring another input of the first input voice instruction has elapsed since the activation signal is outputted.
12. An information processing method performed by an information processing apparatus, the method comprising:
receiving a first input voice instruction and a second input voice instruction from a microphone disposed in a mobile terminal device of the information processing apparatus, the second input voice instruction being subsequent to the first input voice instruction;
operating a first circuitry disposed in the mobile terminal device in an active state;
determining, by the first circuitry, the first input voice instruction as an activation keyword spoken by a registered user and outputting an activation signal when the first input voice instruction is determined as the activation keyword and as having been spoken by the registered user, the activation signal not being outputted when the first input voice instruction is determined as the activation keyword and as having been spoken by an un-registered user;
maintaining a second circuitry disposed in the mobile terminal device in a sleep state while the first circuitry is in the active state until the second circuitry receives the activation signal and activating the second circuitry from the sleep state into the active state, by the activation signal triggered by the spoken activation keyword of the registered user; and
executing an application by the second circuitry corresponding to the second voice input instruction after entering the active state, the second input voice instruction being spoken by the un-registered user, wherein the un-registered user is a different person than the registered user.
13. A non-transitory computer-readable medium including computer program instructions, which when executed by an information processing apparatus, cause the information processing apparatus to perform a process, the process comprising:
receiving, from a microphone disposed in a mobile terminal device of the information processing apparatus, a first input voice instruction and a second input voice instruction from a microphone, the second input voice instruction being subsequent to the first input voice instruction;
operating a first circuitry disposed in the mobile terminal device in an active state;
determining, by the first circuitry, the first input voice instruction as an activation keyword spoken by a registered user and outputting an activation signal when the first input voice instruction matches the activation keyword and as determined to have been spoken by the registered user, the activation signal not being outputted when the first input voice instruction is determined as the activation keyword and as having been spoken by an un-registered user;
maintaining a second circuitry disposed in the mobile terminal device in a sleep state while the first circuitry is in the active state until the second circuitry receives the activation signal and activating the second circuitry from the sleep state into the active state, by the activation signal triggered by the spoken activation keyword of the registered user; and
executing an application by the second circuitry corresponding to the second voice instruction after entering the active state, the second input voice instruction being spoken by the un-registered user, wherein the un-registered user is a different person than the registered user.
US13/780,0102013-02-282013-02-28Device and method for activating with voice inputActive2033-11-23US10395651B2 (en)

Priority Applications (7)

Application NumberPriority DateFiling DateTitle
US13/780,010US10395651B2 (en)2013-02-282013-02-28Device and method for activating with voice input
EP21193046.6AEP3989043A1 (en)2013-02-282013-05-08Device and method for activating with voice input
EP17205856.2AEP3324404A1 (en)2013-02-282013-05-08Device and method for activating with voice input
EP18166758.5AEP3379530A1 (en)2013-02-282013-05-08Device and method for activating with voice input
EP13166981.4AEP2772907A1 (en)2013-02-282013-05-08Device for activating with voice input
US16/505,737US10825457B2 (en)2013-02-282019-07-09Device and method for activating with voice input
US17/026,323US11580976B2 (en)2013-02-282020-09-21Device and method for activating with voice input

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
US13/780,010US10395651B2 (en)2013-02-282013-02-28Device and method for activating with voice input

Related Child Applications (1)

Application NumberTitlePriority DateFiling Date
US16/505,737ContinuationUS10825457B2 (en)2013-02-282019-07-09Device and method for activating with voice input

Publications (2)

Publication NumberPublication Date
US20140244269A1 US20140244269A1 (en)2014-08-28
US10395651B2true US10395651B2 (en)2019-08-27

Family

ID=48463724

Family Applications (3)

Application NumberTitlePriority DateFiling Date
US13/780,010Active2033-11-23US10395651B2 (en)2013-02-282013-02-28Device and method for activating with voice input
US16/505,737ActiveUS10825457B2 (en)2013-02-282019-07-09Device and method for activating with voice input
US17/026,323Active2033-11-24US11580976B2 (en)2013-02-282020-09-21Device and method for activating with voice input

Family Applications After (2)

Application NumberTitlePriority DateFiling Date
US16/505,737ActiveUS10825457B2 (en)2013-02-282019-07-09Device and method for activating with voice input
US17/026,323Active2033-11-24US11580976B2 (en)2013-02-282020-09-21Device and method for activating with voice input

Country Status (2)

CountryLink
US (3)US10395651B2 (en)
EP (4)EP3379530A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20210005201A1 (en)*2013-02-282021-01-07Sony CorporationDevice and method for activating with voice input
US11900948B1 (en)*2013-08-012024-02-13Amazon Technologies, Inc.Automatic speaker identification using speech recognition features

Families Citing this family (117)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US9584642B2 (en)*2013-03-122017-02-28Google Technology Holdings LLCApparatus with adaptive acoustic echo control for speakerphone mode
US10304465B2 (en)*2012-10-302019-05-28Google Technology Holdings LLCVoice control user interface for low power mode
US9703350B2 (en)*2013-03-152017-07-11Maxim Integrated Products, Inc.Always-on low-power keyword spotting
US9467785B2 (en)2013-03-282016-10-11Knowles Electronics, LlcMEMS apparatus with increased back volume
US9503814B2 (en)2013-04-102016-11-22Knowles Electronics, LlcDifferential outputs in multiple motor MEMS devices
US20140343949A1 (en)*2013-05-172014-11-20Fortemedia, Inc.Smart microphone device
US9633655B1 (en)2013-05-232017-04-25Knowles Electronics, LlcVoice sensing and keyword analysis
US10028054B2 (en)2013-10-212018-07-17Knowles Electronics, LlcApparatus and method for frequency detection
US20180317019A1 (en)2013-05-232018-11-01Knowles Electronics, LlcAcoustic activity detecting microphone
US9712923B2 (en)2013-05-232017-07-18Knowles Electronics, LlcVAD detection microphone and method of operating the same
US9711166B2 (en)2013-05-232017-07-18Knowles Electronics, LlcDecimation synchronization in a microphone
US10020008B2 (en)2013-05-232018-07-10Knowles Electronics, LlcMicrophone and corresponding digital interface
US20140358535A1 (en)*2013-05-282014-12-04Samsung Electronics Co., Ltd.Method of executing voice recognition of electronic device and electronic device using the same
US9747899B2 (en)2013-06-272017-08-29Amazon Technologies, Inc.Detecting self-generated wake expressions
CN110096253B (en)*2013-07-112022-08-30英特尔公司Device wake-up and speaker verification with identical audio input
US20150031416A1 (en)*2013-07-232015-01-29Motorola Mobility LlcMethod and Device For Command Phrase Validation
US9386370B2 (en)2013-09-042016-07-05Knowles Electronics, LlcSlew rate control apparatus for digital microphones
US9928851B2 (en)*2013-09-122018-03-27Mediatek Inc.Voice verifying system and voice verifying method which can determine if voice signal is valid or not
US9502028B2 (en)2013-10-182016-11-22Knowles Electronics, LlcAcoustic activity detection apparatus and method
US9147397B2 (en)2013-10-292015-09-29Knowles Electronics, LlcVAD detection apparatus and method of operating the same
US8768712B1 (en)*2013-12-042014-07-01Google Inc.Initiating actions based on partial hotwords
EP3084760A4 (en)*2013-12-202017-08-16Intel CorporationTransition from low power always listening mode to high power speech recognition mode
EP2891974A1 (en)*2014-01-062015-07-08Samsung Electronics Co., LtdDisplay apparatus which operates in response to voice commands and control method thereof
US10031000B2 (en)*2014-05-292018-07-24Apple Inc.System on a chip with always-on processor
CN105280180A (en)*2014-06-112016-01-27中兴通讯股份有限公司Terminal control method, device, voice control device and terminal
CN105490989A (en)*2014-09-182016-04-13中兴通讯股份有限公司Method for logging into terminal application program and terminal
US9831844B2 (en)2014-09-192017-11-28Knowles Electronics, LlcDigital microphone with adjustable gain control
CN105677004A (en)*2014-11-182016-06-15中兴通讯股份有限公司Terminal processing method and terminal
US9712915B2 (en)2014-11-252017-07-18Knowles Electronics, LlcReference microphone for non-linear and time variant echo cancellation
KR102299330B1 (en)2014-11-262021-09-08삼성전자주식회사Method for voice recognition and an electronic device thereof
US9812126B2 (en)*2014-11-282017-11-07Microsoft Technology Licensing, LlcDevice arbitration for listening devices
CN104866274B (en)*2014-12-012018-06-01联想(北京)有限公司Information processing method and electronic equipment
US9779725B2 (en)2014-12-112017-10-03Mediatek Inc.Voice wakeup detecting device and method
US9775113B2 (en)2014-12-112017-09-26Mediatek Inc.Voice wakeup detecting device with digital microphone and associated method
US9652017B2 (en)*2014-12-172017-05-16Qualcomm IncorporatedSystem and method of analyzing audio data samples associated with speech recognition
DE112016000287T5 (en)2015-01-072017-10-05Knowles Electronics, Llc Use of digital microphones for low power keyword detection and noise reduction
WO2016118480A1 (en)2015-01-212016-07-28Knowles Electronics, LlcLow power voice trigger for acoustic apparatus and method
US10121472B2 (en)2015-02-132018-11-06Knowles Electronics, LlcAudio buffer catch-up apparatus and method with two microphones
US9866938B2 (en)2015-02-192018-01-09Knowles Electronics, LlcInterface for microphone-to-microphone communications
WO2016157650A1 (en)*2015-03-312016-10-06ソニー株式会社Information processing device, control method, and program
US9866741B2 (en)*2015-04-202018-01-09Jesse L. WobrockSpeaker-dependent voice-activated camera system
US10291973B2 (en)2015-05-142019-05-14Knowles Electronics, LlcSensor device with ingress protection
US9883270B2 (en)2015-05-142018-01-30Knowles Electronics, LlcMicrophone with coined area
US9478234B1 (en)2015-07-132016-10-25Knowles Electronics, LlcMicrophone apparatus and method with catch-up buffer
US10045104B2 (en)2015-08-242018-08-07Knowles Electronics, LlcAudio calibration using a microphone
CN105204743A (en)*2015-09-282015-12-30百度在线网络技术(北京)有限公司Interaction control method and device for speech and video communication
US9894437B2 (en)2016-02-092018-02-13Knowles Electronics, LlcMicrophone assembly with pulse density modulated signal
US9811314B2 (en)2016-02-222017-11-07Sonos, Inc.Metadata exchange involving a networked playback system and a networked microphone system
US9826306B2 (en)2016-02-222017-11-21Sonos, Inc.Default playback device designation
US10095470B2 (en)2016-02-222018-10-09Sonos, Inc.Audio response playback
US10264030B2 (en)2016-02-222019-04-16Sonos, Inc.Networked microphone device control
US9978390B2 (en)2016-06-092018-05-22Sonos, Inc.Dynamic player selection for audio signal processing
US10499150B2 (en)2016-07-052019-12-03Knowles Electronics, LlcMicrophone assembly with digital feedback loop
US10134399B2 (en)2016-07-152018-11-20Sonos, Inc.Contextualization of voice inputs
US10257616B2 (en)2016-07-222019-04-09Knowles Electronics, LlcDigital microphone assembly with improved frequency response and noise characteristics
US10115400B2 (en)2016-08-052018-10-30Sonos, Inc.Multiple voice services
US9942678B1 (en)2016-09-272018-04-10Sonos, Inc.Audio playback settings for voice interaction
US10181323B2 (en)2016-10-192019-01-15Sonos, Inc.Arbitration-based voice recognition
WO2018081278A1 (en)2016-10-282018-05-03Knowles Electronics, LlcTransducer assemblies and methods
WO2018097969A1 (en)*2016-11-222018-05-31Knowles Electronics, LlcMethods and systems for locating the end of the keyword in voice sensing
CN108231089B (en)*2016-12-092020-11-03百度在线网络技术(北京)有限公司Speech processing method and device based on artificial intelligence
CN110100259A (en)2016-12-302019-08-06美商楼氏电子有限公司Microphone assembly with certification
US11183181B2 (en)2017-03-272021-11-23Sonos, Inc.Systems and methods of multiple voice services
US10475449B2 (en)2017-08-072019-11-12Sonos, Inc.Wake-word detection suppression
WO2019051218A1 (en)2017-09-082019-03-14Knowles Electronics, LlcClock synchronization in a master-slave communication system
US10048930B1 (en)2017-09-082018-08-14Sonos, Inc.Dynamic computation of system response volume
US10446165B2 (en)2017-09-272019-10-15Sonos, Inc.Robust short-time fourier transform acoustic echo cancellation during audio playback
US10051366B1 (en)2017-09-282018-08-14Sonos, Inc.Three-dimensional beam forming with a microphone array
US10482868B2 (en)2017-09-282019-11-19Sonos, Inc.Multi-channel acoustic echo cancellation
US10466962B2 (en)2017-09-292019-11-05Sonos, Inc.Media playback system with voice assistance
US11061642B2 (en)2017-09-292021-07-13Knowles Electronics, LlcMulti-core audio processor with flexible memory allocation
US10880650B2 (en)2017-12-102020-12-29Sonos, Inc.Network microphone devices with automatic do not disturb actuation capabilities
US10818290B2 (en)2017-12-112020-10-27Sonos, Inc.Home graph
US11175880B2 (en)2018-05-102021-11-16Sonos, Inc.Systems and methods for voice-assisted media content selection
US10959029B2 (en)2018-05-252021-03-23Sonos, Inc.Determining and adapting to changes in microphone performance of playback devices
US10681460B2 (en)2018-06-282020-06-09Sonos, Inc.Systems and methods for associating playback devices with voice assistant services
US11076035B2 (en)2018-08-282021-07-27Sonos, Inc.Do not disturb feature for audio notifications
US10461710B1 (en)2018-08-282019-10-29Sonos, Inc.Media playback system with maximum volume setting
KR102563314B1 (en)2018-08-302023-08-04삼성전자주식회사Electronic Device and the Method for Generating Short cut of Quick Command
US11438682B2 (en)2018-09-112022-09-06Knowles Electronics, LlcDigital microphone with reduced processing noise
US10587430B1 (en)2018-09-142020-03-10Sonos, Inc.Networked devices, systems, and methods for associating playback devices based on sound codes
US11024331B2 (en)2018-09-212021-06-01Sonos, Inc.Voice detection optimization using sound metadata
US10811015B2 (en)2018-09-252020-10-20Sonos, Inc.Voice detection optimization based on selected voice assistant service
US11145301B1 (en)*2018-09-252021-10-12Amazon Technologies, Inc.Communication with user presence
US11100923B2 (en)2018-09-282021-08-24Sonos, Inc.Systems and methods for selective wake word detection using neural network models
US10692518B2 (en)2018-09-292020-06-23Sonos, Inc.Linear filtering for noise-suppressed speech detection via multiple network microphone devices
US10908880B2 (en)2018-10-192021-02-02Knowles Electronics, LlcAudio signal circuit with in-place bit-reversal
US11899519B2 (en)2018-10-232024-02-13Sonos, Inc.Multiple stage network microphone device with reduced power consumption and processing load
EP3654249A1 (en)2018-11-152020-05-20SnipsDilated convolutions and gating for efficient keyword spotting
US11183183B2 (en)2018-12-072021-11-23Sonos, Inc.Systems and methods of operating media playback systems having multiple voice assistant services
CN109410944B (en)*2018-12-122020-06-09百度在线网络技术(北京)有限公司Voice interaction method, device and terminal
US11132989B2 (en)2018-12-132021-09-28Sonos, Inc.Networked microphone devices, systems, and methods of localized arbitration
US11637546B2 (en)*2018-12-142023-04-25Synaptics IncorporatedPulse density modulation systems and methods
US10602268B1 (en)2018-12-202020-03-24Sonos, Inc.Optimization of network microphone devices using noise classification
KR20200084730A (en)*2019-01-032020-07-13삼성전자주식회사Electronic device and control method thereof
US10867604B2 (en)2019-02-082020-12-15Sonos, Inc.Devices, systems, and methods for distributed voice processing
WO2020204907A1 (en)2019-04-012020-10-08Google LlcAdaptive management of casting requests and/or user inputs at a rechargeable device
US11120794B2 (en)2019-05-032021-09-14Sonos, Inc.Voice assistant persistence across multiple network microphone devices
US10871943B1 (en)2019-07-312020-12-22Sonos, Inc.Noise classification for event detection
US11138969B2 (en)2019-07-312021-10-05Sonos, Inc.Locally distributed keyword detection
US11189286B2 (en)2019-10-222021-11-30Sonos, Inc.VAS toggle based on device orientation
US11200900B2 (en)2019-12-202021-12-14Sonos, Inc.Offline voice control
US11562740B2 (en)2020-01-072023-01-24Sonos, Inc.Voice verification for media playback
US11556307B2 (en)2020-01-312023-01-17Sonos, Inc.Local voice data processing
US11308958B2 (en)2020-02-072022-04-19Sonos, Inc.Localized wakeword verification
CN111369999B (en)*2020-03-122024-05-14北京百度网讯科技有限公司Signal processing method and device and electronic equipment
US11308962B2 (en)2020-05-202022-04-19Sonos, Inc.Input detection windowing
US11482224B2 (en)2020-05-202022-10-25Sonos, Inc.Command keywords with input detection windowing
US12387716B2 (en)2020-06-082025-08-12Sonos, Inc.Wakewordless voice quickstarts
US11698771B2 (en)2020-08-252023-07-11Sonos, Inc.Vocal guidance engines for playback devices
US12283269B2 (en)2020-10-162025-04-22Sonos, Inc.Intent inference in audiovisual communication sessions
US11984123B2 (en)2020-11-122024-05-14Sonos, Inc.Network device interaction by range
CN115484501B (en)*2021-06-162024-12-13广州视源电子科技股份有限公司 Program management and recommendation method, system, device and medium based on wake-up word
EP4564154A3 (en)2021-09-302025-07-23Sonos Inc.Conflict management for wake-word detection processes
EP4409933A1 (en)2021-09-302024-08-07Sonos, Inc.Enabling and disabling microphones and voice assistants
US12327549B2 (en)2022-02-092025-06-10Sonos, Inc.Gatekeeping for voice intent processing
CN117351956A (en)*2023-12-052024-01-05广州一链通互联网科技有限公司 Freight trajectory generation and query methods

Citations (17)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20060074658A1 (en)*2004-10-012006-04-06Siemens Information And Communication Mobile, LlcSystems and methods for hands-free voice-activated devices
JP2009177440A (en)2008-01-242009-08-06Nec CorpMobile phone unit and control method thereof
US20090232471A1 (en)*2008-03-122009-09-17Hitachi, Ltd.Information Recording Apparatus
WO2010078386A1 (en)2008-12-302010-07-08Raymond KoverzinPower-optimized wireless communications device
US20110044438A1 (en)*2009-08-202011-02-24T-Mobile Usa, Inc.Shareable Applications On Telecommunications Devices
US20110145000A1 (en)*2009-10-302011-06-16Continental Automotive GmbhApparatus, System and Method for Voice Dialogue Activation and/or Conduct
WO2012025784A1 (en)2010-08-232012-03-01Nokia CorporationAn audio user interface apparatus and method
US20120259641A1 (en)*1998-12-232012-10-11Nuance Communications, Inc.Methods and apparatus for initiating actions using a voice-controlled interface
US20130132089A1 (en)*2011-01-072013-05-23Nuance Communications, Inc.Configurable speech recognition system using multiple recognizers
US20130179168A1 (en)*2012-01-092013-07-11Samsung Electronics Co., Ltd.Image display apparatus and method of controlling the same
US20130197914A1 (en)*2012-01-262013-08-01Microtechnologies Llc D/B/A MicrotechVoice activated audio control system and associated method of use
US20130223635A1 (en)*2012-02-272013-08-29Cambridge Silicon Radio LimitedLow power audio detection
US20130289994A1 (en)*2012-04-262013-10-31Michael Jack NewmanEmbedded system for construction of small footprint speech recognition with user-definable constraints
US20140229184A1 (en)*2013-02-142014-08-14Google Inc.Waking other devices for additional data
US9396327B2 (en)*2011-05-162016-07-19D2L CorporationSystems and methods for security verification in electronic learning systems and other systems
US9423870B2 (en)*2012-05-082016-08-23Google Inc.Input determination method
US9613623B2 (en)*2011-09-262017-04-04Lg Electronics Inc.User interface method and device comprising repeated output of an audible signal and a visual display and vibration for user notification

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US10395651B2 (en)*2013-02-282019-08-27Sony CorporationDevice and method for activating with voice input

Patent Citations (18)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20120259641A1 (en)*1998-12-232012-10-11Nuance Communications, Inc.Methods and apparatus for initiating actions using a voice-controlled interface
US20060074658A1 (en)*2004-10-012006-04-06Siemens Information And Communication Mobile, LlcSystems and methods for hands-free voice-activated devices
JP2009177440A (en)2008-01-242009-08-06Nec CorpMobile phone unit and control method thereof
US20090232471A1 (en)*2008-03-122009-09-17Hitachi, Ltd.Information Recording Apparatus
US20120010890A1 (en)2008-12-302012-01-12Raymond Clement KoverzinPower-optimized wireless communications device
WO2010078386A1 (en)2008-12-302010-07-08Raymond KoverzinPower-optimized wireless communications device
US20110044438A1 (en)*2009-08-202011-02-24T-Mobile Usa, Inc.Shareable Applications On Telecommunications Devices
US20110145000A1 (en)*2009-10-302011-06-16Continental Automotive GmbhApparatus, System and Method for Voice Dialogue Activation and/or Conduct
WO2012025784A1 (en)2010-08-232012-03-01Nokia CorporationAn audio user interface apparatus and method
US20130132089A1 (en)*2011-01-072013-05-23Nuance Communications, Inc.Configurable speech recognition system using multiple recognizers
US9396327B2 (en)*2011-05-162016-07-19D2L CorporationSystems and methods for security verification in electronic learning systems and other systems
US9613623B2 (en)*2011-09-262017-04-04Lg Electronics Inc.User interface method and device comprising repeated output of an audible signal and a visual display and vibration for user notification
US20130179168A1 (en)*2012-01-092013-07-11Samsung Electronics Co., Ltd.Image display apparatus and method of controlling the same
US20130197914A1 (en)*2012-01-262013-08-01Microtechnologies Llc D/B/A MicrotechVoice activated audio control system and associated method of use
US20130223635A1 (en)*2012-02-272013-08-29Cambridge Silicon Radio LimitedLow power audio detection
US20130289994A1 (en)*2012-04-262013-10-31Michael Jack NewmanEmbedded system for construction of small footprint speech recognition with user-definable constraints
US9423870B2 (en)*2012-05-082016-08-23Google Inc.Input determination method
US20140229184A1 (en)*2013-02-142014-08-14Google Inc.Waking other devices for additional data

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
European Office Action dated Feb. 13, 2018 in European Application No. 17205856.2 (6 pages).
Extended European Search Report dated Jun. 15, 2018 in corresponding European Patent Application No. 18166758.5, 8 pages.
Extended European Search Report dated May 20, 2014 in Patent Application No. 13166981.4.

Cited By (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20210005201A1 (en)*2013-02-282021-01-07Sony CorporationDevice and method for activating with voice input
US11580976B2 (en)*2013-02-282023-02-14Sony CorporationDevice and method for activating with voice input
US11900948B1 (en)*2013-08-012024-02-13Amazon Technologies, Inc.Automatic speaker identification using speech recognition features

Also Published As

Publication numberPublication date
US20140244269A1 (en)2014-08-28
EP3324404A1 (en)2018-05-23
US20190333509A1 (en)2019-10-31
EP2772907A1 (en)2014-09-03
EP3379530A1 (en)2018-09-26
EP3989043A1 (en)2022-04-27
US20210005201A1 (en)2021-01-07
US11580976B2 (en)2023-02-14
US10825457B2 (en)2020-11-03

Similar Documents

PublicationPublication DateTitle
US11580976B2 (en)Device and method for activating with voice input
US11393472B2 (en)Method and apparatus for executing voice command in electronic device
US20200395012A1 (en)Electronic device and method of performing functions of electronic devices by voice therebetween
CN108021305B (en) Method, device and mobile terminal for application association startup
KR102346302B1 (en)Electronic apparatus and Method of operating voice recognition in the electronic apparatus
US11217244B2 (en)System for processing user voice utterance and method for operating same
US11538328B2 (en)Mobile device self-identification system
CN108712566B (en) A kind of voice assistant wake-up method and mobile terminal
CN104008635A (en)Equipment control method and device
CN111149136B (en)Electronic device and method for controlling electronic device
US20150153827A1 (en)Controlling connection of input device to electronic devices
CN104681025B (en)Utilize the order execution system and its method of work of speech recognition
EP3547310A1 (en)Electronic device for processing user voice
CN109446775A (en)A kind of acoustic-controlled method and electronic equipment
US20180367669A1 (en)Input during conversational session
US10311254B2 (en)Electronic apparatus and information access control method thereof
CN106095224B (en)A kind of method and mobile terminal for starting application
CN108806670A (en)Audio recognition method, device and storage medium
CN102917470A (en)Method and device for establishing connection between at least two devices by utilizing sound
US12114377B2 (en)Electronic device and method for connecting device thereof
CN111352667A (en) Information push method and electronic device
CN105553985A (en)Synchronous web access method, network communication equipment and network server
KR101357413B1 (en)Communication apparatus having voice recognition function
CN114415903A (en)Data entry method, device, medium and equipment
JP2021081619A (en)Terminal device for search, search system, and search method

Legal Events

DateCodeTitleDescription
ASAssignment

Owner name:SONY MOBILE COMMUNICATIONS AB, SWEDEN

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TOKUTAKE, KENJI;REEL/FRAME:034021/0843

Effective date:20141009

STPPInformation on status: patent application and granting procedure in general

Free format text:RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

ASAssignment

Owner name:SONY CORPORATION, JAPAN

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SONY MOBILE COMMUNICATIONS AB;REEL/FRAME:048713/0511

Effective date:20190326

STPPInformation on status: patent application and granting procedure in general

Free format text:NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPPInformation on status: patent application and granting procedure in general

Free format text:PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCFInformation on status: patent grant

Free format text:PATENTED CASE

MAFPMaintenance fee payment

Free format text:PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment:4


[8]ページ先頭

©2009-2025 Movatter.jp