Movatterモバイル変換


[0]ホーム

URL:


USRE47049E1 - Microphone array system - Google Patents

Microphone array system
Download PDF

Info

Publication number
USRE47049E1
USRE47049E1US15/293,626US201615293626AUSRE47049EUS RE47049 E1USRE47049 E1US RE47049E1US 201615293626 AUS201615293626 AUS 201615293626AUS RE47049 EUSRE47049 EUS RE47049E
Authority
US
United States
Prior art keywords
sound
sensors
signals
target sound
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US15/293,626
Inventor
Manli Zhu
Qi Li
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
LI CREATIVE TECHNOLOGIES Inc
Li Creative Tech Inc
Original Assignee
Li Creative Tech Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filedlitigationCriticalhttps://patents.darts-ip.com/?family=45870681&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=USRE47049(E1)"Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by Li Creative Tech IncfiledCriticalLi Creative Tech Inc
Priority to US15/293,626priorityCriticalpatent/USRE47049E1/en
Priority to US16/052,623prioritypatent/USRE48371E1/en
Application grantedgrantedCritical
Publication of USRE47049E1publicationCriticalpatent/USRE47049E1/en
Assigned to VOCALIFE LLCreassignmentVOCALIFE LLCASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: LI, QI, ZHU, MANLI
Assigned to LI CREATIVE TECHNOLOGIES INCreassignmentLI CREATIVE TECHNOLOGIES INCASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: VOCALIFE LLC
Activelegal-statusCriticalCurrent
Adjusted expirationlegal-statusCritical

Links

Images

Classifications

Definitions

Landscapes

Abstract

A method and system for enhancing a target sound signal from multiple sound signals is provided. An array of an arbitrary number of sound sensors positioned in an arbitrary configuration receives the sound signals from multiple disparate sources. The sound signals comprise the target sound signal from a target sound source, and ambient noise signals. A sound source localization unit, an adaptive beamforming unit, and a noise reduction unit are in operative communication with the array of sound sensors. The sound source localization unit estimates a spatial location of the target sound signal from the received sound signals. The adaptive beamforming unit performs adaptive beamforming by steering a directivity pattern of the array of sound sensors in a direction of the spatial location of the target sound signal, thereby enhancing the target sound signal and partially suppressing the ambient noise signals, which are further suppressed by the noise reduction unit.

Description

CROSS REFERENCE TO RELATED APPLICATIONS
This application is a reissue application of U.S. patent application Ser. No. 13/049,877, filed Mar. 16, 2011 (now U.S. Pat. No. 8,861,756), which claims the benefit of provisional patent application No. 61/403,952 titled “Microphone array design and implementation for telecommunications and handheld devices”, filed on Sep. 24, 2010 in the United States Patent and Trademark Office.
The specification of the above referenced patent application is incorporated herein by reference in its entirety.
BACKGROUND
Microphones constitute an important element in today's speech acquisition devices. Currently, most of the hands-free speech acquisition devices, for example, mobile devices, lapels, headsets, etc., convert sound into electrical signals by using a microphone embedded within the speech acquisition device. However, the paradigm of a single microphone often does not work effectively because the microphone picks up many ambient noise signals in addition to the desired sound, specifically when the distance between a user and the microphone is more than a few inches. Therefore, there is a need for a microphone system that operates under a variety of different ambient noise conditions and that places fewer constraints on the user with respect to the microphone, thereby eliminating the need to wear the microphone or be in close proximity to the microphone.
To mitigate the drawbacks of the single microphone system, there is a need for a microphone array that achieves directional gain in a preferred spatial direction while suppressing ambient noise from other directions. Conventional microphone arrays include arrays that are typically developed for applications such as radar and sonar, but are generally not suitable for hands-free or handheld speech acquisition devices. The main reason is that the desired sound signal has an extremely wide bandwidth relative to its center frequency, thereby rendering conventional narrowband techniques employed in the conventional microphone arrays unsuitable. In order to cater to such broadband speech applications, the array size needs to be vastly increased, making the conventional microphone arrays large and bulky, and precluding the conventional microphone arrays from having broader applications, for example, in mobile and handheld communication devices. There is a need for a microphone array system that provides an effective response over a wide spectrum of frequencies while being unobtrusive in terms of size.
Hence, there is a long felt but unresolved need for a broadband microphone array and broadband beamforming system that enhances acoustics of a desired sound signal while suppressing ambient noise signals.
SUMMARY OF THE INVENTION
This summary is provided to introduce a selection of concepts in a simplified form that are further described in the detailed description of the invention. This summary is not intended to identify key or essential inventive concepts of the claimed subject matter, nor is it intended for determining the scope of the claimed subject matter.
The method and system disclosed herein addresses the above stated need for enhancing acoustics of a target sound signal received from a target sound source, while suppressing ambient noise signals. As used herein, the term “target sound signal” refers to a sound signal from a desired or target sound source, for example, a person's speech that needs to be enhanced. A microphone array system comprising an array of sound sensors positioned in an arbitrary configuration, a sound source localization unit, an adaptive beamforming unit, and a noise reduction unit, is provided. The sound source localization unit, the adaptive beamforming unit, and the noise reduction unit are in operative communication with the array of sound sensors. The array of sound sensors is, for example, a linear array of sound sensors, a circular array of sound sensors, or an arbitrarily distributed coplanar array of sound sensors. The array of sound sensors herein referred to as a “microphone array” receives sound signals from multiple disparate sound sources. The method disclosed herein can be applied on a microphone array with an arbitrary number of sound sensors having, for example, an arbitrary two dimensional (2D) configuration. The sound signals received by the sound sensors in the microphone array comprise the target sound signal from the target sound source among the disparate sound sources, and ambient noise signals.
The sound source localization unit estimates a spatial location of the target sound signal from the received sound signals, for example, using a steered response power-phase transform. The adaptive beamforming unit performs adaptive beamforming for steering a directivity pattern of the microphone array in a direction of the spatial location of the target sound signal. The adaptive beamforming unit thereby enhances the target sound signal from the target sound source and partially suppresses the ambient noise signals. The noise reduction unit suppresses the ambient noise signals for further enhancing the target sound signal received from the target sound source.
In an embodiment where the target sound source that emits the target sound signal is in a two dimensional plane, a delay between each of the sound sensors and an origin of the microphone array is determined as a function of distance between each of the sound sensors and the origin, a predefined angle between each of the sound sensors and a reference axis, and an azimuth angle between the reference axis and the target sound signal. In another embodiment where the target sound source that emits the target sound signal is in a three dimensional plane, the delay between each of the sound sensors and the origin of the microphone array is determined as a function of distance between each of the sound sensors and the origin, a predefined angle between each of the sound sensors and a first reference axis, an elevation angle between a second reference axis and the target sound signal, and an azimuth angle between the first reference axis and the target sound signal. This method of determining the delay enables beamforming for arbitrary numbers of sound sensors and multiple arbitrary microphone array configurations. The delay is determined, for example, in terms of number of samples. Once the delay is determined, the microphone array can be aligned to enhance the target sound signal from a specific direction.
The adaptive beamforming unit comprises a fixed beamformer, a blocking matrix, and an adaptive filter. The fixed beamformer steers the directivity pattern of the microphone array in the direction of the spatial location of the target sound signal from the target sound source for enhancing the target sound signal, when the target sound source is in motion. The blocking matrix feeds the ambient noise signals to the adaptive filter by blocking the target sound signal from the target sound source. The adaptive filter adaptively filters the ambient noise signals in response to detecting the presence or absence of the target sound signal in the sound signals received from the disparate sound sources. The fixed beamformer performs fixed beamforming, for example, by filtering and summing output sound signals from the sound sensors.
In an embodiment, the adaptive filtering comprises sub-band adaptive filtering. The adaptive filter comprises an analysis filter bank, an adaptive filter matrix, and a synthesis filter bank. The analysis filter bank splits the enhanced target sound signal from the fixed beamformer and the ambient noise signals from the blocking matrix into multiple frequency sub-bands. The adaptive filter matrix adaptively filters the ambient noise signals in each of the frequency sub-bands in response to detecting the presence or absence of the target sound signal in the sound signals received from the disparate sound sources. The synthesis filter bank synthesizes a full-band sound signal using the frequency sub-bands of the enhanced target sound signal. In an embodiment, the adaptive beamforming unit further comprises an adaptation control unit for detecting the presence of the target sound signal and adjusting a step size for the adaptive filtering in response to detecting the presence or the absence of the target sound signal in the sound signals received from the disparate sound sources.
The noise reduction unit suppresses the ambient noise signals for further enhancing the target sound signal from the target sound source. The noise reduction unit performs noise reduction, for example, by using a Wiener-filter based noise reduction algorithm, a spectral subtraction noise reduction algorithm, an auditory transform based noise reduction algorithm, or a model based noise reduction algorithm. The noise reduction unit performs noise reduction in multiple frequency sub-bands employed for sub-band adaptive beamforming by the analysis filter bank of the adaptive beamforming unit.
The microphone array system disclosed herein comprising the microphone array with an arbitrary number of sound sensors positioned in arbitrary configurations can be implemented in handheld devices, for example, the iPad® of Apple Inc., the iPhone® of Apple Inc., smart phones, tablet computers, laptop computers, etc. The microphone array system disclosed herein can further be implemented in conference phones, video conferencing applications, or any device or equipment that needs better speech inputs.
BRIEF DESCRIPTION OF THE DRAWINGS
The foregoing summary, as well as the following detailed description of the invention, is better understood when read in conjunction with the appended drawings. For the purpose of illustrating the invention, exemplary constructions of the invention are shown in the drawings. However, the invention is not limited to the specific methods and instrumentalities disclosed herein.
FIG. 1 illustrates a method for enhancing a target sound signal from multiple sound signals.
FIG. 2 illustrates a system for enhancing a target sound signal from multiple sound signals.
FIG. 3 exemplarily illustrates a microphone array configuration showing a microphone array having N sound sensors arbitrarily distributed on a circle.
FIG. 4 exemplarily illustrates a graphical representation of a filter-and-sum beamforming algorithm for determining output of the microphone array having N sound sensors.
FIG. 5 exemplarily illustrates distances between an origin of the microphone array and sound sensor M1and sound sensor M3in the circular microphone array configuration, when the target sound signal is at an angle θ from the Y-axis.
FIG. 6A exemplarily illustrates a table showing the distance between each sound sensor in a circular microphone array configuration from the origin of the microphone array, when the target sound source is in the same plane as that of the microphone array.
FIG. 6B exemplarily illustrates a table showing the relationship of the position of each sound sensor in the circular microphone array configuration and its distance to the origin of the microphone array, when the target sound source is in the same plane as that of the microphone array.
FIG. 7A exemplarily illustrates a graphical representation of a microphone array, when the target sound source is in a three dimensional plane.
FIG. 7B exemplarily illustrates a table showing delay between each sound sensor in a circular microphone array configuration and the origin of the microphone array, when the target sound source is in a three dimensional plane.
FIG. 7C exemplarily illustrates a three dimensional working space of the microphone array, where the target sound signal is incident at an elevation angle Ψ<Ω
FIG. 8 exemplarily illustrates a method for estimating a spatial location of the target sound signal from the target sound source by a sound source localization unit using a steered response power-phase transform.
FIG. 9A exemplarily illustrates a graph showing the value of the steered response power-phase transform for every 10°.
FIG. 9B exemplarily illustrates a graph representing the estimated target sound signal from the target sound source.
FIG. 10 exemplarily illustrates a system for performing adaptive beamforming by an adaptive beamforming unit.
FIG. 11 exemplarily illustrates a system for sub-band adaptive filtering.
FIG. 12 exemplarily illustrates a graphical representation showing the performance of a perfect reconstruction filter bank.
FIG. 13 exemplarily illustrates a block diagram of a noise reduction unit that performs noise reduction using a Wiener-filter based noise reduction algorithm.
FIG. 14 exemplarily illustrates a hardware implementation of the microphone array system.
FIGS. 15A-15C exemplarily illustrate a conference phone comprising an eight-sensor microphone array.
FIG. 16A exemplarily illustrates a layout of an eight-sensor microphone array for a conference phone.
FIG. 16B exemplarily illustrates a graphical representation of eight spatial regions to which the eight-sensor microphone array ofFIG. 16A responds.
FIGS. 16C-16D exemplarily illustrate computer simulations showing the steering of the directivity patterns of the eight-sensor microphone array ofFIG. 16A in the directions of 15° and 60° respectively, in the frequency range 300 Hz to 5 kHz.
FIGS. 16E-16L exemplarily illustrate graphical representations showing the directivity patterns of the eight-sensor microphone array ofFIG. 16A in each of the eight spatial regions, where each directivity pattern is an average response from 300 Hz to 5000 Hz.
FIG. 17A exemplarily illustrates a graphical representation of four spatial regions to which a four-sensor microphone array for a wireless handheld device responds.
FIGS. 17B-17I exemplarily illustrate computer simulations showing the directivity patterns of the four-sensor microphone array ofFIG. 17A with respect to azimuth and frequency.
FIGS. 18A-18B exemplarily illustrate a microphone array configuration for a tablet computer.
FIG. 18C exemplarily illustrates an acoustic beam formed using the microphone array configuration ofFIGS. 18A-18B according to the method and system disclosed herein.
FIGS. 18D-18G exemplarily illustrate graphs showing processing results of the adaptive beamforming unit and the noise reduction unit for the microphone array configuration ofFIG. 18B, in both a time domain and a spectral domain for the tablet computer.
FIGS. 19A-19F exemplarily illustrate tables showing different microphone array configurations and the corresponding values of delay τn, for the sound sensors in each of the microphone array configurations.
DETAILED DESCRIPTION OF THE INVENTION
FIG. 1 illustrates a method for enhancing a target sound signal from multiple sound signals. As used herein, the term “target sound signal” refers to a desired sound signal from a desired or target sound source, for example, a person's speech that needs to be enhanced. The method disclosed herein provides101 a microphone array system comprising an array of sound sensors positioned in an arbitrary configuration, a sound source localization unit, an adaptive beamforming unit, and a noise reduction unit. The sound source localization unit, the adaptive beamforming unit, and the noise reduction unit are in operative communication with the array of sound sensors. The microphone array system disclosed herein employs the array of sound sensors positioned in an arbitrary configuration, the sound source localization unit, the adaptive beamforming unit, and the noise reduction unit for enhancing a target sound signal by acoustic beam forming in the direction of the target sound signal in the presence of ambient noise signals.
The array of sound sensors herein referred to as a “microphone array” comprises multiple or an arbitrary number of sound sensors, for example, microphones, operating in tandem. The microphone array refers to an array of an arbitrary number of sound sensors positioned in an arbitrary configuration. The sound sensors are transducers that detect sound and convert the sound into electrical signals. The sound sensors are, for example, condenser microphones, piezoelectric microphones, etc.
The sound sensors receive102 sound signals from multiple disparate sound sources and directions. The target sound source that emits the target sound signal is one of the disparate sound sources. As used herein, the term “sound signals” refers to composite sound energy from multiple disparate sound sources in an environment of the microphone array. The sound signals comprise the target sound signal from the target sound source and the ambient noise signals. The sound sensors are positioned in an arbitrary planar configuration herein referred to as a “microphone array configuration”, for example, a linear configuration, a circular configuration, any arbitrarily distributed coplanar array configuration, etc. By employing beamforming according to the method disclosed herein, the microphone array provides a higher response to the target sound signal received from a particular direction than to the sound signals from other directions. A plot of the response of the microphone array versus frequency and direction of arrival of the sound signals is referred to as a directivity pattern of the microphone array.
The sound source localization unit estimates103 a spatial location of the target sound signal from the received sound signals. In an embodiment, the sound source localization unit estimates the spatial location of the target sound signal from the target sound source, for example, using a steered response power-phase transform as disclosed in the detailed description ofFIG. 8.
The adaptive beamforming unit performsadaptive beamforming104 by steering the directivity pattern of the microphone array in a direction of the spatial location of the target sound signal, thereby enhancing the target sound signal, and partially suppressing the ambient noise signals. Beamforming refers to a signal processing technique used in the microphone array for directional signal reception, that is, spatial filtering. This spatial filtering is achieved by using adaptive or fixed methods. Spatial filtering refers to separating two signals with overlapping frequency content that originate from different spatial locations.
The noise reduction unit performs noise reduction by further suppressing105 the ambient noise signals and thereby further enhancing the target sound signal. The noise reduction unit performs the noise reduction, for example, by using a Wiener-filter based noise reduction algorithm, a spectral subtraction noise reduction algorithm, an auditory transform based noise reduction algorithm, or a model based noise reduction algorithm.
FIG. 2 illustrates asystem200 for enhancing a target sound signal from multiple sound signals. Thesystem200, herein referred to as a “microphone array system”, comprises thearray201 of sound sensors positioned in an arbitrary configuration, the soundsource localization unit202, theadaptive beamforming unit203, and thenoise reduction unit207.
Thearray201 of sound sensors, herein referred to as the “microphone array” is in operative communication with the soundsource localization unit202, theadaptive beamforming unit203, and thenoise reduction unit207. Themicrophone array201 is, for example, a linear array of sound sensors, a circular array of sound sensors, or an arbitrarily distributed coplanar array of sound sensors. Themicrophone array201 achieves directional gain in any preferred spatial direction and frequency band while suppressing signals from other spatial directions and frequency bands. The sound sensors receive the sound signals comprising the target sound signal and ambient noise signals from multiple disparate sound sources, where one of the disparate sound sources is the target sound source that emits the target sound signal.
The soundsource localization unit202 estimates the spatial location of the target sound signal from the received sound signals. In an embodiment, the soundsource localization unit202 uses, for example, a steered response power-phase transform, for estimating the spatial location of the target sound signal from the target sound source.
Theadaptive beamforming unit203 steers the directivity pattern of themicrophone array201 in a direction of the spatial location of the target sound signal, thereby enhancing the target sound signal and partially suppressing the ambient noise signals. Theadaptive beamforming unit203 comprises a fixedbeamformer204, a blockingmatrix205, and anadaptive filter206 as disclosed in the detailed description ofFIG. 10. The fixedbeamformer204 performs fixed beamforming by filtering and summing output sound signals from each of the sound sensors in themicrophone array201 as disclosed in the detailed description ofFIG. 4. In an embodiment, theadaptive filter206 is implemented as a set of sub-band adaptive filters. Theadaptive filter206 comprises ananalysis filter bank206a, anadaptive filter matrix206b, and asynthesis filter bank206c as disclosed in the detailed description ofFIG. 11.
Thenoise reduction unit207 further suppresses the ambient noise signals for further enhancing the target sound signal. Thenoise reduction unit207 is, for example, a Wiener-filter based noise reduction unit, a spectral subtraction noise reduction unit, an auditory transform based noise reduction unit, or a model based noise reduction unit.
FIG. 3 exemplarily illustrates a microphone array configuration showing amicrophone array201 havingN sound sensors301 arbitrarily distributed on acircle302 with a diameter “d”, where “N” refers to the number ofsound sensors301 in themicrophone array201. Consider an example where N=4, that is, there are four sound sensors301 M0, M1, M2, and M3in themicrophone array201. Each of thesound sensors301 is positioned at an acute angle “Φn” from a Y-axis, where Φ1≥0 and n=0, 1, 2, . . . N−1. In an example, the sound sensor301 M0is positioned at an acute angle Φ0from the Y-axis; the sound sensor301 M1is positioned at an acute angle Φ1from the Y-axis; the sound sensor301 M2is positioned at an acute angle Φ2from the Y-axis; and the sound sensor301 M3is positioned at an acute angle Φ3from the Y-axis. A filter-and-sum beamforming algorithm determines the output “y” of themicrophone array201 havingN sound sensors301 as disclosed in the detailed description ofFIG. 4.
FIG. 4 exemplarily illustrates a graphical representation of the filter-and-sum beamforming algorithm for determining the output of themicrophone array201 havingN sound sensors301. Consider an example where the target sound signal from the target sound source is at an angle θ with a normalized frequency w. The microphone array configuration is arbitrary in a two dimensional plane, for example, a circular array configuration where the sound sensors301 M0, M1, M2, . . . , MN, MN−1of themicrophone array201 are arbitrarily positioned on acircle302. The sound signals received by each of thesound sensors301 in themicrophone array201 are inputs to themicrophone array201. Theadaptive beamforming unit203 employs the filter-and-sum beamforming algorithm that applies independent weights to each of the inputs to themicrophone array201 such that directivity pattern of themicrophone array201 is steered to the spatial location of the target sound signal as determined by the soundsource localization unit202.
The output “y” of themicrophone array201 havingN sound sensors301 is the filter-and-sum of the outputs of theN sound sensors301. That is, y=Σn=0N−1wnTxn, where xnis the output of the (n+1)thsound sensor301, and wnTdenotes a transpose of a length-L filter applied to the (n+1)thsound sensor301.
The spatial directivity pattern H (ω, θ) for the target sound signal from angle θ with normalized frequency w is defined as:
H(ω,θ)=Y(ω,θ)X_(ω,θ)=n=0N-1Wn(ω)Xn(ω,θ)X_(ω,θ)(1)
whereX is the signal received at the origin of thecircular microphone array201 and W is the frequency response of the real-valued finite impulse response (FIR) filter w. If the target sound source is far enough away from themicrophone array201, the difference between the signal received by the (n+1)thsound sensor301 “xn” and the origin of themicrophone array201 is a delay τn; that is, Xn(ω,τ)=X(ω, θ)e−jωτn.
FIG. 5 exemplarily illustrates distances between an origin of themicrophone array201 and the sound sensor301 M1and the sound sensor301 M3in the circular microphone array configuration, when the target sound signal is at an angle θ from the Y-axis. Themicrophone array system200 disclosed herein can be used with an arbitrary directivity pattern for arbitrarily distributedsound sensors301. For any specific microphone array configuration, the parameter that is defined to achieve beamformer coefficients is the value of delay τnfor eachsound sensor301. To define the value of τn, an origin or a reference point of themicrophone array201 is defined; and then the distance dnbetween eachsound sensor301 and the origin is measured, and then the angle Φnof eachsound sensor301 biased from a vertical axis is measured.
For example, the angle between the Y-axis and the line joining the origin and the sound sensor301 M0is Φ0, the angle between the Y-axis and the line joining the origin and the sound sensor301 M1is Φ1, the angle between the Y-axis and the line joining the origin and the sound sensor301 M2is Φ2, and the angle between the Y-axis and the line joining the origin and the sound sensor301 M3is Φ3. The distance between the origin ◯ and the sound sensor301 M1, and the origin ◯ and the sound sensor301 M3when the incoming target sound signal from the target sound source is at an angle θ from the Y-axis is denoted as τ1and τ3, respectively.
For purposes of illustration, the detailed description refers to a circular microphone array configuration; however, the scope of themicrophone array system200 disclosed herein is not limited to the circular microphone array configuration but may be extended to include a linear array configuration, an arbitrarily distributed coplanar array configuration, or a microphone array configuration with any arbitrary geometry.
FIG. 6A exemplarily illustrates a table showing the distance between eachsound sensor301 in a circular microphone array configuration from the origin of themicrophone array201, when the target sound source is in the same plane as that of themicrophone array201. The distance measured in meters and the corresponding delay (τ) measured in number of samples is exemplarily illustrated inFIG. 6A. In an embodiment where the target sound source that emits the target sound signal is in a two dimensional plane, the delay (τ) between each of thesound sensors301 and the origin of themicrophone array201 is determined as a function of distance (d) between each of thesound sensors301 and the origin, a predefined angle (Φ) between each of thesound sensors301 and a reference axis (Y) as exemplarily illustrated inFIG. 5, and an azimuth angle (θ) between the reference axis (Y) and the target sound signal. The determined delay (τ) is represented in terms of number of samples.
If the target sound source is far enough from themicrophone array201, the time delay between the signal received by the (n+1)thsound sensor301 “xn,” and the origin of themicrophone array201 is herein denoted as “t” measured in seconds. The sound signals received by themicrophone array201, which are in analog form are converted into digital sound signals by sampling the analog sound signals at a particular frequency, for example, 8000 Hz. That is, the number of samples in each second is 8000. The delay τ can be represented as the product of the sampling frequency (fs) and the time delay (t). That is, τ=fs*t. Therefore, the distance between thesound sensors301 in themicrophone array201 corresponds to the time used for the target sound signal to travel the distance and is measured by the number of samples within that time period.
Consider an example where “d” is the radius of thecircle302 of the circular microphone array configuration, “fs” is the sampling frequency, and “c” is the speed of sound.FIG. 6B exemplarily illustrates a table showing the relationship of the position of eachsound sensor301 in the circular microphone array configuration and its distance to the origin of themicrophone array201, when the target sound source is in the same plane as that of themicrophone array201. The distance measured in meters and the corresponding delay (τ) measured in number of samples is exemplarily illustrated inFIG. 6B.
The method of determining the delay (τ) enables beamforming for arbitrary numbers ofsound sensors301 and multiple arbitrary microphone array configurations. Once the delay (τ) is determined, themicrophone array201 can be aligned to enhance the target sound signal from a specific direction.
Therefore, the spatial directivity pattern H can be re-written as:
H(ω,θ)=Σn=0N−1Wn(ω)e−jωτn(θ)=wTg(ω,θ)  (2)
where wT=[w0T, w1T, w2T, w3T, . . . , wN−1T] and g(ω,θ)={gi(ω, θ)}i=1 . . . NL={e−jω(k+τn(θ))}i=1 . . . NLis the steering vector, i=1 . . . NL, and k=mod(i−1,L) and n=floor ((i−1)/L).
FIGS. 7A-7C exemplarily illustrate an embodiment of amicrophone array201 when the target sound source is in a three dimensional plane. In an embodiment where the target sound source that emits the target sound signal is in a three dimensional plane, the delay (τ) between each of thesound sensors301 and the origin of themicrophone array201 is determined as a function of distance (d) between each of thesound sensors301 and the origin, a predefined angle (Φ) between each of thesound sensors301 and a first reference axis (Y), an elevation angle (Ψ) between a second reference axis (Z) and the target sound signal, and an azimuth angle (θ) between the first reference axis (Y) and the target sound signal. The determined delay (τ) is represented in terms of number of samples. The determination of the delay enables beamforming for arbitrary numbers of thesound sensors301 and multiple arbitrary configurations of themicrophone array201.
Consider an example of a microphone array configuration with four sound sensors301 M0, M1, M2, and M3.FIG. 7A exemplarily illustrates a graphical representation of amicrophone array201, when the target sound source in a three dimensional plane. As exemplarily illustrated inFIG. 7A, the target sound signal from the target sound source is received from the direction (Ψ, θ) with reference to the origin of themicrophone array201, where Ψ is the elevation angle and θ is the azimuth.
FIG. 7B exemplarily illustrates a table showing delay between eachsound sensor301 in a circular microphone array configuration and the origin of themicrophone array201, when the target sound source is in a three dimensional plane. The target sound source in a three dimensional plane emits a target sound signal from a spatial location (Ψ, θ). The distances between the origin ◯ and the sound sensors301 M0, M1, M2, and M3when the incoming target sound signal from the target sound source is at an angle (Ψ, θ) from the Z-axis and the Y-axis respectively, are denoted as τ0, τ1, τ2, and τ3respectively. When the spatial location of the target sound signal moves from the location Ψ=90° to a location Ψ=0°, sin(Ψ) changes from 1 to 0, and as a result, the difference between eachsound sensor301 in themicrophone array201 becomes smaller and smaller. When Ψ=0°, there is no difference between thesound sensors301, which implies that the target sound signal reaches eachsound sensor301 at the same time. Taking into account that the sample delay between thesound sensors301 can only be an integer, the range where all thesound sensors301 are identical is determined.
FIG. 7C exemplarily illustrates a three dimensional working space of themicrophone array201, where the target sound signal is incident at an elevation angle Ψ<Ω, where Ω is a specific angle and is a variable representing the elevation angle. When the target sound signal is incident at an elevation angle Ψ<Ω, all four sound sensors301 M0, M1, M2, and M3receive the same target sound signal for 0°<0<360°. The delay τ is a function of both the elevation angle Ψ and the azimuth angle θ. That is, τ=τ(θ, Ψ). As used herein, Ω refers to the elevation angle such that all τi(θ, Ω) are equal to each other, where i=0, 1, 2, 3, etc. The value of Ω is determined by the sample delay between each of thesound sensors301 and the origin of themicrophone array201. Theadaptive beamforming unit203 enhances sound from this range and suppresses sound signals from other directions, for example, S1and S2treating them as ambient noise signals.
Consider a least mean square solution for beamforming according to the method disclosed herein. Let the spatial directivity pattern be 1 in the passband and 0 in the stopband. The least square cost function is defined as:
J(w)=ΩpΘpH(ω,θ)-12dωdθ+αΩsΘsH(ω,θ)2dωdθ=ΩpΘpH(ω,θ)2dωdθ+αΩsΘxH(ω,θ)2dωdθ-2ΩpΘpRe(H(ω,θ))dωdθ+ΩpΘp1dωdθ(3)
Replacing
|H(ω,θ)|2=wTg(ω,θ)gH(ω,θ)w=wT(GR(ω,θ)+jG1(ω,θ))w=wTGR(ω,θ)w and Re(H(ω,θ))=wTgR(ω,θ), J(ω) becomes
J(ω)=wTQw−2wTα+d, where
Q=∫ΩP73PGR(ω,θ)dωdθ+αθΩSΘSGR(ω,θ)dωdθ
α=∫ΩPΘPgR(ω,θ)dωdθ
d=∫ΩPΘP1dωdθ
where gR(ω,θ)=cos [w(k+τn)] and GR(ω,θ)=cos [w(k−1+τn−τm)].
When ∂J/∂w=0, the cost function J is minimized. The least-square estimate of w is obtained by:
w=Q−1α  (5)
Applying linear constrains Cw=b, the spatial response is further constrained to a predefined value b at angle θfusing following equation:
[gRT(ωstart,θf)gRT(ωend,θf)]w=[bstartbend](6)
Now, the design problem becomes:
minwwTQw-2wTa+dsubjecttoCw=b(7)
and the solution of the constrained minimization problem is equal to:
w=Q−1CT(CQ−1CT)−1(b−CQ−1α)+Q−1α  (8)
where w is the filter parameter for the designedadaptive beamforming unit203.
In an embodiment, the beamforming is performed by a delay-sum method. In another embodiment, the beamforming is performed by a filter-sum method.
FIG. 8 exemplarily illustrates a method for estimating a spatial location of the target sound signal from the target sound source by the soundsource localization unit202 using a steered response power-phase transform (SRP-PHAT). The SRP-PHAT combines the advantages of sound source localization methods, for example, the time difference of arrival (TDOA) method and the steered response power (SRP) method. The TDOA method performs the time delay estimation of the sound signals relative to a pair of spatially separatedsound sensors301. The estimated time delay is a function of both the location of the target sound source and the position of each of thesound sensors301 in themicrophone array201. Because the position of each of thesound sensors301 in themicrophone array201 is predefined, once the time delay is estimated, the location of the target sound source can be determined. In the SRP method, a filter-and-sum beamforming algorithm is applied to themicrophone array201 for sound signals in the direction of each of the disparate sound sources. The location of the target sound source corresponds to the direction in which the output of the filter-and-sum beamforming has the largest response power. The TDOA based localization is suitable under low to moderate reverberation conditions. The SRP method requires shorter analysis intervals and exhibits an elevated insensitivity to environmental conditions while not allowing for use under excessive multi-path. The SRP-PHAT method disclosed herein combines the advantages of the TDOA method and the SRP method, has a decreased sensitivity to noise and reverberations compared to the TDOA method, and provides more precise location estimates than existing localization methods.
For direction i (0≤t≤360), the delay Ditis calculated801 between the tthpair of the sound sensors301 (t=1: all pairs). The correlation value corr(Dit) between the tthpair of thesound sensors301 corresponding to the delay of Ditis then calculated802. For the direction i (0≤i≤360), the correlation value is given803 by:
CORRi=t=1ALLPAIRcorr(Dit)
Therefore, the spatial location of the target sound signal is given804 by:
S=argmax0<=i<=360CORRi.
FIGS. 9A-9B exemplarily illustrate graphs showing the results of sound source localization performed using the steered response power-phase transform (SRP-PHAT).FIG. 9A exemplarily illustrates a graph showing the value of the SRP-PHAT for every 10° The maximum value corresponds to the location of the target sound signal from the target sound source.FIG. 9B exemplarily illustrates a graph representing the estimated target sound signal from the target sound source and a ground truth.
FIG. 10 exemplarily illustrates a system for performing adaptive beamforming by theadaptive beamforming unit203. The algorithm for fixed beamforming is disclosed with reference to equations (3) through (8) in the detailed description ofFIG. 4,FIGS. 6A-6B, andFIGS. 7A-7C, which is extended herein to adaptive beamforming. Adaptive beamforming refers to a beamforming process where the directivity pattern of themicrophone array201 is adaptively steered in the direction of a target sound signal emitted by a target sound source in motion. Adaptive beamforming achieves better ambient noise suppression than fixed beamforming. This is because the target direction of arrival, which is assumed to be stable in fixed beamforming, changes with the movement of the target sound source. Moreover, the gains of thesound sensors301 which are assumed uniform in fixed beamforming, exhibit significant distribution. All these factors reduce speech quality. On the other hand, adaptive beamforming adaptively performs beam steering and null steering; therefore, the adaptive beamforming method is more robust against steering error caused by the array imperfection mentioned above.
As exemplarily illustrated inFIG. 10, theadaptive beamforming unit203 disclosed herein comprises a fixedbeamformer204, a blockingmatrix205, an adaptation control unit208, and anadaptive filter206. The fixedbeamformer204 adaptively steers the directivity pattern of themicrophone array201 in the direction of the spatial location of the target sound signal from the target sound source for enhancing the target sound signal, when the target sound source is in motion. Thesound sensors301 in themicrophone array201 receive the sound signals S1, . . . , S4, which comprise both the target sound signal from the target sound source and the ambient noise signals. The received sound signals are fed as input to the fixedbeamformer204 and the blockingmatrix205. The fixedbeamformer204 outputs a signal “b”. In an embodiment, the fixedbeamformer204 performs fixed beamforming by filtering and summing output sound signals from thesound sensors301. The blockingmatrix205 outputs a signal “z” which primarily comprises the ambient noise signals. The blockingmatrix205 blocks the target sound signal from the target sound source and feeds the ambient noise signals to theadaptive filter206 to minimize the effect of the ambient noise signals on the enhanced target sound signal.
The output “z” of the blockingmatrix205 may contain some weak target sound signals due to signal leakage. If the adaptation is active when the target sound signal, for example, speech is present, the speech is cancelled out with the noise. Therefore, the adaptation control unit208 determines when the adaptation should be applied. The adaptation control unit208 comprises a targetsound signal detector208a and a step size adjusting module208b. The targetsound signal detector208a of the adaptation control unit208 detects the presence or absence of the target sound signal, for example, speech. The step size adjusting module208b adjusts the step size for the adaptation process such that when the target sound signal is present, the adaptation is slow for preserving the target sound signal, and when the target sound signal is absent, adaptation is quick for better cancellation of the ambient noise signals.
Theadaptive filter206 is a filter that adaptively updates filter coefficients of theadaptive filter206 so that theadaptive filter206 can be operated in an unknown and changing environment. Theadaptive filter206 adaptively filters the ambient noise signals in response to detecting presence or absence of the target sound signal in the sound signals received from the disparate sound sources. Theadaptive filter206 adapts its filter coefficients with the changes in the ambient noise signals, thereby eliminating distortion in the target sound signal, when the target sound source and the ambient noise signals are in motion. In an embodiment, the adaptive filtering is performed by a set of sub-band adaptive filters using sub-band adaptive filtering as disclosed in the detailed description ofFIG. 11.
FIG. 11 exemplarily illustrates a system for sub-band adaptive filtering. Sub-band adaptive filtering involves separating a full-band signal into different frequency ranges called sub-bands prior to the filtering process. The sub-band adaptive filtering using sub-band adaptive filters lead to a higher convergence speed compared to using a full-band adaptive filter. Moreover, thenoise reduction unit207 disclosed herein is developed in a sub-band, whereby applying sub-band adaptive filtering provides the same sub-band framework for both beamforming and noise reduction, and thus saves on computational cost.
As exemplarily illustrated inFIG. 11, theadaptive filter206 comprises ananalysis filter bank206a, anadaptive filter matrix206b, and asynthesis filter bank206c. Theanalysis filter bank206a splits the enhanced target sound signal (b) from the fixedbeamformer204 and the ambient noise signals (z) from the blockingmatrix205 exemplarily illustrated inFIG. 10 into multiple frequency sub-bands. Theanalysis filter bank206a performs an analysis step where the outputs of the fixedbeamformer204 and the blockingmatrix205 are split into frequency sub bands. The sub-bandadaptive filter206 typically has a shorter impulse response than its full band counterpart. The step size of the sub-bands can be adjusted individually for each sub-band by the step-size adjusting module208b, which leads to a higher convergence speed compared to using a full band adaptive filter.
Theadaptive filter matrix206b adaptively filters the ambient noise signals in each of the frequency sub-bands in response to detecting the presence or absence of the target sound signal in the sound signals received from the disparate sound sources. Theadaptive filter matrix206b performs an adaptation step, where theadaptive filter206 is adapted such that the filter output only contains the target sound signal, for example, speech. Thesynthesis filter bank206c synthesizes a full-band sound signal using the frequency sub-bands of the enhanced target sound signal. Thesynthesis filter bank206c performs a synthesis step where the sub-band sound signal is synthesized into a full-band sound signal. Since the noise reduction and the beamforming are performed in the same sub-band framework, the noise reduction as disclosed in the detailed description ofFIG. 13, by thenoise reduction unit207 is performed prior to the synthesis step, thereby reducing computation.
In an embodiment, theanalysis filter bank206a is implemented as a perfect-reconstruction filter bank, where the output of thesynthesis filter bank206c after the analysis and synthesis steps perfectly matches the input to theanalysis filter bank206a. That is, all the sub-bandanalysis filter banks206a are factorized to operate on prototype filter coefficients and a modulation matrix is used to take advantage of the fast Fourier transform (FFT). Both analysis and synthesize steps require performing frequency shifts in each sub-band, which involves complex value computations with cosines and sinusoids. The method disclosed herein employs the FFT to perform the frequency shifts required in each sub-band, thereby minimizing the amount of multiply-accumulate operations. The implementation of the sub-bandanalysis filter bank206a as a perfect-reconstruction filter bank ensures the quality of the target sound signal by ensuring that the sub-bandanalysis filter banks206a do not distort the target sound signal itself.
FIG. 12 exemplarily illustrates a graphical representation showing the performance of a perfect-reconstruction filter bank. The solid line represents the input signal to theanalysis filter bank206a, and the circles represent the output of thesynthesis filter bank206c after analysis and synthesis. As exemplarily illustrated inFIG. 12, the output of thesynthesis filter bank206c perfectly matches the input, and is therefore referred to as the perfect-reconstruction filter bank.
FIG. 13 exemplarily illustrates a block diagram of anoise reduction unit207 for performing noise reduction using, for example, a Wiener-filter based noise reduction algorithm. Thenoise reduction unit207 performs noise reduction for further suppressing the ambient noise signals after adaptive beamforming, for example, by using a Wiener-filter based noise reduction algorithm, a spectral subtraction noise reduction algorithm, an auditory transform based noise reduction algorithm, or a model based noise reduction algorithm. In an embodiment, thenoise reduction unit207 performs noise reduction in multiple frequency sub-bands employed by ananalysis filter bank206a of theadaptive beamforming unit203 for sub-band adaptive beamforming.
In an embodiment, the noise reduction is performed using the Wiener-filter based noise reduction algorithm. Thenoise reduction unit207 explores the short-term and long-term statistics of the target sound signal, for example, speech, and the ambient noise signals, and the wide-band and narrow-band signal-to-noise ratio (SNR) to support a Wiener gain filtering. Thenoise reduction unit207 comprises a target sound signal statistics analyzer207a, anoise statistics analyzer207b, a signal-to-noise ratio (SNR) analyzer207c, and a Wiener filter207d. The target soundsignal statistics analyzer207a explores the short-term and long-term statistics of the target sound signal, for example, speech. Similarly, thenoise statistics analyzer207b explores the short-term and long-term statistics of the ambient noise signals. The SNR analyzer207c of thenoise reduction unit207 explores the wide-band and narrow-band signal-to-noise ratio (SNR). After the spectrum of noisy-speech passes through the Wiener filter207d, an estimation of the clean-speech spectrum is generated. Thesynthesis filter bank206c, by an inverse process of theanalysis filter bank206a, reconstructs the signals of the clean speech into a full-band signal, given the estimated spectrum of the clean speech.
FIG. 14 exemplarily illustrates a hardware implementation of themicrophone array system200 disclosed herein. The hardware implementation of themicrophone array system200 disclosed in the detailed description ofFIG. 2 comprises themicrophone array201 having an arbitrary number ofsound sensors301 positioned in an arbitrary configuration,multiple microphone amplifiers1401, one or moreaudio codecs1402, a digital signal processor (DSP)1403, aflash memory1404, one ormore power regulators1405 and1406, abattery1407, a loudspeaker or aheadphone1408, and acommunication interface1409. Themicrophone array201 comprises, for example, four or eightsound sensors301 arranged in a linear or a circular microphone array configuration. Themicrophone array201 receives the sound signals.
Consider an example where themicrophone array201 comprises foursound sensors301 that pick up the sound signals. Fourmicrophone amplifiers1401 receive the output sound signals from the foursound sensors301. Themicrophone amplifiers1401 also referred to as preamplifiers provide a gain to boost the power of the received sound signals for enhancing the sensitivity of thesound sensors301. In an example, the gain of the preamplifiers is 20 dB.
Theaudio codec1402 receives the amplified output from themicrophone amplifiers1401. Theaudio codec1402 provides an adjustable gain level, for example, from about −74 dB to about 6 dB. The received sound signals are in an analog form. Theaudio codec1402 converts the four channels of the sound signals in the analog form into digital sound signals. The pre-amplifiers may not be required for some applications. Theaudio codec1402 then transmits the digital sound signals to theDSP1403 for processing of the digital sound signals. TheDSP1403 implements the soundsource localization unit202, theadaptive beamforming unit203, and thenoise reduction unit207.
After the processing, theDSP1403 either stores the processed signal from theDSP1403 in a memory device for a recording application, or transmits the processed signal to thecommunication interface1409. The recording application comprises, for example, storing the processed signal onto the memory device for the purposes of playing back the processed signal at a later time. Thecommunication interface1409 transmits the processed signal, for example, to a computer, the internet, or a radio for communicating the processed signal. In an embodiment, themicrophone array system200 disclosed herein implements a two-way communication device where the signal received from thecommunication interface1409 is processed by theDSP1403 and the processed signal is then played through the loudspeaker or theheadphone1408.
Theflash memory1404 stores the code for theDSP1403 and compressed audio signals. When themicrophone array system200 boots up, theDSP1403 reads the code from theflash memory1404 into an internal memory of theDSP1403 and then starts executing the code. In an embodiment, theaudio codec1402 can be configured for encoding and decoding audio or sound signals during the start up stage by writing to registers of theDSP1403. For an eight-sensor microphone array201, two four-channel audio codec1402 chips may be used. Thepower regulators1405 and1406, for example,linear power regulators1405 and switchpower regulators1406 provide appropriate voltage and current supply for all the components, for example,201,1401,1402,1403, etc., mechanically supported and electrically connected on a circuit board. A universal serial bus (USB) control is built into theDSP1403. Thebattery1407 is used for powering themicrophone array system200.
Consider an example where themicrophone array system200 disclosed herein is implemented on a mixed signal circuit board having a six-layer printed circuit board (PCB). Noisy digital signals easily contaminate the low voltage analog sound signals from thesound sensors301. Therefore, the layout of the mixed signal circuit board is carefully partitioned to isolate the analog circuits from the digital circuits. Although both the inputs and outputs of themicrophone amplifiers1401 are in analog form, themicrophone amplifiers1401 are placed in a digital region of the mixed signal circuit board because of theirhigh power consumption1401 and switch amplifier nature.
Thelinear power regulators1405 are deployed in an analog region of the mixed signal circuit board due to the low noise property exhibited by thelinear power regulators1405. Five power regulators, for example,1405 are designed in themicrophone array system200 circuits to ensure quality. Theswitch power regulators1406 achieve an efficiency of about 95% of the input power and have high output current capacity; however their outputs are too noisy for analog circuits. The efficiency of thelinear power regulators1405 is determined by the ratio of the output voltage to the input voltage, which is lower than that of theswitch power regulators1406 in most cases. The regulator outputs utilized in themicrophone array system200 circuits are stable, quiet, and suitable for the low power analog circuits.
In an example, themicrophone array system200 is designed with amicrophone array201 having dimensions of 10 cm×2.5 cm×1.5 cm, a USB interface, and an assembled PCB supporting themicrophone array201 and aDSP1403 having a low power consumption design devised for portable devices, a four-channel codec1402, and aflash memory1404. TheDSP1403 chip is powerful enough to handle theDSP1403 computations in themicrophone array system200 disclosed herein. The hardware configuration of this example can be used for any microphone array configuration, with suitable modifications to the software. In an embodiment, theadaptive beamforming unit203 of themicrophone array system200 is implemented as hardware with software instructions programmed on theDSP1403. TheDSP1403 is programmed for beamforming, noise reduction, echo cancellation, and USB interfacing according to the method disclosed herein, and fine tuned for optimal performance.
FIGS. 15A-15C exemplarily illustrate aconference phone1500 comprising an eight-sensor microphone array201. The eight-sensor microphone array201 comprises eightsound sensors301 arranged in a configuration as exemplarily illustrated inFIG. 15A. A top view of theconference phone1500 comprising the eight-sensor microphone array201 is exemplarily illustrated inFIG. 15A. A front view of theconference phone1500 comprising the eight-sensor microphone array201 is exemplarily illustrated inFIG. 15B. Aheadset1502 that can be placed in abase holder1501 of theconference phone1500 having the eight-sensor microphone array201 is exemplarily illustrated inFIG. 15C. In addition to aconference phone1500, themicrophone array system200 disclosed herein with broadband beamforming can be configured for a mobile phone, a tablet computer, etc., for speech enhancement and noise reduction.
FIG. 16A exemplarily illustrates a layout of an eight-sensor microphone array201 for aconference phone1500. Consider an example of acircular microphone array201 in which eightsound sensors301 are mounted on the surface of theconference phone1500 as exemplarily illustrated inFIG. 15A. Theconference phone1500 has aremovable handset1502 on top, and hence themicrophone array system200 is configured to accommodate thehandset1502 as exemplarily illustrated inFIGS. 15A-15C. In an example, thecircular microphone array201 has a diameter of about four inches. Eightsound sensors301, for example, microphones, M0, M1, M2, M3, M4, M5, M6, and M7are distributed along acircle302 on theconference phone1500. Microphones M4-M7are separated by 90 degrees from each other, and microphones M0-M3are rotated counterclockwise by 60 degrees from microphone M4-M7respectively.
FIG. 16B exemplarily illustrates a graphical representation of eight spatial regions to which the eight-sensor microphone array201 ofFIG. 16A responds. The space is divided into eight spatial regions with equal spaces centered at 15°, 60°, 105°, 150°, 195°, 240°, 285°, and 330° respectively. Theadaptive beamforming unit203 configures the eight-sensor microphone array201 to automatically point to one of these eight spatial regions according to the location of the target sound signal from the target sound source as estimated by the soundsource localization unit202.
FIGS. 16C-16D exemplarily illustrate computer simulations showing the steering of the directivity patterns of the eight-sensor microphone array201 ofFIG. 16A, in the directions 15° and 60° respectively, in the frequency range 300 Hz to 5 kHz.FIG. 16C exemplarily illustrates the computer simulation result showing the directivity pattern of themicrophone array201 when the target sound signal is received from the target sound source in the spatial region centered at 15°.
The computer simulation for verifying the performance of theadaptive beamforming unit203 when the target sound signal is received from the target sound source in the spatial region centered at 15° uses the following parameters:
Sampling frequency fs=16 k,
FIR filter taper length L=20
Passband (Θp, Ωp)={300-5000 Hz, −5°-35°}, designed spatial directivity pattern is 1.
Stopband (Θs, Ωs)={300˜5000 Hz,−180°˜−15°+45°˜180°}, the designed spatial directivity pattern is 0.
It can be seen that the directivity pattern of themicrophone array201 in the spatial region centered at 15° is enhanced while the sound signals from all other spatial regions are suppressed.
FIG. 16D exemplarily illustrates the computer simulation result showing the directivity pattern of themicrophone array201 when the target sound signal is received from the target sound source in the spatial region centered at 60°. The computer simulation for verifying the performance of theadaptive beamforming unit203 when the target sound signal is received from the target sound source in the spatial region centered at 60° uses the following parameters:
Sampling frequency fs=16 k,
FIR filter taper length L=20
Passband (Θp, Ωp)={300-5000 Hz, 40°-80°}, designed spatial directivity pattern is 1.
Stopband (Θs, Ωs)={300˜5000 Hz, −180°˜30°+90°˜180°}, the designed spatial directivity pattern is 0.
It can be seen that the directivity pattern of themicrophone array201 in the spatial region centered at 60° is enhanced while the sound signals from all other spatial regions are suppressed. The other six spatial regions have similar parameters. Moreover, in all frequencies, the main lobe has the same level, which means the target sound signal has little distortion in frequency.
FIGS. 16E-16L exemplarily illustrate graphical representations showing the directivity patterns of the eight-sensor microphone array201 ofFIG. 16A in each of the eight spatial regions, where each directivity pattern is an average response from 300 Hz to 5000 Hz. The main lobe is about 10 dB higher than the side lobe, and therefore the ambient noise signals from other directions are highly suppressed compared to the target sound signal in the pass direction. Themicrophone array system200 calculates the filter coefficients for the target sound signal, for example, speech signals from eachsound sensor301 and combines the filtered signals to enhance the speech from any specific direction. Since speech covers a large range of frequencies, the method andsystem200 disclosed herein covers broadband signals from 300 Hz to 5000 Hz.
FIG. 16E exemplarily illustrates a graphical representation showing the directivity pattern of the eight-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 15°.FIG. 16F exemplarily illustrates a graphical representation showing the directivity pattern of the eight-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 60°.FIG. 16G exemplarily illustrates a graphical representation showing the directivity pattern of the eight-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 105°.FIG. 16H exemplarily illustrates a graphical representation showing the directivity pattern of the eight-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 150°.FIG. 16I exemplarily illustrates a graphical representation showing the directivity pattern of the eight-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 195°.FIG. 16J exemplarily illustrates a graphical representation showing the directivity pattern of the eight-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 240°.FIG. 16K exemplarily illustrates a graphical representation showing the directivity pattern of the eight-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 285°.FIG. 16L exemplarily illustrates a graphical representation showing the directivity pattern of the eight-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 330°. Themicrophone array system200 disclosed herein enhances the target sound signal from each of the directions 15°, 60°, 105°, 150°, 195°, 240°, 285°, and 330°, while suppressing the ambient noise signals from the other directions.
Themicrophone array system200 disclosed herein can be implemented for a square microphone array configuration and a rectangular array configuration where asound sensor301 is positioned in each corner of the four-cornered array. Themicrophone array system200 disclosed herein implements beamforming from plane to three dimensional sound sources.
FIG. 17A exemplarily illustrates a graphical representation of four spatial regions to which a four-sensor microphone array201 for a wireless handheld device responds. The wireless handheld device is, for example, a mobile phone. Consider an example where themicrophone array201 comprises foursound sensors301, for example, microphones, uniformly distributed around acircle302 having diameter equal to about two inches. This configuration is identical to positioning foursound sensors301 or microphones on four corners of a square. The space is divided into four spatial regions with equal space centered at −90°, 0°, 90°, and 180° respectively. Theadaptive beamforming unit203 configures the four-sensor microphone array201 to automatically point to one of these spatial regions according to the location of the target sound signal from the target sound source as estimated by the soundsource localization unit202.
FIGS. 17B-17I exemplarily illustrate computer simulations showing the directivity patterns of the four-sensor microphone array201 ofFIG. 17A with respect to azimuth and frequency. The results of the computer simulations performed for verifying the performance of theadaptive beamforming unit203 of themicrophone array system200 disclosed herein for a sampling frequency fs=16 k and FIR filter taper length L=20, are as follows:
For the spatial region centered at 0°:
Passband (Θp, Ωp)={300-4000 Hz, −20°-20°}, designed spatial directivity pattern is 1.
Stopband (Θ, Ωs)={300˜4000 Hz, −180°˜−30°+30°˜180°}, the designed spatial directivity pattern is 0.
For the spatial region centered at 90°:
Passband (Θp, Ωp)={300-4000 Hz, 70°-110°}, designed spatial directivity pattern is 1.
Stopband (Θs, Ωs)={300˜4000 Hz, −180°˜60°+120°˜180°}, the designed spatial directivity pattern is 0. The directivity patterns for the spatial regions centered at −90° and 180° are similarly obtained.
FIG. 17B exemplarily illustrates the computer simulation result representing a three dimensional (3D) display of the directivity pattern of the four-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at −90°.FIG. 17C exemplarily illustrates the computer simulation result representing a 2D display of the directivity pattern of the four-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at −90°.
FIG. 17D exemplarily illustrates the computer simulation result representing a 3D display of the directivity pattern of the four-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 0°.FIG. 17E exemplarily illustrates the computer simulation result representing a 2D display of the directivity pattern of the four-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 0°.
FIG. 17F exemplarily illustrates the computer simulation result representing a 3D display of the directivity pattern of the four-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 90°.FIG. 17G exemplarily illustrates the computer simulation result representing a 2D display of the directivity pattern of the four-sensor microphone array201 when the target sound signal is received from the target sound source in the spatial region centered at 90°.
FIG. 17H exemplarily illustrates the computer simulation result representing a 3D display of the directivity pattern of the four-sensor microphone array201 when the target sound source is received from the target sound source in the spatial region centered at 180°.FIG. 17I exemplarily illustrates the computer simulation result representing a 2D display of the directivity pattern of the four-sensor microphone array201 when the target sound source is received from the target sound source in the spatial region centered at 180°. The 3D displays of the directivity patterns inFIG. 17B,FIG. 17D,FIG. 17F, andFIG. 17H demonstrate that the passbands have the same height. The 2D displays of the directivity patterns inFIG. 17C,FIG. 17E,FIG. 17G, andFIG. 17I demonstrate that the passbands have the same width along the frequency and demonstrates the broadband properties of themicrophone array201.
FIGS. 18A-18B exemplarily illustrates a microphone array configuration for a tablet computer. In this example, foursound sensors301 of themicrophone array201 are positioned on aframe1801 of the tablet computer, for example, the iPad® of Apple Inc. Geometrically, thesound sensors301 are distributed on thecircle302 as exemplarily inFIG. 18B. The radius of thecircle302 is equal to the width of the tablet computer. The angle θ between the sound sensors301 M2and M3is determined to avoid spatial aliasing up to 4000 Hz. This microphone array configuration enhances a front speaker's voice and suppresses background ambient noise. Theadaptive beamforming unit203 configures themicrophone array201 to form anacoustic beam1802 pointing frontwards using the method andsystem200 disclosed herein. The target sound signal, that is, the front speaker's voice within the range of Φ<30° is enhanced compared to the sound signals from other directions.
FIG. 18C exemplarily illustrates anacoustic beam1802 formed using the microphone array configuration ofFIGS. 18A-18B according to the method andsystem200 disclosed herein.
FIGS. 18D-18G exemplarily illustrates graphs showing processing results of theadaptive beamforming unit203 and thenoise reduction unit207 for the microphone array configuration ofFIG. 18B, in both a time domain and a spectral domain for the tablet computer. Consider an example where a speaker is talking in front of the tablet computer with ambient noise signals on the side.FIG. 18D exemplarily illustrates a graph showing the performance of themicrophone array201 before performing beamforming and noise reduction with a signal-to-noise ratio (SNR) of 15 dB.FIG. 18E exemplarily illustrates a graph showing the performance of themicrophone array201 after performing beamforming and noise reduction, according to the method disclosed herein, with an SNR of 15 dB.FIG. 18F exemplarily illustrates a graph showing the performance of themicrophone array201 before performing beamforming and noise reduction with an SNR of 0 dB.FIG. 18G exemplarily illustrates a graph showing the performance of themicrophone array201 after performing beamforming and noise reduction, according to the method disclosed herein, with an SNR of 0 dB.
It can be seen fromFIGS. 18D-18G that the performance graph is noisier for themicrophone array201 before the beamforming and noise reduction is performed. Therefore, theadaptive beamforming unit203 and thenoise reduction unit207 of themicrophone array system200 disclosed herein suppresses ambient noise signals while maintaining the clarity of the target sound signal, for example, the speech signal.
FIGS. 19A-19F exemplarily illustrate tables showing different microphone array configurations and the corresponding values of delay τnfor thesound sensors301 in each of the microphone array configurations. The broadband beamforming method disclosed herein can be used formicrophone arrays201 with arbitrary numbers ofsound sensors301 and arbitrary locations of thesound sensors301. Thesound sensors301 can be mounted on surfaces or edges of any speech acquisition device. For any specific microphone array configuration, the only parameter that needs to be defined to achieve the beamformer coefficients is the value of τnfor eachsound sensor301 as disclosed in the detailed description ofFIG. 5,FIGS. 6A-6B, andFIGS. 7A-7C and as exemplarily illustrated inFIGS. 19A-19F. In an example, the microphone array configuration exemplarily illustrated inFIG. 19F is implemented on a handheld device for hands-free speech acquisition. In a hands-free and non-close talking scenario, a user prefers to talk in distance rather than speaking close to thesound sensor301 and may want to talk while watching a screen of the handheld device. Themicrophone array system200 disclosed herein allows the handheld device to pick up sound signals from the direction of the speaker's mouth and suppress noise from other directions. The method andsystem200 disclosed herein may be implemented on any device or equipment, for example, a voice recorder where a target sound signal or speech needs to be enhanced.
The foregoing examples have been provided merely for the purpose of explanation and are in no way to be construed as limiting of the present invention disclosed herein. While the invention has been described with reference to various embodiments, it is understood that the words, which have been used herein, are words of description and illustration, rather than words of limitation. Further, although the invention has been described herein with reference to particular means, materials and embodiments, the invention is not intended to be limited to the particulars disclosed herein; rather, the invention extends to all functionally equivalent structures, methods and uses, such as are within the scope of the appended claims. Those skilled in the art, having the benefit of the teachings of this specification, may affect numerous modifications thereto and changes may be made without departing from the scope and spirit of the invention in its aspects.

Claims (35)

We claim:
1. A method for enhancing a target sound signal from a plurality of sound signals, comprising:
providing a microphone array system comprising an array of sound sensors positioned in an arbitrary a linear, circular, or other configuration, a sound source localization unit, an adaptive beamforming unit, and a noise reduction unit, wherein said sound source localization unit, said adaptive beamforming unit, and said noise reduction unit are integrated in a digital signal processor, and wherein said sound source localization unit, said adaptive beamforming unit, and said noise reduction unit are in operative communication with said array of said sound sensors;
receiving said sound signals from a plurality of disparate sound sources by said sound sensors, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
determining a delay between each of said sound sensors and an origin of said array of said sound sensors as a function of distance between each of said sound sensors and said origin, a predefined angle between each of said sound sensors and a reference axis, and an azimuth angle between said reference axis and said target sound signal, when said target sound source that emits said target sound signal is in a two dimensional plane, wherein said delay is represented in terms of number of samples, and wherein said determination of said delay enables beamforming for arbitrary numbers of said array of sound sensors and in a plurality of arbitrary configurations of said array of said sound sensors;
estimating a spatial location of said target sound signal from said received sound signals by said sound source localization unit;
performing adaptive beamforming for steering a directivity pattern of said array of said sound sensors in a direction of said spatial location of said target sound signal by said adaptive beamforming unit, wherein said adaptive beamforming unit enhances said target sound signal and partially suppresses said ambient noise signals; and
suppressing said ambient noise signals by said noise reduction unit for further enhancing said target sound signal.
2. The method ofclaim 1, wherein said spatial location of said target sound signal from said target sound source is estimated using a steered response power-phase transform by said sound source localization unit.
3. The method ofclaim 1, wherein said adaptive beamforming comprises:
providing a fixed beamformer, a blocking matrix, and an adaptive filter in said adaptive beamforming unit;
steering said directivity pattern of said array of said sound sensors in said direction of said spatial location of said target sound signal from said target sound source by said fixed beamformer for enhancing said target sound signal, when said target sound source is in motion;
feeding said ambient noise signals to said adaptive filter by blocking said target sound signal received from said target sound source using said blocking matrix; and
adaptively filtering said ambient noise signals by said adaptive filter in response to detecting one of presence and absence of said target sound signal in said sound signals received from said disparate sound sources.
4. The method ofclaim 3, wherein said fixed beamformer performs fixed beamforming by filtering and summing output sound signals from said sound sensors.
5. The method ofclaim 3, wherein said adaptive filtering comprises sub-band adaptive filtering performed by said adaptive filter, wherein said sub-band adaptive filtering comprises:
providing an analysis filter bank, an adaptive filter matrix, and a synthesis filter bank in said adaptive filter;
splitting said enhanced target sound signal from said fixed beamformer and said ambient noise signals from said blocking matrix into a plurality of frequency sub-bands by said analysis filter bank;
adaptively filtering said ambient noise signals in each of said frequency sub-bands by said adaptive filter matrix in response to detecting one of presence and absence of said target sound signal in said sound signals received from said disparate sound sources; and
synthesizing a full-band sound signal using said frequency sub-bands of said enhanced target sound signal by said synthesis filter bank.
6. The method ofclaim 3, wherein said adaptive beamforming further comprises detecting said presence of said target sound signal by an adaptation control unit provided in said adaptive beamforming unit and adjusting a step size for said adaptive filtering in response to detecting one of said presence and said absence of said target sound signal in said sound signals received from said disparate sound sources.
7. The method ofclaim 1, wherein said noise reduction unit performs noise reduction by using one of a Wiener-filter based noise reduction algorithm, a spectral subtraction noise reduction algorithm, an auditory transform based noise reduction algorithm, and a model based noise reduction algorithm.
8. The method ofclaim 1, wherein said noise reduction unit performs noise reduction in a plurality of frequency sub-bands, wherein said frequency sub-bands are employed by an analysis filter bank of said adaptive beamforming unit for sub-band adaptive beamforming.
9. A system for enhancing a target sound signal from a plurality of sound signals, comprising:
an array of sound sensors positioned in an arbitrary a linear, circular, or other configuration, wherein said sound sensors receive said sound signals from a plurality of disparate sound sources, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
a digital signal processor, said digital signal processor comprising:
a sound source localization unit that estimates a spatial location of said target sound signal from said received sound signals, by determining a delay between each of said sound sensors and an origin of said array of said sound sensors as a function of distance between each of said sound sensors and said origin, a predefined angle between each of said sound sensors and a reference axis, and an azimuth angle between said reference axis and said target sound signal, when said target sound source that emits said target sound signal is in a two dimensional plane, wherein said delay is represented in terms of number of samples, and wherein said determination of said delay enables beamforming for arbitrary numbers of said array of sound sensors and in a plurality of arbitrary configurations of said array of said sound sensors;
an adaptive beamforming unit that steers directivity pattern of said array of said sound sensors in a direction of said spatial location of said target sound signal, wherein said adaptive beamforming unit enhances said target sound signal and partially suppresses said ambient noise signals; and
a noise reduction unit that suppresses said ambient noise signals for further enhancing said target sound signal.
10. The system ofclaim 9, wherein said sound source localization unit estimates said spatial location of said target sound signal from said target sound source using a steered response power-phase transform.
11. The system ofclaim 9, wherein said adaptive beamforming unit comprises:
a fixed beamformer that steers said directivity pattern of said array of said sound sensors in said direction of said spatial location of said target sound signal from said target sound source for enhancing said target sound signal, when said target sound source is in motion;
a blocking matrix that feeds said ambient noise signals to an adaptive filter by blocking said target sound signal received from said target sound source; and
said adaptive filter that adaptively filters said ambient noise signals in response to detecting one of presence and absence of said target sound signal in said sound signals received from said disparate sound sources.
12. The system ofclaim 11, wherein said fixed beamformer performs fixed beamforming by filtering and summing output sound signals from said sound sensors.
13. The system ofclaim 11, wherein said adaptive filter comprises a set of sub-band adaptive filters comprising:
an analysis filter bank that splits said enhanced target sound signal from said fixed beamformer and said ambient noise signals from said blocking matrix into a plurality of frequency sub-bands;
an adaptive filter matrix that adaptively filters said ambient noise signals in each of said frequency sub-bands in response to detecting one of presence and absence of said target sound signal in said sound signals received from said disparate sound sources; and
a synthesis filter bank that synthesizes a full-band sound signal using said frequency sub-bands of said enhanced target sound signal.
14. The system ofclaim 9, wherein said adaptive beamforming unit further comprises an adaptation control unit that detects said presence of said target sound signal and adjusts a step size for said adaptive filtering in response to detecting one of said presence and said absence of said target sound signal in said sound signals received from said disparate sound sources.
15. The system ofclaim 9, wherein said noise reduction unit is one of a Wiener-filter based noise reduction unit, a spectral subtraction noise reduction unit, an auditory transform based noise reduction unit, and a model based noise reduction unit.
16. The system ofclaim 9, further comprising one or more audio codecs that convert said sound signals in an analog form of said sound signals into digital sound signals and reconverts said digital sound signals into said analog form of said sound signals.
17. The system ofclaim 9, wherein said noise reduction unit performs noise reduction in a plurality of frequency sub-bands employed by an analysis filter bank of said adaptive beamforming unit for sub-band adaptive beamforming.
18. The system ofclaim 9, wherein said array of said sound sensors is one of a linear array of said sound sensors, a circular array of said sound sensors, and an arbitrarily distributed coplanar array of said sound sensors.
19. The method ofclaim 1, wherein said delay (τ) is determined by a formula τ=fs*t, wherein fsis a sampling frequency and t is a time delay calculated based on said number of samples within a time period and a time delay for said target sound signal to travel said distance between each of said sound sensors in said microphone array and said origin of said array of said sound sensors, and wherein said distance between said each of said sound sensors in the microphone array and said origin of said array of said sound sensors can be same or different.
20. A method for enhancing a target sound signal from a plurality of sound signals, comprising:
providing a microphone array system comprising an array of sound sensors positioned in an arbitrary a linear, circular, or other configuration, a sound source localization unit, an adaptive beamforming unit, and a noise reduction unit, wherein said sound source localization unit, said adaptive beamforming unit, and said noise reduction unit are integrated in a digital signal processor, and wherein said sound source localization unit, said adaptive beamforming unit, and said noise reduction unit are in operative communication with said array of said sound sensors;
receiving said sound signals from a plurality of disparate sound sources by said sound sensors, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
determining a delay between each of said sound sensors and an origin of said array of said sound sensors as a function of distance between each of said sound sensors and said origin, a predefined angle between each of said sound sensors and a first reference axis, an elevation angle between a second reference axis and said target sound signal, and an azimuth angle between said first reference axis and said target sound signal, when said target sound source that emits said target sound signal is in a three dimensional plane, wherein said delay is represented in terms of number of samples, and wherein said determination of said delay enables beamforming for arbitrary numbers of said array of sound sensors and in a plurality of arbitrary configurations of said array of said sound sensors;
estimating a spatial location of said target sound signal from said received sound signals by said sound source localization unit;
performing adaptive beamforming for steering a directivity pattern of said array of said sound sensors in a direction of said spatial location of said target sound signal by said adaptive beamforming unit, wherein said adaptive beamforming unit enhances said target sound signal and partially suppresses said ambient noise signals; and
suppressing said ambient noise signals by said noise reduction unit for further enhancing said target sound signal.
21. A system for enhancing a target sound signal from a plurality of sound signals, comprising:
an array of sound sensors positioned in an arbitrary a linear, circular, or other configuration, wherein said sound sensors receive said sound signals from a plurality of disparate sound sources, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
a digital signal processor, said digital signal processor comprising:
a sound source localization unit that estimates a spatial location of said target sound signal from said received sound signals by determining a delay between each of said sound sensors and an origin of said array of said sound sensors as a function of distance between each of said sound sensors and said origin, a predefined angle between each of said sound sensors and a first reference axis, an elevation angle between a second reference axis and said target sound signal, and an azimuth angle between said first reference axis and said target sound signal, when said target sound source that emits said target sound signal is in a three dimensional plane, wherein said delay is represented in terms of number of samples, and wherein said determination of said delay enables beamforming for arbitrary numbers of said array of sound sensors and in a plurality of arbitrary configurations of said array of said sound sensors;
an adaptive beamforming unit that steers directivity pattern of said array of said sound sensors in a direction of said spatial location of said target sound signal, wherein said adaptive beamforming unit enhances said target sound signal and partially suppresses said ambient noise signals; and
a noise reduction unit that suppresses said ambient noise signals for further enhancing said target sound signal.
22. A method for enhancing a target sound signal from a plurality of sound signals, comprising:
providing a microphone array system comprising an array of sound sensors, a sound source localization unit, a beamforming unit, and a noise reduction unit, wherein said sound source localization unit, said beamforming unit, and said noise reduction unit are integrated in a digital signal processor, and wherein said sound source localization unit, said beamforming unit, and said noise reduction unit are in operative communication with said array of said sound sensors;
receiving said sound signals from a plurality of disparate sound sources by said sound sensors, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
determining a delay between each of said sound sensors and a reference point of said array of said sound sensors as a function of distance between each of said sound sensors and said reference point, a predefined angle between each of said sound sensors and a reference axis, and an azimuth angle between said reference axis and said target sound signal, when said target sound source that emits said target sound signal is in a two dimensional plane, wherein said delay is represented in terms of number of samples, and wherein said determination of said delay enables beamforming for two or more of said sound sensors;
estimating a spatial location of said target sound signal from said received sound signals by said sound source localization unit;
performing beamforming for steering a directivity pattern of said array of said sound sensors in a direction of said spatial location of said target sound signal by said beamforming unit, wherein said beamforming unit enhances said target sound signal and partially suppresses said ambient noise signals; and
suppressing said ambient noise signals by said noise reduction unit for further enhancing said target sound signal.
23. The method of claim 22, wherein said beamforming comprises:
providing a fixed beamformer, a blocking matrix, and an adaptive filter in said beamforming unit;
steering said directivity pattern of said array of said sound sensors in said direction of said spatial location of said target sound signal from said target sound source by said fixed beamformer for enhancing said target sound signal, when said target sound source is in motion;
feeding said ambient noise signals to said adaptive filter by blocking said target sound signal received from said target sound source using said blocking matrix; and
adaptively filtering said ambient noise signals by said adaptive filter in response to detecting one of presence and absence of said target sound signal in said sound signals received from said disparate sound sources.
24. The method of claim 23, wherein said beamforming further comprises detecting said presence of said target sound signal by an adaptation control unit provided in said beamforming unit and adjusting a step size for said adaptive filtering in response to detecting one of said presence and said absence of said target sound signal in said sound signals received from said disparate sound sources.
25. The method of claim 22, wherein said noise reduction unit performs noise reduction in a plurality of frequency sub-bands, wherein said frequency sub-bands are employed by an analysis filter bank of said beamforming unit for sub-band adaptive beamforming.
26. A system for enhancing a target sound signal from a plurality of sound signals, comprising:
an array of sound sensors, wherein said sound sensors receive said sound signals from a plurality of disparate sound sources, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
a digital signal processor, said digital signal processor comprising:
a sound source localization unit that estimates a spatial location of said target sound signal from said received sound signals, by determining a delay between each of said sound sensors and a reference point of said array of said sound sensors as a function of distance between each of said sound sensors and said reference point, a predefined angle between each of said sound sensors and a reference axis, and an azimuth angle between said reference axis and said target sound signal, when said target sound source that emits said target sound signal is in a two dimensional plane, wherein said delay is represented in terms of number of samples, and wherein said determination of said delay enables beamforming for two or more of said sound sensors;
a beamforming unit that steers directivity pattern of said array of said sound sensors in a direction of said spatial location of said target sound signal, wherein said beamforming unit enhances said target sound signal and partially suppresses said ambient noise signals; and
a noise reduction unit that suppresses said ambient noise signals for further enhancing said target sound signal.
27. The system of claim 26, wherein said beamforming unit further comprises an adaptation control unit that detects said presence of said target sound signal and adjusts a step size for said adaptive filtering in response to detecting one of said presence and said absence of said target sound signal in said sound signals received from said disparate sound sources.
28. The system of claim 26, wherein said noise reduction unit performs noise reduction in a plurality of frequency sub-bands employed by an analysis filter bank of said beamforming unit for sub-band adaptive beamforming.
29. The system of claim 26, wherein said array of said sound sensors is one of a linear array of said sound sensors, and a circular array of said sound sensors, and other types of array of said sound sensors.
30. A method for enhancing a target sound signal from a plurality of sound signals, comprising:
providing a microphone array system comprising an array of sound sensors, a sound source localization unit, a beamforming unit, and a noise reduction unit, wherein said sound source localization unit, said beamforming unit, and said noise reduction unit are integrated in a digital signal processor, and wherein said sound source localization unit, said beamforming unit, and said noise reduction unit are in operative communication with said array of said sound sensors;
receiving said sound signals from a plurality of disparate sound sources by said sound sensors, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
determining a delay between each of said sound sensors and a reference point of said array of said sound sensors as a function of distance between each of said sound sensors and said reference point, a predefined angle between each of said sound sensors and a first reference axis, an elevation angle between a second reference axis and said target sound signal, and an azimuth angle between said first reference axis and said target sound signal, when said target sound source that emits said target sound signal is in a three dimensional plane, wherein said delay is represented in terms of number of samples, and wherein said determination of said delay enables beamforming for two or more of said sound sensors:
estimating a spatial location of said target sound signal from said received sound signals by said sound source localization unit;
performing beamforming for steering a directivity pattern of said array of said sound sensors in a direction of said spatial location of said target sound signal by said beamforming unit, wherein said beamforming unit enhances said target sound signal and partially suppresses said ambient noise signals; and
suppressing said ambient noise signals by said noise reduction unit for further enhancing said target sound signal.
31. A system for enhancing a target sound signal from a plurality of sound signals, comprising:
an array of sound sensors, wherein said sound sensors receive said sound signals from a plurality of disparate sound sources, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
a digital signal processor, said digital signal processor comprising:
a sound source localization unit that estimates a spatial location of said target sound signal from said received sound signals by determining a delay between each of said sound sensors and a reference point of said array of said sound sensors as a function of distance between each of said sound sensors and said reference point, a predefined angle between each of said sound sensors and a first reference axis, an elevation angle between a second reference axis and said target sound signal, and an azimuth angle between said first reference axis and said target sound signal, when said target sound source that emits said target sound signal is in a three dimensional plane, wherein said delay is represented in terms of number of samples, and wherein said determination of said delay enables beamforming for two or more of said sound sensors;
a beamforming unit that steers directivity pattern of said array of said sound sensors in a direction of said spatial location of said target sound signal, wherein said beamforming unit enhances said target sound signal and partially suppresses said ambient noise signals; and
a noise reduction unit that suppresses said ambient noise signals for further enhancing said target sound signal.
32. A system for enhancing a target sound signal from a plurality of sound signals, comprising:
an array of sound sensors, wherein said sound sensors receive said sound signals from a plurality of disparate sound sources, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
a digital signal processor, said digital signal processor comprising:
a sound source localization unit that estimates a spatial location of said target sound signal from said received sound signals by determining a delay between each of said sound sensors and a reference point of said array of said sound sensors as a function of distance between each of said sound sensors and said reference point and an angle of each of said sound sensors biased from a reference axis;
a beamforming unit that enhances said target sound signal and suppresses said ambient noise signals; and
a noise reduction unit that suppresses said ambient noise signals.
33. A system for enhancing a target sound signal from a plurality of sound signals, comprising:
an array of sound sensors, wherein said sound sensors receive said sound signals from a plurality of disparate sound sources, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
a digital signal processor, said digital signal processor comprising:
a sound source localization unit that estimates a spatial location of said target sound signal from said received sound signals by determining a delay between each of said sound sensors and a reference point of said array of said sound sensors as a function of distance between each of said sound sensors and said reference point, a predefined angle between each of said sound sensors and a reference axis and an azimuth angle between said reference axis and said target sound signal;
a beamforming unit that enhances said target sound signal and suppresses said ambient noise signals; and
a noise reduction unit that suppresses said ambient noise signals.
34. A system for enhancing a target sound signal from a plurality of sound signals, comprising:
an array of sound sensors, wherein said sound sensors receive said sound signals from a plurality of disparate sound sources, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
a digital signal processor, said digital signal processor comprising:
a sound source localization unit that estimates a spatial location of said target sound signal from said received sound signals by determining a delay between each of said sound sensors and a reference point of said array of said sound sensors as a function of distance between each of said sound sensors and said reference point, a predefined angle between each of said sound sensors and a first reference axis, an elevation angle between a second reference axis and said target sound signal and an azimuth angle between said first reference axis and said target sound signal;
a beamforming unit that enhances said target sound signal and suppresses said ambient noise signals; and
a noise reduction unit that suppresses said ambient noise signals.
35. A system for enhancing a target sound signal from a plurality of sound signals, comprising:
an array of sound sensors positioned in a non-circular configuration, wherein said sound sensors receive said sound signals from a plurality of disparate sound sources, wherein said received sound signals comprise said target sound signal from a target sound source among said disparate sound sources, and ambient noise signals;
a digital signal processor, said digital signal processor comprising:
a sound source localization unit that estimates a spatial location of said target sound signal from said received sound signals by determining a delay between each of said sound sensors and a reference point of said array of said sound sensors as a function of distance between each of said sound sensors and said reference point and an angle of each of said sound sensors biased from a reference axis, wherein said distance between each of said sound sensors and said reference point varies from a minimum value to a maximum value, and wherein said minimum value corresponds to zero and said maximum value is defined based on a limitation associated with size of said system;
a beamforming unit that enhances said target sound signal and suppresses said ambient noise signals; and
a noise reduction unit that suppresses said ambient noise signals.
US15/293,6262010-09-242016-10-14Microphone array systemActive2033-05-18USRE47049E1 (en)

Priority Applications (2)

Application NumberPriority DateFiling DateTitle
US15/293,626USRE47049E1 (en)2010-09-242016-10-14Microphone array system
US16/052,623USRE48371E1 (en)2010-09-242018-08-02Microphone array system

Applications Claiming Priority (3)

Application NumberPriority DateFiling DateTitle
US40395210P2010-09-242010-09-24
US13/049,877US8861756B2 (en)2010-09-242011-03-16Microphone array system
US15/293,626USRE47049E1 (en)2010-09-242016-10-14Microphone array system

Related Parent Applications (1)

Application NumberTitlePriority DateFiling Date
US13/049,877ReissueUS8861756B2 (en)2010-09-242011-03-16Microphone array system

Related Child Applications (1)

Application NumberTitlePriority DateFiling Date
US13/049,877ContinuationUS8861756B2 (en)2010-09-242011-03-16Microphone array system

Publications (1)

Publication NumberPublication Date
USRE47049E1true USRE47049E1 (en)2018-09-18

Family

ID=45870681

Family Applications (3)

Application NumberTitlePriority DateFiling Date
US13/049,877CeasedUS8861756B2 (en)2010-09-242011-03-16Microphone array system
US15/293,626Active2033-05-18USRE47049E1 (en)2010-09-242016-10-14Microphone array system
US16/052,623Active - Reinstated2033-05-18USRE48371E1 (en)2010-09-242018-08-02Microphone array system

Family Applications Before (1)

Application NumberTitlePriority DateFiling Date
US13/049,877CeasedUS8861756B2 (en)2010-09-242011-03-16Microphone array system

Family Applications After (1)

Application NumberTitlePriority DateFiling Date
US16/052,623Active - Reinstated2033-05-18USRE48371E1 (en)2010-09-242018-08-02Microphone array system

Country Status (1)

CountryLink
US (3)US8861756B2 (en)

Cited By (83)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20180188347A1 (en)*2016-03-302018-07-05Yutou Technology (Hangzhou) Co., Ltd.Voice direction searching system and method thereof
US20190110153A1 (en)*2017-08-302019-04-11Harman International Industries, IncorporatedEnvironment discovery via time-synchronized networked loudspeakers
US20190324117A1 (en)*2018-04-242019-10-24Mediatek Inc.Content aware audio source localization
US10957428B2 (en)2017-08-102021-03-23Nuance Communications, Inc.Automated clinical documentation system and method
US11043207B2 (en)2019-06-142021-06-22Nuance Communications, Inc.System and method for array data simulation and customized acoustic modeling for ambient ASR
US11100923B2 (en)2018-09-282021-08-24Sonos, Inc.Systems and methods for selective wake word detection using neural network models
US11120794B2 (en)2019-05-032021-09-14Sonos, Inc.Voice assistant persistence across multiple network microphone devices
US11132989B2 (en)2018-12-132021-09-28Sonos, Inc.Networked microphone devices, systems, and methods of localized arbitration
US11133018B2 (en)2016-06-092021-09-28Sonos, Inc.Dynamic player selection for audio signal processing
US11137979B2 (en)2016-02-222021-10-05Sonos, Inc.Metadata exchange involving a networked playback system and a networked microphone system
US11138975B2 (en)2019-07-312021-10-05Sonos, Inc.Locally distributed keyword detection
US11138969B2 (en)2019-07-312021-10-05Sonos, Inc.Locally distributed keyword detection
US11159880B2 (en)2018-12-202021-10-26Sonos, Inc.Optimization of network microphone devices using noise classification
US11175880B2 (en)2018-05-102021-11-16Sonos, Inc.Systems and methods for voice-assisted media content selection
US11175888B2 (en)2017-09-292021-11-16Sonos, Inc.Media playback system with concurrent voice assistance
US11183181B2 (en)2017-03-272021-11-23Sonos, Inc.Systems and methods of multiple voice services
US11183183B2 (en)2018-12-072021-11-23Sonos, Inc.Systems and methods of operating media playback systems having multiple voice assistant services
US11184704B2 (en)2016-02-222021-11-23Sonos, Inc.Music service selection
US11184969B2 (en)2016-07-152021-11-23Sonos, Inc.Contextualization of voice inputs
US11189286B2 (en)2019-10-222021-11-30Sonos, Inc.VAS toggle based on device orientation
US11197096B2 (en)2018-06-282021-12-07Sonos, Inc.Systems and methods for associating playback devices with voice assistant services
US11200894B2 (en)2019-06-122021-12-14Sonos, Inc.Network microphone device with command keyword eventing
US11200889B2 (en)2018-11-152021-12-14Sonos, Inc.Dilated convolutions and gating for efficient keyword spotting
US11200900B2 (en)2019-12-202021-12-14Sonos, Inc.Offline voice control
US11216480B2 (en)2019-06-142022-01-04Nuance Communications, Inc.System and method for querying data points from graph data structures
US11222103B1 (en)2020-10-292022-01-11Nuance Communications, Inc.Ambient cooperative intelligence system and method
US11222716B2 (en)2018-03-052022-01-11Nuance CommunicationsSystem and method for review of automated clinical documentation from recorded audio
US11227679B2 (en)2019-06-142022-01-18Nuance Communications, Inc.Ambient clinical intelligence system and method
US11240621B2 (en)2020-04-112022-02-01LI Creative Technologies, Inc.Three-dimensional audio systems
US11250382B2 (en)2018-03-052022-02-15Nuance Communications, Inc.Automated clinical documentation system and method
US11302326B2 (en)2017-09-282022-04-12Sonos, Inc.Tone interference cancellation
US11308962B2 (en)2020-05-202022-04-19Sonos, Inc.Input detection windowing
US11308961B2 (en)2016-10-192022-04-19Sonos, Inc.Arbitration-based voice recognition
US11308958B2 (en)2020-02-072022-04-19Sonos, Inc.Localized wakeword verification
US11316865B2 (en)2017-08-102022-04-26Nuance Communications, Inc.Ambient cooperative intelligence system and method
US11315556B2 (en)2019-02-082022-04-26Sonos, Inc.Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification
US11343614B2 (en)2018-01-312022-05-24Sonos, Inc.Device designation of playback and network microphone device arrangements
US11354092B2 (en)2019-07-312022-06-07Sonos, Inc.Noise classification for event detection
US11361756B2 (en)2019-06-122022-06-14Sonos, Inc.Conditional wake word eventing based on environment
US11380322B2 (en)2017-08-072022-07-05Sonos, Inc.Wake-word detection suppression
US11405430B2 (en)2016-02-222022-08-02Sonos, Inc.Networked microphone device control
US11432030B2 (en)2018-09-142022-08-30Sonos, Inc.Networked devices, systems, and methods for associating playback devices based on sound codes
US11451908B2 (en)2017-12-102022-09-20Sonos, Inc.Network microphone devices with automatic do not disturb actuation capabilities
US11482978B2 (en)2018-08-282022-10-25Sonos, Inc.Audio notifications
US11482224B2 (en)2020-05-202022-10-25Sonos, Inc.Command keywords with input detection windowing
US11500611B2 (en)2017-09-082022-11-15Sonos, Inc.Dynamic computation of system response volume
US11501773B2 (en)2019-06-122022-11-15Sonos, Inc.Network microphone device with command keyword conditioning
US11501795B2 (en)2018-09-292022-11-15Sonos, Inc.Linear filtering for noise-suppressed speech detection via multiple network microphone devices
US11513763B2 (en)2016-02-222022-11-29Sonos, Inc.Audio response playback
US11516610B2 (en)2016-09-302022-11-29Sonos, Inc.Orientation-based playback device microphone selection
US11514898B2 (en)2016-02-222022-11-29Sonos, Inc.Voice control of a media playback system
US11515020B2 (en)2018-03-052022-11-29Nuance Communications, Inc.Automated clinical documentation system and method
US11531520B2 (en)2016-08-052022-12-20Sonos, Inc.Playback device supporting concurrent voice assistants
US11531807B2 (en)2019-06-282022-12-20Nuance Communications, Inc.System and method for customized text macros
US11538451B2 (en)2017-09-282022-12-27Sonos, Inc.Multi-channel acoustic echo cancellation
US11551700B2 (en)2021-01-252023-01-10Sonos, Inc.Systems and methods for power-efficient keyword detection
US11551690B2 (en)2018-09-142023-01-10Sonos, Inc.Networked devices, systems, and methods for intelligently deactivating wake-word engines
US11556307B2 (en)2020-01-312023-01-17Sonos, Inc.Local voice data processing
US11556306B2 (en)2016-02-222023-01-17Sonos, Inc.Voice controlled media playback system
US11563842B2 (en)2018-08-282023-01-24Sonos, Inc.Do not disturb feature for audio notifications
US11562740B2 (en)2020-01-072023-01-24Sonos, Inc.Voice verification for media playback
US11589329B1 (en)2010-12-302023-02-21Staton Techiya LlcInformation processing using a population of data acquisition devices
US11641559B2 (en)2016-09-272023-05-02Sonos, Inc.Audio playback settings for voice interaction
US11646045B2 (en)2017-09-272023-05-09Sonos, Inc.Robust short-time fourier transform acoustic echo cancellation during audio playback
US11646023B2 (en)2019-02-082023-05-09Sonos, Inc.Devices, systems, and methods for distributed voice processing
US11664023B2 (en)2016-07-152023-05-30Sonos, Inc.Voice detection by multiple devices
US20230169956A1 (en)*2019-05-032023-06-01Sonos, Inc.Locally distributed keyword detection
US11670408B2 (en)2019-09-302023-06-06Nuance Communications, Inc.System and method for review of automated clinical documentation
US11676590B2 (en)2017-12-112023-06-13Sonos, Inc.Home graph
US11698771B2 (en)2020-08-252023-07-11Sonos, Inc.Vocal guidance engines for playback devices
US11715489B2 (en)2018-05-182023-08-01Sonos, Inc.Linear filtering for noise-suppressed speech detection
US11727936B2 (en)2018-09-252023-08-15Sonos, Inc.Voice detection optimization based on selected voice assistant service
US11727919B2 (en)2020-05-202023-08-15Sonos, Inc.Memory allocation for keyword spotting engines
US11790937B2 (en)2018-09-212023-10-17Sonos, Inc.Voice detection optimization using sound metadata
US11792590B2 (en)2018-05-252023-10-17Sonos, Inc.Determining and adapting to changes in microphone performance of playback devices
US11890168B2 (en)2022-03-212024-02-06Li Creative Technologies Inc.Hearing protection and situational awareness system
US11899519B2 (en)2018-10-232024-02-13Sonos, Inc.Multiple stage network microphone device with reduced power consumption and processing load
US11984123B2 (en)2020-11-122024-05-14Sonos, Inc.Network device interaction by range
US12047753B1 (en)2017-09-282024-07-23Sonos, Inc.Three-dimensional beam forming with a microphone array
US12283269B2 (en)2020-10-162025-04-22Sonos, Inc.Intent inference in audiovisual communication sessions
US12327549B2 (en)2022-02-092025-06-10Sonos, Inc.Gatekeeping for voice intent processing
US12327556B2 (en)2021-09-302025-06-10Sonos, Inc.Enabling and disabling microphones and voice assistants
US12387716B2 (en)2020-06-082025-08-12Sonos, Inc.Wakewordless voice quickstarts

Families Citing this family (99)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN102306496B (en)*2011-09-052014-07-09歌尔声学股份有限公司Noise elimination method, device and system of multi-microphone array
US8983089B1 (en)2011-11-282015-03-17Rawles LlcSound source localization using multiple microphone arrays
WO2013093565A1 (en)2011-12-222013-06-27Nokia CorporationSpatial audio processing apparatus
US9437213B2 (en)*2012-03-052016-09-06Malaspina Labs (Barbados) Inc.Voice signal enhancement
US9354295B2 (en)*2012-04-132016-05-31Qualcomm IncorporatedSystems, methods, and apparatus for estimating direction of arrival
US20130343549A1 (en)*2012-06-222013-12-26Verisilicon Holdings Co., Ltd.Microphone arrays for generating stereo and surround channels, method of operation thereof and module incorporating the same
US9384737B2 (en)*2012-06-292016-07-05Microsoft Technology Licensing, LlcMethod and device for adjusting sound levels of sources based on sound source priority
US9232310B2 (en)*2012-10-152016-01-05Nokia Technologies OyMethods, apparatuses and computer program products for facilitating directional audio capture with multiple microphones
US9078057B2 (en)*2012-11-012015-07-07Csr Technology Inc.Adaptive microphone beamforming
US9595997B1 (en)*2013-01-022017-03-14Amazon Technologies, Inc.Adaption-based reduction of echo and noise
US9294839B2 (en)2013-03-012016-03-22Clearone, Inc.Augmentation of a beamforming microphone array with non-beamforming microphones
US10750132B2 (en)*2013-03-142020-08-18Pelco, Inc.System and method for audio source localization using multiple audio sensors
US20140270219A1 (en)*2013-03-152014-09-18CSR Technology, Inc.Method, apparatus, and manufacture for beamforming with fixed weights and adaptive selection or resynthesis
CN104065798B (en)*2013-03-212016-08-03华为技术有限公司Audio signal processing method and equipment
US9294858B2 (en)*2014-02-262016-03-22Revo Labs, Inc.Controlling acoustic echo cancellation while handling a wireless microphone
US9716946B2 (en)*2014-06-012017-07-25Insoundz Ltd.System and method thereof for determining of an optimal deployment of microphones to achieve optimal coverage in a three-dimensional space
US10149047B2 (en)*2014-06-182018-12-04Cirrus Logic Inc.Multi-aural MMSE analysis techniques for clarifying audio signals
KR102208477B1 (en)2014-06-302021-01-27삼성전자주식회사Operating Method For Microphones and Electronic Device supporting the same
US9883314B2 (en)2014-07-032018-01-30Dolby Laboratories Licensing CorporationAuxiliary augmentation of soundfields
TWI584657B (en)*2014-08-202017-05-21國立清華大學A method for recording and rebuilding of a stereophonic sound field
KR102174850B1 (en)*2014-10-312020-11-05한화테크윈 주식회사Environment adaptation type beam forming apparatus for audio
US9747367B2 (en)2014-12-052017-08-29Stages LlcCommunication system for establishing and providing preferred audio
US9654868B2 (en)2014-12-052017-05-16Stages LlcMulti-channel multi-domain source identification and tracking
US10609475B2 (en)2014-12-052020-03-31Stages LlcActive noise control and customized audio system
WO2016093855A1 (en)*2014-12-122016-06-16Nuance Communications, Inc.System and method for generating a self-steering beamformer
US9565493B2 (en)2015-04-302017-02-07Shure Acquisition Holdings, Inc.Array microphone system and method of assembling the same
US9554207B2 (en)2015-04-302017-01-24Shure Acquisition Holdings, Inc.Offset cartridge microphones
JP6131989B2 (en)*2015-07-072017-05-24沖電気工業株式会社 Sound collecting apparatus, program and method
US9823893B2 (en)2015-07-152017-11-21International Business Machines CorporationProcessing of voice conversations using network of computing devices
WO2017033513A1 (en)*2015-08-242017-03-02ソニー株式会社Information processing device, information processing method, and program
EP3370437A4 (en)*2015-10-262018-10-17Sony CorporationSignal processing device, signal processing method, and program
CN108353225B (en)*2015-10-302020-09-22三菱电机株式会社 Hands-free controls
KR102502601B1 (en)*2015-11-272023-02-23삼성전자주식회사Electronic device and controlling voice signal method
JP2017102085A (en)*2015-12-042017-06-08キヤノン株式会社Information processing apparatus, information processing method, and program
US11064291B2 (en)2015-12-042021-07-13Sennheiser Electronic Gmbh & Co. KgMicrophone array system
US9894434B2 (en)2015-12-042018-02-13Sennheiser Electronic Gmbh & Co. KgConference system with a microphone array system and a method of speech acquisition in a conference system
US9820042B1 (en)2016-05-022017-11-14Knowles Electronics, LlcStereo separation and directional suppression with omni-directional microphones
US20170365255A1 (en)2016-06-152017-12-21Adam KupryjanowFar field automatic speech recognition pre-processing
TWI579833B (en)*2016-06-222017-04-21瑞昱半導體股份有限公司Signal processing device and signal processing method
CN107889022B (en)*2016-09-302021-03-23松下电器产业株式会社Noise suppression device and noise suppression method
US10945080B2 (en)2016-11-182021-03-09Stages LlcAudio analysis and processing system
US9980042B1 (en)2016-11-182018-05-22Stages LlcBeamformer direction of arrival and orientation analysis system
US9980075B1 (en)2016-11-182018-05-22Stages LlcAudio source spatialization relative to orientation sensor and output
US10367948B2 (en)2017-01-132019-07-30Shure Acquisition Holdings, Inc.Post-mixing acoustic echo cancellation systems and methods
JP7051876B6 (en)2017-01-272023-08-18シュアー アクイジッション ホールディングス インコーポレイテッド Array microphone module and system
US10366702B2 (en)2017-02-082019-07-30Logitech Europe, S.A.Direction detection device for acquiring and processing audible input
US10362393B2 (en)2017-02-082019-07-23Logitech Europe, S.A.Direction detection device for acquiring and processing audible input
US10229667B2 (en)2017-02-082019-03-12Logitech Europe S.A.Multi-directional beamforming device for acquiring and processing audible input
US10366700B2 (en)2017-02-082019-07-30Logitech Europe, S.A.Device for acquiring and processing audible input
US20180317006A1 (en)2017-04-282018-11-01Qualcomm IncorporatedMicrophone configurations
US10334360B2 (en)*2017-06-122019-06-25Revolabs, IncMethod for accurately calculating the direction of arrival of sound at a microphone array
US11081125B2 (en)*2017-06-132021-08-03Sandeep Kumar ChintalaNoise cancellation in voice communication systems
US10187721B1 (en)*2017-06-222019-01-22Amazon Technologies, Inc.Weighing fixed and adaptive beamformers
WO2019073804A1 (en)*2017-10-112019-04-18ソニー株式会社Sound source direction estimating device and method, and program
US11565365B2 (en)*2017-11-132023-01-31Taiwan Semiconductor Manufacturing Co., Ltd.System and method for monitoring chemical mechanical polishing
CN108109617B (en)*2018-01-082020-12-15深圳市声菲特科技技术有限公司Remote pickup method
DE102018107579B4 (en)*2018-03-292020-07-02Tdk Corporation Microphone array
CN108319155A (en)*2018-04-242018-07-24苏州宏云智能科技有限公司Wireless intelligent house terminal control unit
CN110441738B (en)*2018-05-032023-07-28阿里巴巴集团控股有限公司Method, system, vehicle and storage medium for vehicle-mounted voice positioning
DE102018110759A1 (en)*2018-05-042019-11-07Sennheiser Electronic Gmbh & Co. Kg microphone array
CN112335261B (en)2018-06-012023-07-18舒尔获得控股公司Patterned microphone array
US11297423B2 (en)2018-06-152022-04-05Shure Acquisition Holdings, Inc.Endfire linear array microphone
US10939030B2 (en)*2018-09-072021-03-02Canon Kabushiki KaishaVideo audio processing system and method of controlling the video audio processing system
US11310596B2 (en)2018-09-202022-04-19Shure Acquisition Holdings, Inc.Adjustable lobe shape for array microphones
US11109133B2 (en)2018-09-212021-08-31Shure Acquisition Holdings, Inc.Array microphone module and system
US20200184994A1 (en)*2018-12-072020-06-11Nuance Communications, Inc.System and method for acoustic localization of multiple sources using spatial pre-filtering
CN109803171B (en)*2019-02-152023-10-24深圳市锐明技术股份有限公司Monitoring camera for displaying voice position and control method thereof
US11558693B2 (en)2019-03-212023-01-17Shure Acquisition Holdings, Inc.Auto focus, auto focus within regions, and auto placement of beamformed microphone lobes with inhibition and voice activity detection functionality
WO2020191380A1 (en)2019-03-212020-09-24Shure Acquisition Holdings,Inc.Auto focus, auto focus within regions, and auto placement of beamformed microphone lobes with inhibition functionality
CN113841419B (en)2019-03-212024-11-12舒尔获得控股公司 Ceiling array microphone enclosure and associated design features
CN114051738B (en)2019-05-232024-10-01舒尔获得控股公司 Steerable speaker array, system and method thereof
WO2020243471A1 (en)2019-05-312020-12-03Shure Acquisition Holdings, Inc.Low latency automixer integrated with voice and noise activity detection
US11226396B2 (en)2019-06-272022-01-18Gracenote, Inc.Methods and apparatus to improve detection of audio signatures
CN110364161A (en)*2019-08-222019-10-22北京小米智能科技有限公司Method, electronic equipment, medium and the system of voice responsive signal
EP4018680A1 (en)2019-08-232022-06-29Shure Acquisition Holdings, Inc.Two-dimensional microphone array with improved directivity
US10887709B1 (en)*2019-09-252021-01-05Amazon Technologies, Inc.Aligned beam merger
WO2021087377A1 (en)2019-11-012021-05-06Shure Acquisition Holdings, Inc.Proximity microphone
CN111025233B (en)*2019-11-132023-09-15阿里巴巴集团控股有限公司Sound source direction positioning method and device, voice equipment and system
US11552611B2 (en)2020-02-072023-01-10Shure Acquisition Holdings, Inc.System and method for automatic adjustment of reference gain
US11277689B2 (en)2020-02-242022-03-15Logitech Europe S.A.Apparatus and method for optimizing sound quality of a generated audible signal
US11025324B1 (en)*2020-04-152021-06-01Cirrus Logic, Inc.Initialization of adaptive blocking matrix filters in a beamforming array using a priori information
USD944776S1 (en)2020-05-052022-03-01Shure Acquisition Holdings, Inc.Audio device
US11706562B2 (en)2020-05-292023-07-18Shure Acquisition Holdings, Inc.Transducer steering and configuration systems and methods using a local positioning system
JP7618995B2 (en)*2020-10-072025-01-22ヤマハ株式会社 Microphone Array System
US12395809B2 (en)2020-12-032025-08-19Dolby Laboratories Licensing CorporationAudibility at user location through mutual device audibility
CN112767908B (en)*2020-12-292024-05-21安克创新科技股份有限公司Active noise reduction method based on key voice recognition, electronic equipment and storage medium
CN112684412B (en)*2021-01-122022-09-13中北大学 A method and system for sound source localization based on pattern clustering
EP4285605A1 (en)2021-01-282023-12-06Shure Acquisition Holdings, Inc.Hybrid audio beamforming system
US11636842B2 (en)*2021-01-292023-04-25Iyo Inc.Ear-mountable listening device having a microphone array disposed around a circuit board
CN113393859B (en)*2021-06-092024-08-09湛江港(集团)股份有限公司Storage yard gridding sound source positioning analysis method and system
WO2023059655A1 (en)2021-10-042023-04-13Shure Acquisition Holdings, Inc.Networked automixer systems and methods
US12250526B2 (en)2022-01-072025-03-11Shure Acquisition Holdings, Inc.Audio beamforming with nulling control system and methods
CN115061087B (en)*2022-05-272024-05-14上海事凡物联网科技有限公司Signal processing method, DOA estimation method and electronic equipment
CN116055869B (en)*2022-05-302023-10-20荣耀终端有限公司Video processing method and terminal
CN114863943B (en)*2022-07-042022-11-04杭州兆华电子股份有限公司Self-adaptive positioning method and device for environmental noise source based on beam forming
CN114858271B (en)*2022-07-052022-09-23杭州兆华电子股份有限公司Array amplification method for sound detection
US12401942B1 (en)2023-05-252025-08-26Amazon Technologies, Inc.Group beam selection and beam merging
CN116953615B (en)*2023-08-042024-04-12中国水利水电科学研究院Networking detection positioning technology for termite nest of dam
SE2351451A1 (en)*2023-12-182025-06-19Monava AbAdaptive Beamforming Monitoring Method and System for Acoustic Sensor Arrays

Citations (14)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US5315562A (en)*1992-10-231994-05-24Rowe, Deines Instruments Inc.Correlation sonar system
US20030204397A1 (en)*2002-04-262003-10-30Mitel Knowledge CorporationMethod of compensating for beamformer steering delay during handsfree speech recognition
US20040161121A1 (en)*2003-01-172004-08-19Samsung Electronics Co., LtdAdaptive beamforming method and apparatus using feedback structure
US7039199B2 (en)2002-08-262006-05-02Microsoft CorporationSystem and process for locating a speaker using 360 degree sound source localization
US20070076898A1 (en)2003-11-242007-04-05Koninkiljke Phillips Electronics N.V.Adaptive beamformer with robustness against uncorrelated noise
WO2008041878A2 (en)2006-10-042008-04-10Micronas NitSystem and procedure of hands free speech communication using a microphone array
US20080232607A1 (en)*2007-03-222008-09-25Microsoft CorporationRobust adaptive beamforming with enhanced noise suppression
US20090141907A1 (en)*2007-11-302009-06-04Samsung Electronics Co., Ltd.Method and apparatus for canceling noise from sound input through microphone
US20090279714A1 (en)2008-05-062009-11-12Samsung Electronics Co., Ltd.Apparatus and method for localizing sound source in robot
US20090304200A1 (en)2008-06-092009-12-10Samsung Electronics Co., Ltd.Adaptive mode control apparatus and method for adaptive beamforming based on detection of user direction sound
WO2010020162A1 (en)*2008-08-192010-02-25深圳华为通信技术有限公司Method, communication device and communication system for controlling sound focusing
US20100241426A1 (en)*2009-03-232010-09-23Vimicro Electronics CorporationMethod and system for noise reduction
US20110096915A1 (en)*2009-10-232011-04-28Broadcom CorporationAudio spatialization for conference calls with multiple and moving talkers
US20110317522A1 (en)*2010-06-282011-12-29Microsoft CorporationSound source localization based on reflections and room estimation

Family Cites Families (69)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US5825898A (en)1996-06-271998-10-20Lamar Signal Processing Ltd.System and method for adaptive interference cancelling
US6236862B1 (en)*1996-12-162001-05-22Intersignal LlcContinuously adaptive dynamic signal separation and recovery system
US6198693B1 (en)1998-04-132001-03-06Andrea Electronics CorporationSystem and method for finding the direction of a wave source using an array of sensors
US7068801B1 (en)1998-12-182006-06-27National Research Council Of CanadaMicrophone array diffracting structure
CA2354858A1 (en)2001-08-082003-02-08Dspfactory Ltd.Subband directional audio signal processing using an oversampled filterbank
CA2399159A1 (en)2002-08-162004-02-16Dspfactory Ltd.Convergence improvement for oversampled subband adaptive filters
EP1538867B1 (en)2003-06-302012-07-18Nuance Communications, Inc.Handsfree system for use in a vehicle
WO2005006808A1 (en)2003-07-112005-01-20Cochlear LimitedMethod and device for noise reduction
EP1633121B1 (en)2004-09-032008-11-05Harman Becker Automotive Systems GmbHSpeech signal processing with combined adaptive noise reduction and adaptive echo compensation
US7970151B2 (en)2004-10-152011-06-28Lifesize Communications, Inc.Hybrid beamforming
US20060245601A1 (en)2005-04-272006-11-02Francois MichaudRobust localization and tracking of simultaneously moving sound sources using beamforming and particle filtering
JP4973657B2 (en)2006-04-202012-07-11日本電気株式会社 Adaptive array control device, method, program, and adaptive array processing device, method, program
US8233353B2 (en)2007-01-262012-07-31Microsoft CorporationMulti-sensor sound source localization
EP2026597B1 (en)2007-08-132009-11-11Harman Becker Automotive Systems GmbHNoise reduction by combined beamforming and post-filtering
KR20090128221A (en)2008-06-102009-12-15삼성전자주식회사 Sound source location estimation method and system according to the method
EP2197219B1 (en)2008-12-122012-10-24Nuance Communications, Inc.Method for determining a time delay for time delay compensation
US9973848B2 (en)2011-06-212018-05-15Amazon Technologies, Inc.Signal-enhancing beamforming in an augmented reality environment
US8694522B1 (en)2012-03-282014-04-08Amazon Technologies, Inc.Context dependent recognition
US20130265276A1 (en)2012-04-092013-10-10Amazon Technologies, Inc.Multiple touch sensing modes
US9354731B1 (en)2012-06-202016-05-31Amazon Technologies, Inc.Multi-dimension touch input
US8855295B1 (en)2012-06-252014-10-07Rawles LlcAcoustic echo cancellation using blind source separation
US8885815B1 (en)2012-06-252014-11-11Rawles LlcNull-forming techniques to improve acoustic echo cancellation
US9373338B1 (en)2012-06-252016-06-21Amazon Technologies, Inc.Acoustic echo cancellation processing based on feedback from speech recognizer
US9767828B1 (en)2012-06-272017-09-19Amazon Technologies, Inc.Acoustic echo cancellation using visual cues
US9229526B1 (en)2012-09-102016-01-05Amazon Technologies, Inc.Dedicated image processor
US9423886B1 (en)2012-10-022016-08-23Amazon Technologies, Inc.Sensor connectivity approaches
US9332167B1 (en)2012-11-202016-05-03Amazon Technologies, Inc.Multi-directional camera module for an electronic device
US9658738B1 (en)2012-11-292017-05-23Amazon Technologies, Inc.Representation management on an electronic device
US9689960B1 (en)2013-04-042017-06-27Amazon Technologies, Inc.Beam rejection in multi-beam microphone systems
US8953777B1 (en)2013-05-302015-02-10Amazon Technologies, Inc.Echo path change detector with robustness to double talk
US9747899B2 (en)2013-06-272017-08-29Amazon Technologies, Inc.Detecting self-generated wake expressions
US9978387B1 (en)2013-08-052018-05-22Amazon Technologies, Inc.Reference signal generation for acoustic echo cancellation
US9473646B1 (en)2013-09-162016-10-18Amazon Technologies, Inc.Robust acoustic echo cancellation
US8983057B1 (en)2013-09-202015-03-17Amazon Technologies, Inc.Step size control for acoustic echo cancellation
US9591404B1 (en)2013-09-272017-03-07Amazon Technologies, Inc.Beamformer design using constrained convex optimization in three-dimensional space
US9704478B1 (en)2013-12-022017-07-11Amazon Technologies, Inc.Audio output masking for improved automatic speech recognition
US10147441B1 (en)2013-12-192018-12-04Amazon Technologies, Inc.Voice controlled system
US9319782B1 (en)2013-12-202016-04-19Amazon Technologies, Inc.Distributed speaker synchronization
US9319783B1 (en)2014-02-192016-04-19Amazon Technologies, Inc.Attenuation of output audio based on residual echo
US9432768B1 (en)2014-03-282016-08-30Amazon Technologies, Inc.Beam forming for a wearable computer
US9336767B1 (en)2014-03-282016-05-10Amazon Technologies, Inc.Detecting device proximities
US9363616B1 (en)2014-04-182016-06-07Amazon Technologies, Inc.Directional capability testing of audio devices
US9432769B1 (en)2014-07-302016-08-30Amazon Technologies, Inc.Method and system for beam selection in microphone array beamformers
US9677986B1 (en)2014-09-242017-06-13Amazon Technologies, Inc.Airborne particle detection with user device
US9456276B1 (en)2014-09-302016-09-27Amazon Technologies, Inc.Parameter selection for audio beamforming
US9390723B1 (en)2014-12-112016-07-12Amazon Technologies, Inc.Efficient dereverberation in networked audio systems
US9940949B1 (en)2014-12-192018-04-10Amazon Technologies, Inc.Dynamic adjustment of expression detection criteria
US9661438B1 (en)2015-03-262017-05-23Amazon Technologies, Inc.Low latency limiter
US9431982B1 (en)2015-03-302016-08-30Amazon Technologies, Inc.Loudness learning and balancing system
US9734845B1 (en)2015-06-262017-08-15Amazon Technologies, Inc.Mitigating effects of electronic audio sources in expression detection
US9516410B1 (en)2015-06-292016-12-06Amazon Technologies, Inc.Asynchronous clock frequency domain acoustic echo canceller
US9678559B1 (en)2015-09-182017-06-13Amazon Technologies, Inc.Determining a device state based on user presence detection
US9589575B1 (en)2015-12-022017-03-07Amazon Technologies, Inc.Asynchronous clock frequency domain acoustic echo canceller
US9747920B2 (en)2015-12-172017-08-29Amazon Technologies, Inc.Adaptive beamforming to create reference channels
US9820036B1 (en)2015-12-302017-11-14Amazon Technologies, Inc.Speech processing of reflected sound
US9614486B1 (en)2015-12-302017-04-04Amazon Technologies, Inc.Adaptive gain control
US9997151B1 (en)2016-01-202018-06-12Amazon Technologies, Inc.Multichannel acoustic echo cancellation for wireless applications
US9659555B1 (en)2016-02-092017-05-23Amazon Technologies, Inc.Multichannel acoustic echo cancellation
US9653060B1 (en)2016-02-092017-05-16Amazon Technologies, Inc.Hybrid reference signal for acoustic echo cancellation
US10109294B1 (en)2016-03-252018-10-23Amazon Technologies, Inc.Adaptive echo cancellation
US9754605B1 (en)2016-06-092017-09-05Amazon Technologies, Inc.Step-size control for multi-channel acoustic echo canceller
US9818425B1 (en)2016-06-172017-11-14Amazon Technologies, Inc.Parallel output paths for acoustic echo cancellation
US10726835B2 (en)2016-12-232020-07-28Amazon Technologies, Inc.Voice activated modular controller
US10237647B1 (en)2017-03-012019-03-19Amazon Technologies, Inc.Adaptive step-size control for beamformer
US10147439B1 (en)2017-03-302018-12-04Amazon Technologies, Inc.Volume adjustment for listening environment
US10229698B1 (en)2017-06-212019-03-12Amazon Technologies, Inc.Playback reference signal-assisted multi-microphone interference canceler
US10304475B1 (en)2017-08-142019-05-28Amazon Technologies, Inc.Trigger word based beam selection
US9966059B1 (en)2017-09-062018-05-08Amazon Technologies, Inc.Reconfigurale fixed beam former using given microphone array
US9973849B1 (en)2017-09-202018-05-15Amazon Technologies, Inc.Signal quality beam selection

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US5315562A (en)*1992-10-231994-05-24Rowe, Deines Instruments Inc.Correlation sonar system
US20030204397A1 (en)*2002-04-262003-10-30Mitel Knowledge CorporationMethod of compensating for beamformer steering delay during handsfree speech recognition
US7039199B2 (en)2002-08-262006-05-02Microsoft CorporationSystem and process for locating a speaker using 360 degree sound source localization
US20040161121A1 (en)*2003-01-172004-08-19Samsung Electronics Co., LtdAdaptive beamforming method and apparatus using feedback structure
US20070076898A1 (en)2003-11-242007-04-05Koninkiljke Phillips Electronics N.V.Adaptive beamformer with robustness against uncorrelated noise
WO2008041878A2 (en)2006-10-042008-04-10Micronas NitSystem and procedure of hands free speech communication using a microphone array
US20080232607A1 (en)*2007-03-222008-09-25Microsoft CorporationRobust adaptive beamforming with enhanced noise suppression
US20090141907A1 (en)*2007-11-302009-06-04Samsung Electronics Co., Ltd.Method and apparatus for canceling noise from sound input through microphone
US20090279714A1 (en)2008-05-062009-11-12Samsung Electronics Co., Ltd.Apparatus and method for localizing sound source in robot
US20090304200A1 (en)2008-06-092009-12-10Samsung Electronics Co., Ltd.Adaptive mode control apparatus and method for adaptive beamforming based on detection of user direction sound
WO2010020162A1 (en)*2008-08-192010-02-25深圳华为通信技术有限公司Method, communication device and communication system for controlling sound focusing
US20110135125A1 (en)*2008-08-192011-06-09Wuzhou ZhanMethod, communication device and communication system for controlling sound focusing
US20100241426A1 (en)*2009-03-232010-09-23Vimicro Electronics CorporationMethod and system for noise reduction
US20110096915A1 (en)*2009-10-232011-04-28Broadcom CorporationAudio spatialization for conference calls with multiple and moving talkers
US20110317522A1 (en)*2010-06-282011-12-29Microsoft CorporationSound source localization based on reflections and room estimation

Cited By (149)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US11589329B1 (en)2010-12-302023-02-21Staton Techiya LlcInformation processing using a population of data acquisition devices
US11736860B2 (en)2016-02-222023-08-22Sonos, Inc.Voice control of a media playback system
US11863593B2 (en)2016-02-222024-01-02Sonos, Inc.Networked microphone device control
US11726742B2 (en)2016-02-222023-08-15Sonos, Inc.Handling of loss of pairing between networked devices
US11832068B2 (en)2016-02-222023-11-28Sonos, Inc.Music service selection
US12047752B2 (en)2016-02-222024-07-23Sonos, Inc.Content mixing
US11405430B2 (en)2016-02-222022-08-02Sonos, Inc.Networked microphone device control
US11184704B2 (en)2016-02-222021-11-23Sonos, Inc.Music service selection
US11750969B2 (en)2016-02-222023-09-05Sonos, Inc.Default playback device designation
US11513763B2 (en)2016-02-222022-11-29Sonos, Inc.Audio response playback
US11556306B2 (en)2016-02-222023-01-17Sonos, Inc.Voice controlled media playback system
US11514898B2 (en)2016-02-222022-11-29Sonos, Inc.Voice control of a media playback system
US11137979B2 (en)2016-02-222021-10-05Sonos, Inc.Metadata exchange involving a networked playback system and a networked microphone system
US11983463B2 (en)2016-02-222024-05-14Sonos, Inc.Metadata exchange involving a networked playback system and a networked microphone system
US11212612B2 (en)2016-02-222021-12-28Sonos, Inc.Voice control of a media playback system
US20180188347A1 (en)*2016-03-302018-07-05Yutou Technology (Hangzhou) Co., Ltd.Voice direction searching system and method thereof
US11133018B2 (en)2016-06-092021-09-28Sonos, Inc.Dynamic player selection for audio signal processing
US11545169B2 (en)2016-06-092023-01-03Sonos, Inc.Dynamic player selection for audio signal processing
US11184969B2 (en)2016-07-152021-11-23Sonos, Inc.Contextualization of voice inputs
US11979960B2 (en)2016-07-152024-05-07Sonos, Inc.Contextualization of voice inputs
US11664023B2 (en)2016-07-152023-05-30Sonos, Inc.Voice detection by multiple devices
US11531520B2 (en)2016-08-052022-12-20Sonos, Inc.Playback device supporting concurrent voice assistants
US11641559B2 (en)2016-09-272023-05-02Sonos, Inc.Audio playback settings for voice interaction
US11516610B2 (en)2016-09-302022-11-29Sonos, Inc.Orientation-based playback device microphone selection
US11308961B2 (en)2016-10-192022-04-19Sonos, Inc.Arbitration-based voice recognition
US11727933B2 (en)2016-10-192023-08-15Sonos, Inc.Arbitration-based voice recognition
US12217748B2 (en)2017-03-272025-02-04Sonos, Inc.Systems and methods of multiple voice services
US11183181B2 (en)2017-03-272021-11-23Sonos, Inc.Systems and methods of multiple voice services
US11900937B2 (en)2017-08-072024-02-13Sonos, Inc.Wake-word detection suppression
US11380322B2 (en)2017-08-072022-07-05Sonos, Inc.Wake-word detection suppression
US11114186B2 (en)2017-08-102021-09-07Nuance Communications, Inc.Automated clinical documentation system and method
US11257576B2 (en)2017-08-102022-02-22Nuance Communications, Inc.Automated clinical documentation system and method
US12008310B2 (en)2017-08-102024-06-11Microsoft Licensing Technology, LLCAutomated clinical documentation system and method
US11316865B2 (en)2017-08-102022-04-26Nuance Communications, Inc.Ambient cooperative intelligence system and method
US11853691B2 (en)2017-08-102023-12-26Nuance Communications, Inc.Automated clinical documentation system and method
US11101023B2 (en)2017-08-102021-08-24Nuance Communications, Inc.Automated clinical documentation system and method
US11101022B2 (en)2017-08-102021-08-24Nuance Communications, Inc.Automated clinical documentation system and method
US11074996B2 (en)2017-08-102021-07-27Nuance Communications, Inc.Automated clinical documentation system and method
US11043288B2 (en)2017-08-102021-06-22Nuance Communications, Inc.Automated clinical documentation system and method
US11322231B2 (en)2017-08-102022-05-03Nuance Communications, Inc.Automated clinical documentation system and method
US11482308B2 (en)*2017-08-102022-10-25Nuance Communications, Inc.Automated clinical documentation system and method
US11605448B2 (en)2017-08-102023-03-14Nuance Communications, Inc.Automated clinical documentation system and method
US11482311B2 (en)2017-08-102022-10-25Nuance Communications, Inc.Automated clinical documentation system and method
US10978187B2 (en)2017-08-102021-04-13Nuance Communications, Inc.Automated clinical documentation system and method
US10957427B2 (en)2017-08-102021-03-23Nuance Communications, Inc.Automated clinical documentation system and method
US11295838B2 (en)2017-08-102022-04-05Nuance Communications, Inc.Automated clinical documentation system and method
US11295839B2 (en)2017-08-102022-04-05Nuance Communications, Inc.Automated clinical documentation system and method
US10957428B2 (en)2017-08-102021-03-23Nuance Communications, Inc.Automated clinical documentation system and method
US10412532B2 (en)*2017-08-302019-09-10Harman International Industries, IncorporatedEnvironment discovery via time-synchronized networked loudspeakers
US20190110153A1 (en)*2017-08-302019-04-11Harman International Industries, IncorporatedEnvironment discovery via time-synchronized networked loudspeakers
US11500611B2 (en)2017-09-082022-11-15Sonos, Inc.Dynamic computation of system response volume
US11646045B2 (en)2017-09-272023-05-09Sonos, Inc.Robust short-time fourier transform acoustic echo cancellation during audio playback
US11302326B2 (en)2017-09-282022-04-12Sonos, Inc.Tone interference cancellation
US11769505B2 (en)2017-09-282023-09-26Sonos, Inc.Echo of tone interferance cancellation using two acoustic echo cancellers
US12047753B1 (en)2017-09-282024-07-23Sonos, Inc.Three-dimensional beam forming with a microphone array
US12236932B2 (en)2017-09-282025-02-25Sonos, Inc.Multi-channel acoustic echo cancellation
US11538451B2 (en)2017-09-282022-12-27Sonos, Inc.Multi-channel acoustic echo cancellation
US11893308B2 (en)2017-09-292024-02-06Sonos, Inc.Media playback system with concurrent voice assistance
US11288039B2 (en)2017-09-292022-03-29Sonos, Inc.Media playback system with concurrent voice assistance
US11175888B2 (en)2017-09-292021-11-16Sonos, Inc.Media playback system with concurrent voice assistance
US11451908B2 (en)2017-12-102022-09-20Sonos, Inc.Network microphone devices with automatic do not disturb actuation capabilities
US11676590B2 (en)2017-12-112023-06-13Sonos, Inc.Home graph
US11343614B2 (en)2018-01-312022-05-24Sonos, Inc.Device designation of playback and network microphone device arrangements
US11689858B2 (en)2018-01-312023-06-27Sonos, Inc.Device designation of playback and network microphone device arrangements
US11270261B2 (en)2018-03-052022-03-08Nuance Communications, Inc.System and method for concept formatting
US11494735B2 (en)2018-03-052022-11-08Nuance Communications, Inc.Automated clinical documentation system and method
US11250382B2 (en)2018-03-052022-02-15Nuance Communications, Inc.Automated clinical documentation system and method
US11250383B2 (en)2018-03-052022-02-15Nuance Communications, Inc.Automated clinical documentation system and method
US11515020B2 (en)2018-03-052022-11-29Nuance Communications, Inc.Automated clinical documentation system and method
US11222716B2 (en)2018-03-052022-01-11Nuance CommunicationsSystem and method for review of automated clinical documentation from recorded audio
US11295272B2 (en)2018-03-052022-04-05Nuance Communications, Inc.Automated clinical documentation system and method
US20190324117A1 (en)*2018-04-242019-10-24Mediatek Inc.Content aware audio source localization
US12360734B2 (en)2018-05-102025-07-15Sonos, Inc.Systems and methods for voice-assisted media content selection
US11175880B2 (en)2018-05-102021-11-16Sonos, Inc.Systems and methods for voice-assisted media content selection
US11797263B2 (en)2018-05-102023-10-24Sonos, Inc.Systems and methods for voice-assisted media content selection
US11715489B2 (en)2018-05-182023-08-01Sonos, Inc.Linear filtering for noise-suppressed speech detection
US11792590B2 (en)2018-05-252023-10-17Sonos, Inc.Determining and adapting to changes in microphone performance of playback devices
US11696074B2 (en)2018-06-282023-07-04Sonos, Inc.Systems and methods for associating playback devices with voice assistant services
US11197096B2 (en)2018-06-282021-12-07Sonos, Inc.Systems and methods for associating playback devices with voice assistant services
US11482978B2 (en)2018-08-282022-10-25Sonos, Inc.Audio notifications
US11563842B2 (en)2018-08-282023-01-24Sonos, Inc.Do not disturb feature for audio notifications
US11432030B2 (en)2018-09-142022-08-30Sonos, Inc.Networked devices, systems, and methods for associating playback devices based on sound codes
US11551690B2 (en)2018-09-142023-01-10Sonos, Inc.Networked devices, systems, and methods for intelligently deactivating wake-word engines
US11778259B2 (en)2018-09-142023-10-03Sonos, Inc.Networked devices, systems and methods for associating playback devices based on sound codes
US11790937B2 (en)2018-09-212023-10-17Sonos, Inc.Voice detection optimization using sound metadata
US12230291B2 (en)2018-09-212025-02-18Sonos, Inc.Voice detection optimization using sound metadata
US12165651B2 (en)2018-09-252024-12-10Sonos, Inc.Voice detection optimization based on selected voice assistant service
US11727936B2 (en)2018-09-252023-08-15Sonos, Inc.Voice detection optimization based on selected voice assistant service
US11100923B2 (en)2018-09-282021-08-24Sonos, Inc.Systems and methods for selective wake word detection using neural network models
US12165644B2 (en)2018-09-282024-12-10Sonos, Inc.Systems and methods for selective wake word detection
US11790911B2 (en)2018-09-282023-10-17Sonos, Inc.Systems and methods for selective wake word detection using neural network models
US11501795B2 (en)2018-09-292022-11-15Sonos, Inc.Linear filtering for noise-suppressed speech detection via multiple network microphone devices
US12062383B2 (en)2018-09-292024-08-13Sonos, Inc.Linear filtering for noise-suppressed speech detection via multiple network microphone devices
US11899519B2 (en)2018-10-232024-02-13Sonos, Inc.Multiple stage network microphone device with reduced power consumption and processing load
US11200889B2 (en)2018-11-152021-12-14Sonos, Inc.Dilated convolutions and gating for efficient keyword spotting
US11741948B2 (en)2018-11-152023-08-29Sonos Vox France SasDilated convolutions and gating for efficient keyword spotting
US11557294B2 (en)2018-12-072023-01-17Sonos, Inc.Systems and methods of operating media playback systems having multiple voice assistant services
US11183183B2 (en)2018-12-072021-11-23Sonos, Inc.Systems and methods of operating media playback systems having multiple voice assistant services
US11538460B2 (en)2018-12-132022-12-27Sonos, Inc.Networked microphone devices, systems, and methods of localized arbitration
US11132989B2 (en)2018-12-132021-09-28Sonos, Inc.Networked microphone devices, systems, and methods of localized arbitration
US11540047B2 (en)2018-12-202022-12-27Sonos, Inc.Optimization of network microphone devices using noise classification
US11159880B2 (en)2018-12-202021-10-26Sonos, Inc.Optimization of network microphone devices using noise classification
US11646023B2 (en)2019-02-082023-05-09Sonos, Inc.Devices, systems, and methods for distributed voice processing
US11315556B2 (en)2019-02-082022-04-26Sonos, Inc.Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification
US11798553B2 (en)2019-05-032023-10-24Sonos, Inc.Voice assistant persistence across multiple network microphone devices
US20230169956A1 (en)*2019-05-032023-06-01Sonos, Inc.Locally distributed keyword detection
US11120794B2 (en)2019-05-032021-09-14Sonos, Inc.Voice assistant persistence across multiple network microphone devices
US11771866B2 (en)*2019-05-032023-10-03Sonos, Inc.Locally distributed keyword detection
US11200894B2 (en)2019-06-122021-12-14Sonos, Inc.Network microphone device with command keyword eventing
US11854547B2 (en)2019-06-122023-12-26Sonos, Inc.Network microphone device with command keyword eventing
US11361756B2 (en)2019-06-122022-06-14Sonos, Inc.Conditional wake word eventing based on environment
US11501773B2 (en)2019-06-122022-11-15Sonos, Inc.Network microphone device with command keyword conditioning
US11216480B2 (en)2019-06-142022-01-04Nuance Communications, Inc.System and method for querying data points from graph data structures
US11227679B2 (en)2019-06-142022-01-18Nuance Communications, Inc.Ambient clinical intelligence system and method
US11043207B2 (en)2019-06-142021-06-22Nuance Communications, Inc.System and method for array data simulation and customized acoustic modeling for ambient ASR
US11531807B2 (en)2019-06-282022-12-20Nuance Communications, Inc.System and method for customized text macros
US20210358481A1 (en)*2019-07-312021-11-18Sonos, Inc.Locally distributed keyword detection
US11710487B2 (en)2019-07-312023-07-25Sonos, Inc.Locally distributed keyword detection
US11354092B2 (en)2019-07-312022-06-07Sonos, Inc.Noise classification for event detection
US11551669B2 (en)*2019-07-312023-01-10Sonos, Inc.Locally distributed keyword detection
US11714600B2 (en)2019-07-312023-08-01Sonos, Inc.Noise classification for event detection
US11138975B2 (en)2019-07-312021-10-05Sonos, Inc.Locally distributed keyword detection
US12211490B2 (en)2019-07-312025-01-28Sonos, Inc.Locally distributed keyword detection
US11138969B2 (en)2019-07-312021-10-05Sonos, Inc.Locally distributed keyword detection
US11670408B2 (en)2019-09-302023-06-06Nuance Communications, Inc.System and method for review of automated clinical documentation
US11189286B2 (en)2019-10-222021-11-30Sonos, Inc.VAS toggle based on device orientation
US11862161B2 (en)2019-10-222024-01-02Sonos, Inc.VAS toggle based on device orientation
US11200900B2 (en)2019-12-202021-12-14Sonos, Inc.Offline voice control
US11869503B2 (en)2019-12-202024-01-09Sonos, Inc.Offline voice control
US11562740B2 (en)2020-01-072023-01-24Sonos, Inc.Voice verification for media playback
US11556307B2 (en)2020-01-312023-01-17Sonos, Inc.Local voice data processing
US11961519B2 (en)2020-02-072024-04-16Sonos, Inc.Localized wakeword verification
US11308958B2 (en)2020-02-072022-04-19Sonos, Inc.Localized wakeword verification
US11240621B2 (en)2020-04-112022-02-01LI Creative Technologies, Inc.Three-dimensional audio systems
US11611840B2 (en)2020-04-112023-03-21LI Creative Technologies, Inc.Three-dimensional audio systems
US11308962B2 (en)2020-05-202022-04-19Sonos, Inc.Input detection windowing
US11482224B2 (en)2020-05-202022-10-25Sonos, Inc.Command keywords with input detection windowing
US11727919B2 (en)2020-05-202023-08-15Sonos, Inc.Memory allocation for keyword spotting engines
US11694689B2 (en)2020-05-202023-07-04Sonos, Inc.Input detection windowing
US12387716B2 (en)2020-06-082025-08-12Sonos, Inc.Wakewordless voice quickstarts
US11698771B2 (en)2020-08-252023-07-11Sonos, Inc.Vocal guidance engines for playback devices
US12283269B2 (en)2020-10-162025-04-22Sonos, Inc.Intent inference in audiovisual communication sessions
US11222103B1 (en)2020-10-292022-01-11Nuance Communications, Inc.Ambient cooperative intelligence system and method
US12424220B2 (en)2020-11-122025-09-23Sonos, Inc.Network device interaction by range
US11984123B2 (en)2020-11-122024-05-14Sonos, Inc.Network device interaction by range
US11551700B2 (en)2021-01-252023-01-10Sonos, Inc.Systems and methods for power-efficient keyword detection
US12327556B2 (en)2021-09-302025-06-10Sonos, Inc.Enabling and disabling microphones and voice assistants
US12327549B2 (en)2022-02-092025-06-10Sonos, Inc.Gatekeeping for voice intent processing
US11890168B2 (en)2022-03-212024-02-06Li Creative Technologies Inc.Hearing protection and situational awareness system

Also Published As

Publication numberPublication date
USRE48371E1 (en)2020-12-29
US8861756B2 (en)2014-10-14
US20120076316A1 (en)2012-03-29

Similar Documents

PublicationPublication DateTitle
USRE48371E1 (en)Microphone array system
KR101566649B1 (en)Near-field null and beamforming
US11800281B2 (en)Pattern-forming microphone array
US8098844B2 (en)Dual-microphone spatial noise suppression
US9966059B1 (en)Reconfigurale fixed beam former using given microphone array
US9094496B2 (en)System and method for stereophonic acoustic echo cancellation
US9020163B2 (en)Near-field null and beamforming
US10229698B1 (en)Playback reference signal-assisted multi-microphone interference canceler
US6584203B2 (en)Second-order adaptive differential microphone array
US20190208318A1 (en)Microphone array auto-directive adaptive wideband beamforming using orientation information from mems sensors
KR101470262B1 (en)Systems, methods, apparatus, and computer-readable media for multi-microphone location-selective processing
US20180350381A1 (en)System and method of noise reduction for a mobile device
US10341759B2 (en)System and method of wind and noise reduction for a headphone
US20140003635A1 (en)Audio signal processing device calibration
US20080175408A1 (en)Proximity filter
CN103339961A (en)Apparatus and method for spatially selective sound acquisition by acoustic triangulation
US20160140949A1 (en)Wrist wearable apparatuses and methods with desired signal extraction
WO2007059255A1 (en)Dual-microphone spatial noise suppression
PriyankaA review on adaptive beamforming techniques for speech enhancement
Gaubitch et al.On near-field beamforming with smartphone-based ad-hoc microphone arrays
CN115515038B (en)Beam forming method, device and equipment
US20240249742A1 (en)Partially adaptive audio beamforming systems and methods
CN110099330A (en)A kind of earphone, earphone system and earphone charging system
Liu et al.Simulation of fixed microphone arrays for directional hearing aids
Šarić et al.Performance analysis of MVDR beamformer applied on an end-fire microphone array composed of unidirectional microphones

Legal Events

DateCodeTitleDescription
ASAssignment

Owner name:VOCALIFE LLC, TEXAS

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ZHU, MANLI;LI, QI;REEL/FRAME:049770/0418

Effective date:20190131

IPRAia trial proceeding filed before the patent and appeal board: inter partes review

Free format text:TRIAL NO: IPR2020-00864

Opponent name:AMAZON.COM, INC. AND AMAZON.COM SERVICES, INC.

Effective date:20200423

IPRAia trial proceeding filed before the patent and appeal board: inter partes review

Free format text:TRIAL NO: IPR2022-00004

Opponent name:GOOGLE LLC

Effective date:20211007

MAFPMaintenance fee payment

Free format text:PAYMENT OF MAINTENANCE FEE, 8TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2552); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

Year of fee payment:8

IPRAia trial proceeding filed before the patent and appeal board: inter partes review

Free format text:TRIAL NO: IPR2022-00469

Opponent name:SONOS, INC.

Effective date:20220119

ASAssignment

Owner name:LI CREATIVE TECHNOLOGIES INC, NEW JERSEY

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VOCALIFE LLC;REEL/FRAME:068463/0912

Effective date:20240831


[8]ページ先頭

©2009-2025 Movatter.jp