Movatterモバイル変換


[0]ホーム

URL:


EP0852052B1 - System for adaptively filtering audio signals to enhance speech intelligibility in noisy environmental conditions - Google Patents

System for adaptively filtering audio signals to enhance speech intelligibility in noisy environmental conditions
Download PDF

Info

Publication number
EP0852052B1
EP0852052B1EP96931552AEP96931552AEP0852052B1EP 0852052 B1EP0852052 B1EP 0852052B1EP 96931552 AEP96931552 AEP 96931552AEP 96931552 AEP96931552 AEP 96931552AEP 0852052 B1EP0852052 B1EP 0852052B1
Authority
EP
European Patent Office
Prior art keywords
noise
frame
filter
speech
estimate
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
EP96931552A
Other languages
German (de)
French (fr)
Other versions
EP0852052A1 (en
Inventor
Torbjörn W. SÖLVE
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ericsson Inc
Original Assignee
Ericsson Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ericsson IncfiledCriticalEricsson Inc
Publication of EP0852052A1publicationCriticalpatent/EP0852052A1/en
Application grantedgrantedCritical
Publication of EP0852052B1publicationCriticalpatent/EP0852052B1/en
Anticipated expirationlegal-statusCritical
Expired - Lifetimelegal-statusCriticalCurrent

Links

Images

Classifications

Definitions

Landscapes

Description

  • The present invention relates to noise reductionsystems, and in particular, to an adaptive speechintelligibility enhancement system for use in portabledigital radio telephones.
  • The cellular telephone industry has made phenomenalstrides in commercial operations in the United States as well as the rest of the world. Demand for cellularservices in major metropolitan areas is outstrippingcurrent system capacity. Assuming this trendcontinues, cellular telecommunications will reach eventhe smallest rural markets. Consequently, cellularcapacity must be increased while maintaining highquality service at a reasonable cost. One importantstep towards increasing capacity is the conversion ofcellular systems from analog to digital transmission.This conversion is also important because the firstgeneration of personal communication networks (PCNs),employing low cost, pocket-size, cordless telephonesthat can be easily carried and used to make or receivecalls in the home, office, street, car, etc., willlikely be provided by cellular carriers using the nextgeneration digital cellular infrastructure.
  • Digital communication systems take advantage ofpowerful digital signal processing techniques. Digitalsignal processing refers generally to mathematical andother manipulation of digitized signals. For example,after converting (digitizing) an analog signal intodigital form, that digital signal may be filtered,amplified, and attenuated using simple mathematicalroutines in a digital signal processor (DSP).Typically, DSPs are manufactured as high speedintegrated circuits so that data processing operationscan be performed essentially in real time. DSPs mayalso be used to reduce the bit transmission rate ofdigitized speech which translates into reduced spectraloccupancy of the transmitted radio signals and increased system capacity. For example, if speechsignals are digitized using 14-bit linear Pulse CodeModulation (PCM) and sampled at an 8 KHz rate, a serialbit rate of 112 Kbits/sec is produced. Moreover, bytaking mathematical advantage of redundancies and otherpredicable characteristics of human speech, voicecoding techniques can be used to compress the serialbit rate from 112 Kbits/sec to 7.95 Kbits/sec toachieve a 14:1 reduction in bit transmission rate.Reduced transmission rates translate into moreavailable bandwidth.
  • One popular speech compression technique adopted inthe United States by the TIA for use as the digitalstandard for the second generation of cellulartelephone systems (i.e., IS-54) is vector sourcebookexcited linear predictive coding (VSELP).Unfortunately, when audio signals including speech,mixed with high levels of ambient noise (particularly"colored noise"), are coded/compressed using VSELP,undesirable audio signal characteristics may be part ofthe result. For example, if a digital mobile telephoneis used in a noisy environment (e.g. inside a movingautomobile), both ambient noise and desired speech arecompressed using the VSELP encoding algorithm andtransmitted to a base station where the compressedsignal is decoded and reconstituted into audiblespeech. When the background noise is reconstitutedinto an analog format, undesirable, audible distortionof the noise, and occasionally in the speech, is introduced. This distortion is very annoying to theaverage listener.
  • The distortion is caused in large part by theenvironment in which the mobile telephones are used.Mobile telephones are typically used in a vehicle'sinterior where there is often ambient noise produced bythe vehicle's engine and surrounding vehicular traffic.This ambient noise in the vehicle's interior istypically concentrated in the low audible frequencyrange and the magnitude of the noise can vary due tosuch factors as the speed and acceleration of thevehicle and the extent of the surrounding vehiculartraffic. This type of low frequency noise also has thetendency of significantly decreasing theintelligibility of the speech coming from the speakingperson in the car environment. The decrease in speechintelligibility caused by low frequency noise can beparticularly significant in communication systemsdeploying a VSELP vocoder, but can also occur incommunication systems that do not include a VSELPvocoder.
  • The influence of the ambient noise on the mobiletelephone can also be affected by the manner in whichthe mobile telephone is used. In particular, themobile telephone may be used in a hands-free mode wherethe telephone user talks on the telephone while themobile telephone is in a cradle. This frees thetelephone user's hands to drive but also increases thedistance that the telephone user's audible words musttravel before reaching the microphone input of the mobile telephone. This increased distance between theuser and the mobile telephone, along with the varyingambient noise, can result in noise being a significantportion of the total power spectral energy of the audiosignal inputted into the mobile telephone.
  • Prior art disclosure contained inEP 0 645 756,EP0 558 312,EP 0 665 530, DE 4 012 349, U.S. Patent Nos.4,811,404, 4,461,025, and 5,251,263 all disclose mannersby which to filter unwanted signal components.
  • In theory, various signal processing algorithms couldbe implemented using digital signal processors to filterthe VSELP encoded background noise. These solutions,however, often require significant digital signalprocessing overhead, measured in terms of millions ofinstructions executed per second (MIPS), which consumesvaluable processing time, memory space, and powerconsumption. Each of these signal processing resources,however, is limited in portable radiotelephones. Hence,simply increasing the processing burden of the DSP is notan optimal solution for minimizing VSELP encoded and othertypes of background noise.
  • The present invention provides a method and an apparatus for selectively altering a frame of a digital signal according toclaims 1 and 9.
  • The present invention provides an adaptive noisereduction system that reduces the undesirablecontributions of encoded background noise while bothminimizing any negative impact on the quality of theencoded speech and minimizing any increased drain ondigital signal processor resources. The method and systemof the present invention increases the intelligibility ofthe speech in a digitized audio signal by passing framesof the digitized audio signal through a filter circuit.The filter circuit functions as an adjustable, high-pass filter which filters aportion of the digitized signal in a low audiblefrequency range and passes the portion of the digitizedsignal falling in higher frequency ranges. Because thenoise in a vehicle tends to be concentrated in a lowaudible frequency range and only a relatively smallportion of the intelligibility content of speech fallswithin this low frequency range, the filter circuitfilters a large segment of the noise in the digitizedaudio signal while only filtering less importantsegments of the speech. This results in a relativelylarger portion of the noise energy being removedcompared to the portion of the speech energy removed.By adaptively adjusting and selecting the frequencyresponse curve of the filter circuit, the amount ofspeech filtered is limited and has a minimal affect onthe intelligibility of the speech outputted by theradio.
  • A filter control circuit is used to adjust thefilter circuit to exhibit different frequency responsecurves as a function of a noise estimate and/or aspectral profile result corresponding to the noise inthe audio signal. The noise estimate and/or thespectral profile result are adjusted on a frame-by-framebasis for the digital signal and as a function ofspeech detection. If speech is not detected, the noiseestimate and/or spectral profile result is updated forthe current frame. If speech is detected, the noiseestimate and/or spectral profile result is leftunadjusted.
  • In a first embodiment, the filter circuitcalculates noise estimates for the frames of thedigitized audio signals. The noise estimatescorrespond to the amount of background noise in theframes of the digitized audio signals. As the relativeamount of background noise to speech in a low frequencyrange of speech increases, the noise estimatesincrease. The filter control circuit uses the noiseestimates to adjust the filter circuit to filter largerportions of the low frequency range of speech as therelative amount of background noise to speech in a lowfrequency range of speech increases. When nobackground noise is present, no portion of the speechsignal is filtered. Larger portions of noise andspeech information are extracted when there is a higherlevel of background noise. Because noise tends to beconcentrated in a low frequency range and only arelatively small portion of the intelligibility contentof speech falls within this low frequency range, theoverall intelligibility of the audio signal can beincreased by increasing the portion of low frequencyenergy being filtered as the noise estimates increase.
  • In a second embodiment, a modified filter controlcircuit is used to adjust the filter circuit to exhibitdifferent frequency response curves as a function of anoise profile of the noise estimate over a selectedfrequency range in the audio signal. The filtercontrol circuit includes a spectral analyzer fordetermining a noise profile estimate as a function ofthe detection speech. A noise profile estimate is determined for a current frame and compared to areference noise profile. Based on this comparison, thefilter circuit is adaptively adjusted to extractvarying amounts of low frequency energy from thecurrent frame.
  • The adaptive noise reduction system according tothe present invention may be advantageously applied totelecommunication systems in which portable/mobileradio transceivers communicate over RF channels witheach other or with fixed telephone line subscribers.Each transceiver includes an antenna, a receiver forconverting radio signals received over an RF channelvia the antenna into analog audio signals, and atransmitter. The transmitter includes a coder-decoder(codec) for digitizing analog audio signals to betransmitted into frames of digitized speechinformation, the speech information including bothspeech and background noise. A digital signalprocessor processes a current frame based on anestimate of the background noise and the detection ofspeech in the current frame to minimize backgroundnoise. A modulator modulates an RF carrier with theprocessed frame of digitized speech information forsubsequent transmission via the antenna.
  • These and other features and advantages of thepresent invention will be readily apparent to one ofordinary skill in the art from the following written description, read in conjunction with the drawings, inwhich:
    • FIGURE 1 is a general functional block diagram ofthe present invention;
    • FIGURE 2 illustrates the frame and slot structureof the U.S. digital standard IS-54 for cellular radiocommunications;
    • FIGURE 3 is a block diagram of a first preferredembodiment of the present invention implemented using adigital signal processor;
    • FIGURE 4 is a functional block diagram of anexemplary embodiment of the present invention in one ofplural portable radio transceivers in atelecommunication system;
    • FIGURES 5A and 5B is a flow chart which illustratesfunctions/operations performed by the digital signalprocessor in implementing the first preferredembodiment of the present invention;
    • FIGURE 6A is a graph illustrating a first exampleof an attenuation vs. frequency characteristic of afilter circuit according to the first preferredembodiment of the present invention;
    • FIGURE 6B is a graph illustrating a second exampleof an attenuation vs. frequency characteristic of afilter circuit according to the first preferredembodiment of the present invention;
    • FIGURE 7 is an example look-up table accessible bythe filter control circuit of the first preferredembodiment of the present invention;
    • FIGURES 8A and 8B are graphs illustrating theamplitude vs. frequency characteristics of exampleinput audio signals;
    • FIGURES 9A and 9B are graphs illustrating theamplitude vs. frequency characteristics of the inputaudio signals of Figures 8A and 8B, respectively, afterhaving been filtered by the filter circuit of thepresent invention;
    • FIGURE 10 is a block diagram of a second preferredembodiment of the present invention implemented using adigital signal processor;
    • FIGURE 11 is a flow chart, corresponding to theflow chart of Figure 5B, which illustratesfunctions/operations performed by the digital signalprocessor in implementing the second preferredembodiment of the present invention; and
    • FIGURE 12 is an example look-up table accessible bythe filter control circuit of the second preferredembodiment of the present invention.
    • In the following description, for purposes ofexplanation and not limitation, specific details areset forth, such as particular circuits, circuitcomponents, techniques, flow charts, etc. in order toprovide a thorough understanding of the invention.However, it will be apparent to one skilled in the artthat the present invention may be practiced in otherembodiments that depart from these specific details.In other instances, detailed descriptions of well known methods, devices, and circuits are omitted so as not toobscure the description of the present invention withunnecessary details.
    • Figure 1 is a general block diagram of the adaptivenoise reduction system 100 according to the presentinvention. Adaptivenoise reduction system 100includes afilter control circuit 105 connected to afilter circuit 115.Filter control circuit 105generates a filter control signal for a current frameof a digitized audio signal. The filter control signalis outputted to thefilter circuit 115, and thefiltercircuit 115 adjusts in response to the filter controlsignal to exhibit a high-pass frequency response curveselected based on the filter control signal. Theadjustedfilter circuit 115 filters the current frameof the digitized audio signal. The filtering signal isprocessed by avoice coder 120 to produce a codedsignal representing the digitized audio signal.
    • In an exemplary embodiment of the invention appliedto portable/mobile radio telephone transceivers in acellular telecommunications system, Figure 2illustrates the time division multiple access (TDMA)frame structure employed by the IS-54 standard fordigital cellular telecommunications. A "frame" is atwenty millisecond time period which includes onetransmit block TX, one receive block RX, and a signalstrength measurement block used for mobile-assistedhand-off (MAHO). The two consecutive frames shown inFigure 2 are transmitted in a forty millisecond timeperiod. Digitized speech and background noise information is processed and filtered on a frame-by-framebasis as further described below.
    • Preferably, the functions of thefilter controlcircuit 105,filter circuit 115, andvoice coder 120shown in Figure 1 are implemented with a high speeddigital signal processor. One suitable digital signalprocessor is the TMS320C53 DSP available from TexasInstruments. The TMS320C53 DSP includes on a singleintegrated chip a sixteen-bit microprocessor, on-chipRAM for storing data such as speech frames to beprocessed, ROM for storing various data processingalgorithms including the VSELP speech compressionalgorithm, and other algorithms to be described belowfor implementing the functions performed by thefiltercontrol circuit 105 and thefilter circuit 115.
    • A first embodiment of the present invention isshown in Figure 3. In the first embodiment, thefiltercircuit 115 is adjusted as a function of backgroundnoise estimates determined by the filter controlcircuit. Frames of pulse code modulated (PCM) audioinformation are sequentially stored in the DSP's on-chipRAM. The audio information could be digitizedusing other digitization techniques. Each PCM frame isretrieved from a DSP on-chip RAM and processed byframeenergy estimator 210, and stored temporarily intemporary frame store 220. The energy of the currentframe determined byframe energy estimator 210 isprovided tonoise estimator 230 andspeech detector 240function blocks.Speech detector 240 indicates thatspeech is present in the current frame when the frame energy estimate exceeds the sum of the previous noiseestimate and a speech threshold. If thespeechdetector 240 determines that no speech is present, thedigital signal processor 200 calculates an updatednoise estimate as a function of the previous noiseestimate and the current frame energy (block 230).
    • The updated noise estimate is outputted to afilterselector 235.Filter selector 235 generates a filtercontrol signal based on the noise estimate. In thepreferred embodiment, thefilter selector 235 accessesa look-up table in generating the filter controlsignal. The look-up table includes a series of filtercontrol values that are each matched with a noiseestimate or range of noise estimates. A filter controlvalue from a look-up table is selected based on theupdated noise estimate and this filter control value isrepresented by a filter control signal outputted to afilter bank 265 for thefilter circuit 115. Tostabilize the process and avoid accessive switchingbetween different filters a hangover time of N framesis set upon the selection of a new filter. A newfilter can only be selected every N frames, where N isan integer greater than one and preferably greater than10.
    • Thefilter circuit 115 is adjusted in response tothe filter control signal to exhibit a high-passfrequency response curve that corresponds with theinputted filter control signal and noise estimate.Various different types of filter circuits well knownin prior art can be utilized to exhibit selected frequency response curves in response to the filtercontrol signal. These prior art filters include IIRfilters such as Butterworth, Chebyshev (Tschebyscheff) orelliptic filters. IIR filters are preferable to FIRfilters, which also can be used, due to lower processingrequirements.
    • The filtered signal is processed by avoice coder 120which is used to compress the bit rate of the filteredsignal. In the preferred embodiments, thevoice coder 120uses vector sourcebook excited linear predictive coding(VSELP) to code the audio signal. Other voice codingtechniques and algorithms such as code excited linearpredictive (CELP) codings, residual pulse excited linearpredictive (RPE-LTP) coding, improved multiband excited(IMBE) coding can be used. By filtering the frames ofaudio signals in accordance with the present inventionbefore voice coding, background noise is minimized whichsubstantially reduces any undesired noise effects in thespeech when it is reconstituted. It also prevents thespeech from being "drowned" in low frequent noise.
    • Thedigital signal processor 200 described inconjunction with Figure 3 can be used, for example, in thetransceiver of a digital portable/mobile radiotelephoneused in a radio telecommunications system. Figure 4illustrates one such digital radio transceiver which maybe used in a cellular telecommunications network.
    • Audio signals including speech and background noiseare input in amicrophone 400 to a coder-decoder (codec)402 which preferably is an application specific integratedcircuit (ASIC). The band limited audio signals detectedatmicrophone 400 are sampled by thecodec 402 at a rateof 8,000 samples per second and blocked into frames.Accordingly, each twenty millisecond frame includes 160speech samples. These samples are quantized and convertedinto a coded digital format such as 14-bit linear PCM.Once 160 samples of digitized speech for a current frameare stored in a transmitDSP 200 in on-chip RAM 202, thetransmitDSP 200 performs channel encoding functions, theframe energy estimation, noise estimation, speechdetection, FFT, filter functions and digital speechcoding/compression in accordance with the VSELP algorithm,as described above in conjunction with Figure 3.
    • Asupervisory microprocessor 432 controls the overalloperation of all of the components in the transceivershown in Figure 4. The filtered PCM data stream generatedby transmitDSP 200 is provided for quadrature modulationand transmission. To this end, anASIC gate array 404generates in-phase (I) and quadrature (Q) channels ofinformation based upon the filtered PCM data stream fromDSP 200. The I and Q bit streams are processed by matched, low pass filters 406and 408 and passed onto IQ mixers inbalanced modulator410. Areference oscillator 412 and amultiplier 414provide a transmit intermediate frequency (IF). The Isignal is mixed with in-phase IF, and the Q signal ismixed with quadrature IF (i.e., the in-phase IF delayedby 90 degrees by phase shifter 416). The mixed I and Qsignals are summed, converted "up" to an RF channelfrequency selected bychannel synthesizer 430, andtransmitted via duplexer 420 andantenna 422 over theselected radio frequency channel.
    • On the receive side, signals received viaantenna422 and duplexer 420 are down converted from theselected receive channel frequency in amixer 424 to afirst IF frequency using a local oscillator signalsynthesized bychannel synthesizer 430 based on theoutput ofreference oscillator 428. The output of thefirst IFmixer 424 is filtered and down converted infrequency to a second IF frequency based on anotheroutput fromchannel synthesizer 430 anddemodulator426. A receivegate array 434 then converts the secondIF signal into a series of phase samples and a seriesof frequency samples. The receiveDSP 436 performsdemodulation, filtering, gain/attenuation, channeldecoding, and speech expansion on the received signals.The processed speech data are then sent tocodec 402and converted to baseband audio signals for drivingloudspeaker 438.
    • The operations performed by thedigital signalprocessor 200 for implementing the functions offilter control circuit 105,filter circuit 115, andvoicecoder 120 will now be described in conjunction with theflow chart illustrated in Figures 5A and 5B.Frameenergy estimator 210 determines the energy in eachframe of audio signals.Frame energy estimator 210determines the energy of the current frame bycalculating the sum of the squared values of each PCMsample in the frame (step 505). Since there are 160samples per twenty millisecond frame for an 8000samples per second sampling rate, 160 squared PCMsamples are summed. Expressed mathematically, theframe energy estimate is determined according toequation 1 below:
      Figure 00170001
      The frame energy value calculated for the current frameis stored in the on-chip RAM 202 of DSP 200 (step 510).
    • The functions ofspeech detector 240 includefetching a noise estimate previously determined bynoise estimator 230 from the on-chip RAM of DSP 200(step 515). Of course, when the transceiver isinitially powered up, no noise estimate will exist.Decision block 520 anticipates this situation andassigns a noise estimate instep 525. Preferably, anarbitrarily high value, e.g. 20 dB above normal speechlevels, is assigned as the noise estimate in order toforce an update of the noise estimate value as will bedescribed below. The frame energy determined byframe energy estimator 210 is retrieved from the on-chip RAM202 of DSP 200 (block 530). A decision is made inblock 535 as to whether the frame energy estimateexceeds the sum of the retrieved noise estimate plus apredetermined speech threshold value, as shown inequation 2 below:frame energy estimate > (noise estimate + speech threshold)
    • The speech threshold value may be a fixed valuedetermined empirically to be larger than short termenergy variations of typical background noise and may,for example, be set to 9 dB. In addition, the speechthreshold value may be adaptively modified to reflectchanging speech conditions such as when the speakerenters a noisier or quieter environment. If the frameenergy estimate exceeds the sum inequation 2, a flagis set inblock 570 that speech exists. Ifspeechdetector 240 detects that speech exists, thennoiseestimator 230 is bypassed and the noise estimatecalculated for the previous frame in the digitizedaudio is retrieved and used as the current noiseestimate. Conversely, if the frame energy estimate isless than the sum inequation 2, the speech flag isreset inblock 540.
    • Other systems for detecting speech in a currentframe can also be used. For example, the EuropeanTelecommunications Standards Institute (ETSI) hasdeveloped a standard for voice activity detection (VAD)in the Global System for Mobile communications (GSM)system and is described in the ETSI Reference: RE/SMG-020632P which is incorporated by reference. Thisstandard could be used for speech detection in thepresent invention and is incorporated by reference.
    • If speech does not exist, the noise estimationupdate routine ofnoise estimator 230 is executed. Inessence, the noise estimate is a running average of theframe energy during periods of no speech. As describedabove, if the initial start-up noise estimate is chosensufficiently high, speech is not detected, and thespeech flag will be reset thereby forcing an update ofthe noise estimate.
    • In the noise estimation routine followed bynoiseestimator 230, a difference/error delta (Δ) isdetermined inblock 545 between the frame noise energygenerated byframe energy estimator 210 and a noiseestimate previously calculated bynoise estimator 230in accordance with the following equation:Δ = current frame energy - previous noise estimateA determination is made indecision block 550 whether Δexceeds zero. If Δ is negative, as occurs for highvalues of the noise estimate, then the noise estimateis recalculated inblock 560 in accordance with thefollowing equation:noise estimate = previous noise estimate + Δ/2Since Δ is negative, this results in a downwardcorrection of the noise estimate. The relatively largestep size of Δ/2 is chosen to rapidly correct for decreasing noise levels. However, if the frame energyexceeds the noise estimate, providing a Δ greater thanzero, the noise is updated inblock 555 in accordancewith the following equation:noise estimate = previous noise estimate + Δ/256Since Δ is positive, the noise estimate must beincreased. However, a smaller step size of Δ/256 (ascompared to Δ/2) is chosen to gradually increase thenoise estimate and provide substantial immunity totransient noise.
    • The noise estimate calculated for the current frameis outputted to thefilter selector 235. In the firstpreferred embodiment,filter selector 235 accesses alook-up table and uses the current noise estimate toselect a filter control value (Step 572). The filtercircuit 115 (in Step 574) is then adjusted as a functionof the selected filter control value to exhibit afrequency response curve intended to increase the amountof noise filtered as the noise estimate and backgroundnoise increases. The PCM samples stored in DSP RAM arethen passed through the adjustedfilter circuit 265 tofilter the PCM samples in order to remove noise (Step576). The filtered PCM samples are then processed byvoice coder 120 (step 578), and the coded samples arethen outputted to RF transmit circuits (Step 580).
    • Figures 6A and 6B show examples of how thefiltercircuit 115 adjusts to exhibit different frequencyresponse curves F1-F4 for different filter control signals inputted to thefilter circuit 115. As shown inFigure 6A, thefilter circuit 115 can be selected toexhibit a series of different frequency response curveswith the frequency response curves F1-F4 having cut-offfrequencies F1c-F4c, respectively. The cut-offfrequencies offilter circuit 115 may range in thepreferred embodiment from 300 Hz to 800 Hz. As thenoise estimates increase, thefilter circuit 115 isdesigned to exhibit frequency response curves havinghigher cut-off frequencies. The higher cut-offfrequencies result in a larger portion of frame energyfalling within the lower frequency range of speech beingextracted by thefilter circuit 115.
    • Likewise, as shown in Figure 6B, thefilter circuit115 can be selected to exhibit a series of differentfrequency response curves F1-F4 with each frequencyresponse curve having a different slope and the samecut-off frequencies. The cut-off frequency forfrequency response curves F1-F4 is in the above-mentionedrange. As the noise estimate increases, thefilter circuit 115 is adjusted to exhibit frequencyresponse curves having steeper slopes. The steeperslopes result in a larger portion of frame energyfalling within the lower frequency range of speech beingextracted by thefilter circuit 115.
    • Thefilter circuit 115 filters the current frames asa function of the noise estimate calculated for thecurrent frame. The current frame is filtered so thatthe noise is reduced and a major portion of the speechis passed. The major portion of speech which is passed unfiltered provides for recognizable speech output withonly a minimal reduction in the quality of the speechsignal. A combination of different cutoff frequenciesand different slopes could be used for adaptivelyextracting selected portions of frame energy fallingwithin a low frequency range of speech.
    • Figure 7 depicts an example look-up table accessedbyfilter selector 235 in order to select one of thefilter response curves F1-F4 forfilter circuit 115.The look-up table includes a series of potential noiseestimates N1-Nn and filter control values F1-Fn thatcorrespond with potential response curves that areexhibitable by thefilter circuit 115. Noise estimatesN1-Nn can each represent a range of noise estimates andare each matched with a particular filter control valueF1-F4. Thefilter control circuit 105 generates afilter control signal by calculating a noise estimateand retrieving from the look-up table the filter controlvalue associated therewith.
    • Figures 8A & B and 9A & B show how the audio signalfor two frames are each adaptively filtered to providean improved audio signal outputted to the RFtransmitter. Figures 8A and 8B show a first frame and asecond frame of an audio signal containing speechcomponents s1 and s2 and noise components n1 and n2,respectively. As shown, the noise energy n1 and n2 inboth frames is concentrated in a low audible frequencyrange, while the speech energy s1 and s2 is concentratedin a higher audible frequency range. Figure 9A showsthe noise signal n1 and speech signal s1 for the first frame after filtering. Figure 9B shows the noise signaln2 and speech signal s2 for the second frame afterfiltering.
    • The adaptive audionoise reduction system 100, asdiscussed, is designed to account for the difference innoise level between the first frame and the second frameby adjusting thefilter control circuit 105 based on acalculated noise estimate for the current frame. Forexample, a noise estimate N1 and a spectral profile S1is calculated byfilter control circuit 105 and a filtercontrol value of F1 is selected for the first frame. Inthe preferred embodiment, thefilter circuit 115 isadjusted based on filter control value F1 and exhibits afrequency response curve F1 having a cut-off frequencyFlc, as shown in Figure 6A. The first frame is passedthrough this adjustedfilter circuit 115. Thefiltercircuit 115 is selected so that a large portion of thenoise n1 and only a small portion of speech s1 fallsbelow the cut-off frequency F1c of the frequencyresponse curve F1. This results in noise n1 beingeffectively filtered and only a relatively insignificantportion of speech s1 being filtered. The filtered audiosignal of the first frame is shown in Figure 9A.
    • In the second frame shown in Figure 8b, a higherbackground noise is present, and assuming speech is notdetected, a higher noise estimate n2 is calculated byfilter control circuit 105. A higher correspondingfilter control value F2 is determined for the secondframe based on the higher noise estimate. In the firstpreferred embodiment, thefilter circuit 115 is adjusted in response to the higher filter control value F2 toexhibit a frequency response curve having a higher cut-offfrequency F2c, as shown in Figure 6A. Thesubsequent frame of audio signal is passed through theadjustedfilter circuit 115. Because the cut-offfrequency F2c of the frequency response curve F2 ishigher for the subsequent frame, a larger portion ofboth the noise n2 and speech s2 is filtered. Theportion of speech s2 filtered is still relativelyinsignificant to the intelligibility informationcontained by the frame so that there is only minimalaffect on the speech. The disadvantage of filtering alarger portion of the speech s2 is offset by theadvantage of the increased removal of noise n2 from thesecond frame. The filtered spectral portion of thespeech does not significantly contribute to theintelligibility of the speech. The filtered audiosignal of the second frame is shown in Figure 9B.
    • A second preferred embodiment of adaptivenoisereduction system 100 is shown in Figures 10-12. In thesecond preferred embodiment, thefilter control circuit105 adjusts thefilter circuit 115 as a function ofnoise profile estimates. A noise profile estimate iscalculated for each frame and is compared to a referencenoise profile. Based on this comparison, thefiltercircuit 115 is adaptively adjusted to extract varyingamounts of low frequency energy from the current frame.
    • Referring to Figure 10, aDSP 200 configuredaccording to the second preferred embodiment is shown.As shown, thefilter control circuit 105 includes aspectral analyzer 270, in addition toframe energyestimator 210,noise estimator 230,speech detector 240,andfilter selector 235 which are described with respectto the first preferred embodiment. Thefilter controlcircuit 105 determines noise estimates and detectsspeech for the received frames as described for thefirst embodiment and shown in flow charts 5A and 5B.Upon speech detection for a current frame, thespectralanalyzer 270 updates the noise profile estimate and usesthe noise profile estimate in adjusting thefiltercircuit 115.
    • Referring to Figure 11, the steps of updating thenoise profile estimate and adjusting thefilter circuit115 is shown. Figure 11 shows the steps performed byspectral analyzer 270 incorporated into the overallprocess previously described in the flow charts ofFigures 5A and 5B for the first preferred embodiment.
    • When speech is not detected for the current frame,thespectral analyzer 270 first determines a noiseprofile for the current frame (step 600). The noiseprofile determined for the current frame includesenergy calculations for different frequencies (i.e.,frequency bins) within a selected low frequency rangeof speech for the current frame. In the preferredembodiment, the selected frequency range isapproximately 300 to 800 hertz. The noise profile ofthe current frame can be determined by processing thecurrent frame using a Fast Fourier Transform (FFT)having N frequency bins. Processing digital signalsusing an FFT is well-known in the prior art and is advantageous in that very little processing power isrequired where the FFT is limited to a relatively smallnumber of frequency bins such as 32. An FFT having Nfrequency bins produces energy calculations at Ndifferent frequencies. The energy calculations for thefrequency bins falling within the selected frequencyrange form the noise profile for the current frame.
    • To determine the noise profile estimate for thecurrent frame (step 604), the noise profile for thecurrent frame is averaged with a noise profile estimatedetermined for the previous frame of the audio signal.Where no previous noise profile estimate is available,such as after initialization, a stored, initial noiseprofile estimate can be used. The noise profileestimate includes noise energy estimates ei (where i =1,2,...n) located at successively lower frequencies(i.e., e1 is the noise energy estimate for the highestfrequency and en is the noise energy estimate for thelowest frequency in the selected frequency range). Inthe preferred embodiment, each noise energy estimate eicorresponds to an average of the energy calculations ata particular frequency in the selected frequency rangeover a plurality of successive frames in which nospeech was detected. By using a plurality of frames indetermining the noise profile estimate, thefiltercircuit 115 is adjusted on a more gradual basis. Inalternate embodiments, the noise profile estimate canbe equated to the noise profile of the current frame.
    • The energy estimates ei of the noise profileestimate are then compared with a reference noise profile (step 604). The reference noise profileincludes reference energy thresholds eri (where i =1,2,...n) at frequencies corresponding to thefrequencies for noise energy estimates ei of the noiseprofile estimate. The reference energy thresholds erican be determined empirically. The noise energyestimates ei are successively compared to correspondingreference energy thresholds eri from the highestfrequency energy estimate e1 to the lowest frequencyenergy estimate en.
    • More specifically, noise energy estimate e1 isfirst compared to reference noise threshold er1. If e1is greater than reference noise threshold er1, then acomparison value c1 is selected and inputted intofilter selector 235. If noise energy estimate e1 isless than reference noise threshold er1, then noiseenergy estimate e2 (which is a noise energy estimatetaken at a lower frequency than e1) is compared toreference noise threshold er2. If noise energyestimate e2 is greater than reference noise thresholder2, then a comparison value c2 is selected and inputtedto filterselector 235. This comparison process iscontinued until a comparison value ci (where i =1,2,...n) is selected.
    • Thefilter circuit 235 uses the determinedcomparison value ci to determine a filter control value.The filter control value is selected from a look-uptable such as that shown in Figure 12. The look-uptable includes a series of comparison values ci and corresponding filter control values Fi. Thefiltercircuit 115 is adjusted as a function of the selectedfilter control value. Thefilter circuit 115 isadjusted to exhibit a frequency response curve forextracting low frequency energy from the current frame.Thefilter circuit 115 is adjusted to extract increasingamounts of low frequency energy as noise energyestimates at successively higher frequencies surpasstheir corresponding reference energy thresholds. Figure6A and 6B show example frequency response curves forselected filter control values.
    • Use of noise profile estimates helps improve theability to adaptively adjust the filter circuit toextract low frequency energy in a manner to improve theoverall quality of speech. Since the car environment isnot the only environment where a mobiletelecommunications device is used, and therefore thenoise profile in certain situations could be tiltedmore towards higher frequencies, thespectral analyzer270 can be selectively disabled when noise energy inthe low frequencies is small. Also, when a significantportion of the noise frequency spectrum resides inlower frequencies a steeper filtering slope could beapplied even though some processing power may besacrificed. This extra processing requirement is stillfairly small.
    • As is evident from the description above, theadaptive noise filter system of the present invention isimplemented simply and without significant increase inDSP calculations. More complex methods of reducing noise, such as "spectral subtraction," require severalcalculation-relates MIPS and a large amount of memory fordata and program code storage. By comparison, the presentinvention may be implemented using only a fraction of theMIPS and memory required for the "spectral subtraction"algorithm which also introduces more speech distortion.Reduced memory reduces the size of the DSP integratedcircuits; decreased MIPS decreases power consumption.Both of these attributes are desirable for battery-poweredportable/mobile radiotelephones.
    • While the invention has been particularly shown anddescribed with reference to the preferred embodimentsthereof, it is not limited to those embodiments. Forexample, although a DSP is disclosed as performing thefunctions of theframe energy estimator 210,noiseestimator 230,speech detector 240,filter selector 235andfilter circuit 265, these functions could beimplemented using other digital and/or analog components.In addition, anadaptive filtering system 100 could beimplemented where thefilter circuit 115 is adjusted asa function of both noise estimates and noise profileestimates.

    Claims (10)

    1. A method for selectively altering a frame of adigital signal formed of a plurality of successive frames,the digital signal representative of an audio signalreceived at a transmitter, the audio signal formedalternately of a speech component, a noise component, andthe speech component together with the noise component,said method comprising the steps of:
      estimating an energy level (505) of a frame ofthe digital signal;
      determining (535), responsive to the estimatemade during said step of estimating, whether the frame ofthe digital signal includes a speech component;
      updating a noise estimate as a function of apreceding noise estimate and the energy level estimatedduring said step of estimating when said step ofdetermining determines that the frame does not contain a speech component;
      accessing (572) an entry in a look-up tablehaving filter characteristics indexed against levels ofnoise estimates, the entry accessed associated with thenoise estimate updated during said step of updating;
      selecting (574) filter characteristics for afilter such that the filter exhibits a frequency responsecurve having variable gain over different frequencyranges, the filter characteristic selected responsive tothe stored filter characteristics of the entry accessedduring said step of accessing; and
      filtering (576) the frame of the digital datawith the filter which exhibits the filter characteristics,thereby to alter the frame of the digital data responsiveto the filter characteristics.
    2. The method of claim 1 further characterized bythe additional intermediary step of determining (600) anoise profile estimate of the frame of the digital signalif the frame of the digital data is determined not toinclude the speech component.
    3. The method of claim 2 wherein the noise profileestimate determined during said step of determining (600)the noise profile estimate is used during said step ofupdating to update the noise estimate.
    4. The method of claim 1 wherein the look-up tableaccessed during said step of accessing is characterizedby a plurality of entries (C1-CN, F4-FN), each entry ofthe plurality including a separate filter characteristic.
    5. The method of claim 4 wherein the separatefilter characteristics of the plurality of entries of thelook-up table comprise separate high pass filtercharacteristics, each high pass characteristic defined bya separate cut-off frequency (F1c, F2c, F3c, F4c).
    6. The method of claim 4 wherein the separatefilter characteristics of the plurality of entries of thelook-up table comprise separate high pass filtercharacteristics, each high pass filter characteristicdefined by a separate frequency response curve slope (F1,F2, F3, F4).
    7. The method of claim 1 characterized by thefurther step of incrementing a counter value to count eachframe for which an energy level is estimated during saidstep of estimating.
    8. The method of claim 7 wherein said step ofselecting the filter-circuit filter characteristics isperformed when the counter value is incremented each Nthtime, N forming an integer value greater than one.
    9. An apparatus (100; 200) for selectively alteringa frame of a digital signal formed of a plurality ofsuccessive frames, the digital signal representative ofan audio signal received at a transmitter, the audio signal formed alternately of a speech component, a noisecomponent, and the speech component together with thenoise component, said apparatus comprising:
      an energy level estimator (210) coupled toreceive indications of a frame of the digital signal, saidenergy level estimator for estimating an energy level ofthe frame of the digital signal;
      a speech detector (240) coupled to said energylevel estimator, said speech detector for determiningwhether the frame of the digital signal includes a speechcomponent;
      a noise estimator (230) operable when saidspeech detector determines that a frame does not containa speech component, said noise estimator for updating anoise estimate as a function of a preceding noiseestimate, and the energy level estimated by saidestimator;
      a look-up table (FIG. 12) containing a pluralityof entries, each entry indexed against levels of noiseestimates, an entry of said look-up table accessedresponsive to a noise estimate formed by said noiseestimator; and
      a filter (265) coupled to receive the frame ofthe digital data, said filter exhibiting selectable filtercharacteristics enabling the filter to exhibit a frequencyresponse curve having variable gain over differentfrequency ranges, selection of the filter characteristicsof the filter determined responsive to the entry of thelook-up table accessed responsive to the noise estimateupdated by said noise estimator.
    10. The apparatus (100; 200) of claim 9 furthercharacterized by a noise profile estimator (270) fordetermining a noise profile estimate of the frame of thedigital data if the frame of the digital data isdetermined by said speech component determiner not toinclude the speech component.
    EP96931552A1995-09-141996-09-13System for adaptively filtering audio signals to enhance speech intelligibility in noisy environmental conditionsExpired - LifetimeEP0852052B1 (en)

    Applications Claiming Priority (3)

    Application NumberPriority DateFiling DateTitle
    US52800595A1995-09-141995-09-14
    US5280051995-09-14
    PCT/US1996/014665WO1997010586A1 (en)1995-09-141996-09-13System for adaptively filtering audio signals to enhance speech intelligibility in noisy environmental conditions

    Publications (2)

    Publication NumberPublication Date
    EP0852052A1 EP0852052A1 (en)1998-07-08
    EP0852052B1true EP0852052B1 (en)2001-06-13

    Family

    ID=24103874

    Family Applications (1)

    Application NumberTitlePriority DateFiling Date
    EP96931552AExpired - LifetimeEP0852052B1 (en)1995-09-141996-09-13System for adaptively filtering audio signals to enhance speech intelligibility in noisy environmental conditions

    Country Status (15)

    CountryLink
    EP (1)EP0852052B1 (en)
    JP (1)JPH11514453A (en)
    KR (1)KR100423029B1 (en)
    CN (1)CN1121684C (en)
    AU (1)AU724111B2 (en)
    BR (1)BR9610290A (en)
    CA (1)CA2231107A1 (en)
    DE (1)DE69613380D1 (en)
    EE (1)EE03456B1 (en)
    MX (1)MX9801857A (en)
    NO (1)NO981074L (en)
    PL (1)PL185513B1 (en)
    RU (1)RU2163032C2 (en)
    TR (1)TR199800475T1 (en)
    WO (1)WO1997010586A1 (en)

    Cited By (2)

    * Cited by examiner, † Cited by third party
    Publication numberPriority datePublication dateAssigneeTitle
    CN102128976A (en)*2011-01-072011-07-20钜泉光电科技(上海)股份有限公司Energy pulse output method and device of electric energy meter and electric energy meter
    WO2021179045A1 (en)*2020-03-132021-09-16University Of South AustraliaA data processing method

    Families Citing this family (180)

    * Cited by examiner, † Cited by third party
    Publication numberPriority datePublication dateAssigneeTitle
    DE19747885B4 (en)*1997-10-302009-04-23Harman Becker Automotive Systems Gmbh Method for reducing interference of acoustic signals by means of the adaptive filter method of spectral subtraction
    DE69815062T2 (en)*1997-10-312004-02-26Koninklijke Philips Electronics N.V. METHOD AND DEVICE FOR AUDIO REPRESENTATION OF LANGUAGE CODED BY THE LPC PRINCIPLE BY ADDING NOISE SIGNALS
    KR20000074236A (en)*1999-05-192000-12-15정몽규Auto audio volume control means
    US8645137B2 (en)2000-03-162014-02-04Apple Inc.Fast, language-independent method for user authentication by voice
    JP2001318694A (en)*2000-05-102001-11-16Toshiba Corp Signal processing device, signal processing method and recording medium
    US6983242B1 (en)*2000-08-212006-01-03Mindspeed Technologies, Inc.Method for robust classification in speech coding
    KR20030010432A (en)*2001-07-282003-02-05주식회사 엑스텔테크놀러지Apparatus for speech recognition in noisy environment
    RU2221284C2 (en)*2002-02-262004-01-10Дочернее государственное унитарное научно-исследовательское предприятие "АРГУС"Coded speech transmission and reception method
    IL148592A0 (en)2002-03-102002-09-12Ycd Multimedia LtdDynamic normalizing
    KR100978015B1 (en)*2002-07-012010-08-25코닌클리케 필립스 일렉트로닉스 엔.브이. Fixed Spectrum Power Dependent Audio Enhancement System
    CN100477705C (en)*2002-07-012009-04-08皇家飞利浦电子股份有限公司Audio enhancement system, system provided with such system, and distortion signal enhancement method
    DE60222813T2 (en)*2002-07-122008-07-03Widex A/S HEARING DEVICE AND METHOD FOR INCREASING REDEEMBLY
    US7242763B2 (en)2002-11-262007-07-10Lucent Technologies Inc.Systems and methods for far-end noise reduction and near-end noise compensation in a mixed time-frequency domain compander to improve signal quality in communications systems
    KR20040058855A (en)*2002-12-272004-07-05엘지전자 주식회사voice modification device and the method
    DE10305369B4 (en)*2003-02-102005-05-19Siemens Ag User-adaptive method for noise modeling
    US7127076B2 (en)2003-03-032006-10-24Phonak AgMethod for manufacturing acoustical devices and for reducing especially wind disturbances
    DK1339256T3 (en)*2003-03-032018-01-29Sonova Ag Process for the manufacture of acoustic appliances and to reduce wind disturbance
    US7499686B2 (en)*2004-02-242009-03-03Microsoft CorporationMethod and apparatus for multi-sensory speech enhancement on a mobile device
    US7830900B2 (en)2004-08-302010-11-09Qualcomm IncorporatedMethod and apparatus for an adaptive de-jitter buffer
    KR100640865B1 (en)2004-09-072006-11-02엘지전자 주식회사 Method and device to improve voice quality
    US8085678B2 (en)2004-10-132011-12-27Qualcomm IncorporatedMedia (voice) playback (de-jitter) buffer adjustments based on air interface
    EP1840874B1 (en)2005-01-112019-04-10NEC CorporationAudio encoding device, audio encoding method, and audio encoding program
    US8155965B2 (en)*2005-03-112012-04-10Qualcomm IncorporatedTime warping frames inside the vocoder by modifying the residual
    US7680656B2 (en)*2005-06-282010-03-16Microsoft CorporationMulti-sensory speech enhancement using a speech-state model
    WO2007015203A1 (en)*2005-08-022007-02-08Koninklijke Philips Electronics N.V.Enhancement of speech intelligibility in a mobile communication device by controlling the operation of a vibrator in dξpendance of the background noise
    GB2429139B (en)2005-08-102010-06-16Zarlink Semiconductor IncA low complexity noise reduction method
    US8677377B2 (en)2005-09-082014-03-18Apple Inc.Method and apparatus for building an intelligent automated assistant
    KR100667852B1 (en)*2006-01-132007-01-11삼성전자주식회사 Noise canceller and method for portable recorder equipment
    CN101882441B (en)*2006-01-272013-02-27杜比国际公司Efficient filtering with a complex modulated filterbank
    RU2417460C2 (en)*2006-06-052011-04-27Эксаудио АбBlind signal extraction
    US9318108B2 (en)2010-01-182016-04-19Apple Inc.Intelligent automated assistant
    US20080152167A1 (en)*2006-12-222008-06-26Step Communications CorporationNear-field vector signal enhancement
    KR101414233B1 (en)2007-01-052014-07-02삼성전자 주식회사 Apparatus and method for improving intelligibility of speech signal
    KR100883896B1 (en)*2007-01-192009-02-17엘지전자 주식회사 Voice Clarity Improvement Device and Method
    KR100876794B1 (en)*2007-04-032009-01-09삼성전자주식회사 Apparatus and method for improving speech intelligibility in a mobile terminal
    US8977255B2 (en)2007-04-032015-03-10Apple Inc.Method and system for operating a multi-function portable electronic device using voice-activation
    RU2343563C1 (en)*2007-05-212009-01-10Федеральное государственное унитарное предприятие "ПЕНЗЕНСКИЙ НАУЧНО-ИССЛЕДОВАТЕЛЬСКИЙ ЭЛЕКТРОТЕХНИЧЕСКИЙ ИНСТИТУТ" (ФГУП "ПНИЭИ")Way of transfer and reception of coded voice signals
    RU2374763C2 (en)*2007-08-072009-11-27Государственное образовательное учреждение высшего профессионального образования "Уральский государственный университет путей сообщения" (УрГУПС)Method for transfer of voice signals
    WO2009035614A1 (en)*2007-09-122009-03-19Dolby Laboratories Licensing CorporationSpeech enhancement with voice clarity
    EP2232704A4 (en)*2007-12-202010-12-01Ericsson Telefon Ab L MNoise suppression method and apparatus
    EP2232703B1 (en)2007-12-202014-06-18Telefonaktiebolaget LM Ericsson (publ)Noise suppression method and apparatus
    US9330720B2 (en)2008-01-032016-05-03Apple Inc.Methods and apparatus for altering audio output signals
    CN101221767B (en)*2008-01-232012-05-30晨星半导体股份有限公司 Human voice speech enhancement device and method applied thereto
    US8996376B2 (en)2008-04-052015-03-31Apple Inc.Intelligent text-to-speech conversion
    EP2279509B1 (en)2008-04-182012-12-19Dolby Laboratories Licensing CorporationMethod and apparatus for maintaining speech audibility in multi-channel audio with minimal impact on surround experience
    US10496753B2 (en)2010-01-182019-12-03Apple Inc.Automatically adapting user interfaces for hands-free interaction
    US20100030549A1 (en)2008-07-312010-02-04Lee Michael MMobile device having human language translation capability with positional feedback
    WO2010067118A1 (en)2008-12-112010-06-17Novauris Technologies LimitedSpeech recognition involving a mobile device
    DE102009011583A1 (en)2009-03-062010-09-09Krones Ag Method and device for producing and filling thin-walled beverage containers
    US9858925B2 (en)2009-06-052018-01-02Apple Inc.Using context information to facilitate processing of commands in a virtual assistant
    US20120309363A1 (en)2011-06-032012-12-06Apple Inc.Triggering notifications associated with tasks items that represent tasks to perform
    US10241752B2 (en)2011-09-302019-03-26Apple Inc.Interface for a virtual digital assistant
    US10241644B2 (en)2011-06-032019-03-26Apple Inc.Actionable reminder entries
    US9431006B2 (en)2009-07-022016-08-30Apple Inc.Methods and apparatuses for automatic speech recognition
    US10705794B2 (en)2010-01-182020-07-07Apple Inc.Automatically adapting user interfaces for hands-free interaction
    US10553209B2 (en)2010-01-182020-02-04Apple Inc.Systems and methods for hands-free notification summaries
    US10276170B2 (en)2010-01-182019-04-30Apple Inc.Intelligent automated assistant
    US10679605B2 (en)2010-01-182020-06-09Apple Inc.Hands-free list-reading by intelligent automated assistant
    DE112011100329T5 (en)2010-01-252012-10-31Andrew Peter Nelson Jerram Apparatus, methods and systems for a digital conversation management platform
    US8682667B2 (en)2010-02-252014-03-25Apple Inc.User profiling for selecting user specific voice input processing information
    CN102202038B (en)*2010-03-242015-05-06华为技术有限公司Method and system for realizing voice energy display, conference server and terminal
    JP5867389B2 (en)*2010-05-242016-02-24日本電気株式会社 Signal processing method, information processing apparatus, and signal processing program
    CN101859569B (en)*2010-05-272012-08-15上海朗谷电子科技有限公司Method for lowering noise of digital audio-frequency signal
    US8639516B2 (en)2010-06-042014-01-28Apple Inc.User-specific noise suppression for voice quality improvements
    US10762293B2 (en)2010-12-222020-09-01Apple Inc.Using parts-of-speech tagging and named entity recognition for spelling correction
    US9262612B2 (en)2011-03-212016-02-16Apple Inc.Device access using voice authentication
    US10057736B2 (en)2011-06-032018-08-21Apple Inc.Active transport based notifications
    US8994660B2 (en)2011-08-292015-03-31Apple Inc.Text correction processing
    EP3392876A1 (en)*2011-09-302018-10-24Apple Inc.Using context information to facilitate processing of commands in a virtual assistant
    US10134385B2 (en)2012-03-022018-11-20Apple Inc.Systems and methods for name pronunciation
    US9483461B2 (en)2012-03-062016-11-01Apple Inc.Handling speech synthesis of content for multiple languages
    US9280610B2 (en)2012-05-142016-03-08Apple Inc.Crowd sourcing information to fulfill user requests
    US9721563B2 (en)2012-06-082017-08-01Apple Inc.Name recognition system
    CN102737646A (en)*2012-06-212012-10-17佛山市瀚芯电子科技有限公司Real-time dynamic voice noise reduction method for single microphone
    US9495129B2 (en)2012-06-292016-11-15Apple Inc.Device, method, and user interface for voice-activated navigation and browsing of a document
    US9576574B2 (en)2012-09-102017-02-21Apple Inc.Context-sensitive handling of interruptions by intelligent digital assistant
    US9547647B2 (en)2012-09-192017-01-17Apple Inc.Voice-based media searching
    DE212014000045U1 (en)2013-02-072015-09-24Apple Inc. Voice trigger for a digital assistant
    US10652394B2 (en)2013-03-142020-05-12Apple Inc.System and method for processing voicemail
    US9368114B2 (en)2013-03-142016-06-14Apple Inc.Context-sensitive handling of interruptions
    WO2014144579A1 (en)2013-03-152014-09-18Apple Inc.System and method for updating an adaptive speech recognition model
    AU2014233517B2 (en)2013-03-152017-05-25Apple Inc.Training an at least partial voice command system
    CN104095640A (en)*2013-04-032014-10-15达尔生技股份有限公司 Blood oxygen saturation detection method and device
    WO2014197334A2 (en)2013-06-072014-12-11Apple Inc.System and method for user-specified pronunciation of words for speech synthesis and recognition
    WO2014197336A1 (en)2013-06-072014-12-11Apple Inc.System and method for detecting errors in interactions with a voice-based digital assistant
    US9582608B2 (en)2013-06-072017-02-28Apple Inc.Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
    WO2014197335A1 (en)2013-06-082014-12-11Apple Inc.Interpreting and acting upon commands that involve sharing information with remote devices
    DE112014002747T5 (en)2013-06-092016-03-03Apple Inc. Apparatus, method and graphical user interface for enabling conversation persistence over two or more instances of a digital assistant
    US10176167B2 (en)2013-06-092019-01-08Apple Inc.System and method for inferring user intent from speech inputs
    AU2014278595B2 (en)2013-06-132017-04-06Apple Inc.System and method for emergency calls initiated by voice command
    EP2816557B1 (en)*2013-06-202015-11-04Harman Becker Automotive Systems GmbHIdentifying spurious signals in audio signals
    US9697831B2 (en)2013-06-262017-07-04Cirrus Logic, Inc.Speech recognition
    DE112014003653B4 (en)2013-08-062024-04-18Apple Inc. Automatically activate intelligent responses based on activities from remote devices
    US9620105B2 (en)2014-05-152017-04-11Apple Inc.Analyzing audio input for efficient speech and music recognition
    US10592095B2 (en)2014-05-232020-03-17Apple Inc.Instantaneous speaking of content on touch devices
    US9502031B2 (en)2014-05-272016-11-22Apple Inc.Method for supporting dynamic grammars in WFST-based ASR
    US9633004B2 (en)2014-05-302017-04-25Apple Inc.Better resolution when referencing to concepts
    US9785630B2 (en)2014-05-302017-10-10Apple Inc.Text prediction using combined word N-gram and unigram language models
    US9760559B2 (en)2014-05-302017-09-12Apple Inc.Predictive text input
    US10170123B2 (en)2014-05-302019-01-01Apple Inc.Intelligent assistant for home automation
    US10289433B2 (en)2014-05-302019-05-14Apple Inc.Domain specific language for encoding assistant dialog
    US9734193B2 (en)2014-05-302017-08-15Apple Inc.Determining domain salience ranking from ambiguous words in natural speech
    US9842101B2 (en)2014-05-302017-12-12Apple Inc.Predictive conversion of language input
    US10078631B2 (en)2014-05-302018-09-18Apple Inc.Entropy-guided text prediction using combined word and character n-gram language models
    US9430463B2 (en)2014-05-302016-08-30Apple Inc.Exemplar-based natural language processing
    CN110797019B (en)2014-05-302023-08-29苹果公司Multi-command single speech input method
    US9715875B2 (en)2014-05-302017-07-25Apple Inc.Reducing the need for manual start/end-pointing and trigger phrases
    US10659851B2 (en)2014-06-302020-05-19Apple Inc.Real-time digital assistant knowledge updates
    US9338493B2 (en)2014-06-302016-05-10Apple Inc.Intelligent automated assistant for TV user interactions
    EP2980801A1 (en)2014-07-282016-02-03Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V.Method for estimating noise in an audio signal, noise estimator, audio encoder, audio decoder, and system for transmitting audio signals
    US10446141B2 (en)2014-08-282019-10-15Apple Inc.Automatic speech recognition based on user feedback
    US9818400B2 (en)2014-09-112017-11-14Apple Inc.Method and apparatus for discovering trending terms in speech requests
    US10789041B2 (en)2014-09-122020-09-29Apple Inc.Dynamic thresholds for always listening speech trigger
    US10127911B2 (en)2014-09-302018-11-13Apple Inc.Speaker identification and unsupervised speaker adaptation techniques
    US9646609B2 (en)2014-09-302017-05-09Apple Inc.Caching apparatus for serving phonetic pronunciations
    US9668121B2 (en)2014-09-302017-05-30Apple Inc.Social reminders
    US9886432B2 (en)2014-09-302018-02-06Apple Inc.Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
    US10074360B2 (en)2014-09-302018-09-11Apple Inc.Providing an indication of the suitability of speech recognition
    US10552013B2 (en)2014-12-022020-02-04Apple Inc.Data detection
    US9711141B2 (en)2014-12-092017-07-18Apple Inc.Disambiguating heteronyms in speech synthesis
    RU2589298C1 (en)*2014-12-292016-07-10Александр Юрьевич БредихинMethod of increasing legible and informative audio signals in the noise situation
    US9865280B2 (en)2015-03-062018-01-09Apple Inc.Structured dictation using intelligent automated assistants
    US9886953B2 (en)2015-03-082018-02-06Apple Inc.Virtual assistant activation
    US10567477B2 (en)2015-03-082020-02-18Apple Inc.Virtual assistant continuity
    US9721566B2 (en)2015-03-082017-08-01Apple Inc.Competing devices responding to voice triggers
    US9899019B2 (en)2015-03-182018-02-20Apple Inc.Systems and methods for structured stem and suffix language models
    US9842105B2 (en)2015-04-162017-12-12Apple Inc.Parsimonious continuous-space phrase representations for natural language processing
    US10083688B2 (en)2015-05-272018-09-25Apple Inc.Device voice control for selecting a displayed affordance
    US10127220B2 (en)2015-06-042018-11-13Apple Inc.Language identification from short strings
    US10101822B2 (en)2015-06-052018-10-16Apple Inc.Language input correction
    US11025565B2 (en)2015-06-072021-06-01Apple Inc.Personalized prediction of responses for instant messaging
    US10186254B2 (en)2015-06-072019-01-22Apple Inc.Context-based endpoint detection
    US10255907B2 (en)2015-06-072019-04-09Apple Inc.Automatic accent detection using acoustic models
    US10671428B2 (en)2015-09-082020-06-02Apple Inc.Distributed personal assistant
    US10747498B2 (en)2015-09-082020-08-18Apple Inc.Zero latency digital assistant
    US9697820B2 (en)2015-09-242017-07-04Apple Inc.Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
    US11010550B2 (en)2015-09-292021-05-18Apple Inc.Unified language modeling framework for word prediction, auto-completion and auto-correction
    US10366158B2 (en)2015-09-292019-07-30Apple Inc.Efficient word encoding for recurrent neural network language models
    US11587559B2 (en)2015-09-302023-02-21Apple Inc.Intelligent device identification
    US10691473B2 (en)2015-11-062020-06-23Apple Inc.Intelligent automated assistant in a messaging environment
    WO2017081092A1 (en)*2015-11-092017-05-18Nextlink Ipr AbMethod of and system for noise suppression
    US10049668B2 (en)2015-12-022018-08-14Apple Inc.Applying neural network language models to weighted finite state transducers for automatic speech recognition
    US10223066B2 (en)2015-12-232019-03-05Apple Inc.Proactive assistance based on dialog communication between devices
    CN105869650B (en)*2015-12-282020-03-06乐融致新电子科技(天津)有限公司Digital audio data playing method and device
    US10446143B2 (en)2016-03-142019-10-15Apple Inc.Identification of voice inputs providing credentials
    CN106060717A (en)*2016-05-262016-10-26广东睿盟计算机科技有限公司High-definition dynamic noise-reduction pickup
    US9934775B2 (en)2016-05-262018-04-03Apple Inc.Unit-selection text-to-speech synthesis based on predicted concatenation parameters
    US9972304B2 (en)2016-06-032018-05-15Apple Inc.Privacy preserving distributed evaluation framework for embedded personalized systems
    US10249300B2 (en)2016-06-062019-04-02Apple Inc.Intelligent list reading
    US10049663B2 (en)2016-06-082018-08-14Apple, Inc.Intelligent automated assistant for media exploration
    DK179309B1 (en)2016-06-092018-04-23Apple IncIntelligent automated assistant in a home environment
    US10586535B2 (en)2016-06-102020-03-10Apple Inc.Intelligent digital assistant in a multi-tasking environment
    US10509862B2 (en)2016-06-102019-12-17Apple Inc.Dynamic phrase expansion of language input
    US10192552B2 (en)2016-06-102019-01-29Apple Inc.Digital assistant providing whispered speech
    US10067938B2 (en)2016-06-102018-09-04Apple Inc.Multilingual word prediction
    US10490187B2 (en)2016-06-102019-11-26Apple Inc.Digital assistant providing automated status report
    DK179415B1 (en)2016-06-112018-06-14Apple IncIntelligent device arbitration and control
    DK179049B1 (en)2016-06-112017-09-18Apple IncData driven natural language event detection and classification
    DK201670540A1 (en)2016-06-112018-01-08Apple IncApplication integration with a digital assistant
    DK179343B1 (en)2016-06-112018-05-14Apple IncIntelligent task discovery
    US9748929B1 (en)*2016-10-242017-08-29Analog Devices, Inc.Envelope-dependent order-varying filter control
    US10593346B2 (en)2016-12-222020-03-17Apple Inc.Rank-reduced token representation for automatic speech recognition
    CN107039044B (en)*2017-03-082020-04-21Oppo广东移动通信有限公司 A kind of voice signal processing method and mobile terminal
    DK179745B1 (en)2017-05-122019-05-01Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
    DK201770431A1 (en)2017-05-152018-12-20Apple Inc.Optimizing dialogue policy decisions for digital assistants using implicit feedback
    US10157627B1 (en)2017-06-022018-12-18Bose CorporationDynamic spectral filtering
    WO2019187841A1 (en)*2018-03-302019-10-03パナソニックIpマネジメント株式会社Noise reduction device
    RU2680735C1 (en)*2018-10-152019-02-26Акционерное общество "Концерн "Созвездие"Method of separation of speech and pauses by analysis of the values of phases of frequency components of noise and signal
    CN109643554B (en)*2018-11-282023-07-21深圳市汇顶科技股份有限公司Adaptive voice enhancement method and electronic equipment
    US11438452B1 (en)2019-08-092022-09-06Apple Inc.Propagating context information in a privacy preserving manner
    US11501758B2 (en)2019-09-272022-11-15Apple Inc.Environment aware voice-assistant devices, and related systems and methods
    CN119107930A (en)2019-09-272024-12-10苹果公司 Environmentally aware voice assistant device and related systems and methods
    CN111370033B (en)*2020-03-132023-09-22北京字节跳动网络技术有限公司Keyboard sound processing method and device, terminal equipment and storage medium
    CN111402916B (en)*2020-03-242023-08-04青岛罗博智慧教育技术有限公司Voice enhancement system, method and handwriting board
    CN114093391B (en)*2020-07-292025-02-25华为技术有限公司 A method and device for filtering abnormal signals
    CN111916106B (en)*2020-08-172021-06-15牡丹江医学院 A Method to Improve Pronunciation Quality in English Teaching
    CN112927715B (en)*2021-02-262024-06-14腾讯音乐娱乐科技(深圳)有限公司Audio processing method, equipment and computer readable storage medium
    CN114550740B (en)*2022-04-262022-07-15天津市北海通信技术有限公司Voice definition algorithm under noise and train audio playing method and system thereof
    CN116092526A (en)*2023-01-102023-05-09浙江大华技术股份有限公司Signal detection method and device
    CN118411998B (en)*2024-07-022024-09-24杭州知聊信息技术有限公司Audio noise processing method and system based on big data

    Family Cites Families (13)

    * Cited by examiner, † Cited by third party
    Publication numberPriority datePublication dateAssigneeTitle
    US4461025A (en)*1982-06-221984-07-17Audiological Engineering CorporationAutomatic background noise suppressor
    US4630305A (en)*1985-07-011986-12-16Motorola, Inc.Automatic gain selector for a noise suppression system
    US4811404A (en)*1987-10-011989-03-07Motorola, Inc.Noise suppression system
    DE4012349A1 (en)*1989-04-191990-10-25Ricoh KkNoise elimination device for speech recognition system - uses spectral subtraction of sampled noise values from sampled speech values
    RU2010439C1 (en)*1991-07-041994-03-30Войсковая Часть 25871Device for transmission of data over vocoder path
    JP3065739B2 (en)*1991-10-142000-07-17三菱電機株式会社 Voice section detection device
    US5412735A (en)*1992-02-271995-05-02Central Institute For The DeafAdaptive noise reduction circuit for a sound reproduction system
    JPH05259928A (en)*1992-03-091993-10-08Oki Electric Ind Co LtdMethod and device for canceling adaptive control noise
    US5251263A (en)*1992-05-221993-10-05Andrea Electronics CorporationAdaptive noise cancellation and speech enhancement system and apparatus therefor
    JPH0695693A (en)*1992-09-091994-04-08Fujitsu Ten LtdNoise reducing circuit for voice recognition device
    JP3270866B2 (en)*1993-03-232002-04-02ソニー株式会社 Noise removal method and noise removal device
    US5485522A (en)*1993-09-291996-01-16Ericsson Ge Mobile Communications, Inc.System for adaptively reducing noise in speech signals
    US5657422A (en)*1994-01-281997-08-12Lucent Technologies Inc.Voice activity detection driven noise remediator

    Cited By (3)

    * Cited by examiner, † Cited by third party
    Publication numberPriority datePublication dateAssigneeTitle
    CN102128976A (en)*2011-01-072011-07-20钜泉光电科技(上海)股份有限公司Energy pulse output method and device of electric energy meter and electric energy meter
    CN102128976B (en)*2011-01-072013-05-15钜泉光电科技(上海)股份有限公司Energy pulse output method and device of electric energy meter and electric energy meter
    WO2021179045A1 (en)*2020-03-132021-09-16University Of South AustraliaA data processing method

    Also Published As

    Publication numberPublication date
    AU724111B2 (en)2000-09-14
    KR100423029B1 (en)2004-07-01
    AU7078496A (en)1997-04-01
    KR19990044659A (en)1999-06-25
    EE9800068A (en)1998-08-17
    WO1997010586A1 (en)1997-03-20
    PL325532A1 (en)1998-08-03
    DE69613380D1 (en)2001-07-19
    MX9801857A (en)1998-11-29
    RU2163032C2 (en)2001-02-10
    CN1121684C (en)2003-09-17
    EE03456B1 (en)2001-06-15
    TR199800475T1 (en)1998-06-22
    NO981074D0 (en)1998-03-11
    PL185513B1 (en)2003-05-30
    CN1201547A (en)1998-12-09
    EP0852052A1 (en)1998-07-08
    JPH11514453A (en)1999-12-07
    CA2231107A1 (en)1997-03-20
    NO981074L (en)1998-05-13
    BR9610290A (en)1999-03-16

    Similar Documents

    PublicationPublication DateTitle
    EP0852052B1 (en)System for adaptively filtering audio signals to enhance speech intelligibility in noisy environmental conditions
    EP0645756B1 (en)System for adaptively reducing noise in speech signals
    EP1017042B1 (en)Voice activity detection driven noise remediator
    EP0786760B1 (en)Speech coding
    KR100667008B1 (en) Complex Signal Activity Detection for Speech / Noise Classification of Improved Audio Signals
    US5544250A (en)Noise suppression system and method therefor
    US8977556B2 (en)Voice detector and a method for suppressing sub-bands in a voice detector
    EP0784311B1 (en)Method and device for voice activity detection and a communication device
    EP0599664B1 (en)Voice encoder and method of voice encoding
    EP1232496A1 (en)Noise suppression
    EP0699334A1 (en)Method and apparatus for group encoding signals
    WO2000075919A1 (en)Methods and apparatus for generating comfort noise using parametric noise model statistics
    US5666429A (en)Energy estimator and method therefor
    US7889874B1 (en)Noise suppressor
    US5710862A (en)Method and apparatus for reducing an undesirable characteristic of a spectral estimate of a noise signal between occurrences of voice signals
    JP2002169599A (en) Noise suppression method and electronic device
    WO2000007177A1 (en)Communication terminal
    JP2002076960A (en) Noise suppression method and mobile phone
    WO2001041334A1 (en)Method and apparatus for suppressing acoustic background noise in a communication system

    Legal Events

    DateCodeTitleDescription
    PUAIPublic reference made under article 153(3) epc to a published international application that has entered the european phase

    Free format text:ORIGINAL CODE: 0009012

    17PRequest for examination filed

    Effective date:19980210

    AKDesignated contracting states

    Kind code of ref document:A1

    Designated state(s):BE DE DK ES FI FR GB GR IT NL PT SE

    17QFirst examination report despatched

    Effective date:19990517

    RIC1Information provided on ipc code assigned before grant

    Free format text:7G 10L 21/02 A

    GRAGDespatch of communication of intention to grant

    Free format text:ORIGINAL CODE: EPIDOS AGRA

    GRAGDespatch of communication of intention to grant

    Free format text:ORIGINAL CODE: EPIDOS AGRA

    GRAHDespatch of communication of intention to grant a patent

    Free format text:ORIGINAL CODE: EPIDOS IGRA

    GRAHDespatch of communication of intention to grant a patent

    Free format text:ORIGINAL CODE: EPIDOS IGRA

    GRAA(expected) grant

    Free format text:ORIGINAL CODE: 0009210

    AKDesignated contracting states

    Kind code of ref document:B1

    Designated state(s):BE DE DK ES FI FR GB GR IT NL PT SE

    PG25Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code:FR

    Free format text:LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

    Effective date:20010613

    Ref country code:FI

    Free format text:LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

    Effective date:20010613

    Ref country code:BE

    Free format text:LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

    Effective date:20010613

    REFCorresponds to:

    Ref document number:69613380

    Country of ref document:DE

    Date of ref document:20010719

    ITFIt: translation for a ep patent filed
    PG25Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code:SE

    Free format text:LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

    Effective date:20010913

    Ref country code:GB

    Free format text:LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date:20010913

    Ref country code:DK

    Free format text:LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

    Effective date:20010913

    PG25Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code:GR

    Free format text:LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

    Effective date:20010914

    Ref country code:DE

    Free format text:LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

    Effective date:20010914

    PG25Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code:PT

    Free format text:LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

    Effective date:20010917

    PG25Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code:ES

    Free format text:LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

    Effective date:20011220

    ENFr: translation not filed
    PLBENo opposition filed within time limit

    Free format text:ORIGINAL CODE: 0009261

    STAAInformation on the status of an ep patent application or granted ep patent

    Free format text:STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

    GBPCGb: european patent ceased through non-payment of renewal fee

    Effective date:20010913

    26NNo opposition filed
    PGFPAnnual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code:NL

    Payment date:20051229

    Year of fee payment:11

    PGFPAnnual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code:IT

    Payment date:20070926

    Year of fee payment:12

    PG25Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code:NL

    Free format text:LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date:20080401

    NLV4Nl: lapsed or anulled due to non-payment of the annual fee

    Effective date:20080401

    PG25Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code:IT

    Free format text:LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date:20080913


    [8]ページ先頭

    ©2009-2025 Movatter.jp