Movatterモバイル変換


[0]ホーム

URL:


CN106328150A - Bowel sound detection method, device and system under noisy environment - Google Patents

Bowel sound detection method, device and system under noisy environment
Download PDF

Info

Publication number
CN106328150A
CN106328150ACN201610686377.6ACN201610686377ACN106328150ACN 106328150 ACN106328150 ACN 106328150ACN 201610686377 ACN201610686377 ACN 201610686377ACN 106328150 ACN106328150 ACN 106328150A
Authority
CN
China
Prior art keywords
signal
borborygmus
label
sample
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610686377.6A
Other languages
Chinese (zh)
Other versions
CN106328150B (en
Inventor
战鸽
朱斌杰
陈平
应冬文
颜永红
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BEIJING YIMAI MEDICAL TECHNOLOGY CO., LTD.
Shandong Yi Mai Medical Technology Co., Ltd.
Original Assignee
Beijing Yimai Medical Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Yimai Medical Technology Co LtdfiledCriticalBeijing Yimai Medical Technology Co Ltd
Priority to CN201610686377.6ApriorityCriticalpatent/CN106328150B/en
Publication of CN106328150ApublicationCriticalpatent/CN106328150A/en
Application grantedgrantedCritical
Publication of CN106328150BpublicationCriticalpatent/CN106328150B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Classifications

Landscapes

Abstract

The invention discloses bowel sound detection method, device and system under noisy environment. The bowel sound detection method comprises steps of collecting a bowl sound mixed signal of a current user through a sensor, wherein the bowl sound mixed signal comprises a bowl sound signal and an environment interference signal; converting the bowl sound mixed signal to a digital signal; extracting a time-frequency spectrum characteristic of the digital signal; inputting the time-frequency spectrum characteristic of the digital signal into a trained convolution nerve network to process; and detecting a time point when the bowl sound happens so as to distinguish the bowl sound signal and the environment interference signal. The bowel sound detection method, device and system under the noisy environment utilize a difference of the time-frequency spectrum characteristics of the bowl sound signal and the environment interference signal, train a convolution nerve network to distinguish the bowl sound and the interference sound, can finish the detection on the bowl sound under the noisy environment, and can help to improve the accuracy of the bowl sound detection.

Description

Borborygmus detection method under noisy environment, Apparatus and system
Technical field
The present invention relates to processing of biomedical signals technical field, in particular to the borborygmus under a kind of noisy environmentSound detection method, Apparatus and system.
Background technology
Intestinal is pipeline the longest in digestive organs, and major function is digestion and assimilates food.Once intestinal has exception, justDisorder of Digestion and Absorption, and a series of related symptoms may be caused.Therefore, the diagnosis of intestinal tract disease is the most necessary.At intestinalIn the diagnosis and treatment process of disease, borborygmus auscultation belongs to conventional noninvasive test.
In the diagnosis and treatment process of present stage, the collection of borborygmus and distinguish the artificial auscultation depending on doctor.BorborygmusGeneration derive from the motion of intestinal, its appearance in time range has openness, thus the auscultation process to borborygmusIt is generally required to longer time and relatively quiet environment.The method using sensor acquisition signal combination computer-assisted analysisCan be the collection of borborygmus and offer sound assurance is provided, but, acoustical signal noisy in existing environment and patient's bodyThe vibration signal of intracavity often has a similar wave characteristic to borborygmus signal, easily to the collection of borborygmus with distinguish formationSerious interference, causes the accuracy identifying borborygmus under noisy environment poor.
Summary of the invention
In view of this, the purpose of the embodiment of the present invention is to provide under a kind of noisy environment borborygmus detection method, dressPut and system, it is possible to be lifted in noisy environment the accuracy identifying borborygmus signal from collection borborygmus mixed signal.
First aspect, embodiments provides the borborygmus detection method under a kind of noisy environment, including:
By the borborygmus mixed signal of sensor acquisition active user, wherein, described borborygmus mixed signal includes intestinalRing tone signal and environmental disturbances signal;
Described borborygmus mixed signal is converted to digital signal;
Extract the time-frequency feature of described digital signal;
The time-frequency feature of described digital signal is inputted in the convolutional neural networks trained and processes, detect intestinalThe time point that ring sound occurs, thus distinguish described borborygmus signal and described environmental disturbances signal;Wherein, described convolutional Neural netThe training process of network includes:
Borborygmus sample signal and least one set interference sample signal is gathered respectively by described sensor;
Sample signal is disturbed all to be converted into numeral sample signal described borborygmus sample signal and least one set;
Extract the time-frequency feature of described numeral sample signal;
In time-frequency domain, described numeral sample signal is made signal label;Described signal label includes for labelling intestinalRing tone signal borborygmus label of time of occurrence point in described borborygmus sample signal, and for labelling interference signal describedThe interference label of time of occurrence point in interference sample signal;
From described numeral sample signal, borborygmus label letter is extracted according to described borborygmus label and described interference labelNumber and each group interference label signal as training sample;
Using the time-frequency feature of numeral sample signal corresponding for described training sample as training data, by described borborygmusLabel and described interference label are as supervision message, and training is for distinguishing borborygmus signal and the convolutional Neural of various interference signalNetwork.
In conjunction with first aspect, embodiments provide the first possible embodiment of first aspect, wherein, carryThe time-frequency feature taking described numeral sample signal includes:
The described numeral sample signal with sequential is carried out framing and windowing;
Described numeral sample signal after windowing is carried out fast Fourier transform, extracts power spectrum;
Use Gammatone bank of filters that described power spectrum is filtered;Described Gammatone bank of filters is accomplished that onePlanting linear transformation, its impulse respective table is shown as:
gi(t)=Atn-1exp(-2πbit)cos(2πfii),t≥0,1≤i≤N,
Wherein, A represents the constant of regulation ratio, and n represents filter order, biRepresent the rate of decay, fiExpression center frequencyRate, φiRepresenting phase place, N represents number of filter;For i-th wave filter, there is bi=1.019ERB (fi), wherein equivalent rectangularBandwidth ERB (fi) expression formula be
ERB(fi)=24.7×(4.37×fi1000+1),
The coefficient matrix of the filtered described power spectrum through Gammatone bank of filters is carried out discrete cosine changeChange, obtain Gammatone cepstrum coefficient;
Using described Gammatone cepstrum coefficient as the time-frequency feature of described numeral sample signal.
In conjunction with the first possible embodiment of first aspect, embodiments providing first aspect the second canThe embodiment of energy, wherein, makes signal label to described numeral sample signal in time-frequency domain and includes:
In time-frequency domain, the described numeral sample signal that each time point is corresponding is judged;Wherein, each time pointCorresponding described numeral sample signal is the signal frame after described framing and windowing;
When the described signal frame of current point in time has borborygmus signal, borborygmus label is set for described signal frame;When the described signal frame of current point in time has interference signal, interference label is set for described signal frame;Wherein, described borborygmusPhonetic symbol label and described interference label multi-C vector represent;
From described numeral sample signal, borborygmus label letter is extracted according to described borborygmus label and described interference labelNumber and each group interference label signal include as training sample: according to arranging described borborygmus label and the institute of described interference labelState signal frame order, from described numeral sample signal, extract borborygmus label signal and each group of interference label signal as trainingSample.
In conjunction with the embodiment that the second of first aspect or first aspect is possible, embodiments provide first partyThe third possible embodiment in face, wherein: described convolutional neural networks includes input layer, multiple hidden layer, full linking layerAnd output layer.Described hidden layer and full linking layer all contain the parameter of self, and described parameter includes weights and biasing;
Described convolutional neural networks training process use gradient descent method, detailed process includes:
Convolutional neural networks is carried out random initializtion;
Start training, order random-ising by described training sample and described signal label, take out the most at randomTake J training sample and form a sample set as input sample, extract the signal label group corresponding with described input sampleBecome a sub-set of tags, complete being trained on described sample set all inputs sample and one take turns training, complete in instituteBeing trained on sample set is had once to train;
During taking turns training one, in described sample set, all of input sample all carries out propagated forward, passes throughAfter the effect of described convolutional neural networks, the output layer at convolutional neural networks compares with corresponding signal label, calculatesDifference square as square error between output result and corresponding signal label;Obtain the output result of all input samplesSquare error with signal label;
During taking turns training one, utilize described square error to carry out back propagation and parameter updates, including: from describedOutput layer starts, and reversely sequentially passes through each layer, obtains the equivalent error on each layer;Utilize the equivalent error meter on each layerCalculate the gradient of parameter on place layer, utilize the parameter of gradient updating place layer on each layer;
During once training, complete last when taking turns training, calculate the mean error of all described square errors,Described mean error is utilized to judge whether described convolutional neural networks restrains;The stable threshold set is tended at described mean errorTime, determine that described convolutional neural networks reaches convergence, if described convolutional neural networks reaches to restrain, deconditioning;Otherwise openBegin newly once to train, during until the number of times of training or duration reach to set threshold value, deconditioning;
After training stops, using current convolutional neural networks as the convolutional neural networks trained.
In conjunction with the third possible embodiment of first aspect, embodiments provide the 4th kind of first aspectPossible embodiment, wherein, the detailed process that described input sample carries out propagated forward includes:
Described input sample is done computing by the input layer of described convolutional neural networks, and each layer of convolutional neural networks is to upperComputing is done in the output of one layer;
In convolutional neural networks, l layer is output as
xl=f (ul) formula (1)
Wherein f () is activation primitive, ul=Wlxl-1+bl, xl-1It is the output of l-1 layer, the input of l layer, WlAnd blIt is weights and the biasing of l layer respectively;Activation primitive uses sigmoid function or hyperbolic tangent function;
The square error calculated between output result and corresponding signal label includes: input sample, meter for eachCalculating the square error between output result and the corresponding signal label obtained by the output layer of convolutional neural networks, jth is defeatedThe squared error function entering sample is
Wherein, K represents described output result and the dimension of signal label,Represent that jth sample is through convolutional Neural netThe kth dimension of the output result after network,Represent the kth dimension of the signal label that jth sample is corresponding.
In conjunction with the 4th kind of possible embodiment of first aspect, embodiments provide the 5th kind of first aspectPossible embodiment, wherein, described back propagation and parameter update and specifically include:
By described output result with the square error of signal label from the beginning of described output layer, be transferred to convolution the most successivelyEach layer in neutral net, obtains the equivalent error on each layer;Described equivalent error is that square error is to place layer parameterError rate, computing formula is
Wherein, E is the square error of output result, and b is the parameter of convolutional neural networks;
Equivalent error on output layer is
L in formula represents output layer, operative symbolRepresent element multiplication one by one;yLFor the output result of output layer, tLForThe signal label of output layer;
Equivalent error on other layers is
Utilize the equivalent error δ on each layerl, calculate the gradient of parameter on the layer of place, obtain the gradient of weights and biasingIt is respectively as follows:
η is learning rate, arranges different learning rates for different parameters;
Utilize the parameter of the gradient updating place layer of parameter on each layer;Plus place layer in the original parameter of each layerThe gradient of parameter obtains new parameter.
In conjunction with the 4th kind of possible embodiment of first aspect, embodiments provide the 5th kind of first aspectPossible embodiment, wherein, during once training, complete last when taking turns training, calculate all described square mistakesThe mean error of difference, described mean error function is:
Wherein, the number of samples during J represents once training;
At described mean error EJWhen tending to the stable threshold set, determine that described convolutional neural networks reaches to receiveHold back;
If described convolutional neural networks reaches to restrain, deconditioning;Otherwise start newly once to train, update convolution godThrough the parameter of network, gradually minimize EJ, the output result making described convolutional neural networks is close with corresponding signal label.
Second aspect, the embodiment of the present invention also provides for the borborygmus sound detection device under a kind of noisy environment, including:
Convolutional neural networks training module, for the training of convolutional neural networks, concrete training process includes: by sensingDevice gathers borborygmus sample signal and least one set interference sample signal respectively;By described borborygmus sample signal and least one setInterference sample signal is all converted into numeral sample signal;Extract the time-frequency feature of described numeral sample signal;In time-frequency domain,Described numeral sample signal is made signal label;Described signal label includes for labelling borborygmus signal at described borborygmusThe borborygmus label of time of occurrence point in sample signal, and when labelling interference signal occurs in described interference sample signalBetween point interference label;From described numeral sample signal, borborygmus is extracted according to described borborygmus label and described interference labelLabel signal and each group of interference label signal are as training sample;Time-frequency by numeral sample signal corresponding for described training sampleDescribed borborygmus label and described interference label, as training data, are used for distinguishing intestinal by spectrum signature as supervision message, trainingRing tone signal and the convolutional neural networks of various interference signal;
Signal acquisition module, for by the borborygmus mixed signal of sensor acquisition active user, wherein, described borborygmusMixture of tones signal includes borborygmus signal and environmental disturbances signal;
Signal conversion module, for described borborygmus mixed signal is converted to digital signal, and extracts described numeral letterNumber time-frequency feature;
Borborygmus detection module, for inputting the training of described convolutional neural networks by the time-frequency feature of described digital signalThe described convolutional neural networks that module trains processes, detects the time point that borborygmus occurs, thus distinguish describedBorborygmus signal and described environmental disturbances signal.
In conjunction with second aspect, embodiments provide the first possible embodiment of second aspect, wherein, instituteState convolutional neural networks training module to include:
Signal windowing unit, for carrying out framing and windowing to the described numeral sample signal with sequential;
Fourier transform unit, for the described numeral sample signal after windowing is carried out fast Fourier transform, extractsPower spectrum;
Gammatone bank of filters, is used for realizing a kind of linear transformation, filters described power spectrum;Described GammatoneThe impulse respective table of bank of filters is shown as:
gi(t)=Atn-1exp(-2πbit)cos(2πfii),t≥0,1≤i≤N,
Wherein, A represents the constant of regulation ratio, and n represents filter order, biRepresent the rate of decay, fiExpression center frequencyRate, φiRepresenting phase place, N represents number of filter;For i-th wave filter, there is bi=1.019ERB (fi), wherein equivalent rectangularBandwidth ERB (fi) expression formula be
ERB(fi)=24.7×(4.37×fi1000+1);
Discrete cosine transform unit, for the filtered described power spectrum through Gammatone bank of filters beingMatrix number carries out discrete cosine transform, obtains Gammatone cepstrum coefficient;Using described Gammatone cepstrum coefficient as describedThe time-frequency feature of numeral sample signal.
The third aspect, the embodiment of the present invention also provides for the borborygmus detecting system under a kind of noisy environment, including second partyThe borborygmus sound detection device of face offer and sensor;
Described sensor is for gathering borborygmus sample signal and least one set interference sample in neural network training processThis signal;Gathering the borborygmus mixed signal of active user during carrying out borborygmus detection, wherein, described borborygmus mixesSignal includes borborygmus signal and environmental disturbances signal;And the signal gathered is sent to described borborygmus sound detection device.
Borborygmus detection method under the noisy environment that the embodiment of the present invention is provided, Apparatus and system, utilize borborygmusThe difference of performance in time and frequency domain characteristics of signal and environmental disturbances signal, train one convolutional neural networks distinguish borborygmus withInterference tones, can complete the detection to borborygmus in noisy environment, contributes to promoting the accuracy of borborygmus detection.
For making the above-mentioned purpose of the present invention, feature and advantage to become apparent, preferred embodiment cited below particularly, and coordinateAppended accompanying drawing, is described in detail below.
Accompanying drawing explanation
In order to be illustrated more clearly that the technical scheme of the embodiment of the present invention, below by embodiment required use attachedFigure is briefly described, it will be appreciated that the following drawings illustrate only certain embodiments of the present invention, and it is right to be therefore not construed asThe restriction of scope, for those of ordinary skill in the art, on the premise of not paying creative work, it is also possible to according to thisA little accompanying drawings obtain other relevant accompanying drawings.
Fig. 1 shows the flow chart of the borborygmus detection method under a kind of noisy environment that the embodiment of the present invention is provided;
Fig. 2 shows in the borborygmus detection method that the embodiment of the present invention is provided, training convolutional neural networks concreteThe flow chart of method;
Fig. 3 shows in the borborygmus detection method that the embodiment of the present invention is provided, the training process of convolutional neural networksFlow chart;
Fig. 4 shows in the borborygmus detection method that the embodiment of the present invention is provided, the structural representation of convolutional neural networksFigure;
Fig. 5 shows the structural representation of the borborygmus sound detection device under a kind of noisy environment that the embodiment of the present invention is providedFigure;
Fig. 6 shows the structural representation of the borborygmus detecting system under a kind of noisy environment that the embodiment of the present invention is providedFigure.
Detailed description of the invention
For making the purpose of the embodiment of the present invention, technical scheme and advantage clearer, below in conjunction with the embodiment of the present inventionMiddle accompanying drawing, is clearly and completely described the technical scheme in the embodiment of the present invention, it is clear that described embodiment is onlyIt is a part of embodiment of the present invention rather than whole embodiments.Generally real with the present invention illustrated described in accompanying drawing hereinThe assembly executing example can be arranged with various different configurations and design.Therefore, below to the present invention's provided in the accompanying drawingsThe detailed description of embodiment is not intended to limit the scope of claimed invention, but is merely representative of the selected reality of the present inventionExecute example.Based on embodiments of the invention, the institute that those skilled in the art are obtained on the premise of not making creative workThere are other embodiments, broadly fall into the scope of protection of the invention.
At present, the collection of borborygmus and distinguish the artificial auscultation depending on doctor, the auscultation process one to borborygmusAs need longer time and relatively quiet environment.The method using sensor acquisition signal combination computer-assisted analysis canThink that diagnosis and treatment process provides sound assurance, but the interference signal such as the noisy speech in existing environment has with borborygmus signalSimilar wave characteristic, easily to the collection of borborygmus with distinguish and form serious interference.For this problem, at noisy environmentUnder borborygmus detection algorithm have great importance.
Convolutional neural networks is applicable to image, Speech processing, is usually used to structure complexity, nonlinearityDependency relation.Being mainly characterized by it and can go out there is distinctive by there being the training procedure extraction of supervision of convolutional neural networksThe feature combining form of matter, various feature combining form is stored in multi-level convolution kernel.Further, roll up in multilamellarUnder the common effect of lamination, it is possible to achieve to the most abstract of feature and combination, so that the expression energy of multi-level convolution kernelPower is abundanter, extends the suitability of the content that convolutional neural networks is acquired on the basis of finite sample.
Utilize the difference of performance in time and frequency domain characteristics of borborygmus signal and noise signal, train a convolutional Neural netNetwork distinguishes borborygmus and noise, can complete the detection to borborygmus under noise jamming.Convolutional neural networks can be spontaneousExtracting characteristically for distinguishing the key message of borborygmus and voice, the learning outcome form of expression in time-frequency domain is with specialThe form of expression levied is close, it is simple to analyzes and adjusts.
Based on this, the invention provides the borborygmus detection method under a kind of noisy environment, Apparatus and system, can filterNoise in environment, apparent extraction and identification borborygmus, complete the detection to borborygmus in noisy environment.
For ease of the present embodiment is understood, first to the intestinal under a kind of noisy environment disclosed in the embodiment of the present inventionRing sound detection method describes in detail, and Fig. 1 shows the borborygmus under a kind of noisy environment that the embodiment of the present invention is providedThe flow chart of detection method.As it is shown in figure 1, this detection method includes:
Step S101, by the borborygmus mixed signal of sensor acquisition active user, wherein, borborygmus mixed signal bagInclude borborygmus signal and environmental disturbances signal;
Step S102, is converted to digital signal by above-mentioned borborygmus mixed signal;
Step S103, extracts the time-frequency feature of above-mentioned digital signal;
Step S104, inputs the time-frequency feature of digital signal in the convolutional neural networks trained and processes, inspectionMeasure the time point that borborygmus occurs, thus distinguish borborygmus signal and environmental disturbances signal.
Wherein, the concrete grammar of training convolutional neural networks is as in figure 2 it is shown, comprise the steps.
Step S201, gathers borborygmus sample signal and least one set interference sample signal respectively by sensor.
Step S202, disturbs sample signal to be all converted into numeral sample signal borborygmus sample signal and least one set.
Step S203, extracts the time-frequency feature of numeral sample signal.The time-frequency feature that the embodiment of the present invention uses isGammatone cepstrum coefficient, the concrete steps of the Gammatone cepstrum coefficient extracting numeral sample signal include: to when havingThe numeral sample signal of sequence carries out framing and windowing;To the numeral sample signal elder generation zero padding of each frame to N point, N=2i, i is wholeNumber, and i >=8;Then, the numeral sample signal of each frame being carried out windowing or preemphasis processes, windowed function uses Hamming windowOr breathe out peaceful window (hanning) (hamming).
Numeral sample signal after windowing is carried out fast Fourier transform, extracts power spectrum;
Use Gammatone bank of filters that power spectrum is filtered;Described Gammatone bank of filters is accomplished that a kind of lineProperty conversion, its impulse respective table is shown as:
gi(t)=Atn-1exp(-2πbit)cos(2πfii),t≥0,1≤i≤N,
Wherein, A represents the constant of regulation ratio, and n represents filter order, biRepresent the rate of decay, fiExpression center frequencyRate, φiRepresenting phase place, N represents number of filter;For i-th wave filter, there is bi=1.019ERB (fi), wherein equivalent rectangularBandwidth ERB (fi) expression formula be
ERB(fi)=24.7×(4.37×fi1000+1),
The coefficient matrix of the filtered power spectrum through Gammatone bank of filters is carried out discrete cosine transform,To Gammatone cepstrum coefficient;Gammatone cepstrum coefficient combines the auditory properties of human ear, is a kind of audition filtering characteristics,The resolution of low frequency is high, and the resolution of high frequency is suitably compressed.
It should be noted that through above-mentioned steps, borborygmus sample signal c (t) correspondence can be respectively obtainedGammatone cepstrum coefficient S (j) that Gammatone cepstrum coefficient C (j) is corresponding with speech samples signal, both cepstrum coefficientsThe training of convolutional neural networks will be used for as training data.In like manner, the borborygmus collected under noisy environment to be detected mixesClose the Gammatone cepstrum coefficient that signal is obtained by above-mentioned steps, the inspection of borborygmus time of occurrence can be used for as featureSurvey.
Step S204, in time-frequency domain, makes signal label to numeral sample signal;Signal label includes for labelling intestinalRing tone signal borborygmus label of time of occurrence point in borborygmus sample signal, and for labelling interference signal at interference sampleThe interference label of time of occurrence point in signal.The detailed process of this step includes:
In time-frequency domain, the numeral sample signal that each time point is corresponding is judged;Wherein, each time point is correspondingNumeral sample signal be the signal frame after framing and windowing;
When the signal frame of current point in time has borborygmus signal, borborygmus label is set for signal frame;When currentBetween point signal frame in have interference signal time, for signal frame, interference label is set.
Wherein, borborygmus label and interference label multi-C vector represent.If only one group interference signal, such as interference letterNumber being voice signal, signal label can use bivector to represent, has [1,0] to be carved with borborygmus when representing this certain moment tSound occurs, [0,1] is carved with voice and is occurred when representing this.Note time index t here be no longer concrete sampled signal timeBetween index, but the time sequencing index of the Gammatone cepstrum coefficient obtained through step S203, i.e. Gammatone cepstrumT frame coefficient time point in time sequencing in coefficient.If having many group interference signals, need to solve many classification problems,Label vector dimension can be increased, keep the value result of element in vector and the corresponding relation of classification results.
Step S205, according to borborygmus label and interference label extract from numeral sample signal borborygmus label signal withEach group interference label signal is as training sample.Detailed process includes: according to arranging borborygmus label and the signal of interference labelFrame sequential, extracts borborygmus label signal and each group of interference label signal as training sample from numeral sample signal.TrainingThe form of sample is continuous d frame Gammatone cepstrum coefficient matrix, and a frame at this matrix center is with borborygmus or voiceCepstrum coefficient.After all training samples have extracted, the Gammatone cepstrum coefficient not being extracted is considered not comprise borborygmusAnd voice, the most it is not used for training convolutional neural networks.Sample one the training sample set of composition being extracted, internalPut in order and only represent the order being extracted, the most corresponding concrete time time point.Corresponding, the mark of successive frame orderThe label that only marked borborygmus and voice appearance in label is extracted, and forms tag set, and remaining label is not used.ByThis, be available for two class sample of signal of training convolutional neural networks and corresponding label.In like manner, adopt under noisy environmentThe borborygmus mixed signal that collection arrives, it is possible to sample drawn set.
Step S206, using the time-frequency feature of numeral sample signal corresponding for training sample as training data, by borborygmusPhonetic symbol label and interference label are as supervision message, and training is for distinguishing borborygmus signal and the convolutional Neural net of various interference signalNetwork.
The structure of described convolutional neural networks as shown in Figure 4, including input layer, multiple hidden layer, full linking layer and defeatedGo out layer.Hidden layer and full linking layer all contain the parameter of self, and described parameter includes weights and biasing.Convolutional neural networks hiddenHide layer comprise alternately arranged two convolutional layer and two down-sampling layers, convolutional layer and down-sampling layer all comprise self weights andBiasing.Convolutional layer, by the convolutional calculation of convolution kernel with input, obtains an output from the block of input every time, passes through convolution kernelTraversal in input obtains complete output.Described convolution kernel is the weights of convolutional layer.Down-sampling layer is by designingProportionality coefficient, input is compressed.
Convolutional neural networks training process as it is shown on figure 3, detailed process includes:
Step S2061, carries out random initializtion to convolutional neural networks;Except weights and the biasing of convolutional neural networks needOutside initializing, it is even more important that need the quantity of the setting network degree of depth and convolution kernel.The present embodiment uses typical caseConfiguration, along with raising and the increase of training sample of learning tasks complexity, can suitably increase the degree of depth and the convolution of networkThe quantity of core.Meanwhile, the specification of convolution kernel is also important influence factor, it is proposed that the limit of convolution kernel in design ground floor convolutional layerLength is general more than sample time span, is so conducive to the feature representation form of convolutional neural networks acquistion global sense;
Step S2062, starts training, order random-ising by training sample and signal label, the most randomExtract J training sample and form a sample set as input sample, extract the signal label composition corresponding with inputting sampleOne sub-set of tags, completes being trained on sample set all inputs sample and one takes turns training, complete at all samplesBeing trained in subset is once trained;
Step S2063, during taking turns training one, in sample set, all of input sample all carries out forward direction biographyBroadcasting, after the effect of convolutional neural networks, the output layer at convolutional neural networks compares with corresponding signal label, meterCalculate difference between output result and corresponding signal label square as square error;Obtain the output knot of all input samplesFruit and the square error of signal label;
Square error cost function is defined as
EJ=12Σj=1JΣk=1K(ykj-tkj)2,
Number of samples during wherein J represents once training, K represents the dimension of output and label,Represent jth sample warpCross the kth dimension of the output of convolutional neural networks,Represent the kth dimension of the label that jth sample is corresponding.The target of training is intended to moreThe parameter of new network so that network output and label closer to, namely minimize EJ.During for one of them sample, thenThe error function of jth sample is
Ej=12Σk=1K(ykj-tkj)2.
In definition neutral net, l layer is output as
xl=f (ul) wherein ul=Wlxl-1+bl,
Here f () is activation primitive, xl-1It is the output of l-1 layer, the namely input of l layer, WlAnd blIt is respectivelyThe weights of l layer and biasing.Activation primitive can have a variety of, usually sigmoid function or hyperbolic tangent function,Sigmoid function is by output squeezing to [0,1], and hyperbolic tangent function is by output squeezing to [-1,1].By training data normalizingCancellation average and variance are the distribution form of 1, can strengthen convergence during stochastic gradient descent.The most permissibleRealizing propagated forward, the output of last layer is done computing by each layer, obtains exporting result, sample through nonlinear activation primitiveInformation is successively transmitted, and last output result is i.e. to inputting the predictive value that sample is borborygmus or voice.
Step S2064, during taking turns training one, utilizes square error to carry out back propagation and parameter updates, including:From the beginning of output layer, reversely sequentially pass through each layer, obtain the equivalent error on each layer;Utilize the equivalent error on each layerCalculate the gradient of parameter on place layer, utilize the parameter of gradient updating place layer on each layer;
Back propagation and parameter renewal process include:
The rate of change of neural network parameter is defined as by error
∂E∂b=∂E∂u∂u∂b=∂E∂u·1=δ,
Then the back propagation on output layer is
L layer i.e. output layer, operative symbol thereinRepresenting element multiplication one by one, the back propagation on other layers is
By the error rate δ on each layerlThe gradient of each weights and biasing can be obtained
ΔWl=-η∂E∂Wl=-ηxl-1(δl)T,
Δbl=-η∂E∂bl=-ηδl,
η therein is learning rate, and the parameter that can be different arranges different learning rates, utilizes gradient descent method to update ginsengDuring number, the gradient of parameter is added in original parameter and obtains new parameter.
Output at convolutional layer be multiple input convolution combination result, be represented by
xjl=f(Σi∈Mjxil-1*kijl+bjl),
WhereinRepresent the jth dimension output on l layer, MjRepresent input set,Represent in input set one concreteInput,Represent the weights contacting this input on l layer with jth dimension output,Represent corresponding biasing.Before and after convolutional layer withDown-sampling layer is connected, and back propagation and parameter on convolutional layer update the inverse process with down-sampling layer.In the embodiment of the present inventionDown-sampling layer weights useRepresenting, down-sampling factor n represents, down-sampling process will the block weighted average of n × n.By mistakeDifference rate of change is when down-sampling layer back propagation, it is only necessary to be once multiplied available with the weights participating in during propagated forward calculatingThe above error rate on a convolutional layer.According to aforementioned back propagation, can obtain the error rate on convolutional layer is
Up () therein represent up-sampling calculate, it is simply that by the object tools on a point to the block carrying out down-samplingIn the matrix that size is identical, this process is also referred to as Kronecker and amasss, and is represented by
up(x)≡x⊗1n×n,
N therein is exactly the factor during down-sampling calculates.Then, can the error rate of change to biasing on this convolutional layer
∂E∂bj=Σu,v(δjl)uv,
The block position of down-sampling, the error rate of change to convolution kernel is carried out when what wherein u, v represented is propagated forwardFor
∂E∂kijl=Σu,v(δjl)uv(pil-1)uv,
WhereinIt isIn with convolution kernelBlock by element multiplication.The thus obtained error rate of change to parameterSubstitute into the formula in back-propagation process and calculate the gradient of parameter, and then undated parameter.
It is output as on down-sampling layer
xjl=f(βjldown(xjl-1)+bjl),
Wherein down () represents that down-sampling calculates, by same in two dimensions for input under the control of down-sampling factor of nTime be compressed into original 1/n.When l+1 layer is convolutional layer, can be byMatrix according to the whole inverted arrangements of ranks order, withCarry out complete convolution algorithm, the result of complete convolution again withElement multiplication one by one, can obtainWhat is called is rolled up completelyLong-pending, it is convolution again after zero padding on boundary position, thus can obtain identical with down-sampling layer output sizePass throughThe error rate of change to parameter on down-sampling layer can be obtained
∂E∂bj=Σu,v(δjl)uv,
And then can be with undated parameter.
Step S2065, during once training, complete last when taking turns training, calculate the flat of all square errorsAll error, utilizes mean error to judge whether convolutional neural networks restrains;When mean error tends to the stable threshold set, reallyDetermine convolutional neural networks and reach convergence, if convolutional neural networks reaches to restrain, deconditioning;Otherwise return step S2602,Start newly once to train, during until the number of times of training or duration reach to set threshold value, deconditioning;
Selection for the condition of convergence is not unique, and the stable threshold of mean error can regard concrete application to be needed reallyFixed, it is also possible to the number of times trained by setting carrys out the time of controlled training neutral net.
Step S2066, after training stops, using current convolutional neural networks as the convolutional neural networks trained.
In other embodiments, it would however also be possible to employ other time and frequency domain characteristics such as amplitude spectrum, power spectrum etc., concrete processMethod belongs to common knowledge, does not repeats at this.
Corresponding with the borborygmus detection method under above-mentioned noisy environment, the embodiment of the present invention additionally provides a kind of noisy ringBorborygmus sound detection device under border.As it is shown in figure 5, this borborygmus sound detection device, including such as lower module:
Convolutional neural networks training module 501, for the training of convolutional neural networks, concrete training process is examined with borborygmusIn survey method, the training process of convolutional neural networks is identical, does not repeats them here.
Signal acquisition module 502, for by the borborygmus mixed signal of sensor acquisition active user;
Signal conversion module 503, for borborygmus mixed signal being converted to digital signal, and extract digital signal timeSpectrum signature;
Borborygmus detection module 504, for inputting convolutional neural networks training module by the time-frequency feature of digital signalThe convolutional neural networks trained processes, detects the time point that borborygmus occurs, thus distinguish borborygmus signal andEnvironmental disturbances signal.
Wherein, convolutional neural networks training module 501 includes:
Signal windowing unit, for carrying out framing and windowing to the numeral sample signal with sequential;
Fourier transform unit, for the numeral sample signal after windowing carries out fast Fourier transform, extracts powerSpectrum;
Gammatone bank of filters, is used for realizing a kind of linear transformation, filters power spectrum;Concrete methods of realizing is upperThe borborygmus detection method stated is it is stated that repeat no more.
Discrete cosine transform unit, for the coefficient square to the filtered power spectrum through Gammatone bank of filtersBattle array carries out discrete cosine transform, obtains Gammatone cepstrum coefficient.
Further embodiment of this invention additionally provides the borborygmus detecting system under a kind of noisy environment, shown in Figure 6, bagInclude the borborygmus sound detection device 62 in above-described embodiment and sensor 64.Sensor 64 is for adopting in neural network training processCollection borborygmus sample signal and least one set interference sample signal;The intestinal of active user is gathered during carrying out borborygmus detectionRing mixture of tones signal, wherein, borborygmus mixed signal includes borborygmus signal and environmental disturbances signal;And the signal of collection is sent outDeliver to borborygmus sound detection device.Wherein, the concrete structure of borborygmus sound detection device 62 can use the structure shown in Fig. 5.
Those skilled in the art is it can be understood that arrive, for convenience and simplicity of description, and the system of foregoing descriptionWith the specific works process of device, it is referred to the corresponding process in preceding method embodiment, does not repeats them here.
Borborygmus detection method under the noisy environment that the embodiment of the present invention is provided, Apparatus and system, it is adaptable at noiseDetection to borborygmus in heterocycle border, utilizes the difference of performance in time and frequency domain characteristics of borborygmus signal and environmental disturbances signalDifferent, borborygmus signal can be identified from multiple interference signal quickly and accurately.
The above, the only detailed description of the invention of the present invention, but protection scope of the present invention is not limited thereto, and anyThose familiar with the art, in the technical scope that the invention discloses, can readily occur in change or replace, should containCover within protection scope of the present invention.Therefore, protection scope of the present invention should described be as the criterion with scope of the claims.

Claims (10)

Convolutional neural networks training module, for the training of convolutional neural networks, concrete training process includes: divided by sensorCai Ji borborygmus sample signal and least one set interference sample signal;Described borborygmus sample signal and least one set are disturbedSample signal is all converted into numeral sample signal;Extract the time-frequency feature of described numeral sample signal;In time-frequency domain, to instituteState numeral sample signal and make signal label;Described signal label includes for labelling borborygmus signal at described borborygmus sampleThe borborygmus label of time of occurrence point in signal, and disturb signal time of occurrence point in described interference sample signal for labellingInterference label;From described numeral sample signal, borborygmus label is extracted according to described borborygmus label and described interference labelSignal and each group of interference label signal are as training sample;The time-frequency spectrum of numeral sample signal corresponding for described training sample is specialLevy as training data, be used for distinguishing borborygmus as supervision message, training using described borborygmus label and described interference labelSignal and the convolutional neural networks of various interference signal;
CN201610686377.6A2016-08-182016-08-18Borborygmus sound detection method, apparatus and system under noisy environmentActiveCN106328150B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN201610686377.6ACN106328150B (en)2016-08-182016-08-18Borborygmus sound detection method, apparatus and system under noisy environment

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN201610686377.6ACN106328150B (en)2016-08-182016-08-18Borborygmus sound detection method, apparatus and system under noisy environment

Publications (2)

Publication NumberPublication Date
CN106328150Atrue CN106328150A (en)2017-01-11
CN106328150B CN106328150B (en)2019-08-02

Family

ID=57744871

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN201610686377.6AActiveCN106328150B (en)2016-08-182016-08-18Borborygmus sound detection method, apparatus and system under noisy environment

Country Status (1)

CountryLink
CN (1)CN106328150B (en)

Cited By (19)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN106683680A (en)*2017-03-102017-05-17百度在线网络技术(北京)有限公司Speaker recognition method and device and computer equipment and computer readable media
CN106691498A (en)*2017-02-062017-05-24宁波江丰生物信息技术有限公司Borborygmus processing system
CN107545890A (en)*2017-08-312018-01-05桂林电子科技大学A kind of sound event recognition method
CN107961032A (en)*2018-01-042018-04-27哈尔滨工业大学Gurgling sound Time-domain Method of Extraction based on stethoscope array precorrection
CN108175436A (en)*2017-12-282018-06-19北京航空航天大学A kind of gurgling sound intelligence automatic identifying method
CN109192226A (en)*2018-06-262019-01-11深圳大学A kind of signal processing method and device
CN109620154A (en)*2018-12-212019-04-16平安科技(深圳)有限公司Borborygmus voice recognition method and relevant apparatus based on deep learning
CN110037733A (en)*2019-04-012019-07-23四川大学华西医院 A portable program-controlled wireless body sound monitoring system
CN110141266A (en)*2019-05-072019-08-20南京大学 Bowel sounds detection method based on wearable body sound capture technology
CN110179492A (en)*2019-05-082019-08-30四川新源生物电子科技有限公司Gurgling sound intelligent recognition algorithm based on auto-adaptive doublethreshold
CN110192892A (en)*2019-05-082019-09-03四川新源生物电子科技有限公司A kind of wireless gurgling sound analyzer
CN110398647A (en)*2019-06-262019-11-01深圳供电局有限公司Transformer state monitoring method
CN110432924A (en)*2019-08-062019-11-12杭州智团信息技术有限公司Borborygmus sound detection device, method and electronic equipment
CN110488278A (en)*2019-08-202019-11-22深圳锐越微技术有限公司Doppler radar signal kind identification method
CN113066483A (en)*2019-12-312021-07-02南昌航空大学 A Generative Adversarial Network Speech Enhancement Method Based on Sparse Continuity Constraints
CN113905662A (en)*2019-04-052022-01-07高丽大学校产学协力团 Digestive tract scanning device, body scanning device, body scanning method and acoustic digestive organ monitoring system
CN115040155A (en)*2021-03-092022-09-13浙江荷清柔性电子技术有限公司Bowel sound recognition method and device, electronic equipment and computer readable storage medium
CN116052725A (en)*2023-03-312023-05-02四川大学华西医院 A fine-grained bowel sound recognition method and device based on deep neural network
CN118370556A (en)*2024-06-212024-07-23长春理工大学Intelligent monitoring waist and abdomen belt for borborygmus and data processing system thereof

Citations (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US6056703A (en)*1996-04-032000-05-02Rush Presbyterian-St Luke's Medical CenterMethod and apparatus for characterizing gastrointestinal sounds
CN1994231A (en)*2006-01-062007-07-11财团法人工业技术研究院 Auscultation device and method for eliminating noise
CN104305961A (en)*2014-10-202015-01-28清华大学Bowel sounds monitoring and recognizing system
CN104811276A (en)*2015-05-042015-07-29东南大学DL-CNN (deep leaning-convolutional neutral network) demodulator for super-Nyquist rate communication
CN104966517A (en)*2015-06-022015-10-07华为技术有限公司Voice frequency signal enhancement method and device
CN105393252A (en)*2013-04-182016-03-09数字标记公司 Physiological Data Acquisition and Analysis

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US6056703A (en)*1996-04-032000-05-02Rush Presbyterian-St Luke's Medical CenterMethod and apparatus for characterizing gastrointestinal sounds
CN1994231A (en)*2006-01-062007-07-11财团法人工业技术研究院 Auscultation device and method for eliminating noise
CN105393252A (en)*2013-04-182016-03-09数字标记公司 Physiological Data Acquisition and Analysis
CN104305961A (en)*2014-10-202015-01-28清华大学Bowel sounds monitoring and recognizing system
CN104811276A (en)*2015-05-042015-07-29东南大学DL-CNN (deep leaning-convolutional neutral network) demodulator for super-Nyquist rate communication
CN104966517A (en)*2015-06-022015-10-07华为技术有限公司Voice frequency signal enhancement method and device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
张和华等: "肠鸣音信号的自适应滤波及其特征提取方法研究", 《中国医学物理学杂志》*

Cited By (28)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN106691498A (en)*2017-02-062017-05-24宁波江丰生物信息技术有限公司Borborygmus processing system
US10957339B2 (en)2017-03-102021-03-23Baidu Online Network Technology (Beijing) Co., Ltd.Speaker recognition method and apparatus, computer device and computer-readable medium
CN106683680A (en)*2017-03-102017-05-17百度在线网络技术(北京)有限公司Speaker recognition method and device and computer equipment and computer readable media
CN107545890A (en)*2017-08-312018-01-05桂林电子科技大学A kind of sound event recognition method
CN108175436A (en)*2017-12-282018-06-19北京航空航天大学A kind of gurgling sound intelligence automatic identifying method
CN107961032A (en)*2018-01-042018-04-27哈尔滨工业大学Gurgling sound Time-domain Method of Extraction based on stethoscope array precorrection
CN109192226A (en)*2018-06-262019-01-11深圳大学A kind of signal processing method and device
CN109620154A (en)*2018-12-212019-04-16平安科技(深圳)有限公司Borborygmus voice recognition method and relevant apparatus based on deep learning
CN110037733A (en)*2019-04-012019-07-23四川大学华西医院 A portable program-controlled wireless body sound monitoring system
CN110037733B (en)*2019-04-012024-04-02四川大学华西医院Portable program-controlled wireless body sound monitoring system
CN113905662A (en)*2019-04-052022-01-07高丽大学校产学协力团 Digestive tract scanning device, body scanning device, body scanning method and acoustic digestive organ monitoring system
CN110141266A (en)*2019-05-072019-08-20南京大学 Bowel sounds detection method based on wearable body sound capture technology
CN110141266B (en)*2019-05-072021-04-27南京大学 Bowel sound detection method based on wearable body sound capture technology
CN110192892B (en)*2019-05-082021-12-14四川新源生物电子科技有限公司Wireless bowel sound analyzer
CN110192892A (en)*2019-05-082019-09-03四川新源生物电子科技有限公司A kind of wireless gurgling sound analyzer
CN110179492A (en)*2019-05-082019-08-30四川新源生物电子科技有限公司Gurgling sound intelligent recognition algorithm based on auto-adaptive doublethreshold
CN110179492B (en)*2019-05-082021-12-14四川新源生物电子科技有限公司Intelligent bowel sound identification algorithm based on self-adaptive double thresholds
CN110398647A (en)*2019-06-262019-11-01深圳供电局有限公司Transformer state monitoring method
CN110432924A (en)*2019-08-062019-11-12杭州智团信息技术有限公司Borborygmus sound detection device, method and electronic equipment
CN110432924B (en)*2019-08-062021-10-22杭州智团信息技术有限公司Bowel sound detection device and method and electronic equipment
CN110488278A (en)*2019-08-202019-11-22深圳锐越微技术有限公司Doppler radar signal kind identification method
CN110488278B (en)*2019-08-202021-07-27深圳锐越微技术有限公司Doppler radar signal type identification method
CN113066483B (en)*2019-12-312024-01-30广州航海学院Sparse continuous constraint-based method for generating countermeasure network voice enhancement
CN113066483A (en)*2019-12-312021-07-02南昌航空大学 A Generative Adversarial Network Speech Enhancement Method Based on Sparse Continuity Constraints
CN115040155A (en)*2021-03-092022-09-13浙江荷清柔性电子技术有限公司Bowel sound recognition method and device, electronic equipment and computer readable storage medium
CN116052725A (en)*2023-03-312023-05-02四川大学华西医院 A fine-grained bowel sound recognition method and device based on deep neural network
CN118370556A (en)*2024-06-212024-07-23长春理工大学Intelligent monitoring waist and abdomen belt for borborygmus and data processing system thereof
CN118370556B (en)*2024-06-212024-08-16长春理工大学 An intelligent bowel sound monitoring waist and abdomen belt and its data processing system

Also Published As

Publication numberPublication date
CN106328150B (en)2019-08-02

Similar Documents

PublicationPublication DateTitle
CN106328150A (en)Bowel sound detection method, device and system under noisy environment
CN107061996B (en)A kind of water supply line leakage detecting and locating method
CN110245608B (en)Underwater target identification method based on half tensor product neural network
US4980917A (en)Method and apparatus for determining articulatory parameters from speech data
Allen et al.Signal analysis: time, frequency, scale, and structure
CN105488466B (en)A kind of deep-neural-network and Acoustic Object vocal print feature extracting method
Chatterjee et al.Extraction of binary black hole gravitational wave signals from detector data using deep learning
CN110786850B (en)Electrocardiosignal identity recognition method and system based on multi-feature sparse representation
EP3309782A1 (en)Method, device and system for noise suppression
Jorgensen et al.Web browser control using EMG based sub vocal speech recognition
CN109740523A (en) A Power Transformer Fault Diagnosis Method Based on Acoustic Features and Neural Networks
CN108648748A (en) Acoustic event detection method in hospital noise environment
CN112802484A (en)Panda sound event detection method and system under mixed audio frequency
CN110246490A (en)Voice keyword detection method and relevant apparatus
CN106847309A (en)A kind of speech-emotion recognition method
CN109409308A (en)A method of the birds species identification based on birdvocalization
CN109524020A (en)A kind of speech enhan-cement processing method
CN106941005A (en)A kind of vocal cords method for detecting abnormality based on speech acoustics feature
CN112633175A (en)Single note real-time recognition algorithm based on multi-scale convolution neural network under complex environment
CN113763966B (en)End-to-end text irrelevant voiceprint recognition method and system
CN101533068A (en)Analog-circuit fault diagnosis method based on DAGSVC
CN108922513A (en)Speech differentiation method, apparatus, computer equipment and storage medium
CN113435276A (en)Underwater sound target identification method based on antagonistic residual error network
CN102664010A (en)Robust speaker distinguishing method based on multifactor frequency displacement invariant feature
CN112257741B (en) A Generative Counterfeit Image Detection Method Based on Complex Neural Networks

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
C10Entry into substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant
TR01Transfer of patent right
TR01Transfer of patent right

Effective date of registration:20191113

Address after:100098 room 4b13, floor 4, building 1, yard 13, Dazhongsi, Haidian District, Beijing

Co-patentee after:Shandong Yi Mai Medical Technology Co., Ltd.

Patentee after:BEIJING YIMAI MEDICAL TECHNOLOGY CO., LTD.

Address before:100000, room 2, building 108, No. 2301, Haidian District, Beijing, Zhichun Road

Patentee before:BEIJING YIMAI MEDICAL TECHNOLOGY CO., LTD.


[8]ページ先頭

©2009-2025 Movatter.jp