Movatterモバイル変換


[0]ホーム

URL:


CN110808052A - Voice recognition method and device and electronic equipment - Google Patents

Voice recognition method and device and electronic equipment
Download PDF

Info

Publication number
CN110808052A
CN110808052ACN201911103239.0ACN201911103239ACN110808052ACN 110808052 ACN110808052 ACN 110808052ACN 201911103239 ACN201911103239 ACN 201911103239ACN 110808052 ACN110808052 ACN 110808052A
Authority
CN
China
Prior art keywords
voice data
voice
sub
user
filtering
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911103239.0A
Other languages
Chinese (zh)
Inventor
郑楚升
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Rui Yun Technology Co Ltd
Original Assignee
Shenzhen Rui Yun Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Rui Yun Technology Co LtdfiledCriticalShenzhen Rui Yun Technology Co Ltd
Priority to CN201911103239.0ApriorityCriticalpatent/CN110808052A/en
Publication of CN110808052ApublicationCriticalpatent/CN110808052A/en
Pendinglegal-statusCriticalCurrent

Links

Images

Classifications

Landscapes

Abstract

The invention provides a voice recognition method, a voice recognition device and electronic equipment. The voice recognition method comprises the following steps: acquiring voice data of a user and age information of the user; determining the frequency range information of the voice of the user according to the age information; dividing the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information; setting a filtering algorithm corresponding to each sub-frequency band; filtering the voice data of the corresponding sub-frequency band by adopting a corresponding filtering algorithm to obtain the filtered optimized voice data; and performing voice recognition on the optimized voice data. The invention can improve the accuracy of voice recognition.

Description

Voice recognition method and device and electronic equipment
Technical Field
The invention relates to the field of voice recognition, in particular to a voice recognition method, a voice recognition device and electronic equipment.
Background
Pitch is the sound level, which depends on frequency. The long, thick sounding body vibrates slowly, and the short, thin sounding body vibrates quickly. The height of the voice is related to the length, thickness and tightness of vocal cords of a person. Since the voices of children when speaking, the voices of middle-aged and elderly people among adults, and the like are different from each other, the common speech recognition model has a high error rate when being used for recognizing the voices of users of different ages, thereby reducing the recognition accuracy.
Disclosure of Invention
The embodiment of the invention provides a voice recognition method, a voice recognition device and electronic equipment, which have the beneficial effect of improving the recognition accuracy.
The invention provides a voice recognition method, which comprises the following steps:
acquiring voice data of a user and age information of the user;
determining the frequency range information of the voice of the user according to the age information;
dividing the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information;
setting a filtering algorithm corresponding to each sub-frequency band;
filtering the voice data of the corresponding sub-frequency band by adopting a corresponding filtering algorithm to obtain the filtered optimized voice data;
and performing voice recognition on the optimized voice data.
The present invention also provides a speech recognition apparatus comprising:
the first acquisition module is used for acquiring voice data of a user and age information of the user;
the second acquisition module is used for determining the frequency range information of the voice of the user according to the age information;
a dividing module, configured to divide a corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information;
the setting module is used for setting a filtering algorithm corresponding to each sub-frequency band;
the filtering module is used for filtering the voice data of the corresponding sub-frequency band by adopting a corresponding filtering algorithm to obtain the filtered optimized voice data;
and the voice recognition module is used for carrying out voice recognition on the optimized voice data.
The invention also provides an electronic device, which comprises a processor and a memory, wherein the memory is stored with a computer program, and the processor is used for the voice recognition method by calling the computer program stored in the memory.
The invention obtains the voice data of the user and the age information of the user; determining the frequency range information of the voice of the user according to the age information; dividing the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information; setting a filtering algorithm corresponding to each sub-frequency band; filtering the voice data of the corresponding sub-frequency band by adopting a corresponding filtering algorithm to obtain the filtered optimized voice data; performing voice recognition on the optimized voice data; because the voice data is filtered according to the age information and then subjected to voice recognition, the accuracy of the voice recognition is improved, and the voice recognition efficiency is improved.
Drawings
FIG. 1 is a flow chart of a speech recognition method according to an embodiment of the present invention;
FIG. 2 is a flowchart of a speech recognition method according to a second embodiment of the present invention;
fig. 3 is a schematic structural diagram of a speech recognition apparatus according to a third embodiment of the present invention.
Fig. 4 is a schematic diagram of a preferred structure of a speech recognition apparatus according to a third embodiment of the present invention.
Detailed Description
Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the accompanying drawings are illustrative only for the purpose of explaining the present invention, and are not to be construed as limiting the present invention.
In the description of the present invention, it should be noted that, unless otherwise explicitly specified or limited, the terms "mounted," "connected," and "connected" are to be construed broadly, e.g., as meaning either a fixed connection, a removable connection, or an integral connection; may be mechanically connected, may be electrically connected or may be in communication with each other; either directly or indirectly through intervening media, either internally or in any other relationship. The specific meanings of the above terms in the present invention can be understood by those skilled in the art according to specific situations.
In the present invention, unless otherwise expressly stated or limited, "above" or "below" a first feature means that the first and second features are in direct contact, or that the first and second features are not in direct contact but are in contact with each other via another feature therebetween. Also, the first feature being "on," "above" and "over" the second feature includes the first feature being directly on and obliquely above the second feature, or merely indicating that the first feature is at a higher level than the second feature. A first feature being "under," "below," and "beneath" a second feature includes the first feature being directly under and obliquely below the second feature, or simply meaning that the first feature is at a lesser elevation than the second feature.
The following disclosure provides many different embodiments or examples for implementing different features of the invention. To simplify the disclosure of the present invention, the components and arrangements of specific examples are described below. Of course, they are merely examples and are not intended to limit the present invention. Furthermore, the present invention may repeat reference numerals and/or letters in the various examples, such repetition is for the purpose of simplicity and clarity and does not in itself dictate a relationship between the various embodiments and/or configurations discussed. In addition, the present invention provides examples of various specific processes and materials, but one of ordinary skill in the art may recognize applications of other processes and/or uses of other materials.
Referring to fig. 1, fig. 1 is a flowchart illustrating a speech recognition method according to an embodiment of the present invention.
The speech recognition method of the embodiment comprises the following steps:
s101, voice data of a user and age information of the user are obtained.
The voice data may be voice uploaded by a user or voice collected in real time, and the voice data is voice of a user speaking so as to be distinguished from non-user voices such as environmental sounds or animal sounds.
And S102, determining the frequency range information of the voice of the user according to the age information.
Wherein, the age information of the user can be inquired in the registration information.
For example, users are roughly classified into children, young adults, middle-aged adults, and elderly adults, and the users are respectively associated with different age groups. Each age group of speech corresponds to a different frequency range. Such as children, young adults, middle-aged adults, and elderly people, are in different frequency ranges.
And S103, dividing the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information.
Wherein, the voice of each age group corresponds to different frequency range information. And the frequency bandwidth of the speech of people of different ages is different, and the sub-frequency band can be divided according to the total width of the frequency band.
And S104, setting a filtering algorithm corresponding to each sub-frequency band.
For example, a voice data optimization table may be preset, where multiple sub-bands are stored in the voice data optimization table, and the voice data of different sub-bands correspond to different filtering algorithms.
And the filtering algorithm corresponding to each sub-band is executed by adopting a Gaussian band-pass filter corresponding to the sub-band. Of course, a gaussian band-pass filter with multiple sub-bands may be provided.
And S105, filtering the voice data of the corresponding sub-frequency band by adopting a corresponding filtering algorithm to obtain the filtered optimized voice data.
In this step, each speech data segment may be filtered using a corresponding subband filtering algorithm. Specifically, each speech data segment may be filtered using a gaussian band-pass filter for the corresponding sub-band.
And S106, performing voice recognition on the optimized voice data.
In this step, existing common speech recognition algorithms can be used for speech recognition.
It is understood that in order to improve recognition accuracy, different speech recognition databases may be set according to users of different age groups to improve matching accuracy and efficiency.
The invention obtains the voice data of the user and the age information of the user; determining the frequency range information of the voice of the user according to the age information; dividing the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information; setting a filtering algorithm corresponding to each sub-frequency band; filtering the voice data of the corresponding sub-frequency band by adopting a corresponding filtering algorithm to obtain the filtered optimized voice data; performing voice recognition on the optimized voice data; because the voice data is filtered according to the age information and then subjected to voice recognition, the accuracy of the voice recognition is improved, and the voice recognition efficiency is improved.
Referring to fig. 2, fig. 2 is a flowchart of a speech recognition method according to a second embodiment of the present invention.
The speech recognition method of the embodiment comprises the following steps:
s201, voice data of a user and age information of the user are obtained.
The voice data may be voice uploaded by a user or voice collected in real time, and the voice data is voice of a user speaking so as to be distinguished from non-user voices such as environmental sounds or animal sounds.
In one embodiment, the step S201 may include:
s2011, voice data of a user is acquired;
s2012, acquiring the registration information of the user to acquire the age information of the user.
For example, the user may register personal information in a preset application program, upload voice through the application program, or send voice collection prompt information to the user through the application program, and then collect voice in real time through the microphone. The personal information includes information such as age information, sex, and contact information. The age information of the user can be obtained by acquiring personal information on the application program.
Of course, in other embodiments, the eyeball information of the user may also be collected by the camera, and the age of the user may be obtained by comparing the eyeball information with the preset eyeball information list. Eyeball information corresponding to different age groups is stored in the preset eyeball information list.
And S202, determining the frequency range information of the voice of the user according to the age information.
For example, users are roughly classified into children, young adults, middle-aged adults, and elderly adults, and the users are respectively associated with different age groups. Each age group of speech corresponds to a different frequency range. Such as children, young adults, middle-aged adults, and elderly people, are in different frequency ranges.
And S203, dividing the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information.
Wherein, the voice of each age group corresponds to different frequency range information. And the frequency bandwidth of the speech of people of different ages is different, and the sub-frequency band can be divided according to the total width of the frequency band.
And S204, setting a filtering algorithm corresponding to each sub-frequency band.
For example, a voice data optimization table may be preset, where multiple sub-bands are stored in the voice data optimization table, and the voice data of different sub-bands correspond to different filtering algorithms.
And the filtering algorithm corresponding to each sub-band is executed by adopting a Gaussian band-pass filter corresponding to the sub-band. Of course, in one embodiment, a gaussian band-pass filter with multiple sub-bands may be provided.
S205, dividing the voice data into a plurality of voice data segments, wherein each voice data segment is positioned in one of the plurality of sub-frequency bands;
in this step, the voice data may be divided into a plurality of voice data segments according to the band information of the voice data.
S206, filtering the voice data segments by adopting a corresponding filtering algorithm according to the sub-band corresponding to each voice data segment to obtain the filtered optimized voice data;
and acquiring a sub-band corresponding to the voice data segment, and filtering the voice data segment by adopting a filtering algorithm corresponding to the sub-band to obtain the filtered optimized voice data.
And S207, performing voice recognition on the optimized voice data.
In this step, existing common speech recognition algorithms can be used for speech recognition.
It can be understood that, in order to improve the recognition accuracy, the invention sets different voice recognition databases according to users of different age groups so as to improve the matching accuracy and efficiency.
In some embodiments, the step S207 may include:
s2071, acquiring a corresponding voice recognition database according to the age information of the user;
s2072, performing voice recognition on the optimized voice data according to the voice recognition database.
Specifically, in some embodiments, the step S2072 may specifically be:
(1) extracting the voice feature of the voice data of each sub-frequency band by using a Gaussian band-pass filter corresponding to each sub-frequency band;
(2) and extracting the amplitude spectrum of the voice characteristic, training a convolutional neural network by using the amplitude spectrum larger than a threshold value, and performing voice recognition on the optimized voice data by using the trained convolutional neural network and based on the corresponding voice recognition database.
Therefore, the voice data of the user and the age information of the user are obtained; determining the frequency range information of the voice of the user according to the age information; dividing the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information; setting a filtering algorithm corresponding to each sub-frequency band; dividing the voice data into a plurality of voice data segments, wherein each voice data segment is positioned in one of the plurality of sub-frequency bands, filtering the voice data segments by adopting a corresponding filtering algorithm according to the sub-frequency band corresponding to each voice data segment to obtain filtered optimized voice data, and performing voice recognition on the optimized voice data; on the basis of the previous embodiment, the voice data is divided into a plurality of voice data sections according to the frequency range, and then the voice data sections are filtered according to the filtering algorithm corresponding to each voice data section, so that the accuracy of voice recognition is further improved.
Referring to fig. 3, fig. 3 is a schematic structural diagram of a speech recognition device according to a third embodiment of the present invention.
Thespeech recognition apparatus 20 includes: a first obtainingmodule 21, a second obtainingmodule 22, a dividingmodule 23, asetting module 24, afiltering module 25 and avoice recognition module 26.
The first obtainingmodule 21 is used for obtaining voice data of a user and age information of the user.
The voice data may be voice uploaded by a user or voice collected in real time, and the voice data is voice of a user speaking so as to be distinguished from non-user voices such as environmental sounds or animal sounds.
With reference to fig. 4, the first obtainingmodule 21 includes: afirst acquisition unit 211 and asecond acquisition unit 212;
afirst acquisition unit 211 for acquiring voice data of a user;
a second obtainingunit 212, configured to obtain registration information of the user to obtain age information of the user.
(II) the second obtainingmodule 22 is used for determining the frequency range information of the voice of the user according to the age information.
Wherein, the age information of the user can be inquired in the registration information.
For example, users are roughly classified into children, young adults, middle-aged adults, and elderly adults, and the users are respectively associated with different age groups. Each age group of speech corresponds to a different frequency range. Such as children, young adults, middle-aged adults, and elderly people, are in different frequency ranges.
And (iii) thedividing module 23 is configured to divide the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information.
Wherein, the voice of each age group corresponds to different frequency range information. And the frequency bandwidth of the speech of people of different ages is different, and the sub-frequency band can be divided according to the total width of the frequency band.
And (iv) thesetting module 24 is configured to set a filtering algorithm corresponding to each of the sub-bands.
For example, a voice data optimization table may be preset, where multiple sub-bands are stored in the voice data optimization table, and the voice data of different sub-bands correspond to different filtering algorithms.
And the filtering algorithm corresponding to each sub-band is executed by adopting a Gaussian band-pass filter corresponding to the sub-band. Of course, a gaussian band-pass filter with multiple sub-bands may be provided.
Thefiltering module 25 is configured to filter the voice data of the corresponding sub-band by using a corresponding filtering algorithm to obtain the filtered optimized voice data;
each speech data segment may be filtered using a corresponding subband filtering algorithm. Specifically, each speech data segment may be filtered using a gaussian band-pass filter for the corresponding sub-band.
With reference to fig. 4, in an embodiment, thefiltering module 25 may include: a dividingunit 251 and afiltering unit 252.
A dividingunit 251, configured to divide the voice data into a plurality of voice data segments, where each voice data segment is located in one of the plurality of sub-bands;
thefiltering unit 252 is configured to filter the voice data segments by using a corresponding filtering algorithm according to the sub-band corresponding to each voice data segment, so as to obtain the filtered optimized voice data.
The filtering algorithm corresponding to each sub-band in thefiltering unit 252 is performed by using a gaussian band-pass filter corresponding to the sub-band.
And (six) thevoice recognition module 26 is used for performing voice recognition on the optimized voice data.
In one embodiment, thespeech recognition module 26 is configured to perform speech recognition using conventional speech recognition algorithms.
Referring to fig. 4, thespeech recognition module 26 may include: a database acquisition unit 261 and a speech recognition unit 262.
A database obtaining unit 261, configured to obtain a corresponding voice recognition database according to the age information of the user;
a speech recognition unit 262, configured to perform speech recognition on the optimized speech data according to the speech recognition database.
The speech recognition unit 262 is specifically configured to: extracting the voice feature of the voice data of each sub-frequency band by using a Gaussian band-pass filter corresponding to each sub-frequency band; and extracting the amplitude spectrum of the voice features, and training a convolutional neural network by using the amplitude spectrum larger than a threshold value so as to perform voice recognition on the optimized voice data by using the trained convolutional neural network.
Therefore, the voice data of the user and the age information of the user are obtained; determining the frequency range information of the voice of the user according to the age information; dividing the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information; setting a filtering algorithm corresponding to each sub-frequency band; filtering the voice data of the corresponding sub-frequency band by adopting a corresponding filtering algorithm to obtain the filtered optimized voice data; performing voice recognition on the optimized voice data; because the voice data is filtered according to the age information and then subjected to voice recognition, the accuracy of the voice recognition is improved, and the voice recognition efficiency is improved.
The invention also provides an electronic device comprising a processor and a memory. The processor is electrically connected with the memory. The processor is a control center of the electronic equipment, connects various parts of the whole server by using various interfaces and lines, executes various functions of the electronic equipment and processes data by running or calling a computer program stored in the memory and calling data stored in the memory, thereby integrally monitoring the electronic equipment.
In this embodiment, a processor of the electronic device loads instructions corresponding to processes of one or more computer programs into a memory according to the following steps, and the processor executes the computer programs stored in the memory, so as to implement various functions: acquiring voice data of a user and age information of the user; determining the frequency range information of the voice of the user according to the age information; dividing the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information; setting a filtering algorithm corresponding to each sub-frequency band; filtering the voice data of the corresponding sub-frequency band by adopting a corresponding filtering algorithm to obtain the filtered optimized voice data; and performing voice recognition on the optimized voice data.
The present invention also provides a storage medium having a computer program stored therein, which, when run on a computer, causes the computer to perform the method of any of the above embodiments, thereby implementing various functions: acquiring voice data of a user and age information of the user; determining the frequency range information of the voice of the user according to the age information; dividing the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information; setting a filtering algorithm corresponding to each sub-frequency band; filtering the voice data of the corresponding sub-frequency band by adopting a corresponding filtering algorithm to obtain the filtered optimized voice data; and performing voice recognition on the optimized voice data.
Therefore, the voice data of the user and the age information of the user are obtained; determining the frequency range information of the voice of the user according to the age information; dividing the corresponding frequency band into a plurality of sub-frequency bands according to the frequency range information; setting a filtering algorithm corresponding to each sub-frequency band; filtering the voice data of the corresponding sub-frequency band by adopting a corresponding filtering algorithm to obtain the filtered optimized voice data; performing voice recognition on the optimized voice data; because the voice data is filtered according to the age information and then subjected to voice recognition, the accuracy of the voice recognition is improved, and the voice recognition efficiency is improved.
In the description herein, references to the description of the terms "one embodiment," "certain embodiments," "an illustrative embodiment," "an example," "a specific example," or "some examples," etc., mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the invention. In this specification, schematic representations of the above terms do not necessarily refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.
In summary, although the present invention has been described with reference to the preferred embodiments, the above-described preferred embodiments are not intended to limit the present invention, and those skilled in the art can make various changes and modifications without departing from the spirit and scope of the present invention, therefore, the scope of the present invention shall be determined by the appended claims.

Claims (10)

CN201911103239.0A2019-11-122019-11-12Voice recognition method and device and electronic equipmentPendingCN110808052A (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN201911103239.0ACN110808052A (en)2019-11-122019-11-12Voice recognition method and device and electronic equipment

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN201911103239.0ACN110808052A (en)2019-11-122019-11-12Voice recognition method and device and electronic equipment

Publications (1)

Publication NumberPublication Date
CN110808052Atrue CN110808052A (en)2020-02-18

Family

ID=69502461

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN201911103239.0APendingCN110808052A (en)2019-11-122019-11-12Voice recognition method and device and electronic equipment

Country Status (1)

CountryLink
CN (1)CN110808052A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN111583495A (en)*2020-04-302020-08-25厦门快商通科技股份有限公司Remote access control system based on voice recognition and authorization method thereof
CN111583496A (en)*2020-04-302020-08-25厦门快商通科技股份有限公司Remote access control system based on voice recognition and control method thereof

Citations (7)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20100324898A1 (en)*2006-02-212010-12-23Sony Computer Entertainment Inc.Voice recognition with dynamic filter bank adjustment based on speaker categorization
CN101944359A (en)*2010-07-232011-01-12杭州网豆数字技术有限公司Voice recognition method facing specific crowd
CN105895105A (en)*2016-06-062016-08-24北京云知声信息技术有限公司Speech processing method and device
KR20180074210A (en)*2016-12-232018-07-03삼성전자주식회사Electronic device and voice recognition method of the electronic device
CN109271128A (en)*2018-09-042019-01-25Oppo广东移动通信有限公司Audio setting method, device, electronic equipment and storage medium
CN110197666A (en)*2019-05-302019-09-03广东工业大学A kind of audio recognition method neural network based, device
CN110427468A (en)*2019-07-102019-11-08深圳市一恒科电子科技有限公司A kind of learning method and learning machine based on children's cloud service

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20100324898A1 (en)*2006-02-212010-12-23Sony Computer Entertainment Inc.Voice recognition with dynamic filter bank adjustment based on speaker categorization
CN101944359A (en)*2010-07-232011-01-12杭州网豆数字技术有限公司Voice recognition method facing specific crowd
CN105895105A (en)*2016-06-062016-08-24北京云知声信息技术有限公司Speech processing method and device
KR20180074210A (en)*2016-12-232018-07-03삼성전자주식회사Electronic device and voice recognition method of the electronic device
CN109271128A (en)*2018-09-042019-01-25Oppo广东移动通信有限公司Audio setting method, device, electronic equipment and storage medium
CN110197666A (en)*2019-05-302019-09-03广东工业大学A kind of audio recognition method neural network based, device
CN110427468A (en)*2019-07-102019-11-08深圳市一恒科电子科技有限公司A kind of learning method and learning machine based on children's cloud service

Cited By (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN111583495A (en)*2020-04-302020-08-25厦门快商通科技股份有限公司Remote access control system based on voice recognition and authorization method thereof
CN111583496A (en)*2020-04-302020-08-25厦门快商通科技股份有限公司Remote access control system based on voice recognition and control method thereof

Similar Documents

PublicationPublication DateTitle
CN108922538B (en)Conference information recording method, conference information recording device, computer equipment and storage medium
US8595005B2 (en)System and method for recognizing emotional state from a speech signal
US6691090B1 (en)Speech recognition system including dimensionality reduction of baseband frequency signals
CN109036382B (en) An audio feature extraction method based on KL divergence
CN110047510A (en)Audio identification methods, device, computer equipment and storage medium
CN103730112B (en)Multi-channel voice simulation and acquisition method
CN103117061A (en)Method and device for identifying animals based on voice
CN110795996B (en)Method, device, equipment and storage medium for classifying heart sound signals
CN108847253B (en)Vehicle model identification method, device, computer equipment and storage medium
CN112820275A (en) An automatic monitoring method for analyzing abnormality of suckling piglets based on sound signals
Li et al.Robust speaker identification using an auditory-based feature
CN112908344A (en)Intelligent recognition method, device, equipment and medium for bird song
WO2023284813A1 (en)Deep learning algorithm-based snore monitoring method and system, and corresponding electric bed control method and system
CN110808052A (en)Voice recognition method and device and electronic equipment
Wang et al.Suppression by selecting wavelets for feature compression in distributed speech recognition
Gupta et al.Robust speech recognition using wavelet coefficient features
Murugaiya et al.Probability enhanced entropy (PEE) novel feature for improved bird sound classification
MouradSpeech enhancement based on stationary bionic wavelet transform and maximum a posterior estimator of magnitude-squared spectrum
CN111968651A (en)WT (WT) -based voiceprint recognition method and system
Nirjon et al.sMFCC: exploiting sparseness in speech for fast acoustic feature extraction on mobile devices--a feasibility study
CN113593604A (en)Method, device and storage medium for detecting audio quality
CN107993666A (en)Audio recognition method, device, computer equipment and readable storage medium storing program for executing
Meng et al.What is learnt by the LEArnable front-end (LEAF)? Adapting per-channel energy normalisation (PCEN) to noisy conditions
TW202143215A (en)Speech enhancement system based on deep learning
Singh et al.A novel algorithm using MFCC and ERB gammatone filters in speech recognition

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
RJ01Rejection of invention patent application after publication
RJ01Rejection of invention patent application after publication

Application publication date:20200218


[8]ページ先頭

©2009-2025 Movatter.jp