Movatterモバイル変換


[0]ホーム

URL:


CN107316638A - A kind of poem recites evaluating method and system, a kind of terminal and storage medium - Google Patents

A kind of poem recites evaluating method and system, a kind of terminal and storage medium
Download PDF

Info

Publication number
CN107316638A
CN107316638ACN201710504389.7ACN201710504389ACN107316638ACN 107316638 ACN107316638 ACN 107316638ACN 201710504389 ACN201710504389 ACN 201710504389ACN 107316638 ACN107316638 ACN 107316638A
Authority
CN
China
Prior art keywords
poem
voice
recited
recites
word
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710504389.7A
Other languages
Chinese (zh)
Inventor
高强
吴凡
夏龙
阎鹏
邓澍军
郭常圳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing ape force Education Technology Co., Ltd
Original Assignee
Beijing Chalk Future Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Chalk Future Technology Co LtdfiledCriticalBeijing Chalk Future Technology Co Ltd
Priority to CN201710504389.7ApriorityCriticalpatent/CN107316638A/en
Publication of CN107316638ApublicationCriticalpatent/CN107316638A/en
Pendinglegal-statusCriticalCurrent

Links

Classifications

Landscapes

Abstract

The application provides a kind of poem and recites evaluating method and system, a kind of terminal and storage medium, and methods described includes:Client determines poem to be recited;Obtain the voice of the poem to be recited;The voice is pre-processed;The pretreated voice is uploaded onto the server;The voice is converted into word using the identification model pre-established in the server;The word is compared with the poem to be recited, the score of the word is calculated according to default code of points;The score is fed back into client.

Description

A kind of poem recites evaluating method and system, a kind of terminal and storage medium
Technical field
The present invention relates to speech recognition and evaluation technology field, more particularly to a kind of poem recite evaluating method and system,A kind of terminal and storage medium.
Background technology
The subject user predominantly students in middle and primary schools that current poem is recited, general speech recognition technology is directed to other usersWhen be difficult the effect for being optimal sounding data as students in middle and primary schools, the recognition accuracy that poem is recited is than relatively low;IfWant to reach that higher recognition accuracy is accomplished by substantial amounts of training data, used training data is substantially by professional MikeWhat wind was recorded, then by the long training time, training method is also more complicated;Nowadays in mobile Internet epoch, userWhen progress poem recites detection, most of is all to record the poem back of the body using sound pick-up outfits such as PC microphone, mobile microphonesVoice is readed aloud, detection is then identified, using existing speech recognition technology to this poem by amateur microphone recordsWord recites speech recognition inaccurately, and needs the noise individually recited poem in voice to process, and increases extra denoisingThe sonication time, cost is high, the response time is long.
The content of the invention
In view of this, the application provides a kind of poem and recites evaluating method and system, a kind of terminal and storage medium, with realityNow the content that can be recited under environment poem of reciting in face of a variety of user groups and complexity is quick and precisely recognized and examinedMeasure and recite error rate.
On the one hand, the application provides a kind of poem and recites evaluating method, including:
It is determined that poem to be recited;
Obtain the voice of the poem to be recited;
The voice is pre-processed;
The pretreated voice is uploaded onto the server;
The voice is converted into word by the identification model pre-established in the server;
The word is compared with the poem to be recited, the word is calculated according to default code of pointsScore;
The score is fed back into client.
Alternatively, carrying out pretreatment to the voice includes:
To the voice burst;
The voice of burst is compressed.
Alternatively, it is described the pretreated voice is uploaded onto the server after also include:
To the speech decompression uploaded onto the server;
The voice after decompression is subjected to Jing Yin detection.
Alternatively, the step of pre-establishing the identification model includes:
Initial speech recognition network is built,
The speech recognition network receives voice training data;
By the wherein one voice training data conversion received into word;
Calculate the error between the word of conversion and the true word of the voice training data;
If the error is more than or equal to predetermined threshold value, the parameter of speech recognition network according to the error transfer factor,Then perform again it is described by the wherein one voice training data conversion received into word the step of, continue to train institute predicateSound recognizes network;
If the error is less than predetermined threshold value, the speech recognition network training terminates.
Alternatively, the construction voice training data include:
Original poem voice is obtained, the original poem voice is regard as training data;
The noise that adulterated in the original poem voice generates new training data.
Alternatively, the original poem voice includes the poem voice of different tonequality and/or alternative sounds feature.
Alternatively, the construction voice training data also include:
Original non-poem voice is obtained, the original non-poem voice is regard as training data.
Alternatively, the identification model includes hidden Markov model (Hidden Markov Model, HMM) and/or deepSpend neural network model (DNN).
Alternatively, it is described that the voice is converted into word using the identification model pre-established in the server, alsoIncluding:Error correction is carried out to the word after conversion using language model.
Alternatively, using language model to after conversion the word carry out error correction after, in addition to:Utilize probabilistic modelThe word after conversion is carried out into result with the true word of the poem to be recited to align.
Alternatively, the probabilistic model is used for judging that the back of the body is recited, recites again or omitted to the poem content recited as orderRead aloud.
Alternatively, the true word of the word after conversion and the poem to be recited is tied using probabilistic modelFruit is alignd, in addition to:
If the poem recited is correct, correct poem will be recited and be shown as the first color,
If the poem mistake recited, the poem for reciting mistake is shown as second color different from the first color;
If the poem recited again is correct, the poem for having been shown as the second color is revised as the first color;
In the event of the poem recited is omitted and the poem part recited is correct, then it will omit the poem recited and be shown as theSecond colors, will recite correct poem and are shown as the first color.
On the other hand, the application provides a kind of poem and recites evaluating system, including client and server, the clientIncluding:
Selecting module, for determining poem to be recited;
Acquisition module, the voice for obtaining the poem to be recited;
Pretreatment module, for being pre-processed to the voice;
Uploading module, for the pretreated voice to be uploaded onto the server;
The server includes:
Modular converter, for the voice to be converted into word using the identification model pre-established in the server;
Scoring modules, for the word to be compared with the poem to be recited, according to default code of points meterCalculate the score of the word;
Feedback module, for the score to be fed back into client.
Alternatively, the pretreatment module includes:
Burst module, for the voice burst;
Compression module, for the voice of burst to be compressed.
Alternatively, the uploading module also includes:
Decompression module, for the speech decompression to uploading onto the server;
Detection module, for the voice after decompression to be carried out into Jing Yin detection.
Alternatively, the identification model includes:
Identification network struction module, the speech recognition network initial for building,
Receiving module, voice training data are received for the speech recognition network;
Identification conversion module, for wherein one voice training data conversion will receiving into word;
Computing module, calculates the error between the word of conversion and the true word of the voice training data;
Judge module,
If the error is more than or equal to predetermined threshold value, the parameter of speech recognition network according to the error transfer factor,Then perform again it is described by the wherein one voice training data conversion received into word the step of, continue to train institute predicateSound recognizes network;If the error is less than predetermined threshold value, the speech recognition network training terminates.
Alternatively, the server also includes:
Poem voice acquisition module, for obtaining original poem voice, regard the original poem voice as training data;
Noise poem module, new training data is generated for the noise that adulterated in the original poem voice.
Alternatively, the original poem voice includes the poem voice of different tonequality and/or alternative sounds feature.
Alternatively, the server also includes:
Non- poem voice acquisition module, for obtaining original non-poem voice, regard the original non-poem voice as instructionPractice data.
Alternatively, the identification model includes hidden Markov model (Hidden Markov Model, HMM) and/or deepSpend neural network model (DNN).
Alternatively, the modular converter includes:Correction module, for being entered using language model to the word after conversionRow error correction.
Alternatively, the correction module also includes:Alignment module, for utilizing probabilistic model by the word after conversionResult is carried out with the true word of the poem to be recited to align.
Alternatively, the probabilistic model is used for judging that the back of the body is recited, recites again or omitted to the poem content recited as orderRead aloud.
Alternatively, the alignment module includes:
Order recites module, if correct for the poem recited, will recite correct poem and be shown as the first color,
If the poem mistake recited, the poem for reciting mistake is shown as second color different from the first color;
Again module is recited, if correct for the poem recited again, the poem of the second color will be had been shown asIt is revised as the first color;
Module is recited in omission, for correct in the event of the poem part omitted the poem recited and recited, then will be omittedThe poem recited is shown as the second color, will recite correct poem and is shown as the first color.
On the other hand, the application provides a kind of terminal, including processor and memory, and the memory storage has computerInstruction, the processor calls the computer instruction and performs following steps:
Client determines poem to be recited;
Obtain the voice of the poem to be recited;
The voice is pre-processed;
The pretreated voice is uploaded onto the server.
On the other hand, the application provides a kind of storage medium, and be stored with computer instruction, the computer instruction perform withLower step:
Client determines poem to be recited;
Obtain the voice of the poem to be recited;
The voice is pre-processed;
The pretreated voice is uploaded onto the server.
A kind of poem that the present patent application is provided recites evaluating method based on being pre-processed to the voice for reciting poem, carriesThe uploading speed of high voice;Voice is identified using the identification module pre-established and is converted into word, passes through default scoringRule calculates word score, so as to detect to recite the error rate of the voice of poem so that the poem voice recited can enterRecognition accuracy is improved while row Real time identification.
Brief description of the drawings
Fig. 1 recites the flow chart of evaluating method for the poem that the embodiment of the application one is provided;
Fig. 2 for the poem that the embodiment of the application one is provided recite evaluating method to voice carry out pretreatment process figure;
Fig. 3 for the poem that the embodiment of the application one is provided recite evaluating method at the voice that uploads onto the serverThe flow chart of reason;
Fig. 4 recites the flow being trained in evaluating method to identification model for the poem that the embodiment of the application one is providedFigure;
Fig. 5 is the flow chart that the embodiment of the application one is trained to HMM-DNN models;
Fig. 6 recites the structural representation of evaluating system for a kind of poem that the embodiment of the application one is provided;
The hardware architecture diagram for the electronic equipment that Fig. 7 provides for the embodiment of the application one;
Fig. 8 recites client after evaluating method is evaluated and tested for the use poem that the embodiment of the application one is provided and is presented to useThe interface at family.
Embodiment
A kind of poem provided in an embodiment of the present invention recites evaluating method and system, a kind of terminal and storage medium, firstThe voice for obtaining poem to be recited is pre-processed, identification model is then based on and text conversion is carried out to the voice so thatThe evaluation and test that poem is recited is more quick and accurate.Below in conjunction with the accompanying drawings, embodiments of the present invention and implementation process are done in detailExplanation.
Existing speech recognition technology need substantial amounts of training speech data and longer training time can be only achieved compared withGood speech recognition, the speech recognition that can only for special equipment record low for the speech recognition accuracy of different usersEffect better, and needs individually to carry out noise processed to the voice for needing to recognize, cost is big, the response time is longer.
The deficiency of evaluating method is recited based on poem in the prior art, a kind of poem that the application is provided recites evaluating methodAnd system, a kind of terminal and storage medium, it is capable of the ancient poetry recorded to different user using common sound pick-up outfit of efficiently and accuratelyThe voice of word is identified.
Referring to Fig. 1, the present embodiment provides a kind of poem and recites evaluating method, including step 101 is to step 107.
Step 101:It is determined that poem to be recited.
In the present embodiment, the poem is classic poetry, and poem is a kind of exclusive style of Chinese, there is special form and rhythmRule;Poem can be divided into pre-Tang poetry and the class of "modern style" poetry, referring to innovations in classical poetry during the Tang Dynasty, marked by strict tonal patterns and rhyme schemes two by musical note point;Pre-Tang poetry and "modern style" poetry, referring to innovations in classical poetry during the Tang Dynasty, marked by strict tonal patterns and rhyme schemes are the concepts formed the Tang Dynasty, are from poemMusical note angle divide;It can be divided into epic, lyric by content, see off poem, frontier poem, Natural environment, poems on history(singing of history poem), mourning poems, object-ode poems, army's poem etc..Pre-Tang poetry has《The Book of Songs》《The Songs of the South》《Music Bureau》《The Chinese is assigned》《Northern and Southern Dynasties' folk song》Deng."modern style" poetry, referring to innovations in classical poetry during the Tang Dynasty, marked by strict tonal patterns and rhyme schemes is usually the poem of four lines, regulated verse, extended form of regulated verse (long rule);User selects to determine what is recited in the application program of clientClassic poetry full text or classic poetry paragraph, start to recite.
Step 102:Obtain the voice of the poem to be recited.
In the present embodiment, the voice that above-mentioned user recites poem is obtained in real time.
Step 103:The voice is pre-processed.
In the present embodiment, above-mentioned voice is pre-processed, farthest realizes the real-time that the voice is uploaded.
Step 104:The pretreated voice is uploaded onto the server.
Step 105:The voice is converted into word by the identification model pre-established in the server.
In the present embodiment, the identification model trained is pre-established on the server, by pretreated voiceWord is identified into the identification model of the server.
Step 106:The word is compared with the poem to be recited, institute is calculated according to default code of pointsState the score of word.
Step 107:The score is fed back into client.
A kind of poem that the present embodiment is provided is recited evaluating method and pre-processed based on the voice to reciting poem, improvesThe uploading speed of voice;Voice is identified using the identification module pre-established and is converted into word, is advised by default scoringWord score is then calculated, so as to detect to recite the error rate of the voice of poem so that the poem voice recited can be carried outRecognition accuracy is improved while Real time identification.
Referring to Fig. 2, the process pre-processed in the embodiment of the application one to the voice includes step 201 to step202。
Step 201:To the voice burst.
In the present embodiment, the burst of the voice, compression and upload are all sightless to user, and user uses clientApplication program (for example, APP on the smart mobile phone) progress at end is continual to recite, and uniquely visible is exactly real-time feedback knotReally;The audio that smart mobile phone is recorded is lossless audio, and being uploaded directly into server needs to expend larger network traffics, in order toThe network flow consumption of smart mobile phone is reduced, we have carried out burst to audio and compression is carried out again before audio is uploadedPass server.
In the present embodiment, it is into isometric fragment, in order to maximum by phonetic segmentation to the voice burstReal-time is realized, in the case where not influenceing speech recognition speed, speech audio is sliced into as far as possible short fragment, by institute's predicateSound is controlled within 1 in the speed that the identification model is recognized, as long as speed is less than 1, it becomes possible to realize good real-time.
Step 202:The voice of burst is compressed.
In the present embodiment, the voice of burst is compressed, is not lose the information of useful information or lossUnder the conditions of insignificant, with Digital Signal Processing, raw tone audio stream is compressed, also referred to as compressed encoding;ThisEmbodiment carrys out compressed encoding raw tone audio using advanced message coding (Advanced Audio Coding, AAC) formStream.
Referring to Fig. 3, in the embodiment of the application one to the pretreated voice is uploaded onto the server after also include stepRapid 301 to step 302.
Step 301:To the speech decompression uploaded onto the server.
In the present embodiment, the server is after the upload compress speech file is received, it is necessary to described in compressionSpeech audio is carried out in decompression, the present embodiment using advanced message coding (Advanced Audio Coding, AAC)Inverse transformation decompression is carried out to the speech audio stream.
Step 302:The voice after decompression is subjected to Jing Yin detection.
In the present embodiment, user carry out being likely to forget when classic poetry is recited occur the behavior recited of pause orIt is that other uncertain factors cause user carrying out the vacancy of voice occur when classic poetry is recited, can produces substantial amounts of this whenJing Yin, i.e., noiseless fragment, this part audio is without being handled, and in order to reduce unnecessary processing time, we makeCarry out Jing Yin detection with the disaggregated model based on deep learning to calculate, the disaggregated model based on deep learning can be quick and preciselyDistinguish normal sound and Jing Yin, the silence clip detected is not done recognition processing.
Referring to Fig. 4, the training of identification model described in the embodiment of the application one is illustrated, and a plurality of language is used in practical applicationSound training data is trained to the identification model, every time wherein one voice training data of input, when this voice trainingData will input next voice training data after terminating to identification model training and the identification model is instructedPractice, until all voice training data have been fully entered or when the identification model reaches goal-selling to the identificationThe training of model terminates, and the training process includes step 401 and arrives step 408.
Step 401:Build initial speech recognition network.
In the present embodiment, it is the parameter for initializing identification model to build initial speech recognition network.
Step 402:The speech recognition network receives voice training data.
In the present embodiment, the voice training data are numbered, it is to avoid the voice training Data duplication occur defeatedEnter.
Step 403:By the wherein one voice training data conversion received into word.
Step 404:Calculate the error between the word of conversion and the true word of the voice training data.
Step 405:Judge whether the error is less than predetermined threshold value, if so, step 406 is performed, if it is not, then performing stepRapid 408.
Step 406:The parameter of speech recognition network according to the error transfer factor.
Step 407:Whether be the last item voice training data, if so, performing step if judging the voice training data408, if it is not, then performing step 403.
Step 408:The speech recognition network training terminates.
In the present embodiment, the construction voice training data include:
Original poem voice is obtained, the original poem voice is regard as training data;It is described to obtain original poem voiceIt can be the speech audio got by equipment such as smart mobile phone microphone, non-smart mobile phone microphones.
The noise that adulterated in the original poem voice generates new training data;In order to allow the classic poetry of user to reciteSpeech recognition can also keep higher accuracy rate in the case where there is noise circumstance, and the training data uses original poem voice, isAny Denoising disposal is not carried out to the speech audio got, but is directly entered using the voice audio data with noiseRow training, distinguished in identification model identification process by the method for pattern-recognition which be noise, which be to recite ancient poetrySound.
In the present embodiment, the original poem voice includes the poem voice of different tonequality and/or alternative sounds feature;PoemThe user group that word recites evaluation and test uses mainly for middle and primary schools, in order to adapt to the characteristic voice of students in middle and primary schools, the voice instructionIt is largely the voice audio data for gathering real middle and primary schools user to practice data;But it is due to the custom that different users speaksDifferent with sounding feature, tonequality and sound characteristic are had any different, in order to improve the generalization ability of identification model, voice training numberThe original poem voice is expanded according to using the enhanced method of data, the voice audio data of different characteristics is generated, increasedPlus the voice training data.
In the present embodiment, the construction voice training data also include:
Original non-poem voice is obtained, the original non-poem voice is regard as training data;In the training data alsoAdd the speech audio of some non-smart mobile phone microphone records, the speech audio of non-students in middle and primary schools user, the non-ancient poetry of contentSpeech audio under the speech audio of word, non-noise environment, improves the generalization ability of identification model so that identification model has moreGood robustness.
Referring to Fig. 5, provided in the embodiment of the application one and hidden Markov model is used with a voice training data instance(Hidden Markov Model, HMM) and/or deep neural network model (DNN) training pattern include step 501 and arrive step508。
In the present embodiment, the identification model include hidden Markov model (Hidden Markov Model, HMM) and/Or deep neural network model (DNN).
Step 501:Initialize HMM-DNN model parameters.
Step 502:Speech data is received from training data.
Step 503:Wherein one speech data received is converted into word.
Step 504:Calculate the error between the word changed out after speech recognition and real language and characters.
Step 505:Judge whether the error is less than predetermined threshold value, if so, step 506 is performed, if it is not, then performing stepRapid 508.
Step 506:The parameter of HMM-DNN models according to the error transfer factor.
Step 507:Whether be the last item speech data, if so, step 508 is performed, if not if judging the speech dataIt is then to perform step 503.
Step 508:The HMM-DNN model trainings terminate.
In the present embodiment, training data is to be expanded on the basis of seed voice using the enhanced method of data's;Seed speech data has the characteristics that:1) seed speech data is the real speech voice data collected from application;2)In order to adapt to the speech audio of smart mobile phone microphone records, our key data be all by smart mobile phone (including but notIt is limited to Android intelligent, iphone) collect;3) in order to adapt to the characteristic voice of students in middle and primary schools, our main languageSound data are all the voice audio datas of the real middle and primary schools user produced on line;4) real recite in speech data is adulteratedVarious ambient noises, can also keep higher accurate in a noisy environment to allow classic poetry to recite language identificationRate, does not carry out any Denoising disposal to audio, but is directly trained using the voice data with noise, passes through mouldThe sound that formula knows method for distinguishing to distinguish noise He recite ancient poetry;The custom and sounding feature spoken due to different users are differedSample, employs data enhanced method and seed data is expanded, generate the sound of different characteristics.
In order to improve the voice for also having added some non-smart mobile phone microphone records in the generalization ability of model, training dataSpeech audio under audio, the speech audio of non-students in middle and primary schools user, the speech audio of the non-classic poetry of content, non-noise environment,So that HMM-DNN models have more preferable robustness.
The embodiment of the application one, word is converted into using the identification model pre-established in the server by the voiceAfterwards, in addition to:Error correction is carried out to the word after conversion using language model.
Language model, which is one, can deduce the probability distribution of next word, i.e., next word is probably the mould of what wordType.
The problem of speech recognition modeling of reality there may be:1) pronunciation that speech recognition modeling is identified, may not beIt is absolutely accurate, even if but it is inaccurate, it is also in most cases an approximate pronunciation;2) pure speech recognition mouldThe recognition result of type and may not meet correct syntax gauge.
It is adjusted, can be corrected by context and syntax rule by mistake by natural language model in the present embodimentThe pronunciation of identification, method of adjustment includes but is not limited to:Merge the language material under several scenes and train the language model come so thatThe correction result of pronunciation has more robustness;It has adjusted weight of the language model in speech recognition process so that recognition result is morePlus precisely.
In the embodiment of the application one, using language model to after conversion the word carry out error correction after, in addition to:ProfitThe word after conversion is carried out into result with the true word of the poem to be recited with probabilistic model to align.
The result of speech recognition and the content that user truly to be recited are gone word for word to be alignd, it is that order is carried on the back to judge userThe content to be carried on the back is readed aloud, recited again, omitted and recite or recite mistake, then poem is recited using default code of pointsContent carries out evaluation and test marking.
In the embodiment of the application one, the probabilistic model is used for judging that the poem content recited is recited as order, carried on the back againRead aloud or omit and recite.
In true recite, sequentially recite, recite again and omit that to recite the frequency that three behaviors occur different, mostlyIt is that order is recited in the case of number, recites again and omission recites first using a kind of combination mankind in fewer generation, the present embodimentTest the bayesian probability model of knowledge to predict these three behaviors so that the accuracy rate of alignment is optimal.
In the embodiment of the application one, using probabilistic model by the true of the word after conversion and the poem to be recitedWord carries out result alignment, in addition to:
If the poem recited is correct, correct poem will be recited and be shown as the first color;
If the poem mistake recited, the poem for reciting mistake is shown as second color different from the first color;
If the poem recited again is correct, the poem for having been shown as the second color is revised as the first color;
In the event of the poem recited is omitted and the poem part recited is correct, then it will omit the poem recited and be shown as theSecond colors, will recite correct poem and are shown as the first color.
Referring to Fig. 6, the application provides a kind of poem and recites evaluating system, including client 601 and server 602, describedClient 601 includes:
Selecting module 611, for determining poem to be recited;
Acquisition module 612, the voice for obtaining the poem to be recited;
Pretreatment module 613, for being pre-processed to the voice;
Uploading module 614, for the pretreated voice to be uploaded onto the server;
The server 602 includes:
Modular converter 621, it is written for being changed the voice using the identification model pre-established in the serverWord;
Scoring modules 622, for the word to be compared with the poem to be recited, according to default code of pointsCalculate the score of the word;
Feedback module 623, for the score to be fed back into client.
Alternatively, the pretreatment module 613 includes:
Burst module, for the voice burst;
Compression module, for the voice of burst to be compressed.
Alternatively, the uploading module 614 also includes:
Decompression module, for the speech decompression to uploading onto the server;
Detection module, for the voice after decompression to be carried out into Jing Yin detection.
Alternatively, the identification model includes:
Identification network struction module, the speech recognition network initial for building,
Receiving module, voice training data are received for the speech recognition network;
Identification conversion module, for wherein one voice training data conversion will receiving into word;
Computing module, calculates the error between the word of conversion and the true word of the voice training data;
Judge module, if the error is more than or equal to predetermined threshold value, the speech recognition net according to the error transfer factorThe parameter of network, then perform again it is described by the wherein one voice training data conversion received into word the step of, continueTrain the speech recognition network;If the error is less than predetermined threshold value, the speech recognition network training terminates.
Alternatively, the server also includes:
Poem voice acquisition module, for obtaining original poem voice, regard the original poem voice as training data;
Noise poem module, new training data is generated for the noise that adulterated in the original poem voice.
Alternatively, the original poem voice includes the poem voice of different tonequality and/or alternative sounds feature.
Alternatively, the server also includes:
Non- poem voice acquisition module, for obtaining original non-poem voice, regard the original non-poem voice as instructionPractice data.
Alternatively, the identification model includes hidden Markov model (Hidden Markov Model, HMM) and/or deepSpend neural network model (DNN).
Alternatively, the modular converter includes:Correction module, for being entered using language model to the word after conversionRow error correction.
Alternatively, the correction module also includes:Alignment module, for utilizing probabilistic model by the word after conversionResult is carried out with the true word of the poem to be recited to align.
Alternatively, the probabilistic model is used for judging that the back of the body is recited, recites again or omitted to the poem content recited as orderRead aloud.
Alternatively, the alignment module includes:
Order recites module, if correct for the poem recited, will recite correct poem and be shown as the first color,
If the poem mistake recited, the poem for reciting mistake is shown as second color different from the first color;
Again module is recited, if correct for the poem recited again, the poem of the second color will be had been shown asIt is revised as the first color;
Module is recited in omission, for correct in the event of the poem part omitted the poem recited and recited, then will be omittedThe poem recited is shown as the second color, will recite correct poem and is shown as the first color.
The DNN in HMM-DNN models used in identification model in the application can be various deep learning networks;HMM-DNN models can with hidden Markov model and gauss hybrid models (Hidden Markov Model, HMM andGaussian Mixture Model, i.e. HMM-GMM) model substitutes, can also be substituted with pure deep learning model.
The present embodiment provides a kind of terminal, including processor and memory, and the memory storage has computer instruction, instituteProcessor is stated to call the computer instruction and perform following steps:
Client determines poem to be recited;
Obtain the voice of the poem to be recited;
The voice is pre-processed;
The pretreated voice is uploaded onto the server.
A kind of exemplary scheme of above-mentioned terminal for the present embodiment.It should be noted that the technical scheme of the terminal withThe technical scheme that above-mentioned poem recites evaluating method belongs to same design, the details that the technical scheme of the terminal is not described in detailContent, may refer to the description that above-mentioned poem recites the technical scheme of evaluating method.
The present embodiment provides a kind of storage medium, and be stored with computer instruction, and the computer instruction performs following steps:
Client determines poem to be recited;
Obtain the voice of the poem to be recited;
The voice is pre-processed;
The pretreated voice is uploaded onto the server.
A kind of exemplary scheme of above-mentioned storage medium for the present embodiment.It should be noted that the skill of the storage mediumThe technical scheme that art scheme recites evaluating method with above-mentioned poem belongs to same design, and the technical scheme of storage medium is not detailedThe detail content of description, may refer to the description that above-mentioned poem recites the technical scheme of evaluating method.
A kind of poem that the present embodiment is provided, which recites evaluating method and system, a kind of terminal and storage medium, following excellentPoint:
1. cost is low, the dedicated voice identification technology recited for classic poetry, training method is relatively easy, training needsData volume is relatively small, the training time is short;
2. speed is fast, progress detection in real time and feedback are recited classic poetry, and the speed of speech recognition is less than 1;
3. effect is good, (include but is not limited to various Android phone equipment, iphone hands in various Smartphone devicesMachine equipment), under various noise circumstances (including but is not limited to quiet environment, road, subway, coffee shop etc.), classic poetry recites inspectionAccuracy rate is surveyed more than 97%.
Fig. 7 is the hardware architecture diagram for the electronic equipment that the poem that the embodiment of the present application is provided recites evaluating method, such asShown in Fig. 7, the electronic equipment includes:
In one or more processors 710 and memory 720, Fig. 7 by taking a processor 710 as an example.
The equipment that execution poem recites evaluating method can also include:Input unit 730 and output device 740.
Processor 710, memory 720, input unit 730 and output device 740 can pass through bus or other modesIn connection, Fig. 7 exemplified by being connected by bus 750.
Memory 720 is as a kind of non-volatile computer readable storage medium storing program for executing, available for storage non-volatile software journeyIt is corresponding that poem in sequence, non-volatile computer executable program and module, such as the embodiment of the present application recites evaluating methodProgrammed instruction/module (for example, modules shown in accompanying drawing 6).Processor 710 is stored in memory 720 by operationNon-volatile software program, instruction and module, so that various function application and the data processing of execute server, that is, realizeThe poem of above method embodiment recites evaluating method.
Memory 720 can include storing program area and storage data field, wherein, storing program area can store operation systemApplication program required for system, at least one function;Storage data field can store the use institute that evaluating system is recited according to poemData of establishment etc..In addition, memory 720 can include high-speed random access memory, non-volatile memories can also be includedDevice, for example, at least one disk memory, flush memory device or other non-volatile solid state memory parts.In some embodimentsIn, memory 720 is optional including the memory remotely located relative to processor 710, and these remote memories can pass through netNetwork is connected to poem and recites evaluating system.The example of above-mentioned network include but is not limited to internet, intranet, LAN,Mobile radio communication and combinations thereof.
Input unit 730 can receive the numeral or character information of input, and produce the use that evaluating system is recited with poemThe key signals input that family is set and function control is relevant.Output device 740 may include the display devices such as display screen.
One or more of modules are stored in the memory 720, when by one or more of processorsDuring 710 execution, the poem performed in above-mentioned any means embodiment recites evaluating method.
The said goods can perform the method that the embodiment of the present application is provided, and possesses the corresponding functional module of execution method and hasBeneficial effect.Not ins and outs of detailed description in the present embodiment, reference can be made to the method that the embodiment of the present application is provided.
The electronic equipment of the embodiment of the present invention exists in a variety of forms, includes but is not limited to:
(1) mobile communication equipment:The characteristics of this kind equipment is that possess mobile communication function, and to provide speech, dataCommunicate as main target.This Terminal Type includes:Smart mobile phone, multimedia handset, feature mobile phone, and low-end mobile phone etc..
(2) super mobile personal computer equipment:This kind equipment belongs to the category of personal computer, there is calculating and processing work(Can, typically also possess mobile Internet access characteristic.This Terminal Type includes:Palm PC (PDA Personal DigitalAssistant), mobile internet device (MID, Mobile Internet Device) and Ultra-Mobile PC(UMPC, Ultra-mobile Personal Computer) equipment etc..
(3) portable entertainment device:This kind equipment can show and play content of multimedia.The kind equipment includes:Audio,Video player, handheld device, e-book, and intelligent toy and portable car-mounted navigation equipment.
(4) server:The equipment for providing the service of calculating, the composition of server is total including processor, hard disk, internal memory, systemLine etc., server is similar with general computer architecture, but is due to need to provide highly reliable service, therefore in processing energyRequire higher in terms of power, stability, reliability, security, scalability, manageability.
(5) other electronic systems with data interaction function.
, can be by it in multiple embodiments provided herein, it should be understood that disclosed system and methodIts mode is realized.For example, the embodiment of detecting system described above is only schematical, for example, the moduleDivide, only a kind of division of logic function there can be other dividing mode when actually realizing, such as multiple module or componentsAnother system can be combined or be desirably integrated into, or some features can be ignored, or do not perform.It is another, it is shown orThe coupling each other discussed or direct-coupling or communication linkage can be by some interfaces, the INDIRECT COUPLING of module or logicalLetter link, can be electrical, machinery or other forms.
The module illustrated as separating component can be or may not be it is physically separate, it is aobvious as moduleThe part shown can be or may not be physical module, you can with positioned at a place, or can also be distributed to multipleOn mixed-media network modules mixed-media.Some or all of module therein can be selected to realize the mesh of this embodiment scheme according to the actual needs's.
In addition, each functional module in each embodiment of the invention can be integrated in a processing module, can alsoThat modules are individually physically present, can also two or more modules be integrated in a module.Above-mentioned integrated mouldBlock can both be realized in the form of hardware, it would however also be possible to employ the form of software function module is realized.
If the integrated module is realized using in the form of software function module and as independent production marketing or usedWhen, it can be stored in a computer read/write memory medium.Understood based on such, technical scheme is substantiallyThe part contributed in other words to prior art or all or part of the technical scheme can be in the form of software productsEmbody, the computer software product is stored in a storage medium, including some instructions are to cause a computerEquipment (can be personal computer, server, or network equipment etc.) performs the complete of each embodiment methods described of the inventionPortion or part steps.And foregoing storage medium includes:USB flash disk, mobile hard disk, read-only storage (ROM, Read-OnlyMemory), random access memory (RAM, Random Access Memory), magnetic disc or CD etc. are various can store journeyThe medium of sequence code.
It should be noted that for foregoing each method embodiment, for simplicity description, therefore it is all expressed as a series ofCombination of actions, but those skilled in the art should know, the present invention is not limited by described sequence of movement becauseAccording to the present invention, some steps can use other orders or carry out simultaneously.Secondly, those skilled in the art should also knowKnow, embodiment described in this description belongs to preferred embodiment, and involved action and module might not all be this hairsNecessary to bright.
In the above-described embodiments, the description to each embodiment all emphasizes particularly on different fields, and does not have the portion being described in detail in some embodimentPoint, it may refer to the associated description of other embodiments.
Present invention disclosed above preferred embodiment is only intended to help and illustrates the present invention.Alternative embodiment is not detailedAll details of narration, it is only described embodiment that the invention is not limited yet.Obviously, according to the content of this specification,It can make many modifications and variations.This specification is chosen and specifically describes these embodiments, is to preferably explain the present inventionPrinciple and practical application so that skilled artisan can be best understood by and utilize the present invention.The present invention is onlyLimited by claims and its four corner and equivalent.

Claims (26)

CN201710504389.7A2017-06-282017-06-28A kind of poem recites evaluating method and system, a kind of terminal and storage mediumPendingCN107316638A (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN201710504389.7ACN107316638A (en)2017-06-282017-06-28A kind of poem recites evaluating method and system, a kind of terminal and storage medium

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN201710504389.7ACN107316638A (en)2017-06-282017-06-28A kind of poem recites evaluating method and system, a kind of terminal and storage medium

Publications (1)

Publication NumberPublication Date
CN107316638Atrue CN107316638A (en)2017-11-03

Family

ID=60181280

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN201710504389.7APendingCN107316638A (en)2017-06-282017-06-28A kind of poem recites evaluating method and system, a kind of terminal and storage medium

Country Status (1)

CountryLink
CN (1)CN107316638A (en)

Cited By (24)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN108231090A (en)*2018-01-022018-06-29深圳市酷开网络科技有限公司Text reading level appraisal procedure, device and computer readable storage medium
CN108320734A (en)*2017-12-292018-07-24安徽科大讯飞医疗信息技术有限公司Audio signal processing method and device, storage medium, electronic equipment
CN108389440A (en)*2018-03-152018-08-10广东小天才科技有限公司Voice playing method and device based on microphone and voice playing equipment
CN109086387A (en)*2018-07-262018-12-25上海慧子视听科技有限公司A kind of audio stream methods of marking, device, equipment and storage medium
CN110010123A (en)*2018-01-162019-07-12上海异构网络科技有限公司English phonetic word pronunciation learning evaluation system and method
CN110010157A (en)*2019-03-272019-07-12广东小天才科技有限公司 Test method, device, equipment and storage medium
CN110060663A (en)*2019-04-282019-07-26北京云迹科技有限公司A kind of method, apparatus and system of answer service
CN110085210A (en)*2019-03-152019-08-02平安科技(深圳)有限公司Interactive information test method, device, computer equipment and storage medium
CN110148413A (en)*2019-05-212019-08-20科大讯飞股份有限公司Speech evaluating method and relevant apparatus
CN110176235A (en)*2019-05-232019-08-27腾讯科技(深圳)有限公司Methods of exhibiting, device, storage medium and the computer equipment of speech recognition text
CN110309350A (en)*2018-03-212019-10-08腾讯科技(深圳)有限公司Recite processing method, system, device, medium and the electronic equipment of task
CN110310086A (en)*2019-06-062019-10-08安徽淘云科技有限公司Auxiliary recites based reminding method, equipment and storage medium
CN110322895A (en)*2018-03-272019-10-11亿度慧达教育科技(北京)有限公司Speech evaluating method and computer storage medium
CN110349583A (en)*2019-07-152019-10-18高磊A kind of Game education method and system based on speech recognition
CN110706710A (en)*2018-06-252020-01-17普天信息技术有限公司Voice recognition method and device, electronic equipment and storage medium
CN111128181A (en)*2019-12-092020-05-08科大讯飞股份有限公司Recitation question evaluation method, device and equipment
CN111383630A (en)*2020-03-042020-07-07广州优谷信息技术有限公司Text recitation evaluation method and device and storage medium
CN111444679A (en)*2020-03-272020-07-24北京小米松果电子有限公司Poetry generation method and device, electronic equipment and storage medium
CN111640342A (en)*2018-09-042020-09-08北京猿力教育科技有限公司Online examination method and device, computing device and storage medium
CN111833850A (en)*2019-04-182020-10-27霍尼韦尔国际公司 Method and system for cockpit speech recognition acoustic model training using multi-level corpus data augmentation
CN112151018A (en)*2019-06-102020-12-29阿里巴巴集团控股有限公司 Voice evaluation and voice recognition method, device, equipment and storage medium
CN113421463A (en)*2021-06-152021-09-21读书郎教育科技有限公司System and method for assisting in reciting lessons
CN113506572A (en)*2021-07-082021-10-15东北师范大学Portable real-time feedback language learning system
CN113808590A (en)*2021-07-272021-12-17深圳市星范儿文化科技有限公司 Recitation processing method, device and readable storage medium based on intelligent reading booth

Citations (21)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN1664923A (en)*2005-03-282005-09-07何宏山Internet voice stenography method
CN101146292A (en)*2007-10-262008-03-19中兴通讯股份有限公司A method and mobile phone device for voice SMS editing
CN101281534A (en)*2008-05-282008-10-08叶睿智Method for searching multimedia resource based on audio content retrieval
CN101739870A (en)*2009-12-032010-06-16深圳先进技术研究院Interactive language learning system and method
CN102063903A (en)*2010-09-252011-05-18中国科学院深圳先进技术研究院Speech interactive training system and speech interactive training method
CN102426015A (en)*2011-09-062012-04-25深圳市凯立德科技股份有限公司Search method of navigation system interest points, and position service terminal
CN102456344A (en)*2010-10-222012-05-16中国电信股份有限公司System and method for analyzing customer behavior characteristic based on speech recognition technique
CN102543082A (en)*2012-01-192012-07-04北京赛德斯汽车信息技术有限公司Voice operation method for in-vehicle information service system adopting natural language and voice operation system
CN102568475A (en)*2011-12-312012-07-11安徽科大讯飞信息科技股份有限公司System and method for assessing proficiency in Putonghua
CN102833633A (en)*2012-09-042012-12-19深圳创维-Rgb电子有限公司System and method for controlling television voice
CN104318921A (en)*2014-11-062015-01-28科大讯飞股份有限公司Voice section segmentation detection method and system and spoken language detecting and evaluating method and system
CN104464755A (en)*2014-12-022015-03-25科大讯飞股份有限公司Voice evaluation method and device
CN104732977A (en)*2015-03-092015-06-24广东外语外贸大学On-line spoken language pronunciation quality evaluation method and system
CN105302795A (en)*2015-11-112016-02-03河海大学Chinese text verification system and method based on Chinese vague pronunciation and voice recognition
CN105374356A (en)*2014-08-292016-03-02株式会社理光Speech recognition method, speech assessment method, speech recognition system, and speech assessment system
CN105845138A (en)*2016-03-252016-08-10乐视控股(北京)有限公司Voice signal processing method and apparatus
US20160306605A1 (en)*2000-01-312016-10-20Callahan Cellular L.L.C.Apparatus and methods of delivering music and information
CN106059895A (en)*2016-04-252016-10-26上海云睦网络科技有限公司Collaborative task generation method, apparatus and system
CN106157974A (en)*2015-04-072016-11-23富士通株式会社Text recites quality assessment device and method
CN106652622A (en)*2017-02-072017-05-10广东小天才科技有限公司Text training method and device
CN106686223A (en)*2016-12-192017-05-17中国科学院计算技术研究所 Auxiliary dialogue system, method and smart phone for deaf-mute and normal people

Patent Citations (21)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20160306605A1 (en)*2000-01-312016-10-20Callahan Cellular L.L.C.Apparatus and methods of delivering music and information
CN1664923A (en)*2005-03-282005-09-07何宏山Internet voice stenography method
CN101146292A (en)*2007-10-262008-03-19中兴通讯股份有限公司A method and mobile phone device for voice SMS editing
CN101281534A (en)*2008-05-282008-10-08叶睿智Method for searching multimedia resource based on audio content retrieval
CN101739870A (en)*2009-12-032010-06-16深圳先进技术研究院Interactive language learning system and method
CN102063903A (en)*2010-09-252011-05-18中国科学院深圳先进技术研究院Speech interactive training system and speech interactive training method
CN102456344A (en)*2010-10-222012-05-16中国电信股份有限公司System and method for analyzing customer behavior characteristic based on speech recognition technique
CN102426015A (en)*2011-09-062012-04-25深圳市凯立德科技股份有限公司Search method of navigation system interest points, and position service terminal
CN102568475A (en)*2011-12-312012-07-11安徽科大讯飞信息科技股份有限公司System and method for assessing proficiency in Putonghua
CN102543082A (en)*2012-01-192012-07-04北京赛德斯汽车信息技术有限公司Voice operation method for in-vehicle information service system adopting natural language and voice operation system
CN102833633A (en)*2012-09-042012-12-19深圳创维-Rgb电子有限公司System and method for controlling television voice
CN105374356A (en)*2014-08-292016-03-02株式会社理光Speech recognition method, speech assessment method, speech recognition system, and speech assessment system
CN104318921A (en)*2014-11-062015-01-28科大讯飞股份有限公司Voice section segmentation detection method and system and spoken language detecting and evaluating method and system
CN104464755A (en)*2014-12-022015-03-25科大讯飞股份有限公司Voice evaluation method and device
CN104732977A (en)*2015-03-092015-06-24广东外语外贸大学On-line spoken language pronunciation quality evaluation method and system
CN106157974A (en)*2015-04-072016-11-23富士通株式会社Text recites quality assessment device and method
CN105302795A (en)*2015-11-112016-02-03河海大学Chinese text verification system and method based on Chinese vague pronunciation and voice recognition
CN105845138A (en)*2016-03-252016-08-10乐视控股(北京)有限公司Voice signal processing method and apparatus
CN106059895A (en)*2016-04-252016-10-26上海云睦网络科技有限公司Collaborative task generation method, apparatus and system
CN106686223A (en)*2016-12-192017-05-17中国科学院计算技术研究所 Auxiliary dialogue system, method and smart phone for deaf-mute and normal people
CN106652622A (en)*2017-02-072017-05-10广东小天才科技有限公司Text training method and device

Cited By (30)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN108320734A (en)*2017-12-292018-07-24安徽科大讯飞医疗信息技术有限公司Audio signal processing method and device, storage medium, electronic equipment
CN108231090A (en)*2018-01-022018-06-29深圳市酷开网络科技有限公司Text reading level appraisal procedure, device and computer readable storage medium
CN110010123A (en)*2018-01-162019-07-12上海异构网络科技有限公司English phonetic word pronunciation learning evaluation system and method
CN108389440A (en)*2018-03-152018-08-10广东小天才科技有限公司Voice playing method and device based on microphone and voice playing equipment
CN110309350A (en)*2018-03-212019-10-08腾讯科技(深圳)有限公司Recite processing method, system, device, medium and the electronic equipment of task
CN110309350B (en)*2018-03-212023-09-01腾讯科技(深圳)有限公司Processing method, system, device, medium and electronic equipment for recitation tasks
CN110322895A (en)*2018-03-272019-10-11亿度慧达教育科技(北京)有限公司Speech evaluating method and computer storage medium
CN110706710A (en)*2018-06-252020-01-17普天信息技术有限公司Voice recognition method and device, electronic equipment and storage medium
CN109086387A (en)*2018-07-262018-12-25上海慧子视听科技有限公司A kind of audio stream methods of marking, device, equipment and storage medium
CN111640342A (en)*2018-09-042020-09-08北京猿力教育科技有限公司Online examination method and device, computing device and storage medium
CN110085210A (en)*2019-03-152019-08-02平安科技(深圳)有限公司Interactive information test method, device, computer equipment and storage medium
CN110085210B (en)*2019-03-152023-10-13平安科技(深圳)有限公司Interactive information testing method and device, computer equipment and storage medium
CN110010157A (en)*2019-03-272019-07-12广东小天才科技有限公司 Test method, device, equipment and storage medium
CN111833850B (en)*2019-04-182025-05-27霍尼韦尔国际公司 Method and system for cockpit speech recognition acoustic model training using multi-stage corpus data augmentation
CN111833850A (en)*2019-04-182020-10-27霍尼韦尔国际公司 Method and system for cockpit speech recognition acoustic model training using multi-level corpus data augmentation
CN110060663A (en)*2019-04-282019-07-26北京云迹科技有限公司A kind of method, apparatus and system of answer service
CN110148413A (en)*2019-05-212019-08-20科大讯飞股份有限公司Speech evaluating method and relevant apparatus
CN110148413B (en)*2019-05-212021-10-08科大讯飞股份有限公司Voice evaluation method and related device
CN110176235A (en)*2019-05-232019-08-27腾讯科技(深圳)有限公司Methods of exhibiting, device, storage medium and the computer equipment of speech recognition text
CN110310086A (en)*2019-06-062019-10-08安徽淘云科技有限公司Auxiliary recites based reminding method, equipment and storage medium
CN112151018B (en)*2019-06-102024-10-29阿里巴巴集团控股有限公司Speech evaluation and speech recognition method, device, equipment and storage medium
CN112151018A (en)*2019-06-102020-12-29阿里巴巴集团控股有限公司 Voice evaluation and voice recognition method, device, equipment and storage medium
CN110349583A (en)*2019-07-152019-10-18高磊A kind of Game education method and system based on speech recognition
CN111128181A (en)*2019-12-092020-05-08科大讯飞股份有限公司Recitation question evaluation method, device and equipment
CN111383630A (en)*2020-03-042020-07-07广州优谷信息技术有限公司Text recitation evaluation method and device and storage medium
CN111444679B (en)*2020-03-272024-05-24北京小米松果电子有限公司Poem generation method and device, electronic equipment and storage medium
CN111444679A (en)*2020-03-272020-07-24北京小米松果电子有限公司Poetry generation method and device, electronic equipment and storage medium
CN113421463A (en)*2021-06-152021-09-21读书郎教育科技有限公司System and method for assisting in reciting lessons
CN113506572A (en)*2021-07-082021-10-15东北师范大学Portable real-time feedback language learning system
CN113808590A (en)*2021-07-272021-12-17深圳市星范儿文化科技有限公司 Recitation processing method, device and readable storage medium based on intelligent reading booth

Similar Documents

PublicationPublication DateTitle
CN107316638A (en)A kind of poem recites evaluating method and system, a kind of terminal and storage medium
CN109949783B (en) Song synthesis method and system
CN108847215B (en)Method and device for voice synthesis based on user timbre
CN113393828B (en) A speech synthesis model training method, speech synthesis method and device
EP3803846B1 (en)Autonomous generation of melody
CN107039050B (en)Automatic testing method and device for voice recognition system to be tested
CN108242234B (en)Speech recognition model generation method, speech recognition model generation device, storage medium, and electronic device
CN102664016B (en)Singing evaluation method and system
CN111433847B (en) Voice conversion method and training method, intelligent device and storage medium
CN107871496B (en)Speech recognition method and device
CN105869641A (en)Speech recognition device and speech recognition method
CN107146497A (en) A Grading System for Piano Examination
CN108206027A (en)A kind of audio quality evaluation method and system
CN112992109B (en)Auxiliary singing system, auxiliary singing method and non-transient computer readable recording medium
CN103594087A (en)Method and system for improving oral evaluation performance
CN104505103B (en)Voice quality assessment equipment, method and system
CN112908308B (en)Audio processing method, device, equipment and medium
CN111916054A (en)Lip-based voice generation method, device and system and storage medium
CN110379411A (en)For the phoneme synthesizing method and device of target speaker
CN110019962A (en)A kind of generation method and device of video official documents and correspondence information
CN114299918A (en) Acoustic model training and speech synthesis method, device and system, and storage medium
CN113870897A (en)Audio data teaching evaluation method and device, equipment, medium and product thereof
CN112863476A (en)Method and device for constructing personalized speech synthesis model, method and device for speech synthesis and testing
CN100589180C (en) A Speech Recognition Approach Using Multimodal Variational Derivation of Switched State-Space Models
CN105895079A (en)Voice data processing method and device

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
CB02Change of applicant information

Address after:Units F01-03 and 05-10 on the 6th floor of No.1 Building, No.8 Courtyard, Guangshun South Street, Chaoyang District, Beijing

Applicant after:Beijing Ape Power Future Technology Co., Ltd.

Address before:Room A116, Floor 2, 88 Xiangshan Road, Haidian District, Beijing

Applicant before:Beijing chalk Future Technology Co., Ltd.

CB02Change of applicant information
TA01Transfer of patent application right

Effective date of registration:20200506

Address after:100102 unit F01, 5th floor and unit 04, F01, 6th floor, building 1, yard 8, Guangshun South Street, Chaoyang District, Beijing

Applicant after:Beijing ape force Education Technology Co., Ltd

Address before:Units F01-03 and 05-10 on the 6th floor of No.1 Building, No.8 Courtyard, Guangshun South Street, Chaoyang District, Beijing

Applicant before:Beijing Ape Power Future Technology Co.,Ltd.

TA01Transfer of patent application right
RJ01Rejection of invention patent application after publication

Application publication date:20171103

RJ01Rejection of invention patent application after publication

[8]ページ先頭

©2009-2025 Movatter.jp