Disclosure of Invention
In view of the above problems in the prior art, a system and a method for evaluating a singing song are provided to facilitate evaluation of the singing song.
The specific technical scheme is as follows:
the invention comprises an evaluation system for singing songs, which is characterized by comprising the following steps:
the processing module is used for processing the type of each original song to extract a first characteristic parameter of the original song;
the training module is connected with the processing module and is used for respectively training the sound speed and the sound intensity change of each segmented audio frequency of each original song so as to respectively extract second characteristic parameters of the sound speed and the sound intensity change of each segmented audio frequency;
the algorithm module is respectively connected with the processing module and the training module and is used for calculating a final score according to the first characteristic parameter, the second characteristic parameter and a third characteristic parameter of a corresponding time period of a singing audio;
and the evaluation module is connected with the algorithm module and provides corresponding evaluation according to the final score.
Preferably, the processing module comprises:
a collecting unit for collecting each original song;
the classification unit is connected with the collection unit and is used for classifying the type of each collected original song;
the first training unit is connected with the classification unit and used for training the classified original song;
the first extraction unit is connected with the first training unit and is used for extracting a first characteristic parameter according to a training result of the first training unit;
and the first computing unit is connected with the first extracting unit and is used for computing the first characteristic parameter and the third characteristic parameter of the singing audio in a corresponding time period so as to obtain a singing type matching rate.
Preferably, the training module comprises:
a segmentation unit, configured to segment each original song according to a beat of the music score to form the segmented audio;
and the second training unit is connected to the segmenting unit and is used for respectively training the sound speed and the sound intensity change in the segmented audio so as to respectively extract the second characteristic parameters of the sound speed and the sound intensity change of each segmented audio.
Preferably, the algorithm module comprises:
a first comparing unit, configured to compare the second characteristic parameter of the change of the speed of sound and the sound intensity with the third characteristic parameter of the corresponding time interval of the singing audio, so as to obtain a speed of sound matching rate and a sound intensity change matching rate;
the second calculation unit is used for calculating the pitch value of each beat of the singing audio;
the second comparison unit is connected with the second calculation unit and is used for comparing the pitch value of each beat of the singing audio with the pitch value of each beat of the original song to obtain a correct pitch score;
and the third calculating unit is connected with the second comparing unit and is used for calculating the final score according to the singing type matching rate, the sound speed matching rate, the sound intensity change matching rate and the correct pitch score.
Preferably, the system further comprises a suggestion module, wherein the suggestion module is connected with the evaluation module and is used for providing corresponding suggestions for singers after the evaluations are provided according to the singing type matching rate, the sound speed matching rate, the sound intensity change matching rate and the pitch correct score.
The invention also provides an evaluation method of singing songs, which comprises the evaluation system of the singing songs, and the evaluation method comprises the following steps:
step S1, processing the type of each original song by adopting a processing module to extract a first characteristic parameter of the original song;
step S2, a training module is adopted for respectively training the sound speed and the sound intensity change of each segmented audio of each original song so as to respectively extract second characteristic parameters of the sound speed and the sound intensity change of each segmented audio;
step S3, an algorithm module is adopted for calculating a final score according to the first characteristic parameter, the second characteristic parameter and a third characteristic parameter of a corresponding time interval of a singing audio;
and step S4, adopting an evaluation module for providing corresponding evaluation according to the final score.
Preferably, the step S1 includes:
step S10, a collecting unit is adopted for collecting each original song;
step S11, adopting a classification unit for classifying the type of each collected original song;
step S12, a first training unit is adopted for training the classified original song;
step S13, a first extraction unit is adopted for extracting a first characteristic parameter according to the training result of the first training unit;
step S14, a first calculating unit is adopted to calculate the first characteristic parameter and the third characteristic parameter of the singing audio corresponding time interval, so as to obtain a singing type matching rate.
Preferably, the step S2 includes:
step S20, a segmentation unit is adopted for segmenting each original song according to the beat of the music score to form the segmented audio;
step S21, a second training unit is adopted to train the speed of sound and the change in sound intensity in the segmented audio respectively, so as to extract the second characteristic parameters of the speed of sound and the change in sound intensity of each segmented audio respectively.
Preferably, the step S3 includes:
step S30, a first comparing unit is adopted to compare the second characteristic parameter of the change of the speed of sound and the sound intensity with the third characteristic parameter of the corresponding time interval of the singing audio, so as to obtain a matching rate of speed of sound and a matching rate of change of sound intensity;
step S31, a second calculating unit is adopted for calculating the pitch value of each beat of the singing audio;
step S32, a second comparison unit is adopted for comparing the pitch value of each beat of the singing audio with the pitch value of each beat of the original song to obtain a correct pitch score;
step S33, a third calculating unit is adopted to calculate the final score according to the singing style matching rate, the sound velocity matching rate, the sound intensity variation matching rate, and the pitch correct score.
Preferably, the method further includes step S5, which is to adopt a suggestion module, and after providing evaluations according to the singing genre matching rate, the sound velocity matching rate, the sound intensity variation matching rate, and the pitch correct score, provide corresponding suggestions to the singer.
The technical scheme of the invention has the beneficial effects that: the evaluation system and method for singing songs are characterized in that an algorithm module calculates a final score according to a first characteristic parameter obtained by a processing module, a second characteristic parameter obtained after a training module respectively trains the sound speed and the sound intensity change of segmented audio of each original singing song and a third characteristic parameter of a singing audio corresponding time period, and then evaluates the singing songs learned or examined by students, correspondingly provides customized suggestions for the students, and helps the students improve the singing level.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It should be noted that the embodiments and features of the embodiments may be combined with each other without conflict.
The invention is further described with reference to the following drawings and specific examples, which are not intended to be limiting.
The invention provides an evaluation system for singing songs, which comprises the following steps:
theprocessing module 1 is used for processing the type of each original song to extract a first characteristic parameter of the original song;
thetraining module 2 is connected with theprocessing module 1 and is used for respectively training the sound speed and the sound intensity change of each segmented audio frequency of each original song so as to respectively extract second characteristic parameters of the sound speed and the sound intensity change of each segmented audio frequency;
thealgorithm module 3 is respectively connected with theprocessing module 1 and thetraining module 2 and used for calculating a final score according to the first characteristic parameter, the second characteristic parameter and a third characteristic parameter of a corresponding time period of a singing audio;
and theevaluation module 4 is connected with thealgorithm module 3, and provides corresponding evaluation according to the final score.
With the evaluation system provided above, as shown in fig. 1, firstly, all the original songs are classified by theprocessing module 1, the type tempo of each original song can be obtained as, for example, 2/4, 4/4, 3/4, 3/8, 6/8, 9/8, 2/2, 1/4, and the type speed of each original song can also be obtained as, for example, 50-140 beats per minute (where each 10 beats is a classification), so that the type or style of each original song can be obtained as: the original song is cheerful or fluent or moderate or relaxed or hypertonic or intense or hot and the like, for example, the type beat of the original song 'love filling the world' is 4/4, the type speed is 76(70-79) beats per minute, so that the type or style of the original song is acquired as relaxed, and the original song is further processed to extract a first characteristic parameter of the original song for subsequent calculation and scoring.
Further, thetraining module 2 separately trains the sound speed and the sound intensity change of each segmented audio of each original song, and in this embodiment, the training methods adopted are CNN (Convolutional Neural networks) and ResNet (Residual Neural networks), so as to extract a large number of second characteristic parameters of the sound speed and the sound intensity change of each segmented audio respectively.
Further, the first characteristic parameter obtained by theprocessing module 1, the second characteristic parameter obtained by thetraining module 2 and the third characteristic parameter of the singing audio in the corresponding time period are calculated through thealgorithm module 3, so that the final score of the live singing song is obtained.
Furthermore, theevaluation module 4 gives corresponding evaluation to the singer according to the total score obtained by thealgorithm module 3, so as to help the student improve the singing level.
In addition, the first characteristic parameter, the second characteristic parameter, and the third characteristic parameter in the present invention are all high-dimensional characteristic parameters.
In a preferred embodiment, theprocessing module 1 comprises:
a collectingunit 10, for collecting each original song;
the classifyingunit 11 is connected with the collectingunit 10, and is used for classifying the type of each collected original song;
thefirst training unit 12, thefirst training unit 12 is connected to the classifyingunit 11, and is used for training the classified original song;
thefirst extraction unit 13 is connected with thefirst training unit 12, and a first characteristic parameter is extracted according to the training result of thefirst training unit 12;
and a first calculatingunit 14, where the first calculatingunit 14 is connected to the first extractingunit 13, and is configured to calculate the first characteristic parameter and a third characteristic parameter of a corresponding time period of the singing audio, so as to obtain a singing type matching rate.
Specifically, as shown in fig. 2, theprocessing module 1 first collects all original songs through thecollection unit 10, then classifies the type of each collected original song, and obtains the beat and speed of each song to know the type or style of each original song;
further, thefirst training unit 11 trains the classified original song by using CNN (Convolutional Neural Networks) and ResNet (Residual Neural Networks) algorithms, and then thefirst extraction unit 13 extracts a large number of first characteristic parameters.
Further, the first feature parameter and a third feature parameter of a corresponding time period of the singing audio are subjected to cosine distance calculation through the first calculatingunit 14, so that a singing type matching rate is obtained, and a style suggestion is obtained.
In a preferred embodiment, thetraining module 2 comprises:
asegmentation unit 20, configured to segment each original song according to the beat of the music score to form a segmented audio;
and thesecond training unit 21 is connected to the segmentingunit 20, and is configured to train the change of the sound velocity and the sound intensity in the segmented audio respectively, so as to extract second characteristic parameters of the change of the sound velocity and the sound intensity of each segmented audio respectively.
Specifically, as shown in fig. 3, thetraining module 2 first segments each original song according to the beat of the music score by thesegmentation unit 20 to form each small segmented audio, and then separately trains the sound velocity and the sound intensity variation of each small segmented audio by thesecond training unit 21 to extract a large number of second feature parameters of the sound velocity and the sound intensity variation of each segmented audio.
In a preferred embodiment, thealgorithm module 3 comprises:
a first comparingunit 30, configured to compare the second characteristic parameter of the change of the sound speed and the sound intensity with the third characteristic parameter of the corresponding time interval of the singing audio, so as to obtain a sound speed matching rate and a sound intensity change matching rate;
a second calculatingunit 31, for calculating the pitch value of each beat of the singing audio;
thesecond comparison unit 32 is connected to thesecond calculation unit 31, and is used for comparing the pitch value of each beat of the singing audio with the pitch value of each beat of the original singing song to obtain a correct pitch score;
a third calculatingunit 33, the third calculatingunit 33 is connected to the second comparingunit 32, and is used for calculating a final score according to the singing type matching rate, the sound velocity matching rate, the sound intensity variation matching rate and the pitch correct score.
Specifically, as shown in fig. 4, thealgorithm module 3 first compares the second characteristic parameter of the variation of the sound speed and the sound intensity with the third characteristic parameter of the corresponding time period of the singing audio through thefirst comparison unit 30, so as to obtain the sound speed matching rate and the sound intensity variation matching rate.
Further, the second calculatingunit 31 calculates the pitch value of each beat of the singing audio by using FFT (fast Fourier transform), the second comparingunit 32 compares the pitch value of each beat of the singing audio with the pitch value of each beat of the original song, if the pitch value of each beat of the singing audio is correct, the score is 1, and if the pitch value of each beat of the singing audio is incorrect, the score is 0, so as to calculate the correct pitch score, and the third calculatingunit 33 is used to obtain the final score according to the following formula:
wherein S represents the final score;
MaxS represents the total score;
o represents a sound velocity matching rate;
p represents a sound intensity change matching rate;
n represents the pitch correct score;
x represents the number of bars of the song;
y represents the number of beats of the song;
i represents the score weight of the sound speed and the sound intensity change;
j represents the score weight for the pitch correct score.
In a preferred embodiment, the system further comprises asuggestion module 5, wherein thesuggestion module 5 is connected to theevaluation module 4, and is configured to provide corresponding suggestions to the singer after providing evaluations according to the singing type matching rate, the sound velocity matching rate, the sound intensity change matching rate, and the pitch correct score.
Specifically, as shown in fig. 1, after the evaluation is made according to the singing type matching rate, the sound velocity matching rate, the sound intensity variation matching rate and the pitch correct score, the singer is suggested accordingly, for example, by taking the 18nd bar 2 in the song "beijing welcome you", the audio of the original song should be from low to high, but the singer may sing from high to low by mistake, so that the singer is suggested to pay attention to the mastering of the sound intensity, and thus the 2 nd bar is repeatedly exercised for the 18 nd bar, so that the singing level of the singer is improved.
The invention also provides an evaluation method of singing songs, which comprises the evaluation system of the singing songs, and the evaluation method comprises the following steps:
step S1, aprocessing module 1 is adopted to process the type of each original song so as to extract a first characteristic parameter of the original song;
step S2, atraining module 2 is adopted for respectively training the sound speed and the sound intensity change of each segmented audio of each original song so as to respectively extract second characteristic parameters of the sound speed and the sound intensity change of each segmented audio;
step S3, analgorithm module 3 is adopted for calculating a final score according to the first characteristic parameter, the second characteristic parameter and a third characteristic parameter of a corresponding time interval of a singing audio;
and step S4, adopting anevaluation module 4 for proposing corresponding evaluation according to the final score.
Through the evaluation method provided above, as shown in fig. 5, firstly, theprocessing module 1 is adopted to classify all the original songs, and the type tempo of each original song can be obtained as, for example, 2/4, 4/4, 3/4, 3/8, 6/8, 9/8, 2/2, 1/4, or the type speed of each original song can be obtained as, for example, 50-140 beats per minute (where each 10 beats is a classification), so that the type or style of each original song can be obtained as: the original song is cheerful or fluent or moderate or relaxed or hypertonic or intense or hot and the like, for example, the type beat of the original song 'love filling the world' is 4/4, the type speed is 76(70-79) beats per minute, so that the type or style of the original song is acquired as relaxed, and the original song is further processed to extract a first characteristic parameter of the original song for subsequent calculation and scoring.
Further, thetraining module 2 is adopted to separately train the sound speed and the sound intensity change of each segmented audio of each original song, in this embodiment, the adopted training methods are CNN (Convolutional Neural networks) and ResNet (Residual Neural networks Residual error Network), so as to extract a large number of second characteristic parameters of the sound speed and the sound intensity change of each segmented audio respectively.
Further, thealgorithm module 3 is adopted to calculate the first characteristic parameter obtained by theprocessing module 1, the second characteristic parameter obtained by thetraining module 2 and the third characteristic parameter of the singing audio in the corresponding time period, so as to obtain the final score of the live singing song.
Furthermore, anevaluation module 4 is adopted to give corresponding evaluation to the singer according to the total score obtained by thealgorithm module 3, so as to help the student improve the singing level.
In addition, the first characteristic parameter, the second characteristic parameter, and the third characteristic parameter in the present invention are all high-dimensional characteristic parameters.
In a preferred embodiment, step S1 includes:
step S10, a collectingunit 10 is adopted for collecting each original song;
step S11, aclassification unit 11 is adopted for classifying the type of each original song after collection;
step S12, afirst training unit 12 is adopted for training the classified original song;
step S13, a first extractingunit 13 is adopted to extract a first feature parameter according to the training result of thefirst training unit 12;
step S14, a first calculatingunit 14 is adopted to calculate the first characteristic parameter and the third characteristic parameter of the singing audio corresponding time interval, so as to obtain a singing type matching rate.
Specifically, as shown in fig. 6, in step S1, thecollection unit 10 is first used to collect all the original songs, then the types of each collected original song are classified, and the beat and speed of each song are obtained to know which type or style each original song is;
further, thefirst training unit 11 is adopted to train the classified original song by adopting CNN (Convolutional Neural Networks) and ResNet (Residual Neural Networks) algorithms, and then thefirst extraction unit 13 is adopted to extract a large number of first characteristic parameters.
Further, the first calculatingunit 14 is adopted to perform cosine distance calculation on the first characteristic parameter and the third characteristic parameter of the singing audio in the corresponding time period so as to obtain a singing type matching rate and further obtain a style suggestion.
In a preferred embodiment, step S2 includes:
step S20, asegmentation unit 20 is adopted for segmenting each original song according to the beat of the music score to form segmented audio;
step S21, asecond training unit 21 is adopted to train the variation of the sound velocity and the sound intensity in the segmented audio respectively, so as to extract the second characteristic parameters of the variation of the sound velocity and the sound intensity of each segmented audio respectively.
Specifically, as shown in fig. 7, in step S2, thesegmentation unit 20 is first used to segment each original song according to the tempo of the music score to form each small segmented audio, and thesecond training unit 21 is then used to perform separate training on the sound velocity and the sound intensity variation of each small segmented audio respectively to extract a large number of second feature parameters of the sound velocity and the sound intensity variation of each segmented audio respectively.
In a preferred embodiment, step S3 includes:
step S30, using a first comparingunit 30 for comparing the second characteristic parameter of sound velocity and sound intensity variation with the third characteristic parameter of the corresponding time interval of the singing audio to obtain a sound velocity matching rate and a sound intensity variation matching rate;
step S31, using a second calculatingunit 31 for calculating the pitch value of each beat of the singing audio;
step S32, asecond comparison unit 32 is adopted to compare the pitch value of each beat of the singing audio with the pitch value of each beat of the original singing song to obtain a correct pitch score;
step S33, a third calculatingunit 33 is adopted to calculate the final score according to the singing style matching rate, the sound velocity matching rate, the sound intensity variation matching rate and the pitch correct score.
Specifically, as shown in fig. 8, in step S3, thefirst comparison unit 30 is first adopted to compare the second characteristic parameters of the variation of the sound velocity and the sound intensity with the third characteristic parameters of the corresponding time period of the singing audio, so as to obtain the sound velocity matching rate and the sound intensity variation matching rate.
Further, the second calculatingunit 31 calculates a pitch value of each beat of the singing audio by using FFT (fast Fourier transform), the second comparingunit 32 compares the pitch value of each beat of the singing audio with the pitch value of each beat of the original song, if the pitch value of each beat of the singing audio is correct, the score is 1, and if the pitch value of each beat of the singing audio is incorrect, the score is 0, so that a correct pitch score is calculated, and the third calculatingunit 33 obtains the final score according to the following formula:
wherein S represents the final score;
MaxS represents the total score;
o represents a sound velocity matching rate;
p represents a sound intensity change matching rate;
n represents the pitch correct score;
x represents the number of bars of the song;
y represents the number of beats of the song;
i represents the score weight of the sound speed and the sound intensity change;
j represents the score weight for the pitch correct score.
In a preferred embodiment, the method further comprises a step S5 of using asuggestion module 5 for providing corresponding suggestions to the singer after making evaluations according to the singing type matching rate, the sound velocity matching rate, the sound intensity variation matching rate and the pitch correct score.
Specifically, as shown in fig. 5, after the evaluation is made according to the singing type matching rate, the sound velocity matching rate, the sound intensity variation matching rate and the pitch correct score, thesuggestion module 5 is adopted to correspondingly give suggestions to the singer, for example, by taking the 18 thminor bar 2 in the song "beijing welcome you", the audio of the original sung song should be from low to high, but the singer may possibly sing from high to low, so that the singer can be suggested to pay attention to the mastering of the sound intensity, and thus the 2 nd minor bar 2-time exercise is repeated, so that the singing level of the singer is improved.
While the invention has been described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the invention.