Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
The following detailed description of specific implementations of the present invention is provided in conjunction with specific embodiments:
the first embodiment is as follows:
fig. 1 shows an implementation flow of a scene learning content obtaining method provided in a first embodiment of the present invention, and for convenience of description, only parts related to the first embodiment of the present invention are shown, which are detailed as follows:
in step S101, when the microphone apparatus receives a scene learning content acquisition request input by a user, sound in a scene where the user is located is collected and identified.
In the embodiment of the invention, the microphone device can be a wireless microphone, and can also be a mobile device which is convenient for a user to carry and is provided with a microphone, such as a smart watch, a mobile phone, a tablet computer or a learning machine. When a user wants to learn language knowledge related to his/her own scene, for example, learning english knowledge related to ordering and purchase orders at the time of ordering, learning english knowledge related to cut prices at the time of shopping, and sending a scene learning content acquisition request to the microphone device.
Preferably, the reminding time for acquiring the scene learning content is preset, and when the current time is detected to be the reminding time, the user is reminded to send the scene learning content acquisition request, so that the user experience is effectively improved. For example, the user may set the time of three meals a day as the reminding time of the scene learning content acquisition to acquire the scene vocabulary during eating, and may also set the time of shopping one hour before the premise of shopping as the reminding time of the scene learning content acquisition to acquire the scene vocabulary during shopping.
In the embodiment of the invention, when receiving a scene learning content acquisition request of a user, the microphone device collects the sound in the scene where the user is located and identifies the collected sound. The collected sound can be identified through a preset voice identification algorithm or a voice identification chip.
Preferably, a plurality of sounds, such as machine running noise, human walking footsteps, collision sound of objects, and the like, may exist in the daily scene, and the collected sounds are denoised before being recognized, so as to improve the accuracy of subsequent sound recognition.
In step S102, the microphone device extracts words from the recognition result of the sound to obtain scene words corresponding to the scene where the user is located, and sends the scene words to the mobile device of the user.
In the embodiment of the invention, the user vocabulary in the preset user dictionary library can be matched with the voice recognition result, the user vocabulary appearing in the recognition result is obtained according to the matching result of the user vocabulary in the user dictionary library and the recognition result, the user vocabulary with the frequency exceeding the preset frequency threshold value in the voice recognition result is set as the scene vocabulary corresponding to the scene where the user is located, the scene vocabulary can be considered to have certain representativeness for the scene where the user is located, for example, the vocabularies such as 'good looking', 'discount', 'fit' and the like have higher frequency in the shopping scene, and the characteristics of the shopping scene can be obviously embodied.
In the embodiment of the present invention, the user mobile device may be a mobile device such as a mobile phone, a tablet computer, a learning machine, and the like. Preferably, the microphone device extracts words from the recognition result of the sound, when the extracted scene words are sent to the user mobile device, the user mobile device detects a wireless signal (bluetooth signal or Wi-Fi signal) of the microphone device, when the wireless signal of the microphone device is detected, a wireless connection request is sent to the microphone device, and after the wireless connection between the user mobile device and the microphone device is successful, a scene word acquisition request is sent to the microphone device, so that the intelligent degree and efficiency of scene learning content acquisition are effectively improved.
Preferably, a scene vocabulary library corresponding to a scene where the user is located is established, scene vocabularies extracted from a voice recognition result are stored in the scene vocabulary library, and the scene vocabularies in the scene vocabulary library are sent to the user mobile equipment, so that different scene vocabulary libraries are established for different scenes, and the accuracy of acquiring the subsequent scene learning content is effectively improved.
Preferably, when the user dictionary base is constructed, the personal information of the user, such as the age of the user, the grade of the user, the occupation of the user, the hobbies of the user, the language learning stage of the user and the like, is acquired, words suitable for the user are collected according to the personal information of the user, and the words form the user dictionary base, so that the user experience is effectively improved. By way of example, when the user is a pupil and just begins to learn English, simple words of interest to the pupil are collected, such as the names of various toys, the names of various fruits, and so forth.
In step S103, the user mobile device generates scene learning content corresponding to the scene vocabulary when receiving the scene vocabulary.
In the embodiment of the invention, when receiving the scene vocabulary sent by the microphone device, the user mobile device takes the scene vocabulary as the key word, searches the learning contents associated with the key word in a preset language learning library or a preset language learning platform, and sets the learning contents as the scene learning contents corresponding to the scene vocabulary. Preferably, the learning content associated with the keyword includes vocabulary translation, grammar sentence pattern, scene dialogue, language course and the like corresponding to the keyword, thereby providing rich learning resources for the user in a targeted manner.
Optionally, when the user mobile device and the microphone device are the same device (for example, the user mobile device and the microphone device are the same tablet computer), the scene learning content corresponding to the scene vocabulary is generated by the microphone device.
In the embodiment of the invention, the microphone equipment collects the sound of the scene where the user is located, identifies the sound content, extracts the scene vocabulary corresponding to the scene where the user is located from the sound content, sends the scene vocabulary to the user mobile equipment, and the user mobile equipment generates the scene learning content corresponding to the scene vocabulary, thereby providing practical and rich scene learning content for the user by combining the user scene, effectively improving the proximity degree of the learning content and the scene where the user is located, further helping the user improve the learning efficiency and improving the user experience.
Example two:
fig. 2 shows an implementation flow of a scene learning content obtaining method provided by the second embodiment of the present invention, and for convenience of description, only the parts related to the second embodiment of the present invention are shown, which are detailed as follows:
in S201, when the microphone device receives a scene learning content acquisition request input by a user, sound in a scene where the user is located is collected and identified.
In the embodiment of the invention, when a user wants to learn the language knowledge related to the scene where the user is, for example, learn the english knowledge related to ordering and buying orders at the time of ordering, learn the english knowledge related to cutting prices at the time of shopping, and send a scene learning content acquisition request to the microphone device. When receiving a scene learning content acquisition request of a user, the microphone device collects sound in a scene where the user is located and identifies the collected sound.
In step S202, the microphone device extracts words from the recognition result of the sound to obtain scene words corresponding to the scene where the user is located, and sends the scene words to the mobile device of the user.
In the embodiment of the invention, the user vocabulary in the preset user dictionary library can be matched with the voice recognition result, the user vocabulary appearing in the recognition result can be obtained according to the matching result of the user vocabulary in the user dictionary library and the recognition result, the user vocabulary with the frequency exceeding the preset frequency threshold in the voice recognition result is set as the scene vocabulary corresponding to the scene where the user is located, the scene vocabulary can be considered to have certain representativeness for the scene where the user is located, for example, the vocabularies such as 'good looking', 'discount', 'fit' and the like have higher frequency in the shopping scene, and the characteristics of the shopping scene can be obviously embodied.
In the embodiment of the present invention, the user mobile device may be a mobile device such as a mobile phone, a tablet computer, a learning machine, and the like. Preferably, the microphone device extracts words from the recognition result of the sound, and when the extracted scene words are sent to the user mobile device, the user mobile device detects a wireless signal (bluetooth signal or Wi-Fi signal) of the microphone device, and when the wireless signal of the microphone device is detected, a wireless connection request may be sent to the microphone device, and after the wireless connection between the user mobile device and the microphone device is successful, a scene word acquisition request is sent to the microphone device, thereby effectively improving the intelligent degree and efficiency of scene learning content acquisition.
In step S203, the user mobile device identifies a scene type corresponding to the scene vocabulary according to the scene vocabulary and a preset typical scene vocabulary library.
In the embodiment of the present invention, the user mobile device may match the scene vocabulary with the vocabulary in the preset typical scene vocabulary library to identify the scene type corresponding to the scene vocabulary. The scene type can be a eating scene, a shopping scene, a learning scene and the like, and the typical scene vocabulary library stores vocabularies which are easy to be bored by a user in different scene types, such as vocabularies of 'how much money', 'whether big codes exist' and the like in the shopping scene.
In step S204, a preset language learning library or a language learning platform searches for learning content corresponding to a scene type, and sets the learning content corresponding to the scene type as scene learning content corresponding to a scene vocabulary.
In the embodiment of the invention, after determining the scene type corresponding to the scene vocabulary, the user mobile device can search the learning content corresponding to the scene type in the language learning library or the language learning platform by taking the scene type as the keyword, and set the learning content as the scene learning content corresponding to the scene vocabulary. Preferably, the learning content corresponding to the scene type includes vocabulary translation, word pronunciation, scene dialogue, grammar sentence pattern, language course and the like corresponding to the scene type, so as to provide rich learning resources for the user in a targeted manner.
Optionally, when the user mobile device and the microphone device are the same device, the microphone device generates scene learning content corresponding to the scene vocabulary.
In the embodiment of the invention, the microphone equipment collects the sound of the scene where the user is located, identifies the sound content, extracts the scene vocabulary corresponding to the scene where the user is located from the sound content, sends the scene vocabulary to the user mobile equipment, and the user mobile equipment generates the scene learning content corresponding to the scene vocabulary, thereby providing practical and rich scene learning content for the user by combining the user scene, effectively improving the proximity degree of the learning content and the scene where the user is located, further helping the user improve the learning efficiency and improving the user experience.
Example three:
fig. 3 shows a structure of a scene learning content acquiring apparatus according to a third embodiment of the present invention, and for convenience of description, only a part related to the third embodiment of the present invention is shown, where the structure includes:
the collecting and identifyingunit 31 is configured to collect and identify sound in a scene where the user is located when the microphone device receives a scene learning content acquisition request input by the user.
In the embodiment of the invention, the microphone device can be a wireless microphone, and can also be a mobile device which is convenient for a user to carry and is provided with a microphone, such as a smart watch, a mobile phone, a tablet computer or a learning machine. When a user wants to learn language knowledge related to his/her own scene, for example, learning english knowledge related to ordering and purchase orders at the time of ordering, learning english knowledge related to cut prices at the time of shopping, and sending a scene learning content acquisition request to the microphone device.
Preferably, the reminding time for acquiring the scene learning content is preset, and when the current time is detected to be the reminding time, the user is reminded to send the scene learning content acquisition request, so that the user experience is effectively improved. For example, the user may set the time of three meals a day as the reminding time of the scene learning content acquisition to acquire the scene vocabulary during eating, and may also set the time of shopping one hour before the premise of shopping as the reminding time of the scene learning content acquisition to acquire the scene vocabulary during shopping.
In the embodiment of the invention, when receiving a scene learning content acquisition request of a user, the microphone device collects the sound in the scene where the user is located and identifies the collected sound. The collected sound can be identified through a preset voice identification algorithm or a voice identification chip.
Preferably, a plurality of sounds, such as machine running noise, human walking footsteps, collision sound of objects, and the like, may exist in the daily scene, and the collected sounds are denoised before being recognized, so as to improve the accuracy of subsequent sound recognition.
And thevocabulary extracting unit 32 is used for extracting the vocabulary of the voice recognition result by the microphone equipment to obtain the scene vocabulary corresponding to the scene where the user is located, and sending the scene vocabulary to the mobile equipment of the user.
In the embodiment of the invention, the user vocabulary in the preset user dictionary library can be matched with the voice recognition result, the user vocabulary appearing in the recognition result is obtained according to the matching result of the user vocabulary in the user dictionary library and the recognition result, the user vocabulary with the frequency exceeding the preset frequency threshold value in the voice recognition result is set as the scene vocabulary corresponding to the scene where the user is located, the scene vocabulary can be considered to have certain representativeness for the scene where the user is located, for example, the vocabularies such as 'good looking', 'discount', 'fit' and the like have higher frequency in the shopping scene, and the characteristics of the shopping scene can be obviously embodied.
In the embodiment of the present invention, the user mobile device may be a mobile device such as a mobile phone, a tablet computer, a learning machine, and the like. Preferably, the microphone device extracts words from the recognition result of the sound, when the extracted scene words are sent to the user mobile device, the user mobile device detects a wireless signal (bluetooth signal or Wi-Fi signal) of the microphone device, when the wireless signal of the microphone device is detected, a wireless connection request is sent to the microphone device, and after the wireless connection between the user mobile device and the microphone device is successful, a scene word acquisition request is sent to the microphone device, so that the intelligent degree and efficiency of scene learning content acquisition are effectively improved.
Preferably, a scene vocabulary library corresponding to a scene where the user is located is established, scene vocabularies extracted from a voice recognition result are stored in the scene vocabulary library, and the scene vocabularies in the scene vocabulary library are sent to the user mobile equipment, so that different scene vocabulary libraries are established for different scenes, and the accuracy of acquiring the subsequent scene learning content is effectively improved.
Preferably, when the user dictionary base is constructed, the personal information of the user, such as the age of the user, the grade of the user, the occupation of the user, the hobbies of the user, the language learning stage of the user and the like, is acquired, words suitable for the user are collected according to the personal information of the user, and the words form the user dictionary base, so that the user experience is effectively improved. By way of example, when the user is a pupil and just begins to learn English, simple words of interest to the pupil are collected, such as the names of various toys, the names of various fruits, and so forth.
The learningcontent generating unit 33 is configured to generate scene learning content corresponding to the scene vocabulary when the user mobile device receives the scene vocabulary.
In the embodiment of the invention, when receiving the scene vocabulary sent by the microphone device, the user mobile device takes the scene vocabulary as the key word, searches the learning contents associated with the key word in a preset language learning library or a preset language learning platform, and sets the learning contents as the scene learning contents corresponding to the scene vocabulary. Preferably, the learning content associated with the keyword includes vocabulary translation, grammar sentence pattern, scene dialogue, language course and the like corresponding to the keyword, thereby providing rich learning resources for the user in a targeted manner.
Optionally, when the user mobile device and the microphone device are the same device (for example, the user mobile device and the microphone device are the same tablet computer), the scene learning content corresponding to the scene vocabulary is generated by the microphone device.
In the embodiment of the invention, the microphone equipment collects the sound of the scene where the user is located, identifies the sound content, extracts the scene vocabulary corresponding to the scene where the user is located from the sound content, sends the scene vocabulary to the user mobile equipment, and the user mobile equipment generates the scene learning content corresponding to the scene vocabulary, thereby providing practical and rich scene learning content for the user by combining the user scene, effectively improving the proximity degree of the learning content and the scene where the user is located, further helping the user improve the learning efficiency and improving the user experience.
Example four:
fig. 4 shows a structure of a scene learning content acquiring apparatus according to a fourth embodiment of the present invention, and for convenience of description, only a part related to the embodiment of the present invention is shown, where the scene learning content acquiring apparatus includes:
and the acquisition andidentification unit 41 is used for acquiring and identifying the sound in the scene where the user is located when the microphone device receives the scene learning content acquisition request input by the user.
In the embodiment of the invention, when a user wants to learn the language knowledge related to the scene where the user is, a scene learning content acquisition request can be sent to the microphone device. When receiving a scene learning content acquisition request of a user, the microphone device collects sound in a scene where the user is located and identifies the collected sound.
And thevocabulary extracting unit 42 is used for the microphone device to extract the vocabulary of the voice recognition result so as to obtain the scene vocabulary corresponding to the scene where the user is located, and sending the scene vocabulary to the mobile device of the user.
In the embodiment of the invention, the user vocabulary in the preset user dictionary library can be matched with the recognition result of the voice, the user vocabulary appearing in the recognition result can be obtained according to the matching result of the user vocabulary in the user dictionary library and the recognition result, the user vocabulary with the frequency exceeding the preset frequency threshold value in the recognition result of the voice is set as the scene vocabulary corresponding to the scene where the user is located, and the scene vocabulary can be considered to have certain representativeness to the scene where the user is located.
In the embodiment of the present invention, the user mobile device may be a mobile device such as a mobile phone, a tablet computer, a learning machine, and the like. Preferably, the microphone device extracts words from the recognition result of the sound, and when the extracted scene words are sent to the user mobile device, the user mobile device detects a wireless signal (bluetooth signal or Wi-Fi signal) of the microphone device, and when the wireless signal of the microphone device is detected, a wireless connection request may be sent to the microphone device, and after the wireless connection between the user mobile device and the microphone device is successful, a scene word acquisition request is sent to the microphone device, thereby effectively improving the intelligent degree and efficiency of scene learning content acquisition.
And thetype recognition unit 43 is used for recognizing the scene type corresponding to the scene vocabulary by the mobile equipment of the user according to the scene vocabulary and a preset typical scene vocabulary library.
In the embodiment of the present invention, the user mobile device may match the scene vocabulary with the vocabulary in the preset typical scene vocabulary library to identify the scene type corresponding to the scene vocabulary. The scene type can be a eating scene, a shopping scene, a learning scene and the like, and the typical scene vocabulary library stores vocabularies which are easy to be bored by a user in different scene types, such as vocabularies of 'how much money', 'whether big codes exist' and the like in the shopping scene.
And acontent searching unit 44, configured to search a preset language learning library or a language learning platform for learning content corresponding to a scene type, and set the learning content corresponding to the scene type as scene learning content corresponding to a scene vocabulary.
In the embodiment of the invention, after determining the scene type corresponding to the scene vocabulary, the user mobile device can search the learning content corresponding to the scene type in the language learning library or the language learning platform by taking the scene type as the keyword, and set the learning content as the scene learning content corresponding to the scene vocabulary. Preferably, the learning content corresponding to the scene type includes vocabulary translation, word pronunciation, scene dialogue, grammar sentence pattern, language course and the like corresponding to the scene type, so as to provide rich learning resources for the user in a targeted manner.
Optionally, when the user mobile device and the microphone device are the same device, the microphone device generates scene learning content corresponding to the scene vocabulary.
Preferably, as shown in fig. 5, thevocabulary extracting unit 42 includes:
thevocabulary matching unit 521 is used for the microphone equipment to match the user vocabulary in the preset user dictionary library with the recognition result so as to obtain the user vocabulary in the recognition result; and
and thefrequency comparing unit 522 is configured to set the user vocabulary with the occurrence frequency exceeding the preset frequency threshold in the recognition result as the scene vocabulary corresponding to the scene where the user is located.
In the embodiment of the invention, the microphone equipment collects the sound of the scene where the user is located, identifies the sound content, extracts the scene vocabulary corresponding to the scene where the user is located from the sound content, sends the scene vocabulary to the user mobile equipment, and the user mobile equipment generates the scene learning content corresponding to the scene vocabulary, thereby providing practical and rich scene learning content for the user by combining the user scene, effectively improving the proximity degree of the learning content and the scene where the user is located, further helping the user improve the learning efficiency and improving the user experience.
In the embodiment of the present invention, each unit of a scene learning content acquiring apparatus may be implemented by a corresponding hardware or software unit, and each unit may be an independent software or hardware unit, or may be integrated into a software or hardware unit, which is not limited herein.
Example five:
fig. 6 shows a structure of a learning apparatus provided in a fifth embodiment of the present invention, and for convenience of explanation, only a part related to the embodiment of the present invention is shown.
Thelearning device 6 of an embodiment of the present invention comprises aprocessor 60, amemory 61 and acomputer program 62 stored in thememory 61 and executable on theprocessor 60. Theprocessor 60, when executing thecomputer program 62, implements the steps in the various method embodiments described above, such as steps S101 to S103 shown in fig. 1. Alternatively, theprocessor 60, when executing thecomputer program 62, implements the functions of the units in the above-described device embodiments, such as the functions of theunits 31 to 33 shown in fig. 3.
In the embodiment of the invention, the microphone equipment collects the sound of the scene where the user is located, identifies the sound content, extracts the scene vocabulary corresponding to the scene where the user is located from the sound content, sends the scene vocabulary to the user mobile equipment, and the user mobile equipment generates the scene learning content corresponding to the scene vocabulary, thereby providing practical and rich scene learning content for the user by combining the user scene, effectively improving the proximity degree of the learning content and the scene where the user is located, further helping the user improve the learning efficiency and improving the user experience.
Example six:
in an embodiment of the present invention, a computer-readable storage medium is provided, which stores a computer program that, when executed by a processor, implements the steps in the various method embodiments described above, e.g., steps S101 to S103 shown in fig. 1. Alternatively, the computer program may be adapted to perform the functions of the units of the above-described device embodiments, such as the functions of theunits 31 to 33 shown in fig. 3, when executed by the processor.
In the embodiment of the invention, the microphone equipment collects the sound of the scene where the user is located, identifies the sound content, extracts the scene vocabulary corresponding to the scene where the user is located from the sound content, sends the scene vocabulary to the user mobile equipment, and the user mobile equipment generates the scene learning content corresponding to the scene vocabulary, thereby providing practical and rich scene learning content for the user by combining the user scene, effectively improving the proximity degree of the learning content and the scene where the user is located, further helping the user improve the learning efficiency and improving the user experience
The computer readable storage medium of the embodiments of the present invention may include any entity or device capable of carrying computer program code, a recording medium, such as a ROM/RAM, a magnetic disk, an optical disk, a flash memory, or the like.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents and improvements made within the spirit and principle of the present invention are intended to be included within the scope of the present invention.