Disclosure of Invention
The embodiment of the invention aims to provide a data processing method, a device, a system and a server based on voiceprint recognition, which solve the technical problem of data repetition in voice data processing and realize accurate recognition and processing of voice data.
In order to solve the above technical problems, embodiments of the present invention provide the following technical solutions:
in a first aspect, an embodiment of the present invention provides a data processing method based on voiceprint recognition, where the method includes:
the method comprises the steps that a user account is created in advance, and the user account comprises basic account information and preset voiceprint information;
acquiring a voice fragment sent by a user side;
according to the voice fragment, recognizing voiceprint information corresponding to the voice fragment;
and matching the user account corresponding to the voiceprint information according to the voiceprint information, and storing the voice fragment to the matched user account.
In some embodiments, the pre-creating a user account comprises:
receiving basic account information and user voice information sent by a user side;
according to the user voice information, identifying voiceprint information in the user voice information;
and taking the identified voiceprint information as preset voiceprint information, storing the preset voiceprint information in association with the basic account information, and registering to generate a user account.
In some embodiments, the matching, according to the voiceprint information, the user account corresponding to the voiceprint information includes:
matching corresponding preset voiceprint information according to the voiceprint information;
and matching the user account corresponding to the voiceprint information according to the preset voiceprint information.
In some embodiments, the acquiring the voice segment sent by the user side specifically includes:
acquiring voice data of a user in one or more application scenes, wherein the application scenes comprise one or more of a recording scene, a meal ordering scene, a singing scene, a speaking scene, a lecture scene, a talking scene, a reading-aloud scene and a reading-after scene, and the voice data comprises at least one voice fragment;
and acquiring each voice segment in the voice data according to the voice data.
In some embodiments, the acquiring the voice segment sent by the user side specifically includes:
acquiring voice data sent by a plurality of users under the same user account, wherein the voice data comprises at least one voice segment;
and acquiring each voice segment in the voice data according to the voice data.
In some embodiments, the matching, according to the voiceprint information, the user account corresponding to the voiceprint information, and storing the voice fragment in the matched user account includes:
according to the voice data, voice print information of each voice segment in the voice data is recognized;
and respectively matching the corresponding user accounts according to the voiceprint information of each voice fragment, and storing each voice fragment to the corresponding user account.
In some embodiments, the acquiring the voice segment sent by the user side specifically includes:
acquiring voice data sent by a plurality of users under a plurality of user accounts, wherein the voice data comprises at least one voice fragment;
and acquiring each voice segment in the voice data according to the voice data.
In a second aspect, an embodiment of the present invention provides a data processing apparatus based on voiceprint recognition, where the apparatus includes:
the system comprises a user account creating unit, a voice print processing unit and a voice print processing unit, wherein the user account creating unit is used for creating a user account in advance, and the user account comprises basic account information and preset voice print information;
the voice fragment acquisition unit is used for acquiring the voice fragments sent by the user side;
the voice print information identification unit is used for identifying voice print information corresponding to the voice fragment according to the voice fragment;
and the user account matching unit is used for matching the user account corresponding to the voiceprint information according to the voiceprint information and storing the voice fragment to the matched user account.
In some embodiments, the user account creation unit is specifically configured to:
receiving basic account information and user voice information sent by a user side;
according to the user voice information, identifying voiceprint information in the user voice information;
and taking the identified voiceprint information as preset voiceprint information, storing the preset voiceprint information in association with the basic account information, and registering to generate a user account.
In some embodiments, the user account matching unit is specifically configured to:
matching corresponding preset voiceprint information according to the voiceprint information;
and matching the user account corresponding to the voiceprint information according to the preset voiceprint information.
In some embodiments, the voice segment obtaining unit is specifically configured to:
acquiring voice data of a user in one or more application scenes, wherein the application scenes comprise one or more of a recording scene, a meal ordering scene and a follow-up reading scene, and the voice data comprises at least one voice fragment;
and acquiring each voice segment in the voice data according to the voice data.
In a third aspect, an embodiment of the present invention provides a server, including:
at least one processor; and the number of the first and second groups,
a memory communicatively coupled to the at least one processor; wherein,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform a voiceprint recognition based data processing method as described above.
In a fourth aspect, an embodiment of the present invention further provides a data processing system based on voiceprint recognition, where the system includes:
the above-mentioned server;
and each user side is in communication connection with the server and is used for inputting user voice information to the server.
In some embodiments, the server further comprises: the voice print recognition module is used for recognizing voice print information in the voice fragment;
the user side includes: the system comprises an input module and a voice acquisition module, wherein the input module is used for inputting basic account information, and the voice acquisition module is used for acquiring voice data of a user.
In a fifth aspect, the embodiment of the present invention further provides a non-transitory computer-readable storage medium, where the computer-readable storage medium stores computer-executable instructions for enabling a server to execute the data processing method based on voiceprint recognition as described above.
The embodiment of the invention has the beneficial effects that: in contrast to the prior art, an embodiment of the present invention provides a data processing method based on voiceprint recognition, including: the method comprises the steps that a user account is created in advance, and the user account comprises basic account information and preset voiceprint information; acquiring a voice fragment sent by a user side; according to the voice fragment, recognizing voiceprint information corresponding to the voice fragment; and matching the user account corresponding to the voiceprint information according to the voiceprint information, and storing the voice fragment to the matched user account. By matching the corresponding user account and storing the corresponding voice data based on the voiceprint information, the embodiment of the invention solves the technical problem of data repetition in the voice data processing, and realizes accurate recognition and processing of the voice data.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It should be noted that, if not conflicted, the various features of the embodiments of the invention may be combined with each other within the scope of protection of the invention. Additionally, while functional block divisions are performed in apparatus schematics, with logical sequences shown in flowcharts, in some cases, steps shown or described may be performed in sequences other than block divisions in apparatus or flowcharts. The terms "first", "second", "third", and the like used in the present invention do not limit data and execution order, but distinguish the same items or similar items having substantially the same function and action.
Before the present invention is explained in detail, terms and expressions referred to in the embodiments of the present invention are explained, and the terms and expressions referred to in the embodiments of the present invention are applied to the following explanations.
1) Voiceprint: voiceprint, a generic term for speech features that are implied in speech and that characterize and identify a speaker, and for speech models that are built based on these features (parameters).
2) Voiceprint characteristics: voiceprint feature, a parameter that is determined from the speaker's speech and that characterizes the personality of the speaker's speech.
3) And (3) voiceprint recognition: voiceprint recognition (VPR), a process of recognizing the speaker corresponding to the voice according to the Voiceprint characteristics of the voice to be recognized.
4) Crowdsourcing: refers to the practice of a company or organization outsourcing work tasks that have been performed by employees to an unspecified (and often large) mass network in a free-form manner.
Referring to fig. 1, fig. 1 is a schematic diagram of an application environment according to an embodiment of the present invention;
as shown in fig. 1, theapplication environment 100 includes: aclient 11 and aserver 12, wherein theclient 11 is connected to theserver 12 through network communication, such as: theserver 12 is connected by means of 2G, 3G, 4G, 5G, local area network, wide area network, wireless broadband, etc. in a communication manner, theuser terminal 11 is configured to provide basic account information for user input, for example: theuser terminal 11 is further used for collecting voice information of a user and sending the voice information to theserver 12, theserver 12 receives the voice information of the user, recognizes voiceprint information according to the voice information, and matches a corresponding user account according to the voiceprint information.
Referring to fig. 2, fig. 2 is a schematic view illustrating an alternative flow chart of a data processing method based on voiceprint recognition according to an embodiment of the present invention;
as shown in fig. 2, the data processing method based on voiceprint recognition includes:
step S21: the method comprises the steps that a user account is created in advance, and the user account comprises basic account information and preset voiceprint information;
specifically, a user account is created in advance and stored on a server, so that the server stores a voice clip corresponding to the user according to account information of the user account, wherein the account information includes basic account information and preset voiceprint information.
Specifically, please refer to fig. 3, fig. 3 is a detailed flowchart of step S21 in fig. 2;
as shown in fig. 3, the pre-creating a user account includes:
step S211: receiving basic account information and user voice information sent by a user side;
specifically, the user uses the user side, for example: inputting personal basic account information by a mobile terminal and a PC terminal, wherein the basic account information comprises: the system comprises a server, a user terminal, a server, a user interface, a voice recording device and a voice recording device, wherein the server is used for recording voice information of the user, the user terminal is used for recording the voice information of the user, and the voice recording device is used for recording the voice information of the user.
Step S212: according to the user voice information, identifying voiceprint information in the user voice information;
specifically, the server identifies the user voice information according to the received user voice information, for example: and carrying out voiceprint feature extraction on the user voice information, and then carrying out mode recognition, thereby obtaining the voiceprint information in the user voice information and storing the voiceprint information.
Step S213: and taking the identified voiceprint information as preset voiceprint information, storing the preset voiceprint information in association with the basic account information, and registering to generate a user account.
Specifically, the recognized voiceprint information is used as preset voiceprint information, the preset voiceprint information is used for matching voiceprint information corresponding to a voice fragment of a user, the server manages and stores the preset voiceprint information and the basic account information and registers the preset voiceprint information and the basic account information to generate a user account, specifically, the server generates an association table, the preset voiceprint information and the basic account information of each user are correspondingly stored through the association table, and when the voice fragment of the user is obtained, the corresponding preset voiceprint information is matched according to the voiceprint information corresponding to the voice fragment, so that the account information of the user is determined. It can be understood that the preset voiceprint information is generated by a server after a user voice message is first sent by a user and is identified according to the user voice message first sent by the user, and the method further includes: receiving a preset voiceprint information change request sent by a user side, obtaining new user voice information sent by the user side for the second time according to the preset voiceprint information change request, identifying new voiceprint information corresponding to the new user voice information according to the user voice information, taking the new voiceprint information as the new preset voiceprint information, and correspondingly storing the new preset voiceprint information and basic account information of a user in the association table. As shown in Table 1 below, Table 1 is
| User ID | Voiceprint information | Basic account information |
| A | Va1、Va2、、、 | Name: zhang III, mobile phone: 158 a 3344 a, |
| B | Vb1、Vb2、、、 | Name: li IV, mobile phone: 139 a, 3344 a, |
TABLE 1
Step S22: acquiring a voice fragment sent by a user side;
specifically, the voice segment sent by the user side is obtained, and the voice segment may be a voice segment sent by the user in different modes, for example: a user recording a speech in a recording mode, or a speech in a singing mode, and so on.
Step S23: according to the voice fragment, recognizing voiceprint information corresponding to the voice fragment;
specifically, after receiving the voice segment, the server identifies the voiceprint information in the voice segment according to the voice segment.
Step S24: and matching the user account corresponding to the voiceprint information according to the voiceprint information, and storing the voice fragment to the matched user account.
Specifically, the matching, according to the voiceprint information, the user account corresponding to the voiceprint information includes: matching corresponding preset voiceprint information according to the voiceprint information; and matching the user account corresponding to the voiceprint information according to the preset voiceprint information.
The server matches preset voiceprint information corresponding to the voiceprint information according to the voiceprint information, and the preset voiceprint information corresponds to a unique user account, so that the user account corresponding to the voiceprint information is further matched by matching the preset voiceprint information corresponding to the voiceprint information, that is, the user account information and the voiceprint information form one-to-one correspondence, for example: and the basic account information corresponds to the voiceprint information, and if the matching is successful, the voice fragment is stored in the user account corresponding to the voiceprint information.
Referring to fig. 4 again, fig. 4 is an alternative schematic diagram of a data processing method based on voiceprint recognition according to an embodiment of the present invention;
as shown in fig. 4, the data processing method based on voiceprint recognition is applied to a data processing system based on voiceprint recognition, and the system includes: the voice recognition system comprises an input module, a voice acquisition module, a voiceprint recognition module, a processor and a memory;
the method comprises the following steps:
step S401: inputting basic account information of a user;
specifically, the user inputs the basic account information of the user to the processor through the input module, for example: the mobile terminal comprises a server, an input module, a processor and a display module, wherein the server is used for displaying the information of name, age, gender, nickname, mobile phone number, identification number, address, mailbox and the like, the input module can be a mobile terminal, a PC terminal and other equipment, and the processor is a processor of the server.
Step S402: collecting user voice information;
specifically, the user passes through voice acquisition module and gathers user speech information, and will user speech information sends voiceprint recognition module, wherein, voice acquisition module can be recording equipment such as recording pen, cell-phone, record player, microphone, recording equipment connects voiceprint recognition module, voiceprint recognition module can be voiceprint recognition equipment, for example: a voiceprint recognizer.
Step S403: recognizing voiceprint information in user voice information;
specifically, the voiceprint recognition module recognizes voiceprint information in the user voice information and sends the voiceprint information to the processor.
Step S404: using the recognized voiceprint information as preset voiceprint information, and creating a user account;
specifically, the server takes the recognized voiceprint information as preset voiceprint information, binds the preset voiceprint information with the basic account information of the user, generates account information of the user, and further creates a user account.
Step S405: sending basic account information and preset voiceprint information;
specifically, the processor sends the basic account information and the preset voiceprint information to the memory, so that the memory stores the basic account information and the preset voiceprint information.
Step S406: storing the basic account information and the preset voiceprint information in a correlation manner;
specifically, the memory binds and stores the basic account information and the preset voiceprint information in an associated manner, and it can be understood that the memory stores the basic account information and the preset voiceprint information through a database, the database includes an association table, and the basic account information and the preset voiceprint information are stored in an associated manner through the association table.
Referring to fig. 5 again, fig. 5 is an alternative schematic diagram of a data processing method based on voiceprint recognition according to an embodiment of the present invention;
as shown in fig. 5, the data processing method based on voiceprint recognition is applied to a data processing system based on voiceprint recognition, and the system includes: the voice recognition system comprises an input module, a voice acquisition module, a voiceprint recognition module, a processor and a memory;
the method comprises the following steps:
step S501: collecting voice fragments;
specifically, the voice acquisition module acquires a voice segment of the user and sends the voice segment to the voiceprint recognition module.
Step S502: recognizing voiceprint information in the voice fragment;
specifically, the voiceprint recognition module receives a voice fragment sent by the voice acquisition module, recognizes voiceprint information in the voice fragment, and sends the voiceprint information to the processor.
Step S503: calling account information corresponding to the voiceprint information;
specifically, after receiving the voiceprint information, the processor requests voiceprint data in the memory from the memory, matches the voiceprint data in the memory according to the voiceprint information, and if the voiceprint information matches preset voiceprint information in the voiceprint data stored in the memory, determines a user account corresponding to the preset voiceprint information according to the preset voiceprint information, and acquires account information of the user account, for example: basic account information.
Step S504: returning the basic account information;
specifically, the memory returns the basic account information of the user account to the processor.
Step S505: the voice segments are correspondingly sent to the user account;
specifically, the memory sends the voice segment and the user account correspondingly according to the basic account information of the user account, for example: adding the voice segment to a user account identification, wherein the user account identification can be one or more of basic account information of a user account, such as: the memory packs the voice segments and the basic account information into a data packet and sends the data packet to the memory.
Step S506: saving the voice segments to a user account;
specifically, the memory receives a data packet sent by the processor, analyzes the data packet, obtains basic account information of a user included in the data packet, determines a user account according to the basic account information, and stores a voice segment in the data packet to a user account corresponding to the basic account information.
In this embodiment of the present invention, the acquiring the voice fragment sent by the user side specifically includes:
acquiring voice data of a user in one or more application scenes, wherein the application scenes comprise one or more of a recording scene, a meal ordering scene, a singing scene, a speaking scene, a lecture scene, a talking scene, a reading-aloud scene and a reading-after scene, and the voice data comprises at least one voice fragment; and acquiring each voice segment in the voice data according to the voice data.
It can be understood that, in a voice data, there may be a plurality of voice segments, and there is an interval time between each voice segment, or the voice data is continuous voice but there are different users, and each voice segment corresponds to a user, the server segments and obtains at least one voice segment according to the voice data, and recognizes voiceprint information corresponding to the voice segment according to each voice segment.
Referring to fig. 6, fig. 6 is a schematic diagram illustrating a storage process of a speech segment according to an embodiment of the present invention;
the acquiring of the voice fragment sent by the user side specifically includes:
acquiring voice data sent by a plurality of users under the same user account, wherein the voice data comprises at least one voice segment; and acquiring each voice segment in the voice data according to the voice data.
Specifically, under the crowdsourcing platform, the same application scenario, for example: the system comprises a recording scene, a ordering scene, a singing scene, a speech scene, a lecture scene, a talking scene, a reading-aloud scene, a reading-following scene and the like, wherein a plurality of users (all registered users) record the voice, the users record the voice under the same user account, a voice acquisition module acquires voice fragments of the users, a processor identifies voiceprint information in the voice fragments, the corresponding voice fragments are stored under the corresponding user accounts according to the user account information corresponding to the voiceprint information, and the accurate identification and processing of the user recording data by the processor are not influenced.
As shown in fig. 6, the user a account is recorded by the user a, the user B, the user C, and the user D, and corresponds to the voice segment a, the voice segment B, the voice segment C, and the voice segment D, respectively, where each voice segment corresponds to different voiceprint information, for example: the voice segment a corresponds to voiceprint information Va1 and Va2, the voice segment B corresponds to voiceprint information Vb1 and Vb2, the voice segment C corresponds to voiceprint information Vc1 and Vc2, and the voice segment D corresponds to voiceprint information Vd1 and Vd2, and each voice segment is stored in a user account corresponding to the voiceprint information by identifying the corresponding voiceprint information, for example: and storing the voice segment A to an account of the user A, storing the voice segment B to an account of the user B, storing the voice segment C to an account of the user C, and storing the voice segment D to an account of the user D.
In this embodiment of the present invention, the acquiring the voice fragment sent by the user side specifically includes:
acquiring voice data sent by a plurality of users under a plurality of user accounts, wherein the voice data comprises at least one voice fragment; and acquiring each voice segment in the voice data according to the voice data.
Specifically, a plurality of users send voice data under a plurality of user accounts, but each user side does not necessarily correspond to the user account, and there may exist voice data uploaded by other users at the user side of a certain user, so that the system needs to acquire each voice segment in the voice data according to the voice data, recognize voiceprint information according to the voice segment, and match corresponding preset voiceprint information according to the voiceprint information; and matching the user account corresponding to the voiceprint information according to the preset voiceprint information.
Referring to fig. 7, fig. 7 is a schematic diagram illustrating another speech segment storing process according to an embodiment of the present invention;
as shown in fig. 7, a user a uploads a voice segment a, a user B uploads a voice segment B, a user C uploads a voice segment C, and a user D uploads a voice segment D, where the processor respectively identifies corresponding voiceprint information according to the voice segment a, the voice segment B, the voice segment C, and the voice segment D, and matches with a user account corresponding to the voiceprint information, for example: voice segment a corresponds to the account number of user a, voice segment B corresponds to the account number of user B, voice segment C corresponds to the account number of user C, and voice segment D corresponds to the account number of user D, and the voice segments are respectively saved to the corresponding user accounts, for example: and storing the voice segment A to an account of the user A, storing the voice segment B to an account of the user B, storing the voice segment C to an account of the user C, and storing the voice segment D to an account of the user D.
In an embodiment of the present invention, a data processing method based on voiceprint recognition is provided, including: the method comprises the steps that a user account is created in advance, and the user account comprises basic account information and preset voiceprint information; acquiring a voice fragment sent by a user side; according to the voice fragment, recognizing voiceprint information corresponding to the voice fragment; and matching the user account corresponding to the voiceprint information according to the voiceprint information, and storing the voice fragment to the matched user account. By matching the corresponding user account and storing the corresponding voice data based on the voiceprint information, the embodiment of the invention solves the technical problem of data repetition in the voice data processing, and realizes accurate recognition and processing of the voice data.
Referring to fig. 8 again, fig. 8 is a schematic diagram of a data processing apparatus based on voiceprint recognition according to an embodiment of the present invention; the data processing apparatus based on voiceprint recognition can be applied to a server, and as shown in fig. 8, the data processing apparatus 80 based on voiceprint recognition includes:
a useraccount creating unit 81, configured to create a user account in advance, where the user account includes basic account information and preset voiceprint information;
a voicesegment acquiring unit 82, configured to acquire a voice segment sent by a user side;
a voiceprintinformation recognition unit 83, configured to recognize voiceprint information corresponding to the voice segment according to the voice segment;
and the useraccount matching unit 84 is configured to match the user account corresponding to the voiceprint information according to the voiceprint information, and store the voice fragment in the matched user account.
In this embodiment of the present invention, the useraccount creating unit 81 is specifically configured to:
receiving basic account information and user voice information sent by a user side;
according to the user voice information, identifying voiceprint information in the user voice information;
and taking the identified voiceprint information as preset voiceprint information, storing the preset voiceprint information in association with the basic account information, and registering to generate a user account.
In this embodiment of the present invention, the useraccount matching unit 84 is specifically configured to:
matching corresponding preset voiceprint information according to the voiceprint information;
and matching the user account corresponding to the voiceprint information according to the preset voiceprint information.
In this embodiment of the present invention, the voicesegment obtaining unit 82 is specifically configured to:
acquiring voice data of a user in a plurality of application scenes, wherein the application scenes comprise a recording scene, a meal ordering scene and a follow-up reading scene, and the voice data comprises at least one voice fragment;
and acquiring each voice segment in the voice data according to the voice data.
In this embodiment of the present invention, the voicesegment obtaining unit 82 is specifically configured to:
acquiring voice data sent by a plurality of users under the same user account, wherein the voice data comprises at least one voice segment;
and acquiring each voice segment in the voice data according to the voice data.
In this embodiment of the present invention, the useraccount matching unit 84 is specifically configured to:
according to the voice data, voice print information of each voice segment in the voice data is recognized;
and respectively matching the corresponding user accounts according to the voiceprint information of each voice fragment, and storing each voice fragment to the corresponding user account.
In this embodiment of the present invention, the voicesegment obtaining unit 82 is specifically configured to:
acquiring voice data sent by a plurality of users under a plurality of user accounts, wherein the voice data comprises at least one voice fragment;
and acquiring each voice segment in the voice data according to the voice data.
Since the apparatus embodiment and the method embodiment are based on the same concept, the contents of the apparatus embodiment may refer to the method embodiment on the premise that the contents do not conflict with each other, and are not described herein again.
Referring to fig. 9, fig. 9 is a schematic structural diagram of a data processing system based on voiceprint recognition according to an embodiment of the present invention;
as shown in fig. 9, the data processing system 90 based on voiceprint recognition includes: the client 91 and the server 92, wherein the client 91 is communicatively connected to the server 92, in the embodiment of the present invention, the number of the clients 91 may be one, two, or more, and each of the clients 91 is communicatively connected to the server 92. Fig. 9 illustrates an example of a user terminal 91 communicatively connected to a server 92.
Wherein, this user side 91 includes: an input module 911 and a voice acquisition module 912;
specifically, the input module 911 is connected to the processor 921 and configured to input basic account information of the user;
specifically, the voice collecting module 912 is connected to the processor 921 and configured to collect voice data of a user;
the server 92 includes one or more processors 921, a voiceprint recognition module 922, and amemory 923. In fig. 9, one processor 921 is taken as an example.
The processor 921, the voiceprint recognition module 922 and thememory 923 may be connected by a bus or other means, such as the bus connection shown in fig. 9.
The voiceprint recognition module 922 is configured to recognize voiceprint information in the speech segment, and the voiceprint recognition module 922 may be a voiceprint recognizer.
Thememory 923 is a non-volatile computer-readable storage medium and may be configured to store a non-volatile software program, a non-volatile computer-executable program, and modules, such as units (e.g., units shown in fig. 8) corresponding to a data processing method based on voiceprint recognition in the embodiment of the present invention. The processor 921 executes various functional applications and data processing of the data processing method based on voiceprint recognition by running the nonvolatile software program, instructions and modules stored in thememory 923, that is, functions of the respective modules and units of the data processing method based on voiceprint recognition of the above method embodiment and the above device embodiment are realized.
Thememory 923 may include high speed random access memory and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other non-volatile solid state storage device. In some embodiments, thememory 923 may optionally include memory located remotely from the processor 921, which may be connected to the processor 921 via a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The modules are stored in thememory 923, and when executed by the one or more processors 921, perform a voiceprint recognition based data processing method in any of the above-described method embodiments, e.g., perform the steps illustrated in fig. 2 to 5 described above; the functions of the respective modules or units described in fig. 8 can also be implemented.
The user terminal 91 of the embodiment of the present invention exists in various forms, including but not limited to:
(1) a mobile communication device: such devices are characterized by mobile communications capabilities and are primarily targeted at providing voice, data communications. Such electronic devices include smart phones (e.g., iphones), multimedia phones, functional phones, and low-end phones, among others.
(2) The mobile personal computer equipment belongs to the category of personal computers, has calculation and processing functions and generally has the characteristic of mobile internet access. Such electronic devices include: PDA, MID, and UMPC devices, etc., such as ipads.
(3) A portable entertainment device: such devices can display and play video content, and generally also have mobile internet access features. This type of device comprises: video players, handheld game consoles, and intelligent toys and portable car navigation devices.
(4) And other electronic equipment with a video playing function and an internet surfing function.
The server 92 of embodiments of the present invention exists in a variety of forms, including but not limited to:
(1) tower server
The general tower server chassis is almost as large as the commonly used PC chassis, while the large tower chassis is much larger, and the overall dimension is not a fixed standard.
(2) Rack-mounted server
Rack-mounted servers are a type of server that has a standard width of 19 inch racks, with a height of from 1U to several U, due to the dense deployment of the enterprise. Placing servers on racks not only facilitates routine maintenance and management, but also may avoid unexpected failures. First, placing the server does not take up too much space. The rack servers are arranged in the rack in order, and no space is wasted. Secondly, the connecting wires and the like can be neatly stored in the rack. The power line, the LAN line and the like can be distributed in the cabinet, so that the connection lines accumulated on the ground can be reduced, and the accidents such as the electric wire kicking off by feet can be prevented. The specified dimensions are the width (48.26cm ═ 19 inches) and height (multiples of 4.445 cm) of the server. Because of its 19 inch width, a rack that meets this specification is sometimes referred to as a "19 inch rack".
(3) Blade server
A blade server is a HAHD (High Availability High Density) low cost server platform designed specifically for the application specific industry and High Density computer environment, where each "blade" is actually a system motherboard, similar to an individual server. In this mode, each motherboard runs its own system, serving a designated group of different users, without any relationship to each other. Although system software may be used to group these motherboards into a server cluster. In the cluster mode, all motherboards can be connected to provide a high-speed network environment, and resources can be shared to serve the same user group.
(4) Cloud server
The cloud server (ECS) is a computing Service with simplicity, high efficiency, safety, reliability, and flexible processing capability. The management mode is simpler and more efficient than that of a physical server, and a user can quickly create or release any plurality of cloud servers without purchasing hardware in advance. The distributed storage of the cloud server is used for integrating a large number of servers into a super computer, and a large number of data storage and processing services are provided. The distributed file system and the distributed database allow access to common storage resources, and IO sharing of application data files is achieved. The virtual machine can break through the limitation of a single physical machine, dynamically adjust and allocate resources to eliminate single-point faults of the server and the storage equipment, and realize high availability.
Embodiments of the present invention also provide a non-volatile computer storage medium, where the computer storage medium stores computer-executable instructions, which are executed by one or more processors, such as one processor 91 in fig. 9, and enable the one or more processors to perform a data processing method based on voiceprint recognition in any of the above method embodiments, such as performing the above-described steps shown in fig. 2 to 5; the functions of the various units described in fig. 8 may also be implemented.
The above-described embodiments of the apparatus or device are merely illustrative, wherein the unit modules described as separate parts may or may not be physically separate, and the parts displayed as module units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network module units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment.
Through the above description of the embodiments, those skilled in the art will clearly understand that each embodiment can be implemented by software plus a general hardware platform, and certainly can also be implemented by hardware. Based on such understanding, the technical solutions mentioned above may be embodied in the form of a software product, which may be stored in a computer-readable storage medium, such as ROM/RAM, magnetic disk, optical disk, etc., and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute the method according to each embodiment or some parts of the embodiments.
Finally, it should be noted that: the above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; within the idea of the invention, also technical features in the above embodiments or in different embodiments may be combined, steps may be implemented in any order, and there are many other variations of the different aspects of the invention as described above, which are not provided in detail for the sake of brevity; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and the modifications or the substitutions do not make the essence of the corresponding technical solutions depart from the scope of the technical solutions of the embodiments of the present application.