Movatterモバイル変換


[0]ホーム

URL:


CN116055756B - Data processing method, device and server for live broadcast room - Google Patents

Data processing method, device and server for live broadcast room
Download PDF

Info

Publication number
CN116055756B
CN116055756BCN202211706177.4ACN202211706177ACN116055756BCN 116055756 BCN116055756 BCN 116055756BCN 202211706177 ACN202211706177 ACN 202211706177ACN 116055756 BCN116055756 BCN 116055756B
Authority
CN
China
Prior art keywords
data
target
user terminal
language
live video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202211706177.4A
Other languages
Chinese (zh)
Other versions
CN116055756A (en
Inventor
熊威
王思梦
王博
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Construction Bank Corp
CCB Finetech Co Ltd
Original Assignee
China Construction Bank Corp
CCB Finetech Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Construction Bank Corp, CCB Finetech Co LtdfiledCriticalChina Construction Bank Corp
Priority to CN202211706177.4ApriorityCriticalpatent/CN116055756B/en
Publication of CN116055756ApublicationCriticalpatent/CN116055756A/en
Application grantedgrantedCritical
Publication of CN116055756BpublicationCriticalpatent/CN116055756B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Classifications

Landscapes

Abstract

Translated fromChinese

本申请提供了直播间的数据处理方法、装置和服务器,应用于云计算技术领域。基于该方法,云服务器在接收到第一用户终端发起的互动请求之后,可以先根据目标用户对象的对象标识,确定出与目标用户相匹配的目标语言,以及目标用户对象所持有的目标用户终端;根据互动数据的数据类型,从预设的处理规则集中,确定出相匹配的目标处理规则;再基于目标语言,根据目标处理规则,处理互动数据,得到处理后的互动数据;通过目标用户终端的直播间界面在向目标用户对象所展示的当前的直播视频中显示所述处理后的互动数据。从而可以使观众用户在直播间内能够高效、便捷地与使用不同语言的目标用户对象进行多样化的交流互动,提高观众用户在直播间的交互体验。

The present application provides a data processing method, device and server for a live broadcast room, which are applied to the field of cloud computing technology. Based on this method, after receiving an interactive request initiated by a first user terminal, the cloud server can first determine the target language that matches the target user and the target user terminal held by the target user object based on the object identifier of the target user object; determine the matching target processing rule from a preset processing rule set based on the data type of the interactive data; then process the interactive data based on the target language and the target processing rule to obtain the processed interactive data; and display the processed interactive data in the current live video displayed to the target user object through the live broadcast room interface of the target user terminal. This allows the audience users to efficiently and conveniently communicate and interact in a variety of ways with the target user objects using different languages in the live broadcast room, thereby improving the interactive experience of the audience users in the live broadcast room.

Description

Data processing method, device and server of live broadcasting room
Technical Field
The application belongs to the technical field of cloud computing, and particularly relates to a data processing method, device and server of a live broadcasting room.
Background
With the rise and development of the video live broadcast industry, more and more activities such as cross-regional exhibition start to be carried out in the form of online video live broadcast.
The activities such as cross-regional exhibition are held in the form of online video live broadcast, and on one hand, a lot of convenience can be provided for users. On the other hand, the method is limited by the existing data processing method of the live broadcasting room, and because the used languages of users in different areas are different, communication interaction between different users in the same live broadcasting room cannot be better performed, and in addition, the existing live broadcasting room cannot provide a private and safe interaction channel for the users so as to perform communication interaction related to key information, so that the interaction experience of the users is relatively poor.
In view of the above problems, no effective solution has been proposed at present.
Disclosure of Invention
The application provides a data processing method, a data processing device and a data processing server for a live broadcast room, which can enable audience users to efficiently and conveniently interact with target user objects using different languages in a diversified manner in the live broadcast room, and improve interaction experience of the audience users.
The application provides a data processing method of a live broadcasting room, which is applied to a cloud server and comprises the following steps:
Receiving an interaction request initiated by a first user terminal, wherein the interaction request at least carries interaction data and object identifiers of target user objects aimed by the interaction data;
According to the object identification of the target user object, determining a target language matched with the target user and a target user terminal held by the target user object;
according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set;
Processing the interaction data according to a target processing rule based on a target language to obtain processed interaction data;
And displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
In one embodiment, the data type of the interactive data comprises at least one of text data, voice data and expression images.
In one embodiment, when the data type of the interactive data includes voice data, processing the interactive data according to a target processing rule based on a target language to obtain processed interactive data includes:
performing voice recognition on the voice data by utilizing a voice recognition model to obtain corresponding text data, and extracting voice characteristics of a first audience user from the voice data;
Processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
And processing the text data based on the target language by utilizing a preset voice synthesis model according to the voice characteristics of the first audience user to obtain corresponding synthesized voice data serving as the processed interaction data.
In one embodiment, when the data type of the interactive data includes an expression image, processing the interactive data according to a target processing rule based on a target language to obtain processed interactive data includes:
performing character detection on the expression image to determine whether meaningful text characters exist in the expression image;
Under the condition that meaningful text characters exist in the expression image, the image recognition model is utilized to recognize the expression image, so that the text characters in the expression image are extracted as text data;
Processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
constructing annotation data for the expression image according to the text data based on the target language;
And combining the expression image and the annotation data to obtain the expression image carrying the annotation data, and taking the expression image carrying the annotation data as the processed interaction data.
In one embodiment, in the case that the data type of the interactive data includes text data, processing the interactive data according to a target processing rule based on a target language to obtain processed interactive data includes:
Processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
And constructing corresponding barrage data according to the text data of the target language, and taking the barrage data as the processed interactive data.
In one embodiment, the method further comprises:
receiving a private letter request initiated by a first user terminal, wherein the private letter request at least carries an object identifier of a target user object;
And responding to the private message request, and establishing a private data communication channel between the first user terminal and the target user terminal based on a live broadcasting interface according to a related encryption communication protocol, wherein the cloud server displays the processed interactive data in the current live broadcasting video displayed to the target user object through the live broadcasting interface of the target user terminal based on the private data communication channel.
In one embodiment, after establishing the private data communication channel based on the live room interface between the first user terminal and the target user terminal, the method further comprises:
encrypting the processed interaction data by utilizing public key data corresponding to the target user terminal to obtain ciphertext data of the interaction data;
The method comprises the steps of sending ciphertext data of interaction data to a target user terminal through a privacy data communication channel, decrypting the ciphertext data of the interaction data by the target user terminal through the private key data to obtain processed interaction data, and displaying the processed interaction data in a current live video displayed to a target user object through a live broadcasting room interface.
In one embodiment, before receiving the interaction request initiated by the first user terminal, the method further comprises:
Receiving a connection request initiated by a first user terminal;
responding to the connection request, and establishing a first data connection with a first user terminal;
and transmitting the live video stream data to a first user terminal through the first data connection, wherein the first user terminal displays the current live video to a first user through a live broadcasting room interface according to the live video stream data.
In one embodiment, while transmitting live video streaming data to the first user terminal over the first data connection, the method further comprises:
Collecting characteristic parameters of a first user terminal;
Determining a first language matched with a first audience user according to the characteristic parameters of the first user terminal;
detecting whether live video stream data currently transmitted through a first data connection are live video stream data based on a first language;
Under the condition that live video stream data transmitted through the first data connection at present is not based on live video stream data of a first language, determining a cloud CDN cached with the live video stream data of the first language as a first target CDN;
And switching the first data connection to a first target CDN so as to transmit the live video streaming data based on the first language to a first user terminal through the first data connection.
In one embodiment, the characteristic parameter of the first user terminal comprises at least one of an IP address of the first user terminal, a default language parameter of a browser of the first user terminal, cookie data of the first user terminal.
The application also provides a data processing method of the live broadcasting room, which is applied to the first user terminal and comprises the following steps:
Displaying a current live video based on a first language to a first audience user through a live room interface;
receiving interaction data input by a first user through a live broadcasting room interface, and determining a target user object aimed at by the interaction data;
generating a corresponding interaction request according to the interaction data, wherein the interaction request also carries an object identifier of a target user object;
The method comprises the steps of sending an interaction request to a cloud server, determining a target language matched with a target user object by the cloud server according to the object identification of the target user object, processing interaction data based on the target language to obtain processed interaction data, and displaying the processed interaction data in a current live video displayed to the target user object through a live broadcasting interface of a target user terminal by the cloud server.
The application also provides a data processing device of the live broadcasting room, which is applied to the cloud server and comprises the following components:
the system comprises a receiving module, a receiving module and a processing module, wherein the receiving module is used for receiving an interaction request initiated by a first user terminal, and the interaction request at least carries interaction data and object identifiers of target user objects aimed at by the interaction data;
The first determining module is used for determining a target language matched with the target user and a target user terminal held by the target user object according to the object identification of the target user object;
the second determining module is used for determining a matched target processing rule from a preset processing rule set according to the data type of the interactive data;
the processing module is used for processing the interaction data according to target processing rules based on target language to obtain processed interaction data;
and the touch module is used for displaying the processed interaction data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
The application also provides a server comprising a processor and a memory for storing processor executable instructions, which when executed by the processor implement the relevant steps of the data processing method of the live room.
The application further provides a computer readable storage medium, wherein computer instructions are stored on the computer readable storage medium, when the instructions are executed by a processor, the computer readable storage medium comprises the steps of receiving an interaction request initiated by a first user terminal, wherein the interaction request at least carries interaction data and object identifiers of target user objects aimed by the interaction data, determining target languages matched with target users and target user terminals held by the target user objects according to the object identifiers of the target user objects, determining matched target processing rules from a preset processing rule set according to the data types of the interaction data, processing the interaction data according to the target processing rules based on the target languages to obtain processed interaction data, and displaying the processed interaction data in a current live video displayed to the target user objects through a live broadcasting interface of the target user terminals.
The application also provides a computer program product comprising a computer program which, when executed by a processor, implements the relevant steps of the data processing method of the living room.
The method, the device and the server for processing the data in the live broadcasting room are based on the fact that after an interaction request initiated by a first user terminal is received by a cloud server of a cloud live broadcasting service platform, a target language matched with a target user and a target user terminal held by the target user object can be determined according to the object identification of the target user object, the matched target processing rule is determined from a preset processing rule set according to the data type of the interaction data, the interaction data are processed according to the target processing rule based on the target language, the processed interaction data are obtained, and the processed interaction data are displayed in a current live broadcasting video displayed to the target user object through an interface of the live broadcasting room of the target user terminal. Therefore, the audience user can efficiently and conveniently interact with target user objects using different languages in a diversified communication manner in the live broadcasting room, and the interaction experience of the audience user in the live broadcasting room is improved. Furthermore, the cloud server can also establish an exclusive privacy data communication channel with the target user object in the living broadcast room for the audience user according to the specific requirements of the audience user, and the audience user can further directly perform safer and more private communication interaction with the target user object in the living broadcast room interface based on the privacy data communication channel, so that the privacy data in the communication interaction process of the audience user is prevented from being revealed.
Drawings
In order to more clearly illustrate the embodiments of the present application, the drawings that are required to be used in the embodiments will be briefly described, and the drawings described below are only some embodiments described in the present application, and other drawings may be obtained according to the drawings without inventive effort for those skilled in the art.
Fig. 1 is a flow chart of a data processing method of a live broadcast room provided by an embodiment of the present application;
FIG. 2 is a schematic diagram of an embodiment of a method for processing data in a live broadcast room according to an embodiment of the present application in one example scenario;
FIG. 3 is a schematic diagram of an embodiment of a method for processing data in a live broadcast room according to an embodiment of the present application in one example scenario;
FIG. 4 is a schematic diagram of an embodiment of a method for processing data in a live broadcast room according to an embodiment of the present application in one example scenario;
FIG. 5 is a schematic diagram of an embodiment of a method for processing data in a live broadcast room according to an embodiment of the present application in one example scenario;
Fig. 6 is a flow chart of a data processing method of a live broadcast room according to another embodiment of the present application;
FIG. 7 is a schematic diagram of the structural composition of a server according to an embodiment of the present application;
FIG. 8 is a schematic diagram of the structural components of a data processing apparatus of a live broadcast room according to an embodiment of the present application;
fig. 9 is a schematic structural diagram of a data processing apparatus of a live broadcast room according to another embodiment of the present application.
Detailed Description
In order to make the technical solution of the present application better understood by those skilled in the art, the technical solution of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application, and it is apparent that the described embodiments are only some embodiments of the present application, not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, shall fall within the scope of the application.
It should be noted that, in the present specification, information data related to a user is acquired and used on the premise that the user knows and agrees. And, the information data is acquired, stored, used, processed and the like, which meet the relevant regulations of national laws and regulations.
Referring to fig. 1, an embodiment of the present application provides a data processing method in a live broadcast room, where the method is specifically applied to a cloud server side. In particular implementations, the method may include the following:
S101, receiving an interaction request initiated by a first user terminal, wherein the interaction request at least carries interaction data and object identifiers of target user objects aimed by the interaction data;
S102, determining a target language matched with a target user and a target user terminal held by the target user object according to the object identification of the target user object;
s103, determining a matched target processing rule from a preset processing rule set according to the data type of the interactive data;
S104, processing the interaction data according to a target processing rule based on a target language to obtain processed interaction data;
and S105, displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
The first user terminal may be a user terminal held by a first audience user. The first viewer user may be specifically understood as any viewer user currently online in the living room.
The target user object may specifically include a user object that is in the same living room as the first viewer user, and the first viewer user object currently wants to interact separately. In particular, the target user object may include a anchor user and/or other viewer users within the same living room as the first viewer user. The target user object may include one or more user objects.
The live video can be a trans-regional exhibition live video, a trans-regional academic conference live video, a trans-regional commodity communication live video and the like. Of course, the live video listed above is only one illustrative illustration. In specific implementation, the live video may further include live video of other suitable types and content according to specific application scenarios and processing requirements. The present specification is not limited to this.
Based on the above embodiment, the cloud server may process the interaction data of the first audience user for the target user object in the same living broadcast room into the processed interaction data based on the target language matched with the target user object, and then display the processed interaction data in the current living broadcast video displayed to the target user object through the living broadcast room interface of the target user terminal, so that the audience user can efficiently and conveniently perform diversified and language barrier-free interaction communication with the target user object such as the anchor user and/or other audience users designated in the same living broadcast room in the living broadcast room, and the interaction experience of the audience user in the living broadcast room is effectively improved.
In some embodiments, referring to fig. 2, the method for processing data in the live broadcast room may be specifically applied to a cloud server side.
The cloud server specifically comprises a background server which is applied to one side of a cloud live broadcast service platform and can realize functions of data transmission, data processing and the like. Specifically, the cloud server may be, for example, an electronic device having a data operation function, a storage function, and a network interaction function. Or the cloud server may also be a software program running in the electronic device that provides support for data processing, storage, and network interactions. In the present embodiment, the number of servers included in the cloud server is not particularly limited. The cloud server can be one server, several servers or a server cluster formed by several servers.
Specifically, the cloud server may be connected to the live broadcast terminal and the plurality of user terminals in a wired or wireless manner.
The live broadcast terminal and the user terminal can specifically comprise front ends which are applied to the side of a host user and a viewer user and can realize functions of data acquisition, data transmission and the like. Specifically, the live broadcast terminal and the user terminal may be electronic devices such as a desktop computer, a tablet computer, a notebook computer, and a smart phone, for example. Or the live terminal and the user terminal may be software applications capable of running in the electronic device. For example, it may be some live APP running on a smart phone, etc.
Furthermore, the cloud server may be further configured with a plurality of algorithm models, such as a speech recognition model, a translation model, and an image recognition model, and may be further connected with a database, a translation terminal, and a plurality of cloud CDNs, such as a cloud CDN1, a cloud CDN2, a cloud CDNn, and the like.
The cloud CDN (Content Delivery Network) may specifically refer to a cloud technology-based content distribution network. The translation terminal can be a manual translation terminal or an automatic translation terminal based on artificial intelligence.
In specific implementation, referring to fig. 2, a host user may use a live terminal to perform live video broadcast, such as exhibition, product propaganda, and the like. The live broadcast terminal can acquire live broadcast video stream data of the anchor user in real time and upload the live broadcast video stream data to the cloud server.
After receiving the live video stream data, the cloud server can directly forward the live video stream data to the user terminal so that audience users can watch live video based on the language used by the anchor user through a live broadcasting room interface of the user terminal, and can send the live video stream data to the translation terminal so as to translate the live video stream data in real time through the translation terminal, and then buffer the translated live video stream data based on different languages into a corresponding cloud terminal so that audience users using different languages in different areas can relatively synchronously pull and watch live video based on the language used by the audience users in the same live broadcasting room through the user terminal.
Specifically, taking any one of the first audience users in the living broadcast room as an example. The current first audience user views the current live video based on the first language through a live broadcast room interface displayed by the held first user terminal. The first language may be understood as a language matching the first audience user. For example, the native language of the first viewer user, etc.
The first audience user can watch the current live video through the live broadcasting room interface displayed by the first user terminal, and can select the anchor user and/or other audience users in the live broadcasting room as target user objects, and the target user objects are independently used for communication interaction in the live broadcasting room interface.
Specifically, a list of user objects may also be presented in the live room interface. The user object list may specifically include object identifiers of user objects such as anchor users and online audience users in the living broadcast room. The first viewer user may initiate a click operation in the list of user objects specifying a target user object to be specifically interacted with by selecting an object identification of one or more user objects.
After selecting the target user object, the live room interface may pop up the interactive data entry box. Accordingly, the first viewer user may input specific interactive data through the interactive data input box.
The first user terminal can receive the interaction data, determine a target user object aimed by the interaction data, acquire an object identification of the target user object, and further generate a corresponding interaction request. The interaction request at least can carry interaction data and object identification of the target user object.
Then, the first user terminal may send the interaction request to the cloud server. Correspondingly, the cloud server receives and acquires the first interaction request.
In some embodiments, the data type of the interactive data may specifically include at least one of text data, voice data, emoticons, and the like. Of course, it should be noted that the above-listed data types of the interactive data are only illustrative. In specific implementation, the interactive data may further include other suitable types of interactive data according to specific situations and processing requirements. The present specification is not limited to this.
Based on the embodiment, the audience user can freely use the interaction data of one or more different data types to interact with the target user object of the living broadcast room according to specific conditions, so that the diversified interaction requirements of the audience user can be met.
In some embodiments, in implementation, the cloud server may determine, as the target user terminal, a user terminal held by the target user object by querying a user database according to the direction identifier of the target user object.
Furthermore, the cloud server can acquire the characteristic parameters of the target user terminal according to the data connection established by the target user terminal for acquiring the live video data, and determine the language matched with the target user as the target language according to the characteristic parameters of the target user terminal.
In addition, the cloud server can also query the database according to the identification information of the target user object, and determine the target language matched with the target user object according to the user data about the target user object recorded in the database.
In some embodiments, before implementation, corresponding preset processing rules may be configured in advance for each of the interactive data of different data types. Each preset processing rule may include a corresponding algorithm rule and an algorithm model. Further, a corresponding preset processing rule set may be obtained according to combining the plurality of preset processing rules. And the preset processing rule set also stores a matching relation between the preset processing rule and the data type of the interactive data.
In some embodiments, in the implementation, a matched preset processing rule may be determined from a preset processing rule set according to a data type of the interactive data, and the matched preset processing rule is used as the target processing rule.
In some embodiments, referring to fig. 3, in the case where the data type of the interactive data includes voice data, the processing the interactive data according to the target processing rule based on the target language may include the following steps when implementing:
s1, performing voice recognition on the voice data by utilizing a voice recognition model to obtain corresponding text data, and extracting voice characteristics of a first audience user from the voice data;
s2, processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
And S3, processing the text data based on the target language by utilizing a preset voice synthesis model according to the voice characteristics of the first audience user to obtain corresponding synthesized voice data serving as the processed interactive data.
Based on the embodiment, the voice data input by the first audience user can be efficiently converted into the voice data based on the target language, and the voice data accords with the voice characteristics of the first audience user when speaking, so that the synthesized voice data of emotion information such as mood and attitude of the first audience user when speaking can be truly and comprehensively reflected, and then the synthesized voice data is used as the processed interaction data to reach the target user object, so that the target user object can understand related semantic content in the interaction data conveniently and efficiently based on the target language used by the target user object, and meanwhile, the real emotion information of the first audience user can be intuitively felt through the synthesized voice data, and relatively good interaction effect can be obtained.
Wherein the speech features include at least one of pitch, loudness, frequency, timbre, etc.
In the implementation, the voice data is subjected to voice recognition by utilizing the voice recognition model to obtain corresponding text data, and the voice data can be processed by utilizing the voice feature extraction model to extract the voice features when the first audience user inputs the voice data. And processing the text data by utilizing the translation model matched with the target language to obtain translated text data based on the target language. Further, a preset voice synthesis model can be utilized, voice synthesis can be performed based on voice characteristics of a target user by using text data based on a target language, so that the processed interactive data containing semantic content based on the target language and simultaneously capable of conveying real emotion information of a first audience user can be obtained.
In some embodiments, referring to fig. 4, in the case where the data type of the interactive data includes an expression image, the processing the interactive data according to the target processing rule based on the target language may include the following steps when implementing:
s1, carrying out character detection on an expression image to determine whether meaningful text characters exist in the expression image;
S2, under the condition that meaningful text characters exist in the expression image, carrying out recognition processing on the expression image by utilizing an image recognition model so as to extract the text characters in the expression image as text data;
s3, processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
S4, constructing annotation data for the expression image according to the text data based on the target language;
and S5, combining the expression image and the annotation data to obtain the expression image carrying the annotation data, and taking the expression image as the processed interaction data.
Based on the embodiment, the expression image input by the first audience user can be efficiently converted into the processed interactive data carrying the labeling data based on the target language, so that the target user object can conveniently and efficiently understand the true meaning expressed by the expression image sent by the first audience user.
In particular, the emoticons may be left unprocessed in the event that it is determined that there are no meaningful text characters in the emoticons. And directly determining the expression image as the processed interactive data.
The method comprises the steps of determining whether a significant text character exists in an expression image, detecting the text character in the expression image, and judging whether the significant text character exists in the text character in the expression image according to a nonsensical character reference template matched with a first language under the condition that the text character exists in the expression image.
When the expression image and the annotation data are specifically combined, an annotation frame containing the annotation data can be generated, and then the annotation frame and the expression image are spliced, so that the expression image carrying the annotation data can be obtained.
In some embodiments, when the data type of the interactive data includes text data, the processing the interactive data according to the target processing rule based on the target language to obtain the processed interactive data may further include the following when implemented:
S1, processing text data by using a translation model matched with a target language to obtain text data based on the target language;
s2, constructing corresponding barrage data according to the text data of the target language, and using the barrage data as the processed interactive data.
Based on the embodiment, the text data input by the first audience user can be efficiently and accurately converted into the text data based on the target language, so that the target user object can conveniently and rapidly understand the semantic content which is expected to be expressed by the interactive data sent by the first audience user.
In some embodiments, it is contemplated that the timeliness requirements are often high in the actual interaction process, but the precision requirements for semantic content are not very stringent. For example, it is common for different user objects to know the approximate meaning of each other in the communication interaction. Furthermore, in consideration of the fact that a plurality of communication interactions may exist at the same time in the same live broadcast room, if the translation terminal with higher calling precision performs related translation processing on text data, on one hand, the data processing capacity of the translation terminal is too large, and even normal processing of live broadcast video stream data is affected, on the other hand, timeliness of the communication interactions cannot be better met.
Based on the above consideration, in the present specification, text data is processed separately mainly by calling a translation model independent of a relatively simplified version of a translation terminal, so that user objects have higher precision compared with key information of interest during translation processing of the text data, timeliness of interaction is effectively ensured on the premise that basic interaction between the user objects is not affected, data processing burden is reduced, and meanwhile, influence on processing of live video stream data is avoided.
The method comprises the steps of determining receiving time of an interaction request, determining subject information in live video when a first user initiates the interaction request through a first user terminal according to the initiating time, and processing text data by using the translation model matched with the target language based on the subject information as a reference, so that the text data is translated more efficiently on the premise of ensuring that key information related to the subject information has higher precision, and the text data based on the target language, which meets basic interaction requirements, is obtained rapidly.
In some embodiments, referring to fig. 5, when the method is implemented, the following may further be included:
s1, receiving a private letter request initiated by a first user terminal, wherein the private letter request at least carries an object identifier of a target user object;
And S2, responding to the private message request, and establishing a private data communication channel based on a live broadcasting interface between the first user terminal and the target user terminal according to a related encryption communication protocol, wherein the cloud server displays the processed interactive data in the current live broadcasting video displayed to the target user object through the live broadcasting interface of the target user terminal based on the private data communication channel. Wherein the processed interactive data is perceivable only by the first viewer user and the target user object.
The privacy data communication channel is different from a common live broadcasting room barrage public screen. In particular, a privacy dialog box visible only to the first viewer user and the target user object may additionally be presented in the live room interface of the two user terminals establishing the privacy data communication channel. The privacy dialog box is only used for displaying and inputting interactive data which are mutually communicated between two user terminals with privacy data communication channels.
It should be noted that, based on the above-mentioned privacy data communication channel, the processed interactive data displayed in the current live video displayed to the target user object through the live broadcast room interface of the target user terminal is only perceptible to the first viewer by the target user object, but is not perceptible to other user objects in the live broadcast room.
Based on the above embodiment, in some live broadcast scenarios such as a business exhibition, the first audience user can also independently establish an independent and private data communication channel with the target user object in the live broadcast room, so that based on the private data communication channel, under the condition that other user objects cannot perceive, and meanwhile, the first audience user and the target user object are not influenced to watch the current live broadcast video, communication interaction with higher security is performed, and private data involved in the communication process is prevented from being revealed.
In some embodiments, after establishing the private data communication channel based on the live-room interface between the first user terminal and the target user terminal, the method may further include, when embodied, the following:
s1, encrypting the processed interaction data by utilizing public key data corresponding to a target user terminal to obtain ciphertext data of the interaction data;
And S2, sending ciphertext data of the interactive data to the target user terminal through a privacy data communication channel, wherein the target user terminal decrypts the ciphertext data of the interactive data by utilizing the private key data to obtain the processed interactive data, and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting interface.
In the implementation, after establishing a private data communication channel based on a live broadcasting interface between a first user terminal and a target user terminal, the cloud server can also interact with the target user terminal according to a related encryption communication protocol, generate public key data and private key data corresponding to the target user terminal by using a terminal identifier of the target user terminal and a random number generator, and store the public key data therein by the cloud server and store the private key data therein by the target user terminal. Similarly, the cloud server can interact with the first user terminal according to the related encryption communication protocol, generate public key data and private key data corresponding to the first user terminal by using the terminal identification of the first user terminal and the random number generator, and keep the public key data therein by the cloud server and keep the private key data therein by the first user terminal.
Based on the embodiment, the privacy data related to the interaction between the first user terminal and the target user terminal through the privacy data communication channel can be more effectively prevented from being revealed, and the data security during the interaction is better protected.
In some embodiments, before receiving the interaction request initiated by the first user terminal, the method may further include the following when implemented:
S1, receiving a connection request initiated by a first user terminal;
s2, responding to the connection request, and establishing a first data connection with a first user terminal;
And S3, transmitting the live video stream data to a first user terminal through the first data connection, wherein the first user terminal displays the current live video to a first user through a live broadcasting room interface according to the live video stream data.
Based on the embodiment, the cloud server can quickly respond to the connection request initiated by the first user terminal, and the first user terminal can efficiently and stably acquire and display the real-time live video by establishing and using the first data connection.
In some embodiments, while transmitting live video streaming data to the first user terminal through the first data connection, the method may further include the following when implemented:
S1, collecting characteristic parameters of a first user terminal;
S2, determining a first language matched with a first audience user according to the characteristic parameters of the first user terminal;
s3, detecting whether live video stream data currently transmitted through the first data connection are live video stream data based on a first language;
S4, determining a cloud CDN cached with the live video stream data based on the first language as a first target CDN under the condition that the live video stream data currently transmitted through the first data connection is not the live video stream data based on the first language;
and S5, switching the first data connection to a first target CDN so as to transmit the live video stream data based on the first language to the first user terminal through the first data connection.
Based on the embodiment, the cloud server can automatically detect and identify the first language matched with the first audience user, and further can automatically switch to the live video matched with the first audience user based on the first language in time under the condition that the live video currently provided for the first audience user is actively judged to be not matched with the first user, so that the first audience user can obtain better interaction experience in a live broadcast room.
In some embodiments, the characteristic parameter package of the first user terminal may specifically include at least one of an IP address of the first user terminal, a default language parameter of a browser of the first user terminal, cookie data of the first user terminal, and the like.
Based on the above embodiment, the cloud server may accurately and automatically determine the language type matched with the first user by collecting and according to the feature parameters of the first user terminal.
Specifically, the cloud server may collect, through the first data connection, a characteristic parameter of the first user terminal.
In some embodiments, the method may further include the following when implemented:
s1, acquiring live video stream data acquired in real time;
s2, calling a corresponding translation terminal to convert the live video stream data into video stream data based on a plurality of different languages, and respectively caching the video stream data based on the plurality of different languages into a corresponding cloud CDN.
In some embodiments, after transmitting the live video streaming data to the first user terminal through the first data connection, the method may further include the following when implemented:
S1, receiving a language switching request initiated by a first user terminal;
S2, determining a second language customized by the first audience user according to the language switching request;
S3, determining a cloud CDN cached with live video stream data based on a second language as a second target CDN;
And S4, switching the first data connection to a second target CDN so as to transmit the live video stream data based on the second language to the first user terminal through the first data connection.
It can be seen that, according to the data processing method of the live broadcasting room provided by the embodiment of the application, after receiving an interaction request initiated by a first user terminal, a cloud server can determine a target language matched with a target user and a target user terminal held by the target user object according to an object identifier of the target user object, determine a matched target processing rule from a preset processing rule set according to a data type of the interaction data, process the interaction data according to the target processing rule based on the target language to obtain processed interaction data, and display the processed interaction data in a current live broadcasting video displayed to the target user object through a live broadcasting room interface of the target user terminal. Therefore, the audience users can efficiently and conveniently interact with target user objects using different languages in a diversified manner in the live broadcasting room, the interaction experience of the audience users in the live broadcasting room is improved, and the diversified interaction requirements of the audience users are met. Furthermore, the cloud server can also establish an exclusive privacy data communication channel with the target user object in the living broadcast room for the audience user according to the needs of the audience user, and the audience user can further directly perform safer and more private communication interaction with the target user object in the living broadcast room interface based on the privacy data communication channel, so that the privacy data related to the communication interaction of the audience user is prevented from being revealed.
Referring to fig. 6, the embodiment of the application also provides another data processing method of the live broadcasting room, which is applied to the first user terminal. The method can be implemented by the following steps:
S601, displaying a current live video based on a first language to a first audience user through a live room interface;
S602, receiving interaction data input by a first user through a live broadcasting room interface, and determining a target user object aimed at by the interaction data;
s603, generating a corresponding interaction request according to the interaction data, wherein the interaction request also carries an object identifier of a target user object;
The method comprises the steps of S604, sending the interaction request to a cloud server, wherein the cloud server determines a target language matched with a target user object according to the object identification of the target user object, processes interaction data based on the target language to obtain processed interaction data, and further displays the processed interaction data in a current live video displayed to the target user object through a live broadcasting interface of a target user terminal.
Therefore, the data processing method of the live broadcasting room provided by the embodiment of the application can enable audience users to efficiently and conveniently interact with target user objects using different languages in a diversified manner in the live broadcasting room, and improve interaction experience of the audience users. Furthermore, a private data communication channel exclusive to the target user object can be established in the living broadcast room for the audience user according to the needs of the audience user, and the audience user can perform safer and private communication interaction with the target user object in the living broadcast room interface directly based on the private data communication channel, so that the private data of the audience user is prevented from being revealed.
The embodiment of the application also provides a server, which comprises a processor and a memory for storing executable instructions of the processor, wherein the processor can execute the following steps according to the instructions when being implemented, the interactive request initiated by a first user terminal is received, the interactive request at least carries interactive data and object identifiers of target user objects aimed by the interactive data, a target language matched with the target user and the target user terminal held by the target user object are determined according to the object identifiers of the target user objects, the matched target processing rule is determined from a preset processing rule set according to the data type of the interactive data, the interactive data are processed according to the target processing rule based on the target language, and the processed interactive data are displayed in the current live video displayed to the target user object through a live broadcasting interface of the target user terminal.
In order to more accurately complete the above instructions, referring to fig. 7, another specific server is further provided in the embodiment of the present application, where the server includes a network communication port 701, a processor 702, and a memory 703, and the above structures are connected by an internal cable, so that each structure may perform specific data interaction.
The network communication port 701 may be specifically configured to receive an interaction request initiated by a first user terminal, where the interaction request at least carries interaction data and an object identifier of a target user object to which the interaction data is directed.
The processor 702 is specifically configured to determine, according to an object identifier of a target user object, a target language matched with the target user and a target user terminal held by the target user object, determine, according to a data type of the interaction data, a matched target processing rule from a preset processing rule set, process the interaction data according to the target processing rule based on the target language, obtain processed interaction data, and display the processed interaction data in a current live video displayed to the target user object through a live broadcasting interface of the target user terminal.
The memory 703 may be used for storing a corresponding program of instructions.
In this embodiment, the network communication port 701 may be a virtual port that binds with different communication protocols, so that different data may be sent or received. For example, the network communication port may be a port responsible for performing web data communication, a port responsible for performing FTP data communication, or a port responsible for performing mail data communication. The network communication port may also be an entity's communication interface or a communication chip. For example, it may be a wireless mobile network communication chip such as GSM, CDMA, etc., it may also be a Wifi chip, it may also be a bluetooth chip.
In this embodiment, the processor 702 may be implemented in any suitable manner. For example, a processor may take the form of, for example, a microprocessor or processor, and a computer-readable medium storing computer-readable program code (e.g., software or firmware) executable by the (micro) processor, logic gates, switches, application SPECIFIC INTEGRATED Circuits (ASICs), programmable logic controllers, and embedded microcontrollers, among others. The description is not intended to be limiting.
In this embodiment, the memory 703 may include multiple levels, and in a digital system, the memory may be any memory as long as binary data can be stored, in an integrated circuit, a circuit with a memory function without a physical form, such as a RAM, a FIFO, etc., and in a system, a memory device with a physical form, such as a memory bank, a TF card, etc.
The embodiment of the application also provides a user terminal, which comprises a processor and a memory for storing executable instructions of the processor, wherein the processor can be used for displaying a current live video based on a first language to a first audience user through a live broadcasting interface when being implemented, receiving interaction data input by the first user through the live broadcasting interface, determining target user objects aimed at by the interaction data, generating a corresponding interaction request according to the interaction data, wherein the interaction request also carries object identification of the target user objects, sending the interaction request to a cloud server, determining target language matched with the target user objects according to the object identification of the target user objects, processing the interaction data based on the target language, and obtaining the processed interaction data, and the cloud server also displaying the processed interaction data in the current live video displayed to the target user objects through the live broadcasting interface of the target user terminal.
The embodiment of the application also provides a computer readable storage medium based on the data processing method of the live broadcasting room, wherein the computer readable storage medium stores computer program instructions, when the computer program instructions are executed, the method comprises the steps of receiving an interaction request initiated by a first user terminal, wherein the interaction request at least carries interaction data and object identifications of target user objects aimed at by the interaction data, determining a target language matched with the target user according to the object identifications of the target user objects and the target user terminal held by the target user objects, determining matched target processing rules from a preset processing rule set according to the data types of the interaction data, processing the interaction data according to the target processing rules based on the target language, and displaying the processed interaction data in a current live broadcasting video displayed to the target user objects through an interface between the target user terminals.
In the present embodiment, the storage medium includes, but is not limited to, a random access Memory (Random Access Memory, RAM), a Read-Only Memory (ROM), a Cache (Cache), a hard disk (HARD DISK DRIVE, HDD), or a Memory Card (Memory Card). The memory may be used to store computer program instructions. The network communication unit may be an interface for performing network connection communication, which is set in accordance with a standard prescribed by a communication protocol.
In this embodiment, the functions and effects of the program instructions stored in the computer readable storage medium may be explained in comparison with other embodiments, and are not described herein.
The embodiment of the application also provides a computer program product, which comprises a computer program, wherein the computer program when being executed by a processor realizes the following steps of receiving an interaction request initiated by a first user terminal, wherein the interaction request at least carries interaction data and object identifications of target user objects aimed at by the interaction data, determining target languages matched with target users and target user terminals held by the target user objects according to the object identifications of the target user objects, determining matched target processing rules from a preset processing rule set according to the data types of the interaction data, processing the interaction data according to the target processing rules based on the target languages to obtain processed interaction data, and displaying the processed interaction data in a current live video displayed to the target user objects through a live broadcasting interface of the target user terminals.
Referring to fig. 8, on a software level, the embodiment of the application further provides a data processing device of a live broadcast room, which is applied to a cloud server side, and the device specifically may include the following structural modules:
the receiving module 801 may be specifically configured to receive an interaction request initiated by a first user terminal, where the interaction request at least carries interaction data and an object identifier of a target user object to which the interaction data is directed;
The first determining module 802 may be specifically configured to determine, according to an object identifier of a target user object, a target language matched with the target user and a target user terminal held by the target user object;
The second determining module 803 may be specifically configured to determine, according to a data type of the interaction data, a matched target processing rule from a preset processing rule set;
The processing module 804 may be specifically configured to process the interaction data according to a target processing rule based on a target language, to obtain processed interaction data;
the touch module 805 may be specifically configured to display the processed interaction data in a current live video displayed to the target user object through a live room interface of the target user terminal.
In some embodiments, the data type of the interactive data may specifically include at least one of text data, voice data, emoticons, and the like.
In some embodiments, the processing module 804 may be configured to process, when the data type of the interactive data includes voice data, the interactive data according to a target processing rule based on a target language, to obtain processed interactive data, that is, perform voice recognition on the voice data by using a voice recognition model to obtain corresponding text data, extract voice features of a first audience from the voice data, process the text data by using a translation model matched with the target language to obtain text data based on the target language, and process the text data based on the target language according to the voice features of the first audience by using a preset voice synthesis model to obtain corresponding synthesized voice data as the processed interactive data.
In some embodiments, the processing module 804 may be configured to process the interactive data based on a target language according to a target processing rule when the data type of the interactive data includes an expression image, to obtain processed interactive data, to perform character detection on the expression image to determine whether there are meaningful text characters in the expression image, to perform recognition processing on the expression image by using an image recognition model to extract the text characters in the expression image as text data when it is determined that there are meaningful text characters in the expression image, to process the text data by using a translation model matched with the target language to obtain text data based on the target language, to construct annotation data for the expression image according to the text data based on the target language, and to combine the expression image and the annotation data to obtain the expression image carrying the annotation data as the processed interactive data.
In some embodiments, the processing module 804 may be configured to process the interactive data according to a target processing rule based on a target language when the data type of the interactive data includes text data, to obtain processed interactive data, process the text data using a translation model matched with the target language to obtain text data based on the target language, and construct corresponding barrage data according to the text data of the target language as the processed interactive data.
In some embodiments, the device may be further configured to receive a private message request initiated by the first user terminal when the device is implemented, where the private message request at least carries an object identifier of the target user object, and establish a private data communication channel between the first user terminal and the target user terminal based on a live broadcast interface according to an associated encrypted communication protocol in response to the private message request, where the cloud server displays the processed interactive data in a current live broadcast video presented to the target user object through the live broadcast interface of the target user terminal based on the private data communication channel.
In some embodiments, after establishing a private data communication channel between a first user terminal and a target user terminal based on a live broadcasting interface, the device may be further configured to encrypt the processed interaction data with public key data corresponding to the target user terminal to obtain ciphertext data of the interaction data, send the ciphertext data of the interaction data to the target user terminal through the private data communication channel, wherein the target user terminal decrypts the ciphertext data of the interaction data with the private key data to obtain the processed interaction data, and display the processed interaction data in a current live broadcasting video displayed to the target user object through the live broadcasting interface.
In some embodiments, before receiving an interaction request initiated by a first user terminal, the apparatus may be further configured to receive a connection request initiated by the first user terminal when the apparatus is implemented, establish a first data connection with the first user terminal in response to the connection request, and transmit live video streaming data to the first user terminal through the first data connection, where the first user terminal displays a current live video to the first user through a live room interface according to the live video streaming data.
In some embodiments, the method may further comprise acquiring a characteristic parameter of the first user terminal when the live video stream data is transmitted to the first user terminal through the first data connection, determining a first language matching with the first viewer user according to the characteristic parameter of the first user terminal, detecting whether the live video stream data currently transmitted through the first data connection is live video stream data based on the first language, determining a cloud CDN cached with the live video stream data based on the first language as a first target CDN if it is determined that the live video stream data currently transmitted through the first data connection is not live video stream data based on the first language, and switching the first data connection to the first target CDN to transmit the live video stream data based on the first language to the first user terminal through the first data connection.
In some embodiments, the characteristic parameter of the first user terminal may specifically include at least one of an IP address of the first user terminal, a default language parameter of a browser of the first user terminal, cookie data of the first user terminal, and the like.
Referring to fig. 9, the embodiment of the present application further provides another apparatus for processing data in a live broadcast room, which is applied to a side of a first user terminal, and specifically may include the following structural modules:
The display module 901 may be specifically configured to display a current live video based on a first language to a first viewer user through a live room interface;
the receiving module 902 may be specifically configured to receive interaction data input by the first user through the live broadcast room interface, and determine a target user object for which the interaction data is directed;
the generating module 903 may be specifically configured to generate a corresponding interaction request according to the interaction data, where the interaction request further carries an object identifier of a target user object;
The sending module 904 may be specifically configured to send the interaction request to a cloud server, where the cloud server determines a target language matched with a target user object according to an object identifier of the target user object, processes interaction data based on the target language, and obtains processed interaction data, and the cloud server further displays the processed interaction data in a current live video displayed to the target user object through a live broadcasting interface of the target user terminal.
It should be noted that, the units, devices, or modules described in the above embodiments may be implemented by a computer chip or entity, or may be implemented by a product having a certain function. For convenience of description, the above devices are described as being functionally divided into various modules, respectively. Of course, when the present description is implemented, the functions of each module may be implemented in the same piece or pieces of software and/or hardware, or a module that implements the same function may be implemented by a plurality of sub-modules or a combination of sub-units, or the like. The above-described apparatus embodiments are merely illustrative, for example, the division of the units is merely a logical function division, and there may be additional divisions when actually implemented, for example, multiple units or components may be combined or integrated into another system, or some features may be omitted or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be an indirect coupling or communication connection via some interfaces, devices or units, which may be in electrical, mechanical or other form.
Therefore, based on the data processing device of the live broadcasting room provided by the embodiment of the application, the audience user can efficiently and conveniently interact with the target user objects using different languages in a diversified communication manner in the live broadcasting room, and the interaction experience of the audience user is improved. Furthermore, a private data communication channel exclusive to the target user object can be established in the living broadcast room for the audience user according to the needs of the audience user, and the audience user can perform safer and private communication interaction with the target user object in the living broadcast room interface directly based on the private data communication channel, so that the private data of the audience user is prevented from being revealed.
Although the present description provides method operational steps as described in the examples or flowcharts, more or fewer operational steps may be included based on conventional or non-inventive means. The order of steps recited in the embodiments is merely one way of performing the order of steps and does not represent a unique order of execution. When implemented by an apparatus or client product in practice, the methods illustrated in the embodiments or figures may be performed sequentially or in parallel (e.g., in a parallel processor or multi-threaded processing environment, or even in a distributed data processing environment). The terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, it is not excluded that additional identical or equivalent elements may be present in a process, method, article, or apparatus that comprises a described element. The terms first, second, etc. are used to denote a name, but not any particular order.
Those skilled in the art will also appreciate that, in addition to implementing the controller in a pure computer readable program code, it is well possible to implement the same functionality by logically programming the method steps such that the controller is in the form of logic gates, switches, application specific integrated circuits, programmable logic controllers, embedded microcontrollers, etc. Such a controller can be regarded as a hardware component, and means for implementing various functions included therein can also be regarded as a structure within the hardware component. Or even means for achieving the various functions may be regarded as either software modules implementing the methods or structures within hardware components.
The description may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, classes, etc. that perform particular tasks or implement particular abstract data types. The specification may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote computer-readable storage media including memory storage devices.
From the above description of embodiments, it will be apparent to those skilled in the art that the present description may be implemented in software plus a necessary general hardware platform. Based on such understanding, the technical solutions of the present specification may be embodied essentially in the form of a software product, which may be stored in a storage medium, such as a ROM/RAM, a magnetic disk, an optical disk, etc., and include several instructions to cause a computer device (which may be a personal computer, a mobile terminal, a server, or a network device, etc.) to perform the methods described in the various embodiments or portions of the embodiments of the present specification.
Various embodiments in this specification are described in a progressive manner, and identical or similar parts are all provided for each embodiment, each embodiment focusing on differences from other embodiments. The specification is operational with numerous general purpose or special purpose computer system environments or configurations. Such as a personal computer, a server computer, a hand-held or portable device, a tablet device, a multiprocessor system, a microprocessor-based system, a set top box, a programmable electronic device, a network PC, a minicomputer, a mainframe computer, a distributed computing environment that includes any of the above systems or devices, and the like.
Although the present specification has been described by way of example, it will be appreciated by those skilled in the art that there are many variations and modifications to the specification without departing from the spirit of the specification, and it is intended that the appended claims encompass such variations and modifications as do not depart from the spirit of the specification.

Claims (13)

Translated fromChinese
1.一种直播间的数据处理方法,其特征在于,应用于云服务器,包括:1. A data processing method for a live broadcast room, characterized in that it is applied to a cloud server and comprises:接收第一用户终端发起的互动请求;其中,所述互动请求至少携带有互动数据,以及该互动数据所针对的目标用户对象的对象标识;Receiving an interaction request initiated by a first user terminal; wherein the interaction request carries at least interaction data and an object identifier of a target user object targeted by the interaction data;根据目标用户对象的对象标识,确定出与目标用户相匹配的目标语言,以及目标用户对象所持有的目标用户终端;Determine, according to the object identifier of the target user object, a target language matching the target user and a target user terminal held by the target user object;根据互动数据的数据类型,从预设的处理规则集中,确定出相匹配的目标处理规则;According to the data type of the interactive data, a matching target processing rule is determined from a preset processing rule set;基于目标语言,根据目标处理规则,处理所述互动数据,得到处理后的互动数据;Based on the target language and according to the target processing rules, processing the interaction data to obtain processed interaction data;通过目标用户终端的直播间界面在向目标用户对象所展示的当前的直播视频中显示所述处理后的互动数据;Displaying the processed interactive data in the current live video displayed to the target user object through the live broadcast room interface of the target user terminal;在接收第一用户终端发起的互动请求之前,所述方法还包括:接收第一用户终端发起的连线请求;响应所述连线请求,建立与第一用户终端之间的第一数据连接;将直播视频流数据通过所述第一数据连接传输至第一用户终端;其中,所述第一用户终端根据所述直播视频流数据通过直播间界面向第一用户展示当前的直播视频;Before receiving the interaction request initiated by the first user terminal, the method further includes: receiving a connection request initiated by the first user terminal; responding to the connection request, establishing a first data connection with the first user terminal; transmitting the live video stream data to the first user terminal through the first data connection; wherein the first user terminal displays the current live video to the first user through the live room interface according to the live video stream data;在将直播视频流数据通过所述第一数据连接传输至第一用户终端的同时,所述方法还包括:采集第一用户终端的特征参数;根据第一用户终端的特征参数,确定出与第一观众用户相匹配的第一语言;检测当前通过第一数据连接传输的直播视频流数据是否为基于第一语言的直播视频流数据;在确定当前通过第一数据连接传输的直播视频流数据不是基于第一语言的直播视频流数据的情况下,确定出缓存有基于第一语言的直播视频流数据的云端CDN作为第一目标CDN;将第一数据连接切换到第一目标CDN,以将基于第一语言的直播视频流数据通过所述第一数据连接传输至第一用户终端。While transmitting the live video stream data to the first user terminal through the first data connection, the method also includes: collecting characteristic parameters of the first user terminal; determining a first language matching the first audience user based on the characteristic parameters of the first user terminal; detecting whether the live video stream data currently transmitted through the first data connection is live video stream data based on the first language; when it is determined that the live video stream data currently transmitted through the first data connection is not live video stream data based on the first language, determining a cloud CDN that caches live video stream data based on the first language as a first target CDN; switching the first data connection to the first target CDN to transmit the live video stream data based on the first language to the first user terminal through the first data connection.2.根据权利要求1所述的方法,其特征在于,所述互动数据的数据类型包括以下至少之一:文本数据、语音数据、表情图像。2. The method according to claim 1 is characterized in that the data type of the interactive data includes at least one of the following: text data, voice data, and expression images.3.根据权利要求2所述的方法,其特征在于,在所述互动数据的数据类型包括语音数据的情况下,基于目标语言,根据目标处理规则,处理所述互动数据,得到处理后的互动数据,包括:3. The method according to claim 2, characterized in that, when the data type of the interactive data includes voice data, based on the target language and according to the target processing rules, the interactive data is processed to obtain the processed interactive data, comprising:利用语音识别模型对所述语音数据进行语音识别,得到对应的文本数据;并从所述语音数据中提取出第一观众用户的语音特征;Performing speech recognition on the speech data using a speech recognition model to obtain corresponding text data; and extracting speech features of the first audience user from the speech data;利用与目标语言匹配的翻译模型处理文本数据,得到基于目标语言的文本数据;Processing text data using a translation model that matches the target language to obtain text data based on the target language;利用预设的语音合成模型根据第一观众用户的语音特征,处理所述基于目标语言的文本数据,得到对应的合成语音数据,作为所述处理后的互动数据。The target language-based text data is processed according to the speech characteristics of the first audience user using a preset speech synthesis model to obtain corresponding synthesized speech data as the processed interactive data.4.根据权利要求2所述的方法,其特征在于,在所述互动数据的数据类型包括表情图像的情况下,基于目标语言,根据目标处理规则,处理所述互动数据,得到处理后的互动数据,包括:4. The method according to claim 2, characterized in that, when the data type of the interactive data includes an expression image, based on the target language and according to the target processing rule, the interactive data is processed to obtain the processed interactive data, comprising:对表情图像进行字符检测,以确定表情图像中是否存在有意义的文本字符;Performing character detection on the expression image to determine whether there are meaningful text characters in the expression image;在确定表情图像中存在有意义的文本字符的情况下,利用图像识别模型对表情图像进行识别处理,以提取出表情图像中的文本字符作为文本数据;When it is determined that there are meaningful text characters in the expression image, the expression image is recognized by using an image recognition model to extract the text characters in the expression image as text data;利用与目标语言匹配的翻译模型处理文本数据,得到基于目标语言的文本数据;Processing text data using a translation model that matches the target language to obtain text data based on the target language;根据所述基于目标语言的文本数据,构建针对表情图像的标注数据;Constructing annotation data for the expression image according to the text data based on the target language;组合所述表情图像和标注数据,得到携带有标注数据的表情图像,作为所述处理后的互动数据。The expression image and the annotation data are combined to obtain the expression image carrying the annotation data as the processed interaction data.5.根据权利要求2所述的方法,其特征在于,在所述互动数据的数据类型包括文本数据的情况下,基于目标语言,根据目标处理规则,处理所述互动数据,得到处理后的互动数据,包括:5. The method according to claim 2, wherein, when the data type of the interactive data includes text data, processing the interactive data based on the target language and according to the target processing rules to obtain the processed interactive data comprises:利用与目标语言匹配的翻译模型处理文本数据,得到基于目标语言的文本数据;Processing text data using a translation model that matches the target language to obtain text data based on the target language;根据目标语言的文本数据,构建对应的弹幕数据,作为所述处理后的互动数据。According to the text data in the target language, corresponding bullet screen data is constructed as the processed interactive data.6.根据权利要求1所述的方法,其特征在于,所述方法还包括:6. The method according to claim 1, characterized in that the method further comprises:接收第一用户终端发起的私信请求;其中,所述私信请求至少携带有目标用户对象的对象标识;Receiving a private message request initiated by a first user terminal; wherein the private message request carries at least an object identifier of a target user object;响应所述私信请求,根据相关的加密通信协议,建立第一用户终端与目标用户终端之间的基于直播间界面的隐私数据通信通道;其中,所述云服务器基于该隐私数据通信通道通过目标用户终端的直播间界面在向目标用户对象所展示的当前的直播视频中显示所述处理后的互动数据。In response to the private message request, a private data communication channel based on the live broadcast room interface is established between the first user terminal and the target user terminal according to the relevant encrypted communication protocol; wherein the cloud server displays the processed interactive data in the current live broadcast video displayed to the target user object through the live broadcast room interface of the target user terminal based on the private data communication channel.7.根据权利要求6所述的方法,其特征在于,在建立第一用户终端与目标用户终端之间的基于直播间界面的隐私数据通信通道之后,所述方法还包括:7. The method according to claim 6, characterized in that after establishing the private data communication channel based on the live broadcast room interface between the first user terminal and the target user terminal, the method further comprises:利用与目标用户终端对应的公钥数据加密处理后的互动数据,得到互动数据的密文数据;The interactive data is encrypted using the public key data corresponding to the target user terminal to obtain ciphertext data of the interactive data;通过隐私数据通信通道,将互动数据的密文数据发送至目标用户终端;其中,目标用户终端利用私钥数据解密互动数据的密文数据,得到处理后的互动数据;并通过直播间界面在向目标用户对象所展示的当前的直播视频中显示所述处理后的互动数据。The ciphertext data of the interactive data is sent to the target user terminal through a private data communication channel; wherein the target user terminal decrypts the ciphertext data of the interactive data using the private key data to obtain the processed interactive data; and the processed interactive data is displayed in the current live video displayed to the target user object through the live broadcast room interface.8.根据权利要求1所述的方法,其特征在于,第一用户终端的特征参数包括以下至少之一:第一用户终端的IP地址、第一用户终端的浏览器的默认语言参数、第一用户终端的cookie数据。8. The method according to claim 1 is characterized in that the characteristic parameters of the first user terminal include at least one of the following: the IP address of the first user terminal, the default language parameter of the browser of the first user terminal, and the cookie data of the first user terminal.9.一种直播间的数据处理方法,其特征在于,应用于第一用户终端,包括:9. A data processing method for a live broadcast room, characterized in that it is applied to a first user terminal and comprises:通过直播间界面向第一观众用户展示基于第一语言的当前的直播视频;Displaying the current live video based on the first language to the first audience user through the live broadcast room interface;接收第一用户通过直播间界面输入的互动数据;并确定所述互动数据所针对的目标用户对象;Receiving interactive data input by the first user through the live broadcast room interface; and determining the target user object for the interactive data;根据所述互动数据,生成对应的互动请求;其中,所述互动请求还携带有目标用户对象的对象标识;Generate a corresponding interaction request according to the interaction data; wherein the interaction request also carries an object identifier of a target user object;将所述互动请求发送至云服务器;其中,所述云服务器根据目标用户对象的对象标识,确定出与目标用户对象相匹配的目标语言;并基于目标语言处理互动数据,得到处理后的互动数据;所述云服务器还通过目标用户终端的直播间界面在向目标用户对象所展示的当前的直播视频中显示所述处理后的互动数据;The interaction request is sent to a cloud server; wherein the cloud server determines a target language that matches the target user object according to the object identifier of the target user object; and processes the interaction data based on the target language to obtain processed interaction data; the cloud server also displays the processed interaction data in the current live video displayed to the target user object through the live room interface of the target user terminal;其中,所述方法还包括:云服务器接收第一用户终端发起的连线请求;响应所述连线请求,建立与第一用户终端之间的第一数据连接;将直播视频流数据通过所述第一数据连接传输至第一用户终端;所述第一用户终端根据所述直播视频流数据通过直播间界面向第一用户展示当前的直播视频;以及,云服务器还采集第一用户终端的特征参数;根据第一用户终端的特征参数,确定出与第一观众用户相匹配的第一语言;检测当前通过第一数据连接传输的直播视频流数据是否为基于第一语言的直播视频流数据;在确定当前通过第一数据连接传输的直播视频流数据不是基于第一语言的直播视频流数据的情况下,确定出缓存有基于第一语言的直播视频流数据的云端CDN作为第一目标CDN;将第一数据连接切换到第一目标CDN,以将基于第一语言的直播视频流数据通过所述第一数据连接传输至第一用户终端。The method further includes: the cloud server receives a connection request initiated by a first user terminal; in response to the connection request, establishes a first data connection with the first user terminal; transmits live video stream data to the first user terminal through the first data connection; the first user terminal displays the current live video to the first user through a live broadcast room interface based on the live video stream data; and the cloud server also collects characteristic parameters of the first user terminal; determines a first language matching the first audience user based on the characteristic parameters of the first user terminal; detects whether the live video stream data currently transmitted through the first data connection is live video stream data based on the first language; if it is determined that the live video stream data currently transmitted through the first data connection is not live video stream data based on the first language, determines a cloud-based CDN that caches live video stream data based on the first language as the first target CDN; switches the first data connection to the first target CDN to transmit the live video stream data based on the first language to the first user terminal through the first data connection.10.一种直播间的数据处理装置,其特征在于,应用于云服务器,包括:10. A data processing device for a live broadcast room, characterized in that it is applied to a cloud server and comprises:接收模块,用于接收第一用户终端发起的互动请求;其中,所述互动请求至少携带有互动数据,以及该互动数据所针对的目标用户对象的对象标识;A receiving module, configured to receive an interaction request initiated by a first user terminal; wherein the interaction request carries at least interaction data and an object identifier of a target user object targeted by the interaction data;第一确定模块,用于根据目标用户对象的对象标识,确定出与目标用户相匹配的目标语言,以及目标用户对象所持有的目标用户终端;A first determination module is used to determine a target language matching a target user and a target user terminal held by the target user object according to an object identifier of the target user object;第二确定模块,用于根据互动数据的数据类型,从预设的处理规则集中,确定出相匹配的目标处理规则;A second determination module is used to determine a matching target processing rule from a preset processing rule set according to the data type of the interactive data;处理模块,用于基于目标语言,根据目标处理规则,处理所述互动数据,得到处理后的互动数据;A processing module, used for processing the interactive data based on the target language and according to the target processing rules to obtain processed interactive data;触达模块,用于通过目标用户终端的直播间界面在向目标用户对象所展示的当前的直播视频中显示所述处理后的互动数据;A reach module, used for displaying the processed interactive data in the current live video displayed to the target user object through the live broadcast room interface of the target user terminal;在接收第一用户终端发起的互动请求之前,所述装置还用于:接收第一用户终端发起的连线请求;响应所述连线请求,建立与第一用户终端之间的第一数据连接;将直播视频流数据通过所述第一数据连接传输至第一用户终端;其中,所述第一用户终端根据所述直播视频流数据通过直播间界面向第一用户展示当前的直播视频;Before receiving the interaction request initiated by the first user terminal, the device is also used to: receive a connection request initiated by the first user terminal; respond to the connection request and establish a first data connection with the first user terminal; transmit the live video stream data to the first user terminal through the first data connection; wherein the first user terminal displays the current live video to the first user through the live room interface according to the live video stream data;在将直播视频流数据通过所述第一数据连接传输至第一用户终端的同时,所述装置还用于:采集第一用户终端的特征参数;根据第一用户终端的特征参数,确定出与第一观众用户相匹配的第一语言;检测当前通过第一数据连接传输的直播视频流数据是否为基于第一语言的直播视频流数据;在确定当前通过第一数据连接传输的直播视频流数据不是基于第一语言的直播视频流数据的情况下,确定出缓存有基于第一语言的直播视频流数据的云端CDN作为第一目标CDN;将第一数据连接切换到第一目标CDN,以将基于第一语言的直播视频流数据通过所述第一数据连接传输至第一用户终端。While transmitting the live video stream data to the first user terminal through the first data connection, the device is also used to: collect characteristic parameters of the first user terminal; determine a first language matching the first audience user based on the characteristic parameters of the first user terminal; detect whether the live video stream data currently transmitted through the first data connection is live video stream data based on the first language; when it is determined that the live video stream data currently transmitted through the first data connection is not live video stream data based on the first language, determine a cloud CDN that caches live video stream data based on the first language as the first target CDN; switch the first data connection to the first target CDN to transmit the live video stream data based on the first language to the first user terminal through the first data connection.11.一种服务器,其特征在于,包括处理器以及用于存储处理器可执行指令的存储器,所述处理器执行所述指令时实现权利要求1至8中任一项所述方法的步骤。11. A server, comprising a processor and a memory for storing processor-executable instructions, wherein the processor implements the steps of the method according to any one of claims 1 to 8 when executing the instructions.12.一种计算机可读存储介质,其特征在于,其上存储有计算机指令,所述指令被处理器执行时实现权利要求1至8,或9中任一项所述方法的步骤。12. A computer-readable storage medium, characterized in that computer instructions are stored thereon, and when the instructions are executed by a processor, the steps of the method described in any one of claims 1 to 8, or 9 are implemented.13.一种计算机程序产品,其特征在于,包含有计算机程序,所述计算机程序被处理器执行时实现权利要求1至8,或9中任一项所述方法的步骤。13. A computer program product, characterized in that it comprises a computer program, and when the computer program is executed by a processor, the steps of the method according to any one of claims 1 to 8 or 9 are implemented.
CN202211706177.4A2022-12-262022-12-26 Data processing method, device and server for live broadcast roomActiveCN116055756B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202211706177.4ACN116055756B (en)2022-12-262022-12-26 Data processing method, device and server for live broadcast room

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202211706177.4ACN116055756B (en)2022-12-262022-12-26 Data processing method, device and server for live broadcast room

Publications (2)

Publication NumberPublication Date
CN116055756A CN116055756A (en)2023-05-02
CN116055756Btrue CN116055756B (en)2025-06-24

Family

ID=86124786

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202211706177.4AActiveCN116055756B (en)2022-12-262022-12-26 Data processing method, device and server for live broadcast room

Country Status (1)

CountryLink
CN (1)CN116055756B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN108737845A (en)*2018-05-222018-11-02北京百度网讯科技有限公司Processing method, device, equipment and storage medium is broadcast live
CN112527168A (en)*2020-12-242021-03-19杭州网易云音乐科技有限公司Live broadcast interaction method and device, storage medium and electronic equipment
CN113179412A (en)*2021-03-152021-07-27北京城市网邻信息技术有限公司Information processing method and device, electronic equipment and storage medium
CN113301357A (en)*2020-07-272021-08-24阿里巴巴集团控股有限公司Live broadcast method and device and electronic equipment

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US11074398B2 (en)*2018-10-122021-07-27International Business Machines CorporationTracking and managing emoji annotations
CN113014986A (en)*2020-04-302021-06-22北京字节跳动网络技术有限公司Interactive information processing method, device, equipment and medium
CN113301363B (en)*2020-12-292023-04-07阿里巴巴集团控股有限公司Live broadcast information processing method and device and electronic equipment
CN114501042B (en)*2021-12-202024-10-18阿里巴巴(中国)网络技术有限公司Cross-border live broadcast processing method and electronic equipment

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN108737845A (en)*2018-05-222018-11-02北京百度网讯科技有限公司Processing method, device, equipment and storage medium is broadcast live
CN113301357A (en)*2020-07-272021-08-24阿里巴巴集团控股有限公司Live broadcast method and device and electronic equipment
CN112527168A (en)*2020-12-242021-03-19杭州网易云音乐科技有限公司Live broadcast interaction method and device, storage medium and electronic equipment
CN113179412A (en)*2021-03-152021-07-27北京城市网邻信息技术有限公司Information processing method and device, electronic equipment and storage medium

Also Published As

Publication numberPublication date
CN116055756A (en)2023-05-02

Similar Documents

PublicationPublication DateTitle
CN111736787B (en)Screen sharing method and device, storage medium and electronic equipment
CN106658200B (en) Method, device and terminal device for sharing and obtaining live video
CN108924662B (en)Continuous microphone interaction method, device, equipment and storage medium
JP6442441B2 (en) Translation system and translation method
CN105721462B (en)Information pushing method and device
US20160337290A1 (en)Message Push Method and Apparatus
CN112929253B (en)Virtual image interaction method and device
US20150227496A1 (en)Method and system for microblog resource sharing
CN111625308B (en)Information display method and device and electronic equipment
CN113568699B (en)Content display method, device, equipment and storage medium
WO2015043547A1 (en)A method, device and system for message response cross-reference to related applications
US20200244608A1 (en)Communication management system
CN103634690A (en)User information processing method, device and system in smart television
CN110647827A (en)Comment information processing method and device, electronic equipment and storage medium
WO2019011295A1 (en)Minus 1 screen content push and request method, apparatus, server and terminal
WO2020078050A1 (en)Comment information processing method and apparatus, and server, terminal and readable medium
CN108809800A (en)A kind of multimedia data processing method and its equipment, system
CN116264603A (en)Live broadcast information processing method, device, equipment and storage medium
CN113038192A (en)Video processing method and device, electronic equipment and storage medium
CN108574878B (en)Data interaction method and device
CN106055671B (en)Multimedia data processing method and equipment thereof
KR102646077B1 (en)Image advertising intermediation service system
CN104268611A (en)Webpage synchronizing method and device
CN110673886B (en)Method and device for generating thermodynamic diagrams
CN115270022A (en)Information display method and device, electronic equipment and storage medium

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp