Movatterモバイル変換


[0]ホーム

URL:


CN113207042A - Media asset playing method and display equipment - Google Patents

Media asset playing method and display equipment
Download PDF

Info

Publication number
CN113207042A
CN113207042ACN202110479497.XACN202110479497ACN113207042ACN 113207042 ACN113207042 ACN 113207042ACN 202110479497 ACN202110479497 ACN 202110479497ACN 113207042 ACN113207042 ACN 113207042A
Authority
CN
China
Prior art keywords
media asset
source type
media
sub
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110479497.XA
Other languages
Chinese (zh)
Other versions
CN113207042B (en
Inventor
卢可敬
李金昆
穆聪聪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hisense Visual Technology Co Ltd
Original Assignee
Hisense Visual Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hisense Visual Technology Co LtdfiledCriticalHisense Visual Technology Co Ltd
Priority to CN202110479497.XApriorityCriticalpatent/CN113207042B/en
Publication of CN113207042ApublicationCriticalpatent/CN113207042A/en
Application grantedgrantedCritical
Publication of CN113207042BpublicationCriticalpatent/CN113207042B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Images

Classifications

Landscapes

Abstract

The application discloses a media asset playing method and display equipment, after media asset data needing to be played are obtained, the source type of the media asset data is determined so as to determine a media asset server providing media assets, and a corresponding media asset playing strategy is executed. When the source type is the first source type, analyzing the media asset data of the first source type to obtain a broadcast control URL, and calling an audio output interface to play the media asset based on the broadcast control URL; and when the source type is the second source type, analyzing the media asset data of the second source type to obtain a media asset ID, acquiring a broadcast control URL corresponding to the media asset ID, and calling an audio output interface to play the media asset based on the broadcast control URL. Therefore, the method and the display device can judge the source type according to the acquired media asset data, unify the media asset playing strategies required by different media asset servers, perform unified media asset searching and playing control, realize timely playing of media assets, and have good user experience.

Description

Media asset playing method and display equipment
Technical Field
The present application relates to the technical field of media asset playing, and in particular, to a media asset playing method and a display device.
Background
Along with the rapid development of display equipment, the functions of the display equipment are more and more abundant, and the performance of the display equipment is more and more powerful, and at present, the display equipment comprises an intelligent television, a mobile phone, a refrigerator, other products with an intelligent display screen and the like. Applications for realizing different functions are configured in the display device so as to improve the user experience of the display device.
When the application of the media assets is configured in the display equipment, a user can search the media assets in a manual mode or a voice mode (the voice interaction function is configured on the display equipment), and then the audio output interface is called to play the media assets.
Because different media assets may come from different media asset servers, the source types of the media assets are different, and the media assets provided by the different media asset servers need to execute different media asset playing strategies. However, an audio output interface in the existing display device generally only supports a default playing policy, which may be different from the media asset playing policy required by each media asset server, and then after the display device acquires the media assets, a situation that the media assets cannot be played may exist.
Disclosure of Invention
The application provides a media asset playing method and display equipment, which aim to solve the problem that the existing display equipment cannot play media assets of different source types.
In a first aspect, the present application provides a display device comprising:
a display configured to present a user interface;
a controller connected with the display, the controller configured to:
acquiring media asset data corresponding to a user search media asset, wherein the source type of the media asset data is a first source type or a second source type, and the first source type and the second source type are used for representing different media asset servers;
if the source type is a first source type, analyzing the media asset data of the first source type to obtain a broadcast control URL, and calling an audio output interface to play the media asset based on the broadcast control URL;
if the source type is a second source type, analyzing the media asset data of the second source type to obtain a media asset ID, acquiring a broadcast control URL corresponding to the media asset ID, and calling an audio output interface to play the media asset based on the broadcast control URL.
In some embodiments of the present application, when the controller performs the parsing of the asset data of the second source type to obtain the asset ID if the source type is the second source type, the controller is further configured to:
when the source type is a second source type, acquiring content type information in the media asset data of the second source type;
if the content type information is null, analyzing the media asset data to obtain a media asset ID;
and if the content type information is not null, analyzing the media asset data to obtain a shell ID or a sub ID, wherein the shell ID is the ID of the media asset set, and the sub ID is the ID of a sub media asset in the media asset set.
In some embodiments of the present application, the controller, in performing the acquiring the play control URL corresponding to the media asset ID, is further configured to:
if the media asset ID is a shell ID, calling a shell switching sub-interface to convert the shell ID into a sub-ID, and acquiring a broadcast control URL corresponding to the sub-ID;
and if the media asset ID is the sub-ID, acquiring a broadcast control URL corresponding to the sub-ID.
In some embodiments of the present application, the controller, in performing the acquiring the broadcast URL corresponding to the sub-ID, is further configured to:
acquiring the media asset category information in the media asset data of the second source type;
if the media asset class information represents that the media asset is the specific media asset, calling an SDK (software development kit) package of the specific media asset, and acquiring a broadcast control URL (uniform resource locator) corresponding to the sub-ID based on the SDK package;
and if the media asset class information represents that the media assets are not specific media assets, calling a data interface of a media asset server to acquire a broadcast control URL corresponding to the sub-ID.
In some embodiments of the present application, when executing the SDK package invoking the specific media asset, the controller is further configured to, based on the SDK package, obtain a broadcast control URL corresponding to the sub ID:
when the media asset is a specific media asset, acquiring an equipment account ID and a token for data transmission between the display equipment and a media asset server of the specific media asset;
calling the SDK packet of the specific media asset, and transmitting the device account ID and the token into the SDK packet to establish data connection between the display device and the media asset server of the specific media asset;
and initializing the SDK packet, and if the SDK packet is successfully initialized, calling an SDK interface of the specific media resource to acquire a broadcast control URL corresponding to the sub-ID.
In some embodiments of the present application, when the controller performs the parsing of the asset data of the first source type to obtain the broadcast control URL if the source type is the first source type, the controller is further configured to:
when the source type is a first source type, analyzing the media asset data of the first source type to obtain a Uniform Resource Locator (URL) and media asset information for playing the media asset;
and generating a broadcasting control URL required by the audio output interface to execute the broadcasting action based on the URL and the media resource information.
In some embodiments of the present application, the controller, when executing the calling audio output interface to play the media asset based on the play control URL, is further configured to:
acquiring the media asset type information in the media asset data;
if the media asset type information represents that the media asset is of a first media asset type, acquiring at least one media asset corresponding to the broadcast control URL, and calling an audio output interface to broadcast at least one media asset;
and if the media asset type information represents that the media asset is of a second media asset type, calling an audio output interface to play the corresponding media asset based on the next playing control URL after playing the media asset corresponding to the playing control URL.
In some embodiments of the present application, the controller is further configured to, when performing the acquiring of the media asset data corresponding to the user search media asset:
responding to a voice search instruction generated when a user searches media assets in a voice mode, and sending the voice search instruction to a semantic engine, wherein the voice search instruction is used for indicating the semantic engine to acquire media asset data from a media asset server;
and receiving the returned media asset data after the semantic engine responds to the voice search instruction.
In some embodiments of the present application, the controller is further configured to:
when the audio output interface plays the media assets based on the sub-ID, responding to a voice control instruction generated by a user based on a voice mode, and calling a semantic engine to convert the voice control instruction into a voice control text;
if the voice control text is consistent with the sub-ID switching control text, acquiring a next sub-ID, and calling an audio output interface to play the corresponding media asset based on a playing control URL corresponding to the next sub-ID;
and if the voice control text is consistent with the shell ID switching control text, acquiring a next shell ID, acquiring a corresponding sub-ID based on the next shell ID, and calling an audio output interface to play the corresponding media assets based on the playing control URL corresponding to the sub-ID.
In a second aspect, the present application further provides a method for playing media assets, where the method includes:
acquiring media asset data corresponding to a user search media asset, wherein the source type of the media asset data is a first source type or a second source type, and the first source type and the second source type are used for representing different media asset servers;
if the source type is a first source type, analyzing the media asset data of the first source type to obtain a broadcast control URL, and calling an audio output interface to play the media asset based on the broadcast control URL;
if the source type is a second source type, analyzing the media asset data of the second source type to obtain a media asset ID, acquiring a broadcast control URL corresponding to the media asset ID, and calling an audio output interface to play the media asset based on the broadcast control URL.
In a third aspect, the present application further provides a storage medium, where the computer storage medium may store a program, and when the program is executed, the program may implement some or all of the steps in the embodiments of the media asset playing method provided in the present application.
According to the technical scheme, after the media asset data needing to be played is obtained, the source type of the media asset data is determined so as to determine a media asset server providing media assets, and a corresponding media asset playing strategy is executed. When the source type is the first source type, analyzing the media asset data of the first source type to obtain a broadcast control URL, and calling an audio output interface to play the media asset based on the broadcast control URL; and when the source type is the second source type, analyzing the media asset data of the second source type to obtain a media asset ID, acquiring a broadcast control URL corresponding to the media asset ID, and calling an audio output interface to play the media asset based on the broadcast control URL. Therefore, the method and the display device can judge the source type according to the acquired media asset data, unify the media asset playing strategies required by different media asset servers, perform unified media asset searching and playing control, realize timely playing of media assets, and have good user experience.
Drawings
In order to more clearly explain the technical solution of the present application, the drawings needed to be used in the embodiments will be briefly described below, and it is obvious to those skilled in the art that other drawings can be obtained according to the drawings without any creative effort.
FIG. 1 illustrates an operational scenario between a display device and a control apparatus according to some embodiments;
fig. 2 illustrates a hardware configuration block diagram of thecontrol apparatus 100 according to some embodiments;
fig. 3 illustrates a hardware configuration block diagram of thedisplay apparatus 200 according to some embodiments;
FIG. 4 illustrates a software configuration diagram in thedisplay device 200 according to some embodiments;
FIG. 5 illustrates an icon control interface display of an application indisplay device 200, in accordance with some embodiments;
FIG. 6 illustrates a network architecture diagram in accordance with some embodiments;
figure 7 illustrates a block diagram of a media asset processing terminal according to some embodiments;
FIG. 8 illustrates a method flow diagram of a method of media asset playback, according to some embodiments;
FIG. 9 illustrates a data flow diagram of a method of media asset playback according to some embodiments;
fig. 10 illustrates a data flow diagram for SDK package initialization for a dragonfly FM according to some embodiments.
Detailed Description
To make the purpose and embodiments of the present application clearer, the following will clearly and completely describe the exemplary embodiments of the present application with reference to the attached drawings in the exemplary embodiments of the present application, and it is obvious that the described exemplary embodiments are only a part of the embodiments of the present application, and not all of the embodiments.
It should be noted that the brief descriptions of the terms in the present application are only for the convenience of understanding the embodiments described below, and are not intended to limit the embodiments of the present application. These terms should be understood in their ordinary and customary meaning unless otherwise indicated.
The terms "first," "second," "third," and the like in the description and claims of this application and in the above-described drawings are used for distinguishing between similar or analogous objects or entities and not necessarily for describing a particular sequential or chronological order, unless otherwise indicated. It is to be understood that the terms so used are interchangeable under appropriate circumstances.
The terms "comprises" and "comprising," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a product or apparatus that comprises a list of elements is not necessarily limited to all elements expressly listed, but may include other elements not expressly listed or inherent to such product or apparatus.
The term "module" refers to any known or later developed hardware, software, firmware, artificial intelligence, fuzzy logic, or combination of hardware and/or software code that is capable of performing the functionality associated with that element.
FIG. 1 illustrates an operational scenario between a display device and a control apparatus according to some embodiments. As shown in fig. 1, a user may operate thedisplay apparatus 200 through the smart device 300 or thecontrol device 100.
In some embodiments, thecontrol apparatus 100 may be a remote controller, and the communication between the remote controller and the display device includes an infrared protocol communication or a bluetooth protocol communication, and other short-distance communication methods, and controls thedisplay device 200 in a wireless or wired manner. The user may input a user instruction through a key on a remote controller, voice input, control panel input, etc., to control thedisplay apparatus 200.
In some embodiments, the smart device 300 (e.g., mobile terminal, tablet, computer, laptop, etc.) may also be used to control thedisplay device 200. For example, thedisplay device 200 is controlled using an application program running on the smart device.
In some embodiments, thedisplay device 200 may also be controlled in a manner other than thecontrol apparatus 100 and the smart device 300, for example, the voice command control of the user may be directly received by a module configured inside thedisplay device 200 to obtain a voice command, or may be received by a voice control device provided outside thedisplay device 200.
In some embodiments, thedisplay device 200 is also in data communication with aserver 400. Thedisplay device 200 may be allowed to be communicatively connected through a Local Area Network (LAN), a Wireless Local Area Network (WLAN), and other networks. Theserver 400 may provide various contents and interactions to thedisplay apparatus 200. Theserver 400 may be a cluster or a plurality of clusters, and may include one or more types of servers.
Fig. 2 illustrates a hardware configuration block diagram of thecontrol apparatus 100 according to some embodiments. As shown in fig. 2, thecontrol device 100 includes acontroller 110, acommunication interface 130, a user input/output interface 140, a memory, and a power supply. Thecontrol apparatus 100 may receive an input operation instruction from a user and convert the operation instruction into an instruction recognizable and responsive by thedisplay device 200, serving as an interaction intermediary between the user and thedisplay device 200.
Fig. 3 illustrates a hardware configuration block diagram of adisplay device 200 according to some embodiments. In some embodiments, referring to fig. 3, thedisplay apparatus 200 includes at least one of atuner demodulator 210, acommunicator 220, a detector 230, anexternal device interface 240, acontroller 250, adisplay 260, anaudio output interface 270, a memory, a power supply, a user interface.
In some embodiments the controller comprises a processor, a video processor, an audio processor, a graphics processor, a RAM, a ROM, a first interface to an nth interface for input/output.
In some embodiments, thedisplay 260 includes a display screen component for presenting a picture, and a driving component for driving an image display, a component for receiving an image signal from the controller output, performing display of video content, image content, and a menu manipulation interface, and a user manipulation UI interface.
In some embodiments, thedisplay 260 may be a liquid crystal display, an OLED display, and a projection display, and may also be a projection device and a projection screen.
In some embodiments,communicator 220 is a component for communicating with external devices or servers according to various communication protocol types. For example: the communicator may include at least one of a Wifi module, a bluetooth module, a wired ethernet module, and other network communication protocol chips or near field communication protocol chips, and an infrared receiver. Thedisplay apparatus 200 may establish transmission and reception of control signals and data signals with theexternal control apparatus 100 or theserver 400 through thecommunicator 220.
In some embodiments, the user interface may be configured to receive control signals for controlling the apparatus 100 (e.g., an infrared remote control, etc.).
In some embodiments, the detector 230 is used to collect signals of the external environment or interaction with the outside. For example, detector 230 includes a light receiver, a sensor for collecting ambient light intensity; alternatively, the detector 230 includes an image collector, such as a camera, which may be used to collect external environment scenes, attributes of the user, or user interaction gestures, or the detector 230 includes a sound collector, such as a microphone, which is used to receive external sounds.
In some embodiments, theexternal device interface 240 may include, but is not limited to, the following: high Definition Multimedia Interface (HDMI), analog or data high definition component input interface (component), composite video input interface (CVBS), USB input interface (USB), RGB port, and the like. The interface may be a composite input/output interface formed by the plurality of interfaces.
In some embodiments, thetuner demodulator 210 receives broadcast television signals via wired or wireless reception, and demodulates audio/video signals, such as EPG data signals, from a plurality of wireless or wired broadcast television signals.
In some embodiments, thecontroller 250 and themodem 210 may be located in different separate devices, that is, themodem 210 may also be located in an external device of the main device where thecontroller 250 is located, such as an external set-top box.
In some embodiments, thecontroller 250 controls the operation of the display device and responds to user operations through various software control programs stored in memory. Thecontroller 250 controls the overall operation of thedisplay apparatus 200. For example: in response to receiving a user command for selecting a UI object to be displayed on thedisplay 260, thecontroller 250 may perform an operation related to the object selected by the user command.
In some embodiments, the object may be any one of selectable objects, such as a hyperlink, an icon, or other actionable control. The operations related to the selected object are: displaying an operation connected to a hyperlink page, document, image, or the like, or performing an operation of a program corresponding to the icon.
In some embodiments the controller comprises at least one of a Central Processing Unit (CPU), a video processor, an audio processor, a Graphics Processing Unit (GPU), a RAM Random Access Memory (RAM), a ROM (Read-Only Memory), a first to nth interface for input/output, a communication Bus (Bus), and the like.
A CPU processor. For executing operating system and application program instructions stored in the memory, and executing various application programs, data and contents according to various interactive instructions receiving external input, so as to finally display and play various audio-video contents. The CPU processor may include a plurality of processors. E.g. comprising a main processor and one or more sub-processors.
In some embodiments, a graphics processor for generating various graphics objects, such as: icons, operation menus, user input instruction display graphics, and the like. The graphic processor comprises an arithmetic unit, which performs operation by receiving various interactive instructions input by a user and displays various objects according to display attributes; the system also comprises a renderer for rendering various objects obtained based on the arithmetic unit, wherein the rendered objects are used for being displayed on a display.
In some embodiments, the video processor is configured to receive an external video signal, and perform video processing such as decompression, decoding, scaling, noise reduction, frame rate conversion, resolution conversion, and image synthesis according to a standard codec protocol of the input signal, so as to obtain a signal that can be displayed or played on thedirect display device 200.
In some embodiments, the video processor includes a demultiplexing module, a video decoding module, an image synthesis module, a frame rate conversion module, a display formatting module, and the like. The demultiplexing module is used for demultiplexing the input audio and video data stream. And the video decoding module is used for processing the video signal after demultiplexing, including decoding, scaling and the like. And the image synthesis module is used for carrying out superposition mixing processing on the GUI signal input by the user or generated by the user and the video image after the zooming processing by the graphic generator so as to generate an image signal for display. And the frame rate conversion module is used for converting the frame rate of the input video. And the display formatting module is used for converting the received video output signal after the frame rate conversion, and changing the signal to be in accordance with the signal of the display format, such as an output RGB data signal.
In some embodiments, the audio processor is configured to receive an external audio signal, decompress and decode the received audio signal according to a standard codec protocol of the input signal, and perform noise reduction, digital-to-analog conversion, and amplification processing to obtain an audio signal that can be played in the speaker.
In some embodiments, a user may enter user commands on a Graphical User Interface (GUI) displayed ondisplay 260, and the user input interface receives the user input commands through the Graphical User Interface (GUI). Alternatively, the user may input the user command by inputting a specific sound or gesture, and the user input interface receives the user input command by recognizing the sound or gesture through the sensor.
In some embodiments, a "user interface" is a media interface for interaction and information exchange between an application or operating system and a user that enables conversion between an internal form of information and a form that is acceptable to the user. A commonly used presentation form of the User Interface is a Graphical User Interface (GUI), which refers to a User Interface related to computer operations and displayed in a graphical manner. It may be an interface element such as an icon, a window, a control, etc. displayed in the display screen of the electronic device, where the control may include a visual interface element such as an icon, a button, a menu, a tab, a text box, a dialog box, a status bar, a navigation bar, a Widget, etc.
In some embodiments, a system of a display device may include a Kernel (Kernel), a command parser (shell), a file system, and an application program. The kernel, shell, and file system together make up the basic operating system structure that allows users to manage files, run programs, and use the system. After power-on, the kernel is started, kernel space is activated, hardware is abstracted, hardware parameters are initialized, and virtual memory, a scheduler, signals and interprocess communication (IPC) are operated and maintained. And after the kernel is started, loading the Shell and the user application program. The application program is compiled into machine code after being started, and a process is formed.
Fig. 4 illustrates a software configuration diagram in thedisplay device 200 according to some embodiments. Referring to fig. 4, in some embodiments, the system is divided into four layers, which are an Application (Applications) layer (abbreviated as "Application layer"), an Application Framework (Application Framework) layer (abbreviated as "Framework layer"), an Android runtime (Android runtime) and system library layer (abbreviated as "system runtime library layer"), and a kernel layer from top to bottom.
In some embodiments, at least one application program runs in the application program layer, and the application programs may be windows (windows) programs carried by an operating system, system setting programs, clock programs or the like; or an application developed by a third party developer. In particular implementations, the application packages in the application layer are not limited to the above examples.
The framework layer provides an Application Programming Interface (API) and a programming framework for the application. The application framework layer includes a number of predefined functions. The application framework layer acts as a processing center that decides to let the applications in the application layer act. The application program can access the resources in the system and obtain the services of the system in execution through the API interface.
As shown in fig. 4, in the embodiment of the present application, the application framework layer includes a manager (Managers), a Content Provider (Content Provider), and the like, where the manager includes at least one of the following modules: an Activity Manager (Activity Manager) is used for interacting with all activities running in the system; the Location Manager (Location Manager) is used for providing the system service or application with the access of the system Location service; a Package Manager (Package Manager) for retrieving various information related to an application Package currently installed on the device; a Notification Manager (Notification Manager) for controlling display and clearing of Notification messages; a Window Manager (Window Manager) is used to manage the icons, windows, toolbars, wallpapers, and desktop components on a user interface.
In some embodiments, the activity manager is used to manage the lifecycle of the various applications as well as general navigational fallback functions, such as controlling exit, opening, fallback, etc. of the applications. The window manager is used for managing all window programs, such as obtaining the size of a display screen, judging whether a status bar exists, locking the screen, intercepting the screen, controlling the change of the display window (for example, reducing the display window, displaying a shake, displaying a distortion deformation, and the like), and the like.
In some embodiments, the system runtime layer provides support for the upper layer, i.e., the framework layer, and when the framework layer is used, the android operating system runs the C/C + + library included in the system runtime layer to implement the functions to be implemented by the framework layer.
In some embodiments, the kernel layer is a layer between hardware and software. As shown in fig. 4, the core layer includes at least one of the following drivers: audio drive, display driver, bluetooth drive, camera drive, WIFI drive, USB drive, HDMI drive, sensor drive (like fingerprint sensor, temperature sensor, pressure sensor etc.) and power drive etc..
FIG. 5 illustrates an icon control interface display of an application indisplay device 200, according to some embodiments. In some embodiments, the display device may directly enter the interface of the preset vod program after being activated, and the interface of the vod program may include at least anavigation bar 510 and a content display area located below thenavigation bar 510, as shown in fig. 5, where the content displayed in the content display area may change according to the change of the selected control in the navigation bar. The programs in the application program layer can be integrated in the video-on-demand program and displayed through one control of the navigation bar, and can also be further displayed after the application control in the navigation bar is selected.
In some embodiments, the display device may directly enter a display interface of a signal source selected last time after being started, or a signal source selection interface, where the signal source may be a preset video-on-demand program, or may be at least one of an HDMI interface, a live tv interface, and the like, and after a user selects different signal sources, the display may display contents obtained from different signal sources.
FIG. 6 illustrates a network architecture diagram in accordance with some embodiments. Referring to fig. 6, the smart device is configured to receive input information and output a processing result of the information. The voice recognition service equipment is electronic equipment with voice recognition service deployed, the semantic service equipment is electronic equipment with semantic service deployed, and the business service equipment is electronic equipment with business service deployed. The electronic device may include a server, a computer, and the like, and the speech recognition service, the semantic service (also referred to as a semantic engine), and the business service are web services that can be deployed on the electronic device, wherein the speech recognition service is used for recognizing audio as text, the semantic service is used for semantic parsing of the text, and the business service is used for providing specific services such as a weather query service for ink weather, a music query service for QQ music, and the like. In one embodiment, there may be multiple entity service devices deployed with different business services in the architecture shown in fig. 6, and one or more function services may also be aggregated in one or more entity service devices.
In some embodiments, the following describes an example of a process for processing information input to a smart device based on the architecture shown in fig. 6, where the information input to the smart device is a query statement input by voice, the process may include the following three processes:
the voice recognition smart device may upload, after receiving a query sentence input by voice, audio of the query sentence to the voice recognition service device to be returned to the smart device after the audio is recognized as text by the voice recognition service device through the voice recognition service. In one embodiment, before uploading the audio of the query statement to the speech recognition service device, the smart device may perform denoising processing on the audio of the query statement, where the denoising processing may include removing echo and environmental noise.
The intelligent device uploads the text of the query sentence identified by the voice recognition service to the semantic service device, and the semantic service device performs semantic analysis on the text through semantic service to obtain the service field, intention and the like of the text.
According to the semantic parsing result of the text of the query statement, the semantic service equipment issues a gamma query instruction to the corresponding business service equipment to obtain the query result given by the business service. The intelligent device can obtain the query result from the semantic service device and output the query result. As an embodiment, the semantic service device may further send a semantic parsing result of the query statement to the intelligent device, so that the intelligent device outputs a feedback statement in the semantic parsing result.
It should be noted that the architecture shown in fig. 6 is only an example, and is not intended to limit the scope of the present application. In the embodiment of the present application, other architectures may also be adopted to implement similar functions, for example: all or part of the three processes can be completed by the intelligent terminal, and are not described herein.
In some embodiments, a media asset application is configured in the display device, which can provide different forms of media assets for the user, such as novels, books, broadcasts, music, and the like. When the target media assets are obtained, the display equipment sends a media asset request to the media asset server, and after media asset data returned by the media asset server are received, the target media assets are played.
In some embodiments, in order to enable the user to quickly play the media assets in which the user is interested, the target media assets can be obtained in a searching mode. The searching mode comprises a manual mode and a voice mode. When the manual searching is adopted, the user starts the media asset application and searches the target media assets by inputting the keywords.
When searching in a voice mode, a voice interaction function needs to be configured in the display device. At this time, a voice terminal application for recognizing a voice instruction input by the user is configured in the display device. The voice terminal is connected with the semantic engine, the voice terminal application sends the received voice instruction to the semantic engine, the semantic engine converts the voice instruction into a voice text, and the voice text is returned to the voice terminal application.
The voice terminal application sends a media asset searching request to the media asset server based on the voice text, the media asset server searches a result based on the media asset searching request, and returns the searched media asset data of the target media asset to the display device, so that the target media asset is played by the display device. The voice mode search can save the steps of manual operation of a user, and the interaction with the display device is more convenient.
In some embodiments, existing display devices typically support only the assets of a certain asset server, such as only the assets provided by the tenuous news. Therefore, the medium resource request is single, more medium resource requests cannot be provided for the user, and the user experience is not good.
In order to facilitate users to play more media assets by using the display equipment, the application provides the display equipment which can simultaneously support media assets under a plurality of media asset servers, not only support media assets under the situation of broadcast messages, but also support the display equipment platform to own media assets and third-party media assets, and the media assets are diversified. The third-party media assets are media assets provided by a third-party media asset application configured in the display device.
Because media resource playing strategies required by different media resource servers are different, in order to realize timely playing of different media resources under the multimedia resource servers, the display equipment provided by the application is provided with the media resource processing terminal, and the media resource processing terminal is used as a media resource fusion frame to fuse media resources from different sources. The media asset processing terminal can be adapted to different media asset servers, and when the display equipment platform obtains the third-party media assets, the display equipment platform can be in butt joint with the third-party media asset servers. Therefore, the media asset processing terminal can unify media asset playing strategies required by different media asset servers so as to carry out unified media asset searching and playing control, and the user experience is good.
In some embodiments, the different media asset servers can be divided into two categories, namely, the tengchining server and the server on which the display device platform is based. The tengcong jingdong server is used for providing the media resources of the tengcong jingdong; the server on which the display equipment platform is based comprises a display equipment platform and a third-party media asset server, wherein the display equipment platform is used for providing self media assets, and the third-party media asset server is used for providing third-party media assets. Based on this, the source type of the media asset data comprises an announcement category and a display device platform category.
Figure 7 illustrates a block diagram of a funding processing terminal, according to some embodiments. Referring to fig. 7, in order to perform a corresponding play policy based on different source types of assets, when searching for assets in a voice manner, a voice assistant (voice terminal application) configured in the display device sends an asset search request to a display device platform or an announcement server based on a voice instruction input by a user. And after the voice terminal application receives the returned media asset data after the display equipment platform or the announcement server is searched, the media asset data is transmitted to the media asset processing terminal.
When the media asset playing strategies are unified, the media asset processing terminal analyzes the media asset data to determine the media asset classification, namely, whether the source type of the media asset data is an announcement category or a display equipment platform category is determined. After the source type is determined, the media asset data is processed, for example, if the media asset data comes from the display device platform, media asset ID conversion is performed on the cloud server (display device platform) to obtain a broadcast control URL corresponding to the media asset ID, where the broadcast control URL is a network path for playing the media asset. And if the media asset data is from the announcement message, directly acquiring the URL from the type of the media asset data. After the broadcast control URL is acquired, the broadcast control URL is sent to a media asset playing terminal configured in the display device, for example, an audio output interface, and then the audio output interface can be called to play the corresponding media asset.
In some embodiments, the application scenes for searching and playing media assets based on the voice mode can be a display device in a speaker mode, a display device in a local bright screen mode, and a display device (media center) in a normal operation state.
The sound box mode belongs to a mode after the screen is turned off, and different from the common mode, the television screen is turned off, but the host computer still runs. Can realize quick switch screen under this audio amplifier mode, promote user experience. The speaker mode is that after the display device is turned off, the controller of the display device is still in a running state, so that the display device can still normally respond to the control instruction and still has a voice function.
The local bright screen mode is generally configured in an OLED type display device, and is similar to a screen saver mode, but is different from a normal screen saver mode in that a local bright screen of a television screen is used to display prompt contents, and most areas of the screen are in a black screen state. The location of the locally bright screen can be located at a corner of the display screen, such as the lower left corner, lower right corner, upper left corner or upper right corner, and the user interface in this scenario presents a "moon" effect that is bright in one corner. Under this local bright screen mode, display device's controller continuously operates, and the display is local bright screen mode, can realize the show and the quick switch screen of prompt message, promotes user experience. The speaker mode and the local bright screen mode are distinguished and only distinguished, one corner of the user interface in the local bright screen mode is provided with an interface, if the lower left corner is bright screen to display prompt content, the rest part is in a black screen state; and the user interfaces in the sound box mode are all in a black screen state.
The display device in the speaker mode and the display device in the local bright screen mode both can respond to the voice function in the off-screen state, while the display device in the normal operation state (the application of the media asset center normally operates) is in the on-screen state, and at this time, can also respond to the voice function. The user can carry out voice interaction with the display equipment under any scene, and the media asset is searched and played.
In some embodiments, a media asset processing terminal is configured in the display device, and when a user performs voice interaction with the display device to play media assets, the media asset processing terminal can process media asset data obtained based on voice search, and can unify media asset playing strategies required by different media asset servers to perform unified media asset search and play control, so that the user experience is good.
FIG. 8 illustrates a method flow diagram of a method of media asset playback, according to some embodiments; FIG. 9 illustrates a data flow diagram of a method of media asset playback according to some embodiments. An embodiment of the present invention provides a display device, including: a display configured to present a user interface; a controller connected to the display, the controller being configured to perform the following steps when executing the media asset playing method shown in fig. 8 and 9:
s1, acquiring the media asset data corresponding to the user search media asset, wherein the source type of the media asset data is a first source type or a second source type, and the first source type and the second source type are used for representing different media asset servers.
When the user plays the media assets by using the display equipment, the user can manually search the media assets or search the media assets in a voice mode. When a user searches media assets manually, a media asset application configured in the display equipment is started, keywords are manually input in a starting page, and media asset searching is carried out.
The display equipment generates a search request based on the keyword and sends the search request to the media asset server, the media asset server searches corresponding media asset data based on the keyword after receiving the search request, and finally returns the media asset data to the display equipment. The display equipment displays the media asset data in the display, and the playing can be realized after the playing button is clicked. The mode needs manual operation of a user, and the operation flow is more.
In some embodiments, to facilitate fast play of the assets, the user may search the assets in a voice manner. At this time, the controller is further configured to perform the following steps in acquiring the asset data corresponding to the user search asset:
and step 11, receiving a voice search instruction generated when a user searches media assets in a voice mode, and sending the voice search instruction to a semantic engine, wherein the voice search instruction is used for indicating the semantic engine to acquire media asset data from a media asset server.
And step 12, receiving the returned media asset data after the semantic engine responds to the voice search instruction.
When the user performs voice interaction with the display device in the voice function, the voice "i want to listen to XXX" is spoken by voice, for example, the user speaks the voice "i want to listen to the three kingdoms speech". At this time, the voice terminal application configured in the display device recognizes the voice search instruction and sends the voice search instruction to the semantic engine.
And after receiving the voice search instruction, the semantic engine converts the voice search instruction into a voice search text and sends a media asset search request carrying the voice search text to the media asset server. And after receiving the media asset searching request, the media asset server inquires the media asset data matched with the voice searching text, if the media asset data is found, the media asset data corresponding to the voice searching text is returned to the semantic engine, and finally, the semantic engine returns the media asset data to the voice terminal for application. For example, return the asset data of "the three kingdoms rehearsal" to the voice terminal application.
In some embodiments, the media asset server comprises an inventory server and a server upon which the display device platform is based. In order to obtain the media asset data in time, after receiving the voice search instruction, the semantic engine firstly sends a media asset search request to a server based on the display equipment platform, and if the server can provide the media asset data corresponding to the voice search instruction, the media asset data is returned to the semantic engine. If the servers according to the display equipment platform can not provide the media resource data corresponding to the voice search instruction, the semantic engine sends a media resource search request to the Tengchong Dian server so that the Tengchong Dian server provides the media resource data required by the user.
In some embodiments, to facilitate distinguishing the asset data from different servers, a source type field (susort) may be added to the asset data by each asset server when the respective asset server sends the asset data to the semantic engine. For example, the source type field (susort) represents whether the asset data is from the display device platform or the announcement, and if the data corresponding to the source type field is hisenseAudio, the asset data represents the display device platform (including own asset and third party asset); if the data corresponding to the source type field is tencntyaudio, the tencunent jingle is represented.
In some embodiments, to facilitate determining the asset classification, the source type of the asset data is divided into a first source type and a second source type, and the first source type and the second source type are used to characterize different asset servers that provide the asset data. The first source type may characterize that the asset data is from an announcement message appropriate server, and the second source type may characterize that the asset data is from a server upon which the display device platform is based.
Therefore, after the voice terminal application in the display equipment receives the media asset data returned by the semantic engine, the media asset data is sent to a media asset processing terminal configured in the display equipment, and the media asset processing terminal processes the media asset data and plays the media assets in time.
After receiving the media asset data, the media asset processing terminal analyzes the media asset data and acquires data of a source type field so as to judge whether the source of the media asset is a broadcast message or a display equipment platform. Two fields, "child" are parsed out from the media asset data: a source; "data": and (4) data. Based on the information of the two fields, the source of the media assets can be determined, so that whether the media assets needing to be played by the user are vacation news stung media assets or media assets of a display equipment platform can be determined.
For example, if the "data" is hisense audio, determining that the source type of the media asset is a second source type, that is, the media asset is the media asset of the display device platform; if the 'data' is tencentAudio, the source type of the medium resource is determined to be the first source type, namely the medium resource is the appropriate medium resource.
S2, if the source type is the first source type, analyzing the media asset data of the first source type to obtain a broadcast control URL, and calling the audio output interface to play the media asset based on the broadcast control URL.
Based on the foregoing embodiment, the source type field information in the asset data is determined, and if it is determined that the source type of the currently received asset data is the first source type, it indicates that the asset is an asset in which the announcement is granted. The media asset data of the announcement usually carries a URL for playing, so that the media asset data of the first source type can be analyzed to obtain the playing URL. And finally, sending the broadcast control URL to an audio output interface configured in the display equipment, and realizing the playing of the media assets by the audio output interface based on the broadcast control URL.
Since the URL format in the asset data provided by the announcement server may not be consistent with the URL format required by the audio output interface in the display device to implement the playing of the asset, the URLs in the asset data of the first source type need to be re-spliced into the URL of the Json format required by the audio output interface.
In some embodiments, the controller is further configured to, in response to parsing the asset data of the first source type to obtain the broadcast URL, if the source type is the first source type:
step 21, when the source type is the first source type, analyzing the media asset data of the first source type to obtain a URL (uniform resource locator) and media asset information for playing the media asset;
and step 22, generating a broadcasting control URL required by the audio output interface to execute the broadcasting action based on the URL and the media asset information.
When the media asset data received by the display equipment is from an announcement message server, a media asset processing terminal in the display equipment analyzes the media asset data of the first source type to acquire a Uniform Resource Locator (URL) and media asset information for playing media assets; based on the URL and the media information, a playing control URL required by the audio output interface when the media playing is realized is spliced.
In some embodiments, the format of the playback URL for playing back by the audio output interface is the Json format. The asset information may be obtained by parsing asset data, and includes asset type (obj _ child _ type), duration (duration), unit s, asset name (obj name), author (author), album (album), sub id (obj id), shell id (derived), and the like.
The asset type (obj _ child _ type) is "1" for streaming media of station type, "0" for program or reading, etc.; the radio station includes dragonfly FM, phoenix FM or Himalayan radio station program. The program or reading material can be a reading material for words such as novel, commentary, and speaking, and the reading material can be a reading material for words in applications such as listening to books or himalayan. The duration (duration) is the playing time of the media asset. The shell ID refers to the ID of the media asset collection, and the sub-ID refers to the ID of the sub-media assets in the media asset collection. The URL is the URL provided by the Tencent Ding server.
Based on the URL and each piece of media asset information, according to the sequence of the type, duration, name, author, album, sub-ID, shell ID and URL of the media assets, the broadcast control URL in the Json format can be spliced. And sending the broadcast control URL to an audio output interface to realize the playing of the media assets.
S3, if the source type is the second source type, analyzing the media asset data of the second source type to obtain a media asset ID, acquiring a broadcast control URL corresponding to the media asset ID, and calling an audio output interface to play the media asset based on the broadcast control URL.
Based on the foregoing embodiment, the source type field information in the media asset data is determined, and if it is determined that the source type of the currently received media asset data is the second source type, it is described that the media asset is the media asset of the display device platform. The media asset data of the display equipment platform does not include URL, and the URL for realizing media asset playing is stored in the media asset server. The URL and the media asset ID are in one-to-one correspondence, so that the corresponding broadcast control URL can be acquired from the media asset server according to the media asset ID.
At this time, the media asset processing terminal analyzes the media asset data of the second source type to obtain the media asset ID. And sending a request to the media asset server based on the media asset ID to obtain a corresponding broadcast control URL. And the media asset processing terminal sends the broadcast control URL to the audio output interface, and the audio output interface plays the media asset based on the broadcast control URL. The media asset server in the scene refers to a server on which the display device platform is based.
In some embodiments, the form of asset IDs in the asset data varies based on the type of asset. The asset ID refers to the number of the asset.
For the fiction and commentary type media assets, as the same total media asset structure comprises a plurality of child media assets, for example, for the fiction ' three countries ' lecture which has more than 100 collections, the three countries ' lecture itself is a shell and belongs to a father media asset; each set is a sub-asset, for example, the 001 th set of the three kingdoms rehearsal is a sub-asset.
For such a asset type (asset for which obj _ child _ type is "0"), the asset ID includes a shell ID and a sub ID. The total media asset structure is a media asset set which is a father media asset, the ID of the father media asset is a shell ID, and the shell ID is the ID of the media asset set. Each basic unit in the medium resource set is a sub medium resource, the ID of the sub medium resource is a sub ID, namely the sub ID refers to the ID of the sub medium resource in the medium resource set.
Corresponding to the media assets of the radio station and the broadcast class, because the broadcast is a media asset, the shell and the child are not divided. For this asset type (asset whose obj _ child _ type is "1"), the asset ID in the asset data is the ID of the asset.
In some embodiments, to facilitate accurate determination of asset types for assets, a determination may be made based on content type information, which is stored in the asset data.
Therefore, when acquiring various forms of asset IDs, the controller is executing, if the source type is the second source type, parsing the asset data of the second source type to obtain the asset IDs, and is further configured to execute the following steps:
and step 31, when the source type is the second source type, acquiring the content type information in the media asset data of the second source type.
And step 32, if the content type information is empty, analyzing the media asset data to obtain a media asset ID.
And step 33, if the content type information is not null, analyzing the media asset data to obtain a shell ID or a sub ID, wherein the shell ID refers to the ID of the media asset set, and the sub ID refers to the ID of the sub media asset in the media asset set.
In the foregoing embodiment, when it is determined that the asset data is an asset of the second source type, in order to accurately determine the asset type of the asset of the second source type, the asset data may be analyzed to obtain content type information (contentType).
If the content type information (contentType) is empty, it indicates that the asset is an asset without a shell or a sub-shell, and the asset is an independently existing asset, such as broadcast, station, etc. At this time, the media asset data is analyzed, and the obtained ID is the media asset ID of the media asset.
If the content type information (contentType) is not empty, the asset is a cased, sub-cased asset, such as a novel, narrative, or the like. At this time, the media asset data is analyzed, and the obtained ID is the shell ID or the sub ID.
In some embodiments, the shell ID or sub-ID in the asset data is associated with voice search instructions upon voice interaction by the user. If the media assets searched by the user are parent media assets, the media asset data comprises a shell ID; for example, the user speaks the voice "i want to hear the three kingdoms speech", at this time, the media asset "the three kingdoms speech" is a shell, and therefore, the shell ID is correspondingly stored in the media asset data.
If the media assets searched by the user are the sub media assets, the media asset data comprises a shell ID and a sub ID; for example, the user speaks a voice "i want to listen to the 1 st set of the three kingdoms speech meaning", at this time, the user search is more detailed, the media asset "the three kingdoms speech meaning" is a shell, and the media asset "the 1 st set of the three kingdoms speech meaning" is a child, so that the shell ID and the child ID are correspondingly stored in the media asset data.
In some embodiments, if the ID obtained by parsing the media asset data is a shell ID, it indicates that the media asset that the user wants to play is a parent media asset; and if the sub-ID exists in the ID obtained by analyzing the media asset data, the media asset which the user wants to play is the sub-media asset. If the ID obtained by analyzing the media asset data is the media asset ID, no score is given out, and the media asset which the user wants to play is a broadcast, a radio station and the like.
The ID with a shell is different in form from the ID without a shell. For example, the media with a shell score is "the three kingdoms rehearsal", the shell ID may be a shorter combination of numbers, such as "21829", and the child ID may be a longer combination of numbers, such as "136756617". The asset ID without a shell may be in the form of a combination of letters and numbers.
And obtaining three results by judging based on the content type information, and obtaining the media asset ID when the content type information is empty, namely obtaining the corresponding broadcast control URL from the media asset server based on the media asset ID. When the content type information is not empty, if the shell ID is acquired, in order to determine the specific sub-media assets needing to be played, the shell ID needs to be converted into the sub-ID, and then the corresponding playing control URL is acquired from the media asset server based on the sub-ID; if the sub-ID is obtained, the corresponding broadcast control URL can be directly obtained from the media asset server based on the sub-ID.
In some embodiments, when the content type information is empty, because the media asset ID can be directly acquired, in order to avoid an abnormal media asset playing process, in this scenario, the media asset processing terminal configures a foolproof design, that is, when a media asset ID without a shell is acquired, a process of replacing a shell ID with a child ID is also directly executed. At this time, a shell switching sub-interface of the cloud server (a server on which the display device platform is based) is called, the media asset ID is converted to obtain a sub-ID, and then the corresponding broadcast control URL is acquired from the media asset server based on the converted sub-ID. Because the media ID has no shell, even if the shell exchange sub-process is performed, the obtained sub-ID is the same as the media ID, and only form conversion is performed.
In some embodiments, when the content type information is not empty, it indicates that the media asset ID in the media asset data has a shell, and at this time, to prevent an exception from occurring when the shell ID is parsed, it may be determined whether the shell ID exists in the media asset data. And when the shell ID is judged to exist, analyzing the media asset data to acquire the shell ID or the sub-ID.
In judging whether or not the shell ID exists in the asset ID, the judgment may be made based on the content of the content type information (contentType), and the judgment logic mainly distinguishes which is the shell ID and which is the child ID. If the returned contentType is 19006, then there is a shell ID, representing a reading; if the returned contentType is 2001, there is a shell ID representing the parent asset in dragonfly FM. If the returned contentType is 2006, there is a child ID representing the child asset in dragonfly FM.
And after judging that the media asset data comprises the shell ID based on the content type information, analyzing the media asset data, wherein the obtained media asset ID can be the shell ID or the sub ID.
In order to obtain a corresponding broadcast control URL from the media asset server based on the media asset ID obtained from the media asset data, the broadcast control URL is generally obtained by using the sub-ID to obtain the sub-media asset to be played. At this time, the controller is further configured to, in performing acquiring the play control URL corresponding to the asset ID:
and step 34, if the media asset ID is the shell ID, calling a shell exchange sub-interface to convert the shell ID into the sub-ID, and acquiring the broadcast control URL corresponding to the sub-ID.
And step 35, if the media asset ID is the sub-ID, acquiring the broadcast control URL corresponding to the sub-ID.
And when the content type information is not empty, the fact that the media asset ID in the media asset data has a shell score is described. And if the shell ID is analyzed from the media asset data, the media asset required to be played by the user is indicated as a media asset set. In order to make sure the specific sub-assets needing to be played, the shell ID is converted into the sub-ID, and then the corresponding playing control URL is obtained from the media asset server based on the sub-ID.
In order to save time, the media asset processing terminal can acquire a plurality of sub-IDs from the cloud server at one time based on the shell ID. The media asset processing terminal acquires the corresponding sub-IDs from the cloud server based on the shell IDs, and can acquire 20 sub-IDs at the same time. The number of the sub-IDs obtained may be defined by the display device and the cloud server in advance, or may be other values, which is not particularly limited herein.
For example, if the user speaks a voice "i want to hear the three kingdoms speech", at this time, the shell ID exists in the asset data, and the asset processing terminal may simultaneously acquire the sub-ID corresponding to the top 20 sets of sub-assets from the cloud server.
After the sub-IDs of the sub-media assets capable of being played are determined, the playing control URL corresponding to each sub-ID can be acquired from a cloud server (a server on which the display device platform is based) based on the sub-IDs, and one sub-ID corresponds to one playing control URL.
In some embodiments, if the sub-ID is parsed from the asset data, it is indicated that the ID is the ID of the sub-asset, and it is indicated that the asset that the user needs to play is the sub-asset, and the playing can be directly performed. Therefore, the process of replacing the shell with the child is not needed, and the corresponding broadcast control URL is directly obtained from the cloud server based on the child ID.
For example, the user speaks the voice "i want to listen to the 1 st set of the three kingdoms speech", and the shell ID and the sub ID exist in the media asset data correspondingly. At this point, it may be determined that the sub-ID may be parsed from the asset data.
In some embodiments, after the sub ID of the playable sub asset is obtained based on the second source type of asset data, the corresponding play control URL may be obtained from the cloud server, so as to play the corresponding asset.
In some embodiments, since not all data in the server may be open to the display device platform, for example, the server of the dragonfly FM has no open right to the outside, i.e., the display device cannot directly acquire the data related to the dragonfly FM. And for other media resource servers, the display equipment can directly acquire corresponding data.
Therefore, when the corresponding broadcast control URL is acquired from the cloud server based on the sub-ID, it is first determined whether the media asset to be played is dragonfly FM, and the determination is based on the media asset category information in the media asset data.
In some embodiments, the controller, in executing step 34 or step 35, that is, acquiring the broadcast URL corresponding to the sub-ID, is further configured to execute the following steps:
361, obtaining the media asset class information in the media asset data of the second source type;
step 362, if the media asset class information represents that the media asset is the specific media asset, calling an SDK packet of the specific media asset, and acquiring a broadcast control URL corresponding to the sub-ID based on the SDK packet;
and step 363, if the media asset class information represents that the media asset is not the specific media asset, calling a data interface of the media asset server to acquire the broadcast control URL corresponding to the sub-ID.
When the media asset data processed by the media asset processing terminal is the data of the second source type, it is described that the media asset required to be played by the user is provided for the server on which the display device platform is based. And analyzing the media asset data to obtain media asset class information (vendor) in order to determine whether the server of the media asset has the permission of opening to the outside.
Because the display equipment is configured with the self-owned media assets and the third-party media assets, for the self-owned media assets, the display equipment can directly acquire the corresponding broadcast control URL based on the determined sub-ID. However, for some specific third-party assets, if a certain third-party asset server does not have an open right, the display device may not be able to directly obtain corresponding data from the third-party asset server. In order to enable the display equipment platform to realize the butt joint of the display equipment platform and the third-party media asset server when the display equipment platform acquires the third-party media assets, the display equipment platform can acquire the third-party media assets based on the SDK packet of the specific media assets.
And judging whether the media assets required to be played by the user are specific media assets or not according to the media asset category information, wherein the specific media assets can be dragonfly FM. The specific asset means an asset provided by an asset server which does not have data open to the outside.
For example, if the media asset class information (vendor) is 50001, it represents dragonfly FM; vendor 50004, then represents the Beacon; if the vendor is 50005, the lazy people listen to the book; vendor 50006, representing tenuous jingdong; vendor 50009, then represents phoenix FM; and the vector is 50016, and represents himalaya.
If the corresponding data is determined to be 5001 based on the media asset type information, the media asset which the user needs to play is the specific media asset (dragonfly FM). At this time, the display device cannot acquire the broadcast control URL to the media asset audio output interface of the media asset. Therefore, to achieve the acquisition of the broadcast control URL, the broadcast control URL corresponding to the sub-ID may be acquired based on the SDK package of the specific media asset.
Fig. 10 illustrates a data flow diagram for SDK package initialization for a dragonfly FM according to some embodiments. Referring to fig. 10, the controller, in executing the SDK package calling the specific media asset, obtaining the play control URL corresponding to the sub-ID based on the SDK package, is further configured to perform the following steps:
step 3621, when the media asset is a specific media asset, acquiring the device account ID, and displaying a token for data transmission between the device and the media asset server of the specific media asset.
Step 3622, invoking the SDK package of the specific media asset, and transmitting the device account ID and the token into the SDK package to establish a data connection between the display device and the media asset server of the specific media asset.
Step 3623, the SDK package is initialized, and if the SDK package is initialized successfully, the SDK interface of the specific media asset is called to obtain the broadcast control URL corresponding to the sub-ID.
After the SDK package of the specific media asset is obtained, in order to obtain the broadcast control URL corresponding to the sub-ID, it is necessary to first determine whether the SDK package of the specific media asset is initialized successfully. If the initialization is not successful, the initialization is required to be carried out again; if the initialization is successful, the broadcast control URL may be obtained.
In some embodiments, when the broadcast control URL is obtained from a server of a specific media asset, a connection between the display device and the server needs to be established first. For this purpose, it is first determined whether the network of the display device is connected. If not, waiting for the next successful connection of the network; and if the connection is made, acquiring a token for data transmission between the display equipment and the media resource server of the specific media resource.
The display equipment can log in the equipment account ID and establish connection with the specific media asset server through the equipment account ID; the display device may also operate the ID to establish a connection with a particular media asset server. The difference between the two is that the operation ID can only obtain free media resources from a specific media resource server, while the equipment account ID can be used as a user member ID and can obtain paid media resources from the specific media resource server. Therefore, the range of the media asset obtainable based on the device account ID is larger than the range of the media asset obtained based on the operation ID.
When the display equipment is authenticated with the specific media asset server, the equipment account ID and the token are written into the SDK packet of the specific media asset, and after the authentication is completed, the data connection between the display equipment and the media asset server of the specific media asset can be established, and at the moment, the display equipment can communicate with the specific media asset server to acquire related data.
And writing the device account ID and the token pair in the SDK packet of the specific media asset, and initializing the SDK packet. If the initialization is successful, the connection between the display equipment and the specific media asset server is established, at this time, the SDK interface of the specific media asset server is called, and the corresponding broadcast control URL is obtained from the specific media asset server based on the sub-ID.
If the initialization is not successful, a retry is required. To avoid always initializing, a maximum number of retries of 3 may be set. And after the current retry times reach 3 times, stopping the initialization process, and uploading error information to the cloud end to locate the error reason.
In some embodiments, if it is determined that the corresponding data is not 5001, but 5004, 50016, etc., based on the media asset type information, it indicates that the media asset that the user needs to play is not a specific media asset (dragonfly FM), but is bevacard song or himalaya, etc. At this time, the display device may directly obtain data from the media asset server, that is, call a data interface of the media asset server to obtain the broadcast control URL corresponding to the sub ID.
In some embodiments, if the broadcast control URL corresponding to a certain sub-ID is not stored in the media asset server, the URL acquired by the display device is empty, and at this time, in order to avoid that the user cannot know the reason why the sub-media asset is not played, a preset voice may be played through the audio output interface when the URL is not acquired, so as to prompt the user. The preset voice can be 'the playing address is acquired, and the current media assets cannot be played', and the like.
After the display device obtains the broadcast control URL based on the sub-ID, the broadcast control URL can be sent to the audio output interface, and the audio output interface plays the corresponding media assets based on the broadcast control URL.
And for different types of media assets, the playing strategies of the audio output interface are different. If the played media asset type is a radio station, because the radio station is independent media asset, a plurality of radio station programs exist in the same channel in the whole day time, and the radio station programs are provided by a radio station platform, and the corresponding radio station programs are automatically played when the corresponding time is reached. Therefore, in this scenario, the audio output interface will always play all the radio programs in the radio channel.
If the played media asset type is reading materials such as novels and the like, the media asset set and the sub media assets are distinguished, so that the fact that the playing of the media asset set is the playing of the sub media assets is achieved, and the playing of the sub media assets is based on the playing control URL corresponding to the sub ID. Therefore, in this scenario, after one sub-asset playing is completed, the media asset processing terminal needs to be notified to re-acquire the playing control URL.
In some embodiments, the controller in executing step S2 or S3, calling the audio output interface to play the asset based on the play URL, is further configured to execute the steps of:
and step 371, acquiring the media asset type information in the media asset data.
And 372, if the media asset type information represents that the media asset is of the first media asset type, acquiring at least one media asset corresponding to the broadcast control URL, and calling an audio output interface to broadcast the at least one media asset.
And step 373, if the asset type information indicates that the asset is of the second asset type, calling the audio output interface to play the corresponding asset based on the next broadcast control URL after playing the asset corresponding to the broadcast control URL.
In order to accurately determine the play strategy of the audio output interface based on the assets of different asset types, the asset type information (obj _ child _ type) can be obtained from the asset data. The asset data may be data of a first source type or data of a second source type.
If the asset type information (obj _ child _ type) is "1", it may be determined that the asset that the user needs to play is of a first asset type, such as streaming media of a station class, a station including a dragonfly FM, a phoenix FM, or a station program in himalaya, etc. At this time, because the corresponding radio station programs are played in time according to the time sequence under the same radio station channel, at least one media asset under the same broadcast control URL can be played after the audio output interface obtains the broadcast control URL.
If the asset type information (obj _ child _ type) is '0', the asset which the user needs to play can be determined to be a second asset type, such as representing a program or a reading material; the program or reading material can be a word reading material such as novel, narration, and speaking voice. At this time, after the audio output interface receives a broadcast control URL, the corresponding media asset is played. And after the media asset playing is finished, the media asset processing terminal needs to acquire the playing control URL of the next media asset, at this time, the next sub ID can be acquired, the new playing control URL is acquired again, and the audio output interface realizes the playing of the next media asset based on the new playing control URL.
For example, when the audio output interface plays the "three kingdoms rehearsal", if the currently played sub-media asset is the "1 st set of the three kingdoms rehearsal", at this time, the index value index of the sub-ID is 1; after the "three kingdoms rehearsal 1 st episode" is played, the "three kingdoms rehearsal 2 nd episode" needs to be played continuously, and at this time, the index value index of the child ID is 2. Therefore, to play the "2 nd set of the three kingdoms rehearsal", a new broadcast URL needs to be retrieved based on the sub-ID with index of 2. And then, the audio output interface realizes the playing of the '2 nd set of the three kingdoms rehearsal' based on the new playing control URL.
In some embodiments, when the user searches for the media assets in a voice manner and plays the media assets, in the playing process, if the user is not interested in the currently played media assets, the user can perform switching operation on the currently played media assets. The switching operation comprises the switching of the sub-media assets under the same media asset set, and can also be used for switching the media asset set.
In some embodiments, in performing the asset switch play, the controller is further configured to perform the steps of:
and step 41, when the audio output interface plays the media assets based on the sub-ID, responding to a voice control instruction generated by a user based on a voice mode, and calling a semantic engine to convert the voice control instruction into a voice control text.
And 42, if the voice control text is consistent with the sub-ID switching control text, acquiring the next sub-ID, and calling the audio output interface to play the corresponding media asset based on the playing control URL corresponding to the next sub-ID.
And 43, if the voice control text is consistent with the shell ID switching control text, acquiring the next shell ID, acquiring the corresponding sub-ID based on the next shell ID, and calling the audio output interface to play the corresponding media assets based on the playing control URL corresponding to the sub-ID.
When playing the media assets, the audio output interface usually needs to implement the media asset playing based on the playing control URL corresponding to the sub-ID, and at this time, the played media assets are independent media assets (radio station, broadcast, etc.) or sub-media assets (novel, etc.) under the combination of the media assets. The sub-ID in the scene may be a sub-ID obtained by converting a shell ID acquired from media data with a shell or a sub-ID directly acquired from the media data with a shell, or may be a sub-ID obtained by converting a media ID acquired from media data without a shell.
If the user wants to switch the played media assets during the playing process, a voice control instruction can be generated in a voice mode. After the voice terminal application of the display equipment acquires the voice control instruction, the voice control instruction is sent to the semantic engine, and the semantic engine converts the voice control instruction into a voice control text and then returns the voice control text to the voice terminal application.
The display device stores preset control texts for executing different switching playing strategies, wherein the preset control texts comprise a sub-ID switching control text and a shell ID switching control text.
And comparing the voice control text generated by the user with the control text of the preset switching playing strategy, and if the voice control text is consistent with the sub-ID switching control text, indicating that the user wants to switch the next sub-media asset currently playing the sub-media assets. At this time, the media asset processing terminal needs to acquire the sub ID corresponding to the next sub media asset, and re-acquire a new broadcast control URL, and the audio output interface plays the next media asset based on the new broadcast control URL.
For example, when the audio output interface plays "three kingdoms presentation", if the currently played sub-media asset is "1 st set of three kingdoms presentation", at this time, the index value index of the sub-ID is 1. At this time, if the user speaks "next" in voice, the voice control text is consistent with the sub-ID switching control text, and the switching playing of the sub-media asset is required, that is, "the 2 nd set of the three kingdoms' rehearsal" is required to be played. And the media asset processing terminal adds 1 to the index value of the sub-ID, namely, obtains the sub-ID with the index of 2, and obtains a new broadcast control URL based on the sub-ID with the index of 2. And finally, the audio output interface realizes the playing of the '2 nd set of the three kingdoms rehearsal' based on the new playing control URL, so that the switching playing of the sub-media resources is realized.
In some embodiments, if the speech control text is consistent with the shell ID switch control text, it indicates that the user wants to switch the currently playing set of assets. At this time, the media asset processing terminal needs to acquire the next shell ID, acquire the corresponding sub-ID based on the next shell ID again, acquire the corresponding new play control URL from the newly acquired sub-ID, and play the audio output interface based on the new play control URL to realize the playing of the sub-media assets in the next media asset set.
For example, when the user is performing a voice search for "three kingdoms," the set of assets that may be retrieved may be novels, commentary, and the like. When multiple media asset sets are obtained, the playing sequence of each media asset set can be preset, for example, playing the fiction of the three countries is played preferentially.
When the audio output interface plays the "three kingdoms rehearsal", the index value index of the shell ID is 1 at this time. If the currently played sub-media is the three-country rehearsal 1 st set, the user speaks "change one" by voice, the voice control text is consistent with the shell ID switching control text, and the media resource set needs to be switched and played, namely the "three-country rehearsal" needs to be played. And the media asset processing terminal adds 1 to the index value of the shell ID, namely obtains the shell ID with index of 2, then obtains the sub-ID corresponding to the shell ID with index of 2, and re-obtains the new broadcast control URL based on the new sub-ID. And finally, the audio output interface realizes the playing of 'explaining the three kingdoms speech' based on the new playing control URL. If the index of the new sub-ID is 1, the audio output interface realizes the playing of the 1 st set of the three kingdoms of commentary, thereby realizing the switching playing of the media asset set.
In some embodiments, the assets requested to be played by the user in the foregoing embodiments are audio-type assets, and if the assets requested to be played by the user are video-type assets, the processing method after the asset data is acquired is the same as the processing method of the audio-type asset data. The difference lies in that when playing, the audio data of the video media assets are sent to the audio output interface for playing, and the content data are sent to the display for displaying. The video-like assets include audio data and content data.
In some embodiments, if the media requested to be played by the user is a video-type media, and the current operation state of the display device is in the speaker mode or the local screen-on mode, in this scenario, since the display is in the screen-off state, in order to display the content data, the display needs to be controlled to be screen-on first. And displaying the content data in the display after the display is lightened.
Therefore, according to the display device provided by the embodiment of the invention, after the media asset data to be played is acquired, the source type of the media asset data is determined, so as to determine the media asset server providing the media asset, and execute the corresponding media asset playing strategy. When the source type is the first source type, analyzing the media asset data of the first source type to obtain a broadcast control URL, and calling an audio output interface to play the media asset based on the broadcast control URL; and when the source type is the second source type, analyzing the media asset data of the second source type to obtain a media asset ID, acquiring a broadcast control URL corresponding to the media asset ID, and calling an audio output interface to play the media asset based on the broadcast control URL. Therefore, the display equipment judges the source type according to the acquired media asset data, and can unify media asset playing strategies required by different media asset servers so as to carry out unified media asset searching and playing control, realize timely playing of media assets and achieve good user experience.
FIG. 8 illustrates a method flow diagram of a method of media asset playback, according to some embodiments. Referring to fig. 8, an embodiment of the present invention provides a media asset playing method, which is executed by a controller in a display device provided in the foregoing embodiment, and the method includes:
s1, acquiring media asset data corresponding to a user search media asset, wherein the source type of the media asset data is a first source type or a second source type, and the first source type and the second source type are used for representing different media asset servers;
s2, if the source type is a first source type, analyzing the media asset data of the first source type to obtain a broadcast control URL, and calling an audio output interface to play the media asset based on the broadcast control URL;
s3, if the source type is a second source type, analyzing the media asset data of the second source type to obtain a media asset ID, acquiring a broadcast control URL corresponding to the media asset ID, and calling an audio output interface to play the media asset based on the broadcast control URL.
According to the technical scheme, after the media asset data needing to be played is obtained, the source type of the media asset data is determined so as to determine a media asset server providing media assets, and a corresponding media asset playing strategy is executed. When the source type is the first source type, analyzing the media asset data of the first source type to obtain a broadcast control URL, and calling an audio output interface to play the media asset based on the broadcast control URL; and when the source type is the second source type, analyzing the media asset data of the second source type to obtain a media asset ID, acquiring a broadcast control URL corresponding to the media asset ID, and calling an audio output interface to play the media asset based on the broadcast control URL. Therefore, the method and the display device can judge the source type according to the acquired media asset data, unify the media asset playing strategies required by different media asset servers, perform unified media asset searching and playing control, realize timely playing of media assets, and have good user experience.
In a specific implementation manner, the present invention further provides a computer storage medium, where the computer storage medium may store a program, and when the program is executed, the program may include some or all of the steps in each embodiment of the media asset playing method provided by the present invention. The storage medium may be a magnetic disk, an optical disk, a read-only memory (ROM) or a Random Access Memory (RAM).
Those skilled in the art will readily appreciate that the techniques of the embodiments of the present invention may be implemented as software plus a required general purpose hardware platform. Based on such understanding, the technical solutions in the embodiments of the present invention may be essentially or partially implemented in the form of a software product, which may be stored in a storage medium, such as ROM/RAM, magnetic disk, optical disk, etc., and includes several instructions for enabling a computer device (which may be a personal computer, a server, or a network device, etc.) to execute the method according to the embodiments or some parts of the embodiments.
The same and similar parts in the various embodiments in this specification may be referred to each other. Particularly, for the embodiment of the method for playing the media assets, since the method is basically similar to the embodiment of the display device, the description is relatively simple, and relevant points can be referred to the description in the embodiment of the display device.
Finally, it should be noted that: the above embodiments are only used for illustrating the technical solutions of the present application, and not for limiting the same; although the present application has been described in detail with reference to the foregoing embodiments, it should be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; and the modifications or the substitutions do not make the essence of the corresponding technical solutions depart from the scope of the technical solutions of the embodiments of the present application.
The foregoing description, for purposes of explanation, has been presented in conjunction with specific embodiments. However, the illustrative discussions above are not intended to be exhaustive or to limit the embodiments to the precise forms disclosed above. Many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to best explain the principles and the practical application, to thereby enable others skilled in the art to best utilize the embodiments and various embodiments with various modifications as are suited to the particular use contemplated.

Claims (10)

CN202110479497.XA2021-04-302021-04-30Media asset playing method and display equipmentActiveCN113207042B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202110479497.XACN113207042B (en)2021-04-302021-04-30Media asset playing method and display equipment

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202110479497.XACN113207042B (en)2021-04-302021-04-30Media asset playing method and display equipment

Publications (2)

Publication NumberPublication Date
CN113207042Atrue CN113207042A (en)2021-08-03
CN113207042B CN113207042B (en)2022-12-09

Family

ID=77028221

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202110479497.XAActiveCN113207042B (en)2021-04-302021-04-30Media asset playing method and display equipment

Country Status (1)

CountryLink
CN (1)CN113207042B (en)

Citations (11)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN102802031A (en)*2012-07-132012-11-28李映红Interactive system and method in allusion to television programs
CN103024477A (en)*2011-09-262013-04-03中兴通讯股份有限公司Method, device and system for implementing IPTV (internet protocol television) system advertisements
US20150304701A1 (en)*2014-04-172015-10-22Hisense Co., LtdPlay control method and device
CN106791974A (en)*2016-12-272017-05-31Ut斯达康(深圳)技术有限公司Matchmaker money delivery method and the device of a kind of IPTV service system
CN111182357A (en)*2019-09-292020-05-19腾讯科技(深圳)有限公司Internet media playing method, device and system based on interactive network television
CN111385633A (en)*2018-12-272020-07-07Tcl集团股份有限公司Resource searching method based on voice, intelligent terminal and storage medium
CN111405318A (en)*2020-03-242020-07-10聚好看科技股份有限公司 Video display method, device and computer storage medium
CN111935510A (en)*2020-08-182020-11-13海信视像科技股份有限公司Double-browser application loading method and display equipment
CN112000820A (en)*2020-08-102020-11-27海信电子科技(武汉)有限公司Media asset recommendation method and display device
CN112004131A (en)*2020-08-122020-11-27海信电子科技(武汉)有限公司Display system
CN112669838A (en)*2020-12-172021-04-16合肥飞尔智能科技有限公司Intelligent sound box audio playing method and device, electronic equipment and storage medium

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN103024477A (en)*2011-09-262013-04-03中兴通讯股份有限公司Method, device and system for implementing IPTV (internet protocol television) system advertisements
CN102802031A (en)*2012-07-132012-11-28李映红Interactive system and method in allusion to television programs
US20150304701A1 (en)*2014-04-172015-10-22Hisense Co., LtdPlay control method and device
CN106791974A (en)*2016-12-272017-05-31Ut斯达康(深圳)技术有限公司Matchmaker money delivery method and the device of a kind of IPTV service system
CN111385633A (en)*2018-12-272020-07-07Tcl集团股份有限公司Resource searching method based on voice, intelligent terminal and storage medium
CN111182357A (en)*2019-09-292020-05-19腾讯科技(深圳)有限公司Internet media playing method, device and system based on interactive network television
CN111405318A (en)*2020-03-242020-07-10聚好看科技股份有限公司 Video display method, device and computer storage medium
CN112000820A (en)*2020-08-102020-11-27海信电子科技(武汉)有限公司Media asset recommendation method and display device
CN112004131A (en)*2020-08-122020-11-27海信电子科技(武汉)有限公司Display system
CN111935510A (en)*2020-08-182020-11-13海信视像科技股份有限公司Double-browser application loading method and display equipment
CN112669838A (en)*2020-12-172021-04-16合肥飞尔智能科技有限公司Intelligent sound box audio playing method and device, electronic equipment and storage medium

Also Published As

Publication numberPublication date
CN113207042B (en)2022-12-09

Similar Documents

PublicationPublication DateTitle
CN112163086B (en)Multi-intention recognition method and display device
CN111405318B (en) Video display method, device and computer storage medium
CN112885354B (en)Display device, server and display control method based on voice
CN112004157B (en)Multi-round voice interaction method and display device
CN115776585A (en)Display device and content presentation method
CN114116622A (en)Display device and file display method
CN112804567A (en)Display device, server and video recommendation method
CN112885347A (en)Voice control method of display device, display device and server
CN111866568A (en) A display device, server and voice-based video collection acquisition method
CN113490057B (en)Display device and media asset recommendation method
CN112911381B (en)Display device, mode adjustment method, device and medium
CN113079400A (en)Display device, server and voice interaction method
CN111914114B (en) A badcase mining method and electronic device
CN113066491A (en)Display device and voice interaction method
CN112601116A (en)Display device and content display method
CN113207042B (en)Media asset playing method and display equipment
CN113473220B (en)Automatic sound effect starting method and display equipment
CN113035194B (en)Voice control method, display device and server
CN114155846B (en) A semantic slot extraction method and display device
CN113038217A (en)Display device, server and response language generation method
CN115701105A (en) Display device, server and voice interaction method
CN114296842A (en)Display device and scrolling text detection method
CN111914565A (en) Electronic device and method for processing user sentences
CN113076427B (en)Media resource searching method, display equipment and server
CN113825007B (en)Video playing method and device and display equipment

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp