Detailed Description
To make the purpose and embodiments of the present application clearer, the following will clearly and completely describe the exemplary embodiments of the present application with reference to the attached drawings in the exemplary embodiments of the present application, and it is obvious that the described exemplary embodiments are only a part of the embodiments of the present application, and not all of the embodiments.
It should be noted that the brief descriptions of the terms in the present application are only for the convenience of understanding the embodiments described below, and are not intended to limit the embodiments of the present application. These terms should be understood in their ordinary and customary meaning unless otherwise indicated.
The terms "first," "second," "third," and the like in the description and claims of this application and in the above-described drawings are used for distinguishing between similar or analogous objects or entities and not necessarily for describing a particular sequential or chronological order, unless otherwise indicated. It is to be understood that the terms so used are interchangeable under appropriate circumstances.
The terms "comprises" and "comprising," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a product or apparatus that comprises a list of elements is not necessarily limited to all elements expressly listed, but may include other elements not expressly listed or inherent to such product or apparatus.
The term "module" refers to any known or later developed hardware, software, firmware, artificial intelligence, fuzzy logic, or combination of hardware or/and software code that is capable of performing the functionality associated with that element.
Fig. 1 is a schematic diagram of a usage scenario of a display device according to an embodiment. As shown in fig. 1, thedisplay apparatus 200 is also in data communication with aserver 400, and a user can operate thedisplay apparatus 200 through thesmart device 300 or thecontrol device 100.
In some embodiments, thecontrol apparatus 100 may be a remote controller, and the communication between the remote controller and the display device includes at least one of an infrared protocol communication or a bluetooth protocol communication, and other short-distance communication methods, and controls thedisplay device 200 in a wireless or wired manner. The user may control thedisplay apparatus 200 by inputting a user instruction through at least one of a key on a remote controller, a voice input, a control panel input, and the like.
In some embodiments, thesmart device 300 may include any of a mobile terminal 300A, a tablet, a computer, a laptop, an AR/VR device, and the like.
In some embodiments, thesmart device 300 may also be used to control thedisplay device 200. For example, thedisplay device 200 is controlled using an application program running on the smart device.
In some embodiments, thesmart device 300 and the display device may also be used for communication of data.
In some embodiments, thedisplay device 200 may also be controlled in a manner other than thecontrol apparatus 100 and thesmart device 300, for example, the voice instruction control of the user may be directly received by a module configured inside thedisplay device 200 to obtain a voice instruction, or may be received by a voice control apparatus provided outside thedisplay device 200.
In some embodiments, thedisplay device 200 is also in data communication with aserver 400. Thedisplay device 200 may be allowed to be communicatively connected through a Local Area Network (LAN), a Wireless Local Area Network (WLAN), and other networks. Theserver 400 may provide various contents and interactions to thedisplay apparatus 200. Theserver 400 may be a cluster or a plurality of clusters, and may include one or more types of servers.
In some embodiments, software steps executed by one step execution agent may be migrated on demand to another step execution agent in data communication therewith for execution. Illustratively, software steps performed by the server may be migrated to be performed on a display device in data communication therewith, and vice versa, as desired.
Fig. 2 exemplarily shows a block diagram of a configuration of thecontrol apparatus 100 according to an exemplary embodiment. As shown in fig. 2, thecontrol device 100 includes acontroller 110, acommunication interface 130, a user input/output interface 140, a memory, and a power supply. Thecontrol apparatus 100 may receive an input operation instruction from a user and convert the operation instruction into an instruction recognizable and responsive by thedisplay device 200, serving as an interaction intermediary between the user and thedisplay device 200.
In some embodiments, thecommunication interface 130 is used for external communication, and includes at least one of a WIFI chip, a bluetooth module, NFC, or an alternative module.
In some embodiments, the user input/output interface 140 includes at least one of a microphone, a touchpad, a sensor, a key, or an alternative module.
Fig. 3 shows a hardware configuration block diagram of thedisplay apparatus 200 according to an exemplary embodiment.
In some embodiments, thedisplay apparatus 200 includes at least one of atuner demodulator 210, acommunicator 220, adetector 230, anexternal device interface 240, acontroller 250, adisplay 260, anaudio output interface 270, a memory, a power supply, a user interface.
In some embodiments the controller comprises a central processor, a video processor, an audio processor, a graphics processor, a RAM, a ROM, a first interface to an nth interface for input/output.
In some embodiments, thedisplay 260 includes a display screen component for displaying pictures, and a driving component for driving image display, a component for receiving image signals from the controller output, displaying video content, image content, and menu manipulation interface, and a user manipulation UI interface, etc.
In some embodiments, thedisplay 260 may be at least one of a liquid crystal display, an OLED display, and a projection display, and may also be a projection device and a projection screen.
In some embodiments, thetuner demodulator 210 receives broadcast television signals via wired or wireless reception, and demodulates audio/video signals, such as EPG data signals, from a plurality of wireless or wired broadcast television signals.
In some embodiments,communicator 220 is a component for communicating with external devices or servers according to various communication protocol types. For example: the communicator may include at least one of a Wifi module, a bluetooth module, a wired ethernet module, and other network communication protocol chips or near field communication protocol chips, and an infrared receiver. Thedisplay apparatus 200 may establish transmission and reception of control signals and data signals with thecontrol device 100 or theserver 400 through thecommunicator 220.
In some embodiments, thedetector 230 is used to collect signals of the external environment or interaction with the outside. For example,detector 230 includes a light receiver, a sensor for collecting ambient light intensity; alternatively, thedetector 230 includes an image collector, such as a camera, which may be used to collect external environment scenes, attributes of the user, or user interaction gestures, or thedetector 230 includes a sound collector, such as a microphone, which is used to receive external sounds.
In some embodiments, theexternal device interface 240 may include, but is not limited to, the following: high Definition Multimedia Interface (HDMI), analog or data high definition component input interface (component), composite video input interface (CVBS), USB input interface (USB), RGB port, and the like. The interface may be a composite input/output interface formed by the plurality of interfaces.
In some embodiments, thecontroller 250 and themodem 210 may be located in different separate devices, that is, themodem 210 may also be located in an external device of the main device where thecontroller 250 is located, such as an external set-top box.
In some embodiments, thecontroller 250 controls the operation of the display device and responds to user operations through various software control programs stored in memory. Thecontroller 250 controls the overall operation of thedisplay apparatus 200. For example: in response to receiving a user command for selecting a UI object to be displayed on thedisplay 260, thecontroller 250 may perform an operation related to the object selected by the user command.
In some embodiments, the object may be any one of selectable objects, such as a hyperlink, an icon, or other actionable control. The operations related to the selected object are: displaying an operation connected to a hyperlink page, document, image, or the like, or performing an operation of a program corresponding to the icon.
In some embodiments the controller comprises at least one of a Central Processing Unit (CPU), a video processor, an audio processor, a Graphics Processing Unit (GPU), a RAM Random Access Memory (RAM), a ROM (Read-Only Memory), a first to nth interface for input/output, a communication Bus (Bus), and the like.
A CPU processor. For executing operating system and application program instructions stored in the memory, and executing various application programs, data and contents according to various interactive instructions receiving external input, so as to finally display and play various audio-video contents. The CPU processor may include a plurality of processors. E.g. comprising a main processor and one or more sub-processors.
In some embodiments, a graphics processor for generating various graphics objects, such as: at least one of an icon, an operation menu, and a user input instruction display figure. The graphic processor comprises an arithmetic unit, which performs operation by receiving various interactive instructions input by a user and displays various objects according to display attributes; the system also comprises a renderer for rendering various objects obtained based on the arithmetic unit, wherein the rendered objects are used for being displayed on a display.
In some embodiments, the video processor is configured to receive an external video signal, and perform at least one of video processing such as decompression, decoding, scaling, noise reduction, frame rate conversion, resolution conversion, and image synthesis according to a standard codec protocol of the input signal, so as to obtain a signal displayed or played on thedirect display device 200.
In some embodiments, the video processor includes at least one of a demultiplexing module, a video decoding module, an image composition module, a frame rate conversion module, a display formatting module, and the like. The demultiplexing module is used for demultiplexing the input audio and video data stream. And the video decoding module is used for processing the video signal after demultiplexing, including decoding, scaling and the like. And the image synthesis module is used for carrying out superposition mixing processing on the GUI signal input by the user or generated by the user and the video image after the zooming processing by the graphic generator so as to generate an image signal for display. And the frame rate conversion module is used for converting the frame rate of the input video. And the display formatting module is used for converting the received video output signal after the frame rate conversion, and changing the signal to be in accordance with the signal of the display format, such as an output RGB data signal.
In some embodiments, the audio processor is configured to receive an external audio signal, decompress and decode the received audio signal according to a standard codec protocol of the input signal, and perform at least one of noise reduction, digital-to-analog conversion, and amplification processing to obtain a sound signal that can be played in the speaker.
In some embodiments, a user may enter user commands on a Graphical User Interface (GUI) displayed ondisplay 260, and the user input interface receives the user input commands through the Graphical User Interface (GUI). Alternatively, the user may input the user command by inputting a specific sound or gesture, and the user input interface receives the user input command by recognizing the sound or gesture through the sensor.
In some embodiments, a "user interface" is a media interface for interaction and information exchange between an application or operating system and a user that enables conversion between an internal form of information and a form that is acceptable to the user. A commonly used presentation form of the User Interface is a Graphical User Interface (GUI), which refers to a User Interface related to computer operations and displayed in a graphical manner. It may be an interface element such as an icon, a window, a control, etc. displayed in the display screen of the electronic device, where the control may include at least one of an icon, a button, a menu, a tab, a text box, a dialog box, a status bar, a navigation bar, a Widget, etc. visual interface elements.
In some embodiments,user interface 280 is an interface that may be used to receive control inputs (e.g., physical buttons on the body of the display device, or the like).
In some embodiments, a system of a display device may include a Kernel (Kernel), a command parser (shell), a file system, and an application program. The kernel, shell, and file system together make up the basic operating system structure that allows users to manage files, run programs, and use the system. After power-on, the kernel is started, kernel space is activated, hardware is abstracted, hardware parameters are initialized, and virtual memory, a scheduler, signals and interprocess communication (IPC) are operated and maintained. And after the kernel is started, loading the Shell and the user application program. The application program is compiled into machine code after being started, and a process is formed.
As shown in fig. 4, the system of the display device is divided into three layers, i.e., an application layer, a middleware layer and a hardware layer from top to bottom.
The Application layer mainly includes common applications on the television and an Application Framework (Application Framework), wherein the common applications are mainly applications developed based on the Browser, such as: HTML5 APPs; and Native APPs (Native APPs);
an Application Framework (Application Framework) is a complete program model, and has all basic functions required by standard Application software, such as: file access, data exchange, and interfaces to use these functions (toolbars, status lists, menus, dialog boxes).
Native APPs (Native APPs) may support online or offline, message push, or local resource access.
The middleware layer comprises various television protocols, multimedia protocols, system components and other middleware. The middleware can use basic service (function) provided by system software to connect each part of an application system or different applications on a network, and can achieve the purposes of resource sharing and function sharing.
The hardware layer mainly comprises an HAL interface, hardware and a driver, wherein the HAL interface is a unified interface for butting all the television chips, and specific logic is realized by each chip. The driving mainly comprises: audio drive, display driver, bluetooth drive, camera drive, WIFI drive, USB drive, HDMI drive, sensor drive (like fingerprint sensor, temperature sensor, pressure sensor etc.) and power drive etc..
Currently, the display device is integrated with an intelligent voice assistant, and a user can search videos through the intelligent voice assistant by using a remote controller.
However, after a user inputs a name of a video resource that the user wants to search, the conventional display device usually performs a complete machine search, that is, searches videos on multiple types of video applications installed on the display device at the same time, so that the purpose of searching videos on a specific video application cannot be achieved, and the video search experience of the user is poor.
In order to solve the above problems, the present application provides a video search system, such as a frame diagram of the video search system shown in fig. 5, which includes adisplay device 200 and aserver 400. The embodiment of the application relates to a scene of interaction between a display device and a server. The display equipment is provided with a plurality of video applications, and the server is used for identifying the sound data collected by the display equipment and providing video resources of the plurality of video applications.
The process of performing video search by using the video search system of the embodiment specifically includes:
a user inputs sound data to the display equipment, and a sound collector of the display equipment collects the sound data input by the user. The display device may send the transcoded sound data to the server. The sound data at least comprises a video resource name.
After receiving the sound data, the server identifies the data from the sound data, specifically at least identifies the video resource name.
In some embodiments, if the sound data further includes a video application name, and the display device has a video application corresponding to the video application name installed thereon. That is, the server recognizes not only the video resource name from the sound data but also the video application name from the sound data, while the display device is installed with the video application. The display device calls a search interface of the video application to search the video resource corresponding to the video resource name in the server. And after the search is successful, the server feeds the video resource back to the display equipment.
In some embodiments, if the sound data further includes a video application name, and the video application corresponding to the video application name is not installed on the display device. The display device cannot call the search interface of the video application, and likewise cannot search the video resource corresponding to the video resource name in the server, and cannot feed back the video resource to the display device.
Illustratively, when a user inputs sound data "search for video X in video application a", the display device transmits the sound data to the server. The server identifies from the sound data a video asset name video X and a video application name application a.
If the video application A is installed on the display equipment, the display equipment calls a search interface of the video application A to search the video X in the server. And after the video resource of the video X is searched, feeding back the video resource of the video X to the display equipment. Therefore, the purpose of searching the appointed video resources in the appointed video application through the voice assistant is achieved, and the video searching experience of the user is improved.
If the video application a is not installed on the display device, the display device may not call the search interface of the video application a, and thus the video X cannot be searched in the server, and the video resource of the video X cannot be fed back to the display device.
In some embodiments, if the sound data does not include a video application name but only includes a video resource name, and a video application is currently running in the background of the display device, a search interface of the currently running video application is called, and a server is searched for a video resource corresponding to the video application name. And after the video resource corresponding to the video resource name is searched, feeding back the video resource to the display equipment.
In some embodiments, if the sound data does not include a video application name, but only includes a video resource name, and there is no currently running video application in the background of the display device, a whole machine search function is invoked, and the whole machine searches for the video resource.
Illustratively, when a user inputs sound data "search video X", the display apparatus transmits the sound data to the server. The server can only identify the video asset name, video X, from the sound data.
And if the video application B runs in the background of the current display equipment, the display equipment calls a search interface of the video application B to search the video resources of the video X in the server. And after the video resource of the video X is searched, feeding back the video resource to the display equipment.
If no video application runs in the background of the current display equipment and a search interface of a single video application cannot be called, a whole machine search function (searching in all the video applications installed on the display equipment) is called, and the video resource of the video X is searched by the whole machine. And after the video resource of the video X is searched, feeding back the video resource to the display equipment. In this scenario, it is possible to search for video assets for video X in multiple video applications. On the display device, the video resources of the video X searched in different video applications can be displayed in a sequence according to the preference degree of the user to each video application.
In some embodiments, theserver 400 includes aspeech recognition sub-service 400A, aninstruction generation sub-server 400B, and avideo search sub-server 400C. The voice recognition sub-server can be a server of an intelligent voice partner and is used for analyzing voice and semantics and recognizing related instructions. The instruction generation sub-server and the video search sub-server may be local servers for generating related search instructions according to the parsed semantics. The video search sub-server is used for receiving the search request of the display equipment end and feeding back related resources.
The process of performing video search by using the video search system of the embodiment specifically includes:
a user inputs sound data to the display equipment, and a sound collector of the display equipment collects the sound data input by the user. The display device may send the transcoded sound data to the speech recognition sub-server. The sound data at least comprises a video resource name.
After receiving the sound data, the voice recognition sub-server analyzes the sound and the semantics of the sound data, recognizes relevant instruction parameters, and specifically at least recognizes the video resource name.
In some embodiments, if the sound data further includes a video application name, and the display device installs a video application corresponding to the video application name. And after the voice recognition sub-server analyzes the voice data, the video application name is also recognized. And then the voice recognition sub-server sends the recognized video application name and video resource name as well as other relevant parameters (such as executed operation parameters, equipment parameters, language parameters and the like) for forming the instruction to the instruction generation sub-server.
The instruction generation sub-server generates a resource search instruction according to the video application name, the video resource name and other related parameters forming the instruction. Here, the voice recognition sub-server may recognize the relevant command directly from the voice data and transmit the recognized relevant command to the command generation sub-server. And the instruction generation sub-server converts the identified related instruction into a resource searching instruction which can be identified by the display equipment. The specific process of analyzing the voice data and generating the resource search command according to the analyzed data is not limited in the present application.
And the instruction generation sub-server feeds back the generated resource searching instruction to the display equipment. And after receiving the resource searching instruction, the display equipment generates a video searching request according to the resource searching instruction. The display device sends the video search request to the video search sub-server, namely, a search interface of the video application corresponding to the video application name is called, and the video search sub-server searches for the video resource in the video application.
After searching the video resource of the video resource name, the video searching sub-server feeds the video resource back to the display device.
Illustratively, as shown in FIG. 6, when a user inputs sound data "search for video X in video application A", the sound data is displayed on a display device. And the display equipment sends the sound data to the voice recognition sub-server. After receiving the sound data, the voice recognition sub-server recognizes the name of the video resource, namely video X, the name of the video application, namely application A, and other relevant parameters (the operation to be performed is search) from the sound data.
And the voice identification sub-server sends the identified video resource name video X, the video application name application A and other related parameters to the instruction generation sub-server. The instruction generation sub-server generates a resource search instruction according to the identified video resource name video X, the video application name application A and other related parameters: video X is searched in video application a. And the instruction generation sub-server feeds the generated resource searching instruction back to the display equipment.
After receiving the resource search instruction, the display device jumps from the user interface shown in fig. 6 to the user interface shown in fig. 7, where the user interface shown in fig. 7 is the user interface of the video application a.
The specific implementation process is as follows: and the display equipment generates a video search request according to the resource search instruction and sends the video search request to a video search sub-server so as to search for the video X in the video application A in the application A. That is, the search interface of the application a is called on the display device, and the application a is searched for the video X in the video application a.
After the video resource of the video X is searched, the video searching sub-server feeds back the video resource of the video X to the display equipment. As shown in the user interface of fig. 7, the video resources of the searched video X (which may show the video X and other videos related to the video X) are shown to the user.
In some embodiments, if the sound data further includes a video application name, and the display device does not install the video application corresponding to the video application name. And after the voice recognition sub-server analyzes the voice data, the video application name is also recognized. And then the voice recognition sub-server sends the recognized video application name and video resource name as well as other relevant parameters (such as executed operation parameters, equipment parameters, language parameters and the like) for forming the instruction to the instruction generation sub-server.
The instruction generation sub-server generates a resource search instruction according to the video application name, the video resource name and other related parameters forming the instruction. Here, the voice recognition sub-server may recognize the relevant command directly from the voice data and transmit the recognized relevant command to the command generation sub-server. And the instruction generation sub-server converts the identified related instruction into a resource searching instruction which can be identified by the display equipment.
And the instruction generation sub-server feeds back the generated resource searching instruction to the display equipment. At this time, the display device does not install the video application corresponding to the video application name, and therefore the search interface of the video application cannot be called. The corresponding video search request cannot be sent to the video search sub-server, and the video resource cannot be acquired from the video search sub-server.
Illustratively, when the user inputs sound data "search for video X in the video application a", the voice recognition sub-server recognizes relevant parameters of the generation instruction from the sound data, and transmits the relevant parameters to the instruction generation sub-server.
The instruction generation sub-server generates a resource search instruction according to the relevant parameters of the generation instruction: video X is searched in video application a. At this time, the application a is not installed on the display device, and therefore, a search interface of the application a cannot be called, and therefore, a video search request cannot be sent to the video search sub-server. The video search sub-server also cannot feed back the video resources of the search video X.
In some embodiments, if the sound data further includes a video application name, and the display device does not install a video application corresponding to the video application name, a corresponding video search request cannot be sent to the video search sub-server, and the video resource cannot be acquired from the video search sub-server. The controller generates prompt information and simultaneously displays the prompt information on the display. The prompt message may be: the application does not exist and please search among other applications.
In some embodiments, the speech recognition sub-server can only recognize the video asset name from the sound data if the sound data does not contain a video application name and the display device is currently running a video application. And the voice identification sub-server sends the identified video resource name and other relevant parameters generated by the instruction to the instruction generation sub-server. And the instruction generation sub-server generates a resource search instruction according to the video resource name and other related parameters.
And after receiving the resource searching instruction, the display equipment generates a video searching request according to the resource searching instruction, namely, calls a searching interface of the currently running video application. And sending the video search request to a video search sub-server. And searching the video resource corresponding to the video resource name in the currently running video application, and feeding back the video resource obtained by searching to the display equipment.
Illustratively, the user interface shown in fig. 8 is a home page interface of the video application B, including a navigation bar and a recommended video. In the user interface shown in fig. 8, when the user inputs the sound data "search video X", the sound data may be displayed in the user interface. The display device transmits the sound data to the voice recognition sub-server.
And the voice recognition sub-server recognizes relevant parameters for generating the instruction from the voice data and sends the relevant parameters to the instruction generation sub-server. The instruction generation sub-server generates a resource search instruction according to the relevant parameters of the generation instruction: searching for the video X, and sending the resource searching instruction to the display device.
After receiving the resource search instruction, the display device jumps from the user interface shown in fig. 8 to the user interface shown in fig. 9. The specific implementation process is as follows: and the display equipment calls a search interface of the video application B according to the resource search instruction to generate a video search request. Sending the video search request to a video search sub-server, searching the video resource of the video X in the video application B, and feeding back the video resource of the video X obtained by searching to the display equipment. Likewise, the searched video resources related to the video X are presented on the display device.
In some embodiments, the speech recognition sub-server can only recognize the video asset name from the sound data if the video application name is not included in the sound data and the video application is not currently running on the display device. The instruction generating sub-server generates a resource search instruction based on data identified from the sound data.
The display equipment does not have the video application running in the background at present, and meanwhile, the sound data does not contain the video application name, so that the display equipment calls the search interfaces of all the installed video applications to generate a video search request and sends the video search request to the video search sub-server. Therefore, complete machine search is realized, and finally all video resources corresponding to the video resource names obtained through search are fed back to the display equipment.
Illustratively, when a user inputs sound data "search video X", the voice recognition sub-server recognizes relevant parameters of a generation instruction from the sound data, and transmits the relevant parameters to the instruction generation sub-server. The instruction generation sub-server generates a resource search instruction according to the relevant parameters of the generation instruction: video X is searched.
At this time, the specified video application is not included in the search instruction, and the display device does not currently have a video application that is running in the background. And calling the search interfaces of all the installed video applications to generate a video search request, and sending the video search request to the video search sub-server. Therefore, the complete machine searches the video resource of the video X. And finally, feeding back the video resource corresponding to the video X to the display equipment.
In some embodiments, the instruction generation sub-server includes an ApplicationName parameter in the resource search instruction generated from the data identified from the voice data. If the speech recognition sub-server recognizes the name of an application from the voice data, the name of the application is assigned to the ApplicationName parameter. When the display device receives a resource searching instruction, if the value of the ApplicationName field is not null and the value of the ApplicationName field is consistent with the name of a certain application installed on the display device, the application is opened. And simultaneously, sending the name of the video resource to be searched to a search interface of the application, so as to search the video resource corresponding to the name of the video resource in the application.
When the display device receives a resource search instruction, if the value of the ApplicationName field is not null, and there is no application which is consistent with the value of the ApplicationName field in the applications installed on the display device, the application cannot be opened, and the video resource search in the application cannot be realized.
When the display device receives a resource searching instruction, if the value of the ApplicationName field is null, that is, the name of an application is not identified from the sound data, and a certain application runs in the background of the current display device, the currently running video application is opened. And meanwhile, the name of the video resource to be searched is sent to a search interface of the current running application, so that the video resource corresponding to the name of the video resource is searched in the current running application.
After the display device receives the resource searching instruction, if the value of the ApplicationName field is null and no running application exists in the background of the current display device, all video applications installed on the display device are opened. And meanwhile, the name of the video resource to be searched is sent to the search interfaces of all video applications, so that the complete machine search is realized.
An embodiment of the present application provides a video search method, such as a signaling diagram of the video search method shown in fig. 10, where the method includes the following steps:
the method comprises the steps that firstly, sound data are collected by display equipment, and the sound data are voice instructions input by a user through a user input interface. The sound data at least contains a video asset name. And the display equipment transmits the sound data to a server.
And step two, after the server receives the sound data, if the sound data also contains the video application name, and the video application corresponding to the video application name is installed on the display equipment. Searching the video application corresponding to the video application name for the video resource corresponding to the video resource name.
And step three, the server feeds back the video resource corresponding to the video resource name to the display equipment.
In some embodiments, if the video application name is also included in the sound data, and the video application corresponding to the video application name is not installed on the display device. No video asset is fed back to the display device.
In some embodiments, if the video application name is not contained in the sound data, and there is a video application currently running on the display device. Searching for a video resource corresponding to the video resource name in the currently running video application, and feeding back the video resource to the display device.
In some embodiments, if the video application name is not contained in the sound data, and the video application is not currently running on the display device. The whole machine searches for the video resource corresponding to the video resource name and feeds the video resource back to the display device.
Based on the foregoing method embodiment, an embodiment of the present application provides another video search method, such as a signaling diagram of the video search method shown in fig. 11, where the method includes the following steps:
the method comprises the steps that firstly, sound data are collected by display equipment, and the sound data are voice instructions input by a user through a user input interface. The sound data at least contains a video asset name. The display device transmits the sound data to the voice recognition sub-server.
And step two, the voice recognition sub-server recognizes each parameter of the generation instruction from the sound data, wherein the parameter at least comprises a video resource name, and sends each parameter of the generation instruction to the instruction generation sub-server.
And step three, the instruction generation sub-server generates a resource search instruction according to each parameter of the generated instruction and feeds the resource search instruction back to the display equipment.
And step four, the display equipment receives the resource searching instruction, the sound data at the moment also comprises a video application name, if the display equipment is provided with the video application corresponding to the video application name, a video searching request is generated according to the resource searching instruction (a searching interface of the video application corresponding to the video application name is called), and the video searching request is sent to the video searching sub-server.
And step five, after receiving the video search request, the video search sub-server searches the video resources corresponding to the video resource names in the video applications corresponding to the video application names, and feeds the video resources corresponding to the video resource names back to the display equipment.
In some embodiments, if the sound data further includes a video application name, and a video application corresponding to the video application name is not installed in the display device, a search interface of the video application cannot be called, and a video search request cannot be generated according to the resource search instruction.
In some embodiments, if the video application name is not contained in the sound data and the video application is currently running on the display device, the search interface of the currently running video application is called and a video search request is sent to the video search sub-server. And after receiving the video search request, the video search sub-server searches video resources corresponding to the video resource name in the currently running video application and feeds the video resources back to the display equipment.
In some embodiments, if the sound data does not contain a video application name and the display device does not currently have a running video application, the whole machine searches for a video resource corresponding to the video resource name. Specifically, all the search interfaces of the video applications installed in the display device are called, and a video search request is sent to the video search sub-server. And the video searching sub-server searches video resources corresponding to the video resource names in all the video applications and feeds back all the searched video resources corresponding to the video resource names to the display equipment.
The same or similar contents in the embodiments of the present application may be referred to each other, and the related embodiments are not described in detail.
Finally, it should be noted that: the above embodiments are only used for illustrating the technical solutions of the present application, and not for limiting the same; although the present application has been described in detail with reference to the foregoing embodiments, it should be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; and the modifications or the substitutions do not make the essence of the corresponding technical solutions depart from the scope of the technical solutions of the embodiments of the present application.
The foregoing description, for purposes of explanation, has been presented in conjunction with specific embodiments. However, the illustrative discussions above are not intended to be exhaustive or to limit the embodiments to the precise forms disclosed above. Many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to best explain the principles and the practical application, to thereby enable others skilled in the art to best utilize the embodiments and various embodiments with various modifications as are suited to the particular use contemplated.