CROSS-REFERENCE TO RELATED APPLICATIONThis application claims priority under 35 U.S.C. §119(a) from Korean Patent Application No. 10-2013-0165062, filed on Dec. 27, 2013 in the Korean Intellectual Property Office, the entire disclosure of which is hereby incorporated by reference.
TECHNICAL FIELDApparatuses and methods consistent with exemplary embodiments relate to a display apparatus, a server apparatus, and a display system including them, which provide a variety of information in response to a user's voice, and a method for providing contents thereof.
BACKGROUNDWith the development of electronic technologies, various kinds of electronic apparatuses have been developed and distributed. In particular, various types of electronic apparatuses including televisions (TVs) are used in general households. Such electronic apparatuses are equipped with various functions to meet the demands of users.
For example, the TV can access the Internet and support Internet-based services. In addition, users can view many digital broadcasting channels through the TV.
Accordingly, control methods are implemented to more easily use various functions of the display apparatus. For example, display apparatuses that can recognize a user's voice and can be controlled by the user's voice are being developed.
However, the related-art display apparatus only provides a service corresponding to a keyword included in the voice signal uttered by the user. That is, the related-art display apparatus provides limited services.
Therefore, there is a demand for a method for providing various resulting screens according to a user's utterance intention.
SUMMARYOne or more exemplary embodiments may overcome the above disadvantages and other disadvantages not described above. However, it is understood that one or more exemplary embodiment are not required to overcome the disadvantages described above, and may not overcome any of the problems described above.
One or more exemplary embodiments provide a display apparatus, a server apparatus, and a display system including them, which can provide various resulting screens according to a status of a content displayed on the display apparatus, and a method for providing contents thereof.
According to an aspect of an exemplary embodiment, there is provided a display apparatus including a microphone configured to receive a voice signal and a controller configured to generate a resulting screen corresponding to the voice signal different from a first displayed screen and according to a status of the content of the first displayed screen.
The display apparatus may further include a communicator configured to communicate with a server apparatus, the server apparatus being configured to perform a search corresponding to the status information and the text. The controller may transmit status information on the at least one content and a text corresponding to the voice signal to the server apparatus through the communicator, and, when a result of searching corresponding to the status information and the text is received from the server apparatus, the controller may generate the resulting screen according to the result of the searching.
When the voice signal is input, the controller may transmit the voice signal to a text converter through the communicator, and, when the voice signal is converted into the text by the text converter, the controller may receive the text from the text converter through the communicator.
The display apparatus may include a processor to replay a content, where the controller may receive address information for acquiring the designated content from the server apparatus, may acquire the designated content according to the address information, and may control the content processor to replay the acquired content.
When the result of the searching based on the voice signal is received from the server apparatus, the controller may control the graphic processor to generate the resulting screen based on the result of the searching.
The display apparatus may include a processor to replay a content, according to the status of the content on the first displayed screen. When a content designated by the voice signal exists in the first displayed screen, the controller may receive address information for acquiring the designated content from the server apparatus, acquire the designated content according to the address information, replay the acquired content, and display the content on the resulting screen. When the content designated by the voice signal does not exist in the first displayed screen, the controller may receive the result of the searching based on the voice signal from the server apparatus and may generate the resulting screen based on the result of the searching.
The status information may include at least one from among title information, alias title information, and identification information of the content included in the first displayed screen.
The status information may include at least one from among identification information of the at least one content, information on a type of content, content setting information, and information on the number of contents included in the first displayed screen.
According to an aspect of another exemplary embodiment, there is provided a display apparatus configured to be controlled by an input of a voice signal, a text converter configured to convert the voice signal input to the display apparatus into a text, and a server apparatus configured to perform searching according to status information of a screen display on the display apparatus and the text received from the display apparatus, and provide a result of the searching to the display apparatus, such that the display apparatus displays a resulting screen according to the result of the searching.
The text converter may be included in the server apparatus.
The server apparatus receives the status information and the text from the display apparatus simultaneously.
The server apparatus may receive the status information from the display apparatus first and may subsequently receive the text.
The server apparatus may identify contents included in a current screen of the display apparatus based on the status information.
When a content designated by the text exists in the screen of the display apparatus, the server apparatus may search address information for acquiring the designated content, and may transmit the address information to the display apparatus.
When the content designated by the text does not exist in the screen of the display apparatus, the server apparatus may perform searching by using the text and may transmit a result of the searching to the display apparatus.
The input may be received by the display apparatus.
According to an aspect another exemplary embodiment, there is provided a method for providing content information of a display apparatus, the method including displaying a screen including at least one content; receiving a voice signal; and generating a resulting screen corresponding to the voice signal differently according to a status of the content in the screen and displaying the resulting screen.
The method may further include transmitting status information on the at least one content and a text corresponding to the voice signal to a server apparatus, searching the server apparatus corresponding to the transmitted status information and the text, and receiving a result of the searching and generating the resulting screen based on the result of the searching.
The method may further include, when the voice signal is input, transmitting the voice signal to a text converter; and, when the voice signal is converted into the text by the text converter, receiving the text.
The displaying may include, when a content designated by the voice signal exists in the screen, receiving address information for acquiring the designated content from the server apparatus, acquiring the designated content according to the address information, replaying the acquired content, and displaying the content on the resulting screen. The displaying may include, when the content designated by the voice signal does not exist in the screen, receiving a result of searching the content designated by the voice signal from the server apparatus, and generating the resulting screen based on the result of the searching and displaying the resulting screen.
According to another exemplary embodiment, there is provided a display apparatus, including a display configured to display at least one content, a graphic processor configured to generate at least one screen having a content different from the displayed at least one content according to a user input, and a controller configured to control the display to display the generated at least one screen.
The display apparatus may also include a microphone configured to receive the user input, and the controller may control the microphone to send the user input to the graphic processor.
The user input may be a voice signal.
The microphone may include a microphone.
According to another exemplary embodiment, there is provided a method for displaying content on a display, the method including displaying at least one content on a display, receiving an input, generating at least one screen having a content different from the displayed at least one content, and displaying the generated at least one screen on the display.
The input may be a voice signal.
The method may include, when the user input is received, searching a server apparatus according to the user input, wherein the generating is performed according to a result of the searching.
The status information may include at least one of information on a kind of a screen displayed on the display apparatus, screen depth information, title information of the at least one content, alias title information of the at least one content, identification information of the at least one content, and information on the number of contents included in the screen.
According to an aspect of still another exemplary embodiment, there is provided a method for providing content information of a server apparatus, the method including: receiving status information on at least one content in a screen which is being displayed by a display apparatus and a text corresponding to a voice signal input to the display apparatus; and performing searching based on the status information and the text and transmitting a result of the searching to the display apparatus.
The transmitting the result of the searching to the display apparatus includes, when a content designated by the text exists in the screen of the display apparatus, transmitting address information for acquiring the designated content to the display apparatus, and, when the content designated by the text does not exist in the screen of the display apparatus, performing searching on the content designated by the text and transmitting the result of the searching to the display apparatus.
The status information may include at least one of information on a kind of a screen displayed on the display apparatus, screen depth information, title information of the at least one content, alias title information of the at least one content, identification information of the at least one content, and information on the number of contents included in the screen.
According to various exemplary embodiments described above, when a user's voice is input, the display apparatus may configure a resulting screen corresponding to the voice variously according to a status of the display apparatus and may display the resulting screen. Accordingly, a voice control can be performed according to a user's intention.
BRIEF DESCRIPTION OF THE DRAWINGSThe above and/or other aspects will be more apparent by describing in detail exemplary embodiments, with reference to the accompanying drawings, in which:
FIG. 1 is a block diagram illustrating a configuration of a display apparatus according to an exemplary embodiment;
FIGS. 2 to 11 are views to illustrate an operation of a display apparatus according to various exemplary embodiments;
FIG. 12 is a flowchart to illustrate a method for providing content information of a display apparatus according to an exemplary embodiment;
FIG. 13 is a block diagram illustrating a configuration of a display system according to an exemplary embodiment;
FIG. 14 is a block diagram illustrating an overall configuration of a display apparatus according to various exemplary embodiments;
FIG. 15 is a view to illustrate a software structure of a display apparatus according to an exemplary embodiment;
FIG. 16 is a view illustrating a configuration of a server apparatus according to an exemplary embodiment; and
FIG. 17 is a timing chart to illustrate an operation of a display system according to an exemplary embodiment.
DETAILED DESCRIPTIONHereinafter, exemplary embodiments will be described in greater detail with reference to the accompanying drawings.
In the following description, same reference numerals are used for the same elements when they are depicted in different drawings. The matters defined in the description, such as detailed construction and elements, are provided to assist in a comprehensive understanding of exemplary embodiments. Thus, it is apparent that exemplary embodiments can be carried out without those specifically defined matters. Also, functions or elements known in the related art are not described in detail since they would obscure the exemplary embodiments with unnecessary detail.
FIG. 1 is a block diagram illustrating a configuration of a display apparatus according to an exemplary embodiment. Thedisplay apparatus100 ofFIG. 1 may be implemented by using various types of apparatuses such as a TV, a Personal Computer (PC), a laptop PC, a mobile phone, a tablet PC, a Personal Digital Assistant (PDA), an MP3 player, an electronic album, a kiosk, etc. Referring toFIG. 1, thedisplay apparatus100 includes amicrophone110, acontroller120, adisplay130, and agraphic processor140.
Themicrophone110 is an element for receiving a voice signal of a user. The configuration of themicrophone110 may vary according to various exemplary embodiments. For example, themicrophone110 may include a microphone to directly receive the voice or may receive the user voice through a microphone provided on a remote controller. For a voice signal to be inputted, a microphone itself may be used, and a microphone which is included in an audio-visual (AV) apparatus, a computer, a portable computer, a mobile phone, a smart phone and a remote controller may be used.
Thecontroller120 is an element for controlling an overall operation of thedisplay apparatus100.
Thegraphic processor140 is an element for configuring various screens. Specifically, thegraphic processor140 may generate a screen including various objects such as an icon, an image, and a text using an operator (not shown) and a renderer (not shown). The operator calculates attribute values of each object to be displayed, such as coordinates values, a shape, a size, and a color, according to a layout of the screen. The renderer generates screens of various layouts including objects based on the attribute values calculated by the operator. The screen generated by the renderer is provided to thedisplay130.
Thedisplay130 displays various screens generated by thegraphic processor140. When thedisplay apparatus100 is a broadcast receiving apparatus for receiving and processing broadcast signals, thedisplay130 displays a broadcast program screen. When thedisplay apparatus100 downloads contents from various content providing servers such as an Internet Protocol (IP) server, thedisplay130 may display a content playback screen.
In addition, thecontroller120 may control thegraphic processor140 to configure various screens such as a content selection screen through which various contents are selected, a main screen through which various menus are selected, a preference function screen which displays functions frequently used by the user, and a viewing history screen for guiding information on contents that the user has viewed. When a user's voice signal is input while a screen including at least one content is being displayed on thedisplay130, thecontroller120 may control thegraphic processor140 to configure a resulting screen corresponding to the voice signal. The screen generated by thegraphic processor140 may be provided to thedisplay130 and displayed. In this case, thecontroller120 may control thegraphic processor140 to configure a different resulting screen according to information of the content included in the screen.
Although thegraphic processor140 is an element separate from thecontroller120 inFIG. 1, thegraphic processor140 may be implemented as a Graphic Processing Unit (GPU) provided in thecontroller120. However, an extra illustration thereof is omitted.
FIGS. 2 to 11 are views illustrating an operation of a display apparatus according to various exemplary embodiments.
FIG. 2 is a view illustrating the performance of a different operation according to whether a content designated by a user's voice signal is included in a screen or not according to an exemplary embodiment.
As shown inFIG. 2, thedisplay100 may display ascreen210 including information of contents which are currently broadcasted. Such ascreen210 may be called various names such as a content selection screen or an ONTV panel. Referring toFIG. 2, thumbnail images211-1 to211-12 regarding viewable contents provided through a TV broadcast network, an IP network, and other various routes may be displayed on thescreen210 according to a predetermined layout.
When auser10 utters a voice signal including a title of a content in the screen while thescreen210 is being displayed, thecontroller120 displays a resultingscreen220 corresponding to the voice signal based on the voice signal and a status of thedisplay apparatus100. For example, when one of the plurality of thumbnail images211-1 to211-12 (for example,211-9) is a thumbnail image regarding a content titled “AAA”, and theuser10 says “AAA, please.”, thecontroller120 performs a control operation to directly select and replay the AAA content. Specifically, thedisplay apparatus100 may further include a processor (not shown) to replay the contents. Thecontroller120 may control the processor to replay the AAA content.
Accordingly, thedisplay130 displays theplayback screen220 of the AAA content. The AAA content may be provided from various sources. For example, when thedisplay apparatus100 has already stored the AAA content, thedisplay apparatus100 may directly replay the stored AAA content. When the AAA content is a broadcast program provided from a specific broadcasting station through a broadcast channel, thecontroller120 controls a tuner (not shown) to tune to the corresponding broadcast channel and receives the broadcast program through the broadcast channel. When the AAA content is a content provided from a content providing server (not shown), thecontroller120 controls a communicator (not shown) to access the content providing server using address information for acquiring the AAA content, and downloads the AAA content. The address information may be information that is pre-stored by thedisplay apparatus100. However, according to another exemplary embodiment, a server apparatus (not shown) interworking with thedisplay apparatus100 may provide the address information. The processor and the communicator mentioned in the above-described exemplary embodiment will be explained below in detail with reference to the drawings.
Theuser10 may say a title of a content that is not included in the screen.FIG. 3 illustrates a case in which the user says “BBB, please” while thesame screen210 as inFIG. 2 is being displayed. When the BBB content designated by the user's voice signal does not exist in thecurrent screen210, thecontroller120 displays asearch result screen310 which is a result of searching the BBB content. According to an exemplary embodiment, the search may be performed in various apparatuses. For example, thecontroller120 of thedisplay apparatus100 may perform the search or an external server apparatus performs the search and provides a corresponding result to thecontroller120. Thesearch result screen310 ofFIG. 3 includes various items311-1 to311-12 including the keyword BBB.
When one of the items311-1 to311-12 is selected by the user, thecontroller120 may replay a content corresponding to the selected item and output the content through thedisplay130.
According to the exemplary embodiments described above with reference toFIGS. 2 and 3, thedisplay apparatus100 may perform a different control operation according to whether information on the content designated by the user's voice signal is included in the current screen or not. Accordingly, the user may replay a specific content directly or may search for the content.
According to another exemplary embodiment, thedisplay apparatus100 may provide various resulting screens according to a kind of a screen currently displayed.
For example, when the user inputs the voice signal while thescreen210 including information on currently viewable contents is being displayed as shown inFIG. 2, thecontroller120 may perform a different operation according to whether the content is included in thescreen210 or not.
On the other hand, when ascreen410 including information on contents preferred by the user is displayed as shown inFIG. 4, and the user inputs a voice signal designating a specific content included in thescreen410, thecontroller120 may not directly replay the content and may perform a different control operation.
Specifically, as shown inFIG. 4, thedisplay apparatus100 may display thescreen410 including the contents that the user viewed before, relevant contents, similar contents, etc. Such ascreen410 may be called various names, such as a preference content screen, a MYCONTENTS screen, a viewing history content, etc. Referring toFIG. 4, thumbnail images411-1 to411-8 associated with the contents may be displayed on thescreen410 according to a predetermined layout. The layout of thescreen410 ofFIG. 4 may be different from the layout of thescreen210 ofFIGS. 2 and 3.
In this state, when the voice signal uttered by the user is input through themicrophone110, thecontroller120 determines whether information on a content corresponding to the voice signal, namely a thumbnail image is included in thecurrent screen410 or not. When the thumbnail image (for example,411-6) of the content corresponding to the voice signal exists, thecontroller120 may display asearch result screen420 having various contents related to the corresponding content on thedisplay130. Thesearch result screen420 may include information on contents421-1 and421-2 that have been already viewed in relation to the corresponding content, and information on contents421-3 to421-5 that are newly viewable. When the corresponding content is a moving image consisting of a plurality of series, thecontroller120 may distinguish between the series421-1 and421-2 that the user has already viewed and the series421-3,421-4, and421-5 that the user has not viewed. When the user inputs a voice signal (for example, “part3”) to designate a specific part of the content (for example,421-3) while thescreen420 is being displayed, thecontroller120 may directly replay the corresponding content and output the content.
FIG. 5 illustrates a case in which a content designated by a user's voice signal does not exist on thescreen410 which includes the information on the contents preferred by the user. For example, when the user inputs a voice signal “DDD, please” to designate a content (for example, DDD) that does not exist in thescreen410 as shown inFIG. 5, thecontroller120 displays asearch result screen510 regarding the identification word DDD from among the keywords included in the corresponding voice signal. Thesearch result screen510 may include information on contents511-1 to511-nthat have various titles including the text DDD.
FIG. 6 is a view to illustrate a case in which the user inputs a voice signal while aplayback screen610 of a specific content is being displayed. In this case, thecontroller120 determines that a content (for example, DDD) included in the voice signal is not included in the current screen, and displays a search result screen regarding an identification word included in the voice signal. When the voice signal includes the same identification word DDD as inFIG. 5, the samesearch result screen510 as inFIG. 5 may be displayed.
In the above-described exemplary embodiments, when an abbreviation for a title or another name is uttered in addition to the full name of the title, thecontroller120 may display information on the contents corresponding to the abbreviation or another name.
As described above with reference toFIGS. 4 to 6, even when the user inputs the same voice signal, thedisplay apparatus100 may provide various resulting screens according to various conditions such as a type of a screen currently displayed, a content display status of the screen, etc.
In the above-described exemplary embodiments, the user inputs the voice signal to designate a specific content. However, thedisplay apparatus100 may receive a voice signal arbitrarily uttered by the user and may provide various resulting screens according to a meaning of the voice signal.
FIG. 7 is a view to illustrate an operation of the display apparatus according to such an exemplary embodiment.FIG. 7 illustrates a case in which theuser10 inputs a voice signal “What's that?” when acertain content screen710 is being displayed. Thecontroller120 analyzes the user's voice signal and determines its meaning. Accordingly, thecontroller120 determines the object indicated by the expression “that”, and displays aresponse message720 corresponding to a result of the determining.FIG. 7 illustrates a case in which the user inputs the voice signal when an animation content having a penguin character named EEE is displayed. Thecontroller120 may configure theresponse message720 “EEE is a penguin” in response to the user's question, and may display theresponse message720.
The response message may be configured in various ways.
For example, thecontroller120 may search a variety of information such as a title, character names, an additional explanation, etc., of the corresponding content from metadata of the currently displayed content. Thecontroller120 may control thegraphic processor140 to configure theresponse message720 according to the result of the search.
According to another exemplary embodiment, thecontroller120 may detect an image frame that is displayed at the time the user's voice signal is input. Thecontroller120 may detect an edge included in the detected image frame and may search related images corresponding to a shape of an object specified by the edge. When the related image is searched, thecontroller120 may control thegraphic processor140 to configure theresponse message720 to include the related image or to include a text with the related image.
In addition, thecontroller120 may access a specific server apparatus based on address information linked with the corresponding content, and may control thegraphic processor140 to configure theresponse message720 based on a text provided by the server apparatus.
FIG. 8 illustrates a state in which aweb screen810 is displayed. In this state, when the user inputs the voice signal “What's that?”, thecontroller120 may display aresponse message820 based on information from a web server (not shown) which provides the currently displayedweb screen810. InFIG. 8, theresponse message820, displaying information indicating that the currently displayed site is the FFF portal site is illustrated.
For another example, thecontroller120 may control thegraphic processor140 to configure theresponse message820 to include a Uniform Resource Locator (URL) of theweb screen810 currently displayed.FIG. 9 is a view to illustrate an exemplary embodiment in which the user inputs a certain voice signal while a content selection screen, which includes information on at least one content, is being displayed. Referring toFIG. 9, acontent selection screen910 may include thumbnail images911-1 to911-12 relating to viewable contents.
Thecontroller120 may highlight one of the thumbnail images911-1 to911-12. InFIG. 9, the second thumbnail image911-2 is highlighted. When the user manipulates a direction key provided on a remote control apparatus or inputs a voice signal indicating a direction, thecontroller120 may move the highlighted area of the display in a direction corresponding to the direction key or voice signal. In this state, when the user inputs the voice signal “What's that?”, thecontroller120 may control thegraphic processor140 and thedisplay130 to display adetailed information screen920 linked with the currently highlighted thumbnail image911-2. For example, when the corresponding thumbnail image911-2 is a movie content, thedetailed information screen920 may include a variety of information such as a movie title, a language, a running time, a story, a main scene, etc.
In addition, when the user inputs a voice signal “Show me that!”, thecontroller120 may control the processor to directly replay the content corresponding to the currently highlighted thumbnail image911-2 and may display a replay screen thereof on thedisplay130.
According to the various exemplary embodiments described above, the display apparatus can provide various resulting screens according to an operation status of the display apparatus even when the user inputs the same voice signal.
In addition, even when the user inputs the same voice signal, the display apparatus may perform a different control operation according to an operation status of the display apparatus.
FIGS. 10 and 11 illustrate an operation of a display apparatus according to another exemplary embodiment. According to the present exemplary embodiment, the display apparatus may perform a different control operation according to an operation status of the display apparatus even when the user inputs the same voice signal.
InFIGS. 10 and 11, the user inputs a voice signal “Increase, please.” InFIG. 10, ascreen1010 having nothing to do with an audio volume is displayed. Specifically, thedisplay apparatus100 may display thescreen1010 including a plurality of objects1011-1 to1011-8. For example, the content selection screen including the thumbnail images regarding the viewable contents or the screen for the user preference contents may be displayed as shown inFIGS. 2 to 5.
In this state, when the voice signal “Increase, please” is input, thecontroller120 may determine that the intention of theuser10 is to magnify thescreen1010. Accordingly, thecontroller120 may control thegraphic processor140 to configure ascreen1020, to include the objects1011-1 to1011-8 ofscreen1010 in a magnified form. When all of the objects1011-1 to1011-8 cannot be included in the onescreen1020 as a result of the magnifying, the number of objects included in the onescreen1020 may be reduced. InFIG. 10, eight objects1011-1 to1011-8 are included in thepre-magnification screen1010, whereas only four objects1011-1,1011-2,1011-5, and1011-6 are included in thepost-magnification screen1020. When the voice signal “Increase, please” is repeatedly input several times, thecontroller120 may control thegraphic processor140 to continue magnifying the objects1011-1 to1011-8. In addition, when the voice signal is input more than a predetermined number of times, thecontroller120 may control thegraphic processor140 to configure one of the objects1011-1 to1011-8 (for example,1011-1) to be a full screen size, so that the one of the objects1011-1 to1011-8 is displayed on the entire display. Although each object1011-1 to1011-8 is displayed in the form of a thumbnail image inFIG. 10, the objects may be implemented in various forms such as a text, a photo, an image, etc.
FIG. 11 illustrates a state in which a moving image content is replayed and aplayback screen1110 thereof is displayed. In this state, when theuser10 inputs the voice signal “Increase, please”, thecontroller120 may determine that the user's intention is to turn up the audio volume. Accordingly, thecontroller120 may increase the audio volume up to a predetermined level. In addition, thecontroller120 may control thegraphic processor140 to configure a Graphic User Interface (GUI)1120 indicating an audio volume status.
InFIGS. 10 and 11, the voice signal “Increase, please” is input by way of an example. However, when a voice signal “Decrease, please” is input, thecontroller120 may perform a different operation such as reducing a size or turning down an audio volume according to a kind of a screen.
According to the various exemplary embodiments as described above, the display apparatus may perform various operations according to an operation status of the display apparatus, even when the user inputs the same voice signal.
In the above-described exemplary embodiment, thedisplay apparatus100 receives the voice signal and processes the voice signal. However, a broadcast receiving apparatus (for example, a set-top box) without a display may perform the above-described processing operation. When a user's voice signal is input while the display apparatus is displaying a screen including at least one content, the set-top box may configure a different resulting screen, corresponding to the voice signal and according to a content status, and may provide the resulting screen to the display apparatus. Consequently, the resulting screen corresponding to the voice signal is displayed on the display apparatus. The resulting screen configured by the broadcast receiving apparatus may be configured according to various exemplary embodiments as described above with reference toFIGS. 2 to 11. Illustration and explanation of the configuration and operation of the broadcast receiving apparatus will be omitted.
FIG. 12 is a flowchart illustrating a method for providing content information of a display apparatus according to an exemplary embodiment.
Referring toFIG. 12, when thedisplay apparatus100 displays a certain screen (S1210), thedisplay apparatus100 may receive a user's voice signal (S1220). The screen may be implemented in various ways as described above with reference toFIGS. 2 to 11. When the user's voice signal is input, thedisplay apparatus100 may configure a resulting screen corresponding to the voice signal to be displayed differently according to a content status in the screen, and may display the resulting screen (S1230). Such an operation may be independently performed by thedisplay apparatus100. However, this should not be considered as limiting. According to an exemplary embodiment, the above-described operation may be performed in a display system including a text converter, thedisplay apparatus100, and a server apparatus.
FIG. 13 is a block diagram illustrating a configuration of a display system according to an exemplary embodiment. Referring toFIG. 13, thedisplay system1300 includes atext converter1310, adisplay apparatus100, and aserver apparatus1400.
When a voice signal is input, thedisplay apparatus100 may transmit the voice signal to theexternal text converter1310. Thetext converter1310 is an apparatus that converts the voice signal transmitted through thedisplay apparatus100 into a text, and provides the text to thedisplay apparatus100. According to an exemplary embodiment, thetext converter1310 may be implemented by using a separate server apparatus or may be included in theserver apparatus1400.
Thetext converter1310 may recognize the voice by using at least one of various recognition algorithms, such as a dynamic time warping method, a hidden Markov model, a neural network, etc., and may convert the recognized voice into the text. For example, when the hidden Markov model is used, thetext converter1310 models a temporal change and a spectrum change of the user's voice, respectively, and detects a similar word from a pre-stored language database. Accordingly, thetext converter1310, may output the detected word as a text. Thetext converter1310 transmits the text to thedisplay apparatus100. When thedisplay apparatus100 is equipped with a text conversion function as described above, thedisplay apparatus100 does not transmit the voice signal to thetext converter1310, and may instead directly convert the voice signal into the text.
When the text is transmitted from thetext converter1310, thedisplay apparatus100 may perform a control operation corresponding to the text. For example, when the voice signal input by the user includes a pre-defined voice command, thedisplay apparatus100 may perform an operation corresponding to the voice command. For example, when thedisplay apparatus100 is a TV, voice commands may be designated for various operations such as turn on, turn off, volume up, volume down, channel up, channel down, etc., and may be stored. When such a voice command is input, thedisplay apparatus100 may perform an operation matching the voice command.
On the other hand, when the voice command is not consistent with the text, thedisplay apparatus100 provides the text to theserver apparatus1400. In addition, thedisplay apparatus100 may transmit status information regarding a content included in the screen which is displayed at the time the voice signal is input to theserver apparatus1400.
Theserver apparatus1400 may receive the status information and the text from thedisplay apparatus100. The status information and the text may be received simultaneously or may be received in sequence when necessary. For example, theserver apparatus1400 may receive the status information from thedisplay apparatus100 first and then may receive the text.
Theserver apparatus1400 performs searching based on the status information and the text, and transmits a result of the searching to thedisplay apparatus100. The result of the searching may include a variety of information. For example, the result of the searching may include thumbnail images including various contents, such as a keyword included in the user's voice signal, title information, and address information for acquiring content information.
When the result of the searching is received by thedisplay apparatus100, thedisplay apparatus100 may configure a resulting screen based on the result of the searching, and display the resulting screen. The status information may include at least one piece of information from among title information of contents included in the screen, alias title information, and identification information. In addition, the status information may include at least one piece of information from among information on the number of contents included in the screen, information on kinds of contents, and content setting information. The information on the number of contents is information indicating the number of thumbnail images displayed in the screen, and the information on the kinds of contents is information indicating the type of content displayed in the screen from among various contents such as a moving image, a photo, a web page, a TV, a Video On Demand (VOD). The content setting information refers to setting information that is set for each content by the user. For example, the content setting information may be depth information when the content is a 3D content.
For example, thedisplay apparatus100 may transmit the following status:
TV_status: recommend, smart-function, web-browser, etc.
screen_info: OnTV, Mycontent, Apps, Movies, TV, photo, web page, multimedia, account, recommendation bar, multiscreen, main screen, etc.
panel_depth_info: first, second, third, etc.
panel_text: title1, title2, command1, command2, etc.
contents_title: title-1, title-2, title-3 . . . , title-n
alias-title: a-title-1, a-title-2, a-title-3, . . . , title-n
content_num: n
content_id: ID-1, ID-2, . . . ID-n
For another example, when a specific content is replayed, thedisplay apparatus100 may transmit the following status information:
device_status_type: EPG
content_info: content-id-001
content_time: 20130102-10:00
title: EEE is My friend
user_id: KIM
Theserver apparatus1400 may determine a current status of thedisplay apparatus100 by using the status information of thedisplay apparatus100. When a command “AAA, please” is input, theserver apparatus1400 determines whether a content titled AAA exists in the current screen of thedisplay apparatus100 or not by using the title information, the alias title information, and the identification information of the status information. When the AAA content exists, theserver apparatus1400 may directly transmit the AAA content to thedisplay apparatus100. According to another exemplary embodiment, theserver apparatus1400 may transmit address information for acquiring the AAA content to thedisplay apparatus100. When the address information is received, thedisplay apparatus100 may access a content providing server (not shown) based on the address information and may acquire the AAA content.
Thedisplay apparatus100 included in thedisplay system1300 ofFIG. 13 may include the configuration described above with reference toFIG. 1. However, this should not be considered as limiting.
FIG. 14 is a view illustrating a detailed configuration of thedisplay apparatus100 according to the above-described exemplary embodiments, when thedisplay apparatus100 is implemented by using a television.
Referring toFIG. 14, thedisplay apparatus100 includes amicrophone110, acontroller120, adisplay130, agraphic processor140, astorage150, acommunicator165, abroadcast receiver160, aprocessor170, aspeaker180, and a remotecontrol signal receiver190.
Themicrophone110 is an element for receiving a voice signal as described above.
Thegraphic processor140 is an element for configuring various screens as described above.
Thecontroller120 may directly convert the voice signal input through themicrophone110 into a text or may transmit the voice signal to the text converter as shown inFIG. 13 to convert the voice signal into the text.
Thecommunicator165 is an element for communicating with various external apparatuses, including thetext converter1310 and theserver apparatus1400. Thecommunicator165 may perform communication in various kinds of communication methods. Specifically, thecommunicator165 may communicate in various wired and wireless communication methods such as Wi-Fi, Bluetooth, Zigbee, Local Area Network (LAN), 3rdGeneration (3G), 4G, Near Field Communication (NFC), etc. Thecontroller120 may communicate with thetext converter1310 and theserver apparatus1400 through thecommunicator165. For example, when a voice signal is input, thecontroller120 may transmit status information on the screen, which is displayed at the time when the voice signal is input, and a text corresponding to the voice signal, to theserver apparatus1400 through thecommunicator165.
The remotecontrol signal receiver190 is an element for receiving various remote control signals from a remote control apparatus.
Thebroadcast receiver160 is an element for tuning to various broadcasting channels and receiving a broadcast content through the tuned broadcasting channel.
Thecontroller120 controls thebroadcast receiver160 to tune to the broadcasting channel according to a remote control signal received from the remotecontrol signal receiver190.
Theprocessor170 is an element for processing contents. Theprocessor170 may process various contents such as a broadcast content received through thebroadcast receiver160, a content received through other networks such as the Internet, and contents stored in thestorage140, and may replay the contents.
Theprocessor170 may include a video processor and an audio processor. The video processor is an element for processing video data included in the content. The video processor may perform various image processing operations such as decoding, scaling, noise filtering, frame rate conversion, resolution conversion, etc. with respect to the video data. The audio processor is an element for processing audio data included in the content. The audio processor may perform various processing operations such as decoding, amplification, noise filtering, etc. with respect to the audio data.
Thedisplay130 may output the video data processed by theprocessor170.
Thespeaker180 may output the audio data processed by theprocessor170.
Thestorage150 is an element for storing various programs or data and contents necessary for the operation of thedisplay apparatus100.
Thecontroller120 controls the overall operation of thedisplay apparatus100 using programs stored in thestorage150. Thecontroller120 may be configured in various forms. For example, thecontroller120 may include a Random Access Memory (RAM)121, a Read Only Memory (ROM)122, aCPU123, and abus125 as shown inFIG. 14. TheRAM121, theROM122, and theCPU123 may be connected to one another through thebus125.
TheCPU123 accesses thestorage150 and performs booting by using an Operating System (OS) stored in thestorage150. In addition, theCPU123 performs various operations using various programs, contents, data, etc., stored in thestorage150.
TheROM122 stores a set of commands to boot the system. When a turn on command is input and power is supplied, theCPU123 copies the OS stored in thestorage150 into theRAM121 according to a command stored in theROM122, executes the O/S and boots the system. When the booting is completed, theCPU123 copies the various programs stored in thestorage150 into theRAM121, executes the programs copied into theRAM121, and performs various operations.
In addition, although not shown inFIG. 14, thedisplay apparatus100 may further include various elements such as a microphone, a camera, a USB port, an external input port, etc.
FIG. 15 illustrates a structure of software stored in thedisplay apparatus100 according to an exemplary embodiment. Referring toFIG. 15, thestorage150 may store anOS151, a voice signal processing module152-1, a status information processing module152-2, a command processing module152-3, a UI framework152-4, an APP manager152-5, a multimedia framework152-6, adatabase153, and various applications154-1 to154-n.
TheOS151 controls and manages an overall operation of hardware. That is, theOS151 is software that performs a basic function such as hardware management, memory, and security.
The voice signal processing module152-1 is software for converting a voice signal input through themicrophone110 into a text. The text conversion may be performed according to various algorithms as described above.
The status information processing module152-2 is an element for generating status information on various contents included in the screen displayed on thedisplay130. The status information may be configured in various forms as described above.
The command processing module152-3 is an element for parsing a command transmitted from theserver apparatus1400. Thecontroller120 performs an operation corresponding to the command processed by the command processing module152-3.
The UI framework152-4 is a module for providing various UIs to be displayed on thedisplay130. The UI framework152-4 may include an image compositor module to configure various kinds of objects to be included in the screen, a coordinate compositor to calculate coordinates of an object to be displayed, a rendering module to render the configured object on the calculated coordinates, and a 2D/3D UI toolkit to provide a tool for configuring a UI of a 2D or 3D format. Thecontroller120 may display various screens as described above with reference toFIGS. 2 to 11 by executing the UI framework152-3.
The APP manager152-5 is a module for managing execution statuses of various applications installed in thestorage150. When an application execution event is detected, the APP manager152-5 calls an application corresponding to the event and executes the application.
Thedatabase153 includes various voice commands or keywords. In addition, various applications155-1 to155-nmay be stored in thestorage150.
The software structure shown inFIG. 15 is merely an example and is not limited to this. Accordingly, some of the elements may be omitted or changed or a new element may be added when necessary. For example, when thedisplay apparatus100 is not equipped with a text conversion function, the voice signal processing module152-1 and thedatabase153 may be omitted.
As described above, thedisplay apparatus100 may be implemented by using various types of apparatuses in addition to the television. Accordingly, the configuration described inFIGS. 14 and 15 may be changed in various ways according to a type ofdisplay apparatus100.
FIG. 16 is a block diagram illustrating an example of a configuration of the server apparatus included in thedisplay system1300. Referring toFIG. 16, theserver apparatus1400 includes aserver communicator1410 and aserver controller1420.
Theserver communicator1410 is an element for communicating with thedisplay apparatus100. Theserver communicator1410 may receive status information on at least one content in the screen which is being displayed on thedisplay apparatus100, and a text.
Theserver controller1420 may perform a search according to an intended use of thedisplay apparatus100 based on the status information and the text. Specifically, theserver controller1420 may identify contents included in the screen of thedisplay apparatus100 based on the status information. When a content designated by the text exists in the identified contents, theserver controller1420 may search for address information for acquiring the designated content and may transmit the address information to the display apparatus. On the other hand, when the content designated by the text does not exist in the screen of thedisplay apparatus100, theserver controller1420 may perform a search using the text and transmit a result of the search to thedisplay apparatus100.
Theserver controller1420 includes aninput data processor1421, aconversation intention analyzer1422, acontent processor1423, astatus information analyzer1424, acommand generator1425, and anoutput data processor1426.
Theinput data processor1421 is an element for parsing data transmitted from thedisplay apparatus100, and providing the data to each of theelements1422 to1426.
Thestatus information analyzer1422 is an element for receiving status information processed by theinput data processor1421 and analyzing the status information. Thestatus information analyzer1422 may determine what content is included in the screen displayed on thedisplay apparatus100 and what kind of screen is displayed by analyzing the status information.
Theconversation intention analyzer1423 is an element for analyzing a user's intent based on the text received through theserver communicator1410 and a result of the analyzing by thestatus information analyzer1422.
Thecontent processor1424 is an element for generating a query including information on a content designated by the user's voice signal. Theserver communicator1410 may transmit the query to acontent providing server1600 or other databases, and may receive a response signal to the query. The response signal may include a result of a search.
Thecommand generator1425 generates a command corresponding to thedisplay apparatus100 based on the result of the analyzing by theconversation intention analyzer1423 and the response signal.
Theoutput data processor1426 is an element for generating output data including the generated command and the result of the searching. The output data may be generated in a format so that the output data can be processed in the display apparatus.
FIG. 17 is a timing chart to illustrate the operation of the display system according to an exemplary embodiment. Referring toFIG. 17, when a user's voice is input (S1710), thedisplay apparatus100 transmits the voice signal to thetext converter1310.
Thetext converter1310 converts the voice signal into a text (S1720), and transmits the text to the display apparatus100 (S1725). Thedisplay apparatus100 transmits the text and status information to the server apparatus1400 (S1730). The status information may be configured in various ways as described above.
Theserver apparatus1400 analyzes a conversation intention based on the status information (S1735). Theserver apparatus1400 performs searching corresponding to the analyzed conversation intention (S1740). The searching may be performed in various ways as described above with reference toFIGS. 2 to 11.
Theserver apparatus1400 transmits a result of the searching to the display apparatus100 (S1745). The result of the searching may include address information for acquiring various contents. Thedisplay apparatus100 accesses thecontent providing server1600 according to the address information included in the result of the searching. Thedisplay apparatus100 requests content information from thecontent providing server1600.
Thecontent providing server1600 may provide the content or thumbnail images according to the request (S1755).
Thedisplay apparatus100 may configure a resulting screen based on the transmitted data (S1760). The resulting screen may be configured in various ways according to the status of thedisplay apparatus100 and the user's voice signal as shown inFIGS. 2 to 11, and the resulting screen may be displayed.
As described above, thedisplay apparatus100 may display various resulting screens according to the voice signal input by the user and the screen display status of thedisplay apparatus100.
The display apparatus or the method for providing content information of the display system according to the above-described exemplary embodiments may be stored in a non-transitory readable medium. The non-transitory readable medium may be mounted and used in various devices.
The non-transitory readable medium refers to a medium that stores data semi-permanently rather than storing data for a very short time, such as a register, a cache, and a memory, and is readable by an apparatus. Specifically, the non-transitory readable medium may be a compact disc (CD), a digital versatile disk (DVD), a hard disk, a Blu-ray disk, a USB, a memory card, a ROM, etc.
For example, a program code for displaying a screen including at least one content, receiving a voice signal, and configuring a resulting screen corresponding to the voice signal differently according to a status of the content, and displaying the resulting screen, may be stored in a non-transitory readable medium and provided.
The foregoing exemplary embodiments and advantages are merely exemplary and are not to be construed as limiting the present inventive concept. The exemplary embodiments can be readily applied to other types of apparatuses. Also, the description of the exemplary embodiments is intended to be illustrative, and not to limit the scope of the claims, and many alternatives, modifications, and variations will be apparent to those skilled in the art.