Detailed Description
Reference will now be made in detail to embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The embodiments described in the following examples do not represent all embodiments consistent with the present application. But merely as exemplifications of systems and methods consistent with certain aspects of the application, as recited in the claims.
The term "module," as used in various embodiments of the present application, may refer to any known or later developed hardware, software, firmware, artificial intelligence, fuzzy logic, or combination of hardware and/or software code that is capable of performing the functionality associated with that element.
The term "remote control" as used in the embodiments of the present application refers to a component of an electronic device (such as the display device disclosed in the present application) that is capable of wirelessly controlling the electronic device, typically over a short distance. The component may typically be connected to the electronic device using infrared and/or Radio Frequency (RF) signals and/or bluetooth, and may also include functional modules such as WiFi, wireless USB, bluetooth, motion sensors, etc. For example: the hand-held touch remote controller replaces most of the physical built-in hard keys in the common remote control device with the user interface in the touch screen.
The term "gesture" as used in the embodiments of the present application refers to a user behavior used to express an intended idea, action, purpose, or result through a change in hand shape or an action such as hand movement.
The term "hardware system" used in the embodiments of the present application may refer to a physical component having computing, controlling, storing, inputting and outputting functions, which is formed by a mechanical, optical, electrical and magnetic device such as an Integrated Circuit (IC), a Printed Circuit Board (PCB) and the like. In various embodiments of the present application, a hardware system may also be generally referred to as a motherboard (motherboard) or a host chip or controller.
The rotary television is a novel intelligent television and mainly comprises a display and a rotary component. The display can be fixed on a wall or a support through the rotating assembly, and the placement direction of the display can be adjusted through the rotating assembly so as to be rotationally suitable for display pictures with different aspect ratios. For example, in most cases the display is positioned horizontally to display video frames with aspect ratios of 16:9, 18:9, etc. When the aspect ratio of the video frame is 9:16, 9:18, etc., the horizontally placed display needs to be scaled and black areas are displayed on both sides of the display. Therefore, the display can be rotated to a vertically placed state by the rotating assembly to accommodate video pictures of 9:16, 9:18, etc. proportions.
The power-on starting process of the rotating television generally comprises the following steps: a user inputs a starting-up instruction; displaying a startup logo interface; displaying a starting-up animation interface; and displaying a homepage interface. For the startup logo interface, the startup animation interface and the homepage interface, the rotary television can display different contents in different states according to the display in the startup process. However, since the initial rotation state of the display when the operating system is powered on is a landscape state by default, in the process of displaying the power-on logo interface, the power-on animation interface and the homepage interface, the displayed picture rotates when detecting that the displayed picture does not accord with the current rotation state, so as to adapt to the current rotation state. For example, when the mobile terminal is started and operated in the vertical screen state, the display displays the startup logo interface, the startup animation interface and the homepage interface according to the default horizontal screen state, and after displaying for a period of time, the interfaces are rotated into the vertical screen state according to the detected current vertical screen state.
In order to alleviate the influence of the image rotation process on the user viewing process, the application provides a display device and a startup interface display method, wherein the display device has a rotation function, and the startup interface display method can be applied to the display device. The display device may be a rotary television, or other intelligent terminal devices with a screen rotation function, such as a mobile phone, a tablet computer, and the like.
Referring to fig. 1A, an application scenario diagram of a display device according to some embodiments of the present application is provided. As shown in fig. 1A, thecontrol apparatus 100 and thedisplay device 200 may communicate with each other in a wired or wireless manner.
Among them, thecontrol apparatus 100 is configured to control thedisplay device 200, which may receive an operation instruction input by a user and convert the operation instruction into an instruction recognizable and responsive by thedisplay device 200, serving as an intermediary for interaction between the user and thedisplay device 200. Such as: the user operates the channel up/down key on thecontrol device 100, and thedisplay device 200 responds to the channel up/down operation.
Thecontrol device 100 may be aremote controller 100A, which includes infrared protocol communication or bluetooth protocol communication, and other short-distance communication methods, etc. to control thedisplay apparatus 200 in a wireless or other wired manner. The user may input a user instruction through a key on a remote controller, voice input, control panel input, etc., to control thedisplay apparatus 200. Such as: the user can input a corresponding control command through a volume up/down key, a channel control key, up/down/left/right moving keys, a voice input key, a menu key, a power on/off key, etc. on the remote controller, to implement the function of controlling thedisplay device 200.
Thecontrol device 100 may also be an intelligent device, such as amobile terminal 100B, a tablet computer, a notebook computer, and the like. For example, thedisplay device 200 is controlled using an application program running on the smart device. The application program may provide various controls to a user through an intuitive User Interface (UI) on a screen associated with the smart device through configuration.
For example, themobile terminal 100B may install a software application with thedisplay device 200 to implement connection communication through a network communication protocol for the purpose of one-to-one control operation and data communication. Such as: themobile terminal 100B may be caused to establish a control instruction protocol with thedisplay device 200, and the functions of the physical keys as arranged by theremote control 100A may be implemented by operating various function keys or virtual controls of the user interface provided on themobile terminal 100B. The audio and video content displayed on themobile terminal 100B may also be transmitted to thedisplay device 200, so as to implement a synchronous display function.
Thedisplay apparatus 200 may provide a network television function of a broadcast receiving function and a computer support function. The display device may be implemented as a digital television, a web television, an Internet Protocol Television (IPTV), or the like.
Thedisplay device 200 may be a liquid crystal display, an organic light emitting display, a projection device. The specific display device type, size, resolution, etc. are not limited.
Thedisplay apparatus 200 also performs data communication with theserver 300 through various communication means. Here, thedisplay apparatus 200 may be allowed to be communicatively connected through a Local Area Network (LAN), a Wireless Local Area Network (WLAN), and other networks. Theserver 300 may provide various contents and interactions to thedisplay apparatus 200. By way of example, thedisplay device 200 may send and receive information such as: receiving Electronic Program Guide (EPG) data, receiving software program updates, or accessing a remotely stored digital media library. Theservers 300 may be a group or groups of servers, and may be one or more types of servers. Other web service contents such as a video on demand and an advertisement service are provided through theserver 300.
In some embodiments, as shown in FIG. 1B,display device 200 includes arotation assembly 276, acontroller 250, adisplay 275, aterminal interface 278 extending from the gap in the backplane, and arotation assembly 276 coupled to the backplane, therotation assembly 276 configured to rotate thedisplay 275. From the perspective of the front view of the display device, therotating component 276 can rotate the display to a vertical screen state, i.e. a state where the vertical side length of the screen is greater than the horizontal side length, or to a horizontal screen state, i.e. a state where the horizontal side length of the screen is greater than the vertical side length.
Fig. 2 is a block diagram illustrating the configuration of thecontrol device 100. As shown in fig. 2, thecontrol device 100 includes acontroller 110, amemory 120, acommunicator 130, auser input interface 140, auser output interface 150, and apower supply 160.
Thecontroller 110 includes a Random Access Memory (RAM)111, a Read Only Memory (ROM)112, aprocessor 113, a communication interface, and a communication bus. Thecontroller 110 is used to control the operation of thecontrol device 100, as well as the internal components of the communication cooperation, external and internal data processing functions.
Illustratively, when an interaction of a user pressing a key disposed on theremote controller 100A or an interaction of touching a touch panel disposed on theremote controller 100A is detected, thecontroller 110 may control to generate a signal corresponding to the detected interaction and transmit the signal to thedisplay device 200.
And amemory 120 for storing various operation programs, data and applications for driving and controlling thecontrol apparatus 100 under the control of thecontroller 110. Thememory 120 may store various control signal commands input by a user.
Thecommunicator 130 enables communication of control signals and data signals with thedisplay apparatus 200 under the control of thecontroller 110. Such as: thecontrol apparatus 100 transmits a control signal (e.g., a touch signal or a control signal) to thedisplay device 200 via thecommunicator 130, and thecontrol apparatus 100 may receive the signal transmitted by thedisplay device 200 via thecommunicator 130. Thecommunicator 130 may include aninfrared signal interface 131 and a radiofrequency signal interface 132. For example: when the infrared signal interface is used, the user input instruction needs to be converted into an infrared control signal according to an infrared control protocol, and the infrared control signal is sent to thedisplay device 200 through the infrared sending module. The following steps are repeated: when the rf signal interface is used, a user input command needs to be converted into a digital signal, and then the digital signal is modulated according to the rf control signal modulation protocol and then transmitted to thedisplay device 200 through the rf transmitting terminal.
Theuser input interface 140 may include at least one of amicrophone 141, atouch pad 142, asensor 143, a key 144, and the like, so that a user can input a user instruction regarding controlling thedisplay apparatus 200 to thecontrol apparatus 100 through voice, touch, gesture, press, and the like.
Theuser output interface 150 outputs a user instruction received by theuser input interface 140 to thedisplay apparatus 200, or outputs an image or voice signal received by thedisplay apparatus 200. Here, theuser output interface 150 may include anLED interface 151, avibration interface 152 generating vibration, asound output interface 153 outputting sound, adisplay 154 outputting images, and the like. For example, theremote controller 100A may receive an output signal such as audio, video, or data from theuser output interface 150 and display the output signal in the form of an image on thedisplay 154, an audio on thesound output interface 153, or a vibration on thevibration interface 152.
And apower supply 160 for providing operation power support for each element of thecontrol device 100 under the control of thecontroller 110. In the form of a battery and associated control circuitry.
A hardware configuration block diagram of thedisplay device 200 is exemplarily shown in fig. 3. As shown in fig. 3, thedisplay apparatus 200 may include atuner demodulator 210, acommunicator 220, adetector 230, anexternal device interface 240, acontroller 250, amemory 260, auser interface 265, avideo processor 270, adisplay 275, a rotatingassembly 276, anaudio processor 280, an audio output interface 285, and apower supply 290.
Therotating assembly 276 may include a driving motor, a rotating shaft, and the like. Wherein, the driving motor can be connected to thecontroller 250 and output the rotation angle under the control of thecontroller 250; one end of the rotation shaft is connected to a power output shaft of the driving motor, and the other end is connected to thedisplay 275, so that thedisplay 275 can be fixedly mounted on a wall or a bracket through therotation member 276.
Therotating assembly 276 may also include other components, such as a transmission component, a detection component, and the like. Wherein, the transmission component can adjust the rotating speed and the torque output by therotating component 276 through a specific transmission ratio, and can be in a gear transmission mode; the detection means may be composed of a sensor, such as an angle sensor, an attitude sensor, or the like, provided on the rotation shaft. These sensors may detect parameters such as the angle at which therotating assembly 276 is rotated and transmit the detected parameters to thecontroller 250, so that thecontroller 250 can determine or adjust the state of thedisplay apparatus 200 according to the detected parameters. In practice, rotatingassembly 276 may include, but is not limited to, one or more of the components described above.
Thetuner demodulator 210 receives the broadcast television signal in a wired or wireless manner, may perform modulation and demodulation processing such as amplification, mixing, and resonance, and is configured to demodulate, from a plurality of wireless or wired broadcast television signals, an audio/video signal carried in a frequency of a television channel selected by a user, and additional information (e.g., EPG data).
Thetuner demodulator 210 is responsive to the user selected frequency of the television channel and the television signal carried by the frequency, as selected by the user and controlled by thecontroller 250.
Thetuner demodulator 210 can receive a television signal in various ways according to the broadcasting system of the television signal, such as: terrestrial broadcasting, cable broadcasting, satellite broadcasting, internet broadcasting, or the like; and according to different modulation types, a digital modulation mode or an analog modulation mode can be adopted; and can demodulate the analog signal and the digital signal according to the different kinds of the received television signals.
In other exemplary embodiments, thetuning demodulator 210 may also be in an external device, such as an external set-top box. In this way, the set-top box outputs a television signal after modulation and demodulation, and inputs the television signal into thedisplay apparatus 200 through theexternal device interface 240.
Thecommunicator 220 is a component for communicating with an external device or an external server according to various communication protocol types. For example, thedisplay apparatus 200 may transmit content data to an external apparatus connected via thecommunicator 220, or browse and download content data from an external apparatus connected via thecommunicator 220. Thecommunicator 220 may include a network communication protocol module or a near field communication protocol module, such as a WIFI module 221, a bluetoothcommunication protocol module 222, and a wired ethernetcommunication protocol module 223, so that thecommunicator 220 may receive a control signal of thecontrol device 100 according to the control of thecontroller 250 and implement the control signal as a WIFI signal, a bluetooth signal, a radio frequency signal, and the like.
Thedetector 230 is a component of thedisplay apparatus 200 for collecting signals of an external environment or interaction with the outside. Thedetector 230 may include a sound collector 231, such as a microphone, which may be used to receive a user's sound, such as a voice signal of a control instruction of the user to control thedisplay device 200; alternatively, ambient sounds may be collected that identify the type of ambient scene, enabling thedisplay device 200 to adapt to ambient noise.
In some other exemplary embodiments, thedetector 230, which may further include animage collector 232, such as a camera, a video camera, etc., may be configured to collect external environment scenes to adaptively change the display parameters of thedisplay device 200; and the function of acquiring the attribute of the user or interacting gestures with the user so as to realize the interaction between the display equipment and the user.
In some other exemplary embodiments, thedetector 230 may further include a light receiver for collecting the intensity of the ambient light to adapt to the display parameter variation of thedisplay device 200.
In some other exemplary embodiments, thedetector 230 may further include a temperature sensor, such as by sensing an ambient temperature, and thedisplay device 200 may adaptively adjust a display color temperature of the image. For example, when the temperature is higher, thedisplay apparatus 200 may be adjusted to display a color temperature of an image that is cooler; when the temperature is lower, thedisplay device 200 may be adjusted to display a warmer color temperature of the image.
Theexternal device interface 240 is a component for providing thecontroller 250 to control data transmission between thedisplay apparatus 200 and an external apparatus. Theexternal device interface 240 may be connected to an external apparatus such as a set-top box, a game device, a notebook computer, etc. in a wired/wireless manner, and may receive data such as a video signal (e.g., moving image), an audio signal (e.g., music), additional information (e.g., EPG), etc. of the external apparatus.
Theexternal device interface 240 may include: a High Definition Multimedia Interface (HDMI) terminal 241, a Composite Video Blanking Sync (CVBS)terminal 242, an analog ordigital Component terminal 243, a Universal Serial Bus (USB)terminal 244, a Component terminal (not shown), a red, green, blue (RGB) terminal (not shown), and the like.
Thecontroller 250 controls the operation of thedisplay device 200 and responds to the operation of the user by running various software control programs (such as an operating system and various application programs) stored on thememory 260.
As shown in fig. 3, thecontroller 250 includes a Random Access Memory (RAM)251, a Read Only Memory (ROM)252, agraphics processor 253, aCPU processor 254, acommunication interface 255, and acommunication bus 256. The RAM251, the ROM252, thegraphic processor 253, and theCPU processor 254 are connected to each other through acommunication bus 256 through acommunication interface 255.
The ROM252 stores various system control instructions. When thedisplay apparatus 200 starts power-on upon receiving the power-on signal, theCPU processor 254 executes the system control instructions in the ROM252, copies the operating system stored in thememory 260 to the RAM251, and starts running the start-up operating system. After the start of the operating system is completed, theCPU processor 254 copies the various application programs in thememory 260 to the RAM251 and then starts running and starting the various application programs.
And agraphic processor 253 for generating various graphic objects such as icons, operation menus, and user input instruction display graphics, etc. Thegraphic processor 253 may include an operator for performing an operation by receiving various interactive instructions input by a user, and further displaying various objects according to display attributes; and a renderer for generating various objects based on the operator and displaying the rendered result on thedisplay 275.
ACPU processor 254 for executing operating system and application program instructions stored inmemory 260. And according to the received user input instruction, processing of various application programs, data and contents is executed so as to finally display and play various audio-video contents.
In some example embodiments, theCPU processor 254 may comprise a plurality of processors. The plurality of processors may include one main processor and a plurality of or one sub-processor. A main processor for performing some initialization operations of thedisplay apparatus 200 in the display apparatus preload mode and/or operations of displaying a screen in the normal mode. A plurality of or one sub-processor for performing an operation in a state of a standby mode or the like of the display apparatus.
Thecommunication interface 255 may include a first interface to an nth interface. These interfaces may be network interfaces that are connected to external devices via a network.
Thecontroller 250 may control the overall operation of thedisplay apparatus 200. For example: in response to receiving a user input command for selecting a GUI object displayed on thedisplay 275, thecontroller 250 may perform an operation related to the object selected by the user input command.
Where the object may be any one of the selectable objects, such as a hyperlink or an icon. The operation related to the selected object is, for example, an operation of displaying a link to a hyperlink page, document, image, or the like, or an operation of executing a program corresponding to the object. The user input command for selecting the GUI object may be a command input through various input means (e.g., a mouse, a keyboard, a touch panel, etc.) connected to thedisplay apparatus 200 or a voice command corresponding to a voice spoken by the user.
Amemory 260 for storing various types of data, software programs, or applications for driving and controlling the operation of thedisplay device 200. Thememory 260 may include volatile and/or nonvolatile memory. And the term "memory" includes thememory 260, the RAM251 and the ROM252 of thecontroller 250, or a memory card in thedisplay device 200.
In some embodiments, thememory 260 is specifically used for storing an operating program for driving thecontroller 250 of thedisplay device 200; storing various application programs built in thedisplay apparatus 200 and downloaded by a user from an external apparatus; data such as visual effect images for configuring various GUIs provided by thedisplay 275, various objects related to the GUIs, and selectors for selecting GUI objects are stored.
In some embodiments,memory 260 is specifically configured to store drivers fortuner demodulator 210,communicator 220,detector 230,external device interface 240,video processor 270,display 275,audio processor 280, etc., and related data, such as external data (e.g., audio-visual data) received from the external device interface or user data (e.g., key information, voice information, touch information, etc.) received by the user interface.
In some embodiments,memory 260 specifically stores software and/or programs representing an Operating System (OS), which may include, for example: a kernel, middleware, an Application Programming Interface (API), and/or an application program. Illustratively, the kernel may control or manage system resources, as well as functions implemented by other programs (e.g., the middleware, APIs, or applications); at the same time, the kernel may provide an interface to allow middleware, APIs, or applications to access the controller to enable control or management of system resources.
A block diagram of the architectural configuration of the operating system in the memory of thedisplay device 200 is illustrated in fig. 4. The operating system architecture comprises an application layer, a middleware layer and a kernel layer from top to bottom.
The application layer, the application programs built in the system and the non-system-level application programs belong to the application layer. Is responsible for direct interaction with the user. The application layer may include a plurality of applications such as a setup application, a post application, a media center application, and the like. These applications may be implemented as Web applications that execute based on a WebKit engine, and in particular may be developed and executed based on HTML5, Cascading Style Sheets (CSS), and JavaScript.
Here, HTML, which is called HyperText Markup Language (HyperText Markup Language), is a standard Markup Language for creating web pages, and describes the web pages by Markup tags, where the HTML tags are used to describe characters, graphics, animation, sound, tables, links, etc., and a browser reads an HTML document, interprets the content of the tags in the document, and displays the content in the form of web pages.
CSS, known as Cascading Style Sheets (Cascading Style Sheets), is a computer language used to represent the Style of HTML documents, and may be used to define Style structures, such as fonts, colors, locations, etc. The CSS style can be directly stored in the HTML webpage or a separate style file, so that the style in the webpage can be controlled.
JavaScript, a language applied to Web page programming, can be inserted into an HTML page and interpreted and executed by a browser. The interaction logic of the Web application is realized by JavaScript. The JavaScript can package a JavaScript extension interface through a browser, realize the communication with the kernel layer,
the middleware layer may provide some standardized interfaces to support the operation of various environments and systems. For example, the middleware layer may be implemented as multimedia and hypermedia information coding experts group (MHEG) middleware related to data broadcasting, DLNA middleware which is middleware related to communication with an external device, middleware which provides a browser environment in which each application program in the display device operates, and the like.
The kernel layer provides core system services, such as: file management, memory management, process management, network management, system security authority management and the like. The kernel layer may be implemented as a kernel based on various operating systems, for example, a kernel based on the Linux operating system.
The kernel layer also provides communication between system software and hardware, and provides device driver services for various hardware, such as: provide display driver for the display, provide camera driver for the camera, provide button driver for the remote controller, provide wiFi driver for the WIFI module, provide audio driver for audio output interface, provide power management drive for Power Management (PM) module etc..
In FIG. 3,user interface 265, receives various user interactions. Specifically, it is used to transmit an input signal of a user to thecontroller 250 or transmit an output signal from thecontroller 250 to the user. For example, theremote controller 100A may transmit an input signal, such as a power switch signal, a channel selection signal, a volume adjustment signal, etc., input by the user to theuser interface 265, and then the input signal is transferred to thecontroller 250 through theuser interface 265; alternatively, theremote controller 100A may receive an output signal such as audio, video, or data output from theuser interface 265 via thecontroller 250, and display the received output signal or output the received output signal in audio or vibration form.
In some embodiments, a user may enter user commands on a Graphical User Interface (GUI) displayed on thedisplay 275, and theuser interface 265 receives the user input commands through the GUI. Specifically, theuser interface 265 may receive user input commands for controlling the position of a selector in the GUI to select different objects or items. Among these, "user interfaces" are media interfaces for interaction and information exchange between an application or operating system and a user, which enable the conversion between an internal form of information and a form acceptable to the user. A common presentation form of a user interface is a Graphical User Interface (GUI), which refers to a user interface related to computer operations and displayed in a graphical manner. It may be an interface element such as an icon, window, control, etc. displayed in the display of the electronic device, where the control may include a visual interface element such as an icon, control, menu, tab, text box, dialog box, status bar, channel bar, Widget, etc.
Alternatively, the user may input a user command by inputting a specific sound or gesture, and theuser interface 265 receives the user input command by recognizing the sound or gesture through the sensor.
Thevideo processor 270 is configured to receive an external video signal, and perform video data processing such as decompression, decoding, scaling, noise reduction, frame rate conversion, resolution conversion, and image synthesis according to a standard codec protocol of the input signal, so as to obtain a video signal that is directly displayed or played on thedisplay 275.
Illustratively, thevideo processor 270 includes a demultiplexing module, a video decoding module, an image synthesizing module, a frame rate conversion module, a display formatting module, and the like.
The demultiplexing module is configured to demultiplex an input audio/video data stream, where, for example, an input MPEG-2 stream (based on a compression standard of a digital storage media moving image and voice), the demultiplexing module demultiplexes the input audio/video data stream into a video signal and an audio signal.
And the video decoding module is used for processing the video signal after demultiplexing, including decoding, scaling and the like.
And the image synthesis module is used for carrying out superposition mixing processing on the GUI signal input by the user or generated by the user and the video image after the zooming processing by the graphic generator so as to generate an image signal for display.
The frame rate conversion module is configured to convert a frame rate of an input video, for example, convert a frame rate of an input 60Hz video into a frame rate of 120Hz or 240Hz, where a common format is implemented by using, for example, an interpolation frame method.
And a display formatting module for converting the signal output by the frame rate conversion module into a signal conforming to a display format of a display, such as converting the format of the signal output by the frame rate conversion module to output an RGB data signal.
Adisplay 275 for receiving the image signal from thevideo processor 270 and displaying the video content, the image and the menu manipulation interface. The display video content may be from the video content in the broadcast signal received by the tuner-demodulator 210, or from the video content input by thecommunicator 220 or theexternal device interface 240. Thedisplay 275, while displaying a user manipulation interface UI generated in thedisplay apparatus 200 and used to control thedisplay apparatus 200.
And, thedisplay 275 may include a display component for presenting a picture and a driving component for driving the display of an image. Alternatively, a projection device and projection screen may be included, provideddisplay 275 is a projection display.
Theaudio processor 280 is configured to receive an external audio signal, decompress and decode the received audio signal according to a standard codec protocol of the input signal, and perform audio data processing such as noise reduction, digital-to-analog conversion, and amplification processing to obtain an audio signal that can be played by thespeaker 286.
Illustratively,audio processor 280 may support various audio formats. Such as MPEG-2, MPEG-4, Advanced Audio Coding (AAC), high efficiency AAC (HE-AAC), and the like.
The audio output interface 285 is used for receiving an audio signal output by theaudio processor 280 under the control of thecontroller 250, and the audio output interface 285 may include aspeaker 286 or an externalsound output terminal 287, such as an earphone output terminal, for outputting to a generating device of an external device.
In other exemplary embodiments,video processor 270 may comprise one or more chips.Audio processor 280 may also comprise one or more chips. And, in other exemplary embodiments, thevideo processor 270 and theaudio processor 280 may be separate chips or may be integrated with thecontroller 250 in one or more chips.
And apower supply 290 for supplying power supply support to thedisplay apparatus 200 from the power input from the external power source under the control of thecontroller 250. Thepower supply 290 may be a built-in power supply circuit installed inside thedisplay apparatus 200 or may be a power supply installed outside thedisplay apparatus 200.
For therotatable display device 200, in practical applications, therotating component 276 can bring thedisplay 275 to adjust at any rotation angle to change the posture of thedisplay 275. In general, therotation assembly 276 may rotate thedisplay 275 in a plane (i.e., a vertical plane viewed from the front of the user) in which the display screen is located, thereby causing thedisplay 275 to be in different rotational states.
The rotation state is a plurality of specific states in which thedisplay 275 is located, and may be set in a plurality of different forms according to the posture of thedisplay 275. For example, landscape state, portrait state, tilt state, etc. The horizontal screen state and the vertical screen state are rotation states used by most users, and can be respectively suitable for a horizontal media asset scene and a vertical media asset scene. Therefore, in some embodiments of the present application, the landscape state and the portrait state may be referred to as a standard state. While the tilted state is generally a state in which thedisplay 275 is rotated out of position or rotated excessively due to a failure of therotating assembly 276, the user rarely rotates thedisplay 275 to the tilted state intentionally. That is, in some embodiments of the present application, the tilted state may also be referred to as a non-standard state.
Wherein the landscape state refers to a state in which the length (width) of thedisplay 275 in the horizontal direction is greater than the length (height) of thedisplay 275 in the vertical direction when viewed from the front of thedisplay 275; the vertical screen state is a state in which the length (width) of thedisplay 275 in the horizontal direction is smaller than the length (height) of thedisplay 275 in the vertical direction when viewed from the front of thedisplay 275. Obviously, the vertical direction means substantially vertical in the present application, and the horizontal direction means substantially horizontal. And the other rotation states except the horizontal screen state and the vertical screen state are the inclined states. The rotation angle of thedisplay 275 is different in different tilted states.
In some embodiments, the display content presented on thedisplay 275 may differ in different rotational states, which may be embodied in specific playscreen content, UI interface layout, and the like. For example, the landscape state of thedevice 200 is displayed in some embodiments of the present application as shown in FIG. 5A. The mode of operation when thedisplay 275 is in the landscape state may be referred to as the landscape viewing mode and the mode of operation when thedisplay 275 is in the portrait state may be referred to as the portrait viewing mode.
Therotating assembly 276 can fix thedisplay device 200 and rotate thedisplay 275 under the control of thecontroller 250, so that thedisplay 275 can be in different rotating states. Therotating assembly 276 may be secured to the back of thedisplay 275, with therotating assembly 276 being secured to a wall or support. Therotating assembly 276 may receive a control command from thecontroller 250 to rotate thedisplay 275 in a vertical plane such that thedisplay 275 is in a landscape state or a portrait state.
Thedisplay 275 may be rotated 90 degrees clockwise or counterclockwise by the rotation of therotation assembly 276, adjusting thedisplay 275 to a portrait position, as shown in FIG. 5B. In the vertical screen state, thedisplay 275 may display the user interface corresponding to the vertical screen state and have the interface layout and the interaction mode corresponding to the vertical screen state. In the vertical screen media asset watching mode, a user can watch vertical screen media assets such as short videos and cartoons. Since thecontroller 250 in thedisplay device 200 is further in communication connection with theserver 300, the media asset data corresponding to the vertical screen can be acquired by calling the interface of theserver 300 in the vertical screen state.
It should be noted that the horizontal screen state is mainly used for displaying horizontal media resources such as dramas and movies, and the vertical screen state is mainly used for displaying vertical media resources such as short videos and cartoons. The horizontal screen state and the vertical screen state are only two different display states, and do not limit the displayed content, for example, vertical media such as short videos, cartoons and the like can still be displayed in the horizontal screen state; horizontal media assets such as TV dramas and movies can still be displayed in the vertical screen state, and only the incompatible display windows need to be compressed and adjusted in the vertical screen state.
The different rotation states can be switched in real time, and since the interface layout and content presented by thedisplay device 200 are different in the different rotation states, the display interface UI may be changed during the switching process. For example, during the process of switching from the landscape screen state to the portrait screen state, the display screen is switched from the landscape mode to the portrait mode so as to adapt to the display scale of the portrait mode. Typically the frame of the switching process may be a flipping action of the display pattern.
The rotational state of thedisplay 275 when thedisplay device 200 is powered on may be a landscape state or a portrait state, and generally depends on the rotational state of thedisplay 275 when thedisplay device 200 is powered off. For example, when the user turns off in the vertical screen state, the rotation state of thedisplay 275 is turned on next time to be in the vertical screen state. For thedisplay 275 in the vertical screen state, the start-up interface and the home interface in the vertical screen state should be displayed, but in the process of starting up thedisplay device 200, the display screen may be turned over only after the home interface is displayed, that is, the default horizontal screen state is started, and after the start-up is completed, the home interface is displayed by switching the display state.
For example, taking an intelligent terminal device based on an android system as an example, after startup, a logo picture is displayed first, and meanwhile, a surfaceFlinger service is initialized into a landscape state, and startup animation is started. And initializing the windows manager into a horizontal screen state, and after the windows manager is initialized, notifying the surfeFinger service, wherein the surfeFinger service stops the boot animation (bootanimation). And then, the App is started to detect the state of the gravity sensor, and when the current rotation state is found to be a vertical screen state, the App performs self-adaptive rotation, so that the homepage interface is rotated to a vertical state.
Therefore, due to the display mode of the starting-up interface, the display contents watched by the user are as follows in sequence: and starting up to display a logo picture, a starting-up animation picture, a horizontal screen homepage interface and a self-adaptive rotation adjusted vertical screen homepage interface. However, since thedisplay 275 is in the vertical screen state when thedisplay device 200 is powered on, all interfaces displayed before the vertical screen homepage interface are in the inclined state, i.e., the startup logo picture, the startup animation picture, and the landscape homepage interface all have the problem of display errors.
In order to solve the above problem of display errors, some embodiments of the present application provide a method for displaying a booting interface, as shown in fig. 6 and 7, the method includes the following steps:
S1: and acquiring a starting-up instruction.
The boot instruction is a control instruction for controlling the boot process, which is input by a user, and may be a physical instruction or a digital instruction according to the boot starting manner. For example, thedisplay device 200 may be powered on and started by a user turning on a power switch (key) on thedisplay device 200, where the power-on command is an electrical signal generated when the user presses the power switch. When the user performs soft-start through a power key on thecontrol device 100 such as a remote controller, thedisplay device 200 may still be started to operate, and the power-on command is a control signal input by thecontrol device 100.
Thecontroller 250 may start running the operating system boot program in response to the boot instruction after acquiring the boot instruction. For example, an initialization process for performing each service, self-detection of the system, running a hardware driver, and starting a display control program, etc. The boot program may consume a certain amount of time to complete the boot program according to different operating systems, and thus during this time, thedisplay device 200 may display the boot interface through thedisplay 275.
In some embodiments, the power-on interface may include a power-on logo interface and a power-on animation interface. The content displayed on the startup sign interface is usually the brand logo of thedisplay device 200, and may be a static or dynamic graphic picture; the content displayed by the boot animation interface may be a specific boot animation set by the provider of thedisplay device 200, or some promotional video pictures. Generally, the boot logo interface is displayed by the display content set in the boot program and along with the running of the boot program. The starting animation interface is displayed by starting animation service, such as bootanimation service, by running corresponding control program.
For example, when the boot animation is a promotional video image, thecontroller 250 calls bootanimation service along with the operation of the boot program, and the bootanimation service acquires the cached video file from the memory and plays the video file. Or after the bootanimation service is called, the bootanimation service acquires the video file from theserver 300 through theconnected server 300 to play.
In order to display the booting interface, thedisplay device 200 may display different types of booting interfaces in different rotation states, that is, the booting interface may also include a horizontal booting interface and a vertical booting interface. Correspondingly, the transverse starting-up interface comprises a transverse starting-up sign interface and a transverse starting-up animation interface; the vertical starting interface comprises a vertical starting sign interface and a vertical starting animation interface.
S2: and responding to the starting-up instruction, and detecting the current rotation state of the display.
After receiving the power-on command, thecontroller 250 may first detect a current rotation state of thedisplay 275, and determine a current posture of thedisplay 275, so as to control the display of the power-on interface according to the current rotation state.
In practice, thecontroller 250 may detect the current rotation state of thedisplay 275 through a posture sensor built in thedisplay device 200. For example, an electronic level may be built in thedisplay device 200, and after receiving the power-on command, thecontroller 250 may detect the current horizontal state of thedisplay 275 through the electronic level, and when detecting that the horizontal state parameter of thecurrent display 275 is within a reasonable error range, determine that the current rotation state of thedisplay 275 is the landscape state; otherwise, when the horizontal state parameter of thecurrent display 275 is detected not to be within a reasonable error range, the current rotation state of thedisplay 275 is determined to be the vertical screen state.
In some embodiments, the position relationship between the gravity direction and thedisplay 275 can also be detected by a gravitational acceleration sensor built in thedisplay device 200 to determine the current rotation state. For example, after receiving the power-on command, thecontroller 250 may detect an angle between the current gravity direction and a width (or short) direction of the screen of thedisplay 275 through the gravity acceleration sensor, so as to determine the current rotation state of thedisplay 275. If the detected gravity direction is perpendicular to the screen widthwise direction of the display 275 (or the detected gravity direction is parallel to the screen widthwise direction of the display 275), determining that the current rotation state of thedisplay 275 is a landscape state; if the detected gravity direction is parallel to the screen width direction of the display 275 (or the detected gravity direction is perpendicular to the screen short side direction of the display 275), it is determined that the current rotation state of thedisplay 275 is the portrait state.
S3: and if the current rotation state of the display is the vertical screen state, controlling the display to display a vertical starting interface and detecting a vertical screen initialization process.
Thecontroller 250 may control thedisplay 275 to display different power-on interfaces according to the detection result of the current rotation state of thedisplay 275. If the current rotation state of thedisplay 275 is the portrait screen state, thedisplay 275 is controlled to display a portrait power-on interface.
The vertical screen initialization process may also be detected after detecting that the current rotational state of thedisplay 275 is a vertical screen state. Because it is different from the default rotational state of thedisplay 275 of the operating system, if the booting is performed in the native booting manner of the operating system, the booting interface and the home interface may be tilted, which may affect the user experience. Therefore, thecontroller 250 will not start according to the native start-up mode of the operating system any more, adjust the display order of the startup interface, and display the homepage interface after detecting that the vertical screen initialization process is completed, thereby avoiding the image turning-over situation.
S4: and after the vertical screen initialization process is completed, closing the vertical starting interface and starting a vertical screen homepage display interface.
In practical application, by detecting the vertical screen initialization process, the vertical screen homepage display interface can be started after the vertical screen initialization process is completed. At this time, because the vertical screen initialization process is completed, the App corresponding to the home page interface no longer detects that the current picture does not conform to the rotation state, that is, the UI is in a locked form, so that the startup interface can be closed to directly start up and display the vertical screen home page interface, and the picture cannot be turned over.
Taking an operating system based on an android platform as an example, in the process of displaying the start-up interface, thecontroller 250 may monitor a vertical screen setting process of the windows manager, so that after the windows manager completes the vertical screen setting, the start-up animation interface is displayed, and after the start-up animation interface is displayed, the vertical screen homepage interface is started to be displayed.
According to the above technical solution, the method for displaying the power-on interface provided in the above embodiment can detect the current rotation state of thedisplay 275 after the power-on instruction is obtained, and can avoid the error of oblique display of the power-on interface by displaying the vertical power-on interface when the current rotation state is the vertical screen state. Meanwhile, in the process of displaying the startup interface, thecontroller 250 may also detect the vertical screen initialization process, so that after the vertical screen initialization process is completed, the vertical startup interface is closed, and the vertical screen homepage display interface is started, so that the homepage interface displayed on thedisplay 275 is directly in a vertical UI layout, the situation of image turning does not occur, and the user experience is improved.
In some embodiments of the present application, the current rotation state of thedisplay 275 may also be detected by an angle sensor built into thedisplay device 200. An angle sensor may be provided on a rotating shaft of the driving motor in therotating assembly 276 to detect a change in the rotational angle of therotating assembly 276. Therefore, as shown in fig. 8, the step of detecting the current rotation state of the display further includes:
S201: acquiring the rotation angle of a driving motor of the rotating assembly;
s202: if the rotation angle is within a preset vertical screen rotation angle interval, determining that the current rotation state of the display is a vertical screen state;
s203: and if the rotation angle is within the preset horizontal screen rotation angle interval, determining that the current rotation state of the display is the horizontal screen state.
In this embodiment, after receiving the power-on command, thecontroller 250 may obtain the rotation angle of the driving motor of therotating component 276 through the angle sensor. The rotation angle of the driving motor is an angle variation of the rotation shaft of the driving motor with respect to a set initial state to reflect the rotated angle of thedisplay 275. For example, if thedisplay 275 is set to have a corresponding posture of 0 in the landscape state, the angle sensor may detect that the rotation angle of the driving motor is 90 degrees when thedisplay 275 is in the portrait state.
The detected rotation angle may be determined by determining which rotation state is equal to the specific value, thereby determining the current rotation state of thedisplay 275. For example, when it is detected that the rotation angle of the driving motor is 0 degrees, it is determined that the current rotation state of thedisplay 275 is the landscape state; when it is detected that the rotation angle of the driving motor is 90 degrees, it is determined that the current rotation state of thedisplay 275 is the portrait screen state.
In practical applications, a certain mechanical return difference exists in therotating assembly 276, and the using environment may have a certain influence on therotating assembly 276, so that when thedisplay 275 is rotated, the rotation angle may not be accurately controlled to a certain set value, and therefore, a rotation angle interval may be preset for each rotation state according to a statistical reasonable error range, so that when the rotation angle of the driving motor is within the preset rotation angle interval, it is determined that thedisplay 275 is in a corresponding rotation state. If the rotation angle is within the preset vertical screen rotation angle interval, determining that the current rotation state of the display is the vertical screen state; and if the rotation angle is within the preset horizontal screen rotation angle interval, determining that the current rotation state of the display is the horizontal screen state.
For example, if the preset vertical screen rotation angle interval is (85 °, 95 °), and the preset horizontal screen rotation angle interval is (-5 °, 5 °), and if the rotation angle of the driving motor of therotating assembly 276 is 87 ° as obtained by the angle sensor, since the rotation angle of 87 ° is within the preset vertical screen rotation angle interval (85 °, 95 °), it is determined that the current rotation state of thedisplay 275 is the vertical screen state.
In some cases, the rotation angle of the driving motor of therotating assembly 276 is not within the vertical screen rotation angle interval and is not within the horizontal screen rotation angle interval. For example, if the angle sensor detects that the rotation angle of the driving motor of therotating assembly 276 is 78 °, it is determined that the current rotation state of thedisplay 275 is not the vertical screen state or the horizontal screen state according to the rotation angle interval shown in the above example. In this case, the operating system may default to the landscape state in order to display the opening machine interface. However, in this display mode, when thedisplay 275 is close to the portrait state, the displayed tilt degree of the booting interface is large, and the display effect is not good.
Thus, in some embodiments, the appropriate rotational state may be determined by setting consecutive portrait and landscape corner intervals. An angle judgment threshold value can be preset, and the corner intervals on the two sides of the threshold value are respectively used as a vertical screen corner interval and a horizontal screen corner interval. For example, if the angle determination threshold is 45 °, the vertical screen angle section is (45 °, 90 °), and the horizontal screen angle section is (0 °, 45 °). When the rotation angle of the driving motor of the rotatingmember 276 is determined to be 78 ° by the angle sensor, it is determined that the rotation angle is within the vertical screen rotation angle interval (45 °, 90 °), thedisplay 275 is more likely to be in the vertical screen state, i.e., the current rotation state of thedisplay 275 is determined to be in the vertical screen state.
In some embodiments, when the power-on interface includes a power-on logo interface and a power-on animation interface, thecontroller 250 may further control thedisplay 275 to display the power-on logo interface after obtaining the power-on command. That is, after the user inputs the power-on command, thedisplay 275 may first display a logo interface to indicate that the power-on program of thedisplay apparatus 200 is currently started.
The Logo interface can be kept displayed until the boot program runs to the display boot animation interface. In some embodiments, as shown in FIG. 9, the power-on animation interface may be displayed according to the following steps:
S301: starting the starting animation interface;
s302: monitoring the initialization process of the vertical screen, and continuously displaying the starting mark interface in the monitoring process;
s303: and displaying the starting animation interface after monitoring that the initialization process of the vertical screen is completed.
Taking an operating system based on an android platform as an example, after thedisplay 275 is detected to be in a vertical screen state by starting, thecontroller 250 initializes the surfacemaker service to be a vertical screen, and then starts to set the windows manager as the vertical screen and starts the boot animation at the same time. After the startup animation is started, animation playing can be performed firstly without using a logo interface, and after the initialization of the windows manager vertical screen is completed, animation playing is performed. At this time, the windows manager and the surfaceFlinger services are initialized and completed in a vertical screen mode, so that the phenomenon of UI rotation cannot occur in the starting animation playing process.
In the above embodiment, by detecting the rotation state of thedisplay 275 and opening the interface display mode in the vertical screen direction when the rotation state is detected to be the vertical screen state, the interface rotation condition caused by changing the rotation state when the interface is started is alleviated, so that the experience is improved. While detecting the rotation state of thedisplay 275, if the detected rotation state is the landscape state, the display of the booting interface may still be completed according to the native boot-up manner of the operating system, that is, in some embodiments of the present application, as shown in fig. 6, if the current rotation state of the display is the landscape state, the booting interface display method further includes the following steps:
S501: controlling the display to display a transverse starting interface and executing a transverse screen initialization process;
s502: and after the horizontal screen initialization process is completed, closing the horizontal starting interface and starting a horizontal screen homepage display interface.
That is, upon detecting that the current rotation state of thedisplay 275 is the landscape state, thecontroller 250 may control thedisplay 275 to display the landscape boot interface first. The horizontal startup interface and the vertical startup interface are different in picture proportion, for example, for a startup logo picture, the startup logo picture can be displayed in a picture or video form, the horizontal logo picture can have the same content as the vertical logo picture, but the picture proportion is different, and the picture proportion of the horizontal logo picture is 16: 9, the picture scale of the vertical logo graph is 9: 16.
similarly, the boot animation can be displayed in different picture scales in the same manner, and different video contents can be displayed in different rotation states. For example, when the boot animation is a promotional video such as an advertisement, different video contents can be displayed in different rotation states, that is, video contents suitable for a landscape application, such as a movie, a tv drama preview, etc., are displayed in a landscape state; and displaying video content suitable for vertical screen application, such as short video advertisements and the like, in the vertical screen state.
While controlling thedisplay 275 to display the landscape boot image, thecontroller 250 may also control the execution of the landscape initialization procedure, such that after the landscape initialization procedure is completed, the landscape boot interface is closed and the landscape home interface is started. Taking an operating system based on an android platform as an example, the horizontal screen initialization process may include: initializing a surfaceFlinger service as a horizontal screen and starting a startup animation interface, and simultaneously setting a windows manager as the horizontal screen and displaying the startup animation interface; and finally, after the windows manager finishes the setting of the horizontal screen, closing the starting animation interface and starting the display horizontal screen homepage interface.
Unlike the vertical screen initialization process, in this embodiment, thecontroller 250 may directly display the power-on logo interface, initialize the surfaceflag service as the horizontal screen, and start the power-on animation interface after detecting that the current rotation state is the horizontal screen state. After the startup animation interface is started, the startup animation interface can be directly displayed without waiting for the completion of the horizontal screen setting process of the windows manager. In the process of displaying the power-on animation interface, thecontroller 250 may further detect a setting process of the windows manager, and may close the power-on animation interface and switch to display a cross screen homepage interface after detecting that the windows manager completes the cross screen setting. Because the current display state is consistent with the default rotation state form of the system, the boot animation can be directly displayed in the boot animation display mode after the boot animation interface is started, the boot logo does not need to be shielded, the display time of the boot logo interface and the boot animation interface can be balanced, and better user experience is brought.
In some embodiments of the present application, as shown in fig. 10, after the current rotation state of the display is the landscape state and the landscape home interface is finally displayed, the method further includes:
s601: detecting a rotation state of the display;
s602: if the rotation state of the display is a vertical screen state, changing the windows manager to be a vertical screen and changing the surfeFinger service to be a vertical screen;
s603: and vertically displaying the UI interface.
After displaying the landscape home interface, thecontroller 250 may also detect the rotation state of thedisplay 275 in real time, and if it is detected that the rotation state of thedisplay 275 is changed to the portrait state at a certain time, the windows manager and the surfefinger service may be changed to the portrait by setting, so as to display a vertical UI interface.
The change of the rotation state due to thedisplay 275 can be input by the user. For example, the user inputs a switching instruction by means of thecontrol device 100, the smart voice system, or the like; thecontroller 250 may control therotation assembly 276 to rotate in response to the switching instruction, thereby adjusting the rotation state of thedisplay 275. When thecontroller 250 detects that thedisplay 275 has been rotated to the portrait state, the windows manager and surfaceflag services are changed to the portrait, and the displayed UI interface is rotated according to the setting results of the windows manager and surfaceflag services.
According to the above technical solution, the power-on interface display method provided by the present application may detect the current rotation state of thedisplay 275 after receiving the power-on instruction, and display the power-on interface in different manners when thedisplay 275 is in different rotation states. And the vertical starting interface can be displayed for the vertical screen state and the vertical screen initialization process is detected, so that after the vertical screen initialization process is completed, the vertical starting interface is closed to start the vertical screen homepage display interface. Therefore, the method can avoid the rotation of the display picture and relieve the problem of display content errors through the shielding of the vertical starting interface in the vertical screen initialization process, and improves the user experience.
Based on the above method for displaying a booting interface, some embodiments of the present application provide a display device, including: adisplay 275, a rotatingassembly 276, and acontroller 250. Thedisplay 275 is configured to display a user interface, including displaying a power-on interface, a play screen, and the like. Therotating assembly 276 is connected to thedisplay 275 and configured to rotate thedisplay 275, so that thedisplay 275 can be in any rotating state, including a landscape state or a portrait state. Bothdisplay 275 androtating assembly 276 are electrically connected tocontroller 250.
Thecontroller 250 is further configured to perform the following program steps:
s1: acquiring a starting-up instruction;
s2: detecting a current rotation state of the display in response to the power-on instruction;
s3: if the current rotation state of the display is a vertical screen state, controlling the display to display a vertical starting interface and detecting a vertical screen initialization process;
s4: and after the vertical screen initialization process is completed, closing the vertical starting interface and starting a vertical screen homepage display interface.
In thedisplay device 200 provided in this embodiment, thecontroller 250 may detect the current rotation state of thedisplay 275 by obtaining a power-on instruction input by a user, and may control thedisplay 275 to display a vertical power-on interface if the current rotation state of thedisplay 275 is a vertical screen state. Meanwhile, thecontroller 250 also detects a vertical screen initialization process in the operating system startup process, and closes the vertical startup interface and starts the vertical screen homepage display interface after detecting that the vertical screen initialization process is completed, so that the problems of rotation of a display frame and error in display content are avoided, and the user experience is improved.
The embodiments provided in the present application are only a few examples of the general concept of the present application, and do not limit the scope of the present application. Any other embodiments extended according to the scheme of the present application without inventive efforts will be within the scope of protection of the present application for a person skilled in the art.