Movatterモバイル変換


[0]ホーム

URL:


EP1228422A1 - Operation method of user interface of hand-held device - Google Patents

Operation method of user interface of hand-held device

Info

Publication number
EP1228422A1
EP1228422A1EP00967937AEP00967937AEP1228422A1EP 1228422 A1EP1228422 A1EP 1228422A1EP 00967937 AEP00967937 AEP 00967937AEP 00967937 AEP00967937 AEP 00967937AEP 1228422 A1EP1228422 A1EP 1228422A1
Authority
EP
European Patent Office
Prior art keywords
display
data
user interface
operation method
hand
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP00967937A
Other languages
German (de)
French (fr)
Inventor
Jukka-Pekka METSÄVAINIO
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Myorigo Oy
Original Assignee
Myorigo Oy
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from FI992191Aexternal-prioritypatent/FI19992191A7/en
Application filed by Myorigo OyfiledCriticalMyorigo Oy
Publication of EP1228422A1publicationCriticalpatent/EP1228422A1/en
Withdrawnlegal-statusCriticalCurrent

Links

Classifications

Definitions

Landscapes

Abstract

An operation method of a user interface of a handheld device, like a palm-size or pocket-size PC, mobile telephone, communicator or equivalent, in which method data on a display of the device is changed for finding or selecting desired data, is characterized in that data (5a - 5e) on the display (2') of the device is changed responsive to certain movements (F) of the device in a three-dimensional space of use thereof, and said movements of the device and the corresponding changes of the data on the display are defined so that a movement and a corresponding change of data on the display have a cause-consequence relationship which is analogous to cause-consequence relationships of the three-dimensional space of perception and action of a user (1).

Description

Operation method of user interface of hand-held device
The invention is related to an operation method of the user interface of a hand-held data-processing device. A hand-held device means here a palm-top or pocket computer. mobile phone, communicator or like device, essential features of which include displaying data on the display of the device and in which data on the display of the device is changed for finding and selecting a desired data.
User interfaces of above mentioned devices are presently mostly of type''point and click". A general operation method of a user interface is browsing and selecting data by means of keys. Palm-top computers are usually managed with a touch sensitive display, a virtual keyboard, a pen. and often some auxiliary hard keys or buttons. Firstly, this kind of user interfaces are difficult to use for small size of the devices and correspondingly the keys or virtual keys. On the other hand, the use becomes more difficult as the devices are continuously provided with new applications and services which are to be used with the same small quantity of small-sized keys. Characteristric of these user interfaces, like user interfaces of computers and other data-processing devices, is generally also that they have their own artificial logics and rules which are fully adopted and known only by a few, normally technically orientated users.
An object of the invention is to present such an operation method of the user interface of hand-held devices which to large extent removes many of the above mentioned problems. To reach this object, the operation method of the user interface of a hand-held device, like palm-top or pocket computer, mobile phone, communicator or like, according to the invention, in which operation method data on the display of the device is changed for finding and selecting a desired data, is characterized in that what is defined by claim 1. Other claims define variable embodiments of the invention.
An advantage of the invention is that a user is able to easily adopt the operating rules of the user interface because they are "natural". The use of the device is also ergonomically easier because there is no need to use small-sized keys or auxiliary devices, for example. The invention and some embodiments thereof are described in further detail in the following with reference to the accompanying drawings, in which:
Figs. 1 to 10 are schematic perspective views presenting examples of the operation method of the user interface of the invention;
Figs. 11 to 13 are schematic front views of a device and data presented on the display thereof and present some further examples of the operation method of the user interface of the invention; Fig. 14 is a schematic presentation of an arrangement of data and operation of an application of the user interface of the invention;
Fig. 15 is a schematic perspective view of operation of a further application of the user interface of the invention; and Fig. 16 is a flow chart presenting a possible realization of the operation method of the user interface according to the invention.
In Figs. 1 to 5, a user is holding a hand-held device 2 in his or her hand 3 and is using a device having a display 2'. The display 2' is a touch sensitive display, and the user is able to give a signal to the device by pressing the display with his or her thumb 4. Movements of the device 2 are detected by an acceleration measurement circuit installed in the device itself, for example, and the data obtained from the circuit is utilized in the operation of the user interface. Some possible ways of realizing the operation method of the user interface according to the invention are considered in further detail below. Figs. 1 to 5 present examples of the operation of the user interface. With reference to Fig. 1, pages 5a, 5b, 5c, 5d and 5e, which may be e.g. weather maps obtained as a result of a web search, are browsed on the display 2' in response to moving the device 2 essentially perpendicularly to the display 2' thereof, as is indicated by arrow F. For the user, it is easy to think that he or she has a pile of pictures 5 a to 5 e ahead of him or her and that by moving in the pile he or she is able to see a picture at which he or she is at each time. In the same way, successive (or piled) pages of a book or other document may be thought to be browsed. The speed of browsing may be made dependent on quickness or intensity of the movement, i.e. on the magnitude of acceleration in the movement. The operation may also be arranged in such a way that a suitable small movement forward or backward and immediate stop bring the next page in the corresponding direction on the display. Furthermore, it may be arranged that a push of the touch sensitive display 2' gives a signal to the display control means in response to which signal changing of the data on the display is stopped. This kind of virtual use of depth increases significantly the capacity of a display having a restricted size.
With reference to Fig. 2, an image on the display 2' is zoomed larger, image 6b, or smaller, image 6a, in response to moving the device 2 in the same way as above, essentially perpendicularly to the display 2' thereof, as is indicated by arrow Z. This operation method of the user interface is fully analogous to human action as he or she wants to examine details of an object, e.g. an image, more closely. Having found a desired magnification, changing of the data on the display may be stopped again by a push of the display 2', for example. The functions of Figs. 1 and 2 may be used alternatively in a user interface according to application, whereby there is no indistinction of what is happening in response to moving the device in this way. The function of Figs. 1 and 2 may also be combined in the operation of a user interface so that, for example, data is first browsed and then zoomed by means of the same movement. Then, it is advantageous to arrange a suitable selector for selecting one or the other of the functions. A selector of zooming may be a push of a certain area on the display, for example, whereby changing over from the zooming mode to the browsing mode may happen automatically by selection of a desired magnification ratio.
In Fig. 3, reference sign 7a indicates image data which is significantly larger than the display. The capacity of the display of a hand-held device is not large enough for large entities of factual or image data. In response to lateral movement of the device in essentially the plane of the display 2', as is indicated by arrows N, E, S and W, data 7b on the display is changed as if a window were moved above a larger image formed by the image data 7a. Having found a desired place, changing of data 7b on the display may again be stopped by a push of the display. It is logical to accomplish the functions of Figs. 2 and 3 together whereby the image data may be examined more thoroughly after zooming by selecting the data in the way described by Fig. 3. Moreover, e.g. number selection may be realized in this way without keys by moving the device above a large virtual keyboard and selecting the desired numbers one by one. This function makes it possible to read full-length web pages with a palm-top computer, for example.
With reference to Fig. 4, in response to a quick rotational movement of the device 2 as if around an axis formed essentially by the edge 9 thereof, i.e. in response to swinging the device like turning page of a book, as is indicated by arrow P, data 8a, 8b, 8c on the display 2' is changed correspondingly, e.g. a page of an electronic book is turned. By the direction of the swing, the direction of turning page is selected. The function may be arranged so that a small swing with quick stopping turns one page in the direction of the swing, and that in response to a more intensive swing, pages are browsed more quickly until the browsing is stopped by a push of the display.
In Fig. 5, an application is illustrated in which any part 9b of image data 9a foirning a panorama picture covering the whole sphere of perceptual space (or respectively of image data forming a cylindrical picture) may be examined at a time. In the operation method of a user interface according to the invention, in response to moving the device 2 as if moving a picture on a corresponding spherical surface or in response to changing orientation of the device, that part 9b of the image data 9a which corresponds to the orientation of the display is obtained on the display 2'. At the same time, stopping and zooming of the image may be applied in the way described above. It may be contemplated also that a panorama picture is rotated by giving a push with a movement of the device to a desired direction, whereby the panorama picture is correspondingly kept moving until moving is stopped by a push of the display, for example.
Figs. 6 to 10 present embodiments of the user interface of the invention in which data on the display is changed by tilting the device. In Fig. 6, an example is presented in which data, illustrated by data Dl, on the display 2' of the device 2 is scrolled in the way indicated by arrow Rl to the direction to which the device is tilted in the way indicated by arrow Tl . The initial position of the device is indicated by broken line and the tilted position by solid line. In other words, in response to tilting the device, the data on the display is rolling or running to the direction to which the device is tilted, which is fully analogous to cause- consequence realtionships of the real world. This may be realized in such a way, for example, that the changing velocity is the grater the more the device is being tilted or is tilted from the initial position. Rolling or running of the data may be stopped by returning the device to the initial position or by means of a key on the device, for example.
Fig. 7 presents an embodiment suitable for moving and searching objects on a map, for example. On the display 2,' a cursor C is placed which keeps its place thereon, and tilting the device in the way indicated by arrow T2 causes that data, illustrated by data D2, starts moving from the tilting direction towards to the cursor C in the way indicated by arrow M2. In other words, the cursor is moving on the data forming a map, for example, to the direction to which the device is tilted. In the example of Fig. 8, data on the display 2' of the device 2 is illustrated by data D3.
As the user is tilting the device 2 from the position indicated by broken line away from himself or herself, like going closer to the data on the display, the data on the display is enlarged, which is illustrated by a graph indicating enlarging of data D3 and arrow Z2. In the example of Fig. 9, data D4 on the display is correspondingly reduced, as is indicated by a graph illustrating the reduction and drawing away and by arrow Z3, by tilting the device towards the user in the way indicated by arrow T4.
The embodiments of Figs. 6, 7, 8 and 9 may be combined in the user interface of a device in such a way, for example, that there is a key on the device for selecting the operation mode (scrolling or zooming) whereby, first, an object is searched by tilting in the scrolling mode, for example, and having found the object, changing over to the zooming mode is made whereby tilting results in zooming the data on the display. Scrolling by tilting may, of course, be combined with zooming in the way described with reference to Fig. 2.
Fig. 10 presents schematically an embodiment of the user interface of the invention for selecting data objects on the display. Data objects D5 to D10 may be alternatives in a menu or selection buttons on a page, for example. In the initial position of the device 2 indicated by broken line, data object D5 selected which is indicated by hatching in the figure. Tilting the device in the way indicated by arrow T5, the selection is moved to the direction of tilting, as is indicated by arrow M3, and is here moved to data object D9. Having selected the desired data object, it may be locked by a key on the device or by a return movement of the device, for example. For the sake of simplicity, tilting the device to only one direction and selection from data objects located one below the other are presented here, but in the same way it is, of course, possible to move between selectable objects also laterally.
Fig. 11 presents a further example of the embodiment described with reference to Fig. 3. Wide data Dl 1 consists of objects which are described by letters arranged to form a matrix. By moving the device in the way indicated by arrows N, E, S and W one is as if moving above the data and looking at it through a window formed by the display 2'. As the desired data, here letter q, is found, it may be selected by means of a key on the device, for example. A comparable function may be realized also so that, instead of moving the device laterally, it is tilted to a direction to which one desires to move on the data.
Fig. 12 presents a further embodiment of the user interface of the invention for selecting an object on the display. On the display 2' of the device 2 there are objects 01 to 09 from which the selection is made. In the initial state at left, the cursor is close to the upper right corner of the display. In this embodiment, cursor C is as if anchored in place in the real world. As one desires to select object 07, the device 2 is moved laterally as if moving a picture under the cursor so that the desired object 07 is coming under the cursor C at right in the figure. There are keys 21, 22 and 23 on the device the operation of which may correspond to the operation of the buttons of a mouse. In fact, the embodiments of Figs. 10 to 12 are solutions which in hand-held devices replace solutions, like a mouse, in conventional computers which are needed for moving, moving a cursor or making selections on the display. Fig. 13 presents an embodiment in which data objects, pages, cards or like, Cl, ... , Ci,
... , Cn are arranged in a stack or one above the other on the display 2" of the device, like e.g. pages of a book in the real world. The data objects may be cards presenting articles of commerce or web pages, for example. The objects are browsed in the way indicated by arrow M4 forward and backword by tilting the device correspondingly either forward in the way indicated by arrow T6 or backward in the way indicated by arrow T7. By means of keys 24 to 26, objects may be selected and files, programs, etc. behind them be opened. There may be several modes in the user interface of a device, like browsing, scrolling, selecting and zooming described above, and e.g. key 24 may be for selection of mode and other keys 25, 26 and 27 may operate like the buttons of a conventional mouse. Fig. 14 presents an application of the user interface of the invention in which data objects are arranged in radial stacks SI 1 to S 18 to extend outwards from the circle of the centre in which a device 2 and a user are. The first page of each stack, e.g. PI 2 or PI 6, defines data therebehind. Front pages may be browsed by turning the device laterally in the way indicated by arrow T8, for example, so that direction A12 and front page P12 are selected. The stack of data objects behind it may be browsed either by moving or tilting the device 2 forward and backward in the way indicated by arrow FI . In this way. e.g. a user interface for electronic shopping may be arranged and make perceiving the places of the products and orientation in a virtual shop easier.
Fig. 15 presents a further developed application of the embodiment described with reference to Fig. 5. Circles Rl and R2 present now image data forming a panorama picture or a 3D picture, in general, covering the whole sphere of the perception space. The image data may be also image data obtained from a video camera connected to the device. Three- dimensional image data D 12 describing a thing, for example, may be brought to this image data and set in a desired place and desired position in "the real surroundings". In response to moving the device 2 in the way described with reference to Fig. 5, e.g. as indicated by arrow T9, both the background data and 3D data D12', D12" describing the thing are changed in accordance with the positions of the device to correspond to viewing directions Al and A2, so that the thing may be examined from different directions in "the real surroundings" thereof. This may be applied e.g. by searching the 3D model of a sofa from a web shop and by placing it in a desired place in an image data presenting one's house whereby it may be examined from different directions to see how the sofa looks like in the real surroundings thereof.
A solution according to the invention may be realized by providing a hand-held device by a multi-axial accelerometer, for example, and with suitable circuits and programs cooperating with the operating system of the device and possibly application programs for processing and interpreting measurement results so that a change of data on the display corresponding to a movement detected by the accelerometer is carried out. E.g. the realization with measurement of accelerations is based on application of as such known technical solutions, and a person skilled in the art, having provided with instructions and specifications, is able to realize the operation of the user interface according to the invention with reasonable efforts.
In small-sized data-processing devices, like palm-top computers, with prior art hardware technology, it is not possible to realize operation systems and applications with a capacity and usability which were even close to the level reached by larger data-processing devices, like desktop or laptop computers. It is also impossible to include conventional storage means, like hard disk, floppy disk or CD-ROM drivers, in hand-held devices. A solution of these problems may be a two-part data-processing device wherein in a hand-hald part there is only a part of necessary circuits and programs in addition to a display. The most of the circuits and programs are in another portable part held by the user, a wireless link connecting this part to the hand-held part. The wireless link may operate with IR or radio frequencies. In this kind of device, it is easy to realize at the same time a system which detects movement or position of the hand-held part in relation to the portable part. The operation method of the user interface according to the invention may then be realized by means of this system.
In the following, a further embodiment of the invention based on detecting accelerations is considered with reference to Fig. 16. The device includes a multi-axial accelerometer and necessary circuits and programs for measurements. The device being switched on, it is monitoring accelerations and in phase 11 is interpreting that an acceleration above a certain threshold is possibly an initial stage of a movement defined in the operation method of the user interface. In response to this, timers TD1 and TD2 are started in phase 12. TD1 sets a very short experimentally determined time of the order of milliseconds from the detected start of the movement, in which the actual direction and magnitude of the acceleration specify the movement which the user gives to the device. TD2 sets a longer time expiring of which stops changing data on the display if no other causes for stopping appeared. At the time of expiration of time TD1, the prevailing acceleration vector is measured in phase 13. In phase 14, it is examined if a detected vector is a defined vector, i.e. if it is corresponding to any movement defined in the operation method of the user interface. If not, the operation returns back to the beginning 10 and phase 11 to monitor accelerations of the device. On the other hand, if a vector is a defined vector, data on the display is changed in phase 15, e.g. is browsed, zoomed, etc., according to a corresponding algorithm. As is described above, e.g. determination of the browsing speed of data on the display on the basis of the magnitude of an acceleration vector may be related to this. In phase 16, it is also monitored at the same time if time TD2 is expired. If so. the operation is forwarded to phase 19 in which changing of data on the display is stopped and the procedure is finished in phase 20. If time TD2 is not expired, it is monitored if any opposite vector to the detected defined vector or any other stopping signal, like a push of the display, appear in phase 17. If an opposite vector or a stopping signal are detected in phase 18, changing data on the display is stopped in phase 19 and the procedure is finished. If no opposite vector or other stopping signals are detected in phase 18, the operation returns back to monitor above mentioned issues.
The operation method of the user interface of the invention may be realized also in other ways than by utilizing acceleration measurements. Any technique by which changes of place and position of the device may be measured may be applied. A solution which may be contemplated is technique in which a place and orientation of the device are detected by sending pulsed DC magnetic field and by measuring it with a detector in the device. With this technique, both a place and position of the device in three-dimensional space may be fully detected. A possible solution is also the technique on which so called optical mouse is based and in which a movement and speed thereof are detected by observing and analyzing any surface in the vicinity, which in this case could be a surface on the user, for example. By combining this with distance measurement, which is easy to realize, and by analyzing changes of distance and relative movement, changes of place, movement and position of a device may be detected.
The invention may, of course, be realized in various ways. It may be contemplated, for example, that for making detection of a defined movement more reliable, an acceleration is measured many times within a short time window, and a defined movement is detected if any one of the measured vectors meets the requirements for detection. Naturally, also a more complicated analysis of a movement based on successive measured acceleration vectors may be contemplaed. A movement of a device may be detected also with an arrangement, for example, in which distances are measured between transmitters and receivers placed in the device, on one hand, and on the user, on the other hand.
A user interface according to the invention may be realized according to use and necessary features of a hand-held device as a suitable combination of the embodiments presented above, the user interface including several operation modes in which the same movement of the device may correspond to different changes of data on the display. As is stated above in relation to some embodiments, there may be a key, for example, on the device for selecting different operation modes. The invention may be varied within the scope of the accompanying claims.

Claims

Claims
1. An operation method of the user interface of a hand-held device, like palm-top or pocket computer, mobile phone, communicator or like, in which operation method data on the display of the device is changed for finding and selecting a desired data- characterized in that: data (5a - 5e; 6a. 6b; 7b; 8a - 8c; 9b; Dl; D2; D3; D4; D5 - D10; Dl l; Cl - Cn; SI 1 - SI 8) on the display (2') of the device is changed in response to defined movements of the device (F; Z; N, E, S, W; P; Tl, T2, T3, T4, T5; T6. T7; T8, FI; Fig. 5) in the three-dimensional space of use thereof; and said movements of the device and the corresponding changes of data on the display are defined so that a movement and a corresponding change of data on the display have a cause-consequence relationship which is analogous to cause-consequence relationships of the three-dimensional space of perception and action of a user (1).
2. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to a movement (F; T6, T7) of the device perpendicularly to the display (2') thereof data objects (5a, 5b, 5c, 5d, 5e; SI, Cl, ..., Ci, ..., Cn) are browsed on the display dependent on the direction of the movement either forward or backward, respectively.
3. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to moving (F) the device forward or backward peφendicularly to the display (2') thereof data objects (5a, 5b, 5c, 5d, 5e) are browsed on the display dependent on the direction of the movement either forward or backward, respectively.
4. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to tilting (T6, T7) the device (2) data objects (SI, Cl, ..., Ci, ..., Cn) placed one above the other are browsed (M4) on the display (2') to a direction corresponding to the direction of tilting.
5. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to moving (Z) the device (2) perpendicularly to the display (2') thereof towards the front of the display data (6a) on the display is zoomed larger (6b).
6. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to moving (Z) the device (2) perpendicularly to the display (2') thereof towards the back of the display data on the display is zoomed smaller.
7. An operation method of the user interface of a hand-held device according to claim 1 , characterized in that in response to moving (N, E, S, W) the device (2) essentially in the direction of the plane of the display (2') thereof it is moved correspondingly (7b) in a displayed data (7a) which forms an image larger than the display.
8. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to moving (N, E, S, W) the device (2) essentially in the direction of the plane of the display (2') thereof data (q) from data (Dl l) larger than the display is selected.
9. An operation method of the user interface of a hand-held device according to claim 1 , characterized in that a cursor (C) is placed on the display (2') and in response to moving (N, E, S, W) the device (2) essentially in the direction of the plane of the display (2') thereof the cursor (C) is moved against the movement for selecting data (01 - 09) on the display as if by moving data on the display in the real world below the cursor keeping its place in the real world (Fig. 12).
10. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to swinging (P) the device (2) like turning page also data (8a, 8b, 8c) on the display (2') is changed in the way corresponding to turning page.
11. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to tilting (Tl) the device (2) data (Dl) on the display (2') is rolled (Ml) to the direction of tilting.
12. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to tilting (T2) the device (2) data (D2) is moved on the display (2') from the direction of tilting towards a cursor (C).
13. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to tilting (T3) the device (2) away from a user data (D3) on the display (2') is enlarged (Z2) and / or in response to tilting (T4) the device (2) towards a user data (D4) (2') is reduced (Z3).
14. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to tilting (T4) the device (2) selection of data (D5) on the display (2') is moved (M3, D9) into the direction of tilting.
15. An operation method of the user interface of a hand-held device according to claim 1, characterized in that data is arranged virtually in stacks (SI 1 - SI 8) on a circle surrounding a user (a device (2)) whereby selection of a stack (SI 2) is carried out by turning (T8) the device and the stack (SI 2) is browswd by moving or tilting (FI) the device into the direction of the stack (SI 2) and away from it, respectively.
16. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in an application (Fig. 5), in which displayed data is a desired part of a panorama picture (9a), in response to turning the device (2) displayed data (9b) is changed so that the change of the viewing direction and the displayed part of the panorama picture (9a) correspond to turning.
17. An operation method of the user interface of a hand-held device according to claim 1 , characterized in that in an application(Fig. 15), in which displayed data is a desired part of a panorama picture or 3D image data (Rl, R2) 3D, data (D12) describing a thing is placed in the image data and in response to turning the device (2) both the panorama picture data or 3D image data (Rl, R2) and 3D image data (D12', D12") are changed on the display (2') to correspond to the viewing direction (Al, A2) corresponding to the orientation of the device (2).
18. An operation method of the user interface of a hand-held device according to claim 1, characterized in that in response to stopping or significant slowing (18) of a movement changing of data on the display is stopped.
19. An operation method of the user interface of a hand-held device according to claim 1, characterized in that changing of data on the display is stopped by giving a signal to the device in other way than by means of movement of the device.
EP00967937A1999-10-122000-10-11Operation method of user interface of hand-held deviceWithdrawnEP1228422A1 (en)

Applications Claiming Priority (5)

Application NumberPriority DateFiling DateTitle
FI992191AFI19992191A7 (en)1999-10-121999-10-12 Handheld device user interface operating method
FI9921911999-10-12
FI20001506AFI20001506A7 (en)1999-10-122000-06-22 Handheld device operating method
FI200015062000-06-22
PCT/FI2000/000871WO2001027735A1 (en)1999-10-122000-10-11Operation method of user interface of hand-held device

Publications (1)

Publication NumberPublication Date
EP1228422A1true EP1228422A1 (en)2002-08-07

Family

ID=26160785

Family Applications (1)

Application NumberTitlePriority DateFiling Date
EP00967937AWithdrawnEP1228422A1 (en)1999-10-122000-10-11Operation method of user interface of hand-held device

Country Status (8)

CountryLink
EP (1)EP1228422A1 (en)
JP (1)JP2003511786A (en)
CN (1)CN1167995C (en)
AU (1)AU7792600A (en)
FI (1)FI20001506A7 (en)
NO (1)NO20021647L (en)
RU (1)RU2242043C2 (en)
WO (1)WO2001027735A1 (en)

Families Citing this family (91)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
JP2002268622A (en)*2001-03-092002-09-20Denso CorpUser interface device of portable terminal device
JP4530593B2 (en)*2001-09-202010-08-25シャープ株式会社 Image display system
FI115861B (en)2001-11-122005-07-29Myorigo Oy Method and apparatus for generating a response
FI20012209A7 (en)*2001-11-142003-06-24Nokia Corp Method for controlling the presentation of information in an electronic device and electronic device
US7714880B2 (en)*2001-11-162010-05-11Honeywell International Inc.Method and apparatus for displaying images on a display
FI20020582A7 (en)*2002-03-262003-09-27Nokia Oyj User interface for a portable communications device
FI20021037A7 (en)2002-05-312003-12-01Nokia Corp Calendar system and method for providing a calendar view
TW200407025A (en)*2002-08-272004-05-01Vitec Co LtdPocket terminal device
EP1588245A2 (en)*2003-01-152005-10-26Koninklijke Philips Electronics N.V.Handheld device with a display screen
US7369102B2 (en)*2003-03-042008-05-06Microsoft CorporationSystem and method for navigating a graphical user interface on a smaller display
US20120081315A1 (en)2010-10-012012-04-05Imerj LLCKeyboard spanning multiple screens
KR100415161B1 (en)*2003-07-012004-01-13(주)두모션Hand held device with three dimensional viewing function by using tilting sensor and system for three-dimensionally displaying image using the same
US7246912B2 (en)2003-10-032007-07-24Nokia CorporationElectroluminescent lighting system
WO2005041167A1 (en)*2003-10-282005-05-06Matsushita Electric Industrial Co., Ltd.Image display device and image display method
US20050212760A1 (en)2004-03-232005-09-29Marvit David LGesture based user interface supporting preexisting symbols
JP2005292893A (en)*2004-03-312005-10-20Nec Access Technica LtdPortable information terminal device
CN1704871B (en)*2004-05-312010-04-28深圳市朗科科技股份有限公司Portable digital devices and input method therefor
EP1759285A1 (en)*2004-06-042007-03-07Philips Intellectual Property & Standards GmbHA hand-held device for content navigation by a user
JP2006079312A (en)*2004-09-092006-03-23Matsushita Electric Ind Co Ltd Portable viewer
GB0422090D0 (en)*2004-10-052004-11-03Symbian Software LtdAn interactive computing device with a configurable user interface
JP2006113859A (en)*2004-10-152006-04-27Nec CorpPortable information terminal and display control method therefor
CN100361051C (en)*2004-12-302008-01-09集嘉通讯股份有限公司Motion analysis device and method for mobile product
KR101002807B1 (en)*2005-02-232010-12-21삼성전자주식회사 Apparatus and method for controlling menu navigation in a terminal displaying a menu screen
KR101025259B1 (en)*2005-05-232011-03-29노키아 코포레이션 Enhanced Pocket Computer and Associated Methods
KR100811160B1 (en)2005-06-022008-03-07삼성전자주식회사Electronic device for inputting command 3-dimensionally
JPWO2007007682A1 (en)2005-07-082009-01-29三菱電機株式会社 Touch panel display device and portable device
US7694234B2 (en)*2005-08-042010-04-06Microsoft CorporationVirtual magnifying glass with on-the fly control functionalities
KR100651368B1 (en)*2005-09-152006-11-29삼성전자주식회사 Image control method according to the movement of mobile terminal
US7431216B2 (en)*2005-11-162008-10-07Sony Ericsson Mobile Communications AbMethods for presenting parameter status information and related portable electronic devices and parameters
JP2009517761A (en)*2005-12-012009-04-30コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Display to assist users in predicting scroll rate changes
CN100429610C (en)*2006-01-192008-10-29宏达国际电子股份有限公司Intuition type fluorescent screen controller
US20070174416A1 (en)*2006-01-202007-07-26France TelecomSpatially articulable interface and associated method of controlling an application framework
KR100877829B1 (en)2006-03-212009-01-12엘지전자 주식회사 A terminal having a scrolling function and a scrolling method thereof
KR20090077755A (en)*2006-09-092009-07-15에프-오리진, 인크. Integrated Pressure Sensing Lens Assembly
JP5158902B2 (en)*2006-09-142013-03-06シャープ株式会社 Electronic device, effective function selection method and program
US7889173B2 (en)*2006-09-142011-02-15Microsoft CorporationDefining user input fields on a portable media device
JP4801623B2 (en)*2006-09-142011-10-26シャープ株式会社 Electronic device and method for selecting effective functions
KR100894146B1 (en)*2007-02-032009-04-22엘지전자 주식회사 Mobile communication terminal and operation control method
CN101330811B (en)*2007-06-222010-12-08鸿富锦精密工业(深圳)有限公司Portable electronic device and operation method thereof
JP5088017B2 (en)*2007-06-282012-12-05ソニー株式会社 Image display apparatus, imaging apparatus, image display method, and program
JP5412812B2 (en)*2007-12-072014-02-12ソニー株式会社 Input device, control device, control system, and handheld device
DE102007059273A1 (en)*2007-12-082009-06-18T-Mobile Internationale Ag Virtual keyboard of a mobile device
US8217964B2 (en)2008-02-142012-07-10Nokia CorporationInformation presentation based on display screen orientation
JP2009265757A (en)*2008-04-222009-11-12Toshiba CorpFoldable portable terminal
WO2009141497A1 (en)*2008-05-222009-11-26Nokia CorporationDevice and method for displaying and updating graphical objects according to movement of a device
JP2010092086A (en)*2008-10-032010-04-22Just Syst CorpUser input apparatus, digital camera, input control method, and input control program
FI20080591A7 (en)*2008-10-242010-07-27Teknillinen Korkeakoulu Method and device for controlling an application
US8645871B2 (en)2008-11-212014-02-04Microsoft CorporationTiltable user interface
US20100146460A1 (en)*2008-12-102010-06-10Sony Ericsson Mobile Communications AbSystem and method for modifying a plurality of key input regions based on detected tilt and/or rate of tilt of an electronic device
JP5618486B2 (en)*2009-01-302014-11-05株式会社東芝 Portable information terminal
JP5357800B2 (en)*2009-02-122013-12-04キヤノン株式会社 Electronic device and control method thereof
JP5223784B2 (en)*2009-06-052013-06-26船井電機株式会社 Mobile terminal device
CN101943988B (en)*2009-07-092013-04-24深圳富泰宏精密工业有限公司System and method for automatically adjusting user interface of electronic device
US9383916B2 (en)2009-09-302016-07-05Microsoft Technology Licensing, LlcDynamic image presentation
JP2011097441A (en)*2009-10-302011-05-12Sony CorpInformation processing apparatus, image display method and computer program
WO2011073557A1 (en)*2009-12-182011-06-23France TelecomMethod for restoring information on a screen of a terminal, and corresponding device, terminal, and computer program
US20110161889A1 (en)*2009-12-302011-06-30Motorola, Inc.User Interface for Electronic Devices
CN102129337A (en)*2010-01-192011-07-20腾讯科技(北京)有限公司Method and device for controlling mobile terminal browser
US9977472B2 (en)*2010-03-192018-05-22Nokia Technologies OyMethod and apparatus for displaying relative motion of objects on graphical user interface
KR101680113B1 (en)*2010-04-222016-11-29삼성전자 주식회사Method and apparatus for providing graphic user interface in mobile terminal
JP2011233064A (en)*2010-04-302011-11-17Sony CorpInformation processor and display screen operation method
US9046992B2 (en)2010-10-012015-06-02Z124Gesture controls for multi-screen user interface
JP2012123451A (en)*2010-12-062012-06-28Sony CorpInformation processor, information processing system and information processing method
KR101740439B1 (en)*2010-12-232017-05-26엘지전자 주식회사Mobile terminal and method for controlling thereof
KR101830962B1 (en)2010-12-292018-02-22삼성전자주식회사Apparatus and method for controlling screen display in portable terminal
KR101766332B1 (en)*2011-01-272017-08-08삼성전자주식회사3d mobile apparatus displaying a plurality of contents layers and display method thereof
KR101864333B1 (en)*2011-03-212018-07-05삼성전자 주식회사Supporting Method For Icon Change Function And Portable Device thereof
JP5684621B2 (en)2011-03-282015-03-18京セラ株式会社 Electronic device, display control method, and display control program
US9182935B2 (en)2011-09-272015-11-10Z124Secondary single screen mode activation through menu option
WO2013048288A2 (en)*2011-09-302013-04-04Miroshnichenko Vladimir VitalievichTouch-sensitive panel
KR101969931B1 (en)*2012-01-102019-04-17삼성전자주식회사Apparatus and method for controlling rotation of display image
JP6019601B2 (en)2012-02-102016-11-02ソニー株式会社 Information processing apparatus, information processing method, and program
JP2014029522A (en)2012-07-062014-02-13Funai Electric Co LtdElectronic information terminal and display method therefor
US9791896B2 (en)*2012-07-132017-10-17Symbol Technologies, LlcDevice and method for performing a functionality
JP5435110B2 (en)*2012-11-292014-03-05日本電気株式会社 Terminal device, display method, and program
CN103179273A (en)*2013-03-112013-06-26广东欧珀移动通信有限公司 A method of operating a mobile phone
JP5516794B2 (en)*2013-05-132014-06-11日本電気株式会社 Portable information terminal, display control method and program
KR102131358B1 (en)*2013-06-172020-07-07삼성전자주식회사User interface device and method of operation of user interface device
CN104407785B (en)*2013-12-102018-06-15贵阳朗玛信息技术股份有限公司A kind of information displaying method and device
JP5639295B2 (en)*2014-03-312014-12-10グリー株式会社 User operation control program, portable device, and user operation control method
JP5640165B2 (en)*2014-03-312014-12-10グリー株式会社 User operation control program, portable device, and user operation control method
CN105288997B (en)*2014-06-242019-08-06腾讯科技(深圳)有限公司Interactive method and apparatus are realized in chessboard interface
CN104216634A (en)*2014-08-272014-12-17小米科技有限责任公司Method and device for displaying manuscript
CN104394312B (en)2014-10-232017-08-22小米科技有限责任公司Filming control method and device
JP5801005B2 (en)*2015-01-142015-10-28京セラ株式会社 Electronic device, display control method, and display control program
JP5801006B2 (en)*2015-01-142015-10-28京セラ株式会社 Electronic device, display control method, and display control program
JP5826415B2 (en)*2015-01-142015-12-02京セラ株式会社 Display control method and display control program
JP2015099602A (en)*2015-01-142015-05-28京セラ株式会社 Display control program
RU2608148C1 (en)*2015-08-202017-01-16Общество с ограниченной ответственностью "1С ВИАРАБЛ" (ООО "1С ВИАРАБЛ")Method, device and system for data input and display on touch screen display
US9697393B2 (en)2015-11-202017-07-04Symbol Technologies, LlcMethods and systems for adjusting mobile-device operating parameters based on housing-support type
CN105807952B (en)*2016-03-072020-01-31联想(北京)有限公司information processing method and electronic equipment

Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
EP0825514A2 (en)*1996-08-051998-02-25Sony CorporationInformation processing device and method for inputting information by operating the overall device with a hand
EP0899650A2 (en)*1997-08-291999-03-03Xerox CorporationA computer user interface using a manipulatory grammar

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
IL92220A (en)*1989-11-061993-02-21Ibm IsraelThree-dimensional computer input device
RU2099775C1 (en)*1996-03-261997-12-20Виктор Алексеевич КузинDevice for information input to controlled object
RU2096821C1 (en)*1996-07-191997-11-20Акционерное общество закрытого типа "Ракурс-3Д"Device for information input into controlled object
SE516552C2 (en)*1997-10-022002-01-29Ericsson Telefon Ab L M Handheld display unit and method for displaying screens

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
EP0825514A2 (en)*1996-08-051998-02-25Sony CorporationInformation processing device and method for inputting information by operating the overall device with a hand
EP0899650A2 (en)*1997-08-291999-03-03Xerox CorporationA computer user interface using a manipulatory grammar

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
FITZMAURICE G.W.: "SITUATED INFORMATION SPACES AND SPATIALLY AWARE PALMTOP", COMMUNICATIONS OF THE ASSOCIATION FOR COMPUTING MACHINERY, ACM, NEW YORK, NY, US, vol. 36, no. 7, 1 July 1993 (1993-07-01), pages 39 - 49, XP000384279, ISSN: 0001-0782*
See also references ofWO0127735A1*

Also Published As

Publication numberPublication date
FI20001506A0 (en)2000-06-22
FI20001506L (en)2001-04-13
FI20001506A7 (en)2001-04-13
JP2003511786A (en)2003-03-25
NO20021647L (en)2002-05-22
NO20021647D0 (en)2002-04-08
WO2001027735A1 (en)2001-04-19
CN1379871A (en)2002-11-13
AU7792600A (en)2001-04-23
CN1167995C (en)2004-09-22
RU2242043C2 (en)2004-12-10

Similar Documents

PublicationPublication DateTitle
EP1228422A1 (en)Operation method of user interface of hand-held device
US9304583B2 (en)Movement recognition as input mechanism
US9880640B2 (en)Multi-dimensional interface
US6396497B1 (en)Computer user interface with head motion input
US7271795B2 (en)Intuitive mobile device interface to virtual spaces
US6288704B1 (en)Motion detection and tracking system to control navigation and display of object viewers
US6115025A (en)System for maintaining orientation of a user interface as a display changes orientation
RU2288512C2 (en)Method and system for viewing information on display
US6567101B1 (en)System and method utilizing motion input for manipulating a display of data
US20060279542A1 (en)Cellular phones and mobile devices with motion driven control
US20020024506A1 (en)Motion detection and tracking system to control navigation and display of object viewers
US20100171691A1 (en)Viewing images with tilt control on a hand-held device
US20140201675A1 (en)Method and mobile device for providing recommended items based on context awareness
US20060061551A1 (en)Motion detection and tracking system to control navigation and display of portable displays including on-chip gesture detection
US20060164382A1 (en)Image manipulation in response to a movement of a display
JP2012514786A (en) User interface for mobile devices
CN1367892A (en) information processing device
CN113253908A (en)Key function execution method, device, equipment and storage medium
Sasaki et al.Hit-wear: A menu system superimposing on a human hand for wearable computers
JP2000020199A (en)Portable information terminal and information collecting system using this
EP1028366A2 (en)Motion driven access to object viewers
HK40051670A (en)Key function executing method, device, equipment and storage medium
CN114138143A (en)Query interface display method and device, terminal and storage medium
Shibuya et al.An intuitive information space navigation method based on the window metaphor

Legal Events

DateCodeTitleDescription
PUAIPublic reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text:ORIGINAL CODE: 0009012

17PRequest for examination filed

Effective date:20020508

AKDesignated contracting states

Kind code of ref document:A1

Designated state(s):AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

AXRequest for extension of the european patent

Free format text:AL;LT;LV;MK;RO;SI

17QFirst examination report despatched

Effective date:20080411

STAAInformation on the status of an ep patent application or granted ep patent

Free format text:STATUS: THE APPLICATION HAS BEEN WITHDRAWN

18WApplication withdrawn

Effective date:20100721


[8]ページ先頭

©2009-2025 Movatter.jp