US20110126097A1 - Information processing apparatus, storage medium having program recorded thereon, and object movement method - Google Patents

Information processing apparatus, storage medium having program recorded thereon, and object movement method Download PDF

Info

Publication number
US20110126097A1
US20110126097A1 US13/054,698 US200913054698A US2011126097A1 US 20110126097 A1 US20110126097 A1 US 20110126097A1 US 200913054698 A US200913054698 A US 200913054698A US 2011126097 A1 US2011126097 A1 US 2011126097A1
Authority
US
United States
Prior art keywords
display
coordinate
contact
contact point
coordinate area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/054,698
Inventor
Kazuki Isono
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NEC Corp
Original Assignee
NEC Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NEC Corp filed Critical NEC Corp
Assigned to NEC CORPORATION reassignment NEC CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ISONO, KAZUKI
Publication of US20110126097A1 publication Critical patent/US20110126097A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0485Scrolling or panning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04845Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04808Several contacts: gestures triggering a specific function, e.g. scrolling, zooming, right-click, when the user establishes several contacts with the surface simultaneously; e.g. using several fingers or a combination of fingers and pen

Definitions

  • the present invention relates to an information processing apparatus having a contact detection means capable of detecting a plurality of contact points, a storage medium on which a program used in the information processing apparatus is recorded, and an object movement method.
  • a contact detection unit called a touch panel or a touch screen is provided on various information processing apparatuses these days. Furthermore, a contact detection means capable of detecting a plurality of contact points has been developed instead of a conventional contact detection means capable of detecting only one contact point.
  • Japanese laid-open patent publications Nos. 2002-304256 and 2007-279638 disclose a touch panel display or a touch panel device capable of detecting a plurality of contact points.
  • An operation method of an information processing apparatus having a contact detection means is designed so as to allow a user to intuitively operate the information processing apparatus. While such an operation method provides a user with intuitive operation, an excessive burden is imposed on the user if the operation (command to a controller, transition of a screen, or the like) is not performed as expected by the user' intuitive image.
  • an information processing apparatus that scrolls displayed contents by detection of contact with a touch panel (determination by touch) or by detection of removal of contact with the touch panel (determination by touch and release) provides a user with sufficient operability on a scroll operation.
  • a scroll operation of only a desired part of displayed information or an operation using a plurality of contact points
  • a (slight) difference is likely to be produced between the setting of the information processing apparatus and the user's intuitive operation. In such a case, the user will repeat intuitive operations.
  • the information processing apparatus does not operate if a scroll operation is set to be activated through the determination by touch and release. In such a case, the user repeats the same operation because no desired operation is performed. Thus, an excessive burden is imposed on the user.
  • the present invention has been made in view of the above drawbacks.
  • the present invention provides an information processing apparatus that can properly detect a user's intuitive operation with use of a contact detection unit capable of detecting a plurality of contact points and can properly move an object.
  • An information processing apparatus includes a display unit operable to display an object, a contact detection unit operable to detect a plurality of contact points on a display screen of the display unit, and a control unit operable to control the object displayed on the display unit so as to be movable with use of the contact detection unit.
  • the control unit detects a first contact point in an active range of the display screen other than a display range of the object with use of the contact detection unit, detects a second contact point in the display range of the object, and further detects a movement instruction to the object, then the control unit controls the object so as to be movable within the active range.
  • an information processing apparatus that can properly detect a user's intuitive operation with use of a contact detection unit capable of detecting a plurality of contact points and can properly move an object.
  • FIG. 1 is a functional block diagram showing an information processing apparatus according to a first embodiment of the present invention.
  • FIG. 2 is a flow chart explanatory of an identification operation of the information processing apparatus of FIG. 1 .
  • FIG. 3 is a functional block diagram showing a mobile terminal as an information processing apparatus according to a second embodiment of the present invention.
  • FIG. 4 is a flow chart explanatory of a scroll operation of the mobile terminal of FIG. 3 .
  • FIGS. 5A and 5B are diagrams showing a transition example of a display screen in the mobile terminal of FIG. 3 in a case of a single touch.
  • FIGS. 6A and 6B are diagrams showing a first transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIGS. 7A and 7B are diagrams showing a second transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIGS. 8A and 8B are diagrams showing a third transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIGS. 9A and 9B are diagrams showing a fourth transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIGS. 10A and 10B are diagrams showing a fifth transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIGS. 11A and 11B are diagrams showing a sixth transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIG. 12 is a partial enlarged view of the display screen shown in FIG. 11 .
  • FIG. 13 is a functional block diagram showing a tablet computer as an information processing apparatus according to a third embodiment of the present invention.
  • FIG. 14 is a flow chart explanatory of a movement operation of the tablet computer of FIG. 13 .
  • FIGS. 15A and 15B are diagrams showing a transition example of a display screen of a spreadsheet program in the tablet computer of FIG. 13 in a case of multiple touches.
  • an information processing apparatus is operated by a program control and is formed as a mobile terminal having a communication function (telephone conversation, mail transmission and reception, and Internet connection), a calendar function, a scheduler function, and the like.
  • a communication function telephone conversation, mail transmission and reception, and Internet connection
  • a calendar function calendar function
  • a scheduler function scheduler function
  • Other examples of the information processing apparatus include a personal computer, a smart phone, a PDA (Personal Digital Assistants), and a PHS (Personal Handyphone System) having a touch screen function.
  • FIG. 1 is a functional block diagram showing an overview of a configuration of an information processing apparatus 100 according to the first embodiment.
  • the information processing apparatus 100 shown in FIG. 1 includes a control unit 110 operable to perform a variety of calculations, a display unit 120 operable to display information directed from the control unit 110 , a contact detection unit 121 operable to detect the presence of a contact point in a coordinate area of the display unit 110 in which the display information is displayed, a coordinate of the contact point (contact coordinate or detection coordinate), and a change (movement) of the detection coordinate, and a storage unit 130 formed by a ROM, a RAM, a HDD, a flash memory, or the like.
  • the storage unit 130 stores therein an operating system, a driver or an application program for the contact detection unit 121 , various types of setting information of the information processing apparatus 100 , and the like.
  • the control unit 110 is operable to control the display unit 120 so as to display first display information in a first display range. Furthermore, the control unit 110 is operable to display second display information in a second display range that is part of the first display range.
  • display range refers to an area (outer frame) in which display information can be displayed.
  • coordinate area refers to an area in which display information has actually been displayed. While the second display range is included in the first display range, a coordinate area in which the second display information has been displayed is not included in a coordinate area in which the first display information has been displayed.
  • control unit 110 can detect, via the contact detection unit 121 , the presence of a contact point in a coordinate area of the display unit 120 in which the first and second display information has been displayed.
  • the control unit 110 can also identify a change (movement) of the detected coordinate (detection coordinate).
  • the control unit 110 can detect a plurality of contact points in a coordinate area in which the first display information is displayed and in a coordinate area in which the second display information is displayed.
  • control unit 110 performs a process of moving (scrolling) only the display information (object) displayed in a coordinate area in which the change of the detection coordinate of the contact point has been detected based upon the change of the detection coordinate.
  • the first display information which is displayed in the first display range, which is the entire display screen or a predetermined area (active range), and the second display information, which is displayed in part of that display range, are displayed on the display unit 120 by the control unit 110 .
  • the second display information may include an image, a figure, a character, or a symbol, or a combination thereof. Furthermore, the second display information may be expressed by any one of an icon, a button, a character string, a text box, and area information visually indicated for providing a command to the control unit 110 , or a combination thereof. Moreover, the second display information including a character string also includes information for managing a selected character string as a bundle after the character string has been selected by range specification.
  • the information processing apparatus 100 can properly detect a user's intuitive operation with use of the contact detection unit 121 capable of detecting a plurality of contact points and can properly move an object.
  • the term “contact point” includes not only a relatively small “point” as in a case where a stylus pen or the like is used, but also a relatively large “area” as in a case where a human finger or the like is used.
  • the control unit 110 of the information processing apparatus 100 produces images of the first display information to be displayed in the first display range and the second display information to be displayed in the first display range of the screen of the display unit 120 and then displays them on the display unit 120 .
  • the control unit 110 detects contact with the first coordinate area in which the first display information has been displayed and the second coordinate area in which the second display information has been displayed (appearance of a contact point), removal of contact (disappearance of the contact point), detection coordinates, and changes of the detection coordinates with use of the contact detection unit 121 .
  • the control unit 110 recognizes the detected contact, removal of contact, and changes of the detection coordinates as determination by touch, determination by touch and release, or a movement instruction as needed and determines an operational instruction to the displayed object or the like.
  • control unit 110 that identifies a movement instruction of the object from changes of detection coordinates of multiple contact points detected with the contact detection unit 121 .
  • the control unit 110 detects a plurality of contact points with the contact detection unit 121 and identifies a user's movement instruction based upon the position coordinates (detection coordinates) of the contact points and the changes of those position coordinates.
  • a movement instruction can be distinguished from other instructions by vectorizing detection coordinates of individual contact points so as to identify movement directions.
  • the movement instruction may be distinguished from other instructions by obtaining a period of time from the detection of contact to the detection of removal of contact with respect to each contact point and relying upon differences of those periods.
  • the operation can be determined to be magnification or reduction of the display screen.
  • the operation can be determined to be a movement instruction.
  • operations may be differentiated by using the size of a contact area.
  • operations may be differentiated by using a difference in size of a contact area between a stylus pen and a finger.
  • operations may be differentiated by using a difference in size of a contact area between a thumb and other fingers.
  • the contact detection unit 121 may be set such that an operation of only an underlay of the display screen or only a fixed operation is allowed around an area at which a finger to hold the information processing apparatus 100 is located.
  • a contact point having a large contact area or a small contact area may be determined to be invalid.
  • operations may be differentiated depending upon whether the information processing apparatus 100 is held by a single hand or by both hands.
  • the program allows the control unit 110 of the information processing apparatus 100 to perform a process of moving only one object while fixing the underlay or the background or perform a process of moving the entire active display screen based upon the positional relationship with contact with other coordinate areas, a duration of contact, and the like.
  • a user's intuitive operation can properly be detected.
  • FIG. 2 is a flow chart explanatory of an identification operation of the information processing apparatus 100 .
  • the identification process of the operation shown in FIG. 2 is to provide a function of moving an object, which is provided as a user interface to a user when the aforementioned information processing apparatus 100 operates.
  • the control unit 110 of the information processing apparatus 100 monitors contact with the display screen of the display unit 120 via the contact detection unit 121 (Step S 201 ).
  • Step S 202 the control unit 110 judges whether a contact point has been detected in only one of the first and second coordinate areas in which the first and second display information has been displayed, or whether a plurality of contact points have been displayed in both of the first and second coordinate areas. Furthermore, in addition to detection of a contact point in the display range of the object, the control unit 110 judges whether a contact point has been detected in an active range of the display screen other than the display range of the object.
  • Detections of contact points in both of the first coordinate area and the second coordinate area are judged to be the same input operation whichever is first detected. Nevertheless, concurrent detections of contact points in the first coordinate area and the second coordinate area may be judged to be different input operations.
  • the control unit 110 monitors detection coordinates of those detected contact points.
  • the control unit 110 judges whether only one of the contact points has moved (a movement (scroll) instruction has been provided).
  • the process proceeds to the next Step S 204 if only one contact point has moved, and the process proceeds to the next Step S 206 if the contact points have moved concurrently (Step S 203 ).
  • the control unit 110 judges whether a change of a detection coordinate of only one of the contact points has been detected.
  • the control unit 110 detects a contact point in an active range of the display screen other than the display range of the object to be moved and judges whether a movement instruction to the object has been detected.
  • control unit 110 If another operation (instruction) that is different from the movement instruction is made, the control unit 110 performs a process corresponding to that operation.
  • the control unit 110 fixes, on the display screen, the display information displayed in the coordinate area in which the other contact point is present, and moves the display information displayed in the coordinate area in which movement of the contact point has been detected (Step S 204 ).
  • the display information displayed in the coordinate area in which a change of the detection coordinate of the contact point has been detected is moved based upon the change of the detection coordinate.
  • an object is controlled so as to be movable within the active range, and an object to which a movement instruction has been provided is moved.
  • control unit 110 monitors the detection coordinate of the detected contact point and judges whether a movement instruction has been provided by a user (Step S 205 ).
  • the control unit 110 moves (scrolls) the entire active display information (Step S 206 ). In other words, if changes of the detection coordinates of all of the contact points have been detected, then the control unit 110 moves the first and second display information based upon the changes of the detection coordinates.
  • the information processing apparatus 100 can properly detect a user's intuitive operation with use of a contact detection unit capable of detecting a plurality of contact points and can thus move display information (object).
  • a mobile terminal 200 will be described as an information processing apparatus according to a second embodiment of the present invention.
  • FIG. 3 is a functional block diagram showing an overview of a configuration of the mobile terminal 200 according to the present embodiment.
  • the mobile terminal 200 includes a radio unit 201 , a broadcast receiver unit 202 , a GPS receiver unit 203 , a camera unit 204 , an acoustical treatment unit 205 , a display unit 206 having a contact detection function, a storage unit 207 , and a control unit 208 .
  • the radio unit 201 is operable to wirelessly transmit information to and receive information from a radio base station apparatus via an antenna.
  • the broadcast receiver unit 202 is operable to receive broadcast signals transmitted from a broadcast station (a ground station, a satellite, or the like) and conducts signal processing of visual data, audio data, information data, and the like, which have been obtained by demodulating the received broadcast signals.
  • the GPS receiver unit 203 is operable to measure periods of time for radio waves emitted from a plurality of GPS satellites to reach the mobile terminal 200 , compute distances between the respective GPS satellites and the mobile terminal 200 , and compute a position coordinate with use of the computed distances.
  • the camera unit 204 is operable to acquire image information (take a photograph).
  • the acoustical treatment unit 205 is operable to process acoustic signals of music, notification sounds, or voice inputted and outputted via a microphone or a speaker.
  • the display unit 206 having a contact detection function is a touch screen (touch panel) having a display function of outputting display information such as images, figures, characters, and symbols (video signal processing) and a switch function of sensing a user's input operation by using pressure, capacitance, or the like (position coordinate detection function and contact detection function).
  • the storage unit 207 stores therein telephone directory information, sent and received mail information, outgoing and incoming call information, content information, application programs, photographed images, various types of setting information of the mobile terminal 200 , and the like.
  • the control unit 208 is operable to control a telephone function, a mail function, an Internet (web) connection function, a camera function, a television receiving function, a GPS (positioning) function, a content playback function, other functions of the terminal, and respective components of the terminal. Furthermore, the control unit 208 is operable to control a display function of the display unit 206 having a contact detection function and display an electronic screen on the display unit 206 . The control unit 208 is operable to detect a user's operational instruction by using the switch function (position coordinate detection function and contact detection function) of the display unit 206 having a contact detection function.
  • FIG. 4 is a flow chart explanatory of a scroll operation of the mobile terminal 200 .
  • the control unit 208 of the mobile terminal 200 monitors contact with the display unit 206 having a contact detection function (Step S 401 ). If any contact is detected, the control unit 208 judges whether the number of contact points is one or more. The process proceeds to the next Step S 403 if a plurality of contact points have been detected, and the process proceeds to Step S 405 if one contact point has been detected (Step S 402 ).
  • control unit 208 judges whether at least one of those contact points is located at a predetermined position (or within an area (movement-fixing coordinate area)).
  • the predetermined position will be discussed later.
  • Step S 404 the process proceeds to the next Step S 404 if any contact has been detected at the preset position, and the process proceeds to Step S 405 if no contact has been detected at the preset position (Step S 403 ).
  • control unit 208 If the control unit 208 detects any contact at the preset position and also detects a movement instruction to an object (corresponding to the second display information), the control unit 208 fixes the background display (corresponding to the first display information) and controls the object so as to be movable within the active range (Step S 404 ).
  • control unit 208 scrolls the entire display based upon the amount of movement of the contact point until the contact is removed (the contact point disappears) (Steps S 205 -S 206 ).
  • the preset position (or area) will be described.
  • the preset position is set within a range of the display screen other than an object to be moved, or within an active range of the display screen other than an object to be moved, or at a position (designated position) that can be contacted (pressed) by a user's hand to hold the mobile terminal 200 .
  • the preset position is set within the active range of the display screen other than the display range of the object, or within the first coordinate area in which the first display information has been displayed, other than the second coordinate area in which the second display information has been displayed and in which the contact point has been detected, or within a designated coordinate area designated by an application program, or within the display range of a movement-fixed object.
  • the position that can be contacted (pressed) is a position that can be reached by a thumb or another specified finger when the mobile terminal 200 is held in a recommended standard manner.
  • the preset position will be exemplified in FIG. 7A and figures following FIG. 7A .
  • FIGS. 5A to 11B are diagrams showing transition examples of the display screen of the mobile terminal 200 .
  • FIGS. 5A to 11B show the entire display screen displayed on the display unit 206 having a contact detection function by the control unit 208 .
  • FIGS. 5A , 6 A, 7 A, 8 A, 9 A, 10 A, and 11 A show the display screen prior to movement (scroll) of an object.
  • FIGS. 5B , 6 B, 7 B, 8 B, 9 B, 10 B, and 11 B show the display screen after movement of an object.
  • FIGS. 5A and 5B are diagrams showing a transition example of the display screen in a case of a single touch.
  • first display information (schedule chart) 501 is displayed in a display screen frame 500 , and second display information (object) 502 is displayed.
  • second display information (object) 502 is displayed.
  • the control unit 208 does not detect any contact point in a first coordinate area in which the first display information 501 has been displayed, but detects a contact point in the second coordinate area in which the second display information 502 has been displayed.
  • the control unit 208 detects a change of a detection coordinate of the contact point as a scroll instruction. That is, FIG. 5A shows an operation from Step S 401 via “No” of Step S 402 to Step S 405 shown in FIG. 4 .
  • FIG. 5B shows a display result in which the control unit 208 has scrolled the first and second display information 501 and 502 , i.e., the entire display information, based upon the change of the detection coordinate as a result of the identification process of FIG. 5A . That is, FIG. 5B shows a state after the operation of Step S 406 shown in FIG. 4 .
  • the schedule chart (first display information 501 ) and the object are scrolled at the same time.
  • FIGS. 6A and 6B are diagrams showing a transition example of the display screen in a case of multiple touches.
  • FIGS. 6A and 6B show an example of screen transition in which the object 502 is controlled so as to be movable in an active range 503 when a contact point is detected in the active range 503 (hatched area) of the display screen other than the display range of the object 502 .
  • FIG. 6A shows a case of detecting a plurality of contact points in a first coordinate area in which the first display information 501 has been displayed and a second coordinate area in which the second display information 502 has been displayed and detecting, as a scroll instruction, a change of a detection coordinate of only the contact point in the second coordinate area, which is one of the contact points. That is, FIG. 6A shows an operation from Step S 401 via “Yes” of Step S 402 to “Yes” of Step S 403 shown in FIG. 4 .
  • FIG. 6B shows a display result in which the control unit 208 has moved the second display information (object) 502 , which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 6A . That is, FIG. 6B shows the operation of Step S 404 shown in FIG. 4 .
  • the object when a contact point is detected in the active range 503 of the display screen other than the display range of the object, and a detection coordinate of that contact point does not change, then the object is controlled so as to be movable within the active range 503 . Accordingly, the object can be moved while the schedule chart is fixed.
  • the control unit 208 may scroll (move) only the first display information 501 based upon the change of the detection coordinate.
  • the contact points are not necessarily required to be present in both of the first coordinate area and the second coordinate area.
  • FIGS. 7A and 7B are diagrams showing another transition example of the display screen in a case of multiple touches.
  • FIGS. 7A and 7B show an example of screen transition in which an object of “Meeting” is controlled so as to be movable with the schedule chart being fixed when a contact is detected at a preset position (designated position, designated coordinate range, movement-fixing coordinate area).
  • FIG. 7A shows a case of detecting a contact point at a preset position (movement-fixing coordinate area) 504 provided in the first coordinate area in which the schedule chart has been displayed, and a contact point in the second coordinate area in which the object has been displayed, and further detecting a change of the detection coordinate of the contact point in the second coordinate area as a movement instruction.
  • a preset position movement-fixing coordinate area
  • FIG. 7B shows a display result in which the control unit 208 has moved only the object, which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 7A . That is, FIG. 7B shows the operation of Step S 404 shown in FIG. 4 .
  • the object when a contact point is detected at the preset position, the object is controlled so as to be movable with the schedule chart being fixed. Accordingly, the object can be moved while the schedule chart is fixed.
  • a control is suitable for a program of displaying the first display information 501 and the second display information 502 in a combined manner (without display control for each object).
  • FIGS. 8A and 8B are diagrams showing still another transition example of the display screen in a case of multiple touches.
  • FIGS. 8A and 8B show an example of screen transition in which only an object is controlled so as to be movable with the schedule chart being fixed when a contact is detected at a preset position as in the case of FIGS. 7A and 7B .
  • FIGS. 8A and 8B differ from FIGS. 7A and 7B in the following points:
  • the movement-fixing coordinate area is set at a portion of the schedule chart in which time is shown.
  • a movement-fixing specified position 505 is set in a visible manner at a lower part of the active range by an application program.
  • FIG. 8A shows a case of detecting a contact point in the movement-fixing specified position 505 provided in the first coordinate area in which the schedule chart has been displayed, detecting a contact point in the second coordinate area in which the object has been displayed, and further detecting a change of the detection coordinate of the detected point in the second coordinate area as a scroll instruction.
  • FIG. 8B shows a display result in which the control unit 208 has moved only the object, which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 8A .
  • the object when a contact point is detected at the movement-fixing specified position 505 , which has been set to be visible, the object is controlled so as to be movable with the schedule chart being fixed. Accordingly, the object can be moved while the schedule chart is fixed.
  • a control is suitable for a program of displaying a number of objects.
  • such a control is suitable for an application program of displaying display information using a layer structure having a number of layers.
  • the control unit sets the object (second display information) on an upper layer of the schedule chart (first display information). Specifically, the second display information is displayed such that it is overlaid on the first display information.
  • FIGS. 9A and 9B are diagrams showing still another transition example of the display screen in a case of multiple touches.
  • FIGS. 9A and 9B show an example of screen transition in which only an object is controlled so as to be movable with the schedule chart being fixed when a contact is detected at a preset position as in the other transition examples of multiple touches.
  • FIGS. 9A and 9B differ from FIGS. 8A and 8B in the following points:
  • the movement-fixing specified position 505 is set at the lower part of the active range by the application program for the schedule chart.
  • a movement-fixing object 506 is set by a program other than the application program for the schedule chart.
  • FIG. 9A shows a case of detecting a contact point (selection) in the coordinate area in which the movement-fixing object 506 has been displayed, which are provided in the first coordinate area in which the schedule chart has been displayed, detecting a contact point in the second coordinate area in which the object to be moved has been displayed, and further detecting a change of the detection coordinate of only the contact point in the second coordinate area as a scroll instruction.
  • FIG. 9B shows a display result in which the control unit 208 has moved only the object, which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 9A .
  • the object when a contact point is detected in the coordinate area in which the movement-fixing object 506 has been displayed by a program other than an active application program, the object is controlled so as to be movable with the schedule chart being fixed. Accordingly, the object can be moved while the schedule chart is fixed.
  • the movement-fixing object 506 is movable on the screen. Furthermore, the movement-fixing object 506 may be displayed in the foreground. With such a control, assuming that the schedule chart is likened to a paper, the movement-fixing object can be likened to a weight or a paperweight. Thus, a user's intuitive operation can properly be detected, and an object provided with a movement instruction can properly be moved.
  • FIGS. 10A and 10B are diagrams showing still another transition example of the display screen in a case of multiple touches.
  • FIGS. 10A and 10B show an example of screen transition in which only an object is controlled so as to be movable with the schedule chart being fixed when a contact is detected at a preset position as in the other transition examples of multiple touches.
  • FIGS. 10A and 10B differ from the transition example of FIGS. 8A and 8B in the following points:
  • the movement-fixing specified position 505 is set at the lower part of the active range by the application program for the schedule chart.
  • an underlay fixing input portion 507 is set in an area outside of the display screen, but within a detection range of the display unit 206 having a contact detection function.
  • FIG. 10A shows a case of detecting contact with the underlay fixing input portion 507 and contact with the second coordinate area in which the object to be moved has been displayed, and further detecting a change of the detection coordinate of only the contact point in the second coordinate area as a scroll instruction.
  • FIG. 10B shows a display result in which the control unit 208 has moved only the object, which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 10A .
  • the object to be moved is controlled so as to be movable in a state in which the underlay (the schedule chart in this example) is fixed. Accordingly, the object can be moved while the underlay is fixed.
  • the underlay fixing input portion can be used to press an edge portion of the paper.
  • FIGS. 11A and 11B are diagrams showing still another transition example of the display screen in a case of multiple touches.
  • FIGS. 11A and 11B show an example of screen transition in which only an object is controlled so as to be movable with the schedule chart being fixed when a plurality of contact points are concurrently detected as in the other transition examples of multiple touches.
  • FIGS. 11A and 11B when a user touches a coordinate area corresponding to the display screen frame 500 and the underlay (the schedule chart in this example) with a finger of a hand holding the mobile terminal 200 at the same time, another object is controlled so as to be movable while the underlay is fixed.
  • a multi-touch screen detectable area which is an area of the display unit 206 having a contact detection function for detecting contact, is provided not only on the display screen, but also on the display screen frame. Furthermore, as shown in FIG. 12 , a contact point continuously extending from the display screen frame 500 to the display screen 509 is deemed contact with the movement-fixing coordinate area.
  • FIG. 11A shows a case of detecting a contact point in the coordinate area corresponding to the display screen frame 500 and the underlay with a finger of a hand holding the mobile terminal 200 , detecting a contact point in the second coordinate area in which the object to be moved has been displayed, and further detecting a change of the detection coordinate of only the contact point in the second coordinate area as a scroll instruction.
  • FIG. 11B shows a display result in which the control unit 208 has moved only the object displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 11A .
  • FIG. 12 is an enlarged view of the display screen shown in FIG. 11 .
  • contact with the movement-fixing coordinate area is detected if a contact point (area) hangs over both of the display screen frame (enclosure) 500 and the display screen 509 .
  • the display screen frame (enclosure) 500 and the underlay are brought into contact with the same finger of the user at the same time, it is determined that the operation is for fixing the underlay.
  • the object to be moved is controlled so as to be movable in a state in which the underlay (the schedule chart in this example) is fixed. Accordingly, the object can be moved while the underlay is fixed.
  • the schedule chart is likened to a paper
  • the paper can be pressed against the housing of the mobile terminal 200 .
  • the underlay can be sandwiched between the housing and the finger. Therefore, a user's intuitive operation can properly be acquired, and an object can properly be moved.
  • an input item of “Meeting” is made movable while the schedule chart (first display information), which is an underlay (background), is fixed (is not allowed to move).
  • the control unit may detect arrival of the object being moved to the vicinity of a lower edge of the display screen and move the information displayed on the edge to the vicinity of a central portion of the active range, or may switch the display information into the next page.
  • a movement-fixing coordinate area is provided in a first display range in which first display information is displayed and/or a second display range in which second display information is displayed.
  • detecting a first contact point in the movement-fixing coordinate area detecting a second contact point in either one of the first coordinate area and the second coordinate area other than the movement-fixing coordinate area, and detecting a change of the detection coordinate of the second contact point, only the first or second display information displayed in the first or second coordinate area in which the second contact point is present can be moved based upon the change of the detection coordinate.
  • the application program is for the schedule chart.
  • the first display information (underlay or background) is not limited to a schedule chart.
  • a map, a calendar, a photograph list, and the like may be displayed as the first display information.
  • the application program may be a document input program or a spreadsheet program. In this case, a selected range of input text or an individual cell can be used as an object. In the case where the present invention is applied to a document input program, a selected part of text can be moved as an object provided with a movement instruction in addition to photographs and figures by a user's intuitive operation.
  • a coordinate area in which a contact point is present may be controlled to be fixed.
  • a predetermined area can be compared with another area away from the predetermined area in a state in which those areas are placed side by side.
  • FIG. 13 is a functional block diagram showing an overview of a configuration of the tablet computer 300 according to the present embodiment.
  • the tablet computer 300 includes a control unit 310 operable to perform a variety of calculations, a RAM 311 operable to temporarily store information, a ROM 312 in which a fundamental control program has been stored, a touch panel display 320 operable to input and output information, a storage unit 330 , and a network interface 313 for communication with the Internet or the like via a network.
  • a control unit 310 operable to perform a variety of calculations
  • a RAM 311 operable to temporarily store information
  • a ROM 312 in which a fundamental control program has been stored
  • a touch panel display 320 operable to input and output information
  • storage unit 330 operable to input and output information
  • a network interface 313 for communication with the Internet or the like via a network.
  • the touch panel display 320 has a display unit 321 and a contact detection unit 322 .
  • the touch panel display 320 displays display information outputted from the control unit 310 , such as images, figures, characters, and symbols, on a display screen of the display unit 321 .
  • the touch panel display 320 is operable to detect an input from a user through contact with the contact detection unit 322 .
  • the storage unit 330 stores therein various application programs 331 , a driver 332 for the touch panel display 320 , various types of data, various types of contents, an OS (Operating System) 333 , and the like.
  • the control unit 310 operates in accordance with instructions from the OS 333 , the driver 332 , and the application programs 331 .
  • Software programs such as the OS 333 are expanded in the RAM 311 as needed and used.
  • the contact detection unit 322 is operable to detect contact with hardware of the contact detection unit 322 (multi-touch screen detectable area) based upon one of the OS 333 , the driver 332 , and the application programs 331 , or a combination thereof.
  • the contact detection unit 322 may use any detection methods as long as it can detect a plurality of contact points (areas).
  • the contact detection unit 322 may be implemented by using any software program.
  • the contact detection unit 322 is not specifically limited as long as it can detect contact.
  • the tablet computer 300 can properly detect a user's intuitive operation with use of the contact detection unit 322 and can properly move an object.
  • the identification process will be described in an example of a spreadsheet program with reference to FIGS. 14 , 15 A, and 15 B.
  • FIG. 14 is a flow chart explanatory of a movement operation of the tablet computer 300 .
  • Processes from S 1401 to S 1403 shown in FIG. 14 are the same as the processes from Step S 401 to S 403 of FIG. 4 . Furthermore, processes of Steps S 1405 and S 1406 are the same as the processes of Steps S 405 and S 406 of FIG. 4 .
  • the control unit 310 of the tablet computer 300 monitors contact with the touch panel display 320 (Step S 1401 ). If any contact is detected, the control unit 310 judges whether there is one contact point or are a plurality of contact points (Step S 1402 ). Furthermore, the control unit 310 detects contact with a display range of another object (another cell) on the display screen other than a display range of an object to be moved (cell to be moved) and judges whether a movement instruction for the object to be moved can be detected (Step S 1403 ).
  • control unit 310 controls the object to be moved so as to be movable while the other object is fixed (Step S 1404 ).
  • Steps S 1405 -S 1406 a predetermined normal operation is performed.
  • FIGS. 15A and 15B are diagrams showing a transition example of a display screen 600 of a spreadsheet program in a case of multiple touches.
  • FIG. 15A shows a case of detecting a plurality of contact points in a first coordinate area in which first display information (table in this example) 601 has been displayed and in a second coordinate area in which second display information (each cell in this example) 602 has been displayed and further detecting a change of a detection coordinate of only the contact point in the second coordinate area as a scroll instruction. That is, FIG. 15A shows an operation from Step S 1401 via “Yes” of S 1402 to “Yes” of Step S 1403 shown in FIG. 14 .
  • FIG. 15B shows a display result in which the control unit 310 has scrolled only the second display information (object), which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 15A . That is, FIG. 15B shows the operation of Step S 1404 shown in FIG. 14 .
  • a cell 604 provided with a movement instruction (cell of 7/10 and 9:00) is controlled so as to be movable while a rectangular area 603 with corners of the upper-left cell and the cell provided with no movement instruction (cell to be fixed) is fixed. Furthermore, if the cell 604 controlled so as to be movable is moved near an edge of the active area, then the first display information is displayed in a divided manner. The cell 604 to be moved is controlled so as to be movable while a frame portion of the spreadsheet program is displayed.
  • the object provided with the movement instruction is controlled so as to be movable within the active range. Accordingly, the user can move information while he/she views and compares the cell being moved (7/10 9:00) and other cells away from that cell being moved with each other in the vertical and horizontal directions.
  • control unit 310 detects contact with the coordinate area in which a plurality of pieces of second display information have been displayed in a frame in which the first display information has been displayed and detects a change of the detection coordinate of only one contact point, then the first display information is displayed in a divided manner. Only the second display information displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected is moved based upon the change of the detection coordinate.
  • an information processing apparatus of the present invention recognizes contact of a user with a screen at a plurality of points and an operation directed to a displayed object. Movement control of an object is performed based upon the recognition result. Therefore, a burden of a user's input operation can be reduced.
  • an information processing apparatus that can properly detect a user's intuitive operation with use of a contact detection unit capable of detecting a plurality of contact points and can properly move an object, a program for implementing such an information processing apparatus, and a computer-readable storage medium in which such a program is recorded.
  • two or more examples of the aforementioned intuitive fixing operations may be used in a combined manner. Such a combination can absorb differences between individual users and further enhance the usability of the users.
  • the present invention is applicable to any information processing apparatus having a contact detection unit capable of detecting a plurality of contact points. Furthermore, the present invention is suitable for an information processing apparatus held by a single hand to operate.

Abstract

An information processing apparatus has a display unit operable to display an object, a contact detection unit operable to detect a plurality of contact points on the display unit, and a control unit operable to control the object displayed on the display unit so as to be movable with use of the contact detection unit. When the control unit detects contact with a display range of the object displayed on the display unit with use of the contact detection unit, detects contact with an active range of a display screen other than the display range of the object, and further detects a movement instruction to the object, then the control unit controls the object so as to be movable within the active range. Thus, use of a contact detection unit capable of a plurality of contact points permits a user's intuitive operation.

Description

    TECHNICAL FIELD
  • The present invention relates to an information processing apparatus having a contact detection means capable of detecting a plurality of contact points, a storage medium on which a program used in the information processing apparatus is recorded, and an object movement method.
  • BACKGROUND ART
  • A contact detection unit called a touch panel or a touch screen is provided on various information processing apparatuses these days. Furthermore, a contact detection means capable of detecting a plurality of contact points has been developed instead of a conventional contact detection means capable of detecting only one contact point.
  • Japanese laid-open patent publications Nos. 2002-304256 and 2007-279638 disclose a touch panel display or a touch panel device capable of detecting a plurality of contact points.
  • DISCLOSURE OF INVENTION
  • An operation method of an information processing apparatus having a contact detection means is designed so as to allow a user to intuitively operate the information processing apparatus. While such an operation method provides a user with intuitive operation, an excessive burden is imposed on the user if the operation (command to a controller, transition of a screen, or the like) is not performed as expected by the user' intuitive image.
  • For example, an information processing apparatus that scrolls displayed contents by detection of contact with a touch panel (determination by touch) or by detection of removal of contact with the touch panel (determination by touch and release) provides a user with sufficient operability on a scroll operation. However, in a case of detailed operations, such as a coordinate movement of only a desired part of displayed information, a scroll operation of only a desired part of displayed information, or an operation using a plurality of contact points, a (slight) difference is likely to be produced between the setting of the information processing apparatus and the user's intuitive operation. In such a case, the user will repeat intuitive operations. For example, even though a user attempts a scroll operation and touches a touch panel, the information processing apparatus does not operate if a scroll operation is set to be activated through the determination by touch and release. In such a case, the user repeats the same operation because no desired operation is performed. Thus, an excessive burden is imposed on the user.
  • As a result, a command to a controller or transition of a screen (such as scrolling) that has not been intended by the user is provided. Therefore, sufficient operability cannot be provided.
  • The present invention has been made in view of the above drawbacks. The present invention provides an information processing apparatus that can properly detect a user's intuitive operation with use of a contact detection unit capable of detecting a plurality of contact points and can properly move an object.
  • An information processing apparatus according to the present invention includes a display unit operable to display an object, a contact detection unit operable to detect a plurality of contact points on a display screen of the display unit, and a control unit operable to control the object displayed on the display unit so as to be movable with use of the contact detection unit. When the control unit detects a first contact point in an active range of the display screen other than a display range of the object with use of the contact detection unit, detects a second contact point in the display range of the object, and further detects a movement instruction to the object, then the control unit controls the object so as to be movable within the active range.
  • According to the present invention, there can be provided an information processing apparatus that can properly detect a user's intuitive operation with use of a contact detection unit capable of detecting a plurality of contact points and can properly move an object.
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 is a functional block diagram showing an information processing apparatus according to a first embodiment of the present invention.
  • FIG. 2 is a flow chart explanatory of an identification operation of the information processing apparatus of FIG. 1.
  • FIG. 3 is a functional block diagram showing a mobile terminal as an information processing apparatus according to a second embodiment of the present invention.
  • FIG. 4 is a flow chart explanatory of a scroll operation of the mobile terminal of FIG. 3.
  • FIGS. 5A and 5B are diagrams showing a transition example of a display screen in the mobile terminal of FIG. 3 in a case of a single touch.
  • FIGS. 6A and 6B are diagrams showing a first transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIGS. 7A and 7B are diagrams showing a second transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIGS. 8A and 8B are diagrams showing a third transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIGS. 9A and 9B are diagrams showing a fourth transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIGS. 10A and 10B are diagrams showing a fifth transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIGS. 11A and 11B are diagrams showing a sixth transition example of the display screen in the mobile terminal of FIG. 3 in a case of multiple touches.
  • FIG. 12 is a partial enlarged view of the display screen shown in FIG. 11.
  • FIG. 13 is a functional block diagram showing a tablet computer as an information processing apparatus according to a third embodiment of the present invention.
  • FIG. 14 is a flow chart explanatory of a movement operation of the tablet computer of FIG. 13.
  • FIGS. 15A and 15B are diagrams showing a transition example of a display screen of a spreadsheet program in the tablet computer of FIG. 13 in a case of multiple touches.
  • BEST MODE FOR CARRYING OUT THE INVENTION
  • An information processing apparatus according to a first embodiment of the present invention will be described with reference to FIGS. 1 and 2. For example, an information processing apparatus according to the present invention is operated by a program control and is formed as a mobile terminal having a communication function (telephone conversation, mail transmission and reception, and Internet connection), a calendar function, a scheduler function, and the like. Other examples of the information processing apparatus include a personal computer, a smart phone, a PDA (Personal Digital Assistants), and a PHS (Personal Handyphone System) having a touch screen function.
  • FIG. 1 is a functional block diagram showing an overview of a configuration of an information processing apparatus 100 according to the first embodiment.
  • The information processing apparatus 100 shown in FIG. 1 includes a control unit 110 operable to perform a variety of calculations, a display unit 120 operable to display information directed from the control unit 110, a contact detection unit 121 operable to detect the presence of a contact point in a coordinate area of the display unit 110 in which the display information is displayed, a coordinate of the contact point (contact coordinate or detection coordinate), and a change (movement) of the detection coordinate, and a storage unit 130 formed by a ROM, a RAM, a HDD, a flash memory, or the like. The storage unit 130 stores therein an operating system, a driver or an application program for the contact detection unit 121, various types of setting information of the information processing apparatus 100, and the like.
  • The control unit 110 is operable to control the display unit 120 so as to display first display information in a first display range. Furthermore, the control unit 110 is operable to display second display information in a second display range that is part of the first display range. Here, the term “display range” refers to an area (outer frame) in which display information can be displayed. The term “coordinate area” refers to an area in which display information has actually been displayed. While the second display range is included in the first display range, a coordinate area in which the second display information has been displayed is not included in a coordinate area in which the first display information has been displayed.
  • Furthermore, the control unit 110 can detect, via the contact detection unit 121, the presence of a contact point in a coordinate area of the display unit 120 in which the first and second display information has been displayed. The control unit 110 can also identify a change (movement) of the detected coordinate (detection coordinate). Moreover, the control unit 110 can detect a plurality of contact points in a coordinate area in which the first display information is displayed and in a coordinate area in which the second display information is displayed. Furthermore, when a change of a detection coordinate of only one of a plurality of contact points is detected, the control unit 110 performs a process of moving (scrolling) only the display information (object) displayed in a coordinate area in which the change of the detection coordinate of the contact point has been detected based upon the change of the detection coordinate.
  • The first display information, which is displayed in the first display range, which is the entire display screen or a predetermined area (active range), and the second display information, which is displayed in part of that display range, are displayed on the display unit 120 by the control unit 110.
  • The second display information (object) may include an image, a figure, a character, or a symbol, or a combination thereof. Furthermore, the second display information may be expressed by any one of an icon, a button, a character string, a text box, and area information visually indicated for providing a command to the control unit 110, or a combination thereof. Moreover, the second display information including a character string also includes information for managing a selected character string as a bundle after the character string has been selected by range specification.
  • With this configuration, the information processing apparatus 100 can properly detect a user's intuitive operation with use of the contact detection unit 121 capable of detecting a plurality of contact points and can properly move an object. In this specification, the term “contact point” includes not only a relatively small “point” as in a case where a stylus pen or the like is used, but also a relatively large “area” as in a case where a human finger or the like is used.
  • Next, operation of the information processing apparatus 100 will be described.
  • In accordance with the application program and the operating system recorded in the storage unit 130, the control unit 110 of the information processing apparatus 100 produces images of the first display information to be displayed in the first display range and the second display information to be displayed in the first display range of the screen of the display unit 120 and then displays them on the display unit 120. Concurrently, in accordance with the driver for the contact detection unit 121, the operating system and application program, the control unit 110 detects contact with the first coordinate area in which the first display information has been displayed and the second coordinate area in which the second display information has been displayed (appearance of a contact point), removal of contact (disappearance of the contact point), detection coordinates, and changes of the detection coordinates with use of the contact detection unit 121. The control unit 110 recognizes the detected contact, removal of contact, and changes of the detection coordinates as determination by touch, determination by touch and release, or a movement instruction as needed and determines an operational instruction to the displayed object or the like.
  • There will be described a process of the control unit 110 that identifies a movement instruction of the object from changes of detection coordinates of multiple contact points detected with the contact detection unit 121.
  • In accordance with various types of software programs, the control unit 110 detects a plurality of contact points with the contact detection unit 121 and identifies a user's movement instruction based upon the position coordinates (detection coordinates) of the contact points and the changes of those position coordinates. A movement instruction can be distinguished from other instructions by vectorizing detection coordinates of individual contact points so as to identify movement directions. Furthermore, the movement instruction may be distinguished from other instructions by obtaining a period of time from the detection of contact to the detection of removal of contact with respect to each contact point and relying upon differences of those periods.
  • For example, when two contact points are concurrently detected, and detection coordinates of those contact points move concurrently, then the operation can be determined to be magnification or reduction of the display screen. Meanwhile, when two contact points are detected at an interval of time, and a detection coordinate of one of those contact points hardly moves, and a detection coordinate of the other contact point moves, then the operation can be determined to be a movement instruction. With such an identifying process, different operations can be performed depending upon a difference in change of detection coordinates of the multiple contact points.
  • Furthermore, operations may be differentiated by using the size of a contact area. For example, operations may be differentiated by using a difference in size of a contact area between a stylus pen and a finger. Moreover, operations may be differentiated by using a difference in size of a contact area between a thumb and other fingers. For example, the contact detection unit 121 may be set such that an operation of only an underlay of the display screen or only a fixed operation is allowed around an area at which a finger to hold the information processing apparatus 100 is located. Furthermore, depending upon the size of the display information in the display screen, a contact point having a large contact area or a small contact area may be determined to be invalid. Additionally, operations may be differentiated depending upon whether the information processing apparatus 100 is held by a single hand or by both hands.
  • In any case, even if a movement instruction is detected by contact with a coordinate area of one object, the program allows the control unit 110 of the information processing apparatus 100 to perform a process of moving only one object while fixing the underlay or the background or perform a process of moving the entire active display screen based upon the positional relationship with contact with other coordinate areas, a duration of contact, and the like. Thus, a user's intuitive operation can properly be detected.
  • FIG. 2 is a flow chart explanatory of an identification operation of the information processing apparatus 100. The identification process of the operation shown in FIG. 2 is to provide a function of moving an object, which is provided as a user interface to a user when the aforementioned information processing apparatus 100 operates.
  • The control unit 110 of the information processing apparatus 100 monitors contact with the display screen of the display unit 120 via the contact detection unit 121 (Step S201).
  • If any contact is detected, the control unit 110 judges whether a plurality of contact points are concurrently present. The process proceeds to the next Step S203 if a plurality of contact points have been detected, and the process proceeds to Step S205 if only one contact point has been detected (Step S202). In other words, the control unit 110 judges whether a contact point has been detected in only one of the first and second coordinate areas in which the first and second display information has been displayed, or whether a plurality of contact points have been displayed in both of the first and second coordinate areas. Furthermore, in addition to detection of a contact point in the display range of the object, the control unit 110 judges whether a contact point has been detected in an active range of the display screen other than the display range of the object.
  • Detections of contact points in both of the first coordinate area and the second coordinate area are judged to be the same input operation whichever is first detected. Nevertheless, concurrent detections of contact points in the first coordinate area and the second coordinate area may be judged to be different input operations.
  • When a plurality of contact points have been detected, the control unit 110 monitors detection coordinates of those detected contact points. The control unit 110 judges whether only one of the contact points has moved (a movement (scroll) instruction has been provided). The process proceeds to the next Step S204 if only one contact point has moved, and the process proceeds to the next Step S206 if the contact points have moved concurrently (Step S203). In other words, the control unit 110 judges whether a change of a detection coordinate of only one of the contact points has been detected. Saying it differently, the control unit 110 detects a contact point in an active range of the display screen other than the display range of the object to be moved and judges whether a movement instruction to the object has been detected.
  • If another operation (instruction) that is different from the movement instruction is made, the control unit 110 performs a process corresponding to that operation.
  • If movement of only one contact point has been detected, the control unit 110 fixes, on the display screen, the display information displayed in the coordinate area in which the other contact point is present, and moves the display information displayed in the coordinate area in which movement of the contact point has been detected (Step S204). In other words, only the display information displayed in the coordinate area in which a change of the detection coordinate of the contact point has been detected is moved based upon the change of the detection coordinate. Saying it differently, an object is controlled so as to be movable within the active range, and an object to which a movement instruction has been provided is moved.
  • In the case of one contact point, the control unit 110 monitors the detection coordinate of the detected contact point and judges whether a movement instruction has been provided by a user (Step S205).
  • If the detected contact points concurrently move in the same direction, or if one contact point has been detected and the detection coordinate thereof changes, the control unit 110 moves (scrolls) the entire active display information (Step S206). In other words, if changes of the detection coordinates of all of the contact points have been detected, then the control unit 110 moves the first and second display information based upon the changes of the detection coordinates.
  • With the above identification operation, the information processing apparatus 100 can properly detect a user's intuitive operation with use of a contact detection unit capable of detecting a plurality of contact points and can thus move display information (object).
  • Next, a mobile terminal 200 will be described as an information processing apparatus according to a second embodiment of the present invention.
  • FIG. 3 is a functional block diagram showing an overview of a configuration of the mobile terminal 200 according to the present embodiment.
  • As shown in FIG. 3, the mobile terminal 200 includes a radio unit 201, a broadcast receiver unit 202, a GPS receiver unit 203, a camera unit 204, an acoustical treatment unit 205, a display unit 206 having a contact detection function, a storage unit 207, and a control unit 208.
  • The radio unit 201 is operable to wirelessly transmit information to and receive information from a radio base station apparatus via an antenna. The broadcast receiver unit 202 is operable to receive broadcast signals transmitted from a broadcast station (a ground station, a satellite, or the like) and conducts signal processing of visual data, audio data, information data, and the like, which have been obtained by demodulating the received broadcast signals. The GPS receiver unit 203 is operable to measure periods of time for radio waves emitted from a plurality of GPS satellites to reach the mobile terminal 200, compute distances between the respective GPS satellites and the mobile terminal 200, and compute a position coordinate with use of the computed distances. The camera unit 204 is operable to acquire image information (take a photograph). The acoustical treatment unit 205 is operable to process acoustic signals of music, notification sounds, or voice inputted and outputted via a microphone or a speaker.
  • The display unit 206 having a contact detection function is a touch screen (touch panel) having a display function of outputting display information such as images, figures, characters, and symbols (video signal processing) and a switch function of sensing a user's input operation by using pressure, capacitance, or the like (position coordinate detection function and contact detection function).
  • The storage unit 207 stores therein telephone directory information, sent and received mail information, outgoing and incoming call information, content information, application programs, photographed images, various types of setting information of the mobile terminal 200, and the like.
  • The control unit 208 is operable to control a telephone function, a mail function, an Internet (web) connection function, a camera function, a television receiving function, a GPS (positioning) function, a content playback function, other functions of the terminal, and respective components of the terminal. Furthermore, the control unit 208 is operable to control a display function of the display unit 206 having a contact detection function and display an electronic screen on the display unit 206. The control unit 208 is operable to detect a user's operational instruction by using the switch function (position coordinate detection function and contact detection function) of the display unit 206 having a contact detection function.
  • Next, an identification process of an operation (instruction) of the mobile terminal 200 will be described with reference to FIG. 4.
  • FIG. 4 is a flow chart explanatory of a scroll operation of the mobile terminal 200.
  • The control unit 208 of the mobile terminal 200 monitors contact with the display unit 206 having a contact detection function (Step S401). If any contact is detected, the control unit 208 judges whether the number of contact points is one or more. The process proceeds to the next Step S403 if a plurality of contact points have been detected, and the process proceeds to Step S405 if one contact point has been detected (Step S402).
  • If a plurality of contact points have been detected, the control unit 208 judges whether at least one of those contact points is located at a predetermined position (or within an area (movement-fixing coordinate area)). The predetermined position will be discussed later.
  • The process proceeds to the next Step S404 if any contact has been detected at the preset position, and the process proceeds to Step S405 if no contact has been detected at the preset position (Step S403).
  • If the control unit 208 detects any contact at the preset position and also detects a movement instruction to an object (corresponding to the second display information), the control unit 208 fixes the background display (corresponding to the first display information) and controls the object so as to be movable within the active range (Step S404).
  • Meanwhile, if one contact point has been detected, the control unit 208 scrolls the entire display based upon the amount of movement of the contact point until the contact is removed (the contact point disappears) (Steps S205-S206).
  • The preset position (or area) will be described. The preset position is set within a range of the display screen other than an object to be moved, or within an active range of the display screen other than an object to be moved, or at a position (designated position) that can be contacted (pressed) by a user's hand to hold the mobile terminal 200. Saying it differently, the preset position is set within the active range of the display screen other than the display range of the object, or within the first coordinate area in which the first display information has been displayed, other than the second coordinate area in which the second display information has been displayed and in which the contact point has been detected, or within a designated coordinate area designated by an application program, or within the display range of a movement-fixed object.
  • The position that can be contacted (pressed) is a position that can be reached by a thumb or another specified finger when the mobile terminal 200 is held in a recommended standard manner. The preset position will be exemplified in FIG. 7A and figures following FIG. 7A.
  • Next, some transition examples of the display screen of the mobile terminal 200 will be described. FIGS. 5A to 11B are diagrams showing transition examples of the display screen of the mobile terminal 200.
  • FIGS. 5A to 11B show the entire display screen displayed on the display unit 206 having a contact detection function by the control unit 208. FIGS. 5A, 6A, 7A, 8A, 9A, 10A, and 11A show the display screen prior to movement (scroll) of an object. On the other hand, FIGS. 5B, 6B, 7B, 8B, 9B, 10B, and 11B show the display screen after movement of an object.
  • FIGS. 5A and 5B are diagrams showing a transition example of the display screen in a case of a single touch.
  • In FIG. 5A, first display information (schedule chart) 501 is displayed in a display screen frame 500, and second display information (object) 502 is displayed. In this state, it is assumed that a user touches a second coordinate area in which the second display information 502 has been displayed with his/her finger and then moves his/her finger downward as indicated by the arrow in FIG. 5A. The control unit 208 does not detect any contact point in a first coordinate area in which the first display information 501 has been displayed, but detects a contact point in the second coordinate area in which the second display information 502 has been displayed. The control unit 208 detects a change of a detection coordinate of the contact point as a scroll instruction. That is, FIG. 5A shows an operation from Step S401 via “No” of Step S402 to Step S405 shown in FIG. 4.
  • FIG. 5B shows a display result in which the control unit 208 has scrolled the first and second display information 501 and 502, i.e., the entire display information, based upon the change of the detection coordinate as a result of the identification process of FIG. 5A. That is, FIG. 5B shows a state after the operation of Step S406 shown in FIG. 4.
  • As shown in FIG. 5A, when the user attempts to move the object in a case where a contact point has been detected only in the display range of the object (the second display information 502), then the schedule chart (first display information 501) and the object are scrolled at the same time.
  • FIGS. 6A and 6B are diagrams showing a transition example of the display screen in a case of multiple touches. FIGS. 6A and 6B show an example of screen transition in which the object 502 is controlled so as to be movable in an active range 503 when a contact point is detected in the active range 503 (hatched area) of the display screen other than the display range of the object 502.
  • FIG. 6A shows a case of detecting a plurality of contact points in a first coordinate area in which the first display information 501 has been displayed and a second coordinate area in which the second display information 502 has been displayed and detecting, as a scroll instruction, a change of a detection coordinate of only the contact point in the second coordinate area, which is one of the contact points. That is, FIG. 6A shows an operation from Step S401 via “Yes” of Step S402 to “Yes” of Step S403 shown in FIG. 4.
  • FIG. 6B shows a display result in which the control unit 208 has moved the second display information (object) 502, which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 6A. That is, FIG. 6B shows the operation of Step S404 shown in FIG. 4.
  • As shown in FIGS. 6A and 6B, when a contact point is detected in the active range 503 of the display screen other than the display range of the object, and a detection coordinate of that contact point does not change, then the object is controlled so as to be movable within the active range 503. Accordingly, the object can be moved while the schedule chart is fixed.
  • When a plurality of contact points are detected in the first coordinate area and the second coordinate area, and a change of the detection coordinate of the contact point in the first coordinate area is detected without detecting a change of the detection coordinate of the contact point in the second coordinate area, then the control unit 208 may scroll (move) only the first display information 501 based upon the change of the detection coordinate. The contact points are not necessarily required to be present in both of the first coordinate area and the second coordinate area.
  • FIGS. 7A and 7B are diagrams showing another transition example of the display screen in a case of multiple touches. FIGS. 7A and 7B show an example of screen transition in which an object of “Meeting” is controlled so as to be movable with the schedule chart being fixed when a contact is detected at a preset position (designated position, designated coordinate range, movement-fixing coordinate area).
  • FIG. 7A shows a case of detecting a contact point at a preset position (movement-fixing coordinate area) 504 provided in the first coordinate area in which the schedule chart has been displayed, and a contact point in the second coordinate area in which the object has been displayed, and further detecting a change of the detection coordinate of the contact point in the second coordinate area as a movement instruction.
  • FIG. 7B shows a display result in which the control unit 208 has moved only the object, which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 7A. That is, FIG. 7B shows the operation of Step S404 shown in FIG. 4.
  • As shown in FIGS. 7A and 7B, when a contact point is detected at the preset position, the object is controlled so as to be movable with the schedule chart being fixed. Accordingly, the object can be moved while the schedule chart is fixed. Such a control is suitable for a program of displaying the first display information 501 and the second display information 502 in a combined manner (without display control for each object).
  • FIGS. 8A and 8B are diagrams showing still another transition example of the display screen in a case of multiple touches. FIGS. 8A and 8B show an example of screen transition in which only an object is controlled so as to be movable with the schedule chart being fixed when a contact is detected at a preset position as in the case of FIGS. 7A and 7B. FIGS. 8A and 8B differ from FIGS. 7A and 7B in the following points: In FIGS. 7A and 7B, the movement-fixing coordinate area is set at a portion of the schedule chart in which time is shown. In FIGS. 8A and 8B, a movement-fixing specified position 505 is set in a visible manner at a lower part of the active range by an application program.
  • FIG. 8A shows a case of detecting a contact point in the movement-fixing specified position 505 provided in the first coordinate area in which the schedule chart has been displayed, detecting a contact point in the second coordinate area in which the object has been displayed, and further detecting a change of the detection coordinate of the detected point in the second coordinate area as a scroll instruction.
  • FIG. 8B shows a display result in which the control unit 208 has moved only the object, which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 8A.
  • As shown in FIGS. 8A and 8B, when a contact point is detected at the movement-fixing specified position 505, which has been set to be visible, the object is controlled so as to be movable with the schedule chart being fixed. Accordingly, the object can be moved while the schedule chart is fixed. Such a control is suitable for a program of displaying a number of objects. Additionally, such a control is suitable for an application program of displaying display information using a layer structure having a number of layers. When a layer structure is used, the control unit sets the object (second display information) on an upper layer of the schedule chart (first display information). Specifically, the second display information is displayed such that it is overlaid on the first display information.
  • FIGS. 9A and 9B are diagrams showing still another transition example of the display screen in a case of multiple touches. FIGS. 9A and 9B show an example of screen transition in which only an object is controlled so as to be movable with the schedule chart being fixed when a contact is detected at a preset position as in the other transition examples of multiple touches. FIGS. 9A and 9B differ from FIGS. 8A and 8B in the following points: In FIGS. 8A and 8B, the movement-fixing specified position 505 is set at the lower part of the active range by the application program for the schedule chart. In FIGS. 9A and 9B, a movement-fixing object 506 is set by a program other than the application program for the schedule chart.
  • FIG. 9A shows a case of detecting a contact point (selection) in the coordinate area in which the movement-fixing object 506 has been displayed, which are provided in the first coordinate area in which the schedule chart has been displayed, detecting a contact point in the second coordinate area in which the object to be moved has been displayed, and further detecting a change of the detection coordinate of only the contact point in the second coordinate area as a scroll instruction.
  • FIG. 9B shows a display result in which the control unit 208 has moved only the object, which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 9A.
  • As shown in FIGS. 9A and 9B, when a contact point is detected in the coordinate area in which the movement-fixing object 506 has been displayed by a program other than an active application program, the object is controlled so as to be movable with the schedule chart being fixed. Accordingly, the object can be moved while the schedule chart is fixed.
  • The movement-fixing object 506 is movable on the screen. Furthermore, the movement-fixing object 506 may be displayed in the foreground. With such a control, assuming that the schedule chart is likened to a paper, the movement-fixing object can be likened to a weight or a paperweight. Thus, a user's intuitive operation can properly be detected, and an object provided with a movement instruction can properly be moved.
  • FIGS. 10A and 10B are diagrams showing still another transition example of the display screen in a case of multiple touches. FIGS. 10A and 10B show an example of screen transition in which only an object is controlled so as to be movable with the schedule chart being fixed when a contact is detected at a preset position as in the other transition examples of multiple touches. FIGS. 10A and 10B differ from the transition example of FIGS. 8A and 8B in the following points: In FIGS. 8A and 8B, the movement-fixing specified position 505 is set at the lower part of the active range by the application program for the schedule chart. In FIGS. 10A and 10B, an underlay fixing input portion 507 is set in an area outside of the display screen, but within a detection range of the display unit 206 having a contact detection function.
  • FIG. 10A shows a case of detecting contact with the underlay fixing input portion 507 and contact with the second coordinate area in which the object to be moved has been displayed, and further detecting a change of the detection coordinate of only the contact point in the second coordinate area as a scroll instruction.
  • FIG. 10B shows a display result in which the control unit 208 has moved only the object, which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 10A.
  • As shown in FIGS. 10A and 10B, when contact with the underlay fixing input portion 507 is detected, the object to be moved is controlled so as to be movable in a state in which the underlay (the schedule chart in this example) is fixed. Accordingly, the object can be moved while the underlay is fixed. With such a control, assuming that the schedule chart is likened to a paper, the underlay fixing input portion can be used to press an edge portion of the paper. Thus, a user's intuitive operation can properly be detected, and an object can properly be moved.
  • FIGS. 11A and 11B are diagrams showing still another transition example of the display screen in a case of multiple touches. FIGS. 11A and 11B show an example of screen transition in which only an object is controlled so as to be movable with the schedule chart being fixed when a plurality of contact points are concurrently detected as in the other transition examples of multiple touches. In the example shown in FIGS. 11A and 11B, when a user touches a coordinate area corresponding to the display screen frame 500 and the underlay (the schedule chart in this example) with a finger of a hand holding the mobile terminal 200 at the same time, another object is controlled so as to be movable while the underlay is fixed. A multi-touch screen detectable area, which is an area of the display unit 206 having a contact detection function for detecting contact, is provided not only on the display screen, but also on the display screen frame. Furthermore, as shown in FIG. 12, a contact point continuously extending from the display screen frame 500 to the display screen 509 is deemed contact with the movement-fixing coordinate area.
  • FIG. 11A shows a case of detecting a contact point in the coordinate area corresponding to the display screen frame 500 and the underlay with a finger of a hand holding the mobile terminal 200, detecting a contact point in the second coordinate area in which the object to be moved has been displayed, and further detecting a change of the detection coordinate of only the contact point in the second coordinate area as a scroll instruction.
  • FIG. 11B shows a display result in which the control unit 208 has moved only the object displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 11A.
  • FIG. 12 is an enlarged view of the display screen shown in FIG. 11. As shown in FIG. 12, contact with the movement-fixing coordinate area is detected if a contact point (area) hangs over both of the display screen frame (enclosure) 500 and the display screen 509. Specifically, when the display screen frame (enclosure) 500 and the underlay are brought into contact with the same finger of the user at the same time, it is determined that the operation is for fixing the underlay.
  • As shown in FIGS. 11A, 11B, and 12, when a user touches the display screen frame and the underlay (the schedule chart in this example) with a finger of a hand holding the mobile terminal 200 at the same time, the object to be moved is controlled so as to be movable in a state in which the underlay (the schedule chart in this example) is fixed. Accordingly, the object can be moved while the underlay is fixed. With such a control, assuming that the schedule chart is likened to a paper, the paper can be pressed against the housing of the mobile terminal 200. In other words, the underlay can be sandwiched between the housing and the finger. Therefore, a user's intuitive operation can properly be acquired, and an object can properly be moved.
  • With the control in the illustrated transition examples, when a plurality of contact points are detected, an input item of “Meeting” (second display information) is made movable while the schedule chart (first display information), which is an underlay (background), is fixed (is not allowed to move). Thus, it is possible to detect a user's intuitive pressing operation, detect a movement (scroll) instruction of display information from the user's operation, and move only the input item.
  • With such a control, when an input item is to be moved, it is possible to omit conventional multiple operations, such as selecting the input item, displaying the detailed information, and changing and determining the detailed information. Unlike a conventional method of moving the input item by changing time-specifying information included in conventional detailed information, the input item can be moved while other schedules that have already been inputted are being confirmed. Accordingly, double-booking of schedules can be prevented.
  • In a case where the object being moved reaches the vicinity of an edge of the display screen, if information displayed on the edge (edge of the display screen of the display information such as the background or the underlay) is controlled so as to move around a central portion of the active range or is controlled so as to move such that movement of the object is not inhibited, then a user's intuitive operation is further facilitated. For example, in a case where display information needs a plurality of screen displays (a plurality of pages), the control unit may detect arrival of the object being moved to the vicinity of a lower edge of the display screen and move the information displayed on the edge to the vicinity of a central portion of the active range, or may switch the display information into the next page.
  • From another point of view with regard to the above transition examples, a movement-fixing coordinate area is provided in a first display range in which first display information is displayed and/or a second display range in which second display information is displayed. In a case of detecting a first contact point in the movement-fixing coordinate area, detecting a second contact point in either one of the first coordinate area and the second coordinate area other than the movement-fixing coordinate area, and detecting a change of the detection coordinate of the second contact point, only the first or second display information displayed in the first or second coordinate area in which the second contact point is present can be moved based upon the change of the detection coordinate.
  • In the above examples, the application program is for the schedule chart. However, the first display information (underlay or background) is not limited to a schedule chart. A map, a calendar, a photograph list, and the like may be displayed as the first display information. Furthermore, the application program may be a document input program or a spreadsheet program. In this case, a selected range of input text or an individual cell can be used as an object. In the case where the present invention is applied to a document input program, a selected part of text can be moved as an object provided with a movement instruction in addition to photographs and figures by a user's intuitive operation.
  • Furthermore, in a case where the background display is separated into a plurality of areas, such as a spreadsheet program, only a coordinate area in which a contact point is present may be controlled to be fixed. In such a case, a predetermined area can be compared with another area away from the predetermined area in a state in which those areas are placed side by side.
  • Next, a tablet computer 300 according to a third embodiment of the present invention will be described.
  • FIG. 13 is a functional block diagram showing an overview of a configuration of the tablet computer 300 according to the present embodiment.
  • As shown in FIG. 13, the tablet computer 300 includes a control unit 310 operable to perform a variety of calculations, a RAM 311 operable to temporarily store information, a ROM 312 in which a fundamental control program has been stored, a touch panel display 320 operable to input and output information, a storage unit 330, and a network interface 313 for communication with the Internet or the like via a network.
  • The touch panel display 320 has a display unit 321 and a contact detection unit 322. The touch panel display 320 displays display information outputted from the control unit 310, such as images, figures, characters, and symbols, on a display screen of the display unit 321. The touch panel display 320 is operable to detect an input from a user through contact with the contact detection unit 322.
  • The storage unit 330 stores therein various application programs 331, a driver 332 for the touch panel display 320, various types of data, various types of contents, an OS (Operating System) 333, and the like.
  • The control unit 310 operates in accordance with instructions from the OS 333, the driver 332, and the application programs 331. Software programs such as the OS 333 are expanded in the RAM 311 as needed and used.
  • Furthermore, the contact detection unit 322 is operable to detect contact with hardware of the contact detection unit 322 (multi-touch screen detectable area) based upon one of the OS 333, the driver 332, and the application programs 331, or a combination thereof. The contact detection unit 322 may use any detection methods as long as it can detect a plurality of contact points (areas). Furthermore, the contact detection unit 322 may be implemented by using any software program. The contact detection unit 322 is not specifically limited as long as it can detect contact.
  • With such a configuration, the tablet computer 300 can properly detect a user's intuitive operation with use of the contact detection unit 322 and can properly move an object.
  • Next, a process of identifying an operation of the tablet computer 300 will be described. The identification process will be described in an example of a spreadsheet program with reference to FIGS. 14, 15A, and 15B.
  • FIG. 14 is a flow chart explanatory of a movement operation of the tablet computer 300.
  • Processes from S1401 to S1403 shown in FIG. 14 are the same as the processes from Step S401 to S403 of FIG. 4. Furthermore, processes of Steps S1405 and S1406 are the same as the processes of Steps S405 and S406 of FIG. 4.
  • The control unit 310 of the tablet computer 300 monitors contact with the touch panel display 320 (Step S1401). If any contact is detected, the control unit 310 judges whether there is one contact point or are a plurality of contact points (Step S1402). Furthermore, the control unit 310 detects contact with a display range of another object (another cell) on the display screen other than a display range of an object to be moved (cell to be moved) and judges whether a movement instruction for the object to be moved can be detected (Step S1403).
  • If contact with the display range of the other object and a movement instruction for the object to be moved are detected, then the control unit 310 controls the object to be moved so as to be movable while the other object is fixed (Step S1404).
  • Meanwhile, in a case where there is one contact point or the like, a predetermined normal operation is performed (Steps S1405-S1406).
  • Here, a movement operation of an application program in which the background (underlay or first display information) is separated into a plurality of areas (objects) will be described with use of a transition example of a display screen of a spreadsheet program.
  • FIGS. 15A and 15B are diagrams showing a transition example of a display screen 600 of a spreadsheet program in a case of multiple touches.
  • FIG. 15A shows a case of detecting a plurality of contact points in a first coordinate area in which first display information (table in this example) 601 has been displayed and in a second coordinate area in which second display information (each cell in this example) 602 has been displayed and further detecting a change of a detection coordinate of only the contact point in the second coordinate area as a scroll instruction. That is, FIG. 15A shows an operation from Step S1401 via “Yes” of S1402 to “Yes” of Step S1403 shown in FIG. 14.
  • FIG. 15B shows a display result in which the control unit 310 has scrolled only the second display information (object), which has been displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected, based upon the change of the detection coordinate as a result of the identification process of FIG. 15A. That is, FIG. 15B shows the operation of Step S1404 shown in FIG. 14.
  • In the screen transition, as shown in FIG. 15B, a cell 604 provided with a movement instruction (cell of 7/10 and 9:00) is controlled so as to be movable while a rectangular area 603 with corners of the upper-left cell and the cell provided with no movement instruction (cell to be fixed) is fixed. Furthermore, if the cell 604 controlled so as to be movable is moved near an edge of the active area, then the first display information is displayed in a divided manner. The cell 604 to be moved is controlled so as to be movable while a frame portion of the spreadsheet program is displayed.
  • As shown in FIGS. 15A and 15B, when contact with the active range of the display screen other than the display range of the object provided with the movement instruction is detected, then the object provided with the movement instruction is controlled so as to be movable within the active range. Accordingly, the user can move information while he/she views and compares the cell being moved (7/10 9:00) and other cells away from that cell being moved with each other in the vertical and horizontal directions.
  • In other words, when the control unit 310 detects contact with the coordinate area in which a plurality of pieces of second display information have been displayed in a frame in which the first display information has been displayed and detects a change of the detection coordinate of only one contact point, then the first display information is displayed in a divided manner. Only the second display information displayed in the coordinate area in which the change of the detection coordinate of the contact point has been detected is moved based upon the change of the detection coordinate.
  • As described above, an information processing apparatus of the present invention recognizes contact of a user with a screen at a plurality of points and an operation directed to a displayed object. Movement control of an object is performed based upon the recognition result. Therefore, a burden of a user's input operation can be reduced.
  • Specifically, according to the present invention, there can be provided an information processing apparatus that can properly detect a user's intuitive operation with use of a contact detection unit capable of detecting a plurality of contact points and can properly move an object, a program for implementing such an information processing apparatus, and a computer-readable storage medium in which such a program is recorded.
  • Furthermore, two or more examples of the aforementioned intuitive fixing operations may be used in a combined manner. Such a combination can absorb differences between individual users and further enhance the usability of the users.
  • Moreover, the specific configuration of the present invention is not limited to the above embodiments. Any modifications would be included in the present invention without departing from the spirit of the present invention.
  • This application claims the benefit of priority from Japanese patent application No. 2008-185622, filed on Jul. 17, 2008, the disclosure of which is incorporated herein in its entirety by reference.
  • INDUSTRIAL APPLICABILITY
  • The present invention is applicable to any information processing apparatus having a contact detection unit capable of detecting a plurality of contact points. Furthermore, the present invention is suitable for an information processing apparatus held by a single hand to operate.

Claims (22)

1. An information processing apparatus comprising:
a display unit operable to display an object;
a contact detection unit operable to detect a plurality of contact points on a display screen of the display unit; and
a control unit operable to control the object displayed on the display unit so as to be movable with use of the contact detection unit,
wherein, in a case of detecting a first contact point in an active range of the display screen other than a display range of the object with use of the contact detection unit, detecting a second contact point in the display range of the object, and further detecting a movement instruction to the object, the control unit controls the object so as to be movable within the active range.
2. An information processing apparatus comprising:
a display unit operable to display first display information displayed in a first display range and second display information displayed in a second display range within the first display range;
a contact detection unit operable to detect contact points in a first coordinate area in which the first display information is displayed and in a second coordinate area in which the second display information is displayed and detect coordinates of the contact points; and
a control unit configured such that, in a case of detecting no contact point in one of the first coordinate area and the second coordinate area but detecting a contact point in the other coordinate area and detecting a change of a coordinate of the detected contact point, the control unit moves both of the first and second display information based upon the change of the coordinate, and
in a case of detecting a plurality of contact points in the first coordinate area and the second coordinate area and detecting a change of a coordinate of only one of the plurality of contact points, the control unit moves the first or second display information displayed in the first or second coordinate area in which the change of the coordinate of the contact point has been detected, based upon the change of the coordinate.
3. The information processing apparatus as recited in claim 2, wherein:
the control unit provides a movement-fixing coordinate area in the first display range, and
in a case of detecting a first contact point in the movement-fixing coordinate area, a second contact point in one of the first coordinate area and the second coordinate area other than the movement-fixing coordinate area, and a change of a coordinate of the second contact point, the control unit moves the first or second display information displayed in the first or second coordinate area in which the second contact point is present, based upon the change of the coordinate.
4. The information processing apparatus as recited in claim 2, wherein:
the control unit displays, as the second display information, one of an icon, a button, a character string, a text box, and area information visually indicated for providing a command to the control unit, or a combination thereof on the display unit.
5. The information processing apparatus as recited in claim 2, wherein:
the control unit displays the first display information and the second display information as combined information.
6. The information processing apparatus as recited in claim 2, wherein:
the control unit sets the second display information as a visible object displayed on an upper layer of the first display information or a visible object displayed such that the second display information is overlaid on the first display information.
7. A computer-readable storage medium having a program recorded thereon for executing a procedure with a control unit, the procedure comprising:
detecting a first contact point in an active range on a display screen other than a display range of an object displayed on a display unit with use of a contact detection unit;
detecting a second contact point in the display range of the object with use of the contact detection unit; and
controlling the object so as to be movable within the active range when a movement instruction to the object is detected.
8. A computer-readable storage medium having a program recorded thereon for executing a procedure with a control unit, the procedure comprising:
detecting contact points in a first coordinate area of a first display range in which first display information is displayed and in a second coordinate area of a second display range in which second display information is displayed within the first display range and detecting coordinates of the contact points; and
in a case of detecting no contact point in one of the first coordinate area and the second coordinate area but detecting a contact point in the other coordinate area, and detecting a change of a coordinate of the detected contact point, moving both of the first and second display information based upon the change of the coordinate, and
in a case of detecting a plurality of contact points in the first coordinate area and the second coordinate area and detecting a change of a coordinate of only one of the plurality of contact points, moving the first or second display information displayed in the first or second coordinate area in which the change of the coordinate of the contact point has been detected, based upon the change of the coordinate.
9. The computer-readable storage medium having a program recorded thereon for executing a procedure with a control unit as recited in claim 8, the procedure comprising:
providing a movement-fixing coordinate area in the first display range, and
in a case of detecting a first contact point in the movement-fixing coordinate area, a second contact point in one of the first coordinate area and the second coordinate area other than the movement-fixing coordinate area, and a change of a coordinate of the second contact point, moving the first or second display information displayed in the first or second coordinate area in which the second contact point is present, based upon the change of the coordinate.
10. The computer-readable storage medium having a program recorded thereon for executing a procedure with a control unit as recited in claim 8, the procedure comprising:
displaying, as the second display information, one of an icon, a button, a character string, a text box, and area information visually indicated for providing a command to the control unit, or a combination thereof on the display unit.
11. The computer-readable storage medium having a program recorded thereon for executing a procedure with a control unit as recited in claim 8, the procedure comprising:
displaying the first display information and the second display information as combined information.
12. The computer-readable storage medium having a program recorded thereon for executing a procedure with a control unit as recited in claim 8, the procedure comprising:
setting the second display information as a visible object displayed on an upper layer of the first display information or a visible object displayed such that the second display information is overlaid on the first display information.
13. The computer-readable storage medium having a program recorded thereon as recited in claim 8, wherein:
the program comprises one of a driver for the contact detection unit, an operating system, and an application program, or a combination thereof.
14. An object movement method comprising:
detecting a first contact point in an active range on a display screen other than a display range of an object displayed on a display unit with use of a contact detection unit;
detecting a second contact point in the display range of the object with use of the contact detection unit; and
controlling the object so as to be movable within the active range when a movement instruction to the object is detected.
15. An object movement method comprising:
detecting contact points in a first coordinate area of a first display range in which first display information is displayed and in a second coordinate area of a second display range in which second display information is displayed within the first display range and detecting coordinates of the contact points; and
in a case of detecting no contact point in one of the first coordinate area and the second coordinate area but detecting a contact point in the other coordinate area, and detecting a change of a coordinate of the detected contact point, moving both of the first and second display information based upon the change of the coordinate, and
in a case of detecting a plurality of contact points in the first coordinate area and the second coordinate area and detecting a change of a coordinate of only one of the plurality of contact points, moving the first or second display information displayed in the first or second coordinate area in which the change of the coordinate of the contact point has been detected, based upon the change of the coordinate.
16. The object movement method as recited in claim 15, further comprising:
providing a movement-fixing coordinate area in the first display range, and
in a case of detecting a first contact point in the movement-fixing coordinate area, a second contact point in one of the first coordinate area and the second coordinate area other than the movement-fixing coordinate area, and a change of a coordinate of the second contact point, moving the first or second display information displayed in the first or second coordinate area in which the second contact point is present, based upon the change of the coordinate.
17. The object movement method as recited in claim 15, further comprising:
displaying, as the second display information, one of an icon, a button, a character string, a text box, and area information visually indicated for providing a command to a control unit, or a combination thereof on the display unit.
18. The object movement method as recited in claim 15, further comprising:
displaying the first display information and the second display information as combined information.
19. The object movement method as recited in claim 15, further comprising:
setting the second display information as a visible object displayed on an upper layer of the first display information or a visible object displayed such that the second display information is overlaid on the first display information.
20. The object movement method as recited in claim 15, wherein:
the object movement method is implemented by one of a driver for a contact detection unit, an operating system, and an application program, or a combination thereof.
21. The computer-readable storage medium having a program recorded thereon as recited in claim 7, wherein:
the program comprises one of a driver for the contact detection unit, an operating system, and an application program, or a combination thereof.
22. The object movement method as recited in claim 14, wherein:
the object movement method is implemented by one of a driver for a contact detection unit, an operating system, and an application program, or a combination thereof.
US13/054,698 2008-07-17 2009-07-15 Information processing apparatus, storage medium having program recorded thereon, and object movement method Abandoned US20110126097A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP2008-185622 2008-07-17
JP2008185622 2008-07-17
PCT/JP2009/063135 WO2010008088A1 (en) 2008-07-17 2009-07-15 Information processing apparatus, storage medium on which program has been recorded, and object shifting method

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2009/063135 A-371-Of-International WO2010008088A1 (en) 2008-07-17 2009-07-15 Information processing apparatus, storage medium on which program has been recorded, and object shifting method

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/338,840 Continuation US9933932B2 (en) 2008-07-17 2014-07-23 Information processing apparatus having a contact detection unit capable of detecting a plurality of contact points, storage medium having program recorded thereon, and object movement method

Publications (1)

Publication Number Publication Date
US20110126097A1 true US20110126097A1 (en) 2011-05-26

Family

ID=41550488

Family Applications (3)

Application Number Title Priority Date Filing Date
US13/054,698 Abandoned US20110126097A1 (en) 2008-07-17 2009-07-15 Information processing apparatus, storage medium having program recorded thereon, and object movement method
US14/338,840 Active 2029-12-06 US9933932B2 (en) 2008-07-17 2014-07-23 Information processing apparatus having a contact detection unit capable of detecting a plurality of contact points, storage medium having program recorded thereon, and object movement method
US15/902,129 Active 2029-12-05 US10656824B2 (en) 2008-07-17 2018-02-22 Information processing apparatus having a contact detection unit capable of detecting a plurality of contact points, storage medium having program recorded thereon, and object movement method

Family Applications After (2)

Application Number Title Priority Date Filing Date
US14/338,840 Active 2029-12-06 US9933932B2 (en) 2008-07-17 2014-07-23 Information processing apparatus having a contact detection unit capable of detecting a plurality of contact points, storage medium having program recorded thereon, and object movement method
US15/902,129 Active 2029-12-05 US10656824B2 (en) 2008-07-17 2018-02-22 Information processing apparatus having a contact detection unit capable of detecting a plurality of contact points, storage medium having program recorded thereon, and object movement method

Country Status (5)

Country Link
US (3) US20110126097A1 (en)
EP (1) EP2306286A4 (en)
JP (6) JP5267827B2 (en)
CN (2) CN104216655B (en)
WO (1) WO2010008088A1 (en)

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100053099A1 (en) * 2008-06-26 2010-03-04 Cirque Corporation Method for reducing latency when using multi-touch gesture on touchpad
US20100306245A1 (en) * 2007-05-07 2010-12-02 Toyota Jidosha Kabushiki Kaisha Navigation system
US20110216095A1 (en) * 2010-03-04 2011-09-08 Tobias Rydenhag Methods, Devices, and Computer Program Products Providing Multi-Touch Drag and Drop Operations for Touch-Sensitive User Interfaces
US20130009904A1 (en) * 2010-03-26 2013-01-10 Kyocera Corporation Portable terminal, display control program and display control method
KR20130070090A (en) * 2011-12-19 2013-06-27 삼성전자주식회사 Method and apparatus for providing multi-touch interaction in portable device
US20130170324A1 (en) * 2012-01-04 2013-07-04 Asustek Computer Inc. Operating method and portable electronic device using the same
US20130203468A1 (en) * 2012-02-07 2013-08-08 Research In Motion Limited Methods and devices for merging contact records
US20140035845A1 (en) * 2012-08-01 2014-02-06 Sony Corporation Display control apparatus, display control method, and computer program
US20140047390A1 (en) * 2012-02-24 2014-02-13 Blackberry Limited Method and apparatus for providing a user interface on a device enabling selection of operations to be performed in relation to content
US20140063321A1 (en) * 2012-08-29 2014-03-06 Canon Kabushiki Kaisha Display control apparatus having touch panel function, display control method, and storage medium
US20140225826A1 (en) * 2011-09-07 2014-08-14 Nitto Denko Corporation Method for detecting motion of input body and input device using same
US20150070284A1 (en) * 2013-09-09 2015-03-12 Samsung Electronics Co. Ltd. Method for differentiation of touch input and visualization of pending touch input
US20150212682A1 (en) * 2014-01-30 2015-07-30 Accompani, Inc. Managing calendar and contact information
US9223483B2 (en) 2012-02-24 2015-12-29 Blackberry Limited Method and apparatus for providing a user interface on a device that indicates content operators
US9250800B2 (en) 2010-02-18 2016-02-02 Rohm Co., Ltd. Touch-panel input device
US9501213B2 (en) * 2011-09-16 2016-11-22 Skadool, Inc. Scheduling events on an electronic calendar utilizing fixed-positioned events and a draggable calendar grid
US20200145361A1 (en) * 2014-09-02 2020-05-07 Apple Inc. Electronic message user interface
US20230230044A1 (en) * 2021-12-30 2023-07-20 Microsoft Technology Licensing, Llc Calendar update using template selections
US11829576B2 (en) 2013-09-03 2023-11-28 Apple Inc. User interface object manipulations in a user interface
US11921926B2 (en) 2018-09-11 2024-03-05 Apple Inc. Content-based tactile outputs
US11936607B2 (en) 2008-03-04 2024-03-19 Apple Inc. Portable multifunction device, method, and graphical user interface for an email client
US11941191B2 (en) 2014-09-02 2024-03-26 Apple Inc. Button functionality

Families Citing this family (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2306286A4 (en) 2008-07-17 2016-05-11 Nec Corp Information processing apparatus, storage medium on which program has been recorded, and object shifting method
JP5501509B2 (en) * 2008-09-03 2014-05-21 キヤノン株式会社 Information processing apparatus, operation method thereof, and program
JP2011227703A (en) * 2010-04-20 2011-11-10 Rohm Co Ltd Touch panel input device capable of two-point detection
JP2011197848A (en) * 2010-03-18 2011-10-06 Rohm Co Ltd Touch-panel input device
US8707174B2 (en) * 2010-02-25 2014-04-22 Microsoft Corporation Multi-screen hold and page-flip gesture
US9454304B2 (en) 2010-02-25 2016-09-27 Microsoft Technology Licensing, Llc Multi-screen dual tap gesture
JP5230684B2 (en) * 2010-05-13 2013-07-10 パナソニック株式会社 Electronic device, display method, and program
JP5418440B2 (en) * 2010-08-13 2014-02-19 カシオ計算機株式会社 Input device and program
JP5615642B2 (en) * 2010-09-22 2014-10-29 京セラ株式会社 Portable terminal, input control program, and input control method
KR102006740B1 (en) 2010-10-20 2019-08-02 삼성전자 주식회사 Method and apparatus for displaying screen in mobile terminal
JP2012247861A (en) * 2011-05-25 2012-12-13 Panasonic Corp Touch screen device, touch operation input method, and program
US8810535B2 (en) 2011-10-18 2014-08-19 Blackberry Limited Electronic device and method of controlling same
EP2584441A1 (en) * 2011-10-18 2013-04-24 Research In Motion Limited Electronic device and method of controlling same
JP5974657B2 (en) * 2012-06-15 2016-08-23 株式会社リコー Information processing apparatus, information processing method, and information processing program
JP5638570B2 (en) * 2012-06-22 2014-12-10 ヤフー株式会社 Image display device, image display method, and image display program
JP2012234569A (en) * 2012-08-09 2012-11-29 Panasonic Corp Electronic apparatus, display method and program
JP2014112335A (en) * 2012-12-05 2014-06-19 Fuji Xerox Co Ltd Information processing device and program
JP2014142750A (en) * 2013-01-23 2014-08-07 Dainippon Printing Co Ltd Ic card having input function and display function
JP6154184B2 (en) * 2013-04-26 2017-06-28 株式会社サミーネットワークス Display control method, display control program, and portable information terminal
WO2015107617A1 (en) * 2014-01-14 2015-07-23 株式会社 東芝 Electronic device, control method and program
JP2015172861A (en) * 2014-03-12 2015-10-01 レノボ・シンガポール・プライベート・リミテッド Method for switching use environment of portable electronic apparatus, portable electronic apparatus, and computer program
WO2016006074A1 (en) * 2014-07-09 2016-01-14 株式会社東芝 Electronic device, method, and program
JP7037306B2 (en) 2016-09-16 2022-03-16 日東電工株式会社 Spiral type membrane element
JP6249073B2 (en) * 2016-09-28 2017-12-20 富士ゼロックス株式会社 Information processing apparatus and program
JP7327368B2 (en) * 2020-12-02 2023-08-16 横河電機株式会社 Apparatus, method and program

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5732227A (en) * 1994-07-05 1998-03-24 Hitachi, Ltd. Interactive information processing system responsive to user manipulation of physical objects and displayed images
US5825352A (en) * 1996-01-04 1998-10-20 Logitech, Inc. Multiple fingers contact sensing method for emulating mouse buttons and mouse operations on a touch sensor pad
US20010024195A1 (en) * 2000-03-21 2001-09-27 Keisuke Hayakawa Page information display method and device and storage medium storing program for displaying page information
US6958749B1 (en) * 1999-11-04 2005-10-25 Sony Corporation Apparatus and method for manipulating a touch-sensitive display panel
US20060001652A1 (en) * 2004-07-05 2006-01-05 Yen-Chang Chiu Method for scroll bar control on a touchpad
US20060026521A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Gestures for touch sensitive input devices
US20060164399A1 (en) * 2005-01-21 2006-07-27 Cheston Richard W Touchpad diagonal scrolling
US20070177803A1 (en) * 2006-01-30 2007-08-02 Apple Computer, Inc Multi-touch gesture dictionary
US20070277126A1 (en) * 2006-05-24 2007-11-29 Ho Joo Park Touch screen device and method of selecting files thereon
US20080165141A1 (en) * 2007-01-05 2008-07-10 Apple Inc. Gestures for controlling, manipulating, and editing of media files using touch sensitive devices
US20080297482A1 (en) * 2007-05-30 2008-12-04 Microsoft Corporation Recognizing selection regions from multiple simultaneous inputs
US20090271723A1 (en) * 2008-04-24 2009-10-29 Nintendo Co., Ltd. Object display order changing program and apparatus
US20100007623A1 (en) * 2008-07-11 2010-01-14 Canon Kabushiki Kaisha Information processing apparatus and method
US20110047459A1 (en) * 2007-10-08 2011-02-24 Willem Morkel Van Der Westhuizen User interface
US8209628B1 (en) * 2008-04-11 2012-06-26 Perceptive Pixel, Inc. Pressure-sensitive manipulation of displayed objects

Family Cites Families (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH03180922A (en) 1989-12-11 1991-08-06 Fujitsu Ltd Touch panel structure
JPH07175587A (en) * 1993-10-28 1995-07-14 Hitachi Ltd Information processor
JPH1173271A (en) * 1997-08-28 1999-03-16 Sharp Corp Instructing device and processor and storage medium
JPH11102274A (en) 1997-09-25 1999-04-13 Nec Corp Scroll device
US20070177804A1 (en) * 2006-01-30 2007-08-02 Apple Computer, Inc. Multi-touch gesture dictionary
JP2000137564A (en) * 1998-11-02 2000-05-16 Pioneer Electronic Corp Picture operating device and its method
JP4542637B2 (en) 1998-11-25 2010-09-15 セイコーエプソン株式会社 Portable information device and information storage medium
JP2000163444A (en) 1998-11-25 2000-06-16 Seiko Epson Corp Portable information device and information storage medium
JP2000163193A (en) * 1998-11-25 2000-06-16 Seiko Epson Corp Portable information equipment and information storage medium
US7003641B2 (en) * 2000-01-31 2006-02-21 Commvault Systems, Inc. Logical view with granular access to exchange data managed by a modular data and storage management system
JP4803883B2 (en) * 2000-01-31 2011-10-26 キヤノン株式会社 Position information processing apparatus and method and program thereof.
JP4686886B2 (en) * 2001-04-06 2011-05-25 ソニー株式会社 Information processing device
US20030117427A1 (en) * 2001-07-13 2003-06-26 Universal Electronics Inc. System and method for interacting with a program guide displayed on a portable electronic device
JP2003173237A (en) 2001-09-28 2003-06-20 Ricoh Co Ltd Information input-output system, program and storage medium
JP3847641B2 (en) 2002-02-28 2006-11-22 株式会社ソニー・コンピュータエンタテインメント Information processing apparatus, information processing program, computer-readable recording medium storing information processing program, and information processing method
US7411575B2 (en) 2003-09-16 2008-08-12 Smart Technologies Ulc Gesture recognition method and touch system incorporating the same
JP4586397B2 (en) * 2004-04-08 2010-11-24 ソニー株式会社 Information processing apparatus and method, and program
JP4045550B2 (en) * 2004-06-28 2008-02-13 富士フイルム株式会社 Image display control apparatus and image display control program
KR100984596B1 (en) * 2004-07-30 2010-09-30 애플 인크. Gestures for touch sensitive input devices
JP2006092025A (en) * 2004-09-21 2006-04-06 Fujitsu Ltd Electronic device, process for control of display screen, and control processing program for display screen
JP2007141177A (en) * 2005-11-22 2007-06-07 Tokai Rika Co Ltd Operation input device
JP3970906B2 (en) 2006-02-03 2007-09-05 株式会社ソニー・コンピュータエンタテインメント Information processing apparatus, information processing program, computer-readable recording medium storing information processing program, and information processing method
JP4810658B2 (en) * 2006-03-10 2011-11-09 国立大学法人広島大学 Contact detection device and contact detection method
JP5210497B2 (en) 2006-04-12 2013-06-12 クラリオン株式会社 Navigation device
JP2008097172A (en) * 2006-10-10 2008-04-24 Sony Corp Display and display method
JP2008185622A (en) 2007-01-26 2008-08-14 Ricoh Co Ltd Color shift detection method and device, and image forming apparatus
JP2009017286A (en) 2007-07-05 2009-01-22 Niigata Seimitsu Kk Am/fm radio receiver and semiconductor integrated circuit for reception used therefor
US9513765B2 (en) * 2007-12-07 2016-12-06 Sony Corporation Three-dimensional sliding object arrangement method and system
EP2306286A4 (en) 2008-07-17 2016-05-11 Nec Corp Information processing apparatus, storage medium on which program has been recorded, and object shifting method
US8407606B1 (en) * 2009-01-02 2013-03-26 Perceptive Pixel Inc. Allocating control among inputs concurrently engaging an object displayed on a multi-touch device
US20120242620A1 (en) * 2011-03-22 2012-09-27 Research In Motion Limited Combined optical navigation and button
US9134893B2 (en) * 2012-12-14 2015-09-15 Barnes & Noble College Booksellers, Llc Block-based content selecting technique for touch screen UI
US9471150B1 (en) * 2013-09-27 2016-10-18 Emc Corporation Optimized gestures for zoom functionality on touch-based device

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5732227A (en) * 1994-07-05 1998-03-24 Hitachi, Ltd. Interactive information processing system responsive to user manipulation of physical objects and displayed images
US5825352A (en) * 1996-01-04 1998-10-20 Logitech, Inc. Multiple fingers contact sensing method for emulating mouse buttons and mouse operations on a touch sensor pad
US6958749B1 (en) * 1999-11-04 2005-10-25 Sony Corporation Apparatus and method for manipulating a touch-sensitive display panel
US20010024195A1 (en) * 2000-03-21 2001-09-27 Keisuke Hayakawa Page information display method and device and storage medium storing program for displaying page information
US20040125081A1 (en) * 2000-03-21 2004-07-01 Nec Corporation Page information display method and device and storage medium storing program for displaying page information
US6765559B2 (en) * 2000-03-21 2004-07-20 Nec Corporation Page information display method and device and storage medium storing program for displaying page information
US20060001652A1 (en) * 2004-07-05 2006-01-05 Yen-Chang Chiu Method for scroll bar control on a touchpad
US20060026521A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Gestures for touch sensitive input devices
US20060164399A1 (en) * 2005-01-21 2006-07-27 Cheston Richard W Touchpad diagonal scrolling
US20070177803A1 (en) * 2006-01-30 2007-08-02 Apple Computer, Inc Multi-touch gesture dictionary
US20070277126A1 (en) * 2006-05-24 2007-11-29 Ho Joo Park Touch screen device and method of selecting files thereon
US20080165141A1 (en) * 2007-01-05 2008-07-10 Apple Inc. Gestures for controlling, manipulating, and editing of media files using touch sensitive devices
US20080297482A1 (en) * 2007-05-30 2008-12-04 Microsoft Corporation Recognizing selection regions from multiple simultaneous inputs
US20110047459A1 (en) * 2007-10-08 2011-02-24 Willem Morkel Van Der Westhuizen User interface
US8209628B1 (en) * 2008-04-11 2012-06-26 Perceptive Pixel, Inc. Pressure-sensitive manipulation of displayed objects
US20090271723A1 (en) * 2008-04-24 2009-10-29 Nintendo Co., Ltd. Object display order changing program and apparatus
US20100007623A1 (en) * 2008-07-11 2010-01-14 Canon Kabushiki Kaisha Information processing apparatus and method
US8253707B2 (en) * 2008-07-11 2012-08-28 Canon Kabushiki Kaisha Information processing apparatus and method
US20120281016A1 (en) * 2008-07-11 2012-11-08 Canon Kabushiki Kaisha Information processing apparatus and method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Jefferson et al., EXTENDING 2D OBJECT ARRANGEMENT WITH PRESSURE-SENSITIVE LAYERING CUES , 04/11/2008, provisional USPTO application SN 61044109 *
Van Der Westhuizen, User Interface, 10/08/2007, Provisional USPTO application SN 60978327 *

Cited By (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100306245A1 (en) * 2007-05-07 2010-12-02 Toyota Jidosha Kabushiki Kaisha Navigation system
US8583676B2 (en) * 2007-05-07 2013-11-12 Toyota Jidosha Kabushiki Kaisha Navigation system
US11936607B2 (en) 2008-03-04 2024-03-19 Apple Inc. Portable multifunction device, method, and graphical user interface for an email client
US8368667B2 (en) * 2008-06-26 2013-02-05 Cirque Corporation Method for reducing latency when using multi-touch gesture on touchpad
US20100053099A1 (en) * 2008-06-26 2010-03-04 Cirque Corporation Method for reducing latency when using multi-touch gesture on touchpad
US9250800B2 (en) 2010-02-18 2016-02-02 Rohm Co., Ltd. Touch-panel input device
US9760280B2 (en) 2010-02-18 2017-09-12 Rohm Co., Ltd. Touch-panel input device
US20110216095A1 (en) * 2010-03-04 2011-09-08 Tobias Rydenhag Methods, Devices, and Computer Program Products Providing Multi-Touch Drag and Drop Operations for Touch-Sensitive User Interfaces
US20130009904A1 (en) * 2010-03-26 2013-01-10 Kyocera Corporation Portable terminal, display control program and display control method
US8941611B2 (en) * 2010-03-26 2015-01-27 Kyocera Corporation Portable terminal, display control program and display control method
US20140225826A1 (en) * 2011-09-07 2014-08-14 Nitto Denko Corporation Method for detecting motion of input body and input device using same
US9501213B2 (en) * 2011-09-16 2016-11-22 Skadool, Inc. Scheduling events on an electronic calendar utilizing fixed-positioned events and a draggable calendar grid
EP2608007A3 (en) * 2011-12-19 2013-09-25 Samsung Electronics Co., Ltd Method and apparatus for providing a multi-touch interaction in a portable terminal
KR101710418B1 (en) * 2011-12-19 2017-02-28 삼성전자주식회사 Method and apparatus for providing multi-touch interaction in portable device
KR20130070090A (en) * 2011-12-19 2013-06-27 삼성전자주식회사 Method and apparatus for providing multi-touch interaction in portable device
US20130170324A1 (en) * 2012-01-04 2013-07-04 Asustek Computer Inc. Operating method and portable electronic device using the same
US20130203468A1 (en) * 2012-02-07 2013-08-08 Research In Motion Limited Methods and devices for merging contact records
US9223483B2 (en) 2012-02-24 2015-12-29 Blackberry Limited Method and apparatus for providing a user interface on a device that indicates content operators
US10698567B2 (en) 2012-02-24 2020-06-30 Blackberry Limited Method and apparatus for providing a user interface on a device that indicates content operators
US9753611B2 (en) * 2012-02-24 2017-09-05 Blackberry Limited Method and apparatus for providing a user interface on a device enabling selection of operations to be performed in relation to content
US20140047390A1 (en) * 2012-02-24 2014-02-13 Blackberry Limited Method and apparatus for providing a user interface on a device enabling selection of operations to be performed in relation to content
US10936153B2 (en) 2012-02-24 2021-03-02 Blackberry Limited Method and apparatus for providing a user interface on a device enabling selection of operations to be performed in relation to content
US20140035845A1 (en) * 2012-08-01 2014-02-06 Sony Corporation Display control apparatus, display control method, and computer program
US9798462B2 (en) * 2012-08-01 2017-10-24 Sony Corporation Display control apparatus, display control method, and computer program
US9313406B2 (en) * 2012-08-29 2016-04-12 Canon Kabushiki Kaisha Display control apparatus having touch panel function, display control method, and storage medium
CN103677621A (en) * 2012-08-29 2014-03-26 佳能株式会社 Display control apparatus having touch panel function and display control method
US20140063321A1 (en) * 2012-08-29 2014-03-06 Canon Kabushiki Kaisha Display control apparatus having touch panel function, display control method, and storage medium
US11829576B2 (en) 2013-09-03 2023-11-28 Apple Inc. User interface object manipulations in a user interface
US9841815B2 (en) * 2013-09-09 2017-12-12 Samsung Electronics Co., Ltd. Method for differentiation of touch input and visualization of pending touch input
US20150070284A1 (en) * 2013-09-09 2015-03-12 Samsung Electronics Co. Ltd. Method for differentiation of touch input and visualization of pending touch input
US20150212682A1 (en) * 2014-01-30 2015-07-30 Accompani, Inc. Managing calendar and contact information
US20200145361A1 (en) * 2014-09-02 2020-05-07 Apple Inc. Electronic message user interface
US11743221B2 (en) * 2014-09-02 2023-08-29 Apple Inc. Electronic message user interface
US11941191B2 (en) 2014-09-02 2024-03-26 Apple Inc. Button functionality
US11921926B2 (en) 2018-09-11 2024-03-05 Apple Inc. Content-based tactile outputs
US20230230044A1 (en) * 2021-12-30 2023-07-20 Microsoft Technology Licensing, Llc Calendar update using template selections

Also Published As

Publication number Publication date
US20180181278A1 (en) 2018-06-28
CN104216655A (en) 2014-12-17
JP2016026352A (en) 2016-02-12
CN102099775B (en) 2014-09-24
CN104216655B (en) 2018-02-16
EP2306286A4 (en) 2016-05-11
JPWO2010008088A1 (en) 2012-01-05
JP5811381B2 (en) 2015-11-11
JP6150082B2 (en) 2017-06-21
JP2015015045A (en) 2015-01-22
WO2010008088A1 (en) 2010-01-21
US10656824B2 (en) 2020-05-19
JP2017168136A (en) 2017-09-21
JP5618106B2 (en) 2014-11-05
EP2306286A1 (en) 2011-04-06
JP2013175228A (en) 2013-09-05
JP5267827B2 (en) 2013-08-21
US20140333580A1 (en) 2014-11-13
JP5787375B2 (en) 2015-09-30
JP2013214308A (en) 2013-10-17
CN102099775A (en) 2011-06-15
JP6369704B2 (en) 2018-08-08
US9933932B2 (en) 2018-04-03

Similar Documents

Publication Publication Date Title
US10656824B2 (en) Information processing apparatus having a contact detection unit capable of detecting a plurality of contact points, storage medium having program recorded thereon, and object movement method
US10521111B2 (en) Electronic apparatus and method for displaying a plurality of images in a plurality of areas of a display
CN108701001B (en) Method for displaying graphical user interface and electronic equipment
KR101510484B1 (en) Mobile Terminal And Method Of Controlling Mobile Terminal
US8825113B2 (en) Portable terminal and driving method of the same
KR101853057B1 (en) Mobile Terminal And Method Of Controlling The Same
EP2667293A2 (en) Mobile terminal and control method thereof
US8819552B2 (en) Method of executing function on standby screen of mobile terminal
KR101688134B1 (en) Mobile terminal having a side touch input device and method for executingfunctions of thereof
EP2284675A2 (en) Method for displaying data and mobile terminal thereof
KR101370173B1 (en) I/O Controller and I/O Controlling Method and Mobile communication terminal
KR20170140702A (en) Mobile terminal
WO2012160867A1 (en) Portable terminal, control method and program
KR101448641B1 (en) Mobile terminal and method of executing function therein
KR101486348B1 (en) Mobile terminal and method of displaying screen therein
KR20170130417A (en) Mobile terminal and control method thereof
JP5510008B2 (en) Mobile terminal device
KR101583802B1 (en) Mobile Terminal And Method Of Controlling Mobile Terminal
KR20150102418A (en) Mobile terminal and controlling method thereof
KR101574116B1 (en) Mobile Terminal And Method Of Controlling Audio Using Same
KR20100038930A (en) Mobile terminal and control method of the same
JP2012238267A (en) Processor, processing method, and program

Legal Events

Date Code Title Description
AS Assignment

Owner name: NEC CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ISONO, KAZUKI;REEL/FRAME:025654/0736

Effective date: 20110111

STCB Information on status: application discontinuation

Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION