INFORMATION PROCESSING DEVICE AND TOUCH OPERATION DETECTION METHOD
An information processing device includes: a display unit having a touch panel on its front surface; a movement amount calculation unit that calculates the movement amount of a touch operation based on a touch point at which the touch operation is performed with respect to the touch panel and a touch release point at which the touch operation is released from the touch panel; an operation determination unit that determines whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and a command recognition unit that recognizes whether a received command is a command corresponding to the depression operation or the gesture operation.
Latest Sony Corporation Patents:
- IMAGE PROCESSING APPARATUS AND METHOD
- Terminal device, information processing device, object identifying method, program, and object identifying system
- Drive circuit, electronic apparatus, and method of controlling drive circuit
- Overcurrent determination circuit and light emission control apparatus
- Non-zero random backoff procedure
The present invention contains subject matter related to Japanese Patent Application JP2007-330849 filed in the Japanese Patent Office on Dec. 21, 2007, the entire contents of which being incorporated herein by reference.
BACKGROUND OF THE INVENTION1. Field of the Invention
The present invention relates to an information processing device and a touch operation detection method, which are suitably applied to, e.g., a portable navigation device.
2. Description of the Related Art
A portable navigation device (hereinafter, referred to as “PND”) is designed to be detachably attached to the dashboard of a vehicle via a cradle.
The PND of such a type serves as a vehicle navigation device when being attached to the dashboard via a cradle and serves as a personal navigation device when being detached from the dashboard.
This type of PND aims more for portability than remote-controllability. Thus, a remote controller is not provided for the PND but a user interface that receives a command from a user via a touch panel provided on the front surface of a liquid crystal display is adopted.
There is also known an electric book display control device equipped with a display integrated tablet on which a user uses his or her finger to perform a pointing operation to realize a page-turning operation (refer to, e.g., Jpn. Pat. Appln. Laid-Open Publication No. 2004-348755).
SUMMARY OF THE INVENTIONThe PND having such a configuration displays a plurality of destination candidates so as to allow a user to select his or her destination. If the destination candidates exist over a plurality of pages, the user needs to perform a page-turning operation in order to determine the destination.
In such a case, where the PND performs a page-turning operation in response to the user's pointing operation using his or her finger in the same manner as in the case of the abovementioned electronic book display control device, the PND often erroneously detects that a depression operation with respect to the candidate destination has been made at the moment when the user falsely touched one of the destination candidates with his or her finger, not a page-turning button. Thus, there is a possibility that the PND cannot correctly reflect the user's intention to perform a pointing operation.
The present invention has been made in view of the above points and an object thereof is to propose an information processing device and a touch operation detection method capable of correctly detect a command issued in response to a user's touch operation.
To solve the above problem, according to an aspect of the present invention, there is provided an information processing device including: a display unit having a touch panel on its front surface; a movement amount calculation unit for calculating a movement amount of a touch operation based on a touch point at which the touch operation is performed with respect to the touch panel and a touch release point at which the touch operation is released from the touch panel; an operation determination unit for determining whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and a command recognition unit for recognizing whether a received command is a command corresponding to the depression operation or the gesture operation.
With the above configuration, the information processing device determines whether a touch operation is a button touch operation or a gesture operation depending on the movement amount of the touch operation, thereby correctly recognizing whether a received command is a command corresponding to the depression operation or the gesture operation.
According to another aspect of the present invention, there is provided a touch operation detection method including: a touch point detection step in which a touch point detection unit detects a touch point at which a touch operation is performed with respect to a touch panel provided on the front surface of a display unit; a touch release point detection step following the touch point detection step, in which a touch release point detection unit detects a touch release point at which the touch operation is released from the touch panel; a movement amount calculation step in which a movement amount calculation unit calculates a movement amount of the touch operation based on the touch point and touch release point; an operation determination step in which an operation determination unit determines whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and a command recognition step in which a command recognition unit recognizes whether a received command is a command corresponding to the depression operation or the gesture operation.
With the above configuration, the touch operation detection method determines whether a touch operation is a button touch operation or a gesture operation depending on the movement amount of the touch operation, thereby correctly recognizing whether a received command is a command corresponding to the depression operation or the gesture operation.
According to the present invention, whether a touch operation is a button touch operation or a gesture operation is determined depending on the movement amount of the touch operation, allowing correct determination on whether a received command is a command corresponding to the depression operation or the gesture operation. Therefore, an information processing device and a touch operation detection method capable of correctly recognizing a command corresponding to a user's touch operation can be realized.
The nature, principle and utility of the invention will become more apparent from the following detailed description when read in conjunction with the accompanying drawings in which like parts are designated by like reference numerals or characters.
In the accompanying drawings:
An embodiment of the present invention will be described in detail below with reference to the accompanying drawings.
(1) Outer Appearance of PNDIn
The PND 1 is held by a cradle 3 attached to a dashboard of a vehicle by a sucker 3A and is electrically connected thereto. In this state, the PND operates by a power supplied from a battery of the vehicle via the cradle 3. Further, when removed from the cradle 3, the PND 1 can operate by a power supplied from a battery incorporated therein.
(2) Circuit Configuration of PNDAs shown in
Concretely, the PND 1 uses a GPS module 12 to demodulate satellite signals S1 from a plurality of GPS satellites received via a GPS antenna 11 and, based on orbital data obtained as a result of the demodulation and distance data between the GPS satellites and vehicle, measures the current position of the vehicle with accuracy and transmits the obtained current position data (current position data S2) to a route search section 13.
The orbital data is detailed orbital information (parameter) representing a detailed orbit of each of the GPS satellites. In order to measure the current position of the vehicle with accuracy, the orbital data needs to be acquired from at least three GPS satellites.
The route search section 13 reads out map data S3 representing the current position of the vehicle and its surrounding area from a map data storage section 14 based on the current position data S2, searches the map data S3 for a driving route from the current position to a destination which has been set by a user, generates a route guide map S4 including the driving route, and transmits the route guide map S4 to the display section 2 constituted by a liquid crystal display.
As shown in
When the controller 16 of the PND 1 receives a command S7 which has been issued in response to depression of an application menu button MB displayed in the lower left of the guide map image G1 via a touch panel 15, the controller 16 reads out application menu screen data S5 from a data storage section 17.
The controller 16 of the PND 1 then displays an application menu screen G2 corresponding to the application menu screen data S5 on the display section 2 as shown in
When detecting a touch operation made to the guidebook button B3 in the application menu screen G2 via the touch panel 15 (
The controller 16 of the PND 1 then displays a book selection screen G3 corresponding to the book selection screen data S6 on the display section 2 as shown in
The book selection screen G3 further displays a “return” button BB1. When the “return” button BB1 is touched by a user, the controller 16 of the PND 1 sets back the display content of the display section 2 from the book selection screen G3 to application menu screen G2.
When detecting, via the touch panel 15, that the “1,000 spots” item KL2 on the book selection screen G3 has been touched by a user, the controller 16 of the PND 1 reads out password requesting screen data S7 from the data storage section 17 if a password has been set for the “1,000 spots” item KL2.
Then, as shown in
Also in this case, when the “return” button BB2 is touched by a user, the controller 16 of the PND 1 sets back the display content of the display section 2 from the password requesting screen G4 to book selection screen G3.
The controller 16 of the PND 1 displays a password input in response to a touch operation with respect to the alphabet key AK or numeric key EK in the password input field IP of the password requesting screen G4 and determines the password displayed in the password input field IP in response to a touch operation with respect to the enter key EK.
When authentication based on the password input to the password input field IP of the password requesting screen G4 is successful, the controller 16 of the PND 1 reads out spot selection screen data S8 from the data storage section 17.
The controller 16 of the PND 1 then displays a spot selection screen G5 corresponding to the spot selection screen data S8 as shown in
Further, the spot selection screen G5 displays a “return” button BB3. When the “return” button BB3 is touched by a user, the controller 16 of the PND 1 sets back the display content of the display section 2 from the spot selection screen G5 to book selection screen G3.
When the controller 16 of the PND 1 detects via the touch panel 15 that e.g., the spot item SA1: “G0001 Jonan Park” displayed on the spot selection screen G5 has been touched, the controller 16 of the PND 1 reads out introduction page screen data S8 about “G0001 Jonan Park” from the data storage section 17.
The controller 16 of the PND 1 then displays an introduction page screen G6 corresponding to the introduction page screen data S8 on the display section 2 as shown in
The introduction page screen G6 further displays, in the lower center thereof, a next button NT1 with an arrow for displaying a next page and a back button BK1 with an arrow for displaying a previous page. When the next button NT1 is touched by a user, the controller 16 of the PND 1 reads out detailed page screen data S9 concerning the “Jonan Park” from the data storage section 17.
The controller 16 of the PND 1 then displays a detailed page screen G7 corresponding to the detailed page screen data S9 on the display section 2 as shown in
The controller 16 of the PND 1 allows a user to visually confirm the detailed content display field DL of the detailed page screen G7 in the manner as described above. As a result, the user can grasp the detailed content of the “G0001 Jonan Park” and determine whether to set or not the “G0001 Jonan Park” as his or her destination or stop-off point.
Also in this case, the detailed page screen G7 displays, in the lower center thereof, a next button NT2 with an arrow for displaying a next page and a back button BK2 with an arrow for displaying a previous page. When the back button BK2 is touched by a user, the controller 16 of the PND 1 sets back the content of the display section 2 to the introduction page screen G6 concerning the “G0001 Jonan Park”.
In addition, when switching from the introduction page screen G6 to detailed page screen G7 (i.e., turn the page) or when switching from the detailed page screen G7 to introduction page screen G6 (i.e., bring back the page), the controller 16 of the PND 1 can perform the switching operation not only by detecting a user's touch operation with respect to the next buttons NT1, NT2 and back buttons BK1, BK2, but also by understanding a command corresponding to a finger gesture with respect to the touch panel 15 of the display section 2. In this case, the controller 16 of the PND 1 can perform the page switching processing according to the command. The details of this command recognition processing will be described below.
(3) Procedure of Command Recognition ProcessingAs shown in
In step SP2, when detecting a touch release operation, i.e., detecting that the user's finger has been separated from the touch panel 15, the controller 16 of the PND 1 proceeds to step SP3.
In step SP3, the controller 16 of the PND 1 calculates the movement amount of the finger on the display section 2 from the touch point detected in step SP1 to touch release point detected in step SP2 and proceeds to step SP4.
In step SP4, the controller 16 of the PND 1 determines whether the movement amount of the finger calculated in step SP3 is not more than a predetermined threshold (e.g., 5 mm). When an affirmative result has been obtained, which means that there is little movement, the controller 16 of the PND 1 determines that a gesture operation based on a drag operation of the finger has not been performed and proceeds to step SP5.
Since the finger motion does not mean a gesture operation, the controller 16 of the PND 1 determines, in step SP5, whether any button (e.g., next button NT1) exists at the touch release point as shown in
When a negative result has been obtained in step SP5, which means that the finger motion means neither a gesture operation nor a button touch operation, the controller 16 of the PND 1 determines that no command has been input and proceeds to step SP8 where the controller 16 of the PND 1 ends this flow without doing anything.
On the other hand, when an affirmative result has been obtained in step SP5, which means that the finger motion means a button touch operation with respect to, e.g., the next button NT1, the controller 16 of the PND 1 proceeds to step SP6.
In step SP6, the controller 16 of the PND 1 recognizes that a command corresponding to the button touch operation with respect to the next button NT1 has been issued and switches the display content from, e.g., the introduction page screen G6 (
When a negative result has been obtained in step SP4, i.e., the movement amount of the finger calculated in step SP3 exceeds a predetermined threshold (e.g., 5 mm), which means that a gesture operation based on a drag operation has been performed such that the finger is dragged from the left to right on the display section 2 as shown in
In step SP7, the controller 16 of the PND 1 recognizes, e.g., a page-turning command corresponding to the pattern of a gesture operation and, according to the page-turning command, switches the display content from the introduction page screen G6 (
When recognizing that a gesture operation of a motion pattern in which the finger is dragged from the left to right on the display section 2 has been performed, the controller 16 of the PND 1 performs the abovementioned page-turning operation, while when recognizing that a gesture operation of a motion pattern in which the finger is dragged from the right to left on the display section 2 has been performed, the controller 16 of the PND 1 performs a page turning back operation to set back the display content from the detailed page screen G7 (
When recognizing that a gesture operation of a motion pattern in which the finger is moved so as to draw a triangle has been performed, the controller 16 of the PND 1 recognizes that the motion means a command to search a driving route from the current position to the home of a user and displays the driving route obtained as a result of the search on the display section 2.
When recognizing that a gesture operation of a motion pattern in which the finger is swirled clockwise has been performed, the controller 16 of the PND 1 enlarges a map currently displayed on the display section 2, while when recognizing that a gesture operation of a motion pattern in which the finger is swirled counterclockwise has been performed, the controller 16 of the PND 1 reduces the map size.
(4) Operation and EffectWith the above configuration, the controller 16 of the PND 1 determines whether a finger motion is a button touch operation or a gesture operation based on a drag operation depending on the movement amount of the finger on the display section 2 from the touch point to touch release point, recognizes a command corresponding to the button touch operation or gesture operation, and performs predetermined processing according to the command.
That is, even if any button displayed on the display section 2 is touched by the finger, the controller 16 of the PND 1 does not determine whether the finger motion is a button touch operation or a gesture operation until the finger is separated from the button.
Therefore, even if a point at which any button exists is accidentally touched by the finger although the user intended to perform a gesture operation, it is possible to prevent the controller 16 of the PND 1 from erroneously determining that the finger motion is a button touch operation.
Further, even if the finger is applied to any button and is moved by an amount not more than a predetermined threshold until it is separated from the display section 2, the controller 16 of the PND 1 does not erroneously determine that the finger motion is a gesture operation but correctly determines that the finger motion is a button touch operation with respect to a button.
According to the above configuration, the controller 16 of the PND 1 correctly determines whether a touch operation is a button touch operation instantaneously made with respect to a button or a gesture operation based on a drag operation to thereby correctly and reliably execute predetermined processing specified by a command corresponding to the button touch operation or gesture operation.
(5) Other EmbodimentsIn the above embodiment, a value of 5 mm is used as a threshold of the movement amount of the finger from its touch point to touch release point to determine whether the finger motion is a button touch operation or a gesture operation. However, the present invention is not limited to this. For example, the value of the threshold may arbitrarily set depending on various factors such as the size of the display area of the display section 2 or size of the button.
Further, in the above embodiment, the pattern of a gesture operation includes the drag-like motion of the finger from the right to left, drag-like motion of the finger from the left to right, motion of the finger such as one by which a triangle is drawn, and clockwise or counterclockwise swirled motion of the finger. However, the present invention is not limited to this. For example, the pattern of a gesture operation may include a motion of the finger such as one by which “x” or “□” is drawn, a finger tapping motion, and other various types of motions.
Further, in the above embodiment, the controller 16 of the PND 1 determines that the finger motion is a gesture operation based on the movement amount of the finger on the display section 2 from the touch point to touch release point irrespective of whether any button displayed on the display section 2 is touched by the finger. However, the present invention is not limited to this but the controller 16 of the PND 1 may determine that the finger motion is a gesture operation based on the movement amount of the finger on the display section 2 where no button is displayed.
Further, in the above embodiment, the controller 16 of the PND 1 executes the command recognition processing procedure (
Further, in the above embodiment, the PND 1 as an information processing device according to the embodiment of the present invention is constituted by the display section 2 serving as a display unit, and controller 16 serving as a movement amount calculation unit, an operation determination unit, and a command recognition unit. Alternatively, however, the information processing device according to the embodiment of the present invention may be constituted by the display unit, movement amount calculation unit, operation determination unit, and command recognition unit including other various circuit configurations.
The information processing device and touch operation detection method can be applied to various electronic apparatuses having a touch panel, such as a dashboard-mounted navigation apparatus other than the PND, a personal computer, a Personal Digital Assistant (PDA), a mobile phone, and a game apparatus.
It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and alterations may occur depending on design requirements and other factors insofar as they are within the scope of the appended claims or the equivalents thereof.
Claims
1. An information processing device comprising:
- display means having a touch panel on its front surface;
- movement amount calculation means for calculating a movement amount of a touch operation based on a touch point at which the touch operation is performed with respect to the touch panel and a touch release point at which the touch operation is released from the touch panel;
- operation determination means for determining whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and
- command recognition means for recognizing whether a received command is a command corresponding to the depression operation or the gesture operation.
2. The information processing device according to claim 1, wherein
- the operation determination means determines that the touch operation is a gesture operation when the movement amount exceeds a predetermined threshold.
3. The information processing device according to claim 2, wherein
- the command recognition means recognizes the command associated with a pattern of the gesture operation.
4. The information processing device according to claim 1, wherein
- the operation determination means determines that, when the movement amount is not more than a predetermined threshold and a button is displayed at the touch release point at which the touch operation is released, the touch operation is the depression operation with respect to the button.
5. A touch operation detection method comprising:
- a touch point detection step in which touch point detection means detects a touch point at which a touch operation is performed with respect to a touch panel provided on the front surface of display means;
- a touch release point detection step following the touch point detection step, in which touch release point detection means detects a touch release point at which the touch operation is released from the touch panel;
- a movement amount calculation step in which movement amount calculation means calculates a movement amount of the touch operation based on the touch point and the touch release point;
- an operation determination step in which operation determination means determines whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and
- a command recognition step in which command recognition means recognizes whether a received command is a command corresponding to the depression operation or the gesture operation.
6. The touch operation detection method according to claim 5, wherein
- the operation determination step determines that the touch operation is the gesture operation when the movement amount exceeds a predetermined threshold.
7. The touch operation detection method according to claim 6, wherein
- the command recognition step recognizes the command associated with a pattern of the gesture operation.
8. The touch operation detection method according to claim 5, wherein
- the operation determination step determines that, when the movement amount is not more than a predetermined threshold and a button is displayed at the touch release point at which the touch operation is released, the touch operation is the depression operation with respect to the button.
9. An information processing device comprising:
- a display unit having a touch panel on its front surface;
- a movement amount calculation unit that calculates a movement amount of a touch operation based on a touch point at which the touch operation is performed with respect to the touch panel and a touch release point at which the touch operation is released from the touch panel;
- an operation determination unit that determines whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and
- a command recognition unit that recognizes whether a received command is a command corresponding to the depression operation or the gesture operation.
Type: Application
Filed: Dec 3, 2008
Publication Date: Jun 25, 2009
Applicant: Sony Corporation (Tokyo)
Inventor: Hirokazu Hashimoto (Kangawa)
Application Number: 12/326,991