Interface control
An improved interface control advantageously may be operated by one hand. A thumbpiece is slidably disposed within a longitudinal arm member which moves in an arcuate path. Placing his or her thumb in the thumbpiece, a user controls the horizontal positioning of a cursor by moving the arm member along the arcuate path. Vertical positioning of the cursor is controlled by sliding the thumbpiece along the length of the arm member. Trigger functions are implemented by exerting a downward force on the thumbpiece. Since the downward force used to implement the trigger function is orthogonal to motions used to control positioning of the cursor irrespective of the particular positions of the arm member and thumbpiece, the disclosed interface control prevents a user from inadvertently altering the positioning of the cursor during implementation of the trigger function. The arm member and sliding thumbpiece emulate the natural pivoting and curling/extending motions of the thumb, thereby resulting in a precise, easy to use, and ergonomically superior interface control.
Latest Immersion Corporation Patents:
- Methods and systems for decoding and rendering a haptic effect associated with a 3D environment
- Systems, devices, and methods for providing actuator braking
- Systems and Methods for Controlling a Multi-Actuator System to Generate a Localized Haptic Effect
- Haptic effect encoding and rendering system
- Apparatus and methods for localizing one or more effects in a haptic interface
The present invention relates to an interface control. More particularly, the present invention relates to an interface control device which allows a user to control the operation of computer applications, machinery, and video games.
BACKGROUND OF THE INVENTIONJoystick controls have been employed in a wide variety of applications, including computer software, industrial machinery, and multimedia interfaces to control the positioning of an object displayed on a screen, such as a cursor or pointer. A typical prior art joystick includes a gimballed stick pivotally coupled to a flat base portion. Angle sensors coupled to the gimballed stick generate position control signals in response to a user pivoting the gimballed stick relative to the base portion. These control signals are used to manipulate the position of the cursor. A depressible switch coupled to the top of the stick is used to generate a trigger control signal for implementing various functions, such as selecting items from a pull down menu or causing a character in a video game to jump.
The structure of these gimballed joystick controllers makes them somewhat difficult to operate. Rotating the arm and wrist to control positioning functions (i.e., pivoting the stick) while pressing downward with the thumb or finger to manipulate trigger functions requires a fair amount of practice and coordination. Further, requiring a user to simultaneously combine these motions may lead to an inadvertent change in the positioning of a cursor while implementing a trigger function. For instance, in a point-and-shoot operation, where a user first positions the cursor onto a target on the display screen and then activates the trigger function, depressing the trigger switch with the thumb or finger often results in slight movements of the arm and/or wrist, thereby causing the cursor to slip off the target. This phenomenon is commonly referred to as cursor creep.
The conventional joystick controller described above has the further disadvantage of undesirably requiring the use of two hands, i.e., one hand to hold the base of the controller and the other hand to operate the controller's stick. The only manner in which these controllers may be operated with one hand is to place the controller on a table or other flat surface.
Other joystick controllers have been developed in response to the above-mentioned problems. One such controller includes a pivoting, handgrip-shaped stick having one or more squeezable trigger switches built into a side portion of the handgrip. The positioning of an image is controlled by pivoting the handgrip, while the trigger functions are controlled by squeezing the trigger switches with the fingers. Although in such a design the positioning controls are somewhat isolated from the trigger function controls (i.e., squeezing the trigger switch with the index finger is not likely to cause as much of an inadvertent change in position as would depressing a trigger switch on the top of the stick with the thumb), cursor creep is nevertheless a problem. Further, such a controller requires the use of two hands or, alternatively, a tabletop support.
Some have attempted to develop a one-handed controller by simply reducing the size of conventional joystick controllers. These controllers fit within a user's hand, where the thumb, resting atop the stick, controls the positioning function. The trigger function is controlled by squeezing a trigger switch located on the side of the controller's stick.
These miniaturized versions of conventional joystick controllers are for the most part clumsy and ineffective. Merely reducing the size of a controller designed for two-handed operation so as to be operated by one hand severely limits the precision with which a user may control a cursor. Further, these miniaturized controllers are ineffective in isolating trigger controls from positioning controls. Indeed, squeezing a trigger switch with, for example, the index finger typically causes the controller stick to move forward, thereby resulting in undesirable vertical cursor creep.
This undesirable interaction between positioning and trigger controls of miniature joystick controllers, coupled with users' complaints of inferior ergonomics, has led others to revert to the more primitive two-handed video game controller shown in FIG. 1. Controller 1 has four keys 2a-2d clustered together in a first portion of controller 1 and three keys 4a-4c grouped together in a second portion of controller 1. Keys 2a-2d control the positioning of a displayed object (such as the hero of the video game) by generating digital positioning signals in response to a user depressing one or more of keys 2a-2d. Keys 4a-4c control various trigger functions (i.e., start-stop, jump, shoot, for example). The controller shown in
Thus, there is a need for joystick controller which may be operated in one hand. There is also a need for a controller having improved precision and ergonomics. Such a device should also isolate positioning and trigger controls, thereby eliminating cursor creep and other inadvertent position control signals produced during activation of trigger functions.
SUMMARY OF THE INVENTIONIn accordance with the present invention, an interface control is disclosed which offers users superior performance and ergonomics. In the preferred embodiment, a thumbpiece is slidably disposed within a longitudinal arm member having a first end pivotally coupled to a fixed point. The arm member may pivot about the fixed point such that a second end portion of the arm member follows an arcuate path having as its center the fixed point. The thumbpiece slides back and forth along the longitudinal axis of the arm member. A first sensor coupled to the arm member in a region proximate to the fixed point senses the angular position of the arm member. A second sensor coupled to the thumbpiece senses the linear movement of the thumbpiece relative to and longitudinally along the arm member. A third sensor coupled to the thumbpiece senses a downward force exerted upon the thumbpiece.
The interface control may, in one embodiment, comfortably rest in the palm of a user's hand. Positioning the fingers along the underside of the interface control, a user places the thumb in the thumbpiece. The user controls the horizontal positioning of, for example, a cursor by causing the arm member to pivot either to the right or to the left about the fixed point. This motion is detected by the first sensor, which in response thereto causes the cursor to move either right or left, respectively, on a display screen. The vertical positioning of the cursor is controlled by sliding the thumbpiece along the length of the arm member. The second sensor detects this linear movement of the thumbpiece and, in response thereto, causes the cursor to move up and down on the display screen. A user may implement trigger functions by exerting a downward force on the thumbpiece. This pressure is detected by the third sensor which, in turn, causes some predetermined function to be implemented on the display screen.
In another embodiment, the longitudinal arm member is disposed within a track defining an arcuate path rather than being coupled to a fixed pivot point. In this embodiment, positioning and trigger functions are controlled in the same manner as described in the preferred embodiment. By moving the arm member along the arcuate path as defined by the track, the thumbpiece follows an arcuate path having as its center a virtual pivot point. A sensor coupled to the arm member senses the arcuate movement of the arm member relative to the interface control and in response thereto generates a horizontal positioning control signal. The vertical positioning of the cursor and trigger functions are implemented as described above in connection with the preferred embodiment.
Embodiments of the present invention isolate the trigger function from the positioning controls. The downward force used to implement a trigger function will always be orthogonal to those motions of the thumb which are used to control the positioning of the cursor, regardless of the positions of the arm member or thumbpiece. This orthogonal relationship prevents a user from inadvertently altering the positioning of the cursor when trying to implement a trigger function.
Embodiments in accordance with the present invention also achieve superior ergonomics. The arm member is preferably of a length approximately equal to that of an adult thumb where different length arm members can be provided for different size hands. Together the arm member, which rotates to form an arcuate path, and the thumbpiece, which slides along the length of the arm member, emulate the natural pivoting and curling/extending motions of the thumb, respectively. The result is a comfortable, precise, and easy to use interface control.
In another embodiment, additional trigger switches are provided within cavities formed in the underside of the interface control. The fingertips of the user's hand, each comfortably nestled within an associated cavity, control the operation of the additional trigger switches, which may be used to implement numerous other functions.
Embodiments of the present invention are usable as an interface between a user and a machine where the machine carries out come predetermined function in response to commands issued by the user. In one embodiment, for instance, the user may control the mechanical operation of construction equipment. In another embodiment, the user may control moveable elements on a display screen, such as a cursor in a software application or an object in a video game.
This invention will be more fully understood in view of the following description taken together with the drawings.
In accordance with the present invention, an interface control is disclosed which allows a user to simultaneously control both a trigger function and the positioning of a cursor through a single contact surface using only the thumb. In the preferred embodiment shown in
First sensor 26 (
Referring to
Thumbpiece 28 (
Shaft 40 (
Second sensor 42 is preferably a linear resistive potentiometer. If a rotary potentiometer is used as second sensor 42, thumbpiece 28 may be coupled to the potentiometer by a conventional rack and pinion gear. It is to be understood, however, that any other sensing device capable of detecting either motion or pressure may be used as first sensor 26 and second sensor 42 in interface control 10. Although first sensor 26 and second sensor 42 preferably produce analog positioning signals, interface control 10 may also employ s sensors which produce digital positioning signals.
A third sensor 44 (
A user cradles hand-held interface control 10 in the palm of his or her hand, positioning his fingers along the underside of bottom plate 14 and resting his thumb on curved contact surface 38 of thumbpiece 38, as shown in FIG. 5. Using only his thumb, the user may control the horizontal and vertical positioning of, for instance, a cursor on a display screen, as well as implement various related functions (such as selecting options from a pull-down menu).
The horizontal positioning of the cursor is manipulated by pivoting arm member 20 about point 22, whereby arm member 20 traces an arcuate path as indicated by arrows 25a, 25b (FIG. 2). First sensor 26, in response to the arcuate motion of arm member 20, generates a first positioning signal which controls the horizontal positioning of a cursor on a display screen indicative of the angular position of arm member 20. For instance, to move the cursor to the right on the display screen, the thumb (resting in thumbpiece 28) is used to move arm member 20 in an arcuate path to the right, as indicated by arrow 25a.
Preferably, arm member 20 is of a length (measured between portions 16a and 16b of side wall 16) approximately equal to that of the user's thumb so that when a user places the thumb on contact surface 38 of thumbpiece 28, the user's large thumb joint is immediately above point 22. The length of arm member 20 may be of varying lengths so as to accommodate different size hands. As discussed above, arm member 20, pivoting about point 22, travels in an arcuate path between sidewalls 16a, 16b of interface control 10. This arcuate path emulates the natural motion of the thumb as it pivots about the large thumb joint, thereby resulting in a natural and ergonomic relationship between the thumb and arm member 20. Hence, interface control 10 uses the natural arcuate motion of a user's thumb to effect linear changes in the horizontal positioning of a cursor on the display screen.
The vertical positioning of the cursor is controlled by sliding thumbpiece 28 along the length of arm member 20 as indicated by arrows 32a, 32b (FIG. 2). Second sensor 42 detects the linear movement of thumbpiece 28 and, in response thereto, generates second positioning signals which control the vertical positioning of the cursor. For instance, to move the cursor up or down on a display screen, the user simply uses his or her thumb to move thumbpiece 28 away from or towards point 22. This movement is easily achieved by extending or curling the thumb. Note that by positioning the fingers on the bottom side of base plate 14, the combined movement of the fingers assist the thumb in the curling and extending motions.
Trigger functions are implemented by exerting a downward pressure (using the thumb) upon thumbpiece 28. Third sensor 44 senses this downward pressure and, in response thereto, generates a trigger control signal. Depending upon the particular application with which interface control 10 is being used, this trigger control signal may implement a variety of functions. For instance, in computer software or interactive systems, this trigger control may select items from a menu. When used with a video game, for example, this trigger control may cause a character to jump.
The downward pressure exerted by the thumb to implement a trigger control signal is always orthogonal to the thumb motions used to control horizontal and vertical positioning, regardless of the angular position of arm member 20 or the linear position of thumbpiece 28. This orthogonal relationship eliminates undesirable interaction between the trigger control and the positioning controls and thus prevents cursor creep. A user, when exerting a downward pressure on thumbpiece 28, is not likely to inadvertently pivot or extend/curl the thumb (which would cause a change in the positioning of the cursor). Thus, by isolating the trigger and positioning controls, as described above, interface control 10 provides advantages over conventional joystick controllers which suffer from cursor creep problems.
In a similar manner, a right-handed user may place one of the three fingers closest to the thumb (of the right hand) in each of cavities 54a, 54b, and 54c to operate associated switches 58a, 58b, 58c, which, like switches 56a, 56b, 56c, generate second, third, and fourth trigger control signals, respectively, for implementing various predetermined functions.
If additional control signals are desired for implementing even more predetermined fractions, control circuitry may be added so that the simultaneous depression of two of more switches 56a, 56b, 56c (or 58a, 58b, 58c for right-handed users) generate these additional control signals. For instance, simultaneously depressing switches 56a and 56b (for left-hand operation) or 58a and 58b (for right-hand operation) may generate a fifth trigger control signal.
The embodiments in accordance with the present invention are advantageous over conventional joystick controllers. Indeed, interface controls 10 and 50 are suitable for one-handed operation, thereby leaving the user's other hand free to perform other tasks. This one-handed operation also eliminates the need for a flat surface, as required when using a mouse or operating conventional joystick interface controls with one hand.
Using the thumb to control positioning functions contributes to the superior ergonomics of interface controls 10 and 50. As mentioned above, conventional joystick controllers require various combinations of wrist and arm movements to control the positioning of a cursor and additional finger or thumb motions to control trigger functions. Such unnatural combinations of motions necessitate training and practice in order for a user to become proficient. This is especially true as the size of such a control is reduced. Unlike conventional controllers, embodiments in accordance with the present invention (1) isolate the horizontal and vertical positioning functions from each other and (2) isolate the positioning functions from the trigger functions while simultaneously allowing both functions to be controlled by a single user contact point. The result is an interface control which allows users to issue vertical and horizontal positioning commands in concert for smooth and precise motions along both axes (i.e., a diagonal motion), as well as implementing trigger functions without inadvertently altering those positioning commands. Further, the interface controls described herein allow users to control both positioning and trigger control functions with simple, intuitive thumb motions.
When a user traces his or her thumb across the tips of the fingers, every joint in his or her hand moves in concert to facilitate the thumb's motion. This opposed position of a user's thumb relative to his or her fingers and palm is utilized by interface controls 10 and 50 to achieve a comfortable and natural interface between the user and a machine (such as a computer). Indeed, by positioning a user's fingers along the bottom surface of base plate 14 and his or her thumb in thumbpiece 28, interface controls 10 and 50 operate in a manner consistent with the thumb's natural motions. By taking advantage of the thumb's full arcuate motion about the large thumb joint and the thumb's excellent linear motion, interface controls in accordance with the present invention allow a user to quickly and easily position a cursor or pointer on a display screen. The superior ergonomics of the above interface controls afford users a high degree of precision and efficiency without the extensive practice and training required of conventional joystick controllers. This accuracy and ease of use makes the present interface controls especially well suited for CAD or any other computer illustration systems.
The frictional coupling between elements of the preferred embodiments may be manipulated to adjust the “feel” of interface controls 10 and 50. For instance, contact surface 38 (
In a similar manner, the frictional coupling between third sensor 44 (
The positioning and trigger functions of interface control 65 are controlled in a manner identical to those of interface control 10 as described above (see
The embodiments described above may be used in virtually any application which requires an interface control between a user and a machine. Embodiments in accordance with the present invention may be used to control the operation of a construction crane or boom. Interface controls 10 and 50 are ideal for replacing the mouse or trackball in computer software applications such as word processing, databases, and spreadsheets. For instance, interface control 50 of
Embodiments of the present invention may also be incorporated into conventional two-banded video game controllers (see
The above described interface controls may be mounted in virtually any enclosure, including (but not limited to) control panels, automobile dashboards, steering wheels, or handgrips of other interface controls. For instance, in one such embodiment, base plate 14 (
In another embodiment, interface controls in accordance with the present invention may be disposed within a control panel such as the dashboard of an automobile, boat, or even an airplane to provide control over certain operations. For example, interface control 10 may be mounted in the control panel of construction equipment to control the operation of a boom or crane. A control panel-mounted interface control 10 could also be used to manually control, for instance, the processing operations of an industrial application or the positioning and firing of lasers in medical applications. These embodiments, like those discussed above, are advantageous since multiple control functions (1) are disposed on a single contact surface and (2) are isolated from one another.
Some of the ergonomic advantages discussed herein may be compromised in order to provide a user interface control capable of controlling positioning in three, rather than two, directions. In one such embodiment in accordance with the present invention, various elements of interface control 10 may be incorporated into the handgrip of a conventional full-size joystick to provide three-dimensional positioning control as well as trigger functions.
A user curls the four fingers of his or her hand around stick 72 and places the thumb in thumbpiece 28 (FIG. 10). The user controls the horizontal and vertical positioning of, for instance, a cursor displayed on a CRT in a conventional manner as described above, i.e., by pivoting stick 72 about the base portion. The user controls the depth positioning of the cursor with the thumb by pivoting arm member 20 about pivot point 22 (see FIG. 2). Trigger functions are activated by pressing downward on thumbpiece 28 (as discussed in reference to interface control 10).
Various forms of feedback may be added to the above described embodiments to provide a user with additional information about the particular application he or she is controlling, as described below in reference to
Referring to
In a similar manner, a second actuator may be provided that in response to a second feedback signal inhibits the movement of thumbpiece 28 along arm member 20, as illustrated in FIG. 12. Arm member 20 has provided therein a sliding bar 90 having a first end coupled to thumbpiece 28. Iron core 92 is coupled to arm member 20 and is positioned in a region proximate to a second end of bar 90. Magnetic coil 94 is wound around ferrous care 92. A second feedback current provided to coil 94 will induce a magnetic attraction between bar 90 and against core 92, thereby resulting in an increased frictional coupling between bar 90 and core 92. This increased frictional coupling resists the sliding motion of thumbpiece 28 along arm member 20. Note that in some embodiments shaft 40 (
The embodiments described above and illustrated in
In other applications, interface control 80 (
The actuators discussed above may comprise a solenoid, a servomotor, or any other suitable device known in the art which generates a force in response to electric signals. The actuators may also employ shape-memory alloys, piezo ceramics, or electro-rheological compounds. Further, motor-type actuators may be employed to augment or restrain motion.
In other embodiments, the actuators discussed above may used to activate and deactivate electrically controlled detents so as to provide tactile click stops in the pivoting motion of arm member 20 (
Embodiments of the present invention may also be equipped with a spring return mechanism. With reference to interface control 10 (FIG. 2), a centering spring may be coupled to arm member 20 which causes arm member 20 to return to its center position whenever arm member 20 has deviated from the center position by exerting pressure on arm member 20. A manually controlled latch may also be provided which engages the centering spring to and disengages the centering spring from arm member 20 so as to turn on and off the centering mechanism. Such a centering mechanism is useful in applications requiring proportional control (i.e., a conventional joystick) rather than absolute control (i.e., a mouse). The centering spring may also be electrically actuated by an external signal from the interfaced device (i.e., computer, video game, and so on). Inclusion of such an electrically actuated spring allows the interfaced device to switch the controller between two modes of operation (spring centering and non-centering), as the particular application may require. In a similar manner, an additional centering spring may be coupled to thumbpiece 28 to provide proportional control in the vertical direction.
While particular embodiments of the present invention have been shown and described, it will be obvious to those skilled in the art that changes and modifications may be made without departing from this invention in its broader aspects and, therefore, the appended claims are to encompass within their scope all such changes and modifications as fall within the true spirit and scope of this invention.
Claims
1. An apparatus for generating at least two control signals, said apparatus comprising:
- a support;
- an arm member disposed within said support, said arm member being moveable in an arcuate path within said support;
- a contact member slidably mounted on said arm member, said contact member being slidable along said arm member;
- a first sensor coupled to said arm member for sensing movement of said arm member along said arcuate path; and
- a second sensor coupled to said contact member for sensing linear movement of said contact member along said arm member.
2. The apparatus of claim 1 further comprising:
- first guide means disposed along said arm member; and
- second guide means disposed on said contact member, said second guide means matingly coupled with said first guide means so as to facilitate the movement of said contact member along said arm member.
3. The apparatus of claim 2 wherein said first guide means comprise first and second grooves horizontally disposed in first and second side walls, respectively, of said arm member, said second guide means comprising first and second flanges disposed on first and second side walls of said contact member, respectively.
4. The apparatus of claim 1 wherein said arm member is pivotally coupled to said support at a pivot point, said arm member moving in said arcuate path as said arm member pivots about said pivot point.
5. The apparatus of claim 1 further comprising:
- a track means disposed within said support, said track means defining said arcuate path; and
- one or more guide elements disposed on an outer surface of said arm member, said one or mere guide elements being matingly coupled to said track means for facilitating the movement of said arm member in said arcuate path.
6. The apparatus of claim 1 wherein said first sensor detects the magnitude and direction of arcuate movement of said arm member along said arcuate path and generates a first control signal in response thereto.
7. The apparatus of claim 1 wherein said second sensor detects the magnitude and direction of linear movement of said contact member and generates a second control signal in response thereto.
8. The apparatus of claim 1 further comprising a third sensor coupled to said arm member, said third sensor detecting a downward pressure on said contact member and in response thereto generating a signal for implementing a first predetermined function.
9. An apparatus for generating at least two control signals, said apparatus comprising:
- a support;
- an arm member having a first end portion pivotally coupled to said housing at a point, said arm member being rotatable about said point;
- a contact member slidably mounted on said arm member, said contact member being slidable along said arm member;
- a first sensor coupled to said arm member for sensing arcuate movement of said arm member; and
- a second sensor coupled to said contact member for sensing linear movement of said contact member along said arm member.
10. The apparatus of claim 9 wherein said support is of a size such that said apparatus is capable of being operated with one hand of a person.
11. The apparatus of claim 9 wherein said first sensor generates a first control signal representing the angle and direction of rotation of said arm member, said first control signal being capable of being used to alter the position of a cursor or pointer in a first direction on a display screen.
12. The apparatus of claim 9 wherein said second sensor generates a second control signal representing the magnitude and direction of linear movement of said contact member, said second control signal bang capable of being used to alter the position of a cursor or pointer in a second direction on a display screen.
13. The apparatus of claim 9 further comprising a third sensor disposed on said arm member, said third sensor detecting a downward pressure on said contact member.
14. The apparatus of claim 13 wherein said third sensor generates a third control signal indicative of said downward pressure on said contact member, said third control signal implementing a first predetermined function.
15. The apparatus of claim 9 further comprising a first actuator, wherein in response to a first feedback signal said first actuator restrains said rotation of said arm member.
16. The apparatus of claim 15 wherein said first actuator is contained within said first sensor.
17. The apparatus of claim 15 further comprising a second actuator, wherein in response to a second feedback signal said second actuator restrains the linear movement of said arm member.
18. The apparatus of claim 17 wherein said second actuator is contained within said second sensor.
19. The apparatus of claim 9 wherein said contact member has a concave upper surface so that a human thumb may matingly situate on said contact member.
20. The apparatus of claim 9 wherein said arm member has disposed therein one or more first guide means, said contact member having disposed thereon one or more second guide means, each of said first guide means being matingly coupled with an associated one of said second guide means for facilitating sliding of said contact member along said arm member.
21. The apparatus of claim 20 wherein said first guide means comprise first and second grooves horizontally disposed in first and second side walls, respectively, of said arm member, said second guide means comprising first and second flanges disposed on first and second side walls of said contact member, respectively.
22. The apparatus of claim 9 wherein said housing has a bottom surface, said apparatus further comprising:
- one or more cavities formed in said bottom surface of said housing;
- one or more switches, each of said switches being associated with and coupled to an associated one of said cavities.
23. The apparatus of claim 22 wherein said switches comprise pressure sensitive devices, the actuation of which generates one or more fourth control signals for implementing one or more second predetermined functions.
24. An interface control device, comprising:
- a support housing configured to be held by one hand of a user;
- a user manipulatable member engageable and moveable by a single thumb of said user;
- at least one sensor coupled to said user manipulatable member and operative to sense movement in a first and second degree of freedom and to generate at least one sensor signal associated with said movement;
- said user manipulatable member moveable in a third degree of freedom and configured to cause a trigger signal to be sent to an application on a computer, said third degree of freedom approximately orthogonal to said first and second degrees of freedom; and
- at least one actuator coupled to said interface control device, said actuator operative to provide a feedback force to said user, wherein said interface control device is configured to be operated by said one hand of a user, and wherein said user manipulatable member is coupled to an arm member having rotary motion about a pivot point to provide motion in one of said first or second degrees of freedom, wherein said actuator is coupled to said arm member to output forces about said pivot point.
25. The interface control device as recited in claim 24 wherein said rotary motion of said arm member is limited to an arcuate path of less than ninety degrees.
26. The interface control device as recited in claim 24 further comprising a second actuator, and wherein said first actuator is grounded to said housing and wherein said second actuator is carried by said arm member.
27. The interface control device as recited in claim 24 wherein said user manipulatable member is a sliding contact member which can be moved in a linear degree of freedom approximately perpendicular to an axis of rotation of said arm member and in substantially the same plane as said rotary motion, thereby providing said motion in one of said first or second degrees of freedom.
28. An interface control device, comprising:
- a support housing configured to be held by one hand of a user;
- a user manipulatable member engageable and moveable by a single thumb of said user;
- at least one sensor coupled to said user manipulatable member and operative to sense movement in a first and second degree of freedom and to generate at least one sensor signal associated with said movement;
- said user manipulatable member moveable in a third degree of freedom and configured to cause a trigger signal to be sent to an application on a computer, said third degree of freedom approximately orthogonal to said first and second degrees of freedom; and
- at least one actuator coupled to said interface control device, said actuator operative to provide a feedback force to said user, wherein said interface control device is configured to be operated by said one hand of a user, and wherein a centering spring bias on said user manipulatable member may be electrically actuated by a signal received from said computer, allowing said interface control device to have a centering mode and a non-centering mode, selected by said computer.
29. An interface control device in communication with a computer for providing positioning signals to said computer for manipulating an image in a computer environment displayed on a screen by said computer, said device comprising:
- a handheld support housing configured to be held by one hand of a user;
- a user manipulatable member coupled to said housing and engageable and moveable by a single thumb of said user in two degrees of freedom relative to said housing, and configured with a contact surface configured to be contacted by said thumb;
- at least one sensor coupled to said user manipulatable member and operative to sense movement of said user manipulatable member in said two degrees of freedom, said sensor operative to provide positioning signals;
- at least one actuator coupled to said interface control device, wherein said actuator is operative to provide a feedback force to said user, and wherein said at least one actuator includes a first brake providing a drag in a first of said two degrees of freedom, and a second computer controlled brake coupled to said user manipulatable member and providing a drag in a second one of said degrees of freedom of said user manipulatable member; and
- a thumb trigger sensor operative to detect a trigger command from said user and to cause a trigger signal to be sent to said computer, said trigger command including a pressing motion by said thumb causing said user manipulatable member to move in a trigger degree of freedom different from said two degrees of freedom, wherein said user manipulatable member is configured to allow said user to control said movement in said two degrees of freedom and perform said trigger command simultaneously using said single thumb on said contact surface, and
- wherein said interface control device is configured to be operated by said one hand of a user, and wherein said user manipulatable member is coupled to an arm member having rotary motion about a pivot point and is a sliding member which can be moved along at least a portion of said arm member in a linear degree of freedom, and wherein said second brake outputs forces in said linear degree of freedom, wherein said first brake is coupled to said arm member to output forces about said pivot point.
30. An interface control device in communication with a computer for providing positioning signals to said computer for positioning an image displayed on a screen said device comprising:
- a support housing configured to be held by one hand of the user;
- a user manipulatable member coupled to said housing and engageable and moveable by a digit of said user in two degrees of freedom relative to said housing while said housing is held by said hand of said user, wherein at least one of said degrees of freedom is a rotary degree of freedom about an axis of rotation;
- a spring return mechanism coupled to said user manipulatable member to provide a centering bias on said user manipulatable member toward a center position of said rotary degree of freedom when said user manipulatable member has been moved from said center position, wherein said spring return mechanism is electrically actuated by an external signal received from said computer, allowing said spring return mechanism to be selectively applied in a centering mode and allowing said spring return mechanism to have no effect in a non-centering model;
- at least one sensor coupled to said user manipulatable member and sensing movement of said user manipulatable member in said two degrees of freedom, said sensor providing positioning signals which control said positioning of said image on said screen;
- at least one actuator coupled to said user manipulatable member; and
- a trigger sensor for detecting a trigger command from said user, said trigger command including a pressing motion causing said user manipulatable member to move in a trigger degree of freedom different from said two degrees of freedom.
31. The interface control device as recited in claim 30 wherein said external signal is controlled by a video game running on said computer.
32. The interface control device as recited in claim 30 wherein said spring return mechanism is coupled to a pivotable arm member providing said rotary degree of freedom, and further comprising a centering spring coupled to said user manipulatable member to provide a centering bias in another of said two degrees of freedom.
33. An interface control feedback device in communication with a computer for providing positioning signals to said computer for manipulating an image in a computer environment displayed on a screen by said computer; said device comprising:
- a support housing configured to be held by one hand of a user;
- a sliding contact member engageable and moveable by a thumb of said user in two degrees of freedom relative to said support housing while said support housing is held by said hand of said user, one of said two degrees of freedom being a linear degree of freedom, wherein said movement in said two degrees of freedom positions said image in two screen dimensions on said display device;
- an arm member coupled to said sliding contact member, said arm member operative to rotationally move about a pivot point to provide motion in one of said two degrees of freedom, wherein said linear degree of freedom is approximately perpendicular to an axis of rotation of said arm member and is in substantially the same plane as said rotary motion;
- at least one sensor coupled to said user manipulatable member and operative to sense movement of said sliding contact member in said two degrees of freedom, said sensor operative to provide positioning signals which control said positioning of said image on said display device;
- at least one actuator coupled to said arm member to output forces about said pivot point, wherein said forces facilitate the selection of options or icons displayed on said display device based on feedback signals generated by an application running on said computer; and
- a trigger sensor for detecting a trigger command from said user, said trigger command including moving said sliding contact member approximately orthogonally to said two degrees of freedom,
- wherein said interface control device is configured to be operated by said one hand of a user.
34. The interface control device as recited in claim 33 further comprising a second actuator to output forces on said sliding contact member in said linear degree of freedom, and wherein said first actuator is grounded to said housing and wherein said second actuator is carried by said arm member.
35. The interface control device as recited in claim 33 wherein said image is a cursor controlled to move in two dimensions of said screen, wherein said cursor can be used to select an icon, wherein said trigger command selects said icon when said cursor is positioned over said icon.
36. The interface control device as recited in claim 33 wherein said image is a video game character provided in a video game environment.
37. An interface control feedback device in communication with a computer for providing positioning signals to said computer for positioning an image displayed on a display device, said device comprising:
- a support housing configured to be held by one hand of a user;
- a user manipulatable member engageable and moveable by a thumb of said user in two degrees of freedom relative to said support housing while said support housing is held by said hand of said user, wherein said movement in said two degrees of freedom positions said image in two screen dimensions on said display device;
- at least one sensor coupled to said user manipulatable member and operative to sense movement of said user manipulatable member in said two degrees of freedom, said sensor operative to provide positioning signals which control said positioning of said image on said display device;
- at least one actuator coupled to said user manipulatable member, wherein said actuator provides a force in at least one of said degrees of freedom of said user manipulatable member, wherein said force facilitates the selection of options or icons displayed on said display device based on feedback signals generated by an application running on said computer, wherein a centering spring bias on said user manipulatable member is electrically actuated by a signal received from said computer in a centering mode, allowing said force feedback device to have said centering mode and a non-centering mode selected by said computer; and
- a trigger sensor for detecting a trigger command from said user, said trigger command including moving said user manipulatable member approximately orthogonally to said two degrees of freedom,
- wherein said interface control device is configured to be operated by said one hand of a user.
38. A device comprising:
- a housing configured to be held in one hand such that no additional support is needed to operate the device;
- a user manipulatable member coupled to said housing and configured to be manipulated by a single digit of a user in two degrees of freedom, wherein said user manipulatable member comprises an arm member operable to rotate in said rotary degree of freedom, and a sliding contact member operable to move in a linear degree of freedom approximately perpendicular to an axis of rotation of said arm member;
- a sensor coupled to said user manipulatable member and operative to sense movement of said user manipulatable member in said two degrees of freedom, wherein one of said two degrees of motion comprises a rotary degree of freedom;
- an actuator operative to provide a feedback force to said user;
- a trigger operative to move in a degree of freedom different from said two degrees of freedom and configured to be actuated by said single digit simultaneously with said user manipulatable object,
- wherein said device is configured to be operated by said one hand of a user.
3157853 | November 1964 | Hirsch |
3220121 | November 1965 | Cutler |
3497668 | February 1970 | Hirsch |
3517446 | June 1970 | Corlyon et al. |
3902687 | September 1975 | Hightower |
3903614 | September 1975 | Diamond et al. |
3919691 | November 1975 | Noll |
4131033 | December 26, 1978 | Wright et al. |
4160508 | July 10, 1979 | Salsbury |
4236325 | December 2, 1980 | Hall et al. |
4398889 | August 16, 1983 | Lam et al. |
4477043 | October 16, 1984 | Repperger |
4489304 | December 18, 1984 | Hayes |
4513235 | April 23, 1985 | Acklam et al. |
4560983 | December 24, 1985 | Williams |
4581491 | April 8, 1986 | Boothroyd |
4599070 | July 8, 1986 | Hladky et al. |
4604016 | August 5, 1986 | Joyce |
4692756 | September 8, 1987 | Clark |
4706294 | November 10, 1987 | Ouchida |
4708656 | November 24, 1987 | de Vries et al. |
4712101 | December 8, 1987 | Culver |
4713007 | December 15, 1987 | Alban |
4724715 | February 16, 1988 | Culver |
4782327 | November 1, 1988 | Kley et al. |
4794384 | December 27, 1988 | Jackson |
4795296 | January 3, 1989 | Jau |
4800721 | January 31, 1989 | Cemenska et al. |
4823634 | April 25, 1989 | Culver |
4825157 | April 25, 1989 | Mikan |
4861269 | August 29, 1989 | Meenen, Jr. |
4868549 | September 19, 1989 | Affinito et al. |
4891764 | January 2, 1990 | McIntosh |
4930770 | June 5, 1990 | Baker |
4934694 | June 19, 1990 | McIntosh |
4935728 | June 19, 1990 | Kley |
4949119 | August 14, 1990 | Moncrief et al. |
4962448 | October 9, 1990 | DeMaio et al. |
4983901 | January 8, 1991 | Lehmer |
5004391 | April 2, 1991 | Burdea |
5007300 | April 16, 1991 | Siva |
5019761 | May 28, 1991 | Kraft |
5022407 | June 11, 1991 | Horch et al. |
5035242 | July 30, 1991 | Franklin |
5038089 | August 6, 1991 | Szakaly |
5044956 | September 3, 1991 | Behensky et al. |
5065145 | November 12, 1991 | Purcell |
5078152 | January 7, 1992 | Bond |
5086296 | February 4, 1992 | Clark |
5095303 | March 10, 1992 | Clark et al. |
5103404 | April 7, 1992 | McIntosh |
5107080 | April 21, 1992 | Rosen |
5142931 | September 1, 1992 | Menahem |
5146566 | September 8, 1992 | Hollis, Jr. et al. |
5184319 | February 2, 1993 | Kramer |
5185561 | February 9, 1993 | Good et al. |
5186629 | February 16, 1993 | Rohen |
5186695 | February 16, 1993 | Mangseth et al. |
5193963 | March 16, 1993 | McAffee et al. |
5202961 | April 13, 1993 | Mills et al. |
5203563 | April 20, 1993 | Loper, III |
5212473 | May 18, 1993 | Louis |
5220260 | June 15, 1993 | Schuler |
5223776 | June 29, 1993 | Radke et al. |
5235868 | August 17, 1993 | Culver |
5237327 | August 17, 1993 | Saitoh et al. |
5240417 | August 31, 1993 | Smithson et al. |
5271290 | December 21, 1993 | Fischer |
5275174 | January 4, 1994 | Cook |
5280276 | January 18, 1994 | Kwok |
5284330 | February 8, 1994 | Carlson |
5293158 | March 8, 1994 | Soma |
5296846 | March 22, 1994 | Ledley |
5296871 | March 22, 1994 | Paley |
5299810 | April 5, 1994 | Pierce |
5309140 | May 3, 1994 | Everett |
5313229 | May 17, 1994 | Gilligan et al. |
5313230 | May 17, 1994 | Venolia et al. |
5317336 | May 31, 1994 | Hall |
5329289 | July 12, 1994 | Sakamoto et al. |
5334027 | August 2, 1994 | Wherlock |
5379663 | January 10, 1995 | Hara |
5381080 | January 10, 1995 | Schnell et al. |
5396266 | March 7, 1995 | Brimhall |
5405152 | April 11, 1995 | Katanics et al. |
5414337 | May 9, 1995 | Schuler |
5450079 | September 12, 1995 | Dunaway |
5457479 | October 10, 1995 | Cheng |
5466213 | November 14, 1995 | Hogan |
5473344 | December 5, 1995 | Bacon et al. |
5474082 | December 12, 1995 | Junker |
5477237 | December 19, 1995 | Parks |
5491477 | February 13, 1996 | Clark et al. |
5492312 | February 20, 1996 | Carlson |
5506605 | April 9, 1996 | Paley |
5512919 | April 30, 1996 | Araki |
5513100 | April 30, 1996 | Parker et al. |
5530455 | June 25, 1996 | Gillick et al. |
5543821 | August 6, 1996 | Marchis et al. |
5547382 | August 20, 1996 | Yamasaki |
5550563 | August 27, 1996 | Matheny et al. |
5576727 | November 19, 1996 | Rosenberg et al. |
5583407 | December 10, 1996 | Yamaguchi |
5587937 | December 24, 1996 | Massie et al. |
5589828 | December 31, 1996 | Armstrong |
5623582 | April 22, 1997 | Rosenberg |
5627531 | May 6, 1997 | Posso et al. |
5629594 | May 13, 1997 | Jacobus et al. |
5631669 | May 20, 1997 | Stobbs et al. |
5642469 | June 24, 1997 | Hannaford |
5643087 | July 1, 1997 | Marcus et al. |
5666473 | September 9, 1997 | Wallace |
5691747 | November 25, 1997 | Amano |
5691898 | November 25, 1997 | Rosenberg et al. |
5709219 | January 20, 1998 | Chen et al. |
5714978 | February 3, 1998 | Yamanaka et al. |
5721566 | February 24, 1998 | Rosenberg et al. |
5724106 | March 3, 1998 | Autry et al. |
5734373 | March 31, 1998 | Rosenberg et al. |
5742278 | April 21, 1998 | Chen |
5766016 | June 16, 1998 | Sinclair |
5769640 | June 23, 1998 | Jacobus et al. |
5781172 | July 14, 1998 | Engel et al. |
5785630 | July 28, 1998 | Bobick et al. |
5790108 | August 4, 1998 | Salcudean et al. |
5816823 | October 6, 1998 | Naimark et al. |
5821920 | October 13, 1998 | Rosenberg et al. |
5825308 | October 20, 1998 | Rosenberg |
5841423 | November 24, 1998 | Carrol, Jr. et al. |
5841428 | November 24, 1998 | Jaeger et al. |
5889670 | March 30, 1999 | Schuler et al. |
5914705 | June 22, 1999 | Johnson et al. |
5956016 | September 21, 1999 | Kuenzner et al. |
6004134 | December 21, 1999 | Marcus et al. |
6069614 | May 30, 2000 | Singhal |
6104382 | August 15, 2000 | Martin et al. |
6107992 | August 22, 2000 | Ishigaki |
6111577 | August 29, 2000 | Zilles et al. |
6184862 | February 6, 2001 | Leiper |
6219034 | April 17, 2001 | Elbing et al. |
6256011 | July 3, 2001 | Culver |
6281883 | August 28, 2001 | Barker |
6300936 | October 9, 2001 | Braun et al. |
6422941 | July 23, 2002 | Thomer et al. |
0265011 | April 1988 | EP |
0349086 | January 1990 | EP |
0607580 | July 1994 | EP |
0626634 | November 1994 | EP |
83-820012 | January 1983 | JP |
59-160229 | September 1984 | JP |
01-003664 | July 1990 | JP |
02-109714 | January 1992 | JP |
04-007371 | August 1993 | JP |
H6-210065 | August 1994 | JP |
05-193862 | January 1995 | JP |
WO92/00559 | January 1992 | WO |
WO95/20788 | August 1995 | WO |
- Schmult, Brian et al., “Application Areas for a Force-Feedback Joystick,” ASME 1993, DSC—vol. 49, pp. 47-54.
- Howe, Robert D., “Task Performance with a Dextrous Teleoperated Hand System,” Proceedings of SPIE, Nov. 1992, vol. 1833, pp. 1-9.
- Russo, Massimo Andrea, “The Design and Implementation of a Three Degree-of-Freedom Force Output Joystick,” Department of Mechanical Engineering, May 11, 1990, pp. 9-40 & 96 & 97.
- Adachi, Yoshitaka et al., “Sensroy Evaluation of Virtual Haptic Push-Buttons,” Technical Research Center Suzuki Motor Corporation.
- Su, S. Augustine et al., “The Virtual Panel Architecture: A 3D Gesture Framework,” IEEE 1993.
- Rosenberg, Louis B., Perceptual Design of A Virtual Rigid Surface Contact, Center for Design Research, Stanford University, Armstrong Laboratory, AL/CF-TR-1995-0029, Apr. 1993.
- Ellis, R.E. et al., “Design and Evaluation of a High-Performance Prototype Planar Haptic Interface,” ASME Dec. 3, 1993, DSC—vol. 49, pp. 55-64.
- Adlestein, Bernard D. et al., “Design and Implementation of a Force Reflecting Manipulandum for Manual Control Research,” 1992, pp. 1-24.
- Adelstein, Bernard D. et al., “A High Performance Two Degree-of-Freedom Kinesthetic Interface,” Massachusetts Institute of Technology 1992, pp. 108-112.
- Minsky, Margaret et al., “Feeling and Seeing: Issues in Force Display,” ACM 1990, pp. 235-242.
- Batter, James J. et al., “Grope-1: A Computer Display to the Sense of Feel,” pp. TA-4-188-TA-4-192.
- Gotow, J.K., et al., “Perception of Mechanical Properties at the Man-Machine Interface,” IEEE 1987, pp. 688-689.
- Atkinston, William D. et al, “Computing with Feeling,” Comput. & Graphics, vol. 2, No. 2-E, pp. 97-103.
- Wiker, Steven F. et al., “Development of Tactile Mice for Blind Access to Computers: Importance of Stimulation Locus, Object Size, and Vibrotactile Display Resolution,” Proceedings of the Human Factors Society 35th Annual Meeting 1991, pp. 708-712.
- Iwata, Hiroo, “Artificial Reality with Force-feedback: Development of Desktop Virtual Space with Compact Master Manipulator,” Computer Graphics, vol. 24, No. 4, 1990, pp. 165-170.
- Hirota, Koichi et al., “Development of Surface Display,” IEEE 0-7803-1363-1, 1993, pp. 256-262.
- Akamatsu, M. et al., “Multimodal Mouse: A Mouse-Type Device with Tactile and Force Display,” Presence, vol. 3, No. 1, 1994, pp. 73-80.
- Millman, P. et al., “Design of a Four Degree-of-Freedom Force-Reflecting Manipulandum with a Specified Force/Torque Workspace,” IEEE CH2969-4, 1991, pp. 1488-1492.
- Brooks, F. et al., “Project GROPE—Haptic Displays for Scientific Visualization,” Computer Graphics, vol. 24, No. 4, 1990, pp. 177-185.
- Kelley, A. J. et al., “MagicMouse: Tactile and Kinesthetic Feedback in the Human-Computer Interface using an Electromagnetically Actuated Input/Output Device,” Dept. of Elec. Eng., Univ. of Brit. Columbia, 1993, pp. 1-27.
- Rosenberg, Louis B., “Virtual Haptic Overlays Enhance Performance in Telepresence Tasks,” SPIE 1994.
- Bostrom, M. et al., “Design of An Interactive Lumbar Puncture Simulator With Tactile Feedback,” IEEE 0-7803-1363, 1993, pp. 280-286.
- Ramstein, Christophe, “The Pantograph: A Large Workspace Haptic Device for a Multimodal Human-Computer Interaction”, Compuiter-Human Interaction, CHI 1994.
- Ouh-young, M. et al., “Creating an Illustion of Feel: Control Issues in Force Display,” Computer Science Dept., Univ. of N. Carolina, 1989, pp. 1-14.
- Hannaford, B. et al., “Force Feedback Cursor Control,” NASA Tech Brief, vol. 13, No. 11, Item #21, 1989, pp. i, 1-4.
- Hayward, V. et al., “Design and Multi-Objective Optimization of a Linkage for a Haptic Interface,” Advances in Robot Kinematics and Computationed Geometry, Kluwer Academic Publishers, 1994, pp. 359-368.
- Iwata, Hiroo, “Pen-base Haptic Virtual Environment”, IEEE, 0-7803-1363-1, pp. 287-292.
- Winey III, C., “Computer Simulated Visual and Tactile Feedback as an Aid to Manipulator and Vehicle Control,” Mass. Inst. of Tech., 1981, pp. 1-79.
- Kilpatrick, P., “The Use of a Kinesthetic Supplement in an Interactive Graphics System,” Univ. of N. Carolina, 1976, pp. 1-175.
- Iwata, “Pen-based Haptic Virtual Environment,” 0-7803-1363-1/93 IEEE, pp. 287-292, 1993.
- Russo, “The Design and Inplementation of a Three Degree of Freedom Force Output Joystick,” MIT Libraries Archives Aug. 14, 1990, pp. 1-131, May 1990.
- Adelstein, “Design and Implementation of a Force Reflecting Manipulandum for Manual Control research,” DSC—vol. 42, Advances in Robotics, Edited by H. Kazerooni, pp. 1-12, 1992.
- Baigrie, “Electric Control Loading-A Low Cost, High Performance Alternative,” Proceedings, pp. 247-254, Nov. 6-8, 1990.
- Brooks et al., “Hand Controllers for Teleoperation-A State-of-the-Art Technology Survey and Evaluation,” JPL Publication 85-11; NASA-CR-175890; N85-28559, pp. 1-84, Mar. 11, 1985.
- Burdea et al., “Distributed Virtual Force Feedback, Lecture Notes for Workshop on Force Display in Virtual Environments and its Application to Robotic Teleoperation,” 1993 IEEE International Conference on Robotics and Automation, pp. 25-44, May 2, 1993.
- Caldwell et al., “Enhanced Tactile Feedback (Tele-Taction) Using a Multi-Functional Sensory System,” 1050-4729/93, pp. 955-960, 1993.
- Gotow et al., “Controlled Impedance Test Apparatus for Studying Human Interpretation of Kinesthetic Feedback,” WA11-11:00, pp. 332-337, 1989.
- Jones et al., “A Perceptual analysis of stiffness,” ISSN 0014-4819 Springer International (Springer-Verlag); Experimental Brain Research, vol. 79, No. 1, pp. 150-156, 1990.
- Ouh-Young, “Force Display in Molecular Docking,” Order No. 9034744, pp. 1-369, 1990.
- Russo, “Controlling Dissipative Magnetic Particle Brakes in Force Reflective Devices,” DSC-vol. 42, Advances in Robotics, pp. 63-70, ASME 1992.
- Snow et al., “Model-X Force-Reflecting-Hand-Controller,” NT Control No. MPO-17851; JPL Case No. 5348, pp. 1-4, Jun. 15, 1989.
- Stanley et al., “Computer Simulation of Interacting Dynamic Mechanical Systems Using Distributed Memory Parallel Processors,” DSC-vol. 42, Advances in Robotics, pp. 55-61, ASME 1992.
- Tadros, “Control System Design for a Three Degree of Freedom Virtual Environment Simulator Using Motor/Brake Pair Actuators”, MIT Archive © Massachusetts Institute of Technology, pp. 1-88, Feb. 1990.
- Adelstein, “A Virtual Environment System For The Study of Human Arm Tremor,” Ph.D Dissertation, Dept. of Mechanical Engineering, MIT, Jun. 1989.
- Bejczy, “Generalization of Bilateral Force-Reflecting Control of Manipulators,”Proceedings Of Fourth CISM-IFToMM, Sep. 8-12, 1981.
- Bejczy, “Sensors, Controls, and Man-Machine Interface for Advanced Teleoperation,”Science, vol. 208, No. 4450, pp. 1327-1335, 1980.
- Eberhardt et al., “OMAR-A Haptic display for speech perception by deaf and deaf-blind individuals,” IEEE Virtual Reality Annual International Symposium, Seattle, WA, Sep. 18-22, 1993.
- Howe, “A Force-Reflecting Teleoperated Hand System for the Study of Tactile Sensing in Precision Manipulation,” Proceedings of the 1992 IEEE International Conference on Robotics and Automation, Nice, France, May 1992.
- IBM Technical Disclosure Bullein, “Mouse Ball-Actuating Device With Force and Tactile Feedback,” vol. 32, No. 9B, Feb. 1990.
- Jacobsen et al., “High Performance, Dextrous Telerobotic Manipulator With Force Reflection,” Intervention/ROV '91 Conference & Exposition, Hollywood, Florida, May 21-23, 1991.
- Kontarinis et al., “Display of High-Frequency Tactile Information to Teleoperators,” Telemanipulator Technology and Space Telerobotics, Won S. Kim, Editor, Proc. SPIE vol. 2057, pp. 40-50, Sep. 7-9, 1993.
- McAffee, “Teleoperator Subsystem/Teleobot Demonsdtrator: Force Reflecting Hand Controller Equipment Manual,” JPL D-5172, pp. 1-50, A1-A36, B1-B5, C1-C36, Jan. 1988.
- Minsky, “Computational Haptics: The Sandpaper System for Synthesizing Texture for a Force-Feedback Display,” Ph. D. Dissertation, MIT, Jun. 1995.
- Patrick et al., “Design and Testing of A Non-reactive, Fingertip, Tactile Display for Interaction with Remote Environments,” Cooperative Intelligent Robotics in Space, Rui J. deFigueiredo et al., Editor, Proc. SPIE vol. 1987, pp. 215-222, 1990.
- Rabinowitz et al., “Multidimensional tactile displays: Identification of vibratory intensity, frequency, and contactor area,” Journal of The Acoustical Society of America, vol. 82, No. 4, Oct. 1987.
- Shimoga, “Finger Force and Touch Feedback Issues in Dexterous Telemanipulation,” Proceedings of Fourth Annual Conference on Intelligent Robotic Systems for Space Expploration, Rensselaer Polytechnic Institute, Sep. 30-Oct. 1, 1992.
- Terry et al., “Tactile Feedback In A Computer Mouse,” Proceedings of Fouteenth Annual Northeast Bioengineering Conference, University of New Hampshire, Mar. 10-11, 1988.
- “Cyberman Technical Specification,” Logitech Cyberman SWIFT Supplement, Apr. 5, 1994.
- Aukstakalnis et al., “Silicon Mirage: The Art and Science of Virtual Reality,” ISBN 0-938151-82-7, pp. 129-180, 1992.
- Bejczy et al., “A Laboratory Breadboard System For Dual-Arm Teleoperation,” SOAR '89 Workshop, JSC, Houston, TX, Jul. 25-27, 1989.
- Bejczy el al., “Kinesthetic Coupling Between Operator and Remote Manipulator,” International Computer Technology Conference, The American Society of Mechanical Engineers, San Francisco, CA, Aug. 12-15, 1980.
- Bejczy, et al., “Universal Computer Control System (UCCS) For Space Telerobots,” CH2413-3/87/0000/0318501.00 1987 IEEE, 1987.
- Eberhardt et al., “Including Dynamic Haptic Perception by The Hand: System Description and Some Results,” DSC-vol. 55-1, Dynamic Systems and Control: vol. 1, ASME 1994.
- Gobel et al., “Tactile Feedback Applied to Computer Mice,” International Journal of Human-Computer Interaction, vol. 7, No. 1, pp. 1-24, 1995.
- Kaczmarek et al., “Tactile Displays,” Virtual Environment Technologies, 1995.
- Marcus, “Touch Feedback in Surgery,” Proceedings of Virtual Reality and Medicine The Cutting Edge, Sep. 8-11, 1994.
- Ouhyoung et al., “The Development of A Low-Cost Force Feedback Joystick and Its Use in the Virtual Reality Environment,” Proceedings of the Third Pacific Conference on Computer Graphics and Applications, Pacific Graphics '95, Seoul, Korea, Aug. 21-24, 1995.
- Ouh-Young, “A Low-Cost Force Feedback Joystick and Its Use in PC Video Games,” IEEE Transactions on Consumer Electronics, vol. 41, No. 3, Aug. 1995.
- Pimentel et al., “Virtual Reality: through the new looking glass,” 2nd Edition; McGraw-Hill, ISBN 0-07-050167-X, pp. 41-202, 1994.
- Scannell, “Taking a Joystick Ride,”Computer Currents, Boston Edition, vol. 9, No. 11, Nov. 1994.
- United States Patent and Tradmark Office , Office Action U.S. Appl. No. 09/875,458, mailed Jun. 28, 2006.
- United States Patent and Tradmark Office , Office Action, U.S. Appl. No. 09/875,458, mailed Jan. 1, 2006.
- United States Patent and Tradmark Office , Office Action, U.S. Appl. No. 09/875,458, mailed Jun. 1, 2005.
- United States Patent and Tradmark Office , Office Action, U.S. Appl. No. 09/875,458, mailed Sep. 29, 2004.
- United States Patent and Tradmark Office , Office Action, U.S. Appl. No. 09/875,458, mailed Feb. 25, 2004.
- United States Patent and Tradmark Office , Office Action, U.S. Appl. No. 09/875,458, mailed Jul. 29, 2003.
- United States Patent and Tradmark Office , Office Action, U.S. Appl. No. 09/875,458, mailed Jan. 13, 2003.
- United States Patent and Tradmark Office , Office Action, U.S. Appl. No. 09/875,458, May 30, 2002.
- United States Patent and Tradmark Office , Office Action, U.S. Appl. No. 11/594,668, mailed Mar. 26, 2010.
- United States Patent and Tradmark Office , Office Action, U.S. Appl. No. 11/594,668, mailed Oct. 20, 2009.
Type: Grant
Filed: Sep 8, 1999
Date of Patent: Mar 1, 2011
Assignee: Immersion Corporation (San Jose, CA)
Inventor: Craig Culver (Woodside, CA)
Primary Examiner: William L Boddie
Attorney: Kilpatrick Stockton LLP
Application Number: 09/393,998
International Classification: G06F 3/033 (20060101); G09G 5/00 (20060101);