Information processing program

- Nintendo Co., Ltd.

A game system comprises image obtaining means, direction calculation means, first rotation means, and display control means. The image obtaining means obtains an image taken by an imaging device. The direction calculation means calculates a direction originally determined from the image of an imaging target included in the image taken by the imaging device. The first rotation means rotates an operation target in accordance with the calculated direction. The display control means generates an image in accordance with the rotation of the operation target performed by the first rotation means and displays the generated image on a screen of a display device.

Skip to: Description  ·  Claims  ·  References Cited  · Patent History  ·  Patent History
Description
CROSS REFERENCE OF RELATED APPLICATION

The disclosure of Japanese Patent Application No. 2005-264207 is incorporated herein by reference. This application also claims the benefit of Provisional Application No. 60/732,648, filed Nov. 3, 2005, the entire contents of which are hereby incorporated by in this application.

TECHNICAL FIELD

The technology herein relates to an information processing program, and more particularly to an information processing program usable for an information processing system operated by a user using an input device including an imaging device.

BACKGROUND AND SUMMARY

Conventionally, technologies for designating a specific position on a display screen using an optical pointing system have been proposed. For example, an optical pointing system described in Japanese Laid-Open Patent Publication No. 6-308879 is used for conferences or the like held for a plurality of viewers. According to this system, a light emitting element is provided in the vicinity of the display screen, and an imaging device using an optic/electric conversion element is built in an indicator. The user can designate an arbitrary position on the display screen using the indicator as follows. The indicator takes an image of the light emitting element by the imaging device and the position designated by the user is calculated based on the obtained image. A mark is displayed at the calculated position, and thus the designated position is indicated with the mark.

A game controller which uses an optical pointing system substantially the same as the above has been proposed (see, for example, Japanese Laid-Open Patent Publication No. 8-71252). The controller has a shape of a gun and is used for a shooting game. The shooting game is for shooting a target displayed on the screen with a gun. The gun is used for designating an arbitrary position (to which the bullet is to hit) on the screen. The gun-shaped controller has a built-in CCD camera, and light emitting elements located at four positions around the screen are imaged by the CCD camera. The controller calculates the rotation angle or the inclining angle of the gun using the obtained images. The controller uses the calculation result to detect whether or not the gun is correctly directed toward the target displayed on the screen, and thus determines whether or not the bullet has hit the target. With such a structure, a game system by which the player performs a game operation by moving the controller (gun) held in his/her hand can be realized.

With the above-described technologies, the input device held in the user's hand (the indicator or the gun-shaped controller) is only used for designating a position on the display screen. Namely, the above-described indicator or gun-shaped controller allows the player to perform only one operation of designating a position on the display screen but not any other operation. For example, when used for a game apparatus, such an input device allows the player to perform only one simple game operation of designating a position on the display screen. Since the game operation is so simple, the game itself is simple and is not entertaining.

Therefore, a feature of non-limiting illustrative embodiments may provide an information processing system allowing the user to perform a new type of operation using an input device held in his/her hand and a program used for such an information processing system.

The non-limiting illustrative embodiments may have the following additional features. The reference numerals, additional explanations and the like in parentheses in this section of the specification indicate the correspondence with the non-limiting illustrative embodiments described later for easier understanding and are limiting in any way.

A first aspect of the non-limiting illustrative embodiments is directed to a computer readable storage medium having stored thereon an information processing program executable by a computer (CPU 10, etc.) of an information processing apparatus (game apparatus 3). The information processing apparatus receives operation data from an operation device (controller 7) including imaging means (imaging element 40) for taking an image of an imaging target (markers 8a and 8b), and displays, on a display device (monitor 2), a virtual space obtained by performing calculation processing on a predetermined operation target (player object 71, key 76, virtual camera 81, or door knobs 85a or 86a) using the operation data. The information processing program causes the computer to execute an obtaining step (S1), a vector calculation step (S16), a first rotation step (S3) and a display step (S4). The obtaining step obtains a taken image taken by the imaging means of the operation device as the operation data. The vector calculation step calculates a two-dimensional vector (vector represented by direction data 57) using a position of the imaging target in the taken image. The first rotation step rotates the operation target in accordance with a calculated value of the two-dimensional vector. The display step displays the virtual space, changed in accordance with the rotation of the operation target in the first rotation step, on a display area of the display device.

According to a second aspect of the non-limiting illustrative embodiments, the vector calculation step may include a first calculation step and a second calculation step. The first calculation step calculates coordinate sets of two predetermined points in the image of the imaging target in a coordinate system corresponding to the taken image. The second calculation step calculates the two-dimensional vector connecting the coordinate sets of the two predetermined points.

A third aspect of the non-limiting illustrative embodiments is directed to a computer readable storage medium having stored thereon an information processing program executable by a computer (CPU 10, etc.) of an information processing apparatus (game apparatus 3). The information processing apparatus receives operation data from an operation device (controller 7) and displays, on a display device (monitor 2), a virtual space obtained by performing calculation processing on a predetermined operation target (player object 71, key 76, virtual camera 81, or door knobs 85a and 86a) using the operation data. The operation device includes imaging means (imaging element 40) for taking an image of an imaging target and first calculation means (image processing circuit 41) for calculating coordinate sets of two predetermined points in the image of the imaging target included in a taken image taken by the imaging means. The information processing program causes the computer to execute an obtaining step (S1), a vector calculation step (S16), a first rotation step (S3) and a display step (S4). The obtaining step obtains the coordinate sets of the two predetermined points as the operation data. The vector calculation step calculates a two-dimensional vector connecting the coordinate sets of the two predetermined points. The first rotation step rotates the operation target in accordance with a calculated value of the two-dimensional vector. The display step displays the virtual space, changed in accordance with the rotation of the operation target in the first rotation step, on a display area of the display device.

A fourth aspect of the non-limiting illustrative embodiments is directed to a computer readable storage medium having stored thereon an information processing program executable by a computer (CPU 10, etc.) of an information processing apparatus (game apparatus 3). The information processing apparatus receives operation data from an operation device (controller 7) and displays, on a display device (monitor 2), a virtual space obtained by performing calculation processing on a predetermined operation target (player object 71, key 76, virtual camera 81, or door knobs 85a and 86b) using the operation data. The operation device includes imaging means (imaging element 40) for taking an image of an imaging target and vector calculation means (image processing circuit 41) for calculating a two-dimensional vector using a position of the imaging target in a taken image taken by the imaging means. The information processing program causes the computer to execute an obtaining step (S1), a first rotation step (S3) and a display step (S4). The obtaining step obtains the two-dimensional vector as the operation data. The first rotation step rotates the operation target in accordance with an obtained value of the two-dimensional vector. The display step displays the virtual space, changed in accordance with the rotation of the operation target in the first rotation step, on a display area of the display device.

According to a fifth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute an indicated coordinate set calculation step (S22, S52, S62). The indicated coordinate set calculation step calculates a predetermined indicated coordinate set (indicated coordinate set data 62) which corresponds to a position on the display area, in correspondence with a position of the image of the imaging target in the taken image obtained in the obtaining step. In this case, the first rotation step rotates the operation target by a calculation using the two-dimensional vector and the indicated coordinate set.

According to a sixth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute an indicated coordinate set calculation step (S22, S52, S62). The indicated coordinate set calculation step calculates a predetermined indicated coordinate set which corresponds to a position on the display area, in correspondence with a coordinate set of an intermediate point between the two predetermined points. In this case, the first rotation step rotates the operation target by a calculation using the two-dimensional vector and the indicated coordinate set.

According to a seventh aspect of the non-limiting illustrative embodiments, the operation data may further include a coordinate set of one point corresponding to a position of the image of the imaging target. In this case, the information processing program further causes the computer to execute an indicated coordinate set calculation step (S22, S52, S62). The indicated coordinate set calculation step calculates a predetermined indicated coordinate set which corresponds to a position on the display area, in correspondence with the coordinate set of the one point. The first rotation step rotates the operation target by a calculation using the two-dimensional vector and the indicated coordinate set.

According to an eighth aspect of the non-limiting illustrative embodiments, the indicated coordinate set calculation step may calculate the indicated coordinate set which corresponds to the position on the display area, in correspondence with a position of the image of the imaging target in the case where the image taken by the imaging means is rotated around the center of the image as an axis and the two-dimensional vector is directed in one certain direction by the rotation.

According to a ninth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute an indicated coordinate set calculation step (S22, S52, S62). The indicated coordinate set calculation step sets a predetermined indicated coordinate set which corresponds to a position on the display area as an indicated coordinate set. The first rotation step rotates the operation target by a calculation using the two-dimensional vector and the indicated coordinate set.

According to a tenth aspect of the non-limiting illustrative embodiments, the virtual space may be a virtual three-dimensional space. In this case, the information processing program further causes the computer to execute a three-dimensional indicated coordinate set setting step (S53). The three-dimensional indicated coordinate set setting step calculates a three-dimensional coordinate set in the virtual space which corresponds to the indicated coordinate set calculated in the indicated coordinate set calculation step and sets the three-dimensional coordinate set as a three-dimensional indicated coordinate set. In this case, the first rotation step rotates the operation target in the three-dimensional space by a calculation using the two-dimensional vector and the three-dimensional indicated coordinate set.

According to an eleventh aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute an object location step (S31, S41, S60). The object location step locates at least one object in the virtual space. The first rotation step rotates any one of the at least one object located in the virtual space as the operation target.

According to a twelfth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute an object location step (S31, S41, S60). The object location step locates at least one object in the virtual space. In this case, the first rotation step may include a determination step (S64) of determining whether or not any one of the at least one object is displayed at the position on the display area which corresponds to the indicated coordinate set; an operation target setting step (S65) of, when it is determined that one of the at least one object is displayed at the position which corresponds to the indicated coordinate set, setting the one object as the operation target; and a step (S21) of rotating the one object which is set in the operation target setting step.

According to a thirteenth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute an object location step (S31, S41, S60). The object location step locates an object as the operation target in the virtual space. In this case, the first rotation step includes an object moving step (S21) of moving the object so as to be displayed at a position of the indicted coordinate set calculated in the indicated coordinate set calculation step; and a step (S22) of rotating the object.

According to a fourteenth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute an object location step (S31, S41, S60). The object location step locates an object as the operation target in the virtual space. In this case, the first rotation step includes an object moving step (S21) of moving the object to the three-dimensional coordinate set (corresponding coordinate set) calculated in the three-dimensional indicated coordinate set setting step; and a step (S22) of rotating the object.

According to a fifteenth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute an object location step (S31, S41, S60). The object location step locates an object as the operation target in the virtual space. In this case, the first rotation step includes a central coordinate set setting step (S53) of setting the indicated coordinate set calculated in the indicated coordinate set calculation step as a central coordinate set of rotation; and a step (S22) of rotating the object around the central coordinate set as the center of rotation.

According to a sixteenth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute an object location step (S31, S41, S60). The object location step locates an object as the operation target in the virtual space. In this case, the first rotation step includes a central coordinate set setting step (S53) of setting the three-dimensional coordinate set calculated in the three-dimensional indicated coordinate set setting step as a central coordinate set of rotation; and a step (S22) of rotating the object in the three-dimensional space around the central coordinate set as the center of rotation.

According to a seventeenth aspect of the non-limiting illustrative embodiments, the virtual space may be a virtual three-dimensional space. In this case, the information processing program further causes the computer to execute a virtual camera setting step (S50). The virtual camera setting step sets a virtual camera directed in a predetermined direction at a predetermined position in the virtual space. The first rotation step rotates the virtual camera as the operation target. The display step displays an image of the virtual space seen from the virtual camera on the display area.

According to an eighteenth aspect of the non-limiting illustrative embodiments, the virtual space may be a virtual three-dimensional space. In this case, the information processing program further causes the computer to execute a virtual camera setting step (S50). The virtual camera setting step sets a virtual camera directed in a predetermined direction at a predetermined position in the virtual space. The first rotation step rotates the virtual camera as the operation target around the three-dimensional coordinate set calculated in the three-dimensional indicated coordinate set setting step as the center of rotation.

According to a nineteenth aspect of the non-limiting illustrative embodiments, the first rotation step may rotate the virtual camera around a position of a point of attention of the virtual camera as the center of rotation.

According to a twentieth aspect of the non-limiting illustrative embodiments, the first rotation step may rotate the virtual camera such that a viewing direction of the virtual camera changes.

According to a twenty-first aspect of the non-limiting illustrative embodiments, the first rotation step may change a posture of the operation target by the rotation.

According to a twenty-second aspect of the non-limiting illustrative embodiments, the first rotation step may move a position of the operation target by rotating the operation target around a predetermined position as the center of rotation.

According to a twenty-third aspect of the non-limiting illustrative embodiments, the first rotation step may determine the posture or the position of the operation target in accordance with a calculated direction of the two-dimensional vector.

According to a twenty-fourth aspect of the non-limiting illustrative embodiments, the first rotation step may includes a determination step of determining whether or not a direction of the two-dimensional vector has exceeded a predetermined range; and a step of, when it is determined in the determination step that the direction of the two-dimensional vector has exceeded the predetermined range, rotating the operation target.

According to a twenty-fifth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute an inclination calculation step (S54). The inclination calculation step obtains a difference between a direction of the two-dimensional vector and a predetermined reference direction as an angle or a vector. In this case, the first rotation step rotates the operation target in accordance with the difference calculated in the inclination calculation step.

According to a twenty-sixth aspect of the non-limiting illustrative embodiments, the first rotation step may include a rotation amount setting step (S55) of setting a rotation amount in accordance with a magnitude of the difference calculated in the inclination calculation step; and a step (S56) of rotating the operation target by the rotation amount.

According to a twenty-seventh aspect of the non-limiting illustrative embodiments, the first rotation step may include a determination step of determining whether or not the difference calculated in the inclination calculation step has exceeded a predetermined value; and a step of, when it is determined in the determination step that the difference has exceeded the predetermined value, rotating the operation target.

According to a twenty-eighth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute a size calculation step (S19). The size calculation step calculates data representing a size of the image of the imaging target in the taken image obtained in the obtaining step from the operation data. In this case, the first rotation step rotates the operation target by a calculation using the two-dimensional vector and the data calculated in the size calculation step.

According to a twenty-ninth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute a length calculation step (S19). The length calculation step calculates data representing a length of a distance between the coordinate sets of the two predetermined points. In this case, the first rotation step rotates the operation target by a calculation using the two-dimensional vector and the data calculated in the length calculation step.

According to a thirtieth aspect of the non-limiting illustrative embodiments, the operation data may further include data representing a length of a distance between the coordinate sets of the two predetermined points in the image of the imaging target included in the taken image taken by the imaging means. In this case, the first rotation step rotates the operation target by a calculation using the two-dimensional vector and the data representing the length of the distance between the coordinate sets of the two predetermined points.

According to a thirty-first aspect of the non-limiting illustrative embodiments, the first rotation step may include a step (S48) of calculating distance data representing a distance between the imaging means and the imaging target from the data calculated in the size calculation step; and a step of rotating the operation target by a calculation using the two-dimensional vector and the distance data.

According to a thirty-second aspect of the non-limiting illustrative embodiments, first rotation step may include a step (S48) of calculating distance data representing a distance between the imaging means and the imaging target from the data calculated in the length calculation step; and a step of rotating the operation target by a calculation using the two-dimensional vector and the distance data.

According to a thirty-third aspect of the non-limiting illustrative embodiments, the first rotation step may include a step (S48) of calculating distance data representing a distance between the imaging means and the imaging target from the data representing the length of the distance between the coordinate sets of the two predetermined points; and a step of rotating the operation target by a calculation using the two-dimensional vector and the distance data.

According to a thirty-fourth aspect of the non-limiting illustrative embodiments, the operation data may further include distance data representing a distance between the imaging means and the imaging target. In this case, the first rotation step rotates the operation target by a calculation using the two-dimensional vector and the distance data.

According to a thirty-fifth aspect of the non-limiting illustrative embodiments, the information processing program may further cause the computer to execute a distance calculation step (S48). The distance calculation step calculates data corresponding to a distance between the imaging means and the imaging target. In this case, the first rotation step rotates the operation target by a calculation using the two-dimensional vector and the data calculated in the distance calculation step.

According to a thirty-sixth aspect of the non-limiting illustrative embodiments, the virtual space may be a virtual three-dimensional space. In this case, the information processing program further causes the computer to execute an object location step. The object location step locates at least one object in the virtual space. In this case, the first rotation step includes a depth setting step of determining a coordinate in a depth direction of the virtual space in correspondence with the distance data; a determination step of determining whether or not there is any one of the at least one object at a position of the coordinate in the depth direction calculated in the depth setting step; an operation target setting step of, when it is determined in the determination step that there is one of the at least one object, setting the one object as the operation target.

According to a thirty-seventh aspect of the non-limiting illustrative embodiments, the virtual space may be a virtual three-dimensional space. In this case, the first rotation step includes a depth setting step (S49) of determining a coordinate in a depth direction of the virtual space in correspondence with the distance data; a moving step of moving a position of the operation target in the depth direction to the coordinate in the depth direction calculated in the depth setting step; and a step of rotating the operation target in accordance with the two-dimensional vector.

According a thirty-eighth aspect of the non-limiting illustrative embodiments, the operation device may comprise inclination determining means (including acceleration sensor 37) for determining an inclination of the operation device. In this case, the information processing program further causes the computer to execute a determination step and a second rotation step. The determination step determines whether or not a two-dimensional vector can be calculated from the taken image in the vector calculation step. The second rotation step, when it is determined in the determination step that the two-dimensional vector cannot be calculated, rotates the operation target in accordance with the inclination determined by the inclination determining means.

According to the first aspect, a two-dimensional vector is calculated in the vector calculation step from the image of the imaging target (taken image). The value of the two-dimensional vector changes in accordance with the rotation state of the operation device including the imaging device. Accordingly, by displaying the operation target so as to rotate in accordance with the direction of the two-dimensional vector, the operation target can be rotated in accordance with the operation of rotating the operation device. Namely, according to the first aspect, a novel operation method by which the user can rotate the operation target by rotating the operation device itself held in his/her hand is provided. Also according to the first aspect, the taken image is obtained from the operation device, and therefore the operation device only needs to have a function of taking an image. Thus, the structure of the operation device can be simplified.

According to the second aspect, the positions of the two predetermined points in the image of the imaging target are calculated. Therefore, a two-dimensional vector can be easily calculated using the positions of the two points.

According to the third aspect, a novel operation method by which the user can rotate the operation target by rotating the operation device itself held in his/her hand is provided. Also according to the third aspect, the positions of the two predetermined points in the image of the imaging target are calculated by the operation device. Therefore, the information processing apparatus can easily calculate a two-dimensional vector using the positions of the two points. Thus, the processing load of the computer of the information processing apparatus can be alleviated.

According to the fourth aspect, a novel operation method by which the user can rotate the operation target by rotating the operation device itself held in his/her hand is provided. Also according to the fourth aspect, a two-dimensional vector is calculated by the operation device. Therefore, the information processing apparatus does not need to execute processing of calculating the two-dimensional vector. Thus, the processing load of the computer of the information processing apparatus can be alleviated.

According to the fifth aspect, a position on the display area is calculated in the indicated coordinate set calculation step from the position of the image of the imaging target in the taken image. Therefore, the user can designate a position on the display area with the operation device usable while being held in his/her hand. In addition, the operation target is rotated using the position designated by the user and the two-dimensional vector. Therefore, a more complicated rotation operation is made possible with the operation device usable while being held in the user's hand, and the user can control the rotation of the operation target more precisely. For example, the user can designate the central position of the rotation, or can rotate the operation target while moving the operation target to a designated position.

According to the sixth aspect, a position on the display area is calculated in the indicated coordinate set calculation step from the position of the image of the imaging target in the taken image. Therefore, the user can designate a position on the display area with the operation device usable while being held in his/her hand. In addition, the operation target is rotated using the position designated by the user and the two-dimensional vector. Therefore, a more complicated rotation operation is made possible with the operation device usable while being held in the user's hand, and the user can control the rotation of the operation target more precisely.

According to the seventh aspect, a position on the display area is calculated in the indicated coordinate set calculation step from the position of the image of the imaging target in the taken image. Therefore, the user can designate a position on the display area with the operation device usable while being held in his/her hand. In addition, the operation target is rotated using the position designated by the user and the two-dimensional vector. Therefore, a more complicated rotation operation is made possible with the operation device usable while being held in the user's hand, and the user can control the rotation of the operation target more precisely.

The position of the image of the imaging target in the taken image is different in accordance with the rotation state of the operation device (see FIG. 15). Namely, even if the operation device indicates one position, the position of the image of the imaging target may be different depending on the rotation state of the operation device. In this case, the position indicated by the operation device (i.e., the position of the indicated coordinate set to be calculated in the indicated coordinate set calculation step) cannot be accurately calculated from the position of the image of the imaging target. By contrast, according to the eighth aspect, the position of the image of the imaging target obtained when the taken image is rotated so as to be directed in one certain direction, is calculated. Therefore, the influence of the rotation state of the operation device can be eliminated, and the position on the display screen or in the virtual space can be accurately calculated from the position of the image of the imaging target in the indicated coordinate set calculation step.

According to the ninth aspect, a position on the display area is calculated in the indicated coordinate set calculation step from the position of the image of the imaging target in the taken image. Therefore, the user can designate a position on the display area with the operation device usable while being held in his/her hand. In addition, the operation target is rotated using the position designated by the user and the two-dimensional vector. Therefore, a more complicated rotation operation is made possible with the operation device usable while being held in the user's hand, and the user can control the rotation of the operation target more precisely.

According to the tenth aspect, the user can move the operation target which is present in the virtual three-dimensional space by a novel operation of rotating the operation device itself. Also according to the tenth aspect, a more complicated rotation operation is made possible with the operation device usable while being held in the user's hand, and the user can control the rotation of the operation target which is present in the virtual three-dimensional space more precisely.

According to the eleventh aspect, the user can move the object which is present in the virtual three-dimensional space by a novel operation of rotating the operation device itself.

According to the twelfth aspect, the object displayed at the position on the display area corresponding to the indicated coordinate set calculated in the indicated coordinate set calculation step is set as the operation target. Therefore, the user can perform an operation of designating the object displayed on the display screen using the operation device. In addition, the user can rotate the designated object using the operation device. Thus, the user can issue two types of instructions, i.e., an instruction regarding the selection of the object as the operation target and an instruction to rotate the selected object, with one type of operation, i.e., an operation on a posture of one operation device while holding the operation device in his/her hand.

According to the thirteenth aspect, the object is moved to the position on the display area corresponding to the indicated coordinate set calculated in the indicated coordinate set calculation step. Therefore, the user can perform an operation of moving the object displayed on the display screen using the operation device. In addition, the user can rotate the designated object using the operation device. Thus, the user can issue two types of instructions, i.e., an instruction to move the object as the operation target and an instruction to rotate the object, with one type of operation, i.e., an operation on a posture of one operation device while holding the operation device in his/her hand.

According to the fourteenth aspect, the object is moved to the position of the three-dimensional coordinate set corresponding to the indicated coordinate set calculated in the three-dimensional indicated coordinate set calculation step. Therefore, the user can perform an operation of moving the object in the virtual three-dimensional space using the operation device. In addition, the user can rotate the designated object using the operation device. Thus, the user can issue two types of instructions, i.e., an instruction to move the object as the operation target and an instruction to rotate the object, with one type of operation, i.e., an operation on a posture of one operation device while holding the operation device in his/her hand.

According to the fifteenth aspect, the user can designate a position on the display screen using the operation device usable while being held in his/her hand. The user can designate the central position of the rotation of the object using the operation device. Since the number of types of rotation is increased, the user can cause the object to perform a wider variety of rotation motions.

According to the sixteenth aspect, the user can designate a position in the virtual three-dimensional space using the operation device usable while being held in his/her hand. The user can designate the central position of the rotation of the object using the operation device. Since the number of types of rotation is increased, the user can cause the object to perform a wider variety of rotation motions.

According to the seventeenth aspect, by rotating the virtual camera in the first rotation step, the position of the viewing point or the viewing direction with respect to the image in the virtual space displayed on the display screen can be changed. The user can change the position of the viewing point or the viewing direction with respect to the image in the virtual space displayed on the display screen using a novel operation of rotating the operation device itself.

According to the eighteenth aspect, the user can designate a position on the display screen using the operation device usable while being held in his/her hand. The user can designate the central position of the rotation of the virtual camera using the operation device. Since the number of types of rotation is increased, the user can cause the virtual camera to perform a wider variety of rotation motions.

According to the nineteenth aspect, the viewing direction with respect to the object which is present at the position of the point of attention can be freely changed by a novel operation of rotating the operation device itself.

According to the twentieth aspect, the viewing direction of the virtual camera can be freely changed by a novel operation of rotating the operation device itself.

According to the twenty-first aspect, the posture of the operation target can be freely changed by a novel operation of rotating the operation device itself.

According to the twenty-second aspect, the operation target can be freely rotated or moved on an arc by a novel operation of rotating the operation device itself.

According to the twenty-third aspect, the operation target can be rotated so as to correspond to the rotation angle of the operation device.

According to the twenty-fourth aspect, when the calculated direction of the two-dimensional vector is within the predetermined range, the operation target is not rotated. If the operation target is rotated as a result of responding too sensitively to the rotation angle of the operation device, the user is required to operate the operation device precisely, which deteriorates the operability of the operation device. By contrast, according to the twenty-fourth aspect, the operation target is not rotated while being in the predetermined range of angles. A so-called “play” margin can be set in the rotation operation on the operation device. Thus, the operability can be improved.

According to the twenty-fifth aspect, the operation target is rotated based on the difference between the direction of the two-dimensional vector and the predetermined reference direction. By setting the direction with which the user feel comfortable as the reference direction, the operability of the operation device can be improved.

According to the twenty-sixth aspect, the rotation amount of the operation target is increased in accordance with the difference between the direction of the two-dimensional vector and the predetermined reference direction. Therefore, the user can control the rotation amount of the operation target by the rotation amount of the operation device. Thus, the user can intuitively and easily control the rotation amount of the operation target.

According to the twenty-seventh aspect, the operation target is not rotated unless the difference between the direction of the two-dimensional vector and the predetermined reference direction exceeds the predetermined range. By setting the range in which the angular difference does not exceed the predetermined range as the so-called “play” range, the operability of the operation device can be improved like in the twenty-fourth aspect.

According to the twenty-eighth aspect, the distance from the imaging means (operation device) to the imaging target can be detected by data representing the size of the image of the imaging target. Therefore, the user can perform an operation of rotating the operation target by moving the operation device itself so as to change the distance. Thus, a more complicated rotation operation is made possible with the operation device usable while being held in the user's hand, and the user can control the rotation of the operation target more precisely.

According to the twenty-ninth aspect, the distance from the imaging means (operation device) to the imaging target can be detected by data representing the length of the distance between the coordinate sets of two points corresponding to the two positions in the imaging target. Therefore, the user can perform an operation of rotating the operation target by moving the operation device itself so as to change the distance. Thus, a more complicated rotation operation is made possible with the operation device usable while being held in the user's hand, and the user can control the rotation of the operation target more precisely.

According to the thirtieth aspect, like in the twenty-eighth and twenty-ninth aspects, a more complicated rotation operation is made possible with the operation device usable while being held in the user's hand, and the user can control the rotation of the operation target more precisely.

According to the thirty-first aspect, the distance from the imaging means to the imaging target can be more accurately reflected on the rotation operation on the operation target.

According to the thirty-second aspect, the distance from the imaging means to the imaging target can be more accurately reflected on the rotation operation on the operation target.

According to the thirty-third aspect, the distance from the imaging means to the imaging target can be more accurately reflected on the rotation operation on the operation target.

According to the thirty-fourth aspect, the distance from the imaging means to the imaging target can be more accurately reflected on the rotation operation on the operation target.

According to the thirty-fifth aspect, data corresponding to the distance from the imaging means (operation device) to the imaging target (length data 59 or distance data 63) is calculated. Therefore, the user can perform an operation of rotating the operation target by moving the operation device itself so as to change the distance. Thus, a more complicated rotation operation is made possible with the operation device usable while being held in the user's hand, and the user can control the rotation of the operation target more precisely.

According to the thirty-sixth aspect, the object which is present at the position corresponding to the coordinate in the depth direction determined in the depth setting step is set as the operation target. Therefore, the user can perform an operation of designating the object in the virtual space using the operation device. In addition, the user can rotate the designated object using the operation device. Thus, the user can issue two types of instructions, i.e., an instruction regarding the selection of the object as the operation target and an instruction to rotate the selected object, with one type of operation, i.e., an operation of moving the operation device itself while holding the operation device in his/her hand.

According to the thirty-seventh aspect, the operation target is moved to the position corresponding to the coordinate in the depth direction determined in the depth setting step. Therefore, the user can perform an operation of moving the operation target in the virtual space using the operation device. In addition, the user can rotate the operation target using the operation device. Thus, the user can issue two types of instructions, i.e., an instruction to move the operation target and an instruction to rotate the operation target, with one type of operation, i.e., an operation of moving the operation device itself while holding the operation device in his/her hand.

According to the thirty-eighth aspect, when the operation device is held by the user in the state of exceeding a range in which the imaging means can take an image of the imaging target (operable range described later), the operation target is rotated in accordance with the inclination determined by the inclination determining means. Therefore, even if the operable range is exceeded while the user is performing an operation of moving the operation device, the rotation of the operation target can be continued.

The above features and aspects may be combined to form yet further embodiments. These and other features, aspects, and advantages of the non-limiting illustrative embodiments will become more apparent from the following detailed description of the non-limiting illustrative embodiments when taken in conjunction with the accompanying drawings.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 is an external view of a game system 1 as an exemplary information processing system according to a non-limiting illustrative embodiment;

FIG. 2 is a functional block diagram of a game apparatus 3;

FIG. 3A is an isometric view of a controller 7 seen from the top rear side thereof;

FIG. 3B is an isometric view of the controller 7 seen from the bottom rear side thereof;

FIG. 4 is a front view of the controller 7;

FIG. 5A is an isometric view illustrating a state where an upper casing of the controller 7 is removed;

FIG. 5B is an isometric view illustrating a state where a lower casing of the controller 7 is removed;

FIG. 6 is a block diagram illustrating a structure of the controller 7;

FIG. 7 shows a general view of a game operation using the controller 7;

FIG. 8 illustrates the viewing angles of markers 8a and 8b and the controller 7;

FIG. 9 shows an example of a taken image including a target image;

FIG. 10 shows a change of the taken image when the position and/or the direction of the controller 7 is changed;

FIG. 11 shows main data stored on a main memory 13 of the game apparatus 3;

FIG. 12 is a flowchart illustrating a flow of game processing executed by the game apparatus 3;

FIG. 13 is a flowchart illustrating a detailed flow of operation state calculation processing (S2) shown in FIG. 12;

FIG. 14 illustrates processing for determining a first coordinate set and a second coordinate set;

FIG. 15 shows the correspondence between the state of the controller 7 and the taken image when the controller 7 is in the respective state;

FIG. 16 is a flowchart illustrating a detailed flow of moving processing (S3) shown in FIG. 12;

FIG. 17A, FIG. 17B and FIG. 17C show examples of a rotation motion of an operation target;

FIG. 18A and FIG. 18B show examples of a rotation motion of an operation target;

FIG. 19 shows an example of a game screen in a first example;

FIG. 20 illustrates an operation of a player object 71 by a rotation operation;

FIG. 21 is a flowchart illustrating a flow of game processing in the first example;

FIG. 22 shows an example of a game screen in a second example;

FIG. 23 is a flowchart illustrating a flow of game processing in the second example;

FIG. 24 is a flowchart illustrating a detailed flow of the moving processing in the second example;

FIG. 25 is a flowchart illustrating a detailed flow of the processing in step S48 shown in FIG. 24;

FIG. 26A and FIG. 26B show the processing in step S48 shown in FIG. 24;

FIG. 27 shows an example of a game screen in a third example;

FIG. 28 shows main data stored on the main memory 13 in the third example;

FIG. 29 is a flowchart illustrating a flow of game processing in the third example;

FIG. 30 is a flowchart illustrating a detailed flow of the moving processing in the third example;

FIG. 31 shows an example of a game screen in a fourth example; and

FIG. 32 is a flowchart illustrating a flow of game processing in the fourth example.

DESCRIPTION OF THE EMBODIMENTS

With reference to FIG. 1, a game system 1 as an example of an information processing system according to a non-limiting illustrative embodiment will be described. FIG. 1 is an external view illustrating the game system 1. In the following exemplary description, the game system 1 is of an installation type.

As shown in FIG. 1, the game system 1 includes an installation type game apparatus (hereinafter, referred to simply as a “game apparatus”) 3, which is connected to a display (hereinafter, referred to as a “monitor”) 2 such as a home-use TV receiver including a speaker 22 via a connection cord, and a controller 7 for giving operation data to the game apparatus 3. Two markers 8a and 8b are provided in the vicinity of the monitor 2 (above the screen of the monitor 2 in FIG. 1). The markers 8a and 8b are specifically infrared LEDs, and each outputs infrared light forward from the monitor 2. The game apparatus 3 is connected to a receiving unit 6 via a connection terminal. The receiving unit 6 receives operation data which is wirelessly transmitted from the controller 7. The controller 7 and the game apparatus 3 are connected to each other by wireless communication. On the game apparatus 3, an optical disc 4 as an example of an exchangeable information storage medium is detachably mounted. The game apparatus 3 has, on a top main surface thereof, a power ON/OFF switch, a game processing reset switch, and an OPEN switch for opening a top lid of the game apparatus 3. When a player presses the OPEN switch, the lid is opened, so that the optical disc 4 is mounted or dismounted.

On the game apparatus 3, an external memory card 5 is detachably mounted when necessary. The external memory card 5 has a backup memory or the like mounted thereon for fixedly storing saved data or the like. The game apparatus 3 executes a game program or the like stored on the optical disc 4 and displays the result on the monitor 2 as a game image. The game apparatus 3 can also reproduce a state of a game played in the past using saved data stored on the memory card 5 and display the game image on the monitor 2. The player playing with the game apparatus 3 can enjoy the game by operating the controller 7 while watching the game image displayed on the display screen of the monitor 2.

The controller 7 wirelessly transmits operation data from a communication section 36 included therein (described later) to the game apparatus 3 connected to the receiving unit 6, using the technology of, for example, Bluetooth (registered trademark). The controller 7 is operation means for operating an operation target (an object displayed on the display screen of the monitor 2 or a virtual camera). The controller 7 includes an operation section having a plurality of operation buttons. As described later in detail, the controller 7 also includes an imaging information calculation section 35 for detecting the position and the posture of the controller 7. The imaging information calculation section 35 takes an image of each of the markers 8a and 8b located in the vicinity of the monitor 2. The game apparatus 3 uses these images to obtain an operation signal corresponding to the position and the posture of the controller 7.

With reference to FIG. 2, a structure of the game apparatus 3 will be described. FIG. 2 is a functional block diagram of the game apparatus 3.

As shown in FIG. 2, the game apparatus 3 includes, for example, a RISC CPU (central processing unit) 10 for executing various types of programs. The CPU 10 executes a start program stored in a boot ROM (not shown) to, for example, initialize memories including a main memory 13, and then executes a game program stored on the optical disc 4 to perform game processing or the like in accordance with the game program. The CPU 10 is connected to a GPU (Graphics Processing Unit) 12, the main memory 13, a DSP (Digital Signal Processor) 14, and an ARAM (Audio RAM) 15 via a memory controller 11. The memory controller 11 is connected to a controller I/F (interface) 16, a video I/F 17, an external memory I/F 18, an audio I/F 19, and a disc I/F 21 via a predetermined bus. The controller I/F 16, the video I/F 17, the external memory I/F 18, the audio I/F 19 and the disc I/F 21 are respectively connected to the receiving unit 6, the monitor 2, the external memory card 5, the speaker 22 and a disc drive 20.

The GPU 12 performs image processing based on an instruction from the CPU 10. The GPU 12 includes, for example, a semiconductor chip for performing calculation processing necessary for displaying 3D graphics. The GPU 12 performs the image processing using a memory dedicated for image processing (not shown) and a part of the storage area of the main memory 13. The GPU 12 generates game image data and a movie to be displayed on the display screen of the monitor 2 using such memories, and outputs the generated data or movie to the monitor 2 via the memory controller 11 and the video I/F 17 as necessary.

The main memory 13 is a storage area used by the CPU 10, and stores a game program or the like necessary for processing performed by the CPU 10 as necessary. For example, the main memory 13 stores a game program read from the optical disc 4 by the CPU 10, various types of data or the like. The game program, the various types of data or the like stored in the main memory 13 are executed by the CPU 10.

The DSP 14 processes sound data or the like generated by the CPU 10 during the execution of the game program. The DSP 14 is connected to the ARAM 15 for storing the sound data or the like. The ARAM 15 is used when the DSP 14 performs predetermined processing (for example, storage of the game program or sound data already read). The DSP 14 reads the sound data stored in the ARAM 15 and outputs the sound data to the speaker 22 included in the monitor 2 via the memory controller 11 and the audio I/F 19.

The memory controller 11 comprehensively controls data transfer, and is connected to the various I/Fs described above. The controller I/F 16 includes, for example, four controller I/Fs, and communicably connects the game apparatus 3 to an external device which is engageable via connectors of the controller I/Fs. For example, the receiving unit 6 is engaged with such a connector and is connected to the game apparatus 3 via the controller I/F 16. As described above, the receiving unit 6 receives the operation data from the controller 7 and outputs the operation data to the CPU 10 via the controller I/F 16. In other embodiments, the game apparatus 3 may include a receiving module for receiving the operation data transmitted from the controller 7, instead of the receiving unit 6. In this case, the operation data received by the receiving module is output to the CPU 10 via a predetermined bus. The video I/F 17 is connected to the monitor 2. The external memory I/F 18 is connected to the external memory card 5 and is accessible to a backup memory or the like provided in the external card 5. The audio I/F 19 is connected to the speaker 22 built in the monitor 2, and is connected such that the sound data read by the DSP 14 from the ARAM 15 or sound data directly output from the disc drive 20 is output from the speaker 22. The disc I/F 21 is connected to the disc drive 20. The disc drive 20 reads data stored at a predetermined reading position of the optical disc 4 and outputs the data to a bus of the game apparatus 3 or the audio I/F 19.

With reference to FIG. 3A through FIG. 7, the controller 7 will be described. FIG. 3A through FIG. 5B are external isometric views of the controller 7. FIG. 3A is an isometric view of the controller 7 seen from the top rear side thereof. FIG. 3B is an isometric view of the controller 7 seen from the bottom rear side thereof. FIG. 4 is a front view of the controller 7.

As shown in FIG. 3A, FIG. 3B and FIG. 4, the controller 7 includes a housing 31 formed by plastic molding or the like. The housing 31 has a generally parallelepiped shape extending in a longitudinal or front-rear direction (the Z-axis direction shown in FIG. 3A). The overall size of the housing 31 is small enough to be held by one hand of an adult or even a child. The player can use the controller 7 to perform a game operation of pressing buttons provided thereon and also to perform a game operation of changing the position or direction of the controller 7 itself. For example, the player can rotate the controller 7 with the longitudinal direction thereof as an axis to move an operation target. The player can change the position indicated by the controller 7 on the display screen to, for example, move the operation target toward the post-change position. The “position indicated by the controller 7 on the display screen” refers to a position at which a phantom straight line extending from a front end of the controller 7 in the longitudinal direction crosses the display screen of the monitor 2. Hereinafter, such a position will be sometimes referred to as an “indicated position” or an “indicated position by the controller 7”. The longitudinal direction of the controller 7 (housing 31) will be sometimes referred to as an “indicated direction”.

The housing 31 has a plurality of operation buttons. Provided on a top surface of the housing 31 are a cross key 32a, an X button 32b, a Y button 32c, a B button 32d, a select switch 32e, a menu switch 32f, and a start switch 32g. On a bottom surface of the housing 31, a recessed portion is formed. On a rear slope surface of the recessed portion, an A button 32i is provided. These buttons and switches are assigned various functions in accordance with the game program executed by the game apparatus 3, but this will not be described in detail. On the top surface of the housing 31, a power switch 32h is provided for remotely turning on or off the game apparatus 3.

The controller 7 has the imaging information calculation section 35 (FIG. 5B). As shown in FIG. 4, a light incident opening 35a of the imaging information calculation section 35 is provided on a front surface of the housing 31. On a rear surface of the housing 31, a connector 33 is provided. The connector 33 is, for example, a 32-pin edge connector, and is used for connecting the controller 7 to another device. In a rear part of the top surface of the housing 31, a plurality of LEDs 34 are provided. The controller 7 is assigned a controller type (number) so as to be distinguishable from the other controllers 7. The LEDs 34 are used for informing the player of the controller type which is currently set to controller 7 that he/she is using. Specifically, when the controller 7 transmits the operation data to the game apparatus 3, one of the plurality of LEDs 34 corresponding to the controller type is lit up.

With reference to FIG. 5A, FIG. 5B and FIG. 6, an internal structure of the controller 7 will be described. FIG. 5A and FIG. 5B illustrate an internal structure of the controller 7. FIG. 5A is an isometric view illustrating a state where an upper casing (a part of the housing 31) of the controller 7 is removed. FIG. 5B is an isometric view illustrating a state where a lower casing (a part of the housing 31) of the controller 7 is removed. FIG. 5B shows a reverse side of a substrate 300 shown in FIG. 5A.

As shown in FIG. 5A, the substrate 300 is fixed inside the housing 31. On a top main surface of the substrate 300, the operation buttons 32a through 32h, an acceleration sensor 37, the LEDs 34, a quartz oscillator 46, a wireless module 44, an antenna 45 and the like are provided. These elements are connected to a microcomputer 42 (see FIG. 6) via lines (not shown) formed on the substrate 300 and the like. The wireless module 44 and the antenna 45 allow the controller 7 to act as a wireless controller. The quartz oscillator 46 generates a reference clock of the microcomputer 42 described later.

As shown in FIG. 5B, at a front edge of a bottom main surface of the substrate 300, the imaging information calculation section 35 is provided. The imaging information calculation section 35 includes an infrared filter 38, a lens 39, an imaging element 40 and an image processing circuit 41 located in this order from the front surface of the controller 7. These elements are attached to the bottom main surface of the substrate 300. At a rear edge of the bottom main surface of the substrate 300, the connector 33 is attached. The operation button 32i is attached on the bottom main surface of the substrate 300 rearward to the imaging information calculation section 35, and cells 47 are accommodated rearward to the operation button 32i. On the bottom main surface of the substrate 300 between the cells 47 and the connector 33, a vibrator 48 is attached. The vibrator 48 may be, for example, a vibration motor or a solenoid. The controller 7 is vibrated by an actuation of the vibrator 48, and the vibration is conveyed to the player holding the controller 7. Thus, a so-called vibration-responsive game is realized.

FIG. 6 is a block diagram showing the structure of the controller 7. The controller 7 includes the communication section 36 and the acceleration sensor 37 in addition to the above-described operation section 32 (operation buttons) and the imaging information calculation section 35.

The imaging information calculation section 35 is a system for analyzing image data taken by imaging means and detecting the position of the center of gravity, the size and the like of an area having a high brightness in the image data. The imaging information calculation section 35 has, for example, a maximum sampling period of about 200 frames per second, and therefore can trace and analyze even a relatively fast motion of the controller 7.

Specifically, the imaging information calculation section 35 includes the infrared filter 38, the lens 39, the imaging element 40 and the image processing circuit 41. The infrared filter 38 allows only infrared light to pass therethrough, among light incident on the front surface of the controller 7. The markers 8a and 8b located in the vicinity of the display screen of the monitor 2 are infrared LEDs for outputting infrared light forward from the monitor 2. Therefore, the provision of the infrared filter 38 allows the image of each of the markers 8a and 8b to be taken more accurately. The lens 39 collects the infrared light which has passed through the infrared filter 38 and outputs the infrared light to the imaging element 40. The imaging element 40 is a solid-state imaging device such as, for example, a CMOS sensor or a CCD. The imaging element 40 takes an image of the infrared light collected by the lens 39. Accordingly, the imaging element 40 takes an image of only the infrared light which has passed through the infrared filter 38 and generates image data. Hereinafter, an image taken by the imaging element 40 will be referred to as a “taken image”. The image data generated by the imaging element 40 is processed by the image processing circuit 41. The image processing circuit 41 calculates the positions of the imaging targets (the markers 8a and 8b) in the taken image, and outputs coordinate sets indicating the respective positions of the markers 8a and 8b in the taken image to the communication section 36. The processing executed by the image processing circuit 41 will be described later in detail.

The acceleration sensor 37 detects an acceleration in three axial directions of the controller 7, i.e., the up-down direction (Y-axis direction in FIG. 3A), the left-right direction (X-axis direction in FIG. 3A), and the front-rear direction (the Z-axis direction in FIG. 3A). The acceleration sensor 37 allows the inclinations of the controller 7 in the X-axis, Y-axis and Z-axis directions to be determined. The game apparatus 3 can also determine the rotation angle of the controller 7 around the Z axis by the acceleration sensor 37, in addition to based on the taken image mentioned above. Data representing the accelerations detected by the acceleration sensor 37 is output to the communication section 36.

As explained above, the controller 7 preferably includes a three-axis, linear acceleration sensor 37 that detects linear acceleration in each of the three axial directions described above. Alternatively, a two axis linear accelerometer that only detects linear acceleration along each of the X-axis and Y-axis (or other pair of axes) may be used in another embodiment depending on the type of control signals desired. As a non-limiting example, the three-axis or two-axis linear accelerometer 37 may be of the type available from Analog Devices, Inc. or STMicroelectronics N.V. Preferably, the acceleration sensor 37 is an electrostatic capacitance or capacitance-coupling type that is based on silicon micro-machined MEMS (microelectromechanical systems) technology. However, any other suitable accelerometer technology (e.g., piezoelectric type or piezoresistance type) now existing or later developed may be used to provide the three-axis or two-axis acceleration sensor 37.

As one skilled in the art understands, linear accelerometers, as used in acceleration sensor 37, are only capable of detecting acceleration along a straight line corresponding to each axis of the acceleration sensor. In other words, the direct output of the acceleration sensor 37 is limited to signals indicative of linear acceleration (static or dynamic) along each of the two or three axes thereof. As a result, the acceleration sensor 37 cannot directly detect movement along a non-linear (e.g. arcuate) path, rotation, rotational movement, angular displacement, tilt, position, attitude or any other physical characteristic.

However, through additional processing of the linear acceleration signals output from the acceleration sensor 37, additional information relating to the controller 7 can be inferred or calculated, as one skilled in the art will readily understand from the description herein. For example, by detecting static, linear acceleration (i.e., gravity), the linear acceleration output of the acceleration sensor 37 can be used to infer or calculate tilt or inclination of the object relative to the gravity vector by correlating tilt angles with detected linear acceleration. In this way, the acceleration sensor 37 can be used in combination with the micro-computer 42 (or another processor) to determine tilt, attitude or position of the controller 7. Similarly, various movements and/or positions of the controller 7 can be calculated or inferred through processing of the linear acceleration signals generated by the acceleration sensor 37 when the controller 7 containing the acceleration sensor 37 is subjected to dynamic accelerations by, for example, the hand of a user. In another embodiment, the acceleration sensor 37 may include an embedded signal processor or other type of dedicated processor for performing any desired processing of the acceleration signals output from the accelerometers therein prior to outputting signals to micro-computer 42. For example, the embedded or dedicated processor could convert the detected acceleration signal to a corresponding tilt angle when the acceleration sensor is intended to detect static acceleration (i.e., gravity).

In another exemplary embodiment, the acceleration sensor 37 may be replaced with a gyro-sensor of any suitable technology incorporating, for example, a rotating or vibrating element. Exemplary MEMS gyro-sensors that may be used in this embodiment are available from Analog Devices, Inc. Unlike the linear acceleration sensor 37, a gyro-sensor is capable of directly detecting rotation (or angular rate) around an axis defined by the gyroscopic element (or elements) therein. Thus, due to the fundamental differences between a gyro-sensor and a linear acceleration sensor (e.g., angle-based vs. vector-based output), corresponding changes need to be made to the processing operations that are performed on the output signals from these devices depending on which device is selected for a particular application. More specifically, when tilt or inclination is calculated using a gyro-sensor instead of the acceleration sensor, significant changes are necessary. Specifically, when using a gyro-sensor, the value of inclination is initialized at the start of detection. Then, data on the angular velocity which is output from the gyroscope is integrated. Furthermore, a change amount in inclination from the value of inclination previously initialized is calculated. In this case, the calculated inclination corresponds to an angle. In contrast, when an acceleration sensor is used, the inclination is calculated by comparing the value of the acceleration of gravity of each axial component with a predetermined reference. Therefore, the calculated inclination can be represented as a vector. Thus, without initialization, an absolute direction can be determined with an accelerometer. The type of the value calculated as an inclination is also very different between a gyroscope and an accelerometer; i.e., the value is an angle when a gyroscope is used and is a vector when an acceleration sensor is used. Therefore, when a gyroscope is used instead of an acceleration sensor or vice versa, data on inclination also needs to be processed by a predetermined conversion that takes into account the fundamental differences between these two devices. Due to the fact that the nature of gyroscopes is known to one skilled in the art, as well as the fundamental differences between linear accelerometers and gyroscopes, further details are not provided herein so as not to obscure the remainder of the disclosure. While gyro-sensors provide certain advantages due to their ability to directly detect rotation, linear acceleration sensors are generally more cost effective when used in connection with the controller applications described herein.

The communication section 36 includes the microcomputer 42, a memory 43, the wireless module 44 and the antenna 45. The microcomputer 42 controls the wireless module 44 for wirelessly transmitting the data obtained by the microcomputer 42 while using the memory 43 as a storage area during processing.

Data which is output from the operation section 32, the acceleration sensor 37 and the imaging information calculation section 35 to the microcomputer 42 is temporarily stored in the memory 43. The wireless transmission from the communication section 36 to the receiving unit 6 is performed at a predetermined time interval. Since game processing is generally performed at a cycle of 1/60 sec., the wireless transmission needs to be performed at a cycle of a shorter time period. At the transmission timing to the receiving unit 6, the microcomputer 42 outputs the data stored in the memory 43 to the wireless module 44 as operation data. The wireless module 44 uses, for example, the Bluetooth (registered trademark) technology to modulate a carrier wave of a predetermined frequency with the operation data and radiate the resultant very weak electric signal from the antenna 45. Namely, the operation data is modulated into a very weak electric signal by the wireless module 44 and transmitted from the controller 7. The very weak electric signal is received by the receiving unit 6 on the side of the game apparatus 3. The received very weak electric signal is demodulated or decoded, so that the game apparatus 3 can obtain the operation data. The CPU 10 of the game apparatus 3 executes the game processing based on the obtained operation data and the game program.

The shape of the controller 7, and the shape, number, position or the like of the operation buttons and switches shown in FIG. 3A through FIG. 5B are merely exemplary non-limiting embodiments, and thus may be altered to form yet further embodiments. The position of the imaging information calculation section 35 in the controller 7 (the light incident opening 35a of the imaging information calculation section 35) does not need to be on the front surface of the housing 31, and may be on another surface as long as light can enter from the outside of the housing 31. In this case, the “indicated direction” is a direction vertical to the light incident opening.

By using the controller 7, the player can perform a game operation, which was not possible conventionally, of changing the position of the controller 7 itself or rotating the controller 7, in addition to the pressing of the operation buttons or switches. Hereinafter, the game operations using the controller 7 will be described.

FIG. 7 is a general view of a game operation using the controller 7. As shown in FIG. 7, when playing the game using the controller 7 with the game system 1, the player holds the controller 7 with one hand (for example, left hand). The markers 8a and 8b are located parallel to the transverse or width direction of the monitor 2. The player holds the controller 7 such that the front surface of the controller 7 (having the light incident opening 35a by which the imaging information calculation section 35 takes the image of each of the markers 8a and 8b) faces the markers 8a and 8b. In this state, the player rotates the controller 7 (as indicated by the arrows in FIG. 7) or changes the position indicated by the controller 7 on the display screen to perform the game operations.

FIG. 8 illustrates the viewing angles of the markers 8a and 8b and the controller 7. As shown in FIG. 8, the markers 8a and 8b each radiate infrared light at a viewing angle θ1. The imaging information calculation section 35 can receive light incident thereon at a viewing angle θ2. For example, the viewing angle θ1 of each of the markers 8a and 8b is 34° (half value angle), and the viewing angle θ2 of the imaging information calculation section 35 is 41°. The player holds the controller 7 at the position and the direction with which the imaging information calculation section 35 can receive the infrared light from both the markers 8a and 8b. Specifically, the player holds the controller 7 in a range in which both the markers 8a and 8b exist in the viewing angle θ2 of the imaging information calculation section 35 and the controller 7 exists in the viewing angle θ1 of the marker 8a and also in the viewing angle θ1 of the marker 8b. The player can perform a game operation by changing the position or the direction of the controller 7 in this range. When the position or the direction of the controller 7 is outside the above-described range, the game operation based on the position and the direction of the controller 7 cannot be performed. Hereinafter, the above-described range will be referred to as an “operable range”.

Where the controller 7 is held in the operable range, the imaging information calculation section 35 takes an image of each of the markers 8a and 8b. The taken image obtained by the imaging element 40 includes an image (target image) of each of the markers 8a and 8b which are imaging targets. FIG. 9 shows an example of the taken image including the target images. Using image data of the taken image including the target images, the image processing circuit 41 calculates a coordinate set representing the position of each of the markers 8a and 8b in the taken image.

The target images appear as high brightness areas in the image data of the taken image. Therefore, the image processing circuit 41 first detects the high brightness areas as candidates of the target images. Next, based on the size of each detected high brightness area, the image processing circuit 41 determines whether or not the high brightness area is a target image. The taken image may include images other than the target images (images 8a′ and 8b′ of the markers 8a and 8b) due to sunlight coming through a window or light of a fluorescent lamp. The determination is performed in order to distinguish the target images 8a′ and 8b′ from the other images so that the target images are accurately detected. Specifically, it is determined whether or not each detected high brightness area has a size within a predetermined size range. When the high brightness area has a size within the predetermined size range, the high brightness area is determined to be a target image; whereas when the high brightness area has a size outside the predetermined size range, the high brightness area is determined not to be a target image.

The image processing circuit 41 calculates the position of a high brightness area which is determined to be a target image as a result of the determination. Specifically, the image processing circuit 41 calculates the position of the center of gravity of the high brightness area. If the target images are accurately detected, two high brightness areas are determined to be target images by the determination. Therefore, two positions are calculated. The position in the taken image is represented with a coordinate system (x-y coordinate system) in which the upper left corner of the taken image is the origin, the downward direction from the origin is a positive y-axis direction, and the rightward direction from the origin is a positive x-axis direction. Accordingly, the image processing circuit 41 outputs data on the two coordinate set values indicating the two positions obtained by the calculation. The output data on the coordinate set values is transmitted to the game apparatus 3 as operation data by the microcomputer 42 as described above.

Using the data on the coordinate set values included in the received operation data, the game apparatus 3 can calculate the indicated position (the position indicated by the controller 7 on the display screen), the rotation angle (posture) of the controller 7 around the indicated direction, and the distance from the controller 7 to each of the markers 8a and 8b. FIG. 10 shows a change of the taken image when the position and/or the direction of the controller 7 is changed. FIG. 10 shows the correspondence between the state of the controller 7 and the taken image obtained when the controller 7 is in the respective state. In FIG. 10, a taken image I1 is obtained when the controller 7 is in state A. In the taken image I1, the target images 8a′ and 8b′ of the markers 8a and 8b are located in the vicinity of the center of the taken image I1. The target images 8a′ and 8b′ are located generally parallel to the X-axis direction.

State B shown in FIG. 10 is obtained by rotating the controller 7 at 90° clockwise around the indicated direction as an axis (around the Z axis) from state A. In this specification, the terms “clockwise” and “counterclockwise” used regarding the controller 7 refer to the rotation directions when seen from behind the controller 7 (from the negative end of the Z-axis direction in FIG. 3A; i.e., from the rear surface of the housing 31). In state B, a taken image I2 is obtained by the imaging information calculation section 35. In the taken image I2, the target images 8a′ and 8b′ have been moved on an arc at 90° counterclockwise from the taken image I1. Accordingly, the posture of the controller 7 with respect to the indicated direction as an axis of rotation is found by detecting the direction of the target images in the taken image (the direction of the image 8b′ with respect to the image 8a′, or the direction of the image 8a′ with respect to the image 8b′).

State C shown in FIG. 10 is obtained by translating the controller 7 rightward (in the positive x-axis direction) from state A. In state C, a taken image I3 is obtained by the imaging information calculation section 35. In the taken image I3, the target images 8a′ and 8b′ have been moved leftward (in the negative x-axis direction) from the taken image I1. In state C, the indicated direction of controller 7 is directed rightward with respect to state A. The indicated direction of controller 7 can be directed rightward by rotating the controller 7 around the Y axis as well as translating the controller 7 rightward. When the controller 7 is rotated around the Y axis, substantially the same taken image as the taken image 13 is obtained. Therefore, when the controller 7 is moved (rotated) so as to direct the indicated direction of the controller 7 rightward, substantially the same taken image as the taken image I3 is obtained. Namely, an image in which the target images 8a′ and 8b′ have been translated is obtained. Accordingly, the indicated direction of the controller 7 can be found by detecting the positions of the target images in the taken image (in the example described below, the position of an intermediate point between the images 8a′ and 8b′). In this specification, the term “intermediate” means substantially exactly halfway between the two points, coordinate sets or images.

State D shown in FIG. 10 is obtained by moving the controller 7 to be farther from the markers 8a and 8b (i.e., translating the controller 7 rearward) from state A. In state D, a taken image I4 is obtained by the imaging information calculation section 35. In the taken image I4, the distance between the target images 8a′ and 8b′ is shorter than in the taken image I1. Accordingly, the moving state of the controller 7 in the indicated direction (the distance from the controller 7 to the markers 8a and 8b) can be found by detecting the distance between the two target images in the taken image (the distance between the images 8a′ and 8b′; when the two target images are considered as one integral target image, the size of the target image).

Next, the game processing executed by the game apparatus 3 will be described in detail. First, main data used for the game processing will be described with reference to FIG. 11. FIG. 11 shows main data stored on the main memory 13 of the game apparatus 3. As shown in FIG. 11, the main memory 13 has stored thereon current operation data 50, previous operation data 53, operation state data 56, operation processing data 60, operation target data 64 and the like. In addition to the data shown in FIG. 11, the main memory 13 has stored thereon other data required for the game processing including data regarding a player character appearing in the game (image data, position data, etc. of the player character) and data regarding the game space (topography data, etc.).

The current operation data 50 is a latest data transmitted from the controller 7. The current operation data 50 includes current first coordinate set data 51 and current second coordinate set data 52. The current first coordinate set data 51 represents a position of the image of one of the two marks (the position in the taken image). The current second coordinate set data 52 represents a position of the image of the other mark (the position in the taken image). The position of the image of each mark is represented with the x-y coordinate system (see FIG. 9). Hereinafter, the coordinate set represented by the current first coordinate set data 51 will be referred to as a “first coordinate set”, and the coordinate set represented by the current second coordinate set data 52 will be referred to as a “second coordinate set”. The first coordinate set and the second coordinate set will be distinguished from each other in order to accurately calculate direction data 57 described later. Either one of the first coordinate set and the second coordinate set may represent the position of the image of the marker 8a or the marker 8b. However, in the case where the first coordinate set represents the image of one of the markers in one frame, the first coordinate set needs to represent the image of the same marker in the next frame (see steps S13 through S15 described later).

The current operation data 50 includes data obtained from the operation section 32 and the acceleration sensor 37 in addition to the coordinate set data obtained from the taken image (the current first coordinate set data 51 and the current second coordinate set data 52). The game apparatus 3 obtains the operation data from the controller 7 at a predetermined time interval (for example, at an interval of 1 frame). The latest data among the operation data obtained is stored on the main memory 13 as the current operation data 50.

The previous operation data 53 includes previous first coordinate set data 54 and previous second coordinate set data 55. In this specification, the term “previous” means “immediately previous”. The previous first coordinate set data 54 is the first coordinate set data obtained immediately before the current first coordinate set data 51. Namely, when new operation data is obtained, the data which has been the current first coordinate set data 51 so far is stored as the previous first coordinate set data 54. One of the two pieces of coordinate set data included in the new operation data is stored on the main memory 13 as the current first coordinate set data 51. Similarly, the previous second coordinate set data 55 is the second coordinate set data obtained immediately before the current second coordinate set data 52. Namely, when new operation data is obtained, the data which has been the current second coordinate set data 52 so far is stored as the previous second coordinate set data 55. The other of the two pieces of coordinate set data included in the new operation data (the data which is not the data representing the first coordinate set) is stored on the main memory 13 as the current second coordinate set data 52.

The operation state data 56 represents an operation state of the controller 7 which is determined based on the taken image. The operation state data 56 represents a position, direction, and size of the target image included in the taken image. The operation state data 56 includes direction data 57, intermediate point data 58, and length data 59. The direction data 57 represents a direction from the first coordinate set to the second coordinate set. In this embodiment, the direction data 57 is data on a two-dimensional vector having a position of the first coordinate set as a start point and a position of the second coordinate set as an end point. The direction data 57 represents a direction of the target image (markers 8a and 8b) in the taken image. The intermediate point data 58 represents a coordinate set of a point intermediate between the first coordinate set and the second coordinate set. In the case where the image of the marker 8a and the image of the marker 8b are considered as one integral target image, the intermediate point data 58 represents a position of the target image. The length data 59 represents a length of a distance between the first coordinate set and the second coordinate set. In the case where the image of the marker 8a and the image of the marker 8b are considered as one integral target image, the length data 59 represents a size of the target image.

The operation processing data 60 is calculated during the process of calculating the operation target data 64 from the operation state data 56. The operation processing data 60 includes direction processing data 61, indicated coordinate set data 62, and distance data 63. The direction processing data 61 is calculated during the process of calculating posture data 65 of the operation target data 64 from the direction data 57. The direction processing data 61 represents, for example, a difference between a predetermined reference direction and a direction represented by the direction data 57 (angular difference). The indicated coordinate set data 62 is calculated from the intermediate point data 58, and represents a coordinate set of an indicated position on the display screen of the monitor 2. The indicated coordinate set data 62 is calculated during the process of calculating position data 66 of the operation target data 64. The distance data 63 represents a distance from the imaging means of the controller 7 (the imaging information calculation section 35) to the imaging target (markers 8a and 8b). The distance data 63 is calculated during the process of calculating the position data 66 from the length data 59. The data 61 through 63 included in the operation processing data 60 are used for calculating the operation target data 64 when necessary, but not necessarily.

The operation target data 64 represents a position and a posture of the operation target on the display screen or in the game space. The “operation target” may be an object displayed on the display screen, an object appearing in the virtual game space, or in the case where a virtual three-dimensional game space is constructed, a virtual camera for displaying the three-dimensional game space on the display screen. The operation target data 64 includes the posture data 65 and the position data 66. The posture data 65 represents a posture of the operation target. The position data 66 represents a position of the operation target in the game space or on the display screen.

Next, the game processing executed by the game apparatus 3 will be described in detail with reference to FIG. 12 through FIG. 18B. FIG. 12 is a flowchart illustrating a flow of the game processing executed by the game apparatus 3. When the game apparatus 3 is turned on the CPU 10 of the game apparatus 3 executes a start program stored on the boot ROM (not shown) to initialize each unit, such as, for example, the main memory 13. The game program stored on the optical disc 4 is read into the main memory 13, and the CPU 10 starts the execution of the game program. The flowchart shown in FIG. 12 illustrates the game processing after the above-described processing is completed. In the following description given with reference to FIG. 12 through FIG. 18B, the game processing regarding the position and the direction of the controller 7 will be explained in detail, and other game processing not directly relevant will be omitted.

With reference to FIG. 12, in step S1, the operation data is obtained from the controller 7. The CPU 10 stores the operation data on the main memory 13 as the current operation data 50. The contents of the data which have been the current first coordinate set data 51 and the current second coordinate set data 52 so far are stored on the main memory 13 as the previous first coordinate set data 54 and the previous second coordinate set data 55. At this point, it has not been determined which of the two pieces of coordinate set data included in the operation data is to be the current first coordinate set data 51 and which is to be the current second coordinate set data 52. Accordingly, at this point, neither the current first coordinate set data 51 nor the current second coordinate set data 52 is stored on the main memory 13, and the two pieces of coordinate set data included in the operation data are stored on the main memory 13 separately from the current first coordinate set data 51 and the current second coordinate set data 52.

The operation data obtained in step S1 includes the coordinate set data representing positions of the markers 8a and 8b in the taken image, and also data representing whether or not each operation button of the controller 7 has been pressed and data representing accelerations of the controller 7 in the up-down direction and the left-right direction. In this embodiment, the controller 7 transmits the operation data at a time interval shorter than one frame, and the CPU 10 obtains the operation data frame by frame. Accordingly, the processing loop of steps S1 through S5 shown in FIG. 12 is repeated frame by frame.

Next in step S2, operation state calculation processing is executed. In the operation state calculation processing, an operation state of controller 7 (a value corresponding to a position, a posture or the like of the controller 7) is calculated based on the operation data from the controller 7. Hereinafter, with reference to FIG. 13 and FIG. 14, the operation state calculation processing will be described in detail.

FIG. 13 is a flowchart illustrating a detailed flow of the operation state calculation processing (S2) shown in FIG. 12. The operation state calculation processing is executed as follows. First in step S11, it is determined whether or not images of the two markers 8a and 8b have been taken by the imaging information calculation section 35 of the controller 7. The determination in step S11 is executed in order to determine whether or not the controller 7 is held in the operable range. When the images of the two markers 8a and 8b have not been taken by the imaging information calculation section 35 (when the controller 7 is not held in the operable range), the operation data from the controller 7 does not include two pieces of coordinate set data. Namely, when the image of only one marker is taken by the imaging information calculation section 35, the image processing circuit 41 outputs only one piece of coordinate set data, and the operation data from the controller 7 includes only one piece of coordinate set data. When the image of neither the marker 8a nor 8b is taken by the imaging information calculation section 35, the image processing circuit 41 does not output any coordinate set data, and the operation data from the controller 7 does not include any coordinate set data. Accordingly, the determination in step S11 can be performed based on whether or not the operation data obtained in step S1 includes two pieces of coordinate set data. When it is determined in step S11 that the images of the two markers 8a and 8b have been taken, processing in step S13 is executed. When it is determined in step S11 that the images of the two markers 8a and 8b have not been taken, processing in step S12 is executed.

In step S12, the contents of the operation state data 56 stored on the main memory 13 are cleared. In the case where the operation state data 56 is not stored on the main memory 13, the operation target is not moved in moving processing described later. Namely, in this embodiment, in the case where the images of two markers 8a and 8b are not taken, no game processing is executed. After step S12, the CPU 10 terminates the operation state calculation processing.

In steps S13 through S15, it is determined which of the two pieces of coordinate set data represents the first coordinate set and which of the data represents the second coordinate set. FIG. 14 illustrates the processing for determining which is the first coordinate set and which is the second coordinate set. In FIG. 14, point P1′ indicated with the dashed circle represents the position of the previous first coordinate set, and point P2′ indicated with the dashed circle represents the position of the previous second coordinate set. Points P1 and P2 indicated with the solid circles show the positions of the coordinate sets represented by the two pieces of coordinate set data obtained at the current time.

Referring to FIG. 14, it cannot be detected, only from the coordinate sets P1 and P2, which of the coordinate sets P1 and P2, represented by the two pieces of coordinate set data obtained at the current time, is the first coordinate set and which is the second coordinate set. Namely, the game apparatus 3 cannot determine, only from the coordinate sets P1 and P2, whether the coordinate set P1′ has moved to become the coordinate set P1 or the coordinate set P2′ has moved to become the coordinate set P1. Either one of the first coordinate set and the second coordinate set may represent the position of the image of the marker 8a or the marker 8b. It is important, however, to correctly detect which data represents the first coordinate set and which data represents the second coordinate set, in order to calculate the direction from the first coordinate set to the second coordinate set (step S16 described later). If the first coordinate set and the second coordinate set are detected oppositely, the direction obtained by the calculation is the opposite. If the previous first coordinate set is detected as the current second coordinate set, for example, the direction calculated in step S16 is incorrect.

In order to avoid this, in this embodiment, the first coordinate set and the second coordinate set are detected based on the distances between the previous coordinate sets and the current coordinate sets. Specifically, among the two coordinate sets newly obtained currently, the coordinate set which is closer to the previous first coordinate set is determined as the first coordinate set, and the other coordinate set is determined as the second coordinate set. In the example shown in FIG. 14, the coordinate set P1 which is closer to the previous coordinate set P1′ is set as the first coordinate set, and the other coordinate set P2 is set as the second coordinate set. In the case where the controller 7 is rotated at an angle greater than 90° from one frame to the next frame, the coordinate set which is farther from the previous first coordinate set is the correct first coordinate set. However, it is considered that the controller 7 cannot actually rotate at an angle greater than 90° during one frame because the time interval at which the coordinate sets are obtained (the time interval at which the imaging information calculation section 35 takes the images) is usually very short (for example, an interval of 1/200 seconds). Therefore, the first coordinate set can be correctly detected by setting the coordinate set closer to the previous first coordinate set as the first coordinate set.

Specifically, in step S13, it is determined whether or not the first coordinate set and the second coordinate set were detected in the previous frame. The determination in step S13 is performed based on whether or not the operation state data 56 is stored on the main memory 13. When the processing in step S12 was executed in the previous frame, the operation state data 56 is not stored on the main memory 13. In this case, at least one of the first coordinate set and the second coordinate set was not obtained in the previous frame, and so the first coordinate set and the second coordinate set cannot be determined using the previous coordinate sets. When the processing in step S13 was executed in the previous frame, the operation state data 56 is stored on the main memory 13. In this case, the first coordinate set and the second coordinate set can be determined using the previous coordinate sets. The processing in step S13 is executed in order to determine whether or not the first coordinate set and the second coordinate set are to be determined using the previous coordinate sets. When it is determined in step S13 that the first coordinate set and the second coordinate set were detected in the previous frame, processing in step S14 is executed. When it is determined in step S13 that at least one of the first coordinate set and the second coordinate set was not detected in the previous frame, processing in step S15 is executed.

In step S14, among the two coordinate sets represented by the two pieces of coordinate set data included in the operation data obtained in step S1, the coordinate set closer to the previous first coordinate set is determined as the first coordinate set. Specifically, the CPU 10 refers to the previous first coordinate set data 54 stored on the main memory 13 and the two pieces of coordinate set data included in the operation data, and specifies, among the coordinate sets represented by the two pieces of coordinate set data, the coordinate set which is closer to the coordinate set represented by the previous first coordinate set data 54. The CPU 10 stores coordinate set data on the specified coordinate set on the main memory 13 as the current first coordinate set data 51, and also stores coordinate set data on the other coordinate set on the main memory 13 as the current second coordinate set data 52.

In step S15, one of the coordinate sets represented by the two pieces of coordinate set data is determined as the first coordinate set in accordance with a predetermined condition. The predetermined condition may be any condition. For example, the CPU 10 determines the coordinate set having a smaller y-axis coordinate value as the first coordinate set, and determines the other coordinate set as the second coordinate set. Data on the coordinate set determined as the first coordinate set is stored on the main memory 13 as the current first coordinate set data 51, and data on the coordinate set determined as the second coordinate set is stored on the main memory 13 as the current second coordinate set data 52.

In step S16, a direction from the first coordinate set to the second coordinate set is determined. Specifically, the CPU 10 refers to the current first coordinate set data 51 and the current second coordinate set data 52 to calculate a vector having a position of the first coordinate set as a start point and a position of the second coordinate set as an end point. Data on the calculated vector is stored on the main memory 13 as the direction data 57. The direction of the calculated vector corresponds to a rotation angle of the controller 7 around the indicated direction. Namely, as a result of the processing in step S16, the rotation of the controller 7 around the indicated direction is calculated.

Next in step S17, a coordinate set of an intermediate point between the first coordinate set and the second coordinate set is calculated. Specifically, the CPU 10 refers to the current first coordinate set data 51 and the current second coordinate set data 52 to calculate the coordinate set of the intermediate point. Data on the calculated coordinate set of the intermediate point is stored on the main memory 13 as the intermediate point data 58. The intermediate point data 58 represents a position of the target image (the markers 8a and 8b) in the taken image.

Next in step S18, the coordinate set of the intermediate point calculated in step S17 is corrected. FIG. 15 illustrates the processing in step S18 shown in FIG. 13. FIG. 15 shows the correspondence between the state of the controller 7 and the taken image obtained when the controller 7 is in the respective state. In FIG. 15, in both state A and state B, the center of the display screen of the monitor 2 is the indicated position by the controller 7. In state A, the top surface of the controller 7 is directed upward (in this embodiment, this state will be referred to as the “reference state”); whereas in state B, the top surface of the controller 7 is directed rightward. Since the markers 8a and 8b are located above the display screen, the positions of the markers 8a and 8b do not match the indicated position (at the center of the display screen). Therefore, in state A, the target images are above the center of the taken image. In state B, the target images are left to the center of the taken image. As seen from FIG. 15, since the positions of the markers 8a and 8b do not match the indicated position (at the center of the display screen), the positions of the target images in the taken image change by the rotation angle of the controller 7 around the indicated direction as an axis. For this reason, the indicated position cannot be accurately calculated based on the positions of the target images.

In order to avoid this, in step S18, the coordinate set of the intermediate point calculated in step S17 is corrected based on the direction calculated in step S16. Specifically, the coordinate set of the intermediate point calculated in step S17 is corrected into the coordinate set obtained in the case where the controller 7 is in the reference state. More specifically, the coordinate set of the intermediate point calculated in step S17 is moved on an arc around the center of the taken image as an axis by an amount corresponding to an angular difference between the direction calculated in step S16 and the direction in the reference state. In state B in FIG. 15, a vector inclined at 90° counterclockwise from the reference state is calculated in step S16. Therefore, it is understood that the controller 7 is inclined at 90° clockwise from the reference state. Accordingly, in step S18, the coordinate set of the intermediate point calculated in step S17 is moved on an arc at 90° clockwise around the center of the taken image as an axis. Data representing the coordinate set of the intermediate point corrected as above in step S18 is stored on the main memory 13 as the updated intermediate point data 58. Based on the corrected coordinate set of the intermediate point obtained in step S18, the game apparatus 3 can calculate the indicated position by the controller 7.

Next in step S19, a length of a distance between the first coordinate set and the second coordinate set is calculated. Specifically, the CPU 10 refers to the current first coordinate set data 51 and the current second coordinate set data 52 to calculate the length. Data on the calculated length is stored on the main memory 13 as the length data 59. The length corresponds to the distance from the controller 7 to the markers 8a and 9b. After step S19, the CPU 10 terminates the operation state calculation processing.

Returning to FIG. 12, in step S3, the moving processing of moving the operation target is executed. With reference to FIG. 16, the moving processing will be described in detail.

FIG. 16 is a flowchart illustrating a detailed flow of the moving processing (S3) shown in FIG. 12. The moving processing is executed as follows. First in step S21, the operation target is rotated in the direction calculated in step S16. As described above, the operation target may be an object appearing in the virtual game space or a virtual camera. Hereinafter, variations of the rotation motion of the operation target will be described with reference to FIG. 17A through FIG. 18B.

FIG. 17A through FIG. 18B show examples of the rotation motion of the operation target. In FIG. 17A through FIG. 18B, the operation target is an object appearing in the game space. A pre-rotation object Ob′ is indicated with the dashed line, and a post-rotation object Ob is indicated with the solid line. The rotation motion of the operation target performed in step S21 may be for changing the position of the operation target on the display screen or in the game space, for changing the posture of the operation target, or for changing both the position and the posture of the operation target. Namely, the rotation motion of the operation target may be, as shown in FIG. 17A, a motion of changing the posture of the operation target without changing the position thereof; as shown in FIG. 17B, a motion of moving the operation target on an arc around a predetermined position P as an axis without changing the posture (direction) thereof; or as shown in FIG. 17C, a motion of moving the operation target on an arc around the predetermined point P as an axis while rotating the operation target, so that both the position and the posture of the operation target are changed. In the case where the operation target is located in the virtual three-dimensional game space, the operation target does not especially need to rotate around an axis vertical to the display screen. FIG. 18A shows the object rotating in the virtual three-dimensional game space. In the case where the object rotates as shown in FIG. 18A, the game space seen from a direction parallel to the Y axis is displayed on the display screen as shown in FIG. 18B. The processing of rotating the object may be processing of rotating the object on the display screen or processing of rotating the object in the game space. In the latter case, the rotation of the object is not easily seen on the display screen.

In step S21, specifically, a posture and/or a position of the operation target is determined in accordance with the direction of the vector represented by the direction data 57 stored on the memory 13 in step S16. More specifically, the determination may be performed as follows. The correspondence between the direction of the vector and the posture and/or the position of the operation target when the vector is inclined in the respective direction is determined in advance. Based on the correspondence, the posture and/or the position of the operation target is determined from the vector. Alternatively, the determination may be performed as follows. An angle made by the vector in the previous frame and the vector in the current frame (angular difference) is calculated, and a changing amount of the posture and/or the position of the operation target is determined based on the angular difference. For example, the operation target may be rotated at an angle in proportion to the angular difference. Data representing the angular difference corresponds to the direction processing data 61 in the operation processing data 60. When the posture and/or the position of the operation target is determined as described above, the contents of the operation target data 64 stored on the main memory 13 are updated. Namely, data representing the posture determined in step S21 is stored on the main memory 13 as the posture data 65, and data representing the position determined in step S21 is stored on the main memory 13 as the position data 66.

In other embodiments, the processing may be set such that in the case where the direction of the vector calculated in step S16 is in a predetermined range of angles, the operation target is not rotated. The reason is that when the operation target rotates as a result of responding too sensitively to the rotation angle of the input device, the operability of the input device may be deteriorated. Depending on the type of, or the situation in, the game, the player may wish to maintain the operation target at a certain posture. If, in such a case, the operation target rotates as a result of responding too sensitively to the rotation angle of the input device, it becomes difficult to maintain the operation target at a certain posture and the operability of the input device is deteriorated. For example, the processing may be set such that the operation target is not rotated when the angle made by the vector and the y axis is in the range of −5° to 5°. Owing to such an arrangement, even if the controller 7 is rotated slightly, the same operation result as in the case where the controller 7 is maintained in a horizontal state is obtained. As described above, the operability of the input device can be improved by setting a so-called “play” margin in the rotation operation performed on the controller 7.

Next in step S22, the operation target is translated. Namely, a post-movement position of the operation target is calculated. When the post-movement position of the operation target is determined in step S21 (when the rotation motion shown in FIG. 17B or FIG. 17C is performed), the processing in step S22 is omitted. When only the posture of the operation target is determined in step S21 (when the rotation motion shown in FIG. 17A is performed), the processing in step S22 is executed.

The translation motion of the operation target in step S22 is determined based on at least one of (i) the coordinate set of the intermediate point corrected in step S18 and (ii) the length of the distance calculated in step S19. Typically, the position of the operation target on the display screen (the position in the game space corresponding to the position on the display screen) is determined based on the coordinate set of the intermediate point corrected in step S18. Based on the distance calculated in step S19, the position of the operation target in the depth direction with respect to the display screen is determined.

The position of the operation target on the display screen is calculated using a function of converting the coordinate set of the intermediate point into a coordinate set on the display screen of the monitor 2. This function converts the coordinate set values of the intermediate point which are calculated from one taken image into coordinate set values corresponding to the position on the display screen which is indicated by the controller 7 (indicated position) when such a taken image is taken. This is performed, for example, as follows. Predetermined reference values are set for both the coordinate set of the intermediate point and the indicated position on the display screen. A displacement of the coordinate set of the intermediate point from the reference values is calculated, and the positive/negative sign of the displacement is inverted and scaled at a predetermined ratio. Thus, a displacement of the current indicated position from the references value is calculated, and the current indicated position is set. This function allows the indicated position on the display screen to be calculated from the coordinate set of the intermediate point. Data representing the indicated position corresponds to the indicated coordinate set data 62 in the operation processing data 60. For converting the coordinate set of the intermediate point into a coordinate set representing the position in the game space (a corresponding coordinate set described later), the indicated position is further converted into the position in the game space corresponding to the indicated position. The “position in the game space corresponding to the indicated position” is a position, in the game space, which is displayed at the indicated position on the display screen.

The position of the operation target in the depth direction with respect to the display screen is calculated based on the length calculated in step S19. Specifically, this is performed as follows. The correspondence between a predetermined position in the depth direction and a predetermined length is determined in advance. The position of the operation target is determined based on the difference between the length calculated in step S19 and the predetermined length. More specifically, the position of the operation target is determined such that as the difference obtained by subtracting the predetermined length from the length calculated in step S19 becomes greater, the position of the operation target becomes deeper. In the case where the difference is a negative value, the position of the operation target is determined to be forward with respect to the predetermined position (i.e., closer to the player in front of the display screen of the monitor 2). Alternatively, the position of the operation target in the depth direction may be determined based on the distance from the imaging means of the controller 7 (the imaging information calculation section 35) to the imaging target (the markers 8a and 8b) (see FIG. 26 described later). Such a distance is calculated from the length calculated in step S19. Data representing such a distance corresponds to the distance data 63 in the operation processing data 60.

Data representing the position of the operation target determined in step S22 is stored on the main memory 13 as the position data 66. In step S22, it is not absolutely necessary to determine both the position of the operation target on the display screen and the position thereof in the depth direction. Either one of the positions may be determined. For example, in the case where the game space is a two-dimensional plane, only the position of the operation target on the display screen may be determined based on the coordinate set of the intermediate point corrected in step S18. After steps S21 and S22, the CPU 10 terminates the moving processing.

In the moving processing in this embodiment, the movement of the operation target is controlled such that the operation target is always moved to the indicated position by the controller 7. In other embodiments, the movement processing may be executed such that when a predetermined operation is performed on the controller 7 (for example, the B button 32d is pressed), the operation target is moved to the indicated position by the controller 7 obtained at the time when such an operation is performed. In the case where the operation of moving the operation target and the operation of rotating the operation target are simultaneously performed as in the above embodiment, the player may inadvertently rotate the controller 7 when intending to move the operation target, resulting in rotating the operation target despite his/her intention. In the case where the processing is set such that the operation target is moved only when necessary, the difficulty in simultaneously moving and rotating the operation target can be alleviated.

Returning to FIG. 12, in step S4 after step S3, display processing is executed. The CPU 10 refers to the operation target data 64 stored on the main memory 13 to generate an image in accordance with the posture and the position determined in step S3. The CPU 10 displays the generated image on the display screen of the monitor 2. For example, when the operation target is an object displayed on the display screen, the object is displayed on the display screen with the posture and the position determined in step S3. For example, in the case where the operation target is a virtual camera, an image of the game space seen from the virtual camera which is set at the posture and the position determined in step S3 is generated, and the generated image is displayed. By repeating the processing in step S4 frame by frame, a moving picture representing the operation target rotating can be displayed.

In step S5, it is determined whether or not to terminate the game. The game is terminated when, for example, a condition for terminating the game (for example, the parameter indicating the stamina of the player character becomes zero) or the player has performed an operation of terminating the game. When it is determined not to terminate the game, the processing is returned to step S1 and the processing in steps S1 through S5 is repeated until it is determined to terminate the game in step S5. When it is determined to terminate the game, the CPU 10 terminates the game processing shown in FIG. 12.

As described above, in this embodiment, a value corresponding to the posture of the controller 7 rotating around the front-rear direction thereof as an axis can be calculated from the taken image obtained by the imaging information calculation section 35. Based on the posture of the controller 7, the operation target can be rotated. With such settings, the player can perform a game operation of rotating the controller 7 around the front-rear direction thereof as an axis so as to cause the operation target to rotate in accordance with the rotation of the controller 7. Thus, the game apparatus 3 can allow the player to perform a new game operation of rotating the operation target by rotating the controller 7.

According to a non-limiting illustrative embodiment, the game apparatus 3 only needs to calculate the rotation of the controller 7 around the indicated direction as an axis. The game apparatus 3 does not need to detect the indicated position by the controller 7 or the distance between the controller 7 and the imaging target (the markers 8a and 8b). In order to calculate the rotation of the controller 7 around the indicated direction thereof as an axis, the game apparatus 3 only needs to obtain information representing at least the direction from the taken image. Therefore, the image processing circuit 41 may output the information representing the direction, instead of the coordinate sets of the above-mentioned two positions. Specifically, the image processing circuit 41 may calculate data on a vector representing the direction from the position of the image of the marker 8a to the position of the image of the marker 8b (or the direction from the position of the image of the marker 8b to the position of the image of the marker 8a). Since this vector only needs to represent the direction, the image processing circuit 41 may output data on a unit vector having a constant size.

The indicated position by the controller 7 can be detected by calculating the position of the target image in the taken image. The position of the target image in the taken image can be represented with the position of the coordinate set of the intermediate point as described in the above embodiment. Accordingly, in the case where the indicated position by the controller 7 is used for moving the operation target, the image processing circuit 41 may output data representing the position of the target image in the taken image and data representing the direction. Namely, the image processing circuit 41 may output the data on the vector representing the direction from the position of the image of the marker 8a to the position of the image of the marker 8b (or the direction from the position of the image of the marker 8b to the position of the image of the marker 8a) and the data on the coordinate set of the intermediate point, instead of outputting the data on the coordinate sets representing the positions of the images of the markers 8a and 8b. In this case, the game apparatus 3 does not need to calculate the direction data or the intermediate point data (steps S16 and S17), the processing load of the game apparatus 3 can be alleviated. The correction processing in step S18 may be executed by the image processing circuit 41 or by the game apparatus 3.

In other embodiments, the imaging information calculation section 35 may not include the image processing circuit 41. In this case, the imaging information calculation section 35 outputs the image data of the taken image to the microcomputer 42, and the image data is transmitted from the controller 7 to the game apparatus 3 as the operation data. The game apparatus 3 calculates the direction data based on the image data. As described above, the processing of calculating the direction data from the image data of the taken image may be partially executed by a dedicated circuit (the image processing circuit 41), entirely executed by such a circuit, or executed by the game apparatus 3 as software processing.

In the above embodiment, the two markers 8a and 8b for outputting infrared light are the imaging targets. Some other elements may be the imaging targets. An imaging target may be anything, an image of which can provide a direction. For example, three or more markers may be imaging targets, or one marker which can provide a direction can be an imaging target. In the case where one marker is the imaging target, it is desirable that the marker has a linear shape or the like, from which two predetermined points can be specified. The reason is that it is necessary to detect the direction from the marker. A circular shape is not desirable. The game apparatus 3 calculates the direction connecting the predetermined two points of the taken image to calculate the direction of the target image in the taken image. The marker is preferably a light emitting element. The marker may output light of other wavelengths than infrared, or may output white light. The frame of the display screen of the monitor 2 may be an imaging target, or the imaging target may be displayed on the display screen. With such a structure, it is not necessary to prepare a marker in addition to the monitor, and the structure of the system can be simplified.

In other embodiments, when the controller 7 goes outside the operable range, the operation target may be rotated using an output from the acceleration sensor. Specifically, when the determination result in step S11 in FIG. 13 is negative, processing of detecting the rotation state of the controller 7 around the indicated direction as an axis is executed using the output from the acceleration sensor, instead of the processing in step S12. The CPU 10 refers to data representing the acceleration included in the operation data to determine the rotation state of the controller 7. In the moving processing in step S3, the operation target is rotated in accordance with the rotation state determined by the above processing. Thus, even if the controller 7 goes outside the operable range, the operation of rotating the operation target can be continued. In the case where the controller 7 goes outside the operable range and then returns to the operable range with the posture thereof being changed, the posture of the controller 7 can be determined using an output from the acceleration. Therefore, the two markers can be each identified.

Now, it is assumed in the above embodiment that the controller 7 goes outside the operable range and then returns to the operable range. In this case, it cannot be determined only from the taken image whether the controller 7 has returned to the operable range in an upward state (with the top surface thereof directed upward) or in a downward state (with the top surface thereof directed downward). The position of the target image in the taken image is different between the upward state and the downward state. Therefore, when the controller 7 returns to the operable range in the downward state, the indicated position cannot be accurately calculated based on the position of the target image.

Therefore, in other embodiments, when the controller 7 returns to the operable range after once going outside, the game apparatus 3 may determine whether or not the controller 7 is in the downward state using the detection result of the acceleration sensor 37. When it is determined that the controller 7 is in the downward state, the correction processing in step S18 is executed. This will be described with reference to the flowchart in FIG. 13. When the determination result of step S13 is negative, the CPU 10 refers to the acceleration data included in the operation data to determine whether or not the controller 7 is in the downward state. When it is determined that the controller 7 is in the downward state, the correction processing in step S18 is executed as follows. The coordinate set of the intermediate point is moved on an arc at an angle obtained by adding 180° to the angular difference between the direction calculated in step S16 and the direction in the reference state. This processing is continued until the next time that the controller 7 goes outside the operable range. Thus, even if the controller 7 returns to the operable range in the downward state, the indicated position by the controller 7 can be accurately calculated.

The above embodiment is applicable to various games in which the controller 7 is rotated around the indicated direction as an axis. Hereinafter, specific games will be described as non-limiting examples.

FIRST EXAMPLE

FIG. 19 shows an example of a game screen in a first example. In FIG. 19, plate-like objects 71 and 72, a ball object 73, and wall objects 74 are located in a two-dimensional game space. The plate-like object 71 is an operation target to be operated by the player. The plate-like object 71 is movable up, down, left and right on the display screen and rotatable by an operation of the player (as indicated by the arrows in FIG. 19). The plate-like object 72 is controlled by the CPU 10 of the game apparatus 3 and is movable up, down, left and right on the display screen and rotatable like the plate-like object 71. Hereinafter, the plate-like object 71 to be operated by the player will be referred to as a “player object”, and the plate-like object 72 to be controlled by the CPU 10 will be referred to as an “opponent object”. The ball object 73 (hereinafter, also referred to simply as a “ball”) moves in the two-dimensional game space on the display screen. The ball 73 linearly moves in the game space, and when colliding against the player object 71, the opponent object 72, or the wall objects 74, bounces back and continues moving. The wall objects 74 are fixedly located at the top and bottom of the display screen. In the game in the first example, two parties fight against each other. More specifically, the game in the first example is a hockey game in which the player operates the player object 71 to hit the ball object 73 with the opponent object 72. The player operates the player object 71 so as to put the ball 73 into the opponent's side (the left end of the display screen, i.e., left to the opponent object 72) while hitting the ball 73 back such that the ball 73 does not reach the player's side (the right end of the display screen, i.e., right to the player object 71).

Next, a method for operating the player object 71 using the controller 7 will be described. The player object 71 moves (translates) up, down, left or right in accordance with a position designation operation performed by the controller 7. The “position designation operation” is an operation of designating the indicated position. Specifically, the position or the direction of the controller 7 is changed to change the position at which the phantom straight line extending from the front end of the controller 7 in the longitudinal direction crosses the display screen of the monitor 2. The indicated position is determined in accordance with the position or the direction of the controller 7. The player can move the indicated position up, down, left or right on the display screen to move the player object 71 up, down, left or right on the display screen. Specifically, the game apparatus 3 controls the operation of the player object 71 to be located at the indicated position.

The player object 71 rotates around the center thereof in accordance with a rotation operation. The “rotation operation” is an operation of rotating the controller 7 around the indicated direction as an axis. FIG. 20 illustrates the operation of the player object 71 provided by the rotation operation. FIG. 20 shows the correspondence among the state of the controller 7, the taken image obtained when the controller 7 is in the respective state, and the posture of the player object 71 when the controller 7 is in the respective state.

In FIG. 20, in state A, the top surface of the controller 7 is directed upward. In a taken image I7 obtained in state A, the images 8a′ and 8b′ of the markers 8a and 8b are located horizontally (the y-axis coordinates of the images 8a′ and 8b′ are the same). In this state, the player object 71 assumes a posture horizontal to the up-down direction of the display screen of the monitor 2. In this example, this state will be referred to as a “reference state”.

In state B shown in FIG. 20, the controller 7 has been rotated counterclockwise at a predetermined angle from state A. In a taken image I8 obtained in state B, the images 8a′ and 8b′ have been moved on an arc clockwise at the predetermined angle from state A. In this state, the player object 71 assumes a posture rotated counterclockwise at the predetermined angle from the up-down direction of the display screen. In state C shown in FIG. 20, the controller 7 has been rotated clockwise at a predetermined angle from state A. In a taken image I9 obtained in state C, the images 8a′ and 8b′ have been moved on an arc counterclockwise at the predetermined angle from state A. In this state, the player object 71 assumes a posture rotated clockwise at the predetermined angle from the up-down direction of the display screen. As described above, the game apparatus 3 controls the posture of the player object 71 to rotate in the direction in which the controller 7 has been rotated from the reference state at the same angle as the controller 7.

Next, the game processing in the first example will be described in detail. In the first example, the length data 59 in the operation state data 56 shown in FIG. 11 is not necessary for the game processing and is not stored on the main memory 13. The direction processing data 61 and the distance data 63 in the operation processing data 60 are not necessary for the game processing and are not stored on the main memory 13. The operation target data 64 shown in FIG. 11 represents a position and a posture of the player object 71 as the operation target. Namely, the posture data 65 represents a posture of the player object 71. Specifically, the posture data 65 represents a vector indicating a direction of the player object 71 on the display screen. This vector is represented with a coordinate system in which the upper left corner of the display screen is the origin, the downward direction from the origin is a positive y′-axis direction, and the rightward direction from the origin is a positive x′-axis direction (x′-y′ coordinate system). In this example, the vector represented by the posture data 65 indicates a direction vertical to the longitudinal direction of the rectangular player object 71. The position data 66 represents a position of the player object 71 on the display screen. A coordinate set representing the position is in the x′-y′ coordinate system. In this example, the position represented by the position data 66 indicates the central position of the player object 71.

In the first example, the data shown in FIG. 11, image data of the objects, data representing the position and the posture of the opponent object 72, data representing the position and the velocity of the ball object 73, and the like are stored on the main memory 13.

FIG. 21 is a flowchart illustrating a flow of the game processing in the first example. In FIG. 21, substantially the same steps as those in the game processing shown in FIG. 12 bear the same step numbers, and detailed descriptions thereof will be omitted. The processing until the processing in the flowchart shown in FIG. 21 is started is the same as in the case of FIG. 12.

The game processing in the first example is executed as follows. First in step S31, initial processing is executed. In the initial processing, the CPU 10 displays a game screen in an initial state. Namely, the objects 71 through 74 are located at respective initial positions. Regarding the operation target data 64, data representing the posture of the player object 71 in the initial state is stored on the main memory 13 as the posture data 65, and data representing the position of the player object 71 in the initial state is stored on the main memory 13 as the position data 66. The CPU 10 displays the game screen in which the objects 71 through 74 are at the respective initial positions on the monitor 2.

After step S31, the processing in step S1 described above is executed, and then the operation state calculation processing in step S2 is executed. In the operation state calculation processing in the first example, the processing of calculating the distance between the two coordinate sets (step S19) is not executed. The operation state calculation processing in the first example is substantially the same as described above with reference to FIG. 12 except for step S19.

Next, the moving processing in step S3 is executed. As described above with reference to FIG. 16, the processing of rotating the player object 71 (step S21) in the direction calculated in accordance with the rotation of the controller 7 around the indicated direction as an axis (around the direction calculated in step S16) is executed. Hereinafter, specific processing in step S21 in the first example will be described.

In step S21 in the first example, the vector representing the direction in the taken image (the direction in the x-y coordinate system) is converted into a vector representing a direction on the display screen (the direction in the x′-y′ coordinate system). The “vector representing the direction in the taken image” is the vector calculated in step S16, which is represented by the direction data 57. By converting the vector into the vector representing the direction on the display screen, the vector representing the direction of the player object 71 is calculated. Specifically, in step S21, the direction of the vector represented by the direction data 57 stored on the main memory 13 is inverted regarding the x axis or the y axis. For example, where the vector represented by the direction data 57 is (vx1, vy1), the CPU 10 calculates the post-conversion vector as (−vx1, vy1). Thus, the vector is inverted regarding the y axis, i.e., converted into the vector representing the direction in the x′-y′ coordinate system. The CPU 10 stores data representing the post-conversion vector on the main memory 13 as the updated posture data 65.

Next in step S22, the player object 71 is translated. Namely, a position of the player object 71 on the display screen is calculated. Specifically, like in step S22 described above, the CPU 10 calculates a coordinate set (indicated coordinate set), in the x′-y′ coordinate system, which represents the indicated position on the display screen, from the coordinate set representing the intermediate point in the x-y coordinate system. Data on the calculated indicated coordinate set is stored on the main memory 13 as the indicated coordinate set data 62. Next, the calculated indicated coordinate set is determined as the position of the player object 71. In the case where the indicated coordinate set is outside a movable range of the player object 71 which is set in the game (for example, the area interposed between the wall objects 74) or the like, the position of the player object 71 is set at a different position from the indicated coordinate set, for example, in the vicinity of the boundary of the movable range. Data on the coordinate set representing the position of the player object 71 on the display screen (indicated coordinate set) determined in step S22 is stored on the memory 13 as the position data 66.

In step S32 after step S3, the opponent object 72 is moved. The CPU 10 determines a position and a posture of the opponent 72 after the movement in accordance with a predetermined algorithm. Next in step S33, the ball 73 is moved. The CPU 10 determines whether or not the ball 73 has collided against the player object 71, the opponent object 72, or the wall objects 74. When the ball 73 has collided, a post-collision moving direction of the ball 73 is determined based on the pre-collision moving direction of the ball 73 and an angle made by the ball and the object against which the ball 73 has collided. When ball 73 did not collide, the moving direction of the ball 73 is maintained. The velocity of the ball 73 may be changed in accordance with the moving velocity or the rotating velocity of the player object 71 or the opponent object 72 at the time of collision.

Next in step S4, the display processing is executed. Namely, the objects 71 through 73 are displayed on the display screen with the new positions and postures determined in steps S3 through S5. By repeating the processing in step S4 frame by frame, the objects 71 through 73 are displayed as moving. Next in step S5, it is determined whether or not to terminate the game. In the first example, this determination is performed based on, for example, the number of times that the ball 73 reached the player's side or the opponent's side.

In the first example described above, the game can be enjoyed with a new game operation of moving or rotating the controller 7 to move or rotate the player object 71.

In the first example, one controller 7 is used to operate only the player object 71. Two controllers 7 may be used to operate two objects (for example, the objects 71 and 72 shown in FIG. 19). In this case, two players can operate the respective objects, each using one controller 7, and fight against each other.

SECOND EXAMPLE

FIG. 22 shows an example of a game screen in a second example. FIG. 22 shows a game scene in which the player can open the door using a key. In FIG. 22, a three-dimensional game space is constructed, and a door 75 and a key 76 are located in the game space. The door 75 has a key hole 75a. The player can move or rotate the key 76 by operating the controller 7. The player performs an operation of moving the key 76 in the three-dimensional game space using the controller 7 and thus puts the key 76 into the key hole 75a. When the player performs an operation of rotating the key 76 in the state where the key 76 is in the key hole 75a, the door 75 opens. An X-Y-Z coordinate system shown in FIG. 22 is a coordinate system in a three-dimensional game space and is not actually displayed on the display screen. In this example, the Z-axis direction is the depth direction, which is parallel to the viewing direction of a virtual camera.

In the second example, the key 76 moves in the three-dimensional game space in accordance with the position designation operation and a depth designation operation performed by the controller 7. The “depth designation operation” is an operation of changing the distance from the controller 7 to the markers 8a and 8b. Specifically, the game apparatus 3 moves the key 76 in the X-axis direction and the Y-axis direction in the game space in accordance with the position designation operation. In accordance with the depth designation operation, the game apparatus 3 moves the key 76 in the Z-axis direction in the game space. The key 76 is rotated by the rotation operation performed by the player. In this example, the key 76 rotates around the Z axis in the game space.

Next, the game processing in the second example will be described in detail. In the second example, the operation target data 64 stored on the main memory 13 (see FIG. 11) represents a position and a posture of the key 76 as the operation target. Namely, the posture data 65 represents a posture of the key 76 rotating around the Z axis. The position data 66 represents a position of the key 76 in the three-dimensional game space. The position data 66 represents a three-dimensional coordinate set. In the second example, the direction processing data 61 is not necessary for the game processing and is not stored on the main memory 13.

FIG. 23 is a flowchart illustrating a flow of the game processing in the second example. In FIG. 23, substantially the same steps as those in the game processing shown in FIG. 12 bear the same step numbers, and detailed descriptions thereof will be omitted. The processing until the processing in the flowchart shown in FIG. 23 is started is the same as in the case of FIG. 12.

The game processing in the second example is executed as follows. First in step S41, initial processing is executed. In the initial processing, the CPU 10 displays a game screen in an initial state. Namely, a three-dimensional game space is constructed, and the door 75, the key 76 and the like are located. The key 76 is located at a predetermined initial position. Regarding the operation target data 64, data representing the posture of the key 76 in the initial state is stored on the main memory 13 as the posture data 65, and data representing the position of the key 76 in the initial state is stored on the main memory 13 as the position data 66. The CPU 10 generates an image of the game space, in which the door 75 and the key 76 are located, seen from a predetermined virtual camera and displays the game screen on the monitor 2.

After step S41, the processing in step S1 described above is executed, and then the operation state calculation processing in step S2 is executed. The processing in steps S1 and S2 is the substantially same as described above with reference to FIG. 12.

Next, the moving processing in step S3 is executed. FIG. 24 is a flowchart illustrating a detailed flow of the moving processing in the second example. The moving processing in the second example is executed as follows. First in step S45, the posture of the key 76 is calculated in accordance with the direction calculated in step S16. Step S45 corresponds to step S21 described above. In the second example, the key 76 is rotated around the depth direction of the display screen as an axis (around the Z axis) in accordance with the rotation of the controller 7 around the indicated direction. Namely, the rotation direction of the controller 7 is the same as the rotation direction of the key 76 displayed on the display screen. Accordingly, the player can feel as if he/she was actually rotating the key 76 by performing an operation of rotating the controller 7.

Next in steps S46 through S49, the position of the key 76 in the game space is determined. The processing in steps S46 through S49 correspond to step S22 described above.

In step S46, the indicated position on the display screen is calculated based on the intermediate point data 58 in the operation state data 56 stored on the main memory 13. A method for calculating the indicated coordinate set is the same as described above regarding step S22. The above-described function is used. Data on the coordinate set representing the indicated position (indicated coordinate set) is stored on the main memory 13 as the indicated coordinate set data 62.

In step S47, the positions of the key 76 in the X-axis direction and the Y-axis direction in the game space are determined based on the indicated position calculated in step S46. More specifically, the position of the key 76 in the X-axis direction in the game space is determined based on the x component of the coordinate set represented by the indicated coordinate set data 62, and the position of the key 76 in the Y-axis direction in the game space is determined based on the y component of the coordinate set. Accordingly, the direction in which the controller 7 is moved is the same as the direction in which the key 76 is moved on the display screen.

In step S48, processing of calculating the distance from the imaging means to the markers 8a and 8b (distance calculation processing) based on the length calculated in step S19 is executed. Hereinafter, the distance calculation processing will be described in detail.

FIG. 25 is a flowchart illustrating a detailed flow of the distance calculation processing in step S48 shown in FIG. 24. FIG. 26A and FIG. 26B illustrate the processing in step S48. FIG. 26A shows an example of a taken image obtained by the controller 7, and FIG. 26B shows the positional relationship between the markers 8a and 8b and the controller 7 when the taken image is obtained. In step S19 described above, a length mi between the two points in the taken image is calculated.

The distance calculation processing is executed as follows. First in step S481, a width w (see FIG. 26B) of an imaging possible range of the imaging element 40 with respect to the positions of the markers 8a and 8b is calculated. Specifically, the CPU 10 calculates the width w from the following equation.
w=wi×m/mi

In the above equation, m is the interval between the markers 8a and 8b (for example, 30 cm), which is a fixed value. wi is the width of the taken image obtained by the imaging element 40 which corresponds to the width w. wi is also a fixed value. The interval m and the width wi are both fixed values and therefore are stored on storage means (not shown) in the game apparatus 3 in advance. The interval m may have any value determined by the player in accordance with the environment; i.e., the player can place the markers 8a and 8b at any interval. In this case, the player may input the interval m.

Next in step S482, a distance d between the markers 8a and 8b and the imaging element 40 (the controller 7) is calculated based on the width w and the viewing angle θ2 of the imaging element 40. Specifically, the CPU 10 calculates the distance d from the following equation.
tan(θ2/2)=(w/2)/d=w/2d

The viewing angle θ2 is a fixed value, and therefore is stored on storage means (not shown) in the game apparatus 3 in advance. Data on the distance d calculated by the above equation is stored on the main memory 13 as the distance data 63. After step S482, the CPU 10 terminates the distance calculation processing.

Returning to FIG. 24, in step S49, the position of the key 76 in the Z-axis direction in the game space is determined based on the distance data 63 calculated in step S48. For example, the position of the key 76 in the Z-axis direction is determined such that as the value of the distance represented by the distance data 63 is greater, the key 76 is located forward with respect to the display screen (i.e., closer to the player in front of the display screen of the monitor 2). Accordingly, with respect to the depth direction as well as the up, down, left and right directions of the display screen, the direction in which the controller 7 is moved is the same as the direction in which the key 76 is moved. With such settings, the player can feel as if he/she was actually moving the key 76 by performing an operation of moving the controller 7. Data on the position of the key 76 determined by the processing in steps S46 through S49 is stored on the main memory 13 as the position data 66. After step S49, the CPU 10 terminates the moving processing.

In the case where the position of the key 76 determined in the moving processing in step S3 is beyond the wall or the door 75, the position data 66 of the key 76 is not updated and the contents of the position data 66 before the moving processing in step S3 are maintained. Namely, in the case where the key 76 collides against the wall or the door 75, the key 76 is not moved. Unless the key 76 is in a predetermined direction around the Z axis, the key 76 does not pass through the key hole 75a.

In step S4 after step S3, the display processing is executed. Namely, the key 76 is displayed on the display screen with the new position and posture determined in step S3. By repeating the processing in step S4 frame by frame, the key 76 is displayed as moving and rotating in accordance with the movement and the rotation of the controller 7.

Next in step S42, it is determined whether or not the key 76 is located in the key hole 75a. This determination can be performed by referring to the position data 66 in the operation target data 64 stored on the main memory 13 and data representing the position of the key hole 75a (stored on the main memory 13). When it is determined in step S42 that the key 76 is located in the key hole 75a, processing in step S43 is executed. When it is determined in step S42 that the key 76 is not located in the key hole 75a, the processing returns to step S1 and the above-described processing is repeated.

In step S43, it is determined whether or not the key 76 is at a predetermined posture around the Z axis. This determination can be performed by referring to the posture data 65 in the operation target data 64 stored on the main memory 13. When it is determined in step S43 that the key 76 is at the predetermined posture, processing in step S44 is executed. When it is determined in step S43 that the key 76 is not at the predetermined posture, the processing returns to step S1 and the above-described processing is repeated.

In step S44, the door 75 is displayed as being opened on the display screen. The processing after step S44 is not directly relevant and thus will not be described in detail. After this, a player character can move beyond the door 75 or the scene beyond the door 75 is displayed on the display screen. So far, the game processing in the second example has been described.

As described above, in the second example, the direction in which the controller 7 is moved or rotated is matched to the direction in which the operation target is moved or rotated. Thus, the player can feel as if he/she was actually moving or rotating the operation target, which was not possible conventionally.

In the above example, the distance between the imaging target (the markers 8a and 8b) and the imaging means (the imaging element 40) is calculated based on the size of the image in the taken image (the total length of the two target images 8a′ and 8b′). In other embodiments, in the case where the strength of the light received by the imaging means is detectable, the distance may be calculated based on the strength of the light from the markers which is received by the imaging means.

THIRD EXAMPLE

Next, the game processing in a third example will be described. In the third example, a three-dimensional game space is constructed, and a virtual camera for generating an image of the game space is set. In the third example, the virtual camera is the operation target. In this example, processing of moving the virtual camera will be mainly described, and other game processing will be omitted.

FIG. 27 shows an example of a game space in the third example. FIG. 27 shows a virtual camera 81, as the operation target, rotating or moving on an arc. In FIG. 27, the reference numeral 81 represents a pre-movement position and direction of the virtual camera, and reference numeral 81′ represents a post-movement position and direction of the virtual camera. In FIG. 27, a coordinate system representing positions in the game space (X-Y-Z coordinate system) is set. In this coordinate system, a ground 83 is parallel to the X axis and the Z axis, and the Y axis is vertical to the ground 83. The virtual camera 81 moves on an arc around a predetermined point in the game space which is designated by the player in accordance with the rotation operation. Specifically, the virtual camera 81 moves on an arc parallel to the X-Z plane without the Y-axis coordinate thereof being changed. The direction of the virtual camera 81 is controlled such that the point of attention of the virtual camera 81 is not changed even when the virtual camera is moved on an arc. In the third example, the only motion performed by the virtual camera is to move on an arc.

Next, the game processing in the third example will be described in detail. FIG. 28 shows main data stored on the main memory 13 in the third example. The length data 59 in the operation state data 56 shown in FIG. 11 is not necessary for the game processing and is not stored on the main memory 13. The distance data 63 in the operation processing data 60 is not necessary for the game processing and is not stored on the main memory 13. The operation target data 64 shown in FIG. 11 represents a position and a posture of the virtual camera as the operation target. Namely, the posture data 65 represents a posture of the virtual camera, and the position data 66 represents a position of the virtual camera in the game space.

In the third example, corresponding coordinate set data 67 is stored on the main memory 13 in addition to the data shown in FIG. 11. The corresponding coordinate set data 67 represents a three-dimensional coordinate set indicating the position in the game space corresponding to the indicated position. In the third example, reference direction data 68 representing a reference direction is stored on the main memory 13 in addition to the data shown in FIG. 11. The “reference direction” is a direction used as the reference with respect to the direction obtained from the taken image. The reference direction is used when determining a rotation amount of the virtual camera, and the rotation amount of the virtual camera is determined based on an angular difference between the reference direction and the current direction (the direction represented by the direction data 57). In the third example, the reference direction is set by the player at the start of the game.

FIG. 29 is a flowchart illustrating a flow of the game processing in the third example. In FIG. 29, substantially the same steps as those in the game processing shown in FIG. 12 bear the same step numbers, and detailed descriptions thereof will be omitted. The processing in the flowchart shown in FIG. 29 is started is the same as in the case of FIG. 12.

The game processing in the third example is executed as follows. First in step S50, a three-dimensional game space is constructed. Namely, the CPU 10 constructs a three-dimensional game space and locates objects in the game space. The CPU 10 also sets a virtual camera at a predetermined position in the game space with a predetermined viewing direction. The predetermined position and the predetermined viewing direction are determined in advance.

Next in step S51, processing of setting the reference direction is executed. Specifically, the CPU 10 urges the player to perform an operation of setting the reference direction through display or the like. In accordance with this, the player holds the controller 7 with an arbitrary posture with respect to the indicated direction and performs an operation of setting the reference direction (for example, presses the A button). When this operation is performed, the CPU 10 calculates the reference direction based on the taken image obtained when the operation is performed, by substantially the same processing as in step S16. Data representing a vector indicating the calculated reference direction is stored on the main memory 13 as the reference direction data 68. In the third example, such processing in step S51 is executed before the processing of moving the virtual camera (steps S1 through S3).

After step S51, the processing in step S1 described above is executed, and then the operation state calculation processing in step S2 is executed. In the operation state calculation processing in the third example, the processing of calculating the distance between the two coordinate sets (step S19) is not executed. The operation state calculation processing in the third example is substantially the same as described above with reference to FIG. 12 except for step S19.

Next, the moving processing in step S3 is executed. FIG. 30 is a flowchart illustrating a detailed flow of the moving processing in the third example. The moving processing in the third example is executed as follows. First in steps S52 and S53, a position of the central point around which the virtual camera moves is determined. First in step S52, the indicated position on the display screen is calculated. A method for calculating the indicated position (indicated coordinate set) is the same as described above regarding steps S46 and S22. Next in step S53, the corresponding coordinate set corresponding to the indicated coordinate set is calculated. In this example, the position of the corresponding coordinate set is the central position around which the virtual camera is moved. Specifically, the position of the corresponding coordinate set may be a three-dimensional position at which a target displayed at the indicated position on the display screen (for example, the ground) exists in the game space, or a three-dimensional position which is set based on the indicated position on the display screen and a predetermined depth direction (the viewing direction of the virtual camera). Data on the corresponding coordinate set calculated in step S53 is stored on the main memory 13 as the corresponding coordinate set data 67. The processing in steps S52 and S53 correspond to step S22 described above.

In steps S54 through S56, processing of rotating the virtual camera in a direction calculated in accordance with the rotation of the controller 7 around the indicated direction (i.e., rotating the virtual camera in the direction calculated in step S16). The processing in steps S54 through S56 corresponds to the processing in step S21 shown in FIG. 2. Hereinafter, specific processing in step S21 in the third example will be described.

First in step S54, the CPU 10 calculates an angle (angular difference) made by the direction calculated in step S16 and the reference direction. Specifically, the CPU 10 calculates an angle made by the vector represented by the reference direction data 67 stored on the main memory 13 and the vector represented by the direction data 57 in the operation state data 56. The obtained angle θ′ is represented in the range of, for example, −180°<θ′≦180°. Data on the calculated angular difference is stored on the main memory 13 as the direction processing data 61.

Next in step S55, the CPU 10 determines a rotation amount of the virtual camera (the rotation angle θ shown in FIG. 27) in accordance with the angle calculated in step S54. For example, the rotation amount of the virtual camera may be the calculated angle, or may be obtained by multiplying the calculated angle by a constant. The virtual camera as the operation target may be set to be rotated only when the angle is a predetermined value or greater. The CPU 10 rotates the virtual camera in a different direction when the calculated angle has a positive value from when the calculated angle has a negative value.

Next in step S56, a post-rotation position and posture of the virtual camera are calculated. Specifically, the CPU 10 calculates a position of the virtual camera after the virtual camera is rotated by the rotation amount determined in step S55 around the central position (the position of the corresponding coordinate set) determined in step S52. As described above, the direction of the virtual camera is controlled such that the point of attention is not changed from before the rotation. Data representing the position of the virtual camera calculated in step S56 is stored on the main memory 13 as the position data 66. Data representing the posture of the virtual camera calculated in step S56 is stored on the main memory 13 as the posture data 65. After step S56, the CPU 10 terminates the moving processing.

In step S4, the display processing is executed using the position and the posture (the viewing direction) of the virtual camera which has been moved in step S3. Namely, an image of the game space seen from the post-movement virtual camera is generated, and the generated image is displayed on the display screen. The processing in step S5 is substantially the same as described above with reference to FIG. 12.

As described above, in the third example, the operation target may be rotated using the direction obtained from the taken image (the direction calculated in step S16) or may be rotated using the angular difference between the direction obtained from the taken image and the reference direction. The center of the rotation may be designated by the player by performing the position designation operation. In other embodiments, the point of attention of the virtual camera may be the center of the rotation. The direction of the virtual camera may be moved such that the corresponding coordinate set corresponding to the indicated coordinate set is the point of attention, in which case the virtual camera is moved on an arc around the point of attention as the center of rotation while the point of attention is fixed. In this case, both the point of attention and the direction of the virtual camera can be simultaneously set. Alternatively, only the posture (direction) of the virtual camera may be changed in accordance with the angular difference without changing the position of the virtual camera.

In the third example, the virtual camera is rotated based on the angular difference. Even in the case where an object appearing in the game space is the operation target, the object can be rotated based on the angular difference. For example, the objects described in the first and second examples (the player object 71 and the key 76) may be rotated based on the angular difference. The virtual camera may be rotated based on a calculation using the direction obtained from the taken image itself, instead of the angular difference.

In other embodiments, an object appearing in the game space may be set as the operation target and translated to the position of the corresponding coordinate set and rotated in the direction represented by the direction data 57.

In the third example, the difference between the direction represented by the direction data 57 and the reference direction is calculated as the angular difference. Alternatively, the difference between the vector represented by the direction data 57 and the vector representing the reference direction may be calculated as a vector. The value of a predetermined component of the calculated vector may be used instead of the angular difference. Thus, the processing cost of calculating the angular difference can be alleviated while allowing controls to be performed based on the inclination of the controller 7.

FOURTH EXAMPLE

FIG. 31 shows an example of a game screen in a fourth example. FIG. 31 shows a game scene in which the player can open the door in the game space. In FIG. 31, a three-dimensional game space is constructed, and two doors 85 and 86 are located in the game space. The door 85 has a door knob 85a, and the door 86 has a door knob 86a. In this example, the two door knobs 85a and 86a are objects, each of which can be an operation target.

On the display screen, a cursor 87 movable by the player is displayed. The player operates the controller 7 to move the cursor 87. The cursor 87 is displayed at the indicated position. The player can use the cursor 87 to designate the object to be set as the operation target (the door knob 85a or 86a). In this example, the object designated by the cursor 87 can be the operation target of the rotation operation.

Next, the game processing in the fourth example will be described in detail. In the fourth example, the operation target data 64 is not stored at the start of the game. The door knobs which can be the operation target is only rotatable and does not move, and therefore the operation target data 64 does not include the position data 66 in the fourth example. In this example, the length data 59 in the operation state data 56 shown in FIG. 11 is not necessary for the game processing and is not stored on the main memory 13. The direction processing data 61 and the distance data 63 in the operation processing data 60 shown in FIG. 11 are not necessary for the game processing and are not stored on the main memory 13.

FIG. 32 is a flowchart illustrating a flow of the game processing in the fourth example. In FIG. 32, substantially the same steps as those in the game processing shown in FIG. 12 bear the same step numbers, and detailed descriptions thereof will be omitted. The processing until the processing in the flowchart shown in FIG. 32 is started is the same as in the case of FIG. 12.

The game processing in the fourth example is executed as follows. First in step S61, initial processing is executed. In the initial processing, the CPU 10 displays a game screen in an initial state. Namely, a three-dimensional game space is constructed, and the doors 85 and 86 and the like are located. The cursor 87 is displayed at a predetermined position on the display screen.

After step S61, the processing in step S1 described above is executed, and then the operation state calculation processing in step S2 is executed. In the operation state calculation processing in the fourth example, the processing of calculating the distance between the two coordinate sets (step S19) is not executed. The operation state calculation processing in the fourth example is substantially the same as described above with reference to FIG. 12 except for step S19.

In the fourth example, after step S2, processing in step S62 is executed. In the step S62, the indicated position is calculated. The processing in step S62 is substantially the same as in step S46. Data on the coordinate set representing the indicated position (indicated coordinate set) is stored on the main memory 13 as the indicated coordinate set data 62. Next in step S63, the position of the cursor 87 is updated in accordance with the indicated coordinate set. The position of the cursor 87 is determined as the position of the indicated coordinate set.

In step S64, it is determined whether or not the cursor 87, the position of which was updated in step S63, designates an object which can be the operation target (the door knob 85a or 86a). Namely, it is determined whether or not the position indicated by the cursor 87 is a display position of the door knob 85a or 86a (i.e., whether or not the cursor 87 is in a predetermined range from the display position of the door knob 85a or 86a). When it is determined in step S64 that the cursor 87 designates an object, processing in step S65 is executed. When it is determined in step S64 that the cursor 87 does not designate an object, the processing in steps S65 and S3 is skipped and the processing in step S4 is executed.

In step S65, the object which is determined to be designated by the cursor 87 in step S64 is set as the operation target. Specifically, data representing a posture of such an object is stored on the main memory 13 as the posture data 65 in the operation target data 64.

In step S3, the moving processing is executed. In the moving processing in the fourth example, as described above, the processing of rotating the operation target (step S21) in a direction calculated in accordance with the rotation of the controller 7 around the indicated direction as an axis (around the direction calculated in step S16) is executed. In the fourth example, the operation target is the object, the posture of which is stored on the main memory 13 as the operation target data 64, i.e., the door knob designated by the cursor 87.

In step S4, the display processing is executed. Specifically, the cursor 87 is displayed at the position updated in step S63. In the case where the processing in step S3 is executed, the door knob 85a or 86a is displayed as being rotated. Next in step S66, it is determined whether or not either the door knob 85a or 86a assumes a predetermined posture. This determination is substantially the same as the processing in step S43 in the second example. When it is determined in step S66 that either the door knob 85a or 86a assumes the predetermined posture, processing in step S67 is executed. When it is determined in step S66 that neither the door knob 85a nor the door knob 86a assumes the predetermined posture, the processing returns to step S1 and the above-described processing is repeated. In step S67, the door having the door knob assuming the predetermined posture is displayed on the display screen as being opened.

As described above, in the fourth example, the operation target which is to be rotated by the player is not limited to a predetermined object and may be an object designated by the player. Alternatively, an object as the operation target may be designated using the indicated position by the controller 7.

In the fourth example, a position on the display screen is designated by the controller 7 to move the cursor to the position, and thus the door knob as the operation target is selected. In other embodiments, any method is usable for designating the position on the display screen. For example, the cursor may be moved by a cross key provided on the controller 7.

In the fourth example, the cursor 87 is set on the display screen and moves on the two-dimensional plane of the display screen. In other embodiments, the cursor 87 may be movable in a three-dimensional game space. The cursor 87 may be moved in the three-dimensional game space in, for example, the same manner as with the virtual camera in the third example. The door knob may be set to be rotatable only when the position of the cursor 87 in the Z-axis direction matches the position of the door knob in the Z-axis direction (or when the distance between the two positions is within a predetermined distance). In this case, the CPU 10 determines whether or not there is any object (door knob) which can be the operation target at the Z-axis direction position of the coordinate set of the cursor 87. When it is determined that there is an object, the object is set as the operation target.

OTHER EXAMPLES

Other non-limiting illustrative embodiments may be usable with various games other than those described above. For example, one non-limiting illustrative embodiment may be applicable to a racing game. Specifically, the racing car operated by the player is set as the operation target, and the direction of the racing car seen from the above is changed by rotating the racing car in accordance with the rotation of the controller 7 around the indicated direction as an axis. Namely, the driving direction of the racing car is changed. Thus, the racing car rotates as shown in FIG. 17A or FIG. 17C. The player can change the driving direction of the racing car by rotating the controller 7 as if he/she was turning the steering wheel of an actual automobile. The player can intuitively operate the racing car.

In the above description, the information processing system and the program may be realized as a game system and a game program. However, other non-limiting illustrative embodiments may be applicable to any information processing system using an input device for determining the position or direction of the input device itself using a taken image.

A non-limiting illustrative embodiment provides a novel operation of an information processing system using an input device which is used by the user while being held in his/her hand, and is applicable to a game apparatus or a game program, for example.

While the non-limiting illustrative embodiments have been described in detail, the foregoing description is in all aspects illustrative and not restrictive. It is understood that numerous other modifications and variations can be devised to form yet more non-limiting illustrative embodiments. The invention is intended to be defined by the claims and to cover all corresponding and equivalent arrangements whether or not specifically disclosed herein.

Claims

1. A non-transitory computer readable storage medium having stored thereon an information processing program executable by a computer of an information processing apparatus for receiving operation data from an operation device including an imager for taking an image of an imaging target and for displaying, on a display device, a virtual space obtained by performing calculation processing on a predetermined operation target using the operation data, the information processing program causing the computer to execute:

obtaining a taken image taken by the imager of the operation device as the operation data;
calculating a two-dimensional vector using a position of the imaging target in the taken image;
rotating the operation target in accordance with a calculated value of the two-dimensional vector; and
displaying the virtual space on a display area of the display device in accordance with the rotated operation target.

2. A storage medium according to claim 1, wherein calculating a two-dimensional vector using a position of the imaging target in the taken image further includes:

calculating coordinate sets of two predetermined points in the image of the imaging target in a coordinate system corresponding to the taken image; and
calculating the two-dimensional vector connecting the coordinate sets of the two predetermined points.

3. A storage medium according to claim 2, wherein:

the information processing program further causes the computer to execute calculating a predetermined indicated coordinate set which corresponds to a position on the display area, in correspondence with a coordinate set of an intermediate point between the two predetermined points; and
wherein the rotating the operation target is further done in accordance with the indicated coordinate set.

4. A storage medium according to claim 3, wherein the indicated coordinate set which corresponds to the position on the display area is calculated in correspondence with a position of the image of the imaging target in the case where the image taken by the imager is rotated around the center of the image as an axis and the two-dimensional vector is directed in one certain direction by the rotation.

5. A storage medium according to claim 3, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to execute calculating a three-dimensional coordinate set in the virtual space which corresponds to the indicated coordinate set and setting the three-dimensional coordinate set as a three-dimensional indicated coordinate set; and
wherein the operation target is rotated in the three-dimensional space by a calculation using the two-dimensional vector and the three-dimensional indicated coordinate set.

6. A storage medium according to claim 5, wherein:

the information processing program further causes the computer to execute locating an object as the operation target in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: moving the object to the calculated three-dimensional coordinate set; and rotating the object.

7. A storage medium according to claim 5, wherein:

the information processing program further causes the computer to execute locating an object as the operation target in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: setting the three-dimensional coordinate set as a central coordinate set of rotation; and rotating the object in the three-dimensional space around the central coordinate set as a center of rotation.

8. A storage medium according to claim 5, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to execute setting a virtual camera directed in a predetermined direction at a predetermined position in the virtual space; and
wherein the operation target is the virtual camera, and is rotated around the three-dimensional coordinate set as a center of rotation.

9. A storage medium according to claim 3, wherein:

the information processing program further causes the computer to execute locating at least one object in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: determining whether or not any one of the at least one object is displayed at the position on the display area which corresponds to the indicated coordinate set; if one of the at least one object is displayed at the position which corresponds to the indicated coordinate set, then setting the one object as the operation target; and rotating the one object.

10. A storage medium according to claim 3, wherein:

the information processing program further causes the computer to execute locating an object as the operation target in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: moving the object so as to be displayed at a position of the calculated indicted coordinate set; and rotating the object.

11. A storage medium according to claim 3, wherein:

the information processing program further causes the computer to execute locating an object as the operation target in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: setting the indicated coordinate set as a central coordinate set of rotation; and rotating the object around the central coordinate set as a center of rotation.

12. A storage medium according to claim 2, wherein:

the information processing program further causes the computer to execute setting a predetermined indicated coordinate set which corresponds to a position on the display area as an indicated coordinate set; and
wherein the rotating the operation target is further done in accordance with the indicated coordinate set.

13. A storage medium according to claim 2, wherein:

the information processing program further causes the computer to execute calculating data representing a length of a distance between the coordinate sets of the two predetermined points; and
wherein the rotating the operation target is further done in accordance with the calculated data representing the calculated length.

14. A storage medium according to claim 13, wherein rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes:

calculating distance data representing a distance between the imager and the imaging target from the calculated data representing the calculated length; and
rotating the operation target by a calculation using the two-dimensional vector and the distance data.

15. A storage medium according to claim 14, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to execute locating at least one object in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: determining a coordinate in a depth direction of the virtual space in correspondence with the distance data; determining whether or not there is any one of the at least one object at a position of the coordinate in the calculated depth direction; and if there is one of the at least one object at a position of the coordinate in the calculated depth direction, then setting the one object as the operation target.

16. A storage medium according to claim 14, wherein:

the virtual space is a virtual three-dimensional space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: determining a coordinate in a depth direction of the virtual space in correspondence with the distance data; moving a position of the operation target in the depth direction to the coordinate in the calculated depth direction; and rotating the operation target in accordance with the two-dimensional vector.

17. A storage medium according to claim 1, wherein:

the information processing program further causes the computer to execute calculating a predetermined indicated coordinate set which corresponds to a position on the display area, in correspondence with a position of the image of the imaging target in the obtained taken image; and
wherein the rotating the operation target is further done in accordance with the indicated coordinate set.

18. A storage medium according to claim 17, wherein the indicated coordinate set which corresponds to the position on the display area is calculated in correspondence with a position of the image of the imaging target in the case where the image taken by the imager is rotated around the center of the image and the two-dimensional vector is directed in one certain direction by the rotation.

19. A storage medium according to claim 17, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to execute calculating a three-dimensional coordinate set in the virtual space which corresponds to the calculated indicated coordinate set and setting the three-dimensional coordinate set as a three-dimensional indicated coordinate set; and
wherein the operation target is rotated in the three-dimensional space by a calculation using the two-dimensional vector and the three-dimensional indicated coordinate set.

20. A storage medium according to claim 19, wherein:

the information processing program further causes the computer to execute locating an object as the operation target in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: moving the object to the calculated three-dimensional coordinate set; and rotating the object.

21. A storage medium according to claim 19, wherein:

the information processing program further causes the computer to execute locating an object as the operation target in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: setting the calculated three-dimensional coordinate set as a central coordinate set of rotation; and rotating the object in the three-dimensional space around the central coordinate set as a center of rotation.

22. A storage medium according to claim 19, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to execute setting a virtual camera directed in a predetermined direction at a predetermined position in the virtual space; and
wherein the operation target is the virtual camera, and is rotated around the three-dimensional coordinate set as a center of rotation.

23. A storage medium according to claim 17, wherein:

the information processing program further causes the computer to execute locating at least one object in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: determining whether or not any one of the at least one object is displayed at the position on the display area which corresponds to the indicated coordinate set; if one of the at least one object is displayed at the position which corresponds to the indicated coordinate set, then setting the one object as the operation target; and rotating the one object.

24. A storage medium according to claim 17, wherein:

the information processing program further causes the computer to execute locating an object as the operation target in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: moving the object so as to be displayed at a position of the calculated indicted coordinate set; and rotating the object.

25. A storage medium according to claim 17, wherein:

the information processing program further causes the computer to execute locating an object as the operation target in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: setting the calculated indicated coordinate set as a central coordinate set of rotation; and rotating the object around the central coordinate set as a center of rotation.

26. A storage medium according to claim 1, wherein:

the information processing program further causes the computer to execute setting a predetermined indicated coordinate set which corresponds to a position on the display area as an indicated coordinate set; and
wherein the rotating the operation target is further done in accordance with the indicated coordinate set.

27. A storage medium according to claim 1, wherein:

the information processing program further causes the computer to execute locating at least one object in the virtual space; and
wherein the operation target is any one of the at least one object located in the virtual space.

28. A storage medium according to claim 1, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to execute setting a virtual camera directed in a predetermined direction at a predetermined position in the virtual space;
the operation target is the virtual camera; and
wherein the virtual space is displayed as seen from the virtual camera on the display area.

29. A storage medium according to claim 28, wherein the virtual camera is rotated around a position of a point of attention of the virtual camera as a center of rotation.

30. A storage medium according to claim 28, wherein the virtual camera is rotated such that a viewing direction of the virtual camera changes.

31. A storage medium according to claim 1, wherein rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes changing a posture of the operation target by the rotation.

32. A storage medium according to claim 31, wherein the posture or the position of the operation target is determined in accordance with a calculated direction of the two-dimensional vector.

33. A storage medium according to claim 1, rotating the operation target further includes moving a position of the operation target around a predetermined position as a center of rotation.

34. A storage medium according to claim 33, wherein a posture or the position of the operation target is determined in accordance with a calculated direction of the two-dimensional vector.

35. A storage medium according to claim 1, wherein rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes:

determining whether or not a direction of the two-dimensional vector has exceeded a predetermined range; and
rotating the operation target when the direction of the two-dimensional vector has exceeded the predetermined range.

36. A storage medium according to claim 1, wherein:

the information processing program further causes the computer to execute obtaining a difference between a direction of the two-dimensional vector and a predetermined reference direction as an angle or a vector; and
wherein the rotating the operation target is further done in accordance with the obtained difference.

37. A storage medium according to claim 36, wherein rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes:

setting a rotation amount in accordance with a magnitude of the difference; and
rotating the operation target by the rotation amount.

38. A storage medium according to claim 36, wherein rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes:

determining whether or not the difference has exceeded a predetermined value; and
rotating the operation target when the difference has exceeded the predetermined value.

39. A storage medium according to claim 1, wherein:

the information processing program further causes the computer to execute calculating data representing a size of the image of the imaging target in the taken image; and
wherein the rotating the operation target is further done in accordance with the calculated data representing a size of the image.

40. A storage medium according to claim 39, wherein rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes:

calculating distance data representing a distance between the imager and the imaging target from the calculated data representing a size of the image; and
rotating the operation target by a calculation using the two-dimensional vector and the distance data.

41. A storage medium according to claim 40, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to execute locating at least one object in the virtual space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: determining a coordinate in a depth direction of the virtual space in correspondence with the distance data; determining whether or not there is any one of the at least one object at a position of the coordinate in the depth direction; and if there is one of the at least one object at a position of the determined coordinate in the depth direction, then setting the one object as the operation target.

42. A storage medium according to claim 40, wherein:

the virtual space is a virtual three-dimensional space; and
rotating the operation target in accordance with a calculated value of the two-dimensional vector further includes: determining a coordinate in a depth direction of the virtual space in correspondence with the distance data; moving a position of the operation target in the depth direction to the coordinate in the depth direction; and rotating the operation target in accordance with the two-dimensional vector.

43. A storage medium according to claim 1, wherein:

the information processing program further causes the computer to execute calculating data corresponding to a distance between the imager and the imaging target; and
wherein the rotating the operation target is further done in accordance with the calculated data corresponding to a distance between the imager and the imaging target a.

44. A storage medium according to claim 1, wherein:

the operation device comprises an inclination processing system for determining an inclination of the operation device; and
the information processing program further causes the computer to execute: asserting a two-dimensional vector calculation that is based on the taken image; and rotating the operation target in accordance with the inclination determined by the inclination processing system when the two-dimensional vector calculation assertion is not successful.

45. A non-transitory computer readable storage medium having stored thereon an information processing program executable by a computer of an information processing apparatus for receiving operation data from an operation device, which includes an imager for taking an image of an imaging target and a processor adapted to calculate coordinate sets of two predetermined points in the image of the imaging target included in a taken image taken by the imager, and for displaying, on a display device, a virtual space obtained by performing calculation processing on a predetermined operation target using the operation data, the information processing program causing the computer to:

obtain the coordinate sets of the two predetermined points as the operation data;
calculate a two-dimensional vector connecting the coordinate sets of the two predetermined points;
rotate the operation target in accordance with a calculated value of the two-dimensional vector; and
display the virtual space, changed in accordance with the rotation of the operation target, on a display area of the display device.

46. A storage medium according to claim 45, wherein:

the information processing program further causes the computer to calculate a predetermined indicated coordinate set which corresponds to a position on the display area, in correspondence with a coordinate set of an intermediate point between the two predetermined points,
wherein the operation target is rotated by a calculation using the two-dimensional vector and the indicated coordinate set.

47. A storage medium according to claim 46, wherein the indicated coordinate set which corresponds to the position on the display area is calculated in correspondence with a position of the image of the imaging target in the case where the image taken by the imager is rotated around the center of the image as an axis and the two-dimensional vector is directed in one certain direction by the rotation.

48. A storage medium according to claim 46, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to calculate a three-dimensional coordinate set in the virtual space which corresponds to the indicated coordinate set;
set the three-dimensional coordinate set as a three-dimensional indicated coordinate set; and
wherein the operation target is rotated in the three-dimensional space by a calculation using the two-dimensional vector and the three-dimensional indicated coordinate set.

49. A storage medium according to claim 48, wherein:

the information processing program further causes the computer to locate an object as the operation target in the virtual space; and
rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes: move the object to the three-dimensional coordinate set; and rotate the object.

50. A storage medium according to claim 48, wherein:

the information processing program further causes the computer to locate an object as the operation target in the virtual space; and
rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes: set the three-dimensional coordinate set as a central coordinate set of rotation; and rotate the object in the three-dimensional space around the central coordinate set as a center of rotation.

51. A storage medium according to claim 48, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to set a virtual camera directed in a predetermined direction at a predetermined position in the virtual space; and
wherein the operation target is the virtual camera, and is rotated around the three-dimensional coordinate set as a center of rotation.

52. A storage medium according to claim 46, wherein:

the information processing program further causes the computer to locate at least one object in the virtual space; and
rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes: determine whether or not any one of the at least one object is displayed at the position on the display area which corresponds to the indicated coordinate set; set the one object as the operation target when one of the at least one object is displayed at the position which corresponds to the indicated coordinate set; and rotate the one object.

53. A storage medium according to claim 46, wherein:

the information processing program further causes the computer to locate an object as the operation target in the virtual space; and
rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes: move the object so as to be displayed at a position of the indicted coordinate set; and rotate the object.

54. A storage medium according to claim 46, wherein:

the information processing program further causes the computer to locate an object as the operation target in the virtual space; and
rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes: set the indicated coordinate as a central coordinate set of rotation; and rotate the object around the central coordinate set as a center of rotation.

55. A storage medium according to claim 45, wherein:

the information processing program further causes the computer to set a predetermined indicated coordinate set which corresponds to a position on the display area as an indicated coordinate set; and
wherein the rotation of the operation target is further done in accordance with the indicated coordinate set.

56. A storage medium according to claim 45, wherein:

the information processing program further causes the computer to locate at least one object in the virtual space; and
wherein the operation target is any one of the at least one object located in the virtual space.

57. A storage medium according to claim 45, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to set a virtual camera directed in a predetermined direction at a predetermined position in the virtual space;
the virtual camera is the operation target; and
the display of the virtual space includes an image seen from the virtual camera on the display area.

58. A storage medium according to claim 57, wherein the virtual camera is rotated around a position of a point of attention of the virtual camera as a center of rotation.

59. A storage medium according to claim 57, wherein the virtual camera is rotated such that a viewing direction of the virtual camera changes.

60. A storage medium according to claim 45, wherein a posture of the operation target is changed by the rotation.

61. A storage medium according to claim 60, wherein the posture or the position of the operation target is determined in accordance with a calculated direction of the two-dimensional vector.

62. A storage medium according to claim 45, wherein a position of the operation target is moved by rotating the operation target around a predetermined position as the center of rotation.

63. A storage medium according to claim 62, wherein a posture or the position of the operation target is determined in accordance with a calculated direction of the two-dimensional vector.

64. A storage medium according to claim 45, wherein rotation of the operation target in accordance with a calculated value of the two-dimensional vector further includes:

determine whether or not a direction of the two-dimensional vector has exceeded a predetermined range; and
rotate the operation target if the direction of the two-dimensional vector has exceeded the predetermined range.

65. A storage medium according to claim 45, wherein:

the information processing program further causes the computer to obtain a difference between a direction of the two-dimensional vector and a predetermined reference direction as an angle or a vector; and
the operation target is rotated in further accordance with the difference.

66. A storage medium according to claim 65, wherein rotation of the operation target in accordance with a calculated value of the two-dimensional vector further includes:

set a rotation amount in accordance with a magnitude of the difference; and
rotate the operation target by the rotation amount.

67. A storage medium according to claim 65, wherein rotation of the operation target in accordance with a calculated value of the two-dimensional vector further includes:

determine whether or not the difference has exceeded a predetermined value; and
rotate the operation target if the difference has exceeded the predetermined value.

68. A storage medium according to claim 45, wherein:

the information processing program further causes the computer to calculate data representing a size of the image of the imaging target in the taken image; and
the operation target is rotated based on a calculation using the two-dimensional vector and the calculated data.

69. A storage medium according to claim 68, wherein rotation of the operation target in accordance with a calculated value of the two-dimensional vector further includes:

calculate distance data representing a distance between the imager and the imaging target from the calculated data; and
the operation target is rotated based on a calculation using the two-dimensional vector and the distance data.

70. A storage medium according to claim 69, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to locate at least one object in the virtual space; and
rotation of the operation target in accordance with a calculated value of the two-dimensional vector further includes: determine a coordinate in a depth direction of the virtual space in correspondence with the distance data; determine whether or not there is any one of the at least one object at a position of the coordinate in the depth direction; and set the one object as the operation target if there is one of the at least one object at a position of the coordinate in the depth direction.

71. A storage medium according to claim 69, wherein:

the virtual space is a virtual three-dimensional space; and
rotation of the operation target in accordance with a calculated value of the two-dimensional vector further includes: determine a coordinate in a depth direction of the virtual space in correspondence with the distance data; move a position of the operation target in the depth direction to the coordinate in the depth direction; and rotate the operation target in accordance with the two-dimensional vector.

72. A storage medium according to claim 45, wherein:

the information processing program further causes the computer to calculate data corresponding to a distance between the imager and the imaging target; and
the operation target is rotated by a calculation using the two-dimensional vector and the calculated data.

73. A storage medium according to claim 45, wherein:

the operation device comprises an inclination determiner to determine an inclination of the operation device; and
the information processing program further causes the computer to: assert a two-dimensional vector calculation that is based on the taken image; and rotate the operation target in accordance with the inclination determined by the inclination determiner when the asserted two-dimensional vector calculation fails.

74. A non-transitory computer readable storage medium having stored thereon an information processing program executable by a computer of an information processing apparatus for receiving operation data from an operation device, which includes an imager for taking an image of an imaging target and vector programmed logic circuitry for calculating a two-dimensional vector using a position of the imaging target in a taken image taken by the imager, and for displaying, on a display device, a virtual space obtained by performing calculation processing on a predetermined operation target using the operation data, the information processing program causing the computer to:

obtain the two-dimensional vector as the operation data;
rotate the operation target in accordance with an obtained value of the two-dimensional vector; and
display the virtual space, changed in accordance with the rotation of the operation target, on a display area of the display device.

75. A storage medium according to claim 74, wherein:

the operation data further includes a coordinate set of one point corresponding to a position of the image of the imaging target;
the information processing program further causes the computer to calculate a predetermined indicated coordinate set which corresponds to a position on the display area, in correspondence with the coordinate set of the one point; and
the operation target is rotated by a calculation using the two-dimensional vector and the indicated coordinate set.

76. A storage medium according to claim 75, wherein the indicated coordinate set which corresponds to the position on the display area is calculate in correspondence with a position of the image of the imaging target in the case where the image taken by the imager is rotated around the center of the image as an axis and the two-dimensional vector is directed in one certain direction by the rotation.

77. A storage medium according to claim 75, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to calculate a three-dimensional coordinate set in the virtual space which corresponds to the indicated coordinate set and set the three-dimensional coordinate set as a three-dimensional indicated coordinate set; and
wherein the operation target is rotated in the three-dimensional space by a calculation using the two-dimensional vector and the three-dimensional indicated coordinate set.

78. A storage medium according to claim 77, wherein:

the information processing program further causes the computer to locate an object as the operation target in the virtual space; and
rotation of the operation target in accordance with an obtained value of the two-dimensional vector includes: move the object to the three-dimensional coordinate set; and rotate the object.

79. A storage medium according to claim 77, wherein:

the information processing program further causes the computer to locate an object as the operation target in the virtual space; and
rotation of the operation target in accordance with an obtained value of the two-dimensional vector includes: set the three-dimensional coordinate set as a central coordinate set of rotation; and rotate the object in the three-dimensional space around the central coordinate set as the center of rotation.

80. A storage medium according to claim 77, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to set a virtual camera directed in a predetermined direction at a predetermined position in the virtual space; and
the virtual camera is rotated as the operation target around the three-dimensional coordinate set as the center of rotation.

81. A storage medium according to claim 75, wherein:

the information processing program further causes the computer to locate at least one object in the virtual space; and
rotation of the operation target in accordance with an obtained value of the two-dimensional vector includes: determine whether or not any one of the at least one object is displayed at the position on the display area which corresponds to the indicated coordinate set; set the one object as the operation target when the at least one object is displayed at the position which corresponds to the indicated coordinate set; and rotate the one object.

82. A storage medium according to claim 75, wherein:

the information processing program further causes the computer to locate an object as the operation target in the virtual space; and
rotation of the operation target in accordance with an obtained value of the two-dimensional vector includes: move the object so as to be displayed at a position of the indicted coordinate set; and rotate the object.

83. A storage medium according to claim 75, wherein:

the information processing program further causes the computer to locate an object as the operation target in the virtual space; and
rotation of the operation target in accordance with an obtained value of the two-dimensional vector includes: set the indicated coordinate set as a central coordinate set of rotation; and rotate the object around the central coordinate set as the center of rotation.

84. A storage medium according to claim 74, wherein:

the information processing program further causes the computer to set a predetermined indicated coordinate set which corresponds to a position on the display area as an indicated coordinate set; and
the operation target is rotated by a calculation using the two-dimensional vector and the indicated coordinate set.

85. A storage medium according to claim 74, wherein:

the information processing program further causes the computer to locate at least one object in the virtual space; and
any one of the at least one object located in the virtual space is the operation target.

86. A storage medium according to claim 74, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to set a virtual camera directed in a predetermined direction at a predetermined position in the virtual space;
the virtual camera is the operation target; and
an image of the virtual space seen from the virtual camera is displayed on the display area.

87. A storage medium according to claim 86, wherein the virtual camera is rotated around a position of a point of attention of the virtual camera as the center of rotation.

88. A storage medium according to claim 86, wherein the virtual camera rotates such that a viewing direction of the virtual camera changes.

89. A storage medium according to claim 74, wherein a posture of the operation target is changed by the rotation.

90. A storage medium according to claim 89, wherein the posture or a position of the operation target is determined in accordance with a calculated direction of the two-dimensional vector.

91. A storage medium according to claim 74, wherein a position of the operation target is moved by rotating the operation target around a predetermined position as the center of rotation.

92. A storage medium according to claim 91, wherein a posture or the position of the operation target is determined in accordance with a calculated direction of the two-dimensional vector.

93. A storage medium according to claim 74, wherein rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes:

determine whether or not a direction of the two-dimensional vector has exceeded a predetermined range; and
rotate the operation target if the direction of the two-dimensional vector has exceeded the predetermined range.

94. A storage medium according to claim 74, wherein:

the information processing program further causes the computer to obtain a difference between a direction of the two-dimensional vector and a predetermined reference direction as an angle or a vector; and
the operation target is rotated in accordance with the difference.

95. A storage medium according to claim 94, wherein rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes:

set a rotation amount in accordance with a magnitude of the difference; and
rotate the operation target by the rotation amount.

96. A storage medium according to claim 94, wherein rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes:

determine whether or not the difference has exceeded a predetermined value; and
rotate the operation target if the difference has exceeded the predetermined value.

97. A storage medium according to claim 74, wherein:

the operation data further includes data representing a length of a distance between the coordinate sets of the two predetermined points in the image of the imaging target included in the taken image taken by the imager; and
the operation target is rotated by a calculation using the two-dimensional vector and the data representing the length of the distance between the coordinate sets of the two predetermined points.

98. A storage medium according to claim 97, wherein rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes:

calculate distance data representing a distance between the imager and the imaging target from the data representing the length of the distance between the coordinate sets of the two predetermined points; and
rotate the operation target by a calculation using the two-dimensional vector and the distance data.

99. A storage medium according to claim 98, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to locate at least one object in the virtual space; and
rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes: determine a coordinate in a depth direction of the virtual space in correspondence with the distance data; determine whether or not there is any one of the at least one object at a position of the coordinate in the depth direction; and set the one object as the operation target when there is one of the at least one object at a position of the coordinate in the depth direction.

100. A storage medium according to claim 98, wherein:

the virtual space is a virtual three-dimensional space; and
rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes: determine a coordinate in a depth direction of the virtual space in correspondence with the distance data; move a position of the operation target in the depth direction to the coordinate in the depth direction; and rotate the operation target in accordance with the two-dimensional vector.

101. A storage medium according to claim 74, wherein:

the operation data further includes distance data representing a distance between the imager and the imaging target; and
the operation target is rotate by a calculation using the two-dimensional vector and the distance data.

102. A storage medium according to claim 101, wherein:

the virtual space is a virtual three-dimensional space;
the information processing program further causes the computer to locate at least one object in the virtual space; and
rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes: determine a coordinate in a depth direction of the virtual space in correspondence with the distance data; determine whether or not there is any one of the at least one object at a position of the coordinate in the depth direction; and set the one object as the operation target when there is one of the at least one object at a position of the coordinate in the depth direction.

103. A storage medium according to claim 101, wherein:

the virtual space is a virtual three-dimensional space; and
rotation of the operation target in accordance with a calculated value of the two-dimensional vector includes: determine a coordinate in a depth direction of the virtual space in correspondence with the distance data; move a position of the operation target in the depth direction to the coordinate in the depth direction; and rotate the operation target in accordance with the two-dimensional vector.

104. A storage medium according to claim 74, wherein:

the information processing program further causes the computer to calculate data corresponding to a distance between the imager and the imaging target; and
the operation target is rotated by a calculation using the two-dimensional vector and the calculated data.

105. A storage medium according to claim 74, wherein:

the operation device comprises an inclination sensor configured to determine an inclination of the operation device; and
the information processing program further causes the computer to: determine an assertion of a two-dimensional vector calculation based on the taken image; and rotate the operation target in accordance with the inclination determined by the inclination sensor when the assertion is not successful.
Referenced Cited
U.S. Patent Documents
3454920 July 1969 Mehr
3474241 October 1969 Kuipers
D220268 March 1971 Kliewer
3660648 May 1972 Kuipers
3973257 August 3, 1976 Rowe
4009619 March 1, 1977 Snyman
4038876 August 2, 1977 Morris
4166406 September 4, 1979 Maughmer
4240638 December 23, 1980 Morrison et al.
4287765 September 8, 1981 Kreft
4303978 December 1, 1981 Shaw et al.
4318245 March 9, 1982 Stowell et al.
4321678 March 23, 1982 Krogmann
4337948 July 6, 1982 Breslow
4342985 August 3, 1982 Desjardins
4402250 September 6, 1983 Baasch
4425488 January 10, 1984 Moskin
4443866 April 17, 1984 Burgiss, Sr.
4450325 May 22, 1984 Luque
4503299 March 5, 1985 Henrard
4514600 April 30, 1985 Lentz
4514798 April 30, 1985 Lesche
4540176 September 10, 1985 Baer
4546551 October 15, 1985 Franks
4558604 December 17, 1985 Auer
4561299 December 31, 1985 Orlando et al.
4578674 March 25, 1986 Baker et al.
4623930 November 18, 1986 Oshima et al.
4672374 June 9, 1987 Desjardins
4739128 April 19, 1988 Grisham
4761540 August 2, 1988 McGeorge
4787051 November 22, 1988 Olson
4816810 March 28, 1989 Moore
4839838 June 13, 1989 LaBiche et al.
4849655 July 18, 1989 Bennett
4851685 July 25, 1989 Dubgen
4862165 August 29, 1989 Gart
4914598 April 3, 1990 Krogmann et al.
4918293 April 17, 1990 McGeorge
4957291 September 18, 1990 Miffitt et al.
4961369 October 9, 1990 McGill
4969647 November 13, 1990 Mical et al.
4988981 January 29, 1991 Zimmerman et al.
4994795 February 19, 1991 MacKenzie
5045843 September 3, 1991 Hansen
D320624 October 8, 1991 Taylor
5059958 October 22, 1991 Jacobs et al.
5062696 November 5, 1991 Oshima et al.
5068645 November 26, 1991 Drumm
D322242 December 10, 1991 Cordell
D325225 April 7, 1992 Adhida
5124938 June 23, 1992 Algrain
5128671 July 7, 1992 Thomas, Jr.
D328463 August 4, 1992 King et al.
5136222 August 4, 1992 Yamamoto
5138154 August 11, 1992 Hotelling
D331058 November 17, 1992 Morales
5175481 December 29, 1992 Kanno
5178477 January 12, 1993 Gambaro
5181181 January 19, 1993 Glynn
5192082 March 9, 1993 Inoue et al.
5202844 April 13, 1993 Kamio et al.
5207426 May 4, 1993 Inoue et al.
D338242 August 10, 1993 Cordell
D340042 October 5, 1993 Copper et al.
5259626 November 9, 1993 Ho
5262777 November 16, 1993 Low et al.
D342256 December 14, 1993 Payne
5280744 January 25, 1994 DeCarlo et al.
D345164 March 15, 1994 Grae
5296871 March 22, 1994 Paley
5307325 April 26, 1994 Scheiber
5309137 May 3, 1994 Kajiwara
5317394 May 31, 1994 Hale et al.
5319387 June 7, 1994 Yoshikawa
5329276 July 12, 1994 Hirabayashi
5332322 July 26, 1994 Gambaro
5339095 August 16, 1994 Redford
D350736 September 20, 1994 Takahashi et al.
D350782 September 20, 1994 Barr
D351430 October 11, 1994 Barr
5357267 October 18, 1994 Inoue
5359321 October 25, 1994 Ribic
5359348 October 25, 1994 Pilcher et al.
5363120 November 8, 1994 Drumm
5369580 November 29, 1994 Monji et al.
H1383 December 6, 1994 Kaplan et al.
5369889 December 6, 1994 Callaghan
5373857 December 20, 1994 Hirabayashi et al.
5396265 March 7, 1995 Ulrich et al.
5421590 June 6, 1995 Robbins
5430435 July 4, 1995 Hoch et al.
D360903 August 1, 1995 Barr et al.
5440326 August 8, 1995 Quinn
5453758 September 26, 1995 Sato
D362870 October 3, 1995 Oikawa
5459489 October 17, 1995 Redford
5469194 November 21, 1995 Clark et al.
5481957 January 9, 1996 Paley et al.
5484355 January 16, 1996 King, II et al.
5485171 January 16, 1996 Copper et al.
5490058 February 6, 1996 Yamasaki et al.
5502486 March 26, 1996 Ueda et al.
5506605 April 9, 1996 Paley
5512892 April 30, 1996 Corballis et al.
5517183 May 14, 1996 Bozeman, Jr.
5523800 June 4, 1996 Dudek
5526022 June 11, 1996 Donahue et al.
5528265 June 18, 1996 Harrison
5531443 July 2, 1996 Cruz
5541860 July 30, 1996 Takei et al.
5551701 September 3, 1996 Bouton et al.
5554033 September 10, 1996 Bizzi
5554980 September 10, 1996 Hashimoto et al.
5561543 October 1, 1996 Ogawa
5563628 October 8, 1996 Stroop
5569085 October 29, 1996 Igarashi et al.
D375326 November 5, 1996 Yokoi et al.
5573011 November 12, 1996 Felsing
5574479 November 12, 1996 Odell
5579025 November 26, 1996 Itoh
D376826 December 24, 1996 Ashida
5587558 December 24, 1996 Matsushima
5594465 January 14, 1997 Poulachon
5598187 January 28, 1997 Ide et al.
5602569 February 11, 1997 Kato
5603658 February 18, 1997 Cohen
5605505 February 25, 1997 Han
5606343 February 25, 1997 Tsuboyama et al.
5611731 March 18, 1997 Bouton et al.
5615132 March 25, 1997 Horton et al.
5621459 April 15, 1997 Ueda et al.
5624117 April 29, 1997 Ohkubo et al.
5627565 May 6, 1997 Morishita et al.
D379832 June 10, 1997 Ashida
5640152 June 17, 1997 Copper
5641288 June 24, 1997 Zaenglein, Jr.
5643087 July 1, 1997 Marcus et al.
5645077 July 8, 1997 Foxlin
5645277 July 8, 1997 Cheng
5666138 September 9, 1997 Culver
5667220 September 16, 1997 Cheng
5670845 September 23, 1997 Grant et al.
5670988 September 23, 1997 Tickle
5676673 October 14, 1997 Ferre et al.
5679004 October 21, 1997 McGowan et al.
5682181 October 28, 1997 Nguyen et al.
5698784 December 16, 1997 Hotelling et al.
5701131 December 23, 1997 Kuga
5702305 December 30, 1997 Norman et al.
5703623 December 30, 1997 Hall et al.
5724106 March 3, 1998 Autry et al.
5726675 March 10, 1998 Inoue
5734371 March 31, 1998 Kaplan
5734373 March 31, 1998 Rosenberg et al.
5734807 March 31, 1998 Sumi
D393884 April 28, 1998 Hayami
5736970 April 7, 1998 Bozeman, Jr.
5739811 April 14, 1998 Rosenberg et al.
5741182 April 21, 1998 Lipps et al.
5742331 April 21, 1998 Uomori et al.
5745226 April 28, 1998 Gigioli, Jr.
D394264 May 12, 1998 Sakamoto et al.
5746602 May 5, 1998 Kikinis
5751273 May 12, 1998 Cohen
5752880 May 19, 1998 Gabai et al.
5757354 May 26, 1998 Kawamura
5757360 May 26, 1998 Nitta et al.
D395464 June 23, 1998 Shiibashi et al.
5764224 June 9, 1998 Lilja et al.
5769719 June 23, 1998 Hsu
5771038 June 23, 1998 Wang
D396468 July 28, 1998 Schindler et al.
5785317 July 28, 1998 Sasaki
D397162 August 18, 1998 Yokoi et al.
5794081 August 11, 1998 Itoh et al.
5796354 August 18, 1998 Cartabiano et al.
5807284 September 15, 1998 Foxlin
5819206 October 6, 1998 Horton
5820462 October 13, 1998 Yokoi et al.
5822713 October 13, 1998 Profeta
5825350 October 20, 1998 Case, Jr. et al.
D400885 November 10, 1998 Goto
5831553 November 3, 1998 Lenssen et al.
5835077 November 10, 1998 Dao
5835156 November 10, 1998 Blonstein et al.
5841409 November 24, 1998 Ishibashi et al.
D402328 December 8, 1998 Ashida
5847854 December 8, 1998 Benson, Jr.
5850624 December 15, 1998 Gard et al.
5854622 December 29, 1998 Brannon
D405071 February 2, 1999 Gambaro
5867146 February 2, 1999 Kim et al.
5874941 February 23, 1999 Yamada
5875257 February 23, 1999 Marrin et al.
D407071 March 23, 1999 Keating
D407761 April 6, 1999 Barr
5897437 April 27, 1999 Nishiumi et al.
5898421 April 27, 1999 Quinn
5900867 May 4, 1999 Schindler et al.
5902968 May 11, 1999 Sato et al.
D410909 June 15, 1999 Tickle
5912612 June 15, 1999 DeVolpi
5919149 July 6, 1999 Allum
5923317 July 13, 1999 Sayler et al.
5926780 July 20, 1999 Fox et al.
5929782 July 27, 1999 Stark et al.
D412940 August 17, 1999 Kato
5947868 September 7, 1999 Dugan
5955713 September 21, 1999 Titus et al.
5955988 September 21, 1999 Blonstein et al.
5956035 September 21, 1999 Scianmanella et al.
5967898 October 19, 1999 Takasaka et al.
5973757 October 26, 1999 Aubuchon et al.
5982352 November 9, 1999 Pryor
5982356 November 9, 1999 Akiyama
5984785 November 16, 1999 Takeda
5986644 November 16, 1999 Herder et al.
5991085 November 23, 1999 Rallison et al.
5999168 December 7, 1999 Rosenberg et al.
6002394 December 14, 1999 Schein et al.
D419199 January 18, 2000 Cordell et al.
D419200 January 18, 2000 Ashida
6010406 January 4, 2000 Kajikawa et al.
6011526 January 4, 2000 Toyoshima et al.
6012980 January 11, 2000 Yoshida et al.
6013007 January 11, 2000 Root et al.
6016144 January 18, 2000 Blonstein et al.
6019680 February 1, 2000 Cheng
6020876 February 1, 2000 Rosenberg et al.
6037882 March 14, 2000 Levy
6044297 March 28, 2000 Sheldon et al.
6049823 April 11, 2000 Hwang
6052083 April 18, 2000 Wilson
6057788 May 2, 2000 Cummings
6058342 May 2, 2000 Orbach et al.
6059576 May 9, 2000 Brann
6069594 May 30, 2000 Barnes et al.
6072467 June 6, 2000 Walker
6072470 June 6, 2000 Ishigaki
6075575 June 13, 2000 Schein et al.
6081819 June 27, 2000 Ogino
6084315 July 4, 2000 Schmitt
6084577 July 4, 2000 Sato et al.
6087950 July 11, 2000 Capan
D429718 August 22, 2000 Rudolph
6110039 August 29, 2000 Oh
6115028 September 5, 2000 Balakrishnan
6137457 October 24, 2000 Tokuhashi et al.
D433381 November 7, 2000 Talesfore
6146278 November 14, 2000 Kobayashi
6148100 November 14, 2000 Anderson et al.
6155926 December 5, 2000 Miyamoto et al.
6160405 December 12, 2000 Needle et al.
6160540 December 12, 2000 Fishkin et al.
6162191 December 19, 2000 Foxlin
6164808 December 26, 2000 Shibata et al.
6171190 January 9, 2001 Thanasack et al.
6176837 January 23, 2001 Foxlin
6181329 January 30, 2001 Stork et al.
6183365 February 6, 2001 Tonomura et al.
6184862 February 6, 2001 Leiper
6184863 February 6, 2001 Silbert et al.
6191774 February 20, 2001 Schena et al.
6198295 March 6, 2001 Hill
6198470 March 6, 2001 Agam et al.
6198471 March 6, 2001 Cook
6200219 March 13, 2001 Rudell et al.
6200253 March 13, 2001 Nishiumi et al.
6201554 March 13, 2001 Lands
6211861 April 3, 2001 Rosenberg et al.
6217450 April 17, 2001 Meredith
6217478 April 17, 2001 Vohmann et al.
D442998 May 29, 2001 Ashida
6225987 May 1, 2001 Matsuda
6226534 May 1, 2001 Aizawa
6238291 May 29, 2001 Fujimoto et al.
6239806 May 29, 2001 Nishiumi et al.
6241611 June 5, 2001 Takeda et al.
6243658 June 5, 2001 Raby
6244987 June 12, 2001 Oshuga et al.
6245014 June 12, 2001 Brainard, II
6264558 July 24, 2001 Nishiumi et al.
6273819 August 14, 2001 Strauss et al.
6280327 August 28, 2001 Leifer et al.
6287198 September 11, 2001 McCauley
6297751 October 2, 2001 Fadavi-Ardekani
6301534 October 9, 2001 McDermott, Jr. et al.
6304250 October 16, 2001 Yang et al.
6315673 November 13, 2001 Kopera et al.
6323614 November 27, 2001 Palazzolo et al.
6323654 November 27, 2001 Needle et al.
6325718 December 4, 2001 Nishiumi et al.
6331841 December 18, 2001 Tokuhashi et al.
6331856 December 18, 2001 Van Hook et al.
6337954 January 8, 2002 Soshi et al.
6347998 February 19, 2002 Yoshitomi et al.
6361507 March 26, 2002 Foxlin
D456410 April 30, 2002 Ashida
6369794 April 9, 2002 Sakurai et al.
6375572 April 23, 2002 Masuyama et al.
6377793 April 23, 2002 Jenkins
6377906 April 23, 2002 Rowe
D456854 May 7, 2002 Ashida
6383079 May 7, 2002 Takeda et al.
6392613 May 21, 2002 Goto
6394904 May 28, 2002 Stalker
D458972 June 18, 2002 Ashida
6400480 June 4, 2002 Thomas
6400996 June 4, 2002 Hoffberg et al.
6409687 June 25, 2002 Foxlin
D459727 July 2, 2002 Ashida
D460787 July 23, 2002 Nishikawa
6415223 July 2, 2002 Lin et al.
6421056 July 16, 2002 Nishiumi et al.
6424333 July 23, 2002 Tremblay
6426719 July 30, 2002 Nagareda et al.
6426741 July 30, 2002 Goldsmith et al.
D462683 September 10, 2002 Ashida
6452494 September 17, 2002 Harrison
6456276 September 24, 2002 Park
D464053 October 8, 2002 Zicolello
D464950 October 29, 2002 Fraquelli
6466198 October 15, 2002 Feinstein
6466831 October 15, 2002 Shibata et al.
6473070 October 29, 2002 Mishra et al.
6473713 October 29, 2002 McCall et al.
6474159 November 5, 2002 Foxlin et al.
6484080 November 19, 2002 Breed
6492981 December 10, 2002 Stork et al.
6496122 December 17, 2002 Sampsell
6518952 February 11, 2003 Leiper
6530838 March 11, 2003 Ha
6538675 March 25, 2003 Aratani et al.
D473942 April 29, 2003 Motoki et al.
6540607 April 1, 2003 Mokris et al.
6540611 April 1, 2003 Nagata
6544124 April 8, 2003 Ireland et al.
6544126 April 8, 2003 Sawano et al.
6545661 April 8, 2003 Goschy et al.
6554781 April 29, 2003 Carter et al.
D474763 May 20, 2003 Tozaki et al.
6565444 May 20, 2003 Nagata et al.
6567536 May 20, 2003 McNitt et al.
6572108 June 3, 2003 Bristow
6577350 June 10, 2003 Proehl et al.
6582299 June 24, 2003 Matsuyama et al.
6582380 June 24, 2003 Kazlausky et al.
6585596 July 1, 2003 Leifer
6590536 July 8, 2003 Walton
6591677 July 15, 2003 Rothuff
6597342 July 22, 2003 Haruta
6597443 July 22, 2003 Boman
6599194 July 29, 2003 Smith et al.
6605038 August 12, 2003 Teller et al.
6608563 August 19, 2003 Weston et al.
6609977 August 26, 2003 Shimizu et al.
6616607 September 9, 2003 Hashimoto et al.
6628257 September 30, 2003 Oka et al.
6634949 October 21, 2003 Briggs et al.
6636826 October 21, 2003 Abe et al.
6650029 November 18, 2003 Johnston
6650313 November 18, 2003 Levine et al.
6650345 November 18, 2003 Saito et al.
6654001 November 25, 2003 Su
6672962 January 6, 2004 Ozaki et al.
6676520 January 13, 2004 Nishiumi
6677990 January 13, 2004 Kawahara
6681629 January 27, 2004 Foxlin et al.
6682351 January 27, 2004 Abraham-Fuchs et al.
6684062 January 27, 2004 Gosior et al.
D486145 February 3, 2004 Kaminski et al.
6686954 February 3, 2004 Kitaguchi et al.
6692170 February 17, 2004 Abir
6693622 February 17, 2004 Shahoian et al.
6712692 March 30, 2004 Basson et al.
6717573 April 6, 2004 Shahoian et al.
6718280 April 6, 2004 Hermann
6725173 April 20, 2004 An et al.
D489361 May 4, 2004 Mori et al.
6736009 May 18, 2004 Schwabe
D491924 June 22, 2004 Kaminski et al.
D492285 June 29, 2004 Ombao et al.
6743104 June 1, 2004 Ota et al.
6747632 June 8, 2004 Howard
6747690 June 8, 2004 Mølgaard
6749432 June 15, 2004 French et al.
6752719 June 22, 2004 Himoto et al.
6753849 June 22, 2004 Curran et al.
6753888 June 22, 2004 Kamiwada et al.
6757068 June 29, 2004 Foxlin
6757446 June 29, 2004 Li et al.
6761637 July 13, 2004 Weston et al.
6765553 July 20, 2004 Odamura
D495336 August 31, 2004 Andre et al.
6786877 September 7, 2004 Foxlin
6796177 September 28, 2004 Mori
6811489 November 2, 2004 Shimizu et al.
6811491 November 2, 2004 Levenberg et al.
6812881 November 2, 2004 Mullaly et al.
6813525 November 2, 2004 Reid et al.
6813584 November 2, 2004 Zhou et al.
6816151 November 9, 2004 Dellinger
6821204 November 23, 2004 Aonuma et al.
6821206 November 23, 2004 Ishida et al.
6836705 December 28, 2004 Hellmann et al.
6836751 December 28, 2004 Paxton et al.
6836971 January 4, 2005 Wan
6842991 January 18, 2005 Levi et al.
6850221 February 1, 2005 Tickle
6850844 February 1, 2005 Walters et al.
6852032 February 8, 2005 Ishino
6856327 February 15, 2005 Choi
D502468 March 1, 2005 Knight et al.
6868738 March 22, 2005 Moscrip et al.
6872139 March 29, 2005 Sato et al.
6873406 March 29, 2005 Hines et al.
D503750 April 5, 2005 Kit et al.
D504677 May 3, 2005 Kaminski et al.
D505424 May 24, 2005 Ashida et al.
6897845 May 24, 2005 Ozawa
6897854 May 24, 2005 Cho et al.
6906700 June 14, 2005 Armstrong
6908388 June 21, 2005 Shimizu et al.
6922632 July 26, 2005 Foxlin
6925410 August 2, 2005 Narayanan
6929543 August 16, 2005 Ueshima et al.
6929548 August 16, 2005 Wang
6933861 August 23, 2005 Wang
6933923 August 23, 2005 Feinstein
6954980 October 18, 2005 Song
6955606 October 18, 2005 Taho et al.
6956564 October 18, 2005 Williams
6967566 November 22, 2005 Weston et al.
6982697 January 3, 2006 Wilson et al.
6984208 January 10, 2006 Zheng
6990639 January 24, 2006 Wilson
6993451 January 31, 2006 Chang et al.
6995748 February 7, 2006 Gordon et al.
6998966 February 14, 2006 Pederson et al.
7000469 February 21, 2006 Foxlin et al.
7002591 February 21, 2006 Leather et al.
7031875 April 18, 2006 Ellenby et al.
7066781 June 27, 2006 Weston
D524298 July 4, 2006 Hedderich et al.
7081051 July 25, 2006 Himoto et al.
7090582 August 15, 2006 Danieli et al.
7098891 August 29, 2006 Pryor
7098894 August 29, 2006 Yang et al.
7102616 September 5, 2006 Sleator
7107168 September 12, 2006 Oystol et al.
D531228 October 31, 2006 Ashida et al.
7115032 October 3, 2006 Cantu et al.
7126584 October 24, 2006 Nishiumi et al.
7127370 October 24, 2006 Kelly et al.
D531585 November 7, 2006 Weitgasser et al.
7133026 November 7, 2006 Horie et al.
7136674 November 14, 2006 Yoshie et al.
7139983 November 21, 2006 Kelts
7140962 November 28, 2006 Okuda et al.
7142191 November 28, 2006 Idesawa et al.
7149627 December 12, 2006 Ockerse et al.
7154475 December 26, 2006 Crew
7155604 December 26, 2006 Kawai
7158118 January 2, 2007 Liberty
7173604 February 6, 2007 Marvit et al.
7176919 February 13, 2007 Drebin et al.
7182691 February 27, 2007 Schena
7183480 February 27, 2007 Nishitani et al.
7184059 February 27, 2007 Fouladi et al.
D543246 May 22, 2007 Ashida et al.
7220220 May 22, 2007 Stubbs et al.
7225101 May 29, 2007 Usuda et al.
7231063 June 12, 2007 Naimark et al.
7233316 June 19, 2007 Smith et al.
7236156 June 26, 2007 Liberty et al.
7239301 July 3, 2007 Liberty et al.
7261690 August 28, 2007 Teller et al.
7262760 August 28, 2007 Liberty
D556201 November 27, 2007 Ashida et al.
7292151 November 6, 2007 Ferguson et al.
7301527 November 27, 2007 Marvit
7301648 November 27, 2007 Foxlin
D556760 December 4, 2007 Ashida et al.
D559847 January 15, 2008 Ashida et al.
D561178 February 5, 2008 Azuma
7335134 February 26, 2008 LaVelle
D563948 March 11, 2008 d'Hore
D567243 April 22, 2008 Ashida et al.
7359121 April 15, 2008 French et al.
RE40324 May 20, 2008 Crawford
7379078 May 27, 2008 Gossweiler et al.
7379566 May 27, 2008 Hildreth
7395181 July 1, 2008 Foxlin
7414611 August 19, 2008 Liberty
7445550 November 4, 2008 Barney et al.
7488231 February 10, 2009 Weston
7500917 March 10, 2009 Barney et al.
7545956 June 9, 2009 Miyahara
7568289 August 4, 2009 Burlingham et al.
7582016 September 1, 2009 Suzuki
7614958 November 10, 2009 Weston et al.
7663509 February 16, 2010 Shen
7672543 March 2, 2010 Hull et al.
7774155 August 10, 2010 Sato et al.
7775882 August 17, 2010 Kawamura et al.
7783076 August 24, 2010 Tsunashima
7796116 September 14, 2010 Salsman et al.
7831064 November 9, 2010 Ohta
7937878 May 10, 2011 Sammut et al.
20010010514 August 2, 2001 Ishino
20010015123 August 23, 2001 Nishitani et al.
20010024973 September 27, 2001 Meredith
20010031662 October 18, 2001 Larian
20010049302 December 6, 2001 Hagiwara
20020024500 February 28, 2002 Howard
20020024675 February 28, 2002 Foxlin
20020028071 March 7, 2002 Mølgaard
20020072418 June 13, 2002 Masuyama et al.
20020075335 June 20, 2002 Rekimoto
20020098887 July 25, 2002 Himoto et al.
20020103026 August 1, 2002 Himoto et al.
20020107069 August 8, 2002 Ishino
20020126026 September 12, 2002 Lee
20020137567 September 26, 2002 Cheng
20020140666 October 3, 2002 Bradski
20020140745 October 3, 2002 Ellenby et al.
20020158843 October 31, 2002 Levine et al.
20020183961 December 5, 2002 French et al.
20030038778 February 27, 2003 Noguera et al.
20030052860 March 20, 2003 Park et al.
20030057808 March 27, 2003 Lee et al.
20030063068 April 3, 2003 Anton et al.
20030069077 April 10, 2003 Korienek
20030107551 June 12, 2003 Dunker
20030144056 July 31, 2003 Leifer et al.
20030193572 October 16, 2003 Wilson et al.
20030195041 October 16, 2003 McCauley
20030204361 October 30, 2003 Townsend et al.
20030216176 November 20, 2003 Shimizu et al.
20030222851 December 4, 2003 Lai et al.
20040028258 February 12, 2004 Naimark et al.
20040034289 February 19, 2004 Teller et al.
20040046736 March 11, 2004 Pryor et al.
20040048666 March 11, 2004 Bagley
20040070564 April 15, 2004 Dawson
20040075650 April 22, 2004 Paul et al.
20040095317 May 20, 2004 Zhang et al.
20040134341 July 15, 2004 Sandoz et al.
20040140954 July 22, 2004 Faeth
20040143413 July 22, 2004 Oystol et al.
20040147317 July 29, 2004 Ito et al.
20040152515 August 5, 2004 Wegmuller et al.
20040174340 September 9, 2004 Bruneau et al.
20040193413 September 30, 2004 Wilson et al.
20040203638 October 14, 2004 Chan
20040204240 October 14, 2004 Barney
20040218104 November 4, 2004 Smith et al.
20040222969 November 11, 2004 Buchenrieder
20040227725 November 18, 2004 Calarco et al.
20040229693 November 18, 2004 Lind et al.
20040233223 November 25, 2004 Schkolne et al.
20040239626 December 2, 2004 Noguera
20040252109 December 16, 2004 Trent et al.
20040254020 December 16, 2004 Dragusin
20040259651 December 23, 2004 Storek
20040268393 December 30, 2004 Hunleth et al.
20050017454 January 27, 2005 Endo et al.
20050020369 January 27, 2005 Davis et al.
20050021048 January 27, 2005 Kuhlman et al.
20050024360 February 3, 2005 Abe et al.
20050032582 February 10, 2005 Mahajan
20050047621 March 3, 2005 Cranfill
20050052415 March 10, 2005 Braun et al.
20050054457 March 10, 2005 Eyestone et al.
20050070359 March 31, 2005 Rodriquez et al.
20050076161 April 7, 2005 Albanna et al.
20050085298 April 21, 2005 Woolston
20050125826 June 9, 2005 Hunleth et al.
20050130739 June 16, 2005 Argentar
20050134555 June 23, 2005 Liao
20050137774 June 23, 2005 Rupp
20050143173 June 30, 2005 Barney et al.
20050170889 August 4, 2005 Lum et al.
20050172734 August 11, 2005 Alsio
20050174324 August 11, 2005 Liberty et al.
20050176485 August 11, 2005 Ueshima
20050179644 August 18, 2005 Alsio
20050210419 September 22, 2005 Kela
20050212749 September 29, 2005 Marvit
20050212750 September 29, 2005 Marvit
20050212751 September 29, 2005 Marvit
20050212752 September 29, 2005 Marvit
20050212753 September 29, 2005 Marvit
20050212754 September 29, 2005 Marvit
20050212755 September 29, 2005 Marvit
20050212756 September 29, 2005 Marvit
20050212757 September 29, 2005 Marvit
20050212758 September 29, 2005 Marvit
20050212759 September 29, 2005 Marvit
20050212760 September 29, 2005 Marvit
20050212764 September 29, 2005 Toba
20050212767 September 29, 2005 Marvit et al.
20050215295 September 29, 2005 Arneson
20050215322 September 29, 2005 Himoto et al.
20050217525 October 6, 2005 McClure
20050233808 October 20, 2005 Himoto et al.
20050239548 October 27, 2005 Ueshima et al.
20050243061 November 3, 2005 Liberty et al.
20050243062 November 3, 2005 Liberty
20050244034 November 3, 2005 Miyahara
20050253806 November 17, 2005 Liberty et al.
20050256675 November 17, 2005 Kurata
20060028446 February 9, 2006 Liberty et al.
20060030385 February 9, 2006 Barney et al.
20060040738 February 23, 2006 Okazaki et al.
20060046849 March 2, 2006 Kovacs
20060092133 May 4, 2006 Touma et al.
20060094502 May 4, 2006 Katayama et al.
20060122474 June 8, 2006 Teller et al.
20060123146 June 8, 2006 Wu et al.
20060139327 June 29, 2006 Dawson et al.
20060148563 July 6, 2006 Yang
20060152487 July 13, 2006 Grunnet-Jepsen et al.
20060152488 July 13, 2006 Salsman et al.
20060152489 July 13, 2006 Sweetser et al.
20060154726 July 13, 2006 Weston et al.
20060178212 August 10, 2006 Penzias
20060205507 September 14, 2006 Ho
20060231794 October 19, 2006 Sakaguchi et al.
20060252477 November 9, 2006 Zalewski et al.
20060256081 November 16, 2006 Zalewski et al.
20060258452 November 16, 2006 Hsu
20060264258 November 23, 2006 Zalewski et al.
20060264260 November 23, 2006 Zalewski et al.
20060282873 December 14, 2006 Zalewski et al.
20060287086 December 21, 2006 Zalewski et al.
20060287087 December 21, 2006 Zalewski et al.
20070015588 January 18, 2007 Matsumoto et al.
20070021208 January 25, 2007 Mao et al.
20070049374 March 1, 2007 Ikeda et al.
20070050597 March 1, 2007 Ikeda et al.
20070052177 March 8, 2007 Ikeda et al.
20070060228 March 15, 2007 Akasaka et al.
20070060391 March 15, 2007 Ikeda et al.
20070066394 March 22, 2007 Ikeda et al.
20070066396 March 22, 2007 Weston et al.
20070072674 March 29, 2007 Ohta et al.
20070072680 March 29, 2007 Ikeda et al.
20070091084 April 26, 2007 Ueshima et al.
20070093291 April 26, 2007 Hulvey
20070159362 July 12, 2007 Shen
20070173705 July 26, 2007 Teller et al.
20070211027 September 13, 2007 Ohta
20070252815 November 1, 2007 Kuo et al.
20070265075 November 15, 2007 Zalewski
20070265076 November 15, 2007 Lin et al.
20070265088 November 15, 2007 Nakada et al.
20070270220 November 22, 2007 Kaneshige et al.
20080014835 January 17, 2008 Weston et al.
20080015017 January 17, 2008 Ashida et al.
20080039202 February 14, 2008 Sawano et al.
20080121782 May 29, 2008 Hotelling et al.
20080273011 November 6, 2008 Lin
20080278445 November 13, 2008 Sweetser et al.
20080280660 November 13, 2008 Ueshima et al.
20090005166 January 1, 2009 Sato
20090051653 February 26, 2009 Barney et al.
20090124165 May 14, 2009 Weston
20090156309 June 18, 2009 Weston et al.
20110132983 June 9, 2011 Sammut et al.
Foreign Patent Documents
1338961 March 2002 CN
1559644 January 2005 CN
3930581 March 1991 DE
19701344 July 1997 DE
19701374 July 1997 DE
19648487 June 1998 DE
19814254 October 1998 DE
19937307 February 2000 DE
10029173 January 2002 DE
10241392 May 2003 DE
10219198 November 2003 DE
1 524 334 March 1977 EP
0 728 503 August 1996 EP
0 835 676 April 1998 EP
0 848 226 June 1998 EP
0 852 961 July 1998 EP
1 062 994 December 2000 EP
1 239 260 September 2002 EP
1 249 260 October 2002 EP
1 279 425 January 2003 EP
1 293 237 March 2003 EP
0993845 December 2005 EP
1524334 September 1978 GB
2 244 546 May 1990 GB
2244546 May 1990 GB
2284478 June 1995 GB
2307133 May 1997 GB
2316482 February 1998 GB
2319374 May 1998 GB
60-077231 May 1985 JP
62-14527 January 1987 JP
03-08103 August 1991 JP
3-059619 November 1991 JP
04-287888 October 1992 JP
5-056191 July 1993 JP
6-50758 February 1994 JP
6-154422 June 1994 JP
03-000028 July 1994 JP
6-190144 July 1994 JP
6-198075 July 1994 JP
3-273531 November 1994 JP
6-308879 November 1994 JP
3-228845 January 1995 JP
7-28591 January 1995 JP
7-44315 February 1995 JP
7044315 February 1995 JP
7-107573 April 1995 JP
07-22312 May 1995 JP
7-115690 May 1995 JP
3-517482 June 1995 JP
7-146123 June 1995 JP
7-200142 August 1995 JP
07-262797 October 1995 JP
7-302148 November 1995 JP
07-318332 December 1995 JP
8-71252 March 1996 JP
8-071252 March 1996 JP
8-095704 April 1996 JP
8-106352 April 1996 JP
08-111144 April 1996 JP
11-114223 April 1996 JP
8-114415 May 1996 JP
8-122070 May 1996 JP
8-152959 June 1996 JP
8-211993 August 1996 JP
8-305355 November 1996 JP
83-35136 December 1996 JP
9-230997 September 1997 JP
9-274534 October 1997 JP
H09-265346 October 1997 JP
09-319510 December 1997 JP
10-021000 January 1998 JP
10-033831 February 1998 JP
10-99542 April 1998 JP
10-154038 June 1998 JP
10-254614 September 1998 JP
11-099284 April 1999 JP
11-506857 June 1999 JP
11-305935 November 1999 JP
2000-270237 September 2000 JP
2000-308756 November 2000 JP
2001-038052 February 2001 JP
30-78268 April 2001 JP
2001-104643 April 2001 JP
03-080103 June 2001 JP
2001-175412 June 2001 JP
2001-251324 September 2001 JP
2001-306245 November 2001 JP
2001-356875 December 2001 JP
2002-062981 February 2002 JP
2002-082751 March 2002 JP
2002-091692 March 2002 JP
2002-153673 May 2002 JP
2002-202843 July 2002 JP
2002-224444 August 2002 JP
2002-232549 August 2002 JP
2002-233665 August 2002 JP
2002-282541 October 2002 JP
2003-053038 February 2003 JP
34-22383 April 2003 JP
2003-208263 July 2003 JP
2003-280785 October 2003 JP
2003-325974 November 2003 JP
2004-062774 February 2004 JP
2004-313429 November 2004 JP
2004-313492 November 2004 JP
2005-20458 January 2005 JP
2005-040493 February 2005 JP
2005-063230 March 2005 JP
2003-140823 April 2006 JP
2006-113019 April 2006 JP
2002-136694 June 2006 JP
2006-136694 June 2006 JP
2006-216569 April 2007 JP
2007-283134 November 2007 JP
9300171 August 1994 NL
2125853 February 1999 RU
2126161 February 1999 RU
2141738 November 1999 RU
94/02931 February 1994 WO
2004/039055 May 1994 WO
96/05766 February 1996 WO
97/09101 March 1997 WO
97/12337 April 1997 WO
97/17598 May 1997 WO
97/28864 August 1997 WO
97/32641 September 1997 WO
98/11528 March 1998 WO
99/58214 November 1999 WO
00/33168 June 2000 WO
00/35345 June 2000 WO
00/47108 August 2000 WO
00/63874 October 2000 WO
01/87426 November 2001 WO
01/91042 November 2001 WO
WO 01/86920 November 2001 WO
02/17054 February 2002 WO
02/34345 May 2002 WO
03/015005 February 2003 WO
03/107260 June 2003 WO
03/088147 October 2003 WO
2004/012130 February 2004 WO
WO 2004/012130 February 2004 WO
2004/051391 June 2004 WO
Other references
  • Japanese Office Action issued on Apr. 25, 2011 in JP 2005-264207.
  • ADXL202 Specification Sheet: Low Cost ±2 g Dual Axis i MEMs® Accelerometer with Digital Output; Analog Devices, Inc., 1998.
  • ADXL330 Specification Sheet: Small, Low Power, 3-Axis ±3 g i MEMs® Accelerometer; Analog Devices, Inc., 2007.
  • Pictures of Microsoft Xwand retrieved on May 13, 2009 from http://www.kf12.com/blogs/uploads/xwand.jpg and http://www.cs.cmu.edu/%7Edwilson/images/xwand.jpg.
  • Wilson, Andrew D., et al.; “Demonstration of the XWand Interface for Intelligent Spaces”; Microsoft Research; UIST '02 Companion; pp. 37-38.
  • Wilson, Daniel, et al.; “Gesture Recognition Using the XWand”; Robotics Institute; Carnegie Mellon University; tech report CMU-RI-TR-04-57; Apr. 2004.
  • Wilson, Andy, “XWand: UI for Intelligent Environments”; Apr. 26, 2004; retrieved May 12, 2009 from http://research.microsoft.com/en-us/um/people/awilson/wand/default.htm.
  • Wilson, Andrew, et al.; “XWand: UI for Intelligent Spaces”; Microsoft Research; CHI 2003, Apr. 5-10, 2003; Ft. Lauderdale, FL.
  • Selectech Air Mouse, Description; retrieved on May 5, 2009 from http://cgi.ebay.com.my/ws/eBayISAPI.dll?ViewItem&item=350096666675&indexURL.
  • Lin et al., “An Assisted User Interface for 3D Examine Mode”, Proceedings of 2002 15th IPRR Conference on Computer Vision, Graphics, and Image Processing, Taiwan 2002, 7 pages, XP-002531989.
  • Bowman, D. A. et al., “An Evaluation of Techniques for Grabbing and Manipulating Remote Objects in Immersive Virtual Environments”, Computer Graphics USA, vol. 24, No. 2, Mar. 1990, pp. 175-183.
  • Althoff et al., “A Generic Approach for Interfacing VRML Browsers to Various Input Devices and Creating Customizable 3D Applications”, Proc. of Intl. Conference on 3D Web Technology, Feb. 24-28, 2002, pp. 67-74.
  • Ware, C. et al., “Exploration and Virtual Camera Control in Virtual Three Dimensional Environments,” Computer Graphics USA, vol. 24, No. 2, Mar. 1990, pp. 175-183.
  • You S. et al., “Fusion of Vision and Gyro Tracking for Robust Augmented Reality Registration”, Proceedings IEEE Virtual Reality 2001 IEEE Comput. Soc. Los Alamitos, CA, USA, 2001, pp. 71-78.
  • English Abstract for Japanese Patent No. JP10021000, published Jan. 23, 1998.
  • English Abstract for Japanese Patent No. JP11053994, published Feb. 26, 1999.
  • English Abstract for Japanese Patent No. JP11099284, published Apr. 13, 1999.
  • English Abstract for Japanese Patent No. JP2001038052, published Feb. 13, 2001.
  • English Abstract for Japanese Patent No. JP2002224444, published Aug. 13, 2002.
  • English Abstract for Japanese Patent No. JP2006136694, published Jun. 1, 2006.
  • English Abstract for Japanese Patent No. WO9732641, published Sep. 12, 1997.
  • Acar, “Robust Micromachined Vibratory Gyroscopes” Dissertation (Dec. 2004).
  • Acar, et al., “Experimental evaluation and comparative analysis of commercial variable-capacitance MEMS accelerometers,” Journal of Micromechanics and Microengineering, vol. 13 (1), pp. 634-645 (May 2003).
  • Achenbach, “Golf's New Measuring Stick,” Golfweek, Jun. 11, 2005, 1 page.
  • ACT Labs: Miacomet Background, 1 page, May 1999, http://www.act-labs.com/realfeelbackground/htm.
  • AirPad Controller Manual (AirPad Corp. 2000).
  • Airpad Motion Reflext Controller for Sony Playstation—Physical Product (AirPad Corp.2000).
  • Albrecht, “An Adaptive Digital Filter to Predict Pilot Head Look Direction for Helmet-mounted Displays,” MS Thesis University of Dayton (1989).
  • Algrain, “Estimation of 3-D Angular Motion Using Gyroscopes and Linear Accelerometers,” IEEE Transactions on Aerospace and Electronic Systems, vol. 27, No. 6, pp. 910-920 (Nov. 1991).
  • Algrain, et al., “Accelerometer Based Line-of-Sight Stabilization Approach for Pointing and Tracking System,” Second IEEE Conference on Control Applications, vol. 1, Issue 13-16 pp. 159-163 (Sep. 1993).
  • Algrain, et al., “Interlaced Kalman Filtering of 3-D Angular Motion Based on Euler's Nonlinear Equations,” IEEE Transactions on Aerospace and Electronic Systems, vol. 30, No. 1 (Jan. 1994).
  • Allen, et al., “A General Method for Comparing the Expected Performance of Tracking and Motion Capture Systems,” {VRST} '05: Proceedings of the ACM symposium on Virtual reality software and technology, pp. 201-210 (Nov. 2005).
  • Allen, et al., “Tracking: Beyond 15 Minutes of Thought,” SIGGRAPH 2001 Course 11 (Course Pack) from Computer Graphics (2001).
  • Alves, “Extended Kalman filtering applied to a full accelerometer strapdown inertial measurement unit,” M.S. Thesis Massachusetts Institute of Technology. Dept. of Aeronautics and Astronautics, Santiago (1992).
  • Analog Devices Data Sheet, “MicroConverter®, Multichannel 12-Bit ADC with Embedded Flash MCU, ADuC812” (2003) (http://www.analog.com/static/imported-files/datasheets/ADUC812.pdf) 60 pages.
  • Analog Devices “ADXL50 Single Axis Accelerometer” (Data Sheet), http://www.analog.com/en/obsolete/adx150/products/product.html (Mar. 1996).
  • Analog Devices “ADXL50 Monolithic Accelerometer with Signal Conditioning” Datasheet (1996).
  • Analog Devices “ADXRS150 ±150°/s Single Chip Yaw Rate Gyro with Signal Conditioning” (Data Sheet), Rev. B (2004).
  • Analog Devices “ADXRS401 ±75°/s Single Chip Yaw Rate Gyro with Signal Conditioning” (Data Sheet), Rev. O (2004).
  • Ang, et al., “Design and Implementation of Active Error Canceling in Hand-held Microsurgical Instrument,” Proceedings of the 2001 IEEE/RSJ International Conference on Intelligent Robots and Systems, vol. 2, (Oct. 2001).
  • Ang, et al., “Design of All-Accelerometer Inertial Measurement Unit for Tremor Sensing in Hand-held Microsurgical Instrument,” Proceedings of the 2003 IEEE International Conference on Robotics & Automation (Sep. 2003).
  • Apostolyuk, Vladislav, “Theory and design of micromechanical vibratory gyroscopes,” MEMS/NEMS Handbook, Springer, 2006, vol. 1, pp. 173-195 (2006).
  • Arcanatech, “IMP User's Guide” (1994).
  • Arcanatech, IMP (Photos) (1994).
  • Ascension Technology, The Bird 6D Input Devices (specification) (1998).
  • “ASCII Grip One Handed Controller,” One Switch-ASCII Grip One Handed Playstation Controller, http://www.oneswitch.org.uk/1/ascii/grip.htm, Jul. 11, 2008, pp. 1-2.
  • “ASCII Grip” One-Handed Controller The Ultimate One-Handed Controller Designed for the Playstation Game Console (ASCII Entertainment 1997).
  • “ASCII/Sammy Grip V2,” One Switch-Accessible Gaming Shop—ASCII Grip V2, http://www.oneswitch.org.uk/1/AGS/AGS-onehand/ascii-grip-v2.html, Jul. 10, 2008, pp. 1-2.
  • ASCII, picture one-handed controller, 2 pages (Feb. 6, 2006).
  • Ashida et al., entitled, “Game Controller,” U.S. Appl. No. 11/790,780, filed Apr. 27, 2007, pending.
  • “At-home fishing” 1 page, Dec. 1996-1999.
  • Ator, “Image-Velocity with Parallel-Slit Reticles,” Journal of the Optical Society of America (Dec. 1963).
  • Azarbayejani, et al, “Real-Time 3-D Tracking of the Human Body,” Proceedings of Image'Com 96 (1996).
  • Azarbayejani, et al., “Visually Controlled Graphics,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 15, No. 6, pp. 602-605 (Jun. 1993).
  • Azuma et al., “Improving Static and Dynamic Registration in an Optical See-Through HMD,” International Conference on Computer Graphics and Interactive Techniques Proceedings of the 21st annual conference on Computer graphics and interactive techniques, pp. 197-204 (1994).
  • Azuma et al., “Making Augmented Reality Work Outdoors Requires Hybrid Tracking, ”Proceedings of the International Workshop on Augmented Reality, San Francisco, CA, Nov. 1, 1998, Bellevue, Washington, pp. 219-224 (1999).
  • Azuma, “Predictive Tracking for Augmented Reality,” Ph.D. Dissertation, University of North Carolina at Chapel Hill (1995).
  • Azuma, et al., “A Frequency-Domain Analysis of Head-Motion Prediction,” Proceedings of SIGGRAPH '94, pp. 401-408 (1995).
  • Azuma, et al., “A motion-stabilized outdoor augmented reality system,” Proceedings of IEEE Virtual Reality '99, Houston, TX (Mar. 1999).
  • Bachmann et al., “Inertial and Magnetic Posture Tracking for Inserting Humans into Networked Virtual Environments,” Virtual Reality Software and Technology archive, Proceedings of the ACM Symposium on Virtual Reality Software and Technology, Baniff, Alberta, Canada, pp. 9-16 (2001).
  • Bachmann et al., “Orientation Tracking for Humans and Robots Using Inertial Sensors” (CIRA '99), Naval Postgraduate School, Monterey, CA (1999).
  • Bachmann, “Inertial and Magnetic Angle Tracking of Limb Segments for Inserting Humans into Synthetic Environments,” Dissertation, Naval Postgraduate School, Monterey, CA (Dec. 2000).
  • Baker et al., “Active Multimodal Control of a Floppy Telescope Structure,” Proc. SPIE, vol. 4825, 74 (Mar. 2003).
  • Balakrishnan, “The Rockin' Mouse: Integral 3D Manipulation on a Plane,” (CHI '97), Univ. Toronto, (1997).
  • Ballagas, et al., Jan, “iStuff: A Physical User Interface Toolkit for Ubiquitous Computer Environments,” Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, vol. 5, No. 1, at 537-44 (ACM) (Apr. 5-10, 2003).
  • Baraff, “An Introduction to Physically Based Modeling,” SIGGRAPH 97 Course Notes (1997).
  • Bass Fishing “Legends of the Lake”, Radica 2 pages, 2002.
  • Baudisch, et al., “Soap: a pointing device that works in mid-air” Proc. UIST (2006).
  • BBN Report, “Virtual Environment Technology for Training (VETT),” The Virtual Environment and Teleoperator Research Consortium (VETREC) (Mar. 1992).
  • Behringer, “Improving Registration Precision Through Visual Horizon Silhouette Matching,” Proceedings of the international workshop on Augmented reality : placing artificial objects in real scenes: placing artificial objects in real scenes, Bellevue, Washington, United States pp. 225-232 (1999).
  • Behringer, “Registration for Outdoor Augmented Reality Applications Using Computer Vision Techniques and Hybrid Sensors,” Virtual Reality, 1999 Proceedings., IEEE Computer Society, pp. 244-261 (1999).
  • Bei Systron Donner Inertial Division, Gyrochip Theory of Operation (2001).
  • BEI, “BEI GyrochipTM Model QRS11 Data Sheet,” BEI Systron Donner Inertial Division, BEI Technologies, Inc., (Sep. 1998).
  • Benbasat, “An Inertial Measurement Unit for User Interfaces,” Massachusetts Institute of Technology Dissertation, (Sep. 2000).
  • Benbasat, et al., “An Inertial Measurement Framework for Gesture Recognition and Applications,” Gesture and Sign Language in Human-Computer Interaction, International Gesture Workshop, GW 2001, London, UK, 2001 Proceedings, LNAI 2298, at 9-20, I. Wachsmuth and T. Sowa (eds.), Springer-Verlag Berlin Heibelberg (2001, 2002).
  • Beuter, A., Publications, University of Quebec at Montreal, http://www.er.uqam.ca/nobel/r11040/publicat.htm (Aug. 2007).
  • BGM-109 Tomahawk, http://en.wikipedia.org/wiki/BGM-109Tomahawk, Wikipedia, Jan. 2009.
  • Bhatnagar, “Position trackers for Head Mounted Display systems: a survey” (Technical Report), University of North Carolina at Chapel Hill (Mar. 1993).
  • Bianchi, “A Tailless Mouse, New cordless Computer Mouse Invented by ArcanaTech.” Inc. Article (Jun. 1992).
  • Bishop, “The Self-Tracker: a Smart Optical Sensor on Silicon,” Ph.D. Dissertation, Univ. of North Carolina at Chapel Hill (1984).
  • Bishop, et al., “Grids Progress Meeting” (Slides), University of North Carolina at Chapel Hill, NC (1998).
  • Bishop, et al., Self-Tracker: Tracking for Hybrid Environments without Infrastructure (1996).
  • Bloomberg: Nintendo Announces Wireless GBA Link, Sep. 2003, 2 pages.
  • Bona, et al., “Optimum Reset of Ship's Inertial Navigation System,” IEEE Transactions on Aerospace and Electronic Systems (1965).
  • Borenstein, et al., “Where am I? Sensors and Methods for Mobile Robot Positioning” (1996).
  • Boser, “3-Axis Accelerometer with Differential Sense Electronics,” http://www.eecs.berkeley.edu/˜boser/pdf/3axis.pdf (1997).
  • Boser, “Accelerometer Design Example: Analog Devices XL-05/5,” http://www.eecs.berkeley.edu/˜boser/pdf/x105.pdf (1996).
  • Bowman et al., 3D User Interfaces: Theory and Practice, Addison-Wesley, Inc., (2005).
  • Bowman et al., “An Introduction to 3-D User Interface Design,” MIT Presence, vol. 10, No. 1, pp. 96-108 (2001).
  • Briefs (New & Improved) (Brief Article), PC Magazine, Oct. 26, 1993.
  • Britton et al., “Making Nested rotations Convenient for the User,” ACM SIGGRAPH Computer Graphics, vol. 12, Issue 3, pp. 222-227 (Aug. 1978).
  • Britton, “A Methodology for the Ergonomic Design of Interactive Computer Graphic Systems, and its Application to Crystallography” (UNC Thesis) (1977).
  • Brownell, Richard: Review of Peripheral-GameCube-G3 Wireless Controller, GAF, Jul. 17, 2003, 2 pages.
  • Buchanan, Levi: “Happy Birthday, Rumble Pak,” IGN.com, Apr. 3, 2008, 2 pages.
  • Business Wire, “Feature/Virtual reality glasses that interface to Sega channel,” Time Warner, TCI: project announced concourrent with COMDEX (Nov. 1994).
  • Business Wire, “Free-space ‘Tilt’ Game Controller for Sony Playstation Uses Scenix Chip; SX Series IC Processes Spatial Data In Real Time For On-Screen” (Dec. 1999).
  • Business Wire, “InterSense Inc. Launches InertiaCube2—The World's Smallest Precision Orientation Sensor With Serial Interface” (Aug. 14, 2001).
  • Business Wire, “Logitech MAGELLAN 3D Controller,” Logitech (Apr. 1997).
  • Business Wire, “Mind Path Introduces GYROPOINT RF Wireless Remote” (Jan. 2000).
  • Business Wire, “Pegasus' Wireless PenCell Writes on Thin Air with ART's Handwriting Recognition Solutions,” Business Editors/High Tech Writers Telecom Israel 2000 Hall 29, Booth 19-20 (Nov. 2000).
  • Business Wire, “RPI ships low-cost pro HMD Plus 3D Mouse and VR PC graphics card system for CES” (Jan. 1995).
  • Buxton, Bill, “Human input/output devices,” In M. Katz (ed.), Technology Forecast: 1995, Menlo Park, C.A.: Price Waterhouse World Firm Technology Center, 49-65 (1994).
  • Buxton, Bill, A Directory of Sources for Input Technologies, http://www.billbuxton.com/InputSources.html, Apr. 2001 (last update 2008).
  • Buxton et al., “A Study in Two-Handed Input,” ACM CHI '86 Proceedings (1986).
  • Byte, “Imp Coexists With Your Mouse,” What's New, ArcanaTec (Jan. 1994).
  • Canaday, R67-26 “The Lincoln Wand,” IEEE Transactions on Electronic Computers, vol. EC-16, No. 2, p. 240 (Apr. 1967).
  • Caruso et al., “New Perspective on Magnetic Field Sensing,” Sensors Magazine (Dec. 1998).
  • Caruso et al., “Vehicle Detection and Compass Applications using AMR Magnetic Sensors,” Honeywell (May 1999).
  • Caruso, “Application of Magnetoresistive Sensors in Navigation Systems,” Sensors and Actuators, SAE SP-1220, pp. 15-21 (Feb. 1997 ).
  • Caruso, “Applications of Magnetic Sensors for Low Cost Compass Systems,” Honeywell, SSEC, http://www.ssec.honeywell.com/magnetic/datasheets/lowcost.pdf (May 1999 ).
  • Chatfield, “Fundamentals of High Accuracy Inertial Navigation,” vol. 174 Progress in Astronautics and Aeronautics, American Institute of Aeronautics and Astronautics, Inc. (1997).
  • Cheng, “Direct interaction with large-scale display systems using infrared laser tracking devices,” ACM International Conference Proceeding Series; vol. 142 (2003).
  • Cho, et al., “Magic Wand: A Hand-Drawn Gesture Input Device in 3-D Space with Inertial Sensors,” Proceedings of the 9th Intl Workshop on Frontiers in Handwriting Recognition (IWFHR-9 2004), IEEE (2004).
  • CNET News.com, http://news.com.com/2300-10433-6070295-2.html?tag=ne.gall.pg, “Nintendo Wii Swings Into Action,” May 25 2006, 1pg.
  • “Coleco Vision: Super Action™ Controller Set,” www.vintagecomputing.com/wp-content/images/retroscan/colecosac1large.jgp. (Sep. 2006).
  • Computer Mouse (Wikipedia) (Jul. 5, 2005).
  • “Controllers—Atari Space Age Joystick,” AtariAge: Have You Played Atari Today? www.atariage.com/controllerpage.html?SystemID=2600&ControllerID=12. (Sep. 2006).
  • “Controllers—Booster Grip,” AtariAge: Have You Played Atari Today? www.atariage.com/controllerpage.html?SystemID=2600&ControllerID=18. (Sep. 2006).
  • Computergram, “RPI Entertainment Pods Improve Virtual Experience” (1995).
  • Cookbook, Numerical Recipes Electronic Edition, http://www.library.cornell.edu/nr/cbookcpdf.html (2000?).
  • Cooke, et al., “NPSNET: flight simulation dynamic modeling using quaternions,” Presence, vol. 1, No. 4,pp. 404-420, MIT Press (1992/1994).
  • Crossan, A. et al.: A General Purpose Control-Based Trajectory Playback for Force-Feedback Systems, niversity of Glasgow, Dept. Computing Science, 4 pages (Feb. 2008).
  • CSIDC Winners—Tablet-PC Classroom System Wins Design Competition, IEEE Computer Society Press, vol. 36 , Issue 8, pp. 15-18 , IEEE Computer Society (Aug. 2003).
  • Cutrone, “Hot products: Gyration GyroPoint Desk, GyroPoint Pro gyroscope-controlled wired and wireless mice” (Computer Reseller News) (Dec. 1995).
  • Cutts, “A Hybrid Image/Inertial System for Wide-Area Tracking” (Internal to UNC-CH Computer Science) (Jun. 1999 ).
  • Cyberglove/Cyberforce, Immersion, Cyberforce CyberGlove Systems “Immersion Ships New Wireless CyberGlove(R) II Hand Motion-Capture Glove; Animators, Designers, and Researchers Gain Enhanced Efficiency and Realism for Animation, Digital Prototyping and Virtual Reality Projects,” Business Wire, Dec. 7, 2005.
  • Deruyck, et al., “An Electromagnetic Position Sensor,” Polhemus Navigation Sciences, Inc., Burlington, VT (Nov. 1973.)
  • Dichtburn, “Camera in Direct3D” Toymaker, Mar. 5, 2005, 5 pages, http://web.archive.org/web/20050206032104/http://toymaker.info/games/html/camera.html.
  • Donelson, et al., “Spatial Management of Information” (1978 ).
  • Eiβele, “Orientation as an additional User Interface in Mixed-Reality Environments,” 1. workshop Ervwiterte und Virtuelle Realität, pp. 79-90. GI-Fachgruppe AR/VR (2007).
  • Electro-Plankton Weblog, http://www.tranism.com/weblog/2005/09/, “This is the Revolution, Nintendo Style,” Sep. 15, 2005, 2 pgs.
  • “Electronic Plastic: BANDAI—Power Fishing”, “Power Fishing Company: BANDAI”, 1984, 1 page, http:/www.handhelden.com/Bandai/PowerFishing.html.
  • Emura, et al., “Sensor Fusion Based Measurement of Human Head Motion,” 3rd IEEE International Workshop on Robot and Human Communication (Jul. 1994).
  • Ewalt, David M., “Nintendo's Wii Is a Revolution,” Review, Forbes.com (Nov. 13, 2006).
  • Fielder, Lauren: “E3 2001: Nintendo unleashes GameCube software, a new Miyamoto game, and more,” GameSpot, May 16, 2001, 2 pages, http://www.gamespot.com/downloads/2761390.
  • Ferrin, “Survey of Helmet Tracking Technologies,” Proc. SPIE vol. 1456, p. 86-94 (Apr. 1991).
  • Fishing Games: The Evolution of Virtual Fishing Games and related Video Games/Computer Games , 15 pages, 2003.
  • Foley et al., “Computer Graphics: Principles and Practice,” Second Edition, 1990.
  • Foremski, T. “Remote Control Mouse Aims at Interactive TV”, Electronics Weekly, Mar. 9, 1994.
  • Foxlin et al., “An Inertial Head-Orientation Tracker with Automatic Drift Compensation for Use with HMD's,” Proceedings of the conference on Virtual reality software and technology, Singapore, Singapore, pp. 159-173 (1994).
  • Foxlin et al., “Miniature 6-DOF Inertial System for Tracking HMDs,” SPIE vol. 3362 (Apr. 1998).
  • Foxlin et al., “Miniaturization, Calibration & Accuracy Evaluation of a Hybrid Self-Tracker,” The Second IEEE and ACM International Symposium on Mixed and Augmented Reality, pp. 151-160 (2003).
  • Foxlin et al., “WearTrack: A Self-Referenced Head and Hand Tracker for Wearable Computers and Portable VR,” International Symposium on Wearable Computers (ISWC 2000), Oct. 16-18, 2000, Atlanta, GA (2000).
  • Foxlin, “FlightTracker: A Novel Optical/Inertial Tracker for Cockpit Enhanced Vision, Symposium on Mixed and Augmented Reality,” Proceedings of the 3rd IEEE/ACM International Symposium on Mixed and Augmented Reality, pp. 212-221 (Nov. 2004).
  • Foxlin, “Generalized architecture for simultaneous localization, auto-calibration, and map-building,” IEEE/RSJ Conf. on Intelligent Robots and Systems, Lausanne, Switzerland (Oct. 2002).
  • Foxlin, “Head-tracking Relative to a Moving Vehicle or Simulator Platform Using Differential Inertial Sensors,” InterSense, Inc., Presented: Helmet and Head-Mounted Displays V, SPIE vol. 4021, AeroSense Symposium, Orlando, FL, Apr. 24-25, 2000.
  • Foxlin, “Inertial Head Tracker Sensor Fusion by a Complementary Separate-bias Kalman Filter,” Proceedings of the IEEE 1996 Virtual Reality Annual International Symposium, pp. 185-194, 267 (1996).
  • Foxlin, “Inertial Head-Tracking,” MS Thesis, Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science (Sep. 1993).
  • Foxlin, “Motion Tracking Requirements and Technologies,” Chapter 7, from Handbook of Virtual Environment Technology, Stanney Kay, Ed. (2002).
  • Foxlin, “Pedestrian Tracking with Shoe-Mounted Inertial Sensors,” IEEE Computer Graphics and Applications, vol. 25, No. 6, pp. 38-46 (Nov. 2005).
  • Foxlin, et al., “Constellation: A Wide-Range Wireless Motion-Tracking System for Augmented Reality and Virtual Set Applications,” ACM SIGGRAPH, pp. 372-378 (1998).
  • Foxlin, et al., “VIS-Tracker: A Wearable Vision-Inertial Self-Tracker,” IEEE Computer Society (2003).
  • Frankie, “E3 2002: Roll O Rama”, IGN: Roll-o-Rama Preview, . 3 pages. E3 Demo of Kirby game (“Roll O Rama”), http://cube.ign.com/objects/482/482164.html, (May 23, 2002).
  • Freiburg Center for Data Analysis and Modeling—Publications, http://www.fdm.uni-freiburg.de/cms/puplications/publications/ (Aug. 2007).
  • Friedmann, et al., “Device Synchronization Using an Optimal Linear Filter,” SI3D '92: Proceedings of the 1992 symposium on Interactive 3D graphics, pp. 57-62 (1992).
  • Friedmann, et al., “Synchronization in virtual realities,” MIT Presence, vol. 1, No. 1, pp. 139-144 (1992).
  • Fröhlich, “The Yo Yo: An interaction device combining elastic and isotonic control,” at http://www.uni-weimar.de/cms/medien/vr/research/hci/3d-handheld-interaction/the-yoyo-a-handheld-device-combining-elastic-and-isotonic-input.html (2003).
  • FrontSide Field Test, “Get This!”, Golf Magazine, Jun. 2005, p. 36.
  • Fuchs, “Intertial Head-Tracking,” Massachusetts Institute of Technology, Sep. 1993.
  • Furniss, Maureen, “Motion Capture,” MoCap MIT (Dec. 1999) 12 pages.
  • “Game Controller” Wikipedia, Aug. 2010, 8 pages, http://en.wikipedia.org/w/index.php?title=Gamecontroller&oldid=21390758.
  • “Game Controller” Wikipedia, Jan. 5, 2005.
  • Gamecubicle, Jim-New Contributor, Nintendo WaveBird Control, http://www.gamecubicle.com/news-nintendogamecubewavebirdcontroller.htm, May 14, 2002.
  • Geen et al.: “MEMS Angular Rate-Sensing Gyroscope” pp. 1-3 (2003).
  • Gelmis, J.: “Ready to Play, The Future Way”, Jul. 23, 1996, Buffalo News.
  • “Get Bass”, Videogame by Sega, The International Arcade Museum and the KLOV, 1998, 4 pages.
  • “Glove-based input interfaces”, Cyberglove/Cyberforce, Jun. 1991, 12 pages, http://www.angelfire.com/ca7/mellott124/glove1.htm.
  • Goschy, “Midway Velocity Controller” (youtube video http://www.youtube.com/watch?v=wjLhSrSxFNw) (Sep. 8, 2007).
  • Grewal et al., “Global Positioning Systems, Inertial Navigation and Integration,” 2001.
  • Grimm et al., “Real-Time Hybrid Pose Estimation from Vision and Inertial Data,” Proceedings, First Canadian Conference on Computer and Robot Vision, pp. 480-486 (2004).
  • Gyration, Inc., GyroRemote and Mobile RF Keyboard User Manual, Saratoga, CA 24 pages, www.theater.stevejenkins.com/docs/GyrationKeyboardManual (Mar. 9, 2011).
  • Gyration, Inc., GyroRemote GP240-01 Professional Series, copyrighted 2003, www.gyration.com.
  • Gyration Inc., “The Magic Inside GyroPoint”.
  • Gyration Ultra Cordless Optical Mouse, Setting Up Ultra Mouse, Gyration Quick Start Card part No. DL00071-0001 Rev. A. Gyration, Inc. (Jun. 2003).
  • Gyration Ultra Cordless Optical Mouse, User Manual, 1-15, Gyration, Inc., Saratoga, CA (2003).
  • Gyration, “Gyration GP110 Ultra Cordless Optical Mouse Data Sheet,” http://www.gyration.com/descriptions/document/GP110-SPEC-EN.pdf (2002).
  • Gyration, “Gyration GP110 Ultra Cordless Optical Mouse User Manual,” http://www.gyration.com/descriptions/document/GP110-MANUAL-EN.pdf (2002).
  • Gyration, “Gyration MicroGyro 100 Developer Kit Data Sheet,” http://web.archive.org/web/19980708122611/www.gyration.com/html/devkit.html (Jul. 1998).
  • Gyration, “Gyration Ultra Cordless Optical Mouse,” photos (2002).
  • Hamilton Institute, http://www.dcs.gla.ac.uk/.about.rod/, R. Murray-Smith (Aug. 2007).
  • Harada, et al., “Portable Absolute Orientation Estimation Device with Wireless Network under Accelerated Situation” Proceedings, 2004 IEEE International Conference on Robotics and Automation, vol. 2, Issue , Apr. 26-May 1, 2004 pp. 1412-1417 vol. 2 (Apr. 2004).
  • Harada, et al., “Portable orientation estimation device based on accelerometers, magnetometers and gyroscope sensors for sensor network,” Proceedings of IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems, MFI2003, pp. 191-196 (Jul. 2003).
  • Hartley, Matt, “Why is the Nintendo Wii So Successful?”, Smarthouse—The Lifestyle Technology Guide Website (Sep. 12, 2007).
  • Haykin, et al., “Adaptive Tracking of Linear Time-Variant Systems by Extended RLS Algorithms, IEEE Transactions on Signal Processing,” vol. 45, No. 5 (May 1997).
  • Heath, “Virtual Reality Resource Guide AI Expert,” v9 n5 p. 32(14) (May 1994).
  • HiBall-3100—“Wide-Area, High-Precision Tracker and 3D Digitizer,” www.3rdtech.com/HiBall.htm 4 pages.
  • Hinckley, Ken, “Haptic Issues for Virtual Manipulation,” Thesis (Dec. 1996).
  • Hinckley, “Synchronous Gestures for Multiple Persons and Computers”, CHI Letters vol. 5 No. 2 (ACM 2003) & Proceedings of the 16th Annual ACM UIST 2003 Symposium on User Interface Software & Technology, at 149-58 (UIST '03 Vancouver BC Canada) (ACM) (Nov. 2003).
  • Hinckley, et al., “Sensing Techniques for Mobile Interaction,” Proceedings of the 13th Annual ACM Symposium on User Interface Software and Technology (San Diego, Cal.), ACM UIST 2000 & Technology, CHI Letters 2 (2), at 91-100 (ACM) (2000).
  • Hinckley, Ken, et al., “The VideoMouse: A Camera-Based Multi-Degree-of-Freedom Input Device,” CHI Letters vol. 1, 1, UIST '99, Asheville, NC, pp. 103-112 (1999).
  • Hinckley. et al. , “A Survey of Design Issues in Spatial Input,” Proceedings of the ACM Symposium on User Interface Software and Technology (1994).
  • Hinkley et al. Stitching: pen gestures that span multiple displays, 2004.
  • Hinkley et al.: Synchronomous gestures for multiple persons and computers, 2003.
  • Hogue, “MARVIN: A Mobile Automatic Realtime Visual and INertial tracking system,” Master's Thesis, York University (2003).
  • Hogue, et al., “An optical-inertial tracking system for fully-enclosed VR displays,” Proceedings of the 1st Canadian Conference on Computer and Robot Vision, pp. 22-29 (May 2004 ).
  • Holden, Maureen K., et al.: Use of Virtual Environments in Motor Learning and Rehabilitation Department of Brain and Cognitive Sciences, Handbook of Virtual Environments: Design, Implementation, and Applications, Chap. 49, pp. 999-1026, Stanney (ed), Lawrence Erlbaum Associates 2002.
  • Hollands, Robin, “Sourceless Trackers,” VR News (Apr. 1995).
  • Holloway, Richard Lee, “Registration Errors in Augmented Reality Systems,” Ph.D. Dissertation, University of North Carolina at Chapel Hill (1995).
  • House, Matthew, Product Description: Hot Wheels Stunt Track Driver, Hot Wheels (Jan. 2000).
  • Hudson Soft, “Brochure of Toukon Road Brave Warrior, Brave Spirits” (1998).
  • Hudson Soft—Screen Shot of Brave Spirits (1998 ).
  • Immersion CyberGlove product, Immersion Corporation, 1990, http://www.cyberglovesystem.com.
  • Inman, “Cheap sensors could capture your every move,” http://technology.newscientist.com/article/dn12963-cheap-sensors-could-capture-your-every-move.html (Nov. 2007 ).
  • InterSense, “InterSense InertiaCube2 Devices,” (Specification) (image) (2001).
  • InterSense, “InterSense InertiaCube2 Manual for Serial Port Model” (2001).
  • InterSense, InterSense IS 900 Technical Overview—Motion Tracking System, 1999.
  • InterSense, “InterSense IS-1200 FlightTracker Prototype Demonstration” (Video) (Nov. 2004).
  • InterSense, “InterSense IS-1200 InertiaHawk Datasheet” (2009).
  • InterSense, “InterSense IS-1200 VisTracker Datasheet” (2007).
  • InterSense, “InterSense IS-1200 VisTracker Devices,” (image) (2007).
  • InterSense, “InterSense IS-900 MicroTraxTM Datasheet” (2007).
  • InterSense, “InterSense IS-900 Systems Datasheet” (2007).
  • InterSense, “InterSense MicroTrax Demo Reel,” http://www.youtube.com/watch?v=O2F4fuCISo (2007).
  • InterSense, “IS-900 Precision Motion Trackers” www.isense.com May 16, 2003.
  • InterSense, “InterSense Motion Trackers” www.isense.com Mar. 12, 1998.
  • InterSense, “InterSence Inc., The New Standard in Motion Tracking” www.isense.com Mar. 27, 2004.
  • InterSense, “IS-900 Precision Motion Trackers” www.isense.com Sep. 10, 2002.
  • InterSense, “InterSense Mobile Mixed Reality Demonstration” (Video), http://www.youtube.com/watch?v=daVdzGK0nUE&feature=channelpage (Oct. 2006).
  • InterSense, “InterSense Motion Gaming Tech Demo,” http://www.youtube.com/watch?v=7-3y5tdju4E, InterSense (Mar. 2008) (Video).
  • InterSense, “IS-1200 VisTracker Augmented Maintenance Demonstration” (Video), http://www.intersense.com/IS-1200Systems.aspx, http://www.youtube.com/watch?v=1MI78s91WQo&feature=channelpage (Jan. 2009).
  • InterSense, “IS-1200 VisTracker Industrial Cart Demonstration” (Video), InterSense http://www.intersense.com/IS-1200Systems.aspx http://www.youtube.com/watch?v=7xKLCvDGMgY&feature=channelpage (Jan. 2008).
  • InterSense, “IS-900 Product Technology Brief,” http://www.intersense.com/uploadedFiles/Products/WhitePapers/IS900TechOverviewEnhanced.pdf (1999).
  • InterSense, Inc., “Comparison of InterSense IS-900 System and Optical Systems,” http://www.intersense.com/uploadedFiles/Products/WhitePapers/Comparison%20of%20InterSense%20IS-900%20System%20and%20optical%20Systems.pdf (Jul. 12, 2004).
  • Interview with Pat Goschy (youtube video http://www.youtube.com/watch?v=oKtZysYGDLE) (Jan. 14, 2008).
  • Izumori et al, High School Algebra: Geometry (1986).
  • Jacob, “Human-Computer Interaction—Input Devices” http://www.cs.tufts.edu/˜jacob/papers/surveys.html, “Human-Computer Interaction: Input Devices,” ACM Computing Surveys, vol. 28, No. 1, pp. 177-179 (Mar. 1996).
  • Jakubowsk, et al., “Increasing Effectiveness of Human Hand Tremor Separation Process by Using Higher-Order Statistics,” Measurement Science Review, vol. 1 (2001).
  • Jakubowski, et al., “Higher Order Statistics and Neural Network for Tremor Recognition,” IEEE Transactions on Biomedical Engineering, vol. 49, No. 2 (Feb. 2002).
  • Ji, H.: “Study on the Infrared Remote-Control Lamp-Gesture Device”, Yingyong Jiguang/Applied Laser Technology, v. 17, n. 5, p. 225-227, Oct. 1997 Language: Chinese—Abstract only.
  • Jian, et al., “Adaptive Noise Cancellation,” Rice University, http://www.ece.rice.edu/.about.klwang/elec434/elec434.htm, (Aug. 2007).
  • Jiang, “Capacitive position-sensing interface for micromachined inertial sensors,” Dissertation at Univ. of Cal. Berkley (2003).
  • Ju, et al., “The Challenges of Designing a User Interface for Consumer Interactive Television Consumer Electronics Digest of Technical Papers.,” IEEE 1994 International Conference on Volume , Issue , Jun. 21-23, 1994 pp. 114-115.
  • Kalawsky, “The Science of Virtual Reality and Virtual Environments,” 1993.
  • Keir, et al., “Gesture-recognition with Non-referenced Tracking,” IEEE Symposium on 3D User Interfaces, pp. 151-158 (Mar. 25-26, 2006).
  • Kennedy, P.J., “Hand-Held Data Input Device,” IBM Technical Disclosure Bulletin, vol. 26, No. 11, pp. 5826-5827 (Apr. 1984).
  • Kessler, et al., “The Simple Virtual Environment Library” (MIT Presence) (2000).
  • Kindratenko, “A Comparison of the Accuracy of an Electromagnetic and a Hybrid Ultrasound—Inertia Position Tracking System,” MIT Presence, vol. 10, No. 6, Dec. 2001, 657-663.
  • Klein et al.,“Tightly Integrated Sensor Fusion for Robust Visual Tracking,” British Machine Vision Computing, vol. 22, No. 10, pp. 769-776 (2004).
  • Kohler, “Triumph of the Wii: How Fun Won Out in the Console Wars,” www.wired.com/print.gaming/hardware/news/2007/06/wii. (Jun. 2007).
  • Kohlhase, “NASA Report, The Voyager Neptune travel guide,” Jet Propulsion Laboratory Publication 89-24, excerpt (Jun. 1989).
  • Krumm, et al.,“How a Smart Environment Can Use Perception,” Ubicomp 2001 (Sep. 2001).
  • Kuipers, Jack B., “SPASYN—An Electromagnetic Relative Position and Orientation Tracking System,” IEEE Transactions on Instrumentation and Measurement, vol. 29, No. 4, pp. 462-466 (Dec. 1980).
  • Kunz, Andreas M. et al., “Design and Construction of a New Haptic Interface,” Proceedings of DETC '00, ASME 2000 Design Engineering Technical Conferences and Computers and Information in Engineering Conference, Baltimore, Maryland (Sep. 10-13, 2000).
  • La Scala, et al., “Design of an Extended Kalman Filter Frequency Tracker,” IEEE Transactions on Signal Processing, vol. 44, No. 3 (Mar. 1996).
  • Larimer et al., “VEWL: A Framework for building a Windowing Interface in a Virtual Environment,” in Proc. of IFIP TC13 Int. Conf. on Human-Computer Interaction Interact'2003 (Zürich, http://people.cs.vt.edu/˜bowman/papers/VEWLfinal.pdf (2003).
  • Laughlin, et al., “Inertial Anular Rate Sensors: Theory and Rate Sensors: Theory and Applications,” Sensors Magazine (Oct. 1992).
  • Lee et al., “Tilta-Pointer: the Free-Space Pointing Device,” Princeton COS 436 Project, http://www.milyehuang.com/cos436/project/specs.html (2004).
  • Lee, et al., “Innovative Estimation Method with Measurement Likelihood for all-Accelerometer Type Inertial Navigation System,” IEEE Transactions on Aerospace and Electronic Systems, vol. 38, No. 1 (Jan. 2002).
  • Lee, et al., “Two-Dimensional Position Detection System with MEMS Accelerometer for Mouse Applications” Design Automation Conference, 2001. Proceedings, pp. 852-857 (Jun. 2001).
  • Leganchuk et al., “Manual and Cognitive Benefits of Two-Handed Input: An Experimental Study,” ACM Transactions on Computer-Human Interaction, vol. 5, No. 4, pp. 326-359 (Dec. 1998).
  • Leonard, “Computer Pointer Controls 3D Images in Free Space,” Electronic Design, pp. 160, 162, 165 , (Nov. 1991).
  • Liang, et al., “On Temporal-Spatial Realism in the Virtual Reality Environment,” ACM 1991 Symposium on User Interface Software and Technology (Nov. 1991).
  • Link, “Field-Qualified Silicon Accelerometers From 1 Milli g to 200,000 g,” Sensors (Mar. 1993).
  • Liu, et al., “Enhanced Fisher Linear Discriminant Models for Face Recognition,” Proc. 14.sup.th International Conference on Pattern Recognition, Queensland, Australia (Aug. 1998).
  • Lobo et al., “Vision and Inertial Sensor Cooperation Using Gravity as a Vertical Reference,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 25, No. 12, pp. 1597-1608 (Dec. 2003).
  • Logitech, Logitech 2D/6D Mouse Devices Specification (1991).
  • Logitech, “Logitech 2D/6D Mouse Technical Reference Manual” (1991).
  • Logitech, Inc., “3D Mouse & Head Tracker Technical Reference Manual” (1992).
  • Logitech WingMan Cordless Rumblepad, Logitech, Press Release Sep. 2, 2001, 2 pages.
  • Louderback, Jim, “Nintendo Wii,” Reviews by PC Magazine, (Nov. 13, 2006).
  • “LPC2104/2105/2106, Single-chip 32-bit microcontrollers; 128 kB ISP/IAP Flash with 64 kB/32 kB/16 kB RAM”, Phillips, Dec. 22, 2004; 32 pages.
  • Luinge, Inertial sensing of human movement, Thesis, University of Twente (2002).
  • Luinge, et al., “Estimation of orientation with gyroscopes and accelerometers,” Proceedings of the First Joint BMES/EMBS Conference, 1999., vol. 2, p. 844 (Oct. 1999).
  • Luthi, P. et al., “Low Cost Inertial Navigation System,” and translation (2000).
  • MacKenzie et al., “A two-ball mouse affords three degrees of freedom,” Extended Abstracts of the CHI '97 Conference on Human Factors in Computing Systems, pp. 303-304. New York: ACM (1997).
  • MacKinlay, “Rapid Controlled Movement Through a Virtual 3D Workspace,” ACM SIGGRAPH Computer Graphics archive, vol. 24 , No. 4, pp. 171-176 (Aug. 1990).
  • MacLean, “Designing with Haptic Feedback”, Proceedings of IEEE Robotics and Automation (ICRA '2000), at 783-88 (Apr. 22-28, 2000).
  • MacLean, Karen, Publications and patents, bibliography (Nov. 2006).
  • Maggioni, C., “A novel gestural input device for virtual reality”, IEEE Virtual Reality Annual International Symposium (Cat. No. 93CH3336-5), 118-24, 1993.
  • Markey et al., “The Mechanics of Inertial Position and Heading Indication,” Massachusetts Institute of Technology, 1961.
  • Marti et al., “Biopsy navigator: a smart haptic interface for interventional radiological gestures”, International Congress Series, vol. 1256, Jun. 2003, 6 pages.
  • Marrin, “Possibilities for the Digital Baton as a General-Purpose Gestural Interface”, Late-Breaking/Short Talks, CHI 97, Mar. 22-27, 1997 (pp. 311-312).
  • Marrin, Teresa et al.: “The Digital Baton: a Versatile Performance Instrument” (1997).
  • Marrin, Teresa: “Toward an Understanding of Musical Gesture: Mapping Expressive Intention with the Digital Baton” (1996).
  • Masliah, “Measuring the Allocation of Control in 6 Degree of Freedom Human-Computer Interaction Tasks,” Proceedings of the SIGCHI conference on Human factors in computing systems, pp. 25-32 (2001).
  • Maybeck, “Stochastic Models, Estimation and Control,” vol. 1, Mathematics in Science and Engineering, vol. 141 (1979).
  • “MEMS enable smart golf clubs” Small Times—MEMS enable smart golf clubs, Jan. 6, 2005, 2 pages.
  • Merians, Alma S. et al.: “Virtual Reality-Augmented Rehabilitation for Patients Following Stroke,” Physical Therapy, vol. 82, No. 9 (Sep. 2002).
  • Merrill, “FlexiGesture: A sensor-rich real-time adaptive gesture and affordance learning platform for lectronic music control,” Thesis, Massachusetts Institute of Technology (Jun. 2004).
  • Meyer et al., “A Survey of Position Tracker,” vol. 1, Issue 2, pp. 173-200, MIT Presence, (1992).
  • Miles, “New pads lack control,” The Times, Dec. 6, 1999.
  • Mizell, “Using Gravity to Estimate Accelerometer Orientation,” IEEE Computer Society (2003).
  • Morgan, C.; “Still chained to the overhead projector instead of the podium? (TV Interactive Corp's LaserMouse Remote Pro infrared mouse) (Clipboard)(Brief Article) (Product Announcement)”, Government Computer News, Jun. 13, 1994.
  • Morris, “Accelerometry—a technique for the measurement of human body movements,” J Biomechanics 6: 729-736 (1973).
  • Moser, “Low Bud Inertial Navigation Platform (2000),” www.tmoser.ch/typo3/11.0.html, Oct. 2008.
  • Mulder, “How to Build an Instrumental Glove Based on the Powerglove Flex Sensors,” PCVR 16, pp. 10-14 (1994).
  • Mulder, “Human movement tracking technology,” School of Kinesiology, Simon Fraser University (Jul. 1994).
  • Myers, et al., “Interacting at a Distance: Measuring the Performance of Laser Pointers and Other Devices,” CHI 2002, (Apr. 2002).
  • N.I.C.E., “The N.I.C.E. Project” (video), (1997) http://www.niceproject.com/.
  • Naimark, et al., “Circular Data Matrix Fiducial System and Robust Image Processing for a Wearable Vision-Inertial Self-Tracker,” Proceedings. International Symposium on Mixed and Augmented Reality, ISMAR (2002).
  • Naimark, et al., “Encoded LED System for Optical Trackers,” Fourth IEEE and ACM International Symposium on Mixed and Augmented Reality, pp. 150-153 (2005).
  • Navarrete, et al., “Eigenspace-based Recognition of Faces: Comparisons and a new Approach,” Image Analysis and Processing (2001).
  • Newswire PR, “Five New Retailers to Carry Gyration's Gyropoint Point and Gyropoint Pro” (1996).
  • Newswire PR, “Three-Axis MEMS-based Accelerometer From STMicroelectronics Targets Handheld Terminals,” STMicro (Feb. 2003).
  • Nichols, “Geospatial Registration of Information for Dismounted Soldiers (GRIDS),” Contractor's Progress, Status, and Management Report (Milestone 3 Report to DARPA ETO) (Oct. 1998).
  • Nintendo, G3 Wireless Controller (Pelican) (2001).
  • Nintendo, Game Boy Advance SP System (2003).
  • Nintendo, GameBoy Color 1998.
  • Nintendo Game Boy, Consumer Information and Precautions Booklet, Nintendo, Jul. 31, 1969.
  • Nintendo, GameCube Controller (2001).
  • Nintendo, GameCube System and Controller (2001).
  • Nintendo, NES Controller (1984).
  • Nintendo, NES Duck Hunt Game (1984).
  • Nintendo, NES System and Controllers (1984).
  • Nintendo, NES Zapper Guns (1984).
  • Nintendo, Nintendo 64 Controller (1996).
  • Nintendo, Nintendo 64 System (N64) (1996).
  • Nintendo, Nintendo 64 System and Controllers (1996).
  • Nintendo, Nintendo Entertainment System (NES) (1984).
  • Nintendo, Nintendo Game Boy Advance (2001).
  • Nintendo, Nintendo Game Boy Advance System (2001).
  • Nintendo, Nintendo Game Boy Advance Wireless Adapter (Sep. 26, 2003).
  • Nintendo, Nintendo Game Boy Color Game Cartridge with Built-In Rumble (Jun. 28, 2009).
  • Nintendo, Nintendo GameBoy Color System (1998).
  • Nintendo, Nintendo GameBoy System (1989).
  • Nintendo, Nintendo GameCube System (2001).
  • Nintendo, Nintendo N64 Controller with Rumble Pack (1996-1997).
  • Nintendo, Nintendo N64 Rumble Packs (1996-1997).
  • Nintendo, Nintendo Super NES (SNES) (1991).
  • Nintendo, Nintendo: Kirby Tilt & Tumble game, packaging and user manual (Aug. 2000-2001).
  • Nintendo, Nintendo: WarioWare: Twisted game, packaging and user manual (2004-2005).
  • Nintendo, Pokemon Pinball (1998).
  • Nintendo, SNES Superscope (1991).
  • Nintendo, SNES System & Controllers (1991).
  • Nintendo, Wavebird Wireless Controllers (May 2002).
  • Nintendo Wavebird Controller, Nintendo, Jun. 2010 Wikipedia Article, http://en.wikipedia.org/wiki/WaveBird.
  • Nintendo, Nintendo Entertainment System Consumer Information and Precautions Booklet, Ninendo of America, Inc. 1992.
  • Nintendo, Nintendo Entertainment System Instruction Nintendo of America, Inc. 1992.
  • Nintendo, Nintendo Entertainment System Booth 2002.
  • Nintendo, Nintendo Entertainment System Layout, May 9, 2002.
  • Nintendo, Nintendo Feature: History of Pokeman Part 2, Official Nintendo Magazine May 17, 2009, http:///www.officialnintendomagazine.co.uk/article.php?id=8576.
  • Nishiyama, “A Nonlinear Filter for Estimating a Sinusoidal Signal and its Parameters in White Noise: On the Case of a Single Sinusoid,” IEEE Transactions on Signal Processing, vol. 45, No. 4 (Apr. 1997).
  • Nishiyama, “Robust Estimation of a Single Complex Sinusoid in White Noise-H.infin. Filtering Approach,” IEEE Transactions on Signal Processing, vol. 47, No. 10 (Oct. 1999).
  • Odell, “An Optical Pointer for Infrared Remote Controllers,” Proceedings of International Conference on Consumer Electronics (1995).
  • Odell, Transcript of Testimony, Investigation No. 337-TA-658, Before the United States International Trade Commission, vol. IV, redacted (May 14, 2009).
  • Ogawa et al., “Wii are the Elite,” GameSpot web site (Feb. 5, 2008).
  • Ojeda et al., “No GPS? No Problem?” University of Michigan Develops Award-Winning Personal Dead-Reackoning (PDR) System for Walking Users, http://www.engin.umich.edu/research/mrl/urpr/InPress/P135.pdf (post 2004).
  • OLPC, “One Laptop Per Child,” wiki.laptop.org/go/OneLaptopperChild (May 2009).
  • Omelyan, “On the numerical integration of motion for rigid polyatomics: The modified quaternion approach” Computers in Physics, vol. 12 No. 1, pp. 97-103 (1998).
  • Ovaska, “Angular Acceleration Measurement: A Review,” Instrumentation and Measurement Technology Conference, Conference Proceedings. IEEE, vol. 2 (Oct. 1998).
  • PAD-Controller and Memory I/F in Playstation (Apr. 17, 1995; Jan. 12, 2002).
  • Pai, et al., “The Tango: A Tangible Tangoreceptive Whole-Hand Interface,” Proceedings of World Haptics and IEEE Eurohaptics Conference, Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems (2005).
  • Paley, W. Bradford, “Interaction in 3D Graphics,” SIGGRAPH Computer Graphics Newsletter, Cricket input sevice Nov. 1998.
  • Paradiso, et al., “Interactive Therapy with Instrumented Footwear,” CHI 2004, Apr. 24-29, 2004, Vienna, Austria.
  • Paradiso, Joseph A., “The Brain Opera Technology: New Instruments and Gestural Sensors for Musical Interaction and Performance” (Nov. 1998) (“Brain Opera Article”).
  • Park, Adaptive control strategies for MEMS gyroscopes (Dissertation), Univ. Cal. Berkley (2000).
  • PC World, “The 20 Most Innovative Products of the Year” (Dec. 27, 2006).
  • Perry Simon: “Nintendo to Launch Wireless Game Boy Adaptor,” Digital Lifestyles, Sep. 26, 2003 http://digital-lifestyles.info/2003/09/26/nintendo-to-launch-wireless-game-boy-adaptor/.
  • Pham, Hubert “Pointing in Intelligent Environments with WorldCursor,” Proceedings of Internet 2003, Andrew Wilson (2003).
  • Phillips, “Forward/Up Directional Incompatibilities During Cursor Placement Within Graphical User Interfaces,” Ergonomics, informaworld.com (May 2005).
  • Phillips, “On the Right Track: A unique optical tracking system gives users greater freedom to explore virtual worlds” (Apr. 2000).
  • Photographs of prior art ASCII Grip V2 Controller, (cited in previous IDS as: ASCII/Sammy Grip V2 One Switch-Accessible Gaming Shop—ASCII Grip V2, http://www.oneswitch.org.uk/1/AGS/AGS-onehand/ascii-grip-v2.html, Jul. 10, 2008, pp. 1-2.)
  • Pierce et al., “Image Plane Interaction Techniques in 3D Immersive Environments,” Proceedings of the 1997 symposium on Interactive 3D graphics, portal.acm.org (1997).
  • Pilcher, “AirMouse Remote Controls,” IEEE Conference on Consumer Electronics (1992).
  • Pique, “Semantics of Interactive Rotations,” Interactive 3D Graphics, Proceedings of the 1986 workshop on Interactive 3D graphics, pp. 259-269 (Oct. 1986).
  • Piyabongkarn, “Development of a MEMS Gyroscope for Absolute Angle Measurement,” IEEE Transactions on Control Systems Technology, vol. 13, Issue 2, pp. 185-195 (Mar. 2005).
  • Piyabongkarn, “Development of a MEMS Gyroscope for Absolute Angle Measurement,” Dissertation, Univ. Minnesota (Nov. 2004).
  • Pokeman Pinball Game, 1999, Wikipedia Article, http://en.wikipedia.org/wiki/Pok?C3?A9monPinball.
  • Polhemus, “Polhemus 3SPACE FASTRACK devices” (image) (2000).
  • Polhemus: “FASTRACK, The Fast and Easy Digital Tracker” copyrighted 2001, Coldiester, Vermont 2 pages.
  • PowerGlove product Program Guide, Mattel, 1989.
  • PowerGlove product, Mattel, 1989 Wikipedia Article.
  • PowerGlove product, Instructions, Mattel, 1989.
  • Pryor et al., “A Reusable Software Architecture for Manual Controller Integration,” IEEE Conf. on Robotics and Automation, Univ of Texas (Apr. 1997).
  • Raab, et al., “Magnetic Position and Orientation Tracking System,” IEEE Transactions on Aerospace and Electronic Systems, vol. AES-15, No. 5, pp. 709-718 (Sep. 1979).
  • Raethjen, et al., “Tremor Analysis in Two Normal Cohorts,” Clinical Neurophysiology 115 (2004).
  • Rebo, “Helmet-mounted virtual environment display system,” Thesis, Air Force Institute of Technology, Defense Technical Information Center (Dec. 1988).
  • Rebo, et al., “Helmet-Mounted Virtual Environment Display System,” Proc. SPIE vol. 1116, pp. 80-84 (Sep. 1989).
  • Regan, “Smart Golf Clubs”, The Baltimore Sun, Jun. 17, 2005, 1 page.
  • Rekimoto, “Tilting Operations for Small Screen Interfaces,” Proceedings of the 9th Annual ACM Symposium On User Interface Software and Technology, pp. 167-168 (1996).
  • Reunert, “Fiber-Optic Gyroscopes: Principles and Applications,” Sensors, (Aug. 1993).
  • Ribo, et al., “Hybrid Tracking for Outdoor Augmented Reality Applications,” IEEE Computer Graphics and Applications, vol. 22, No. 6, pp. 54-63 (Nov./Dec. 2002).
  • Riviere, Cameron, Testimony, Trial Day 5, In the Matter of Certain Video Game Machines and Related Three-Dimensional Pointing Devices, ITC Investigation No. 337-TA-658 (May 15, 2009).
  • Riviere, C., Robotics Institute, http://www.ri.cmu.edu/people/riviere.sub.--cameron.html http://www.ri.cmu.edu/person.html?type=publications&personid=248 (Aug. 2007).
  • Riviere, et al., “Adaptive Canceling of Physiological Tremor for Improved Precision in Microsurgery,” IEEE Transactions on Biomedical Engineering, vol. 45, No. 7 (Jul. 1998).
  • Riviere, et al., “Toward Active Tremor Canceling in Handheld Microsurgical Instruments,” IEEE Transactions on Robotics and Automation, vol. 19, No. 5 (Oct. 2003).
  • Roberts, “The Lincoln Wand,” AFIPS Conference Proceedings, MIT Lincoln Laboratory (1966).
  • Robinett et al., “Implementation of Flying, Scaling, and Grabbing in Virtual Worlds,” ACM Symposium (1992).
  • Robinett et al., “The Visual Display Transformation for Virtual Reality,” University of North Carolina at Chapel Hill (1994).
  • Robotics Research Group, “Robot Design: Robot Manual Controller Design,” The University of Texas of Austin, May 2009.
  • Roetenberg, “Inertial and magnetic sensing of human motion,” Thesis (2006).
  • Roetenberg, et al., “Inertial and Magnetic Sensing of Human Movement Near Ferromagnetic Materials,” Proceedings. The Second IEEE and ACM International Symposium on Mixed and Augmented Reality (Mar. 2003).
  • Rolland, et al., “A Survey of Tracking Technology for Virtual Environments,” University of Central Florida, Center for Research and Education in Optics Lasers (CREOL) (2001).
  • Sakai, et al., “Optical Spatial Filter Sensor for Ground Speed,” Optical Review, vol. 2, No. 1 pp. 65-67 (1994).
  • Satterfield, Shane, E3 2002: Nintendo announces new GameCube games, GameSpot, May 21, 2002, http://wwwgamespot.com/gamecube/action/rollarama/new.html?sid=2866974&comact-convert&omclk=nesfeatures&tag=newsfeatures%Btitle%3B.
  • Savage, Paul G., “Advances in Strapdown Inertial Systems,” Lecture Series Advisory Group for Aerospace Research and Development Neuilly-Sur-Seine (France) (1984).
  • Sawada et al., “A Wearable Attitude-Measurement System Using a Fiberoptic Gyroscope” Massachusetts Institute of Technology, vol. 11, No., Apr. 2002, pp. 109-118.
  • Saxena et al., “In Use Parameter Estimation of Inertial Sensors by Detecting Multilevel Quasi-Static States,” Lecture Notes in Computer Science, 2005—Berlin: Springer-Verlag, (Apr. 2004).
  • Sayed, “A Framework for State-Space Estimation with Uncertain Models,” IEEE Transactions on Automatic Control, vol. 46, No. 7 (Jul. 2001).
  • Sayed, UCLA Adaptive Systems Laboratory—Home Page, UCLA, http://asl.ee.ucla.edu/index.php?option=com.sub.--frontpage&Itemid=1 (Aug. 2007).
  • Schmorrow et al., “The PSI Handbook of Virtual Environments for Training and Education,” vol. 1, 2009.
  • Schofield, Jack et al., Coming up for airpad, The Guardian (Feb. 2000).
  • Sega/Sports Sciences, Inc., “Batter Up, It's a Hit,” Instruction Manual, Optional Equipment Manual (1994).
  • Sega/Sports Sciences, Inc., “Batter Up, It's a Hit,” Photos of baseball ball (1994).
  • Selectech, “Airmouse Remote Control System Model AM-1 User's Guide,” Colchester, VT (Sep. 24, 1991).
  • Selectech, “AirMouse Remote Controls, AirMouse Remote Control Warranty” (1991).
  • Selectech, “Changing Driver Versions on CDTV/AMIGA” (Oct. 17, 1991).
  • Selectech, “Selectech AirMouse Remote Controls, Model # AM-R1,” photographs (1991).
  • Selectech, Facsimile Transmission from Rossner to Monastiero, Airmouse Remote Controls, Colchester, VT (Mar. 25, 1992).
  • Selectech, Selectech AirMouse Devices (image) (1991).
  • Selectech, Software, “AirMouse for DOS and Windows IBM & Compatibles,” “AirMouse Remote Control B0100EN-C, Amiga Driver, CDTV Driver, Version: 1.00,” “AirMouse Remote Control B0100EM-C.1, Apple Macintosh Serial Driver Version: 1.00 (1.01B),” “AirMouse Remote Control B0100EL-B/3.05 DOS Driver Version: 3.0, Windows Driver Version 1.00,” AirMouse Remote Control MS-DOS Driver Version: 3.00/3.05, Windows 3.0 Driver Version: 1.00 (1991).
  • “Self-Contained, Free Standing “Fishing Rod” Fishing Games,” Miacomet and Interact Announce Agreement to Launch Line of Reel Feel™ Sport Controllers, Press Release, May 13, 1999, 4 pages.
  • Seoul National Univ., “EMMU System”—Seoul National Univ Power Point Presentation, www.computer.org/portal/cmsdocsieeecs/ieeecs/education/csidc/CSIDC03Presentations/SNU.ppt (2003).
  • Serial Communication (Wikipedia) (Jul. 2, 2005).
  • Shoemake, Ken, Quaternions, UPenn, Online (Oct. 2006).
  • Simon, et al. “The Yo Yo: A Handheld Combining Elastic and Isotonic Input,” http://www.uni-weimar.de/cms/fileadmin/medien/vr/documents/publications/TheYoYo-Interact2003-Talk.pdf (2003).
  • Simon, et al., “The Yo Yo: A Handheld Device Combining Elastic and Isotonic Input,” Human-Computer Interaction—INTERACT'03, pp. 303-310 (2003).
  • SmartSwing internal drawing, 1 page (2004).
  • SmartSwing, Training Aid, Apr. 2005, Austin, Texas.
  • SmartSwing: “Register to be notified when Smartswing products are available for purchase,” 3 pages, May 2004, retrieved May 19, 2009, http://web.archive.org/web/20040426182437/www.smartswinggolf.com/.
  • SmartSwing: “SmartSwing: Intellegent Golf Clubs that Build a Better Swing,” 2 pages, 2004 retrieved May 19, 2009, http://web.archive.org/web/20040728221951/http://www.smartswinggolf....
  • SmartSwing: “The SmartSwing Learning System Overview,” 3 pages, 2004, retrieved May 19, 2009, http://web.archive.org/web/20040810142134/http://www.smartswinggolf.com/t....
  • SmartSwing: “The SmartSwing Product, 3 pages, 2004, retrieved May 19, 2009, http://web.archive.org/web/200400403204628/http://www.smartswinggolf.com/...”.
  • SmartSwing: The SmartSwing Product Technical Product: Technical Information, 1 page, 2004, retrieved May 19, 2009, http://web.archive.org/web/200400403205906/http://www.smartswinggolf.com/....
  • SmartSwing, Letter from the CEO—pp. 1-3, May 2009.
  • SmartSwing: The SmartSwing Learning System: How it Works, 3 pages, 2004, retrieved May 19, 2009, http://web.archive.org/web/20040403213108/http://www.smartswinggolf.com/....
  • Smith, “Gyrevolution: Orienting the Digital Era,” http://www.gyration.com/images/pdfs/GyrationWhitePaper.pdf (2007).
  • Sorenson, et al., “The Minnesota Scanner: A Prototype Sensor for Three-Dimensional Tracking of Moving Body Segments,” IEEE Transactions on Robotics and Animation (Aug. 1989).
  • Sourceforge.com, “ARToolkit API Documentation” (SourceForge web pages) (2004-2006).
  • Stovall, “Basic Inertial Navigation,” NAWCWPNS TM 8128, Navigation and Data Link Section, Systems Integration Branch (Sep. 1997).
  • Sulic, “Logitech Wingman Cordless Rumblepad Review”, Review at IGN, 4 pages, Jan. 14, 2002.
  • “Superfamicom Grip controller by ASCII,” http://superfami.com/sfcgrip.html, Jul. 10, 2008, pp. 1-2.
  • Sutherland, “A Head-Mounted Three Dimensional Display,” AFIPS '68 (Fall, part I): Proceedings of the Dec. 9-11, 1968, fall joint computer conference, part I, pp. 757-764.
  • Sutherland, Ivan E., “Sketchpad: A Man-Machine Graphical Communication System,” AFIPS '63 (Spring): Proceedings of the May 21-23, 1963, Spring Joint Computer Conference, pp. 329-346.
  • Sweetster, “A Quaternion Algebra Tool Set,” http://world.std.com/%7Esweetser/quaternions/intro/tools/tools.html (Jun. 2005).
  • Swisher “How Science Can Improve Your Golf Game, Your Club is Watching” The Wall Street Journal, Apr. 18, 2005, 1 page.
  • Templeman, James N., “Virtual Locomotion: Walking in Place through Virtual Environments,” Presence, vol. 8 No. 6, pp. 598-617, Dec. 1999.
  • Thinkoptics, Thinkoptics Wavit devices (image) (2007).
  • Timmer, “Data Analysis and Modeling Dynamic Processes in the Life Sciences,” Freiburg Center for Data Analysis and Modeling, http://webber.physik.uni-freiburg.de/ .about.jeti/ (Aug. 2007).
  • Timmer, “Modeling Noisy Time Series: Physiological Tremor,” International Journal of Bifurcation and Chaos, vol. 8, No. 7 (1998).
  • Timmer, et al, “Pathological Tremors: Deterministic Chaos or Nonlinear Stochastic Oscillators?” Chaos, vol. 10, No. 1 (Mar. 2000).
  • Timmer, et al., “Characteristics of Hand Tremor Time Series,” Biological Cybernetics, vol. 70 (1993).
  • Timmer, et al., Cross-Spectral Analysis of Physiological Tremor and Muscle Activity: I Theory and Application to unsynchronized electromyogram, vol. 78 (1998).
  • Timmer, et al., Cross-Spectral Analysis of Physiological Tremor and Muscle Activity: II Application to Synchronized Electromyogram, Biological Cybernetics, vol. 78 (1998).
  • Timmer, et al., “Cross-Spectral Analysis of Tremor Time Series,” International Journal of Bifurcation and Chaos, vol. 10, No. 11 2000.
  • Titterton et al., “Strapdown Inertial Navigation Technology,” pp. 1-56 and pp. 292-321 (May 1997).
  • Traq 3D (Trazer) Product, http://www.exergamefitness.com/traq3d.htm, http://www.trazer.com/, http://www.traq3d.com/ (1997).
  • Traq 3D, “Healthcare” 1 pages, //www.traq3d.com/Healthcare/Healthcare.aspx, 1997.
  • Translation of the brief of BigBen of Oct. 27, 2010 and original German text (Nov. 3, 2010).
  • Translation of the brief of System Com 99 of Oct. 27, 2010 and original German text.
  • Translation of Exhibit B-B01: Cancellation Request of BigBen of Oct. 15, 2010 against German utility model 20 2006 020 818 (UM1) (Oct. 15, 2010) and original German text.
  • Translation of Exhibit B-C01: Cancellation Request of BigBen of Oct. 15, 2010 against German utility model 20 2006 020 819 (UM2) (Oct. 15, 2010) and original German text.
  • Translation of Exhibit B-D01: Cancellation Request of BigBen of Oct. 15, 2010 against German utility model 20 2006 020 820 (UM3) (Oct. 15, 2010) and original German text.
  • Translation of Opposition Brief of BigBen of Sep. 2, 2010 Against European Patent No. EP 1854518.
  • Transmission Mode (Apr. 22, 1999).
  • Ulanoff, Lance, “Nintendo's Wii is the Best Product Ever,” PC Magazine (Jun. 21, 2007).
  • UNC Computer Science Department, “News & Notes from Sitterson Hall,” UNC Computer Science, Department Newsletter, Issue 24, Spring (Apr. 1999).
  • Univ. Illinois At Chicago, “CAVE—A Virtual Reality Theater,” http://www.youtube.com/watch?v=-Sf6bJjwSCE 1993.
  • Univ. Wash., “ARToolkit” (U. Wash. web pages) (1999).
  • Urban, “BAA 96-37 Proposer Information,” DARPA/ETO (1996).
  • US Dynamics Corp, “Spinning Mass Mechanical Gyroscopes” (Aug. 2006).
  • US Dynamics Corp, “The Concept of ‘Rate’ (more particularly, angular rate pertaining to rate gyroscopes) (rate gyro explaination),” (Aug. 2006).
  • US Dynamics Corp, “US Dynamics Model 475 Series Rate Gyroscope Technical Brief—brief discussion on rate gyroscope basics, operation, and uses, and a dissection of the model by major component” (Dec. 2005).
  • US Dynamics Corp, “US Dynamics Rate Gyroscope Interface Brief (rate gyro IO)” (Aug. 2006).
  • VTI, Mindflux-VTi CyberTouch, 1996, http://www.mindflux.com.au/products/vti/cybertouch.html.
  • Van Den Bogaard, “Using linear filters for real-time smoothing of rotational data in virtual reality application,” http://www.science.uva.nl/research/ias/alumni/m.sc.theses/theses/RobvandenBogaard.pdf (Aug. 2004).
  • Van Laerhoven, et al., “Using an Autonomous Cube for Basic Navigation and Input,” Proceedings of the 5th International Conference on Multimodal interfaces, Vancouver, British Columbia, Canada, pp. 203-210 (2003).
  • Van Rheeden, et al., “Noise Effects on Centroid Tracker Aim Point Estimation,” IEEE Trans. on Aerospace and Electronic Systems, vol. 24, No. 2, pp. 177-185 (Mar. 1988).
  • Vaz, et al., “An Adaptive Estimation of Periodic Signals Using a Fourier Linear Combiner,” IEEE Transactions on Signal Processing, vol. 42, Issue 1, pp. 1-10 (Jan. 1994).
  • Verplaetse, “Inertial Proprioceptive Devices: Self-Motion Sensing Toys and Tools,” IBM Systems Journal (Sep. 1996).
  • Verplaetse, “Inertial-Optical Motion-Estimating Camera for Electronic Cinematography,” Masters of Science Thesis, MIT, (1997).
  • Villoria, Gerald, Hands on Roll-O-Rama Game Cube, Game Spot, May 29, 2002, http://www.gamespot.com/gamecube/action/rollorama/news.html?sid=2868421&comact=convert&omclk=newsfeatures&tag=newsfeatures;title;1&m.
  • Virtual Fishing, Operational Manual, 2 pages, Tiger Electronics, Inc., 1998.
  • Virtual Technologies, Inc., Cyberglove brochure, Palo Alto, CA, www.virtex.com. (1999).
  • Vorozcovs, et al.,“The Hedgehog: A Novel Optical Tracking Method for Spatially Immersive Displays,” MIT Presence, vol. 15, No. 1, pp. 108-121 (2006).
  • VR Solutions, “IS-1200”, www.vrs.com.au/motion-tracking/intersense/is-1200.html 2 pages (May 2009).
  • Wang, et al., “Tracking a Head-Mounted Display in a Room-Sized Environment with Head-Mounted Cameras,” SPIE 1990 Technical Symposium on Optical Engineering and Photonics in Aerospace Sensing, vol. 1290, pp. 47-57 (1990).
  • Ward, et al., “A Demonstrated Optical Tracker With Scalable Work Area for Head-Mounted Display Systems,” Symposium on Interactive 3D Graphics, Proceedings of the 1992 Symposium on Interactive 3D Graphics, pp. 43-52, ACM Press, Cambridge, MA (1992).
  • Watt, 3D Computer Graphics, “Three-Dimensional Geometry in Computer Graphics,”, pp. 1-22 Addison-Wesley (1999).
  • Welch et al., HiBall Devices (image) (2002-2006).
  • Welch et al., Motion Tracking: No Silver Bullet, but a Respectable Arsenal IEEE Computer Graphics and pplications, vol. 22, No. 6, pp. 24-38 (Nov. 2002).
  • Welch, “Hybrid Self-Tracker: An Inertial/Optical Hybrid Three-Dimensional Tracking System,” Tech. Report TR95-048, Dissertation Proposal, Univ. of North Carolina at Chapel Hill, Dept. Computer Science, Chapel Hill, N.C. (1995).
  • Welch, “A Self-Contained Wide-Area Tracker Using Sensor Fusion” (2001).
  • Welch, “Hawkeye Zooms in on Mac Screens with Wireless Infrared Penlight Pointer,” MacWeek (May 1993).
  • Welch, et al., “Complementary Tracking and Two-Handed Interaction for Remote 3D Medical Consultation ith a PDA,” Proceedings of Trends and Issues in Tracking for Virtual Environments, Workshop at the IEEE Virtual Reality 2007 Conference (Mar. 2007).
  • Welch, et al., “High-Performance Wide-Area Optical Tracking: The HiBall Tracking System,” MIT Presence: Teleoperators & Virtual Environments (2001).
  • Welch, et al., “SCAAT: Incremental Tracking with Incomplete Information,” Computer Graphics, SIGGRAPH 97 Conference Proceedings, pp. 333-344 (Aug. 1997).
  • Welch, et al., “Source Code for HiBall+Inerital device,” UNC-CH Computer Science (Jun. 1998).
  • Welch, et al., “The HiBall Tracker: High-Performance Wide-Area Tracking for Virtual and Augmented Environments,” ACM SIGGRAPH, Addison-Wesley (1999).
  • Welch, et al., “The High-Performance Wide-Area Optical Tracking : The HiBall Tracking System,” MIT Presence, Presence, vol. 10 , No. 1 (Feb. 2001).
  • Welch, et al., “Tracking for Training in Virtual Environments: Estimating the Pose of People and Devices for Simulation and Assessment,” [J. Cohn, D. Nicholson, and D. Schmorrow, editors, The PSI Handbook of VirtuaI Environments for Training and Education: Developments for the Milita and Beyond, Chap.1, pp. 23-47] (2008).
  • Widrow, et al., “Fundamental Relations Between the LMS Algorithm and the DFT,” IEEE Transactions on Circuits and Systems, vol. 34, No. CAS-7, (Jul. 1987).
  • Wiley, M.: “Nintendo Wavebird Review,” US, Jun. 11, 2002, 21 pages.
  • Williams, et al., “Physical Presence: Palettes In Virtual Spaces,” Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, vol. 3639, No. 374-384 (May 1999).
  • Williams, Robert L. et al., “Implementation and Evaluation of a Haptic Playback System,” vol. 3 No. 3, Haptics-e (2004).
  • Williams, Robert L. et al., “The Virtual Haptic Back Project,” Presented at the IMAGE 2003 Conference, Scottsdale, Arizong (Jul. 14-18, 2003).
  • Wilson, “Wireless User Interface Devices for Connected Intelligent Environments,” Ubicomp 2003 Workshop (2003).
  • Wilson, “WorldCursor: Pointing in Intelligent Environments with the World Cursor,” UIST '03 Companion (Nov. 2003).
  • Wilson, Research page, biography available at http://research.microsoft.com/en-us/um/people/awilson/?0sr=a, Microsoft Corp. (2009).
  • Wilson, Transcript of Testimony, Investigation No. 337-TA-658, Before the United States International Trade Commission, vol. V (May 15, 2009).
  • Wilson, XWand video, http://research.microsoft.com/˜awilson/wand/wand%20video%20768k.WMV (Mar. 2002).
  • Wired Glove, Wikipedia Article, 4 pages, http://en.wikipedia.org/wiki/Wiredglove, (Nov. 18, 2010).
  • Wireless (Wikipedia) (Aug. 12, 2005).
  • Wormell, “Unified Camera, Content and Talent Tracking in Digital Television and Movie Production,” InterSense, Inc. & Mark Read, Hypercube Media Concepts, Inc. Presented: NAB 2000, Las Vegas, NV, Apr. 8-13, 2000.
  • Wormell, et al., “Advancements In 3D Interactive Devices for Virtual Environments,” ACM International Conference Proceeding Series; vol. 39 (2003).
  • Worringham, et al., “Directional Stimulus-Response Compatibility: A Test of Three Alternative Principles,” Ergonomics, vol. 41, Issue 6, pp. 864-880 (Jun. 1998).
  • www.3rdtech.com (2000-2006).
  • Yang, et al., “Implementation and Evaluation of ‘Just Follow Me’: An Immersive, VR-Based, Motion-Training System,” MIT Presence: Teleoperators and Virtual Environments, vol. 11 No. 3, at 304-23 (MIT Press) (Jun. 2002).
  • You, et al., “Hybrid Inertial and Vision Tracking for Augmented Reality Registration,” http://graphics.usc.edu/cgit/pdf/papers/Vr1999.PDF (1999).
  • You, et al., “Orientation Tracking for Outdoor Augmented Reality Registration,” IEEE Computer Graphics and Applications, IEEE, vol. 19, No. 6, pp. 36-42 (Nov. 1999).
  • Youngblut, et al., “Review of Virtual Environment Interface Technology,” Institute for Defense Analyses (Jul. 1996).
  • Yun, et al., “Recent Developments in Silicon Microaccelerometers,” Sensors, University of California at Berkeley (Oct. 1992).
  • Zhai, “Human Performance in Six Degree of Freedom Input Control,” Thesis, University of Toronto (1995).
  • Zhai, “User Performance in Relation to 3D Input Device Design”, Computer Graphics 32(4), Nov. 1998, 15 pages.
  • Zhou, et al., “A survey—Human Movement Tracking and Stroke Rehabilitation,” Technical Report: CSM-420, ISSN 1744-8050, Dept. of Computer Sciences, University of Essex, UK (Dec. 8, 2004).
  • Zhu, et al., “A Real-Time Articulated Human Motion Tracking Using Tri-Axis Inertial/Magnetic Sensors Package,” IEEE Transactions on Neural Systems and Rehabilitation Engineering, vol. 12, No. 2 (Jun. 2004).
  • European Search Report for Application No. EP 07 11 2880, Oct. 18, 2007.
  • Office Action issued in related Chinese patent application 200610111559.7 (Sep. 18, 2009).
  • Office Action issued in related Japanese patent application 2006-216569 (Oct. 20, 2009).
  • Office Action issued in corressonding Japanese patent application 2007-203785 (Oct. 27, 2008).
  • Office Action issued in corresponding Japanese patent application 2008-256858 (Sep. 9, 2010).
  • Notice of Allowance issued on Feb. 22, 2011 in U.S. Appl. No. 11/446,188.
  • Office Action issued Oct. 21, 2009 in U.S. Appl. No. 11/446,188.
  • Office Action issued Feb. 27, 2009 in U.S. Appl. No. 11/446,188.
  • Office Action issued Apr. 30, 2008 in U.S. Appl. No. 11/446,188.
  • Notice of Allowance issued on Apr. 15, 2010 in U.S. Appl. No. 11/532,328.
  • Office Action issued Oct. 6, 2009 in U.S. Appl. No. 11/532,328.
  • Office Action issued Nov. 25, 2008 in U.S. Appl. No. 11/532,328.
  • Office Action issued Oct. 19, 2010 in U.S. Appl. No. 11/446,187.
  • Office Action (Examiner-Initiated Interview Summary) issued Oct. 26, 2009 in U.S. Appl. No. 11/446,187.
  • Office Action issued Aug. 19, 2009 in U.S. Appl. No. 11/446,187.
  • Office Action issued Apr. 29, 2009 in U.S. Appl. No. 11/446,187.
  • Office Action issued Feb. 26, 2009 in U.S. Appl. No. 11/446,187.
  • Office Action issued Jul. 1, 2008 in U.S. Appl. No. 11/404,871.
  • Office Action issued Jun. 14, 2010 in U.S. Appl. No. 11/404,871.
  • Office Action issued Sep. 30, 2009 in U.S. Appl. No. 11/404,871.
  • Office Action issued Mar. 18, 2009 in U.S. Appl. No. 11/404,871.
  • Office Action issued Sep. 4, 2008 in U.S. Appl. No. 11/404,871.
  • Examiner's Answer issued Feb. 22, 2011 in U.S. Appl. No. 11/404,871.
  • Office Action issued Sep. 1, 2010 in U.S. Appl. No. 11/790,780.
  • Office Action issued Mar. 12, 2010 in U.S. Appl. No. 11/790,780.
  • Notice of Allowance issued May 21, 2010 in U.S. Appl. No. 12/222,787.
  • Office Action issued Feb. 5, 2010 in U.S. Appl. No. 12/222,787.
  • Office Action issued Aug. 23, 2010 in U.S. Appl. No. 11/488,135.
  • Office Action issued Apr. 21, 2010 in U.S. Appl. No. 11/488,135.
  • Notice of Allowance issued Sep. 14, 2010 in U.S. Appl. No. 11/764,409.
  • Office Action issued Jan. 6, 2010 in U.S. Appl. No. 11/764,409.
  • Office Action issued Feb. 20, 2009 in U.S. Appl. No. 11/764,409.
  • Office Action issued Oct. 6, 2010 in U.S. Appl. No. 11/404,844.
  • Office Action issued Sep. 18, 2009 in U.S. Appl. No. 11/404,844.
  • Office Action issued Jan. 29, 2009 in U.S. Appl. No. 11/404,844.
  • Office Action issued Sep. 4, 2008 in U.S. Appl. No. 11/404,844.
  • Office Action issued May 3, 2010 in U.S. Appl. No. 11/598,787.
  • Office Action issued Jan. 27, 2011 in U.S. Appl. No. 11/504,086.
  • Office Action issued in U.S. Appl. No. 11/441,146 on Sep. 27, 2010.
  • Office Action issued in U.S. Appl. No. 11/441,146 on Nov. 26, 2010.
  • Office Action issued in U.S. Appl. No. 11/441,146 on May 11, 2011.
  • Advisory Action issued in U.S. Appl. No. 11/441,146 on Sep. 1, 2011.
Patent History
Patent number: 8157651
Type: Grant
Filed: Jun 2, 2006
Date of Patent: Apr 17, 2012
Patent Publication Number: 20070072674
Assignee: Nintendo Co., Ltd. (Kyoto)
Inventors: Keizo Ohta (Kyoto), Kenta Sato (Kyoto)
Primary Examiner: Pierre E Elisca
Attorney: Nixon & Vanderhye P.C.
Application Number: 11/445,280