Patents by Inventor Toshikazu Ohshima
Toshikazu Ohshima has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 7330197Abstract: In exhibition of a synthesized image which is obtained by synthesizing a virtual world image with a real world image observed from a viewpoint position and direction of a user, data representing a position and orientation of a user is acquired, a virtual image is generated based on the data representing the position and orientation of the user, and the virtual image is synthesized with a real image corresponding to the position and orientation of the user. Based on a measurable area of the position and orientation of the user, area data is set. Based on the data representing the position of the user and the area data, notification related to the measurable area is controlled.Type: GrantFiled: December 3, 2004Date of Patent: February 12, 2008Assignee: Canon Kabushiki KaishaInventors: Toshihiro Kobayashi, Toshikazu Ohshima
-
Publication number: 20080030499Abstract: In a mixed-reality presentation system, a handler for manipulating a virtual object is displayed as a virtual object. Then, manipulation of the handler by the hand(s) of an observer is detected as manipulation of the virtual object associated with the handler, and those detection results are reflected in a display of the manipulated virtual object. By dynamically changing virtual objects displaying the handler according to the size of the virtual objects displayed in a view volume of the observer presented with a mixed reality space, the observer can select a desired virtual object with ease.Type: ApplicationFiled: July 24, 2007Publication date: February 7, 2008Applicant: CANON KABUSHIKI KAISHAInventors: Koichiro Wanda, Toshikazu Ohshima
-
Publication number: 20080024523Abstract: A method for combining a real space image with a virtual image, includes: causing an imaging unit to capture an image of a real space; generating an image covering a predetermined space based on a plurality of real images in the captured real space; extracting position information of a light source based on the generated image; and adding a light source or a shadow on the virtual image based on the extracted position information of the light source.Type: ApplicationFiled: July 23, 2007Publication date: January 31, 2008Applicant: CANON KABUSHIKI KAISHAInventors: Kaname Tomite, Toshikazu Ohshima
-
Publication number: 20070258658Abstract: An information processing apparatus connected to a display apparatus, the information processing apparatus comprises an acquisition unit adapted to acquire a real image captured by an imaging apparatus; a generation unit adapted to generate a virtual image based on an orientation of said imaging apparatus; a detection unit adapted to detect a tilt of the real image; a correction unit adapted to correct the real image based on the detected tilt of said real image; a composite unit adapted to generate a composite image based on the corrected real image and the virtual image; and an output unit adapted to output the corrected composite image to said display apparatus.Type: ApplicationFiled: April 25, 2007Publication date: November 8, 2007Inventors: Toshihiro Kobayashi, Toshikazu Ohshima
-
Patent number: 7289130Abstract: A game state manager (201) manages the state of an AR game (information that pertains to rendering of each virtual object (102), the score of a player (101), the AR game round count, and the like). An objective viewpoint video generator (202) generates a video of each virtual object (102) viewed from a camera (103). An objective viewpoint video composition unit (203) generates a composite video of the video of the virtual object (102) and an actually sensed video, and outputs it to a display (106). A subjective viewpoint video generator (212) generates a video of the virtual object (102) viewed from an HMD (107). A subjective viewpoint video composition unit (213) generates a composite video of the video of the virtual object (102) and an actually sensed video, and outputs it to the HMD (107).Type: GrantFiled: September 8, 2000Date of Patent: October 30, 2007Assignee: Canon Kabushiki KaishaInventors: Kiyohide Satoh, Toshikazu Ohshima
-
Publication number: 20070236510Abstract: An acquired real image undergoes conversion based on a first parameter to generate an image corresponding to a designated characteristic. A virtual image is generated on the basis of a second parameter to generate an image corresponding to a designated characteristic. The converted real image is composited with the generated virtual image. The composite image is displayed.Type: ApplicationFiled: March 30, 2007Publication date: October 11, 2007Inventors: Hiroyuki Kakuta, Toshikazu Ohshima
-
Patent number: 7212687Abstract: When a CG object is overlapped onto a photographed real scenery image as a background and an image of a photographed real photographic subject is synthesized in front of the overlapped image, an image in which the photographed real background and the photographic subject are imaged simultaneously without imaging the photographed real background and the photographic subject independently, is synthesized with CG. A photographed real image including a specific image is acquired, and area information representing an area of the specific image from the photographed real image, and an area other than the area of the specific image of the photographed real image is synthesized with a computer graphics image using the area information so detected.Type: GrantFiled: March 27, 2003Date of Patent: May 1, 2007Assignee: Canon Kabushiki KaishaInventor: Toshikazu Ohshima
-
Patent number: 7199807Abstract: A set of objects to be rendered by an identical rendering method is specified from objects which form a virtual space. A hierarchical structure formed by the object included in the specified set is generated. The objects included in the specified set are rendered by the rendering method common to the objects included in this set in accordance with the generated hierarchical structure.Type: GrantFiled: November 5, 2004Date of Patent: April 3, 2007Assignee: Canon Kabushiki KaishaInventors: Toshihiro Kobayashi, Toshikazu Ohshima
-
Publication number: 20060273984Abstract: When a plurality of mixed reality (MR) experiencing persons experience the same mixed reality, there is a possibility that the experiencing persons come into contact with one another. An image processing apparatus capable of reporting to the MR experiencing persons that there is a possibility of contact. Therefore, a real space image that is image captured from the position and orientation of a user's viewpoint is drawn. The position and orientation of the user's viewpoint at this time is detected by a sensor unit. It is determined whether or not the viewpoint position is smaller than or equal to an attention distance at which there is a possibility of contact. If the viewpoint position is smaller than or equal to the attention distance, an attention display for this fact is performed.Type: ApplicationFiled: April 7, 2006Publication date: December 7, 2006Applicant: Canon Kabushiki KaishaInventors: Koichiro Wanda, Toshikazu Ohshima
-
Patent number: 7123214Abstract: It is intended to solve the problem that, when an operator performs an operation using two-dimensional input means while seeing a picture frame seen from a viewpoint of the operator, an input operation is difficult to perform when the direction of the line of sight of the operator is unstable, since a designated direction greatly deviates in a world coordinate system. For that purpose, viewpoint information is detected, and an instruction of the operator for operating a position of a pointer image is input. A designated direction in a pointer coordinate system is obtained in accordance with the operator's instruction, and the pointer image is generated based on the designated direction. The pointer coordinate system is changed from the detected viewpoint information in accordance with a specific instruction of the operator.Type: GrantFiled: March 25, 2003Date of Patent: October 17, 2006Assignee: Canon Kabushiki KaishaInventors: Tsuyoshi Kuroki, Toshikazu Ohshima
-
Publication number: 20060202985Abstract: A viewpoint position/orientation is acquired (S1020). An area of an environment mapping image to be used for environment mapping on the environment mapping target area of a virtual object is specified on the basis of the viewpoint orientation (S1040 and S1060). An image in the specified area is environment-mapped on the environment mapping target area (S1060). An image is generated by seeing, from the viewpoint, a virtual space including the virtual object which has undergone environment mapping.Type: ApplicationFiled: March 7, 2006Publication date: September 14, 2006Applicant: CANON KABUSHIKI KAISHAInventors: Toshihiro Kobayashi, Toshikazu Ohshima
-
Patent number: 7084887Abstract: There is disclosed a marker layout method in a mixed reality space, which can reliably detect markers in units of players, even when a plurality of players share a common mixed reality space. According to this invention, markers to be used by only a given player are laid out at positions that cannot be seen from other players. Real objects used in an application that uses a mixed reality space may be used.Type: GrantFiled: February 18, 2000Date of Patent: August 1, 2006Assignee: Canon Kabushiki KaishaInventors: Kiyohide Sato, Toshikazu Ohshima
-
Publication number: 20060167954Abstract: When a command is a command which requests to establish a communication path with a server apparatus, the communication path with the server apparatus is established (S603), and a sharing mode is set in a public mode (S604). When a command is a command which requests to disconnect the communication path with the server apparatus, the communication path with the server apparatus is disconnected (S606), and the sharing mode is set in a private mode (S607). When a command is a database manipulation command, an event indicating the contents of the manipulation command is generated (S608). If the sharing mode is a public mode, the event is input to a transceiver queue (S610). If the sharing mode is a private mode, the event is input to a receiver queue (409) (S611).Type: ApplicationFiled: February 27, 2004Publication date: July 27, 2006Applicant: Canon Kabushiki KaishaInventors: Masakazu Fujiki, Toshikazu Ohshima
-
Patent number: 7056216Abstract: There is disclosed a user interface apparatus to which the user (player) can easily sensuoulsy become accustomed, and which can accurately recognize instructions (commands) that the user (player) intended. The user interface apparatus has a location/posture sensor (100) attached to the head of the user, and a location/posture sensor (200) attached to his or her hand, analyzes hand action on the basis of the relative position of the location/posture of the hand with respect to those of the head, and estimates the current action state from the analysis result. The apparatus then determines a user instruction (player command) on the basis of the action state.Type: GrantFiled: February 22, 2000Date of Patent: June 6, 2006Assignee: Canon Kabushiki KaishaInventor: Toshikazu Ohshima
-
Publication number: 20060103591Abstract: There is provided a technology for controlling information to be supplied to an observer depending on a state of an apparatus, namely, whether the observer is receiving information. A use state of a head-mounted display device, for example, the position and/or orientation of the head-mounted display device is detected to control the power supply of the head-mounted display device based on the detected use state.Type: ApplicationFiled: November 10, 2005Publication date: May 18, 2006Applicant: Canon Kabushiki KaishaInventors: Kaname Tanimura, Takashi Aso, Toshikazu Ohshima
-
Publication number: 20060089974Abstract: In a system in which a plurality of processes connected via a network hold and use shared data to be shared by these processes, when a manipulation request of the shared data is generated, a manipulation event that represents the manipulation request is issued (T2501). Also, a flushing event is issued in response to a predetermined user's manipulation (T2503). When a manipulation event that responds to the issued manipulation event is received from a server, manipulation of the shared data is executed in accordance with this manipulation event. At this time, issuance of manipulation events based on manipulation requests is inhibited after the flushing event is issued until a flushing event corresponding to the issued flushing event is received (T2506). With this control, in manipulations which are made for the shared data to be shared by the plurality of processes, a subsequent manipulation command is issued after the previously issued manipulation is executed.Type: ApplicationFiled: February 25, 2004Publication date: April 27, 2006Applicant: CANON KABUSHIKI KAISHAInventors: Toshihiro Kobayashi, Toshikazu Ohshima, Masakazu Fujiki
-
Publication number: 20060075058Abstract: In an information processing method for maintaining, in a system in which each of a plurality of processes connected via a network holds and uses shared data to be shared by the processes, consistency of shared data held by the respective processes, when a manipulation request for the shared data is generated, an event indicating that manipulation request is output onto the network. An event as a reply to the event output onto the network is received from the network. A manipulation for the shared data is executed in accordance with the received event. In this way, the manipulation execution timings for shared data to be shared by a plurality of processes can be accurately matched.Type: ApplicationFiled: November 28, 2003Publication date: April 6, 2006Applicant: Canon Kabushiki KaishaInventors: Masakazu Fujiki, Toshikazu Ohshima
-
Publication number: 20060050087Abstract: A three-dimensional computer graphics model which represents a measurable range of a position and orientation sensor is prepared in advance. A composite image is generated by superimposing an image of a virtual object which represents the measurable range of the position and orientation sensor on an actually-sensed image in real space where the position and orientation sensor is placed. The composite image, when displayed, makes it possible to visually recognize the measurable range of the position and orientation sensor.Type: ApplicationFiled: September 2, 2005Publication date: March 9, 2006Applicant: CANON KABUSHIKI KAISHAInventors: Kaname Tanimura, Toshikazu Ohshima
-
Publication number: 20060044327Abstract: Whether or not a masking portion is to be masked is set for an image of each individual object. A first image of a virtual object which is set not to be masked is generated (step S305). A second image formed by masking a masking portion in an image of a virtual object which is set to be masked is generated. A third image formed by superimposing the first and second images is generated, and an image formed by superimposing the third image on an image on a physical space is generated (step S309).Type: ApplicationFiled: June 1, 2005Publication date: March 2, 2006Applicant: Canon Kabushiki KaishaInventors: Yasuhiro Okuno, Toshikazu Ohshima, Kaname Tanimura
-
Publication number: 20050285879Abstract: An information processing method of adjusting the position/orientation of a virtual object combined with a physical object includes steps of: acquiring position information of a virtual index defined on the virtual object; informing a user of the position of the virtual index; acquiring position information of a physical index on the physical object corresponding to the virtual index based on the position information of a pointing unit operated by the user; and defining position/orientation information of the virtual object based on the position information of the virtual index and the position information of the physical index.Type: ApplicationFiled: June 24, 2005Publication date: December 29, 2005Applicant: Canon Kabushiki KaishaInventors: Masahiro Suzuki, Toshikazu Ohshima, Toshihiro Kobayashi