Patents by Inventor Toshikazu Ohshima

Toshikazu Ohshima has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 7330197
    Abstract: In exhibition of a synthesized image which is obtained by synthesizing a virtual world image with a real world image observed from a viewpoint position and direction of a user, data representing a position and orientation of a user is acquired, a virtual image is generated based on the data representing the position and orientation of the user, and the virtual image is synthesized with a real image corresponding to the position and orientation of the user. Based on a measurable area of the position and orientation of the user, area data is set. Based on the data representing the position of the user and the area data, notification related to the measurable area is controlled.
    Type: Grant
    Filed: December 3, 2004
    Date of Patent: February 12, 2008
    Assignee: Canon Kabushiki Kaisha
    Inventors: Toshihiro Kobayashi, Toshikazu Ohshima
  • Publication number: 20080030499
    Abstract: In a mixed-reality presentation system, a handler for manipulating a virtual object is displayed as a virtual object. Then, manipulation of the handler by the hand(s) of an observer is detected as manipulation of the virtual object associated with the handler, and those detection results are reflected in a display of the manipulated virtual object. By dynamically changing virtual objects displaying the handler according to the size of the virtual objects displayed in a view volume of the observer presented with a mixed reality space, the observer can select a desired virtual object with ease.
    Type: Application
    Filed: July 24, 2007
    Publication date: February 7, 2008
    Applicant: CANON KABUSHIKI KAISHA
    Inventors: Koichiro Wanda, Toshikazu Ohshima
  • Publication number: 20080024523
    Abstract: A method for combining a real space image with a virtual image, includes: causing an imaging unit to capture an image of a real space; generating an image covering a predetermined space based on a plurality of real images in the captured real space; extracting position information of a light source based on the generated image; and adding a light source or a shadow on the virtual image based on the extracted position information of the light source.
    Type: Application
    Filed: July 23, 2007
    Publication date: January 31, 2008
    Applicant: CANON KABUSHIKI KAISHA
    Inventors: Kaname Tomite, Toshikazu Ohshima
  • Publication number: 20070258658
    Abstract: An information processing apparatus connected to a display apparatus, the information processing apparatus comprises an acquisition unit adapted to acquire a real image captured by an imaging apparatus; a generation unit adapted to generate a virtual image based on an orientation of said imaging apparatus; a detection unit adapted to detect a tilt of the real image; a correction unit adapted to correct the real image based on the detected tilt of said real image; a composite unit adapted to generate a composite image based on the corrected real image and the virtual image; and an output unit adapted to output the corrected composite image to said display apparatus.
    Type: Application
    Filed: April 25, 2007
    Publication date: November 8, 2007
    Inventors: Toshihiro Kobayashi, Toshikazu Ohshima
  • Patent number: 7289130
    Abstract: A game state manager (201) manages the state of an AR game (information that pertains to rendering of each virtual object (102), the score of a player (101), the AR game round count, and the like). An objective viewpoint video generator (202) generates a video of each virtual object (102) viewed from a camera (103). An objective viewpoint video composition unit (203) generates a composite video of the video of the virtual object (102) and an actually sensed video, and outputs it to a display (106). A subjective viewpoint video generator (212) generates a video of the virtual object (102) viewed from an HMD (107). A subjective viewpoint video composition unit (213) generates a composite video of the video of the virtual object (102) and an actually sensed video, and outputs it to the HMD (107).
    Type: Grant
    Filed: September 8, 2000
    Date of Patent: October 30, 2007
    Assignee: Canon Kabushiki Kaisha
    Inventors: Kiyohide Satoh, Toshikazu Ohshima
  • Publication number: 20070236510
    Abstract: An acquired real image undergoes conversion based on a first parameter to generate an image corresponding to a designated characteristic. A virtual image is generated on the basis of a second parameter to generate an image corresponding to a designated characteristic. The converted real image is composited with the generated virtual image. The composite image is displayed.
    Type: Application
    Filed: March 30, 2007
    Publication date: October 11, 2007
    Inventors: Hiroyuki Kakuta, Toshikazu Ohshima
  • Patent number: 7212687
    Abstract: When a CG object is overlapped onto a photographed real scenery image as a background and an image of a photographed real photographic subject is synthesized in front of the overlapped image, an image in which the photographed real background and the photographic subject are imaged simultaneously without imaging the photographed real background and the photographic subject independently, is synthesized with CG. A photographed real image including a specific image is acquired, and area information representing an area of the specific image from the photographed real image, and an area other than the area of the specific image of the photographed real image is synthesized with a computer graphics image using the area information so detected.
    Type: Grant
    Filed: March 27, 2003
    Date of Patent: May 1, 2007
    Assignee: Canon Kabushiki Kaisha
    Inventor: Toshikazu Ohshima
  • Patent number: 7199807
    Abstract: A set of objects to be rendered by an identical rendering method is specified from objects which form a virtual space. A hierarchical structure formed by the object included in the specified set is generated. The objects included in the specified set are rendered by the rendering method common to the objects included in this set in accordance with the generated hierarchical structure.
    Type: Grant
    Filed: November 5, 2004
    Date of Patent: April 3, 2007
    Assignee: Canon Kabushiki Kaisha
    Inventors: Toshihiro Kobayashi, Toshikazu Ohshima
  • Publication number: 20060273984
    Abstract: When a plurality of mixed reality (MR) experiencing persons experience the same mixed reality, there is a possibility that the experiencing persons come into contact with one another. An image processing apparatus capable of reporting to the MR experiencing persons that there is a possibility of contact. Therefore, a real space image that is image captured from the position and orientation of a user's viewpoint is drawn. The position and orientation of the user's viewpoint at this time is detected by a sensor unit. It is determined whether or not the viewpoint position is smaller than or equal to an attention distance at which there is a possibility of contact. If the viewpoint position is smaller than or equal to the attention distance, an attention display for this fact is performed.
    Type: Application
    Filed: April 7, 2006
    Publication date: December 7, 2006
    Applicant: Canon Kabushiki Kaisha
    Inventors: Koichiro Wanda, Toshikazu Ohshima
  • Patent number: 7123214
    Abstract: It is intended to solve the problem that, when an operator performs an operation using two-dimensional input means while seeing a picture frame seen from a viewpoint of the operator, an input operation is difficult to perform when the direction of the line of sight of the operator is unstable, since a designated direction greatly deviates in a world coordinate system. For that purpose, viewpoint information is detected, and an instruction of the operator for operating a position of a pointer image is input. A designated direction in a pointer coordinate system is obtained in accordance with the operator's instruction, and the pointer image is generated based on the designated direction. The pointer coordinate system is changed from the detected viewpoint information in accordance with a specific instruction of the operator.
    Type: Grant
    Filed: March 25, 2003
    Date of Patent: October 17, 2006
    Assignee: Canon Kabushiki Kaisha
    Inventors: Tsuyoshi Kuroki, Toshikazu Ohshima
  • Publication number: 20060202985
    Abstract: A viewpoint position/orientation is acquired (S1020). An area of an environment mapping image to be used for environment mapping on the environment mapping target area of a virtual object is specified on the basis of the viewpoint orientation (S1040 and S1060). An image in the specified area is environment-mapped on the environment mapping target area (S1060). An image is generated by seeing, from the viewpoint, a virtual space including the virtual object which has undergone environment mapping.
    Type: Application
    Filed: March 7, 2006
    Publication date: September 14, 2006
    Applicant: CANON KABUSHIKI KAISHA
    Inventors: Toshihiro Kobayashi, Toshikazu Ohshima
  • Patent number: 7084887
    Abstract: There is disclosed a marker layout method in a mixed reality space, which can reliably detect markers in units of players, even when a plurality of players share a common mixed reality space. According to this invention, markers to be used by only a given player are laid out at positions that cannot be seen from other players. Real objects used in an application that uses a mixed reality space may be used.
    Type: Grant
    Filed: February 18, 2000
    Date of Patent: August 1, 2006
    Assignee: Canon Kabushiki Kaisha
    Inventors: Kiyohide Sato, Toshikazu Ohshima
  • Publication number: 20060167954
    Abstract: When a command is a command which requests to establish a communication path with a server apparatus, the communication path with the server apparatus is established (S603), and a sharing mode is set in a public mode (S604). When a command is a command which requests to disconnect the communication path with the server apparatus, the communication path with the server apparatus is disconnected (S606), and the sharing mode is set in a private mode (S607). When a command is a database manipulation command, an event indicating the contents of the manipulation command is generated (S608). If the sharing mode is a public mode, the event is input to a transceiver queue (S610). If the sharing mode is a private mode, the event is input to a receiver queue (409) (S611).
    Type: Application
    Filed: February 27, 2004
    Publication date: July 27, 2006
    Applicant: Canon Kabushiki Kaisha
    Inventors: Masakazu Fujiki, Toshikazu Ohshima
  • Patent number: 7056216
    Abstract: There is disclosed a user interface apparatus to which the user (player) can easily sensuoulsy become accustomed, and which can accurately recognize instructions (commands) that the user (player) intended. The user interface apparatus has a location/posture sensor (100) attached to the head of the user, and a location/posture sensor (200) attached to his or her hand, analyzes hand action on the basis of the relative position of the location/posture of the hand with respect to those of the head, and estimates the current action state from the analysis result. The apparatus then determines a user instruction (player command) on the basis of the action state.
    Type: Grant
    Filed: February 22, 2000
    Date of Patent: June 6, 2006
    Assignee: Canon Kabushiki Kaisha
    Inventor: Toshikazu Ohshima
  • Publication number: 20060103591
    Abstract: There is provided a technology for controlling information to be supplied to an observer depending on a state of an apparatus, namely, whether the observer is receiving information. A use state of a head-mounted display device, for example, the position and/or orientation of the head-mounted display device is detected to control the power supply of the head-mounted display device based on the detected use state.
    Type: Application
    Filed: November 10, 2005
    Publication date: May 18, 2006
    Applicant: Canon Kabushiki Kaisha
    Inventors: Kaname Tanimura, Takashi Aso, Toshikazu Ohshima
  • Publication number: 20060089974
    Abstract: In a system in which a plurality of processes connected via a network hold and use shared data to be shared by these processes, when a manipulation request of the shared data is generated, a manipulation event that represents the manipulation request is issued (T2501). Also, a flushing event is issued in response to a predetermined user's manipulation (T2503). When a manipulation event that responds to the issued manipulation event is received from a server, manipulation of the shared data is executed in accordance with this manipulation event. At this time, issuance of manipulation events based on manipulation requests is inhibited after the flushing event is issued until a flushing event corresponding to the issued flushing event is received (T2506). With this control, in manipulations which are made for the shared data to be shared by the plurality of processes, a subsequent manipulation command is issued after the previously issued manipulation is executed.
    Type: Application
    Filed: February 25, 2004
    Publication date: April 27, 2006
    Applicant: CANON KABUSHIKI KAISHA
    Inventors: Toshihiro Kobayashi, Toshikazu Ohshima, Masakazu Fujiki
  • Publication number: 20060075058
    Abstract: In an information processing method for maintaining, in a system in which each of a plurality of processes connected via a network holds and uses shared data to be shared by the processes, consistency of shared data held by the respective processes, when a manipulation request for the shared data is generated, an event indicating that manipulation request is output onto the network. An event as a reply to the event output onto the network is received from the network. A manipulation for the shared data is executed in accordance with the received event. In this way, the manipulation execution timings for shared data to be shared by a plurality of processes can be accurately matched.
    Type: Application
    Filed: November 28, 2003
    Publication date: April 6, 2006
    Applicant: Canon Kabushiki Kaisha
    Inventors: Masakazu Fujiki, Toshikazu Ohshima
  • Publication number: 20060050087
    Abstract: A three-dimensional computer graphics model which represents a measurable range of a position and orientation sensor is prepared in advance. A composite image is generated by superimposing an image of a virtual object which represents the measurable range of the position and orientation sensor on an actually-sensed image in real space where the position and orientation sensor is placed. The composite image, when displayed, makes it possible to visually recognize the measurable range of the position and orientation sensor.
    Type: Application
    Filed: September 2, 2005
    Publication date: March 9, 2006
    Applicant: CANON KABUSHIKI KAISHA
    Inventors: Kaname Tanimura, Toshikazu Ohshima
  • Publication number: 20060044327
    Abstract: Whether or not a masking portion is to be masked is set for an image of each individual object. A first image of a virtual object which is set not to be masked is generated (step S305). A second image formed by masking a masking portion in an image of a virtual object which is set to be masked is generated. A third image formed by superimposing the first and second images is generated, and an image formed by superimposing the third image on an image on a physical space is generated (step S309).
    Type: Application
    Filed: June 1, 2005
    Publication date: March 2, 2006
    Applicant: Canon Kabushiki Kaisha
    Inventors: Yasuhiro Okuno, Toshikazu Ohshima, Kaname Tanimura
  • Publication number: 20050285879
    Abstract: An information processing method of adjusting the position/orientation of a virtual object combined with a physical object includes steps of: acquiring position information of a virtual index defined on the virtual object; informing a user of the position of the virtual index; acquiring position information of a physical index on the physical object corresponding to the virtual index based on the position information of a pointing unit operated by the user; and defining position/orientation information of the virtual object based on the position information of the virtual index and the position information of the physical index.
    Type: Application
    Filed: June 24, 2005
    Publication date: December 29, 2005
    Applicant: Canon Kabushiki Kaisha
    Inventors: Masahiro Suzuki, Toshikazu Ohshima, Toshihiro Kobayashi