Patents by Inventor Scott Merritt
Scott Merritt has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 9402691Abstract: An image guidance system for tracking a surgical instrument during a surgical procedure. The image guidance system includes a plurality of cameras adapted to be located external to a surgical area for capturing images of optically visible patterns. A processing system receives and processes the images to recognize the patterns and triangulate the locations and orientations of each pattern relative to a camera. The processing system uses a reference dataset which defines a reference coordinate system based on alignment to a portion of the oral anatomy. The processing system determines the location and orientation of the tracked instrument based on the reference dataset.Type: GrantFiled: September 16, 2014Date of Patent: August 2, 2016Assignee: X-Nav Technologies, LLCInventors: Scott A. Merritt, Robert W. Emery, III, Edward J. Marandola, Christopher W. Scharff
-
Publication number: 20160151117Abstract: A visual guidance system for use in a surgical procedure. The system includes an image processing unit programmed to receive a three-dimensional image of a surgical area, information on an instrument being used in the procedure, and a surgical plan including a planned trajectory that the instrument should follow. The image processing unit generates a derived representation of the surgical procedure depicting the movement of the instrument on the three-dimensional image. A graphical guidance indicator is disposed on the three-dimensional image with a center point of the graphical guidance indicator centered on the planned trajectory, the graphical guidance indicator being displayed orthogonal to the planned trajectory. The derived image and the graphical guidance indicator depict at least five degrees of freedom of motion of the instrument. The graphical guidance indicator is changed based on the movement of the instrument in accordance with the surgical plan.Type: ApplicationFiled: December 2, 2014Publication date: June 2, 2016Applicant: X-NAV TECHNOLOGIES, INC.Inventors: Jason Gibbs, Scott A. Merritt, Edward J. Marandola, Christopher W. Scharff, Glenn A. Straub, Robert W. Emery, III
-
Publication number: 20160074127Abstract: An image guidance system for tracking a surgical instrument during a surgical procedure. The image guidance system includes a plurality of cameras adapted to be located external to a surgical area for capturing images of optically visible patterns. A processing system receives and processes the images to recognize the patterns and triangulate the locations and orientations of each pattern relative to a camera. The processing system uses a reference dataset to recognize the patterns.Type: ApplicationFiled: September 16, 2014Publication date: March 17, 2016Inventors: Scott A. Merritt, Robert W. Emery, III, Edward J. Marandola, Christopher W. Scharff
-
Publication number: 20160074129Abstract: An image guidance system for tracking a surgical instrument during a surgical procedure. The image guidance system includes a plurality of cameras adapted to be located external to a surgical area for capturing images of optically visible patterns. A processing system receives and processes the images to recognize the patterns and triangulate the locations and orientations of each pattern relative to a camera. The processing system uses a reference dataset which defines a reference coordinate system based on alignment to a portion of the oral anatomy. The processing system determines the location and orientation of the tracked instrument based on the reference dataset.Type: ApplicationFiled: September 16, 2014Publication date: March 17, 2016Inventors: Scott A. Merritt, Robert W. Emery, III, Edward J. Marandola, Christopher W. Scharff
-
Publication number: 20140272773Abstract: An image guidance system for tracking a surgical instrument within the oral cavity. The image guidance system includes a plurality of cameras adapted to be located within the oral cavity to provide intraoral images of optically visible patterns within oral cavity. A processing system receives and processes the intraoral images to recognize patterns and triangulate the locations and orientations of each camera. The processing system uses a reference dataset which defines a reference coordinate system based on alignment to a portion of the oral anatomy. The processing system determines the location and orientation of the tracked instrument based on the reference dataset. In an embodiment, the system includes an oral fixture that is removably attachable to teeth in a patient and is configured to hold one of the cameras.Type: ApplicationFiled: March 13, 2014Publication date: September 18, 2014Applicant: X-Nav Technologies, LLCInventors: Scott A. Merritt, Robert W. Emery, Edward J. Marandola, Christopher W. Scharff
-
Patent number: 8675935Abstract: Fast and continuous registration between two imaging modalities makes it possible to completely determine the rigid transformation between multiple sources at real-time or near real-time frame-rates in order to localize video cameras and register the two sources. A set of reference images are computed or captured within a known environment, with corresponding depth maps and image gradients defining a reference source. Given one frame from a real-time or near-real time video feed, and starting from an initial guess of viewpoint, a real-time video frame is warped to the nearest viewing site of the reference source. An image difference is computed between the warped video frame and the reference image. Steps are repeated for each frame until the viewpoint converges or the next video frame becomes available. The final viewpoint gives an estimate of the relative rotation and translation between the camera at that particular video frame and the reference source.Type: GrantFiled: November 16, 2011Date of Patent: March 18, 2014Assignee: The Penn State Research FoundationInventors: William E. Higgins, Scott A. Merritt, Lav Rai
-
Patent number: 8672836Abstract: Methods and apparatus provide continuous guidance of endoscopy during a live procedure. A data-set based on 3D image data is pre-computed including reference information representative of a predefined route through a body organ to a final destination. A plurality of live real endoscopic (RE) images are displayed as an operator maneuvers an endoscope within the body organ. A registration and tracking algorithm registers the data-set to one or more of the RE images and continuously maintains the registration as the endoscope is locally maneuvered. Additional information related to the final destination is then presented enabling the endoscope operator to decide on a final maneuver for the procedure. The reference information may include 3D organ surfaces, 3D routes through an organ system, or 3D regions of interest (ROIs), as well as a virtual endoscopic (VE) image generated from the precomputed data-set.Type: GrantFiled: January 30, 2008Date of Patent: March 18, 2014Assignee: The Penn State Research FoundationInventors: William E. Higgins, Scott A. Merritt, Lav Rai, Jason D. Gibbs, Kun-Chang Yu
-
Publication number: 20120082351Abstract: Fast and continuous registration between two imaging modalities makes it possible to completely determine the rigid transformation between multiple sources at real-time or near real-time frame-rates in order to localize video cameras and register the two sources. A set of reference images are computed or captured within a known environment, with corresponding depth maps and image gradients defining a reference source. Given one frame from a real-time or near-real time video feed, and starting from an initial guess of viewpoint, a real-time video frame is warped to the nearest viewing site of the reference source. An image difference is computed between the warped video frame and the reference image. Steps are repeated for each frame until the viewpoint converges or the next video frame becomes available. The final viewpoint gives an estimate of the relative rotation and translation between the camera at that particular video frame and the reference source.Type: ApplicationFiled: November 16, 2011Publication date: April 5, 2012Applicant: The Penn State Research FoundationInventors: William E. Higgins, Scott A. Merritt, Lav Rai
-
Patent number: 8064669Abstract: A novel framework for fast and continuous registration between two imaging modalities is disclosed. The approach makes it possible to completely determine the rigid transformation between multiple sources at real-time or near real-time frame-rates in order to localize the cameras and register the two sources. A disclosed example includes computing or capturing a set of reference images within a known environment, complete with corresponding depth maps and image gradients. The collection of these images and depth maps constitutes the reference source. The second source is a real-time or near-real time source which may include a live video feed. Given one frame from this video feed, and starting from an initial guess of viewpoint, the real-time video frame is warped to the nearest viewing site of the reference source. An image difference is computed between the warped video frame and the reference image.Type: GrantFiled: February 7, 2011Date of Patent: November 22, 2011Assignee: The Penn State Research FoundationInventors: William E. Higgins, Scott A. Merritt, Lav Rai
-
Publication number: 20110184238Abstract: A global registration system and method identifies bronchoscope position without the need for significant bronchoscope maneuvers, technician intervention, or electromagnetic sensors. Virtual bronchoscopy (VB) renderings of a 3D airway tree are obtained including VB views of branch positions within the airway tree. At least one real bronchoscopic (RB) video frame is received from a bronchoscope inserted into the airway tree. An algorithm according to the invention is executed on a computer to identify the several most likely branch positions having a VB view closest to the received RB view, and the 3D position of the bronchoscope within the airway tree is determined in accordance with the branch position identified in the VB view. The preferred embodiment involves a fast local registration search over all the branches in a global airway-bifurcation search space, with the weighted normalized sum of squares distance metric used for finding the best match.Type: ApplicationFiled: January 28, 2011Publication date: July 28, 2011Applicant: The Penn State Research FoundationInventors: William E. Higgins, Rahul Khare, Scott A. Merritt
-
Publication number: 20110128352Abstract: A novel framework for fast and continuous registration between two imaging modalities is disclosed. The approach makes it possible to completely determine the rigid transformation between multiple sources at real-time or near real-time frame-rates in order to localize the cameras and register the two sources. A disclosed example includes computing or capturing a set of reference images within a known environment, complete with corresponding depth maps and image gradients. The collection of these images and depth maps constitutes the reference source. The second source is a real-time or near-real time source which may include a live video feed. Given one frame from this video feed, and starting from an initial guess of viewpoint, the real-time video frame is warped to the nearest viewing site of the reference source. An image difference is computed between the warped video frame and the reference image.Type: ApplicationFiled: February 7, 2011Publication date: June 2, 2011Applicant: The Penn State Research FoundationInventors: William E. Higgins, Scott A. Merritt, Lav Rai
-
Patent number: 7889905Abstract: A novel framework for fast and continuous registration between two imaging modalities is disclosed. The approach makes it possible to completely determine the rigid transformation between multiple sources at real-time or near real-time frame-rates in order to localize the cameras and register the two sources. A disclosed example includes computing or capturing a set of reference images within a known environment, complete with corresponding depth maps and image gradients. The collection of these images and depth maps constitutes the reference source. The second source is a real-time or near-real time source which may include a live video feed. Given one frame from this video feed, and starting from an initial guess of viewpoint, the real-time video frame is warped to the nearest viewing site of the reference source. An image difference is computed between the warped video frame and the reference image.Type: GrantFiled: May 19, 2006Date of Patent: February 15, 2011Assignee: The Penn State Research FoundationInventors: William E. Higgins, Scott A. Merritt, Lav Rai
-
Publication number: 20100280365Abstract: A method provides guidance to the physician during a live bronchoscopy or other endoscopic procedures. The 3D motion of the bronchoscope is estimated using a fast coarse tracking step followed by a fine registration step. The tracking is based on finding a set of corresponding feature points across a plurality of consecutive bronchoscopic video frames, then estimating for the new pose of the bronchoscope. In the preferred embodiment the pose estimation is based on linearization of the rotation matrix. By giving a set of corresponding points across the current bronchoscopic video image, and the CT-based virtual image as an input, the same method can also be used for manual registration. The fine registration step is preferably a gradient-based Gauss-Newton method that maximizes the correlation between the bronchoscopic video image and the CT-based virtual image. The continuous guidance is provided by estimating the 3D motion of the bronchoscope in a loop.Type: ApplicationFiled: July 12, 2010Publication date: November 4, 2010Applicant: The Penn State Research FoundationInventors: William E. Higgins, Scott A. Merritt, Lav Rai
-
Patent number: 7756563Abstract: A method provides guidance to the physician during a live bronchoscopy or other endoscopic procedures. The 3D motion of the bronchoscope is estimated using a fast coarse tracking step followed by a fine registration step. The tracking is based on finding a set of corresponding feature points across a plurality of consecutive bronchoscopic video frames, then estimating for the new pose of the bronchoscope. In the preferred embodiment the pose estimation is based on linearization of the rotation matrix. By giving a set of corresponding points across the current bronchoscopic video image, and the CT-based virtual image as an input, the same method can also be used for manual registration. The fine registration step is preferably a gradient-based Gauss-Newton method that maximizes the correlation between the bronchoscopic video image and the CT-based virtual image. The continuous guidance is provided by estimating the 3D motion of the bronchoscope in a loop.Type: GrantFiled: May 19, 2006Date of Patent: July 13, 2010Assignee: The Penn State Research FoundationInventors: William E. Higgins, Scott A. Merritt, Lav Rai
-
Publication number: 20080207997Abstract: Methods and apparatus provide continuous guidance of endoscopy during a live procedure. A data-set based on 3D image data is pre-computed including reference information representative of a predefined route through a body organ to a final destination. A plurality of live real endoscopic (RE) images are displayed as an operator maneuvers an endoscope within the body organ. A registration and tracking algorithm registers the data-set to one or more of the RE images and continuously maintains the registration as the endoscope is locally maneuvered. Additional information related to the final destination is then presented enabling the endoscope operator to decide on a final maneuver for the procedure. The reference information may include 3D organ surfaces, 3D routes through an organ system, or 3D regions of interest (ROIs), as well as a virtual endoscopic (VE) image generated from the precomputed data-set.Type: ApplicationFiled: January 30, 2008Publication date: August 28, 2008Applicant: The Penn State Research FoundationInventors: William E. Higgins, Scott A. Merritt, Lav Rai, Jason D. Gibbs, Kun-Chang Yu
-
Patent number: 7203409Abstract: Systems and methods according to the present invention address this need and others by providing SLD devices and methods for generating optical energy that reduce internal reflections without the use of an absorber region. This can be accomplished by, among other things, adapting the waveguide geometry to dump reflections from the front facet out through the back facet of the device.Type: GrantFiled: August 16, 2004Date of Patent: April 10, 2007Assignee: Covega CorporationInventors: Scott A. Merritt, Peter J. S. Heim
-
Publication number: 20070015997Abstract: A method provides guidance to the physician during a live bronchoscopy or other endoscopic procedures. The 3D motion of the bronchoscope is estimated using a fast coarse tracking step followed by a fine registration step. The tracking is based on finding a set of corresponding feature points across a plurality of consecutive bronchoscopic video frames, then estimating for the new pose of the bronchoscope. In the preferred embodiment the pose estimation is based on linearization of the rotation matrix. By giving a set of corresponding points across the current bronchoscopic video image, and the CT-based virtual image as an input, the same method can also be used for manual registration. The fine registration step is preferably a gradient-based Gauss-Newton method that maximizes the correlation between the bronchoscopic video image and the CT-based virtual image. The continuous guidance is provided by estimating the 3D motion of the bronchoscope in a loop.Type: ApplicationFiled: May 19, 2006Publication date: January 18, 2007Inventors: William Higgins, Scott Merritt, Lav Rai
-
Publication number: 20070013710Abstract: A novel framework for fast and continuous registration between two imaging modalities is disclosed. The approach makes it possible to completely determine the rigid transformation between multiple sources at real-time or near real-time frame-rates in order to localize the cameras and register the two sources. A disclosed example includes computing or capturing a set of reference images within a known environment, complete with corresponding depth maps and image gradients. The collection of these images and depth maps constitutes the reference source. The second source is a real-time or near-real time source which may include a live video feed. Given one frame from this video feed, and starting from an initial guess of viewpoint, the real-time video frame is warped to the nearest viewing site of the reference source. An image difference is computed between the warped video frame and the reference image.Type: ApplicationFiled: May 19, 2006Publication date: January 18, 2007Inventors: William Higgins, Scott Merritt, Lav Rai
-
Patent number: 7158291Abstract: A semiconductor optical amplifier (SOA) has an overall gain that is substantially polarization independent, i.e., less than 1 dB difference between transverse electric (TE) and transverse magnetic (TM) gain. The SOA includes a residual cladding layer having different thicknesses over different portions of the gain section. Over a first portion of the gain section, the residual cladding layer is thinner than over a second portion of the gain section. This results in the first portion providing more gain to optical energy having a TE polarization state than optical energy having a TM polarization state. In the second portion of the gain section, however, more gain is provided to optical energy having a TM polarization state than energy having a TE polarization state. The resulting gain differences can be designed to offset one another so that the output has a gain that is substantially polarization independent.Type: GrantFiled: January 29, 2004Date of Patent: January 2, 2007Assignee: Quantum Photonics, Inc.Inventors: Simarjeet S. Saini, Peter J. S. Heim, Scott A. Merritt, Mario Dagenais
-
Patent number: 7089108Abstract: An apparatus for entering a flight plan into an aircraft navigation system, the apparatus comprising: an acoustic sampler adapted for sampling a microphone signal and generating an acoustic signal; a waypoint identifier adapted for generating an identified waypoint from the acoustic signal and the flight plan; and a navigation interface adapted for incorporating the identified waypoint into the flight plan and for transmitting and receiving navigation data to and from the aircraft navigation system.Type: GrantFiled: March 12, 2004Date of Patent: August 8, 2006Assignee: VoiceFlight Systems, Inc.Inventor: J. Scott Merritt