Patents by Inventor Ryan M. Uhlenbrock

Ryan M. Uhlenbrock has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20220375222
    Abstract: Described is a system and method for accurate image and/or video scene classification. More specifically, described is a system that makes use of a specialized convolutional-neural network (hereafter CNN) based technique for the fusion of bottom-up whole-image features and top-down entity classification. When the two parallel and independent processing paths are fused, the system provides an accurate classification of the scene as depicted in the image or video.
    Type: Application
    Filed: July 14, 2022
    Publication date: November 24, 2022
    Inventors: Ryan M. Uhlenbrock, Deepak Khosla, Yang Chen, Fredy Monterroza
  • Patent number: 11423651
    Abstract: Described is a system and method for accurate image and/or video scene classification. More specifically, described is a system that makes use of a specialized convolutional-neural network (hereafter CNN) based technique for the fusion of bottom-up whole-image features and top-down entity classification. When the two parallel and independent processing paths are fused, the system provides an accurate classification of the scene as depicted in the image or video.
    Type: Grant
    Filed: February 8, 2017
    Date of Patent: August 23, 2022
    Assignee: HRL LABORATORIES, LLC
    Inventors: Ryan M. Uhlenbrock, Deepak Khosla, Yang Chen, Fredy Monterroza
  • Patent number: 11227162
    Abstract: Described is a system for activity and behavior detection in a target system. Raw data extracted from various heterogeneous sources of the target system is fused across spatial and temporal scales into a multi-graph representation. Information flows of the multi-graph representation are analyzed using a set of multi-layer information dynamic measures. Based on the set of multi-layer information dynamic measures, at least one of an economic and social indicator of emerging activity of interest in the target system is derived. The indicator is then used for prediction of future activity of interest in the target system.
    Type: Grant
    Filed: April 25, 2017
    Date of Patent: January 18, 2022
    Assignee: HRL Laboratories, LLC
    Inventors: Tsai-Ching Lu, Kang-Yu Ni, Ryan M. Uhlenbrock
  • Patent number: 11055872
    Abstract: Described is a system for real-time object recognition. The system extracts a candidate target region representing a candidate object from an input image of a scene based on agglomeration of channel features. The candidate target region is classified using a trained convolutional neural network (CNN) classifier, resulting in an initial classified object. A multi-target tracker is used for tracking the classified objects for final classification of each classified object, resulting in a final output, and a device is controlled based on the final output.
    Type: Grant
    Filed: January 30, 2018
    Date of Patent: July 6, 2021
    Assignee: HRL Laboratories, LLC
    Inventors: Yang Chen, Deepak Khosla, Ryan M. Uhlenbrock
  • Patent number: 10997421
    Abstract: Described is a system for visual activity recognition that includes one or more processors and a memory, the memory being a non-transitory computer-readable medium having executable instructions encoded thereon, such that upon execution of the instructions, the one or more processors perform operations including detecting a set of objects of interest in video data and determining an object classification for each object in the set of objects of interest, the set including at least one object of interest. The one or more processors further perform operations including forming a corresponding activity track for each object in the set of objects of interest by tracking each object across frames. The one or more processors further perform operations including, for each object of interest and using a feature extractor, determining a corresponding feature in the video data. The system may provide a report to a user's cell phone or central monitoring facility.
    Type: Grant
    Filed: April 6, 2018
    Date of Patent: May 4, 2021
    Assignee: HRL Laboratories, LLC
    Inventors: Deepak Khosla, Ryan M. Uhlenbrock, Yang Chen
  • Patent number: 10909407
    Abstract: Described is a system for converting a convolutional neural network (CNN) designed and trained for color (RGB) images to one that works on infrared (IR) or grayscale images. The converted CNN comprises a series of convolution layers of neurons arranged in a set kernels having corresponding depth slices. The converted CNN is used for performing object detection. A mechanical component of an autonomous device is controlled based on the object detection.
    Type: Grant
    Filed: March 23, 2018
    Date of Patent: February 2, 2021
    Assignee: HRL Laboratories, LLC
    Inventors: Ryan M. Uhlenbrock, Yang Chen, Deepak Khosla
  • Patent number: 10891488
    Abstract: Described is a system for visual activity recognition. In operation, the system detects a set of objects of interest (OI) in video data and determines an object classification for each object in the set of OI, the set including at least one OI. A corresponding activity track is formed for each object in the set of OI by tracking each object across frames. Using a feature extractor, the system determines a corresponding feature in the video data for each OI, which is then used to determine a corresponding initial activity classification for each OI. One or more OI are then detected in each activity track via foveation, with the initial object detection and foveated object detection thereafter being appended into a new detected-objects list. Finally, a final classification is provided for each activity track using the new detected-objects list and filtering the initial activity classification results using contextual logic.
    Type: Grant
    Filed: January 14, 2019
    Date of Patent: January 12, 2021
    Assignee: HRL Laboratories, LLC
    Inventors: Deepak Khosla, Ryan M. Uhlenbrock, Huapeng Su, Yang Chen
  • Patent number: 10861173
    Abstract: A method includes generating, based on first surface data that includes three-dimensional (3D) point positions corresponding to a first portion of a surface of an object, first hole data that indicates first positions of holes in the first portion of the surface. The method includes generating, based on second surface data corresponding to a second portion of the surface of the object, second hole data that indicates second positions of the holes in the second portion of the surface. The method also includes matching the first positions to the second positions to perform an alignment with respect to the first surface data and the second surface data.
    Type: Grant
    Filed: June 22, 2018
    Date of Patent: December 8, 2020
    Assignee: THE BOEING COMPANY
    Inventors: Ryan M. Uhlenbrock, Deepak Khosla, Anthony W. Baker
  • Patent number: 10803362
    Abstract: A system and method of automated classification of rock types includes: partitioning, by a processing device, an image into partitions; extracting, by the processing device, sub-images from each of the partitions; first-level classifying, by an automated classifier, the sub-images into corresponding first classes; and second-level classifying, by the processing device, the partitions into corresponding second classes by, for each partition of the partitions, selecting a most numerous one of the corresponding first classes of the sub-images extracted from the partition. A method of displaying automated classification results on a display device is provided. The method includes: receiving, by a processing device, an image partitioned into partitions and classified into corresponding classes; and manipulating, by the processing device, the display device to display the image together with visual identification of the partitions and their corresponding classes.
    Type: Grant
    Filed: June 19, 2019
    Date of Patent: October 13, 2020
    Assignee: HRL Laboratories, LLC
    Inventors: Yang Chen, Deepak Khosla, Fredy Monterroza, Ryan M. Uhlenbrock
  • Publication number: 20200219020
    Abstract: Described is a system for structuring rationales for collaborative forecasting between users of a crowdsourcing platform. For a given forecasting question, the system produces a forecasting rationale model from a combination of variables related to users and topics in a discussion of the users' forecasting rationale for making an initial forecast of an event. A relationship between the variables is determined, and based on the relationship between the variables, a prediction of each user's performance in making the initial forecast. Based on the predictions, top performing users and their forecasting rationales are selected, and the forecasting rationales of the top performing users are shared with other users of the crowdsourcing platform, allowing the other users to revise their initial forecasts in response to the shared forecasting rationales, resulting in revised forecasts. A forecast of the event that combines the revised forecasts is then output.
    Type: Application
    Filed: October 2, 2019
    Publication date: July 9, 2020
    Inventors: Robert Giaquinto, Tsai-Ching Lu, Aruna Jammalamadaka, Ryan M. Uhlenbrock
  • Patent number: 10699139
    Abstract: Described is an object recognition system. Using an integral channel features (ICF) detector, the system extracts a candidate target region (having an associated original confidence score representing a candidate object) from an input image of a scene surrounding a platform. A modified confidence score is generated based on a location and height of detection of the candidate object. The candidate target regions are classified based on the modified confidence score using a trained convolutional neural network (CNN) classifier, resulting in classified objects. The classified objects are tracked using a multi-target tracker for final classification of each classified object as a target or non-target. If the classified object is a target, a device can be controlled based on the target.
    Type: Grant
    Filed: February 14, 2019
    Date of Patent: June 30, 2020
    Assignee: HRL Laboratories, LLC
    Inventors: Yang Chen, Deepak Khosla, Ryan M. Uhlenbrock
  • Patent number: 10671917
    Abstract: Described is a system for neural decoding of neural activity. Using at least one neural feature extraction method, neural data that is correlated with a set of behavioral data is transformed into sparse neural representations. Semantic features are extracted from a set of semantic data. Using a combination of distinct classification modes, the set of semantic data is mapped to the sparse neural representations, and new input neural data can be interpreted.
    Type: Grant
    Filed: October 26, 2016
    Date of Patent: June 2, 2020
    Assignee: HRL Laboratories, LLC
    Inventors: Rajan Bhattacharyya, James Benvenuto, Vincent De Sapio, Michael J. O'Brien, Kang-Yu Ni, Kevin R. Martin, Ryan M. Uhlenbrock, Rachel Millin, Matthew E. Phillips, Hankyu Moon, Qin Jiang, Brian L. Burns
  • Patent number: 10540813
    Abstract: An apparatus includes a nearest neighbor search engine configured to receive multiple sets of surface data. Each of the multiple sets includes three-dimensional point positions of a corresponding portion of a surface of an object. The nearest neighbor search engine is configured to perform a nearest neighbor search based on adjacency data indicating pairs of overlapping sets of the multiple sets to identify, in each pair of overlapping sets, pairs of closest corresponding points that are within a threshold distance of each other. The apparatus includes a transformation matrix generator configured to determine one or more transformation matrices to reduce a global distance metric that corresponds to distances between the points of each identified pair of closest corresponding points. The apparatus also includes a data set aligner configured to generate updated 3D point positions by applying the one or more transformation matrices to one or more of the multiple sets.
    Type: Grant
    Filed: August 22, 2018
    Date of Patent: January 21, 2020
    Assignee: THE BOEING COMPANY
    Inventors: Ryan M. Uhlenbrock, Deepak Khosla
  • Publication number: 20190392595
    Abstract: A method includes generating, based on first surface data that includes three-dimensional (3D) point positions corresponding to a first portion of a surface of an object, first hole data that indicates first positions of holes in the first portion of the surface. The method includes generating, based on second surface data corresponding to a second portion of the surface of the object, second hole data that indicates second positions of the holes in the second portion of the surface. The method also includes matching the first positions to the second positions to perform an alignment with respect to the first surface data and the second surface data.
    Type: Application
    Filed: June 22, 2018
    Publication date: December 26, 2019
    Inventors: Ryan M. Uhlenbrock, Deepak Khosla, Anthony W. Baker
  • Patent number: 10410092
    Abstract: A system and method of automated classification of rock types includes: partitioning, by a processing device, an image into partitions; extracting, by the processing device, sub-images from each of the partitions; first-level classifying, by an automated classifier, the sub-images into corresponding first classes; and second-level classifying, by the processing device, the partitions into corresponding second classes by, for each partition of the partitions, selecting a most numerous one of the corresponding first classes of the sub-images extracted from the partition. A method of displaying automated classification results on a display device is provided. The method includes: receiving, by a processing device, an image partitioned into partitions and classified into corresponding classes; and manipulating, by the processing device, the display device to display the image together with visual identification of the partitions and their corresponding classes.
    Type: Grant
    Filed: December 15, 2016
    Date of Patent: September 10, 2019
    Assignee: HRL Laboratories, LLC
    Inventors: Yang Chen, Deepak Khosla, Fredy Monterroza, Ryan M. Uhlenbrock
  • Patent number: 10402699
    Abstract: A method for training an automated classifier of input images includes: receiving, by a processing device, a convolution neural network (CNN) model; receiving, by the processing device, training images and corresponding classes, each of the corresponding classes being associated with several ones of the training images; preparing, by the processing device, the training images, including separating the training images into a training set of the training images and a testing set of the training images; and training, by the processing device, the CNN model utilizing the training set, the testing set, and the corresponding classes to generate the automated classifier.
    Type: Grant
    Filed: December 15, 2016
    Date of Patent: September 3, 2019
    Assignee: HRL Laboratories, LLC
    Inventors: Yang Chen, Deepak Khosla, Fredy Monterroza, Ryan M. Uhlenbrock
  • Publication number: 20190251358
    Abstract: Described is a system for visual activity recognition. In operation, the system detects a set of objects of interest (OI) in video data and determines an object classification for each object in the set of OI, the set including at least one OI. A corresponding activity track is formed for each object in the set of OI by tracking each object across frames. Using a feature extractor, the system determines a corresponding feature in the video data for each OI, which is then used to determine a corresponding initial activity classification for each OI. One or more OI are then detected in each activity track via foveation, with the initial object detection and foveated object detection thereafter being appended into a new detected-objects list. Finally, a final classification is provided for each activity track using the new detected-objects list and filtering the initial activity classification results using contextual logic.
    Type: Application
    Filed: January 14, 2019
    Publication date: August 15, 2019
    Inventors: Deepak Khosla, Ryan M. Uhlenbrock, Huapeng Su, Yang Chen
  • Publication number: 20190180119
    Abstract: Described is an object recognition system. Using an integral channel features (ICF) detector, the system extracts a candidate target region (having an associated original confidence score representing a candidate object) from an input image of a scene surrounding a platform. A modified confidence score is generated based on a location and height of detection of the candidate object. The candidate target regions are classified based on the modified confidence score using a trained convolutional neural network (CNN) classifier, resulting in classified objects. The classified objects are tracked using a multi-target tracker for final classification of each classified object as a target or non-target. If the classified object is a target, a device can be controlled based on the target.
    Type: Application
    Filed: February 14, 2019
    Publication date: June 13, 2019
    Inventors: Yang Chen, Deepak Khosla, Ryan M. Uhlenbrock
  • Publication number: 20190005330
    Abstract: Described is a system and method for accurate image and/or video scene classification. More specifically, described is a system that makes use of a specialized convolutional-neural network (hereafter CNN) based technique for the fusion of bottom-up whole-image features and top-down entity classification. When the two parallel and independent processing paths are fused, the system provides an accurate classification of the scene as depicted in the image or video.
    Type: Application
    Filed: February 8, 2017
    Publication date: January 3, 2019
    Inventors: Ryan M. Uhlenbrock, Deepak Khosla, Yang Chen, Fredy Monterroza
  • Publication number: 20180341832
    Abstract: Described is a system for converting a convolutional neural network (CNN) designed and trained for color (RGB) images to one that works on infrared (IR) or grayscale images. The converted CNN comprises a series of convolution layers of neurons arranged in a set kernels having corresponding depth slices. The converted CNN is used for performing object detection. A mechanical component of an autonomous device is controlled based on the object detection.
    Type: Application
    Filed: March 23, 2018
    Publication date: November 29, 2018
    Inventors: Ryan M. Uhlenbrock, Yang Chen, Deepak Khosla