Patents by Inventor Anoop Cherian

Anoop Cherian has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240069501
    Abstract: A controller for controlling an entity is provided. The controller comprises a memory to store a hierarchical multimodal reinforcement learning (RL) neural network, and a processor. The hierarchical multimodal RL neural network includes a first level controller and two second level controllers. Each of the second level controllers comprise a first sub level controller relating to a first modality and a second sub level controller relating to a second modality. The first modality is different from the second modality. The processor is configured to select one of the two second level controllers to perform a first sub-task relating to a task, using the first level controller, based on input data and a state of the hierarchical multimodal RL neural network. The selected second level controller is configured to determine a set of control actions to perform the first sub-task, and control the entity based on the set of control actions.
    Type: Application
    Filed: August 30, 2022
    Publication date: February 29, 2024
    Inventors: Anoop Cherian Cherian, Sudipta Paul
  • Publication number: 20240046085
    Abstract: An artificial intelligence (AI) low-latency processing system is provided. The low-latency processing system includes a processor; and a memory having instructions stored thereon. The low-latency processing system is configured to collect a sequence of frames jointly including information dispersed among at least some frames in the sequence of frames, execute a timing neural network trained to identify an early subsequence of frames in the sequence of frames including at least a portion of the information indicative of the information, and execute a decoding neural network trained to decode the information from the portion of the information in the subsequence of frames, wherein the timing neural network is jointly trained with the decoding neural network to iteratively identify the smallest number of subframes from the beginning of a training sequence of frames containing a portion of training information sufficient to decode the training information.
    Type: Application
    Filed: August 4, 2022
    Publication date: February 8, 2024
    Applicant: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Chiori Hori, Jonathan Le Roux, Anoop Cherian, 02139 Marks
  • Patent number: 11809988
    Abstract: An artificial intelligence (AI) system that includes a processor configured to execute modules of the AI system. The modules comprise a feature extractor, an adversarial noise generator, a compressor and a classifier. The feature extractor is trained to process input data to extract features of the input data for classification of the input data. The adversarial noise generator is trained to generate noise data for distribution of features of the input data such that a misclassification rate of corrupted features that include the extracted features corrupted with the generated noise data is greater than a misclassification rate of the extracted features. The compressor is configured to compress the extracted features. The compressed features are closer to the extracted features than to the corrupted features. The classifier is trained to classify the compressed features.
    Type: Grant
    Filed: June 22, 2020
    Date of Patent: November 7, 2023
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Anoop Cherian, Aeron Shuchin
  • Publication number: 20230267614
    Abstract: An imaging controller is provided for segmenting instances from depth images including objects to be manipulated by a robot. The imaging controller includes an input interface configured to receive a depth image that includes objects, a memory configured to store instructions and a neural network trained to segment instances from the objects in the depth image, and a processor, coupled with the memory, configured to perform the instructions to segment a pickable instance using the trained neural network.
    Type: Application
    Filed: February 25, 2022
    Publication date: August 24, 2023
    Applicant: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Anoop Cherian, Tim Marks, Alan Sullivan
  • Patent number: 11663798
    Abstract: Present disclosure discloses an image processing system and method for manipulating two-dimensional (2D) images of three-dimensional (3D) objects of a predetermined class (e.g., human faces). A 2D input image of a 3D object of the predetermined class is manipulated by manipulating physical properties of the 3D object, such as a 3D shape of the 3D input object, an albedo of the 3D input object, a pose of the 3D input object, and lighting illuminating the 3D input object. The physical properties are extracted from the 2D input image using a neural network that is trained to reconstruct the 2D input image. The 2D input image is reconstructed by disentangling the physical properties from pixels of the 2D input image using multiple subnetworks. The disentangled physical properties produced by the multiple subnetworks are combined into a 2D output image using a differentiable renderer.
    Type: Grant
    Filed: October 13, 2021
    Date of Patent: May 30, 2023
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Tim Marks, Safa Medin, Anoop Cherian, Ye Wang
  • Patent number: 11651497
    Abstract: System and method for generating verisimilar images from real depth images. Train a generative adversarial neural network (GAN) by accessing test depth images having identical instances as instances of a real depth image. Input the test depth images in the generator to generate estimated depth images representing an implicit three-dimensional model of the object. Input, each estimated depth image into a discriminator to obtain a loss and into a pose encoder to obtain a matching loss. Iteratively repeat processes until the losses are minimized to a threshold, to end training. Identify the instances in the real image using the trained GAN pose encoder, to produce a pose transformation matrix for each instance in the real image. Identify pixels in the depth images corresponding to the instances of the real image and merge the pixels for the depth images to form an instance segmentation map for the real depth image.
    Type: Grant
    Filed: March 25, 2021
    Date of Patent: May 16, 2023
    Inventors: Anoop Cherian, Goncalo José Dias Pais, Tim Marks, Alan Sullivan
  • Patent number: 11635299
    Abstract: A navigation system for providing driving instructions to a driver of a vehicle traveling on a route is provided. The driving instructions are generated by executing a multimodal fusion method that comprises extracting features from sensor measurements, annotating the features with directions for the vehicle to follow the route with respect to objects sensed by the sensors, and encoding the annotated features with a multimodal attention neural network to produce encodings. The encodings are transformed into a common latent space, and the transformed encodings are fused using an attention mechanism producing an encoded representation of the scene. The method further comprises decoding the encoded representation with a sentence generation neural network to generate a driving instruction and submitting the driving instruction to an output device.
    Type: Grant
    Filed: February 6, 2020
    Date of Patent: April 25, 2023
    Inventors: Chiori Hori, Anoop Cherian, Siheng Chen, Tim Marks, Jonathan Le Roux, Takaaki Hori, Bret Harsham, Anthony Vetro, Alan Sullivan
  • Publication number: 20230112302
    Abstract: Present disclosure discloses an image processing system and method for manipulating two-dimensional (2D) images of three-dimensional (3D) objects of a predetermined class (e.g., human faces). A 2D input image of a 3D object of the predetermined class is manipulated by manipulating physical properties of the 3D object, such as a 3D shape of the 3D input object, an albedo of the 3D input object, a pose of the 3D input object, and lighting illuminating the 3D input object. The physical properties are extracted from the 2D input image using a neural network that is trained to reconstruct the 2D input image. The 2D input image is reconstructed by disentangling the physical properties from pixels of the 2D input image using multiple subnetworks. The disentangled physical properties produced by the multiple subnetworks are combined into a 2D output image using a differentiable renderer.
    Type: Application
    Filed: October 13, 2021
    Publication date: April 13, 2023
    Applicant: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Tim Marks, Safa Medin, Anoop Cherian, Ye Wang
  • Patent number: 11582485
    Abstract: Embodiments of the present disclosure discloses a scene-aware video encoder system. The scene-aware encoder system transforms a sequence of video frames of a video of a scene into a spatio-temporal scene graph. The spatio-temporal scene graph includes nodes representing one or multiple static and dynamic objects in the scene. Each node of the spatio-temporal scene graph describes an appearance, a location, and/or a motion of each of the objects (static and dynamic objects) at different time instances. The nodes of the spatio-temporal scene graph are embedded into a latent space using a spatio-temporal transformer encoding different combinations of different nodes of the spatio-temporal scene graph corresponding to different spatio-temporal volumes of the scene. Each node of the different nodes encoded in each of the combinations is weighted with an attention score determined as a function of similarities of spatio-temporal locations of the different nodes in the combination.
    Type: Grant
    Filed: February 7, 2022
    Date of Patent: February 14, 2023
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Anoop Cherian, Chiori Hori, Jonathan Le Roux, Tim Marks, Alan Sullivan
  • Publication number: 20230020834
    Abstract: Embodiments disclose a method and system for a scene-aware audio-video representation of a scene. The scene-aware audio video representation corresponds to a graph of nodes connected by edges. A node in the graph is indicative of the video features of an object in the scene. An edge in the graph connecting two nodes indicates an interaction of the corresponding two objects in the scene. In the graph, at least one or more edges are associated with audio features of a sound generated by the interaction of the corresponding two objects. The graph of the audio-video representation of the scene may be used to perform a variety of different tasks. Examples of the tasks include one or a combination of an action recognition, an anomaly detection, a sound localization and enhancement, a noisy-background sound removal, and a system control.
    Type: Application
    Filed: July 19, 2021
    Publication date: January 19, 2023
    Applicant: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Moitreya Chatterjee, Anoop Cherian, Jonathan Le Roux
  • Publication number: 20220309672
    Abstract: System and method for generating verisimilar images from real depth images. Train a generative adversarial neural network (GAN) by accessing test depth images having identical instances as instances of a real depth image. Input the test depth images in the generator to generate estimated depth images representing an implicit three-dimensional model of the object. Input, each estimated depth image into a discriminator to obtain a loss and into a pose encoder to obtain a matching loss. Iteratively repeat processes until the losses are minimized to a threshold, to end training. Identify the instances in the real image using the trained GAN pose encoder, to produce a pose transformation matrix for each instance in the real image. Identify pixels in the depth images corresponding to the instances of the real image and merge the pixels for the depth images to form an instance segmentation map for the real depth image.
    Type: Application
    Filed: March 25, 2021
    Publication date: September 29, 2022
    Applicant: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Anoop Cherian, Goncalo José Dias Pais, Tim Marks, Alan Sullivan
  • Patent number: 11449985
    Abstract: A system includes an imager, a processor, and an output module. The imager is configured to provide a plurality of tissue images. The processor is coupled to the imager and is configured to receive the plurality of images. The processor is coupled to a memory. The memory has instructions for determining classification of a region of tissue associated with the plurality of tissue images. Determining classification includes fusing discriminator outputs from a region covariance descriptor and from a normalized color histogram discriminator. The output module is coupled to the processor. The output module is configured to provide a three dimensional representation of the tissue.
    Type: Grant
    Filed: December 1, 2017
    Date of Patent: September 20, 2022
    Assignees: Regents of the University of Minnesota, Australian National University
    Inventors: Panagiotis Stanitsas, Anoop Cherian, Vassilios Morellas, Nikolaos Papanikolopoulos, Alexander Truskinovsky
  • Patent number: 11445267
    Abstract: A scene captioning system is provided. The scene captioning system includes an interface configured to acquire a stream of scene data signals including frames and sound data, a memory to store a computer-executable scene captioning model including a scene encoder, a timing decoder, a timing detector, and a caption decoder, wherein the audio-visual encoder is shared by the timing decoder and the timing detector and the caption decoder, and a processor, in connection with the memory. The processor is configured to perform steps of extracting scene features from the scene data signals by use of the audio-visual encoder, determining a timing of generating a caption by use of the timing detector, wherein the timing is arranged an early stage of the stream of scene data signals, and generating the caption based on the scene features by using the caption decoder according to the timing.
    Type: Grant
    Filed: July 23, 2021
    Date of Patent: September 13, 2022
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Chiori Hori, Takaaki Hori, Anoop Cherian, Tim Marks, Jonathan Le Roux
  • Patent number: 11423698
    Abstract: Embodiments of the present disclosure disclose an anomaly detector for detecting an anomaly in a sequence of poses of a human performing an activity. The anomaly detector includes an input interface configured to accept input data indicative of a distribution of the sequence of poses, a memory configured to store a discriminative one-class classifier having a pair of complementary classifiers bounding normal distribution of pose sequences in a reproducing kernel Hilbert space (RKHS), a processor configured to embed the input data into an element of the RKHS and classify the embedded data using the discriminative one-class classifier, and an output interface configured to render a classification result.
    Type: Grant
    Filed: October 26, 2020
    Date of Patent: August 23, 2022
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Anoop Cherian, Jue Wang
  • Publication number: 20220129666
    Abstract: Embodiments of the present disclosure disclose an anomaly detector for detecting an anomaly in a sequence of poses of a human performing an activity. The anomaly detector includes an input interface configured to accept input data indicative of a distribution of the sequence of poses, a memory configured to store a discriminative one-class classifier having a pair of complementary classifiers bounding normal distribution of pose sequences in a reproducing kernel Hilbert space (RKHS), a processor configured to embed the input data into an element of the RKHS and classify the embedded data using the discriminative one-class classifier, and an output interface configured to render a classification result.
    Type: Application
    Filed: October 26, 2020
    Publication date: April 28, 2022
    Applicant: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Anoop Cherian, Jue Wang
  • Patent number: 11264009
    Abstract: A computer-implemented method for training a dialogue response generation system and the dialogue response generation system are provided. The method includes arranging a first multimodal encoder-decoder for the dialogue response generation or video description having a first input and a first output, wherein the first multimodal encoder-decoder has been pretrained by training audio-video datasets with training video description sentences, arranging a second multimodal encoder-decoder for dialog response generation having a second input and a second output, providing first audio-visual datasets with first corresponding video description sentences to the first input of the first multimodal encoder-decoder, wherein the first encoder-decoder generates first output values based on the first audio-visual datasets with the first corresponding description sentences, providing the first audio-visual datasets excluding the first corresponding video description sentences to the second multimodal encoder-decoder.
    Type: Grant
    Filed: September 13, 2019
    Date of Patent: March 1, 2022
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Chiori Hori, Anoop Cherian, Tim Marks, Takaaki Hori
  • Patent number: 11210523
    Abstract: A scene aware dialog system includes an input interface to receive a sequence of video frames, contextual information, and a query and a memory configured to store neural networks trained to generate a response to the input query by analyzing one or combination of input sequence of video frames and the input contextual information. The system further includes a processor configured to detect and classify objects in each video frame of the sequence of video frames; determine relationships among the classified objects in each of the video frame; extract features representing the classified objects and the determined relationships for each of the video frame to produce a sequence of feature vectors; and submit the sequence of feature vectors, the input query and the input contextual information to the neural network to generate a response to the input query.
    Type: Grant
    Filed: February 6, 2020
    Date of Patent: December 28, 2021
    Assignee: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Shijie Geng, Peng Gao, Anoop Cherian, Chiori Hori, Jonathan Le Roux
  • Publication number: 20210397970
    Abstract: An artificial intelligence (AI) system that includes a processor configured to execute modules of the AI system. The modules comprise a feature extractor, an adversarial noise generator, a compressor and a classifier. The feature extractor is trained to process input data to extract features of the input data for classification of the input data. The adversarial noise generator is trained to generate noise data for distribution of features of the input data such that a misclassification rate of corrupted features that include the extracted features corrupted with the generated noise data is greater than a misclassification rate of the extracted features. The compressor is configured to compress the extracted features. The compressed features are closer to the extracted features than to the corrupted features. The classifier is trained to classify the compressed features.
    Type: Application
    Filed: June 22, 2020
    Publication date: December 23, 2021
    Inventors: Anoop Cherian, Aeron Shuchin
  • Publication number: 20210247201
    Abstract: A navigation system configured to provide driving instructions to a driver of a moving vehicle based on real-time description of objects in a scene pertinent to driving the vehicle is provided.
    Type: Application
    Filed: February 6, 2020
    Publication date: August 12, 2021
    Applicant: Mitsubishi ELectric Research Laboratories, Inc.
    Inventors: Chiori Hori, Anoop Cherian, Siheng Chen, Tim Marks, Jonathan Le Roux, Takaaki Hori, Bret Harsham, Anthony Vetro, Alan Sullivan
  • Publication number: 20210248375
    Abstract: A scene aware dialog system includes an input interface to receive a sequence of video frames, contextual information, and a query and a memory configured to store neural networks trained to generate a response to the input query by analyzing one or combination of input sequence of video frames and the input contextual information. The system further includes a processor configured to detect and classify objects in each video frame of the sequence of video frames; determine relationships among the classified objects in each of the video frame; extract features representing the classified objects and the determined relationships for each of the video frame to produce a sequence of feature vectors; and submit the sequence of feature vectors, the input query and the input contextual information to the neural network to generate a response to the input query.
    Type: Application
    Filed: February 6, 2020
    Publication date: August 12, 2021
    Applicant: Mitsubishi Electric Research Laboratories, Inc.
    Inventors: Shijie Geng, Peng Gao, Anoop Cherian, Chiori Hori, Jonathan Le Roux