Patents by Inventor Nishant Shukla

Nishant Shukla has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240114459
    Abstract: The disclosure relates to a 5G or 6G communication system for supporting a higher data transmission rate. A method for optimizing power in a user equipment (UE) is provided. The method includes receiving, from a network, one or more relaxed measurement parameters, determining one of a mobility state or a location of the UE with respect to an edge of a serving cell of a visited public land mobile network (VPLMN) of the network based on one or more relaxed measurement parameters, and deferring a background PLMN (BPLMN) search upon determining one of the UE being in a low mobility state or not being at the edge of the serving cell, and/or deferring a near cell measurement search during a measurement gap of a connected mode discontinuous reception (CDRX) sleep duration of a CDRX state to optimize the power consumption in the UE upon determining one of the UE being in the low mobility state or not being located at the edge of the serving cell.
    Type: Application
    Filed: September 11, 2023
    Publication date: April 4, 2024
    Inventors: Anoop Perumudi VEEDU, Vivek MURUGAIYAN, Nishant ., Kailash Kumar JHA, Siddharth SHUKLA
  • Patent number: 11504856
    Abstract: The present teaching relates to method, system, medium, and implementation for activating an animatronic device. Information about a user is obtained for whom an animatronic device is to be configured to carry out a dialogue with the user. The animatronic device includes a head portion and a body portion and the head portion is configured based on one of a plurality of selectable head portions. One or more preferences of the user are identified from the obtained information and used to select, from the plurality of selectable head portions, a first selected head portion. A configuration of the head portion of the animatronic device is then configured based on the first selected head portion for carrying out the dialogue.
    Type: Grant
    Filed: December 27, 2018
    Date of Patent: November 22, 2022
    Assignee: DMAI, INC.
    Inventor: Nishant Shukla
  • Patent number: 11468894
    Abstract: The present teaching relates to method, system, medium, and implementations for enabling communication with a user. Information representing surrounding of a user engaged in an on-going dialogue is received via the communication platform, wherein the information includes a current response from the user in the on-going dialogue and is acquired from a current scene in which the user is present and captures characteristics of the user and the current scene. Relevant features are extracted from the information. A state of the user is estimated based on the relevant features and a dialogue context surrounding the current scene is determined based on the relevant features. A feedback directed to the current response of the user is generated based on the state of the user and the dialogue context.
    Type: Grant
    Filed: December 27, 2018
    Date of Patent: October 11, 2022
    Assignee: DMAI, INC.
    Inventor: Nishant Shukla
  • Publication number: 20220215678
    Abstract: The present teaching relates to method, system, medium, and implementations for understanding a three dimensional (3D) scene. Image data acquired by a camera at different time instances with respect to the 3D scene are received wherein the 3D scene includes a user or one or more objects. The face of the user is detected and tracked at different time instances. With respect to some of the time instances, a 2D user profile representing a region in the image data occupied by the user is generated based on a corresponding face detected and a corresponding 3D space in the 3D scene is estimated based on calibration parameters associated with the camera. Such estimated 3D space occupied by the user in the 3D scene is used to dynamically update a 3D space occupancy record of the 3D scene.
    Type: Application
    Filed: March 23, 2022
    Publication date: July 7, 2022
    Inventor: Nishant Shukla
  • Patent number: 11308312
    Abstract: The present teaching relates to method, system, medium, and implementations for understanding a three dimensional (3D) scene. Image data acquired by a camera at different time instances with respect to the 3D scene are received wherein the 3D scene includes a user or one or more objects. The face of the user is detected and tracked at different time instances. With respect to some of the time instances, a 2D user profile representing a region in the image data occupied by the user is generated based on a corresponding face detected and a corresponding 3D space in the 3D scene is estimated based on calibration parameters associated with the camera. Such estimated 3D space occupied by the user in the 3D scene is used to dynamically update a 3D space occupancy record of the 3D scene.
    Type: Grant
    Filed: February 15, 2019
    Date of Patent: April 19, 2022
    Assignee: DMAI, INC.
    Inventor: Nishant Shukla
  • Publication number: 20220101856
    Abstract: The present teaching relates to method, system, medium, and implementations for detecting a source of speech sound in a dialogue. A visual signal acquired from a dialogue scene is first received, where the visual signal captures a person present in the dialogue scene. A human lip associated with the person is detected from the visual signal and tracked to detect whether lip movement is observed. If lip movement is detected, a first candidate source of sound is generated corresponding to an area in the dialogue scene where the lip movement occurred.
    Type: Application
    Filed: December 13, 2021
    Publication date: March 31, 2022
    Inventors: Nishant Shukla, Ashwin Dharne
  • Publication number: 20220086272
    Abstract: The present teaching relates to method, system, and medium for cross network communications. Information related to an application running on a user device is first received, which includes a state of the application and sensor data obtained with respect to a user interacting with the application on the user device. A request is sent to an application server for an instruction of a state transition of the application. A light weight model (LWM) for an object involved in the state transition is received and is personalized based on at least one of the sensor data and one or more preferences related to the user to generate a personalized model (PM) for the object, which is then sent to the user device.
    Type: Application
    Filed: November 29, 2021
    Publication date: March 17, 2022
    Inventors: Nawar Rajab, Nishant Shukla
  • Patent number: 11200902
    Abstract: The present teaching relates to method, system, medium, and implementations for detecting a source of speech sound in a dialogue. A visual signal acquired from a dialogue scene is first received, where the visual signal captures a person present in the dialogue scene. A human lip associated with the person is detected from the visual signal and tracked to detect whether lip movement is observed. If lip movement is detected, a first candidate source of sound is generated corresponding to an area in the dialogue scene where the lip movement occurred.
    Type: Grant
    Filed: February 15, 2019
    Date of Patent: December 14, 2021
    Assignee: DMAI, INC.
    Inventors: Nishant Shukla, Ashwin Dharne
  • Patent number: 11190635
    Abstract: The present teaching relates to method, system, and medium for cross network communications. Information related to an application running on a user device is first received, which includes a state of the application and sensor data obtained with respect to a user interacting with the application on the user device. A request is sent to an application server for an instruction of a state transition of the application. A light weight model (LWM) for an object involved in the state transition is received and is personalized based on at least one of the sensor data and one or more preferences related to the user to generate a personalized model (PM) for the object, which is then sent to the user device.
    Type: Grant
    Filed: January 8, 2020
    Date of Patent: November 30, 2021
    Assignee: DMAI, INC.
    Inventors: Nawar Rajab, Nishant Shukla
  • Patent number: 11017779
    Abstract: The present teaching relates to method, system, medium, and implementations for speech recognition. An audio signal is received that represents a speech of a user engaged in a dialogue. A visual signal is received that captures the user uttering the speech. A first speech recognition result is obtained by performing audio based speech recognition based on the audio signal. Based on the visual signal, lip movement of the user is detected and a second speech recognition result is obtained by performing lip reading based speech recognition. The first and the second speech recognition results are then integrated to generate an integrated speech recognition result.
    Type: Grant
    Filed: February 15, 2019
    Date of Patent: May 25, 2021
    Assignee: DMAI, INC.
    Inventors: Nishant Shukla, Ashwin Dharne
  • Patent number: 11017551
    Abstract: The present teaching relates to method, system, medium, and implementations for identifying object of interest. Image data acquired by a camera with respect to a scene are received. One or more users are detected, during a period of time, from the image data who are present at the scene. Three dimensional (3D) gazing rays of the one or more users during the period of time are estimated. One or more intersections of such 3D gazing rays are identified and are used to determine at least one object of interest of the one or more users.
    Type: Grant
    Filed: February 15, 2019
    Date of Patent: May 25, 2021
    Assignee: DMAI, INC.
    Inventor: Nishant Shukla
  • Publication number: 20200145527
    Abstract: The present teaching relates to method, system, and medium for cross network communications. Information related to an application running on a user device is first received, which includes a state of the application and sensor data obtained with respect to a user interacting with the application on the user device. A request is sent to an application server for an instruction of a state transition of the application. A light weight model (LWM) for an object involved in the state transition is received and is personalized based on at least one of the sensor data and one or more preferences related to the user to generate a personalized model (PM) for the object, which is then sent to the user device.
    Type: Application
    Filed: January 8, 2020
    Publication date: May 7, 2020
    Inventors: Nawar Rajab, Nishant Shukla
  • Patent number: 10567570
    Abstract: The present teaching relates to method, system, and medium for cross network communications. Information related to an application running on a user device is first received, which includes a state of the application and sensor data obtained with respect to a user interacting with the application on the user device. A request is sent to an application server for an instruction of a state transition of the application. A light weight model (LWM) for an object involved in the state transition is received and is personalized based on at least one of the sensor data and one or more preferences related to the user to generate a personalized model (PM) for the object, which is then sent to the user device.
    Type: Grant
    Filed: December 27, 2018
    Date of Patent: February 18, 2020
    Assignee: DMAI, INC.
    Inventors: Nawar Rajab, Nishant Shukla
  • Publication number: 20190371318
    Abstract: The present teaching relates to method, system, medium, and implementations for speech recognition in a spoken language. Upon receiving a speech signal representing an utterance of a speaker in one of a plurality of spoken languages, speech recognition is performed based on the speech signal in accordance with a plurality of speech recognition models corresponding to the plurality of spoken languages to generate a plurality of text strings each of which represents a speech recognition result in a corresponding one of the plurality of spoken languages. With respect to each of the plurality of text strings associated with a corresponding spoken language, a likelihood that the utterance is in the corresponding spoken language is computed. A spoken language of the utterance is determined based on the likelihood with respect to each of the plurality of text strings.
    Type: Application
    Filed: February 15, 2019
    Publication date: December 5, 2019
    Inventor: Nishant Shukla
  • Publication number: 20190279642
    Abstract: The present teaching relates to method, system, medium, and implementations for speech recognition. An audio signal is received that represents a speech of a user engaged in a dialogue. A visual signal is received that captures the user uttering the speech. A first speech recognition result is obtained by performing audio based speech recognition based on the audio signal. Based on the visual signal, lip movement of the user is detected and a second speech recognition result is obtained by performing lip reading based speech recognition. The first and the second speech recognition results are then integrated to generate an integrated speech recognition result.
    Type: Application
    Filed: February 15, 2019
    Publication date: September 12, 2019
    Inventors: Nishant Shukla, Ashwin Dharne
  • Publication number: 20190251701
    Abstract: The present teaching relates to method, system, medium, and implementations for identifying object of interest. Image data acquired by a camera with respect to a scene are received. One or more users are detected, during a period of time, from the image data who are present at the scene. Three dimensional (3D) gazing rays of the one or more users during the period of time are estimated. One or more intersections of such 3D gazing rays are identified and are used to determine at least one object of interest of the one or more users.
    Type: Application
    Filed: February 15, 2019
    Publication date: August 15, 2019
    Inventor: Nishant Shukla
  • Publication number: 20190251350
    Abstract: The present teaching relates to method, system, medium, and implementations for determining a type of a scene. Image data acquired by a camera with respect to a scene are received and one or more objects present in the scene are detected therefrom. The detected objects are recognized based on object recognition models. The spatial relationships among the detected objects are then determined based on the image data. The recognized objects and their spatial relationships are then used to infer a type of the scene in accordance with at least one scene context-free grammar model.
    Type: Application
    Filed: February 15, 2019
    Publication date: August 15, 2019
    Inventors: Nishant Shukla, Ashwin Dharne
  • Publication number: 20190251970
    Abstract: The present teaching relates to method, system, medium, and implementations for detecting a source of speech sound in a dialogue. A visual signal acquired from a dialogue scene is first received, where the visual signal captures a person present in the dialogue scene. A human lip associated with the person is detected from the visual signal and tracked to detect whether lip movement is observed. If lip movement is detected, a first candidate source of sound is generated corresponding to an area in the dialogue scene where the lip movement occurred.
    Type: Application
    Filed: February 15, 2019
    Publication date: August 15, 2019
    Inventors: Nishant Shukla, Ashwin Dharne
  • Publication number: 20190251331
    Abstract: The present teaching relates to method, system, medium, and implementations for understanding a three dimensional (3D) scene. Image data acquired by a camera at different time instances with respect to the 3D scene are received wherein the 3D scene includes a user or one or more objects. The face of the user is detected and tracked at different time instances. With respect to some of the time instances, a 2D user profile representing a region in the image data occupied by the user is generated based on a corresponding face detected and a corresponding 3D space in the 3D scene is estimated based on calibration parameters associated with the camera. Such estimated 3D space occupied by the user in the 3D scene is used to dynamically update a 3D space occupancy record of the 3D scene.
    Type: Application
    Filed: February 15, 2019
    Publication date: August 15, 2019
    Inventor: Nishant Shukla
  • Publication number: 20190208052
    Abstract: The present teaching relates to method, system, and medium for cross network communications. Information related to an application running on a user device is first received, which includes a state of the application and sensor data obtained with respect to a user interacting with the application on the user device. A request is sent to an application server for an instruction of a state transition of the application. A light weight model (LWM) for an object involved in the state transition is received and is personalized based on at least one of the sensor data and one or more preferences related to the user to generate a personalized model (PM) for the object, which is then sent to the user device.
    Type: Application
    Filed: December 27, 2018
    Publication date: July 4, 2019
    Inventors: Nawar Rajab, Nishant Shukla