Patents by Inventor Nitin Khandelwal

Nitin Khandelwal has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11948580
    Abstract: Implementations described herein are directed to enabling collaborative ranking of interpretations of spoken utterances based on data that is available to an automated assistant and third-party agent(s), respectively. The automated assistant can determine first-party interpretation(s) of a spoken utterance provided by a user, and can cause the third-party agent(s) to determine third-party interpretation(s) of the spoken utterance provided by the user. In some implementations, the automated assistant can select a given interpretation, from the first-party interpretation(s) and the third-party interpretation(s), of the spoken utterance, and can cause a given third-party agent to satisfy the spoken utterance based on the given interpretation.
    Type: Grant
    Filed: November 29, 2021
    Date of Patent: April 2, 2024
    Assignee: GOOGLE LLC
    Inventors: Akshay Goel, Nitin Khandelwal, Richard Park, Brian Chatham, Jonathan Eccles, David Sanchez, Dmytro Lapchuk
  • Patent number: 11935530
    Abstract: Systems, methods, and apparatus for using a multimodal response in the dynamic generation of client device output that is tailored to a current modality of a client device is disclosed herein. Multimodal client devices can engage in a variety of interactions across the multimodal spectrum including voice only interactions, voice forward interactions, multimodal interactions, visual forward interactions, visual only interactions etc. A multimodal response can include a core message to be rendered for all interaction types as well as one or more modality dependent components to provide a user with additional information.
    Type: Grant
    Filed: November 1, 2021
    Date of Patent: March 19, 2024
    Assignee: GOOGLE LLC
    Inventors: April Pufahl, Jared Strawderman, Harry Yu, Adriana Olmos Antillon, Jonathan Livni, Okan Kolak, James Giangola, Nitin Khandelwal, Jason Kearns, Andrew Watson, Joseph Ashear, Valerie Nygaard
  • Publication number: 20240031339
    Abstract: Implementations described herein utilize an independent server for facilitating secure exchange of data between multiple disparate parties. The independent server receives client data, via an automated assistant application executing at least in part at a client device, that is to be transmitted to a given third-party application. The independent server processes the client data, using a first encoder-decoder model, to generate opaque client data, and transmits the opaque client data to the given third-party application and without transmitting any of the client data. Further, the independent server receives response data, via the given third-party application, that is generated based on the opaque client data and that is to be transmitted back to the client device. The independent server processes the response data, using a second encoder-decoder model, to generate opaque response data, and transmits the opaque response data to the client device and without transmitting any of the response data.
    Type: Application
    Filed: August 23, 2022
    Publication date: January 25, 2024
    Inventors: Akshay Goel, Jonathan Eccles, Nitin Khandelwal, Sarvjeet Singh, David Sanchez, Ashwin Ram
  • Publication number: 20230062201
    Abstract: Implementations described herein are directed to enabling collaborative ranking of interpretations of spoken utterances based on data that is available to an automated assistant and third-party agent(s), respectively. The automated assistant can determine first-party interpretation(s) of a spoken utterance provided by a user, and can cause the third-party agent(s) to determine third-party interpretation(s) of the spoken utterance provided by the user. In some implementations, the automated assistant can select a given interpretation, from the first-party interpretation(s) and the third-party interpretation(s), of the spoken utterance, and can cause a given third-party agent to satisfy the spoken utterance based on the given interpretation.
    Type: Application
    Filed: November 29, 2021
    Publication date: March 2, 2023
    Inventors: Akshay Goel, Nitin Khandelwal, Richard Park, Brian Chatham, Jonathan Eccles, David Sanchez, Dmytro Lapchuk
  • Patent number: 11568869
    Abstract: Implementations include identifying, from a database of entries reflecting past automated assistant commands submitted within a threshold amount of time relative to a current time, particular entries that each reflect corresponding features of a corresponding user submission of a particular command. Further, those implementations include determining that the particular command is a golden command, for a particular automated assistant function, responsive to determining that: at least a threshold percentage of the user submissions of the particular command triggered the particular automated assistant function, and a quantity of the user submission of the particular command satisfies a threshold quantity.
    Type: Grant
    Filed: November 23, 2020
    Date of Patent: January 31, 2023
    Assignee: GOOGLE LLC
    Inventors: Aakash Goel, Tayfun Elmas, Keith Brady, Akshay Jaggi, Ester Lopez Berga, Arne Vansteenkiste, Robin Martinjak, Mahesh Palekar, Krish Narang, Nitin Khandelwal, Pravir Gupta
  • Publication number: 20220207873
    Abstract: A system and methodology provide for annotating videos with entities and associated probabilities of existence of the entities within video frames. A computer-implemented method identifies an entity from a plurality of entities identifying characteristics of video items. The computer-implemented method selects a set of features correlated with the entity based on a value of a feature of a plurality of features, determines a classifier for the entity using the set of features, and determines an aggregation calibration function for the entity based on the set of features. The computer-implemented method selects a video frame from a video item, where the video frame having associated features, and determines a probability of existence of the entity based on the associated features using the classifier and the aggregation calibration function.
    Type: Application
    Filed: December 13, 2021
    Publication date: June 30, 2022
    Inventors: Balakrishnan Varadarajan, George Dan Toderici, Apostol Natsev, Nitin Khandelwal, Sudheendra Vijayanarasimhan, Weilong Yang, Sanketh Shetty
  • Publication number: 20220165259
    Abstract: Implementations include identifying, from a database of entries reflecting past automated assistant commands submitted within a threshold amount of time relative to a current time, particular entries that each reflect corresponding features of a corresponding user submission of a particular command. Further, those implementations include determining that the particular command is a golden command, for a particular automated assistant function, responsive to determining that: at least a threshold percentage of the user submissions of the particular command triggered the particular automated assistant function, and a quantity of the user submission of the particular command satisfies a threshold quantity.
    Type: Application
    Filed: November 23, 2020
    Publication date: May 26, 2022
    Inventors: Aakash Goel, Tayfun Elmas, Keith Brady, Akshay Jaggi, Ester Lopez Berga, Arne Vansteenkiste, Robin Martinjak, Mahesh Palekar, Krish Narang, Nitin Khandelwal, Pravir Gupta
  • Publication number: 20220051675
    Abstract: Systems, methods, and apparatus for using a multimodal response in the dynamic generation of client device output that is tailored to a current modality of a client device is disclosed herein. Multimodal client devices can engage in a variety of interactions across the multimodal spectrum including voice only interactions, voice forward interactions, multimodal interactions, visual forward interactions, visual only interactions etc. A multimodal response can include a core message to be rendered for all interaction types as well as one or more modality dependent components to provide a user with additional information.
    Type: Application
    Filed: November 1, 2021
    Publication date: February 17, 2022
    Inventors: April Pufahl, Jared Strawderman, Harry Yu, Adriana Olmos Antillon, Jonathan Livni, Okan Kolak, James Giangola, Nitin Khandelwal, Jason Kearns, Andrew Watson, Joseph Ashear, Valerie Nygaard
  • Patent number: 11200423
    Abstract: A system and methodology provide for annotating videos with entities and associated probabilities of existence of the entities within video frames. A computer-implemented method identifies an entity from a plurality of entities identifying characteristics of video items. The computer-implemented method selects a set of features correlated with the entity based on a value of a feature of a plurality of features, determines a classifier for the entity using the set of features, and determines an aggregation calibration function for the entity based on the set of features. The computer-implemented method selects a video frame from a video item, where the video frame having associated features, and determines a probability of existence of the entity based on the associated features using the classifier and the aggregation calibration function.
    Type: Grant
    Filed: November 18, 2019
    Date of Patent: December 14, 2021
    Assignee: Google LLC
    Inventors: Balakrishnan Varadarajan, George Dan Toderici, Apostol Natsev, Nitin Khandelwal, Sudheendra Vijayanarasimhan, Weilong Yang, Sanketh Shetty
  • Patent number: 11164576
    Abstract: Systems, methods, and apparatus for using a multimodal response in the dynamic generation of client device output that is tailored to a current modality of a client device is disclosed herein. Multimodal client devices can engage in a variety of interactions across the multimodal spectrum including voice only interactions, voice forward interactions, multimodal interactions, visual forward interactions, visual only interactions etc. A multimodal response can include a core message to be rendered for all interaction types as well as one or more modality dependent components to provide a user with additional information.
    Type: Grant
    Filed: January 18, 2019
    Date of Patent: November 2, 2021
    Assignee: GOOGLE LLC
    Inventors: April Pufahl, Jared Strawderman, Harry Yu, Adriana Olmos Antillon, Jonathan Livni, Okan Kolak, James Giangola, Nitin Khandelwal, Jason Kearns, Andrew Watson, Joseph Ashear, Valerie Nygaard
  • Publication number: 20210166035
    Abstract: A computer-implemented method for selecting representative frames for videos is provided. The method includes receiving a video and identifying a set of features for each of the frames of the video. The features including frame-based features and semantic features. The semantic features identifying likelihoods of semantic concepts being present as content in the frames of the video. A set of video segments for the video is subsequently generated. Each video segment includes a chronological subset of frames from the video and each frame is associated with at least one of the semantic features. The method generates a score for each frame of the subset of frames for each video segment based at least on the semantic features, and selecting a representative frame for each video segment based on the scores of the frames in the video segment. The representative frame represents and summarizes the video segment.
    Type: Application
    Filed: December 14, 2020
    Publication date: June 3, 2021
    Inventors: Sanketh Shetty, Tomas Izo, Min-Hsuan Tsai, Sudheendra Vijayanarasimhan, Apostol Natsev, Sami Abu-El-Haija, George Dan Toderici, Susana Ricco, Balakrishnan Varadarajan, Nicola Muscettola, WeiHsin Gu, Weilong Yang, Nitin Khandelwal, Phuong Le
  • Patent number: 10867183
    Abstract: A computer-implemented method for selecting representative frames for videos is provided. The method includes receiving a video and identifying a set of features for each of the frames of the video. The features including frame-based features and semantic features. The semantic features identifying likelihoods of semantic concepts being present as content in the frames of the video. A set of video segments for the video is subsequently generated. Each video segment includes a chronological subset of frames from the video and each frame is associated with at least one of the semantic features. The method generates a score for each frame of the subset of frames for each video segment based at least on the semantic features, and selecting a representative frame for each video segment based on the scores of the frames in the video segment. The representative frame represents and summarizes the video segment.
    Type: Grant
    Filed: April 23, 2018
    Date of Patent: December 15, 2020
    Assignee: Google LLC
    Inventors: Sanketh Shetty, Tomas Izo, Min-Hsuan Tsai, Sudheendra Vijayanarasimhan, Apostol Natsev, Sami Abu-El-Haija, George Dan Toderici, Susanna Ricco, Balakrishnan Varadarajan, Nicola Muscettola, WeiHsin Gu, Weilong Yang, Nitin Khandelwal, Phuong Le
  • Publication number: 20200082173
    Abstract: A system and methodology provide for annotating videos with entities and associated probabilities of existence of the entities within video frames. A computer-implemented method identifies an entity from a plurality of entities identifying characteristics of video items. The computer-implemented method selects a set of features correlated with the entity based on a value of a feature of a plurality of features, determines a classifier for the entity using the set of features, and determines an aggregation calibration function for the entity based on the set of features. The computer-implemented method selects a video frame from a video item, where the video frame having associated features, and determines a probability of existence of the entity based on the associated features using the classifier and the aggregation calibration function.
    Type: Application
    Filed: November 18, 2019
    Publication date: March 12, 2020
    Inventors: Balakrishnan Varadarajan, George Dan Toderici, Apostol Natsev, Nitin Khandelwal, Sudheendra Vijayanarasimhan, Weilong Yang, Sanketh Shetty
  • Publication number: 20200075002
    Abstract: Systems, methods, and apparatus for using a multimodal response in the dynamic generation of client device output that is tailored to a current modality of a client device is disclosed herein. Multimodal client devices can engage in a variety of interactions across the multimodal spectrum including voice only interactions, voice forward interactions, multimodal interactions, visual forward interactions, visual only interactions etc. A multimodal response can include a core message to be rendered for all interaction types as well as one or more modality dependent components to provide a user with additional information.
    Type: Application
    Filed: January 18, 2019
    Publication date: March 5, 2020
    Inventors: April Pufahl, Jared Strawderman, Harry Yu, Adriana Olmos Antillon, Jonathan Livni, Okan Kolak, James Giangola, Nitin Khandelwal, Jason Kearns, Andrew Watson, Joseph Ashear, Valerie Nygaard
  • Patent number: 10482328
    Abstract: A system and methodology provide for annotating videos with entities and associated probabilities of existence of the entities within video frames. A computer-implemented method identifies an entity from a plurality of entities identifying characteristics of video items. The computer-implemented method selects a set of features correlated with the entity based on a value of a feature of a plurality of features, determines a classifier for the entity using the set of features, and determines an aggregation calibration function for the entity based on the set of features. The computer-implemented method selects a video frame from a video item, where the video frame having associated features, and determines a probability of existence of the entity based on the associated features using the classifier and the aggregation calibration function.
    Type: Grant
    Filed: October 2, 2017
    Date of Patent: November 19, 2019
    Assignee: Google LLC
    Inventors: Balakrishnan Varadarajan, George Dan Toderici, Apostol Natsev, Nitin Khandelwal, Sudheendra Vijayanarasimhan, Weilong Yang, Sanketh Shetty
  • Patent number: 10216842
    Abstract: An initial plurality of search results is obtained based on a search query pertaining to a topic of interest. A determination is made of whether to perform clustering on the initial plurality of search results. Upon determining not to perform clustering on the search result, a user interface with the initial plurality of search results is provided. Upon determining to perform clustering on the initial plurality of search results, a user interface with a modified plurality of search results is provided. The user interface is to position a first result in the modified plurality of search results adjacent to a position of a second search result in the modified plurality of search results. The first search result is associated with a channel and the second search result is associated with the channel.
    Type: Grant
    Filed: June 3, 2013
    Date of Patent: February 26, 2019
    Assignee: GOOGLE LLC
    Inventors: Minya Dai, Nitin Khandelwal, Nitin Jindal, Neal Norwitz, Matthew Smith, Trevor O'Brien
  • Publication number: 20180239964
    Abstract: A computer-implemented method for selecting representative frames for videos is provided. The method includes receiving a video and identifying a set of features for each of the frames of the video. The features including frame-based features and semantic features. The semantic features identifying likelihoods of semantic concepts being present as content in the frames of the video. A set of video segments for the video is subsequently generated. Each video segment includes a chronological subset of frames from the video and each frame is associated with at least one of the semantic features. The method generates a score for each frame of the subset of frames for each video segment based at least on the semantic features, and selecting a representative frame for each video segment based on the scores of the frames in the video segment. The representative frame represents and summarizes the video segment.
    Type: Application
    Filed: April 23, 2018
    Publication date: August 23, 2018
    Inventors: Sanketh Shetty, Tomas Izo, Min-Hsuan Tsai, Sudheendra Vijayanarasimhan, Apostol Natsev, Sami Abu-El-Haija, George Dan Toderici, Susanna Ricco, Balakrishnan Varadarajan, Nicola Muscettola, WeiHsin Gu, Weilong Yang, Nitin Khandelwal, Phuong Le
  • Patent number: 9959322
    Abstract: A search result is modified to include a predetermined number of channels within a predetermined number of a plurality of entries in the search result. The search result is generated in response to a search query. The search result is modified to comprise a predetermined number of channels within a predetermined number of the plurality of entries. The modified search result is updated based on the plurality of entries and a subset of the updated modified search result is selected to be presented in response to the search query. The subset of the updated modified search result comprises the channel and at least one selection of data content.
    Type: Grant
    Filed: May 13, 2016
    Date of Patent: May 1, 2018
    Assignee: GOOGLE LLC
    Inventors: Nitin Khandelwal, Sreedhar Mukkamalla, Nitin Jindal, Fabio Lopiano, Nicola Muscettola
  • Patent number: 9953222
    Abstract: A computer-implemented method for selecting representative frames for videos is provided. The method includes receiving a video and identifying a set of features for each of the frames of the video. The features including frame-based features and semantic features. The semantic features identifying likelihoods of semantic concepts being present as content in the frames of the video. A set of video segments for the video is subsequently generated. Each video segment includes a chronological subset of frames from the video and each frame is associated with at least one of the semantic features. The method generates a score for each frame of the subset of frames for each video segment based at least on the semantic features, and selecting a representative frame for each video segment based on the scores of the frames in the video segment. The representative frame represents and summarizes the video segment.
    Type: Grant
    Filed: September 8, 2015
    Date of Patent: April 24, 2018
    Assignee: Google LLC
    Inventors: Sanketh Shetty, Tomas Izo, Min-Hsuan Tsai, Sudheendra Vijayanarasimhan, Apostol Natsev, Sami Abu-El-Haija, George Dan Toderici, Susanna Ricco, Balakrishnan Varadarajan, Nicola Muscettola, WeiHsin Gu, Weilong Yang, Nitin Khandelwal, Phuong Le
  • Publication number: 20180025228
    Abstract: A system and methodology provide for annotating videos with entities and associated probabilities of existence of the entities within video frames. A computer-implemented method identifies an entity from a plurality of entities identifying characteristics of video items. The computer-implemented method selects a set of features correlated with the entity based on a value of a feature of a plurality of features, determines a classifier for the entity using the set of features, and determines an aggregation calibration function for the entity based on the set of features. The computer-implemented method selects a video frame from a video item, where the video frame having associated features, and determines a probability of existence of the entity based on the associated features using the classifier and the aggregation calibration function.
    Type: Application
    Filed: October 2, 2017
    Publication date: January 25, 2018
    Inventors: Balakrishnan Varadarajan, George Dan Toderici, Apostol Natsev, Nitin Khandelwal, Sudheendra Vijayanarasimhan, Weilong Yang, Sanketh Shetty