Patents by Inventor Victor Carbune

Victor Carbune has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230031521
    Abstract: Techniques are described herein for enabling an automated assistant to adjust its behavior depending on a detected age range and/or “vocabulary level” of a user who is engaging with the automated assistant. In various implementations, data indicative of a user's utterance may be used to estimate one or more of the user's age range and/or vocabulary level. The estimated age range/vocabulary level may be used to influence various aspects of a data processing pipeline employed by an automated assistant. In various implementations, aspects of the data processing pipeline that may be influenced by the user's age range/vocabulary level may include one or more of automated assistant invocation, speech-to-text (“STT”) processing, intent matching, intent resolution (or fulfillment), natural language generation, and/or text-to-speech (“TTS”) processing. In some implementations, one or more tolerance thresholds associated with one or more of these aspects, such as grammatical tolerances, vocabularic tolerances, etc.
    Type: Application
    Filed: October 10, 2022
    Publication date: February 2, 2023
    Inventors: Pedro Gonnet Anders, Victor Carbune, Daniel Keysers, Thomas Deselaers, Sandro Feuz
  • Patent number: 11568870
    Abstract: Implementations relate to an automated assistant that can respond to communications received via a third party application and/or other third party communication modality. The automated assistant can determine that the user is participating in multiple different conversations via multiple different third party communication services. In some implementations, conversations can be processed to identify particular features of the conversations. When the automated assistant is invoked to provide input to a conversation, the automated assistant can compare the input to the identified conversation features in order to select the particular conversation that is most relevant to the input. In this way, the automated assistant can assist with any of multiple disparate conversations that are each occurring via a different third party application.
    Type: Grant
    Filed: December 2, 2020
    Date of Patent: January 31, 2023
    Assignee: Google LLC
    Inventors: Victor Carbune, Matthew Sharifi
  • Publication number: 20230025709
    Abstract: Systems and methods for providing dialog data, from an initially invoked automated assistant to a subsequently invoked automated assistant. A first automated assistant may be invoked by a user utterance, followed by a dialog with the user that is processed by the first automated assistant. During the dialog, a request to transfer dialog data to a second automated assistant is received. The request may originate with the user, by the first automated assistant, and/or by the second automated assistant. Once authorized, the first automated assistant provides the previous dialog data to the second automated assistant. The second automated assistant performs one or more actions based on the dialog data.
    Type: Application
    Filed: November 22, 2021
    Publication date: January 26, 2023
    Inventors: Matthew Sharifi, Victor Carbune
  • Patent number: 11562748
    Abstract: Techniques are described herein for detecting and suppressing commands in media that may trigger another automated assistant. A method includes: determining, for each of a plurality of automated assistant devices in an environment that are each executing at least one automated assistant, an active capability of the automated assistant device; initiating playback of digital media by an automated assistant; in response to initiating playback, processing the digital media to identify an audio segment in the digital media that, upon playback, is expected to trigger activation of at least one automated assistant executing on at least one of the plurality of automated assistant devices in the environment, based on the active capability of the at least one of the plurality of automated assistant devices; and in response to identifying the audio segment in the digital media, modifying the digital media to suppress the activation of the at least one automated assistant.
    Type: Grant
    Filed: December 1, 2020
    Date of Patent: January 24, 2023
    Assignee: Google LLC
    Inventors: Matthew Sharifi, Victor Carbune
  • Patent number: 11557278
    Abstract: A method includes receiving audio data corresponding to an utterance spoken by a user that includes a command for a digital assistant to perform a long-standing operation, activating a set of one or more warm words associated with a respective action for controlling the long-standing operation, and associating the activated set of one or more warm words with only the user. While the digital assistant is performing the long-standing operation, the method includes receiving additional audio data corresponding to an additional utterance, identifying one of the warm words from the activated set of warm words, and performing speaker verification on the additional audio data. The method further includes performing the respective action associated with the identified one of the warm words for controlling the long-standing operation when the additional utterance was spoken by the same user that is associated with the activated set of one or more warm words.
    Type: Grant
    Filed: December 10, 2020
    Date of Patent: January 17, 2023
    Assignee: Google LLC
    Inventors: Matthew Sharifi, Victor Carbune
  • Patent number: 11558546
    Abstract: Implementations set forth herein relate to an automated assistant that can control a camera according to one or more conditions specified by a user. A condition can be satisfied when, for example, the automated assistant detects a particular environment feature is apparent. In this way, the user can rely on the automated assistant to identify and capture certain moments without necessarily requiring the user to constantly monitor a viewing window of the camera. In some implementations, a condition for the automated assistant to capture media data can be based on application data and/or other contextual data that is associated with the automated assistant. For instance, a relationship between content in a camera viewing window and other content of an application interface can be a condition upon which the automated assistant captures certain media data using a camera.
    Type: Grant
    Filed: November 24, 2020
    Date of Patent: January 17, 2023
    Assignee: Google LLC
    Inventors: Felix Weissenberger, Balint Miklos, Victor Carbune, Matthew Sharifi, Domenico Carbotta, Ray Chen, Kevin Fu, Bogdan Prisacari, Fo Lee, Mucun Lu, Neha Garg, Jacopo Sannazzaro Natta, Barbara Poblocka, Jae Seo, Matthew Miao, Thomas Qian, Luv Kothari
  • Patent number: 11557300
    Abstract: Techniques are described herein for detecting and handling failures in other automated assistants. A method includes: executing a first automated assistant in an inactive state at least in part on a computing device operated by a user; while in the inactive state, determining, by the first automated assistant, that a second automated assistant failed to fulfill a request of the user; in response to determining that the second automated assistant failed to fulfill the request of the user, the first automated assistant processing cached audio data that captures a spoken utterance of the user comprising the request that the second automated assistant failed to fulfill, or features of the cached audio data, to determine a response that fulfills the request of the user; and providing, by the first automated assistant to the user, the response that fulfills the request of the user.
    Type: Grant
    Filed: November 2, 2020
    Date of Patent: January 17, 2023
    Assignee: GOOGLE LLC
    Inventors: Victor Carbune, Matthew Sharifi
  • Patent number: 11557293
    Abstract: Some implementations process, using warm word model(s), a stream of audio data to determine a portion of the audio data that corresponds to particular word(s) and/or phrase(s) (e.g., a warm word) associated with an assistant command, process, using an automatic speech recognition (ASR) model, a preamble portion of the audio data (e.g., that precedes the warm word) and/or a postamble portion of the audio data (e.g., that follows the warm word) to generate ASR output, and determine, based on processing the ASR output, whether a user intended the assistant command to be performed. Additional or alternative implementations can process the stream of audio data using a speaker identification (SID) model to determine whether the audio data is sufficient to identify the user that provided a spoken utterance captured in the stream of audio data, and determine if that user is authorized to cause performance of the assistant command.
    Type: Grant
    Filed: May 17, 2021
    Date of Patent: January 17, 2023
    Assignee: GOOGLE LLC
    Inventors: Victor Carbune, Matthew Sharifi, Ondrej Skopek, Justin Lu, Daniel Valcarce, Kevin Kilgour, Mohamad Hassan Rom, Nicolo D'Ercole, Michael Golikov
  • Publication number: 20230005480
    Abstract: A method includes receiving a first instance of raw audio data corresponding to a voice-based command and receiving a second instance of the raw audio data corresponding to an utterance of audible contents for an audio-based communication spoken by a user. When a voice filtering recognition routine determines to activate voice filtering for at least the voice of the user, the method also includes obtaining a respective speaker embedding of the user and processing, using the respective speaker embedding, the second instance of the raw audio data to generate enhanced audio data for the audio-based communication that isolates the utterance of the audible contents spoken by the user and excludes at least a portion of the one or more additional sounds that are not spoken by the user The method also includes executing.
    Type: Application
    Filed: September 9, 2022
    Publication date: January 5, 2023
    Applicant: Google LLC
    Inventors: Matthew Sharifi, Victor Carbune
  • Publication number: 20220414093
    Abstract: Methods, systems, and computer readable media related to generating a combined search query based on search parameters of a current search query of a user and search parameters of one or more previously submitted search quer(ies) of the user that are determined to be of the same line of inquiry as the current search query. Two or more search queries may be determined to share a line of inquiry when it is determined that they are within a threshold level of semantic similarity to one another. Once a shared line of inquiry has been identified and a combined search query generated, users may interact with the search parameters and/or the search results to update the search parameters of the combined search query.
    Type: Application
    Filed: September 6, 2022
    Publication date: December 29, 2022
    Inventors: Matthew Sharifi, Victor Carbune
  • Publication number: 20220412768
    Abstract: A first computing device may implement a method for sharing a navigation session to minimize driver distraction. The method includes requesting access to a shared navigation session synchronized with a navigation session at a second computing device indicative of a navigation route, and the navigation session includes route data. In response to receiving access to the shared navigation session, the first computing device may receive synchronized route data indicative of the navigation route at the second computing device. The method further comprises obtaining one or more first navigation instructions corresponding to the synchronized route data including at least one navigation instruction that augments or is output in a different manner than one or more second navigation instructions output at the second computing device, and outputting the one or more first navigation instructions.
    Type: Application
    Filed: December 16, 2020
    Publication date: December 29, 2022
    Inventors: Matthew Sharifi, Victor Carbune
  • Patent number: 11537931
    Abstract: The present disclosure provides an on-device machine learning platform that enables sharing of machine-learned models between applications on a computing device. For example, a first application which has a machine-learned model for a specific task can expose the model to other applications through a system level application programming interface (API) for the other applications to use. Communications using the API can be handled by the on-device machine learning platform. In some implementations, some exchange of resources (e.g., computing resources) can be provided so that the first application is compensated for sharing the machine-learned model (e.g., on a per model invocation basis).
    Type: Grant
    Filed: November 29, 2017
    Date of Patent: December 27, 2022
    Assignee: GOOGLE LLC
    Inventors: Sandro Feuz, Victor Carbune
  • Patent number: 11532313
    Abstract: Some implementations relate to performing speech biasing, NLU biasing, and/or other biasing based on historical assistant interaction(s). It can be determined, for one or more given historical interactions of a given user, whether to affect future biasing for (1) the given user account, (2) additional user account(s), and/or (3) the shared assistant device as a whole. Some implementations disclosed herein additionally and/or alternatively relate to: determining, based on utterance(s) of a given user to a shared assistant device, an association of first data and second data; storing the association as accessible to a given user account of the given user; and determining whether to store the association as also accessible by additional user account(s) and/or the shared assistant device.
    Type: Grant
    Filed: August 27, 2020
    Date of Patent: December 20, 2022
    Assignee: GOOGLE LLC
    Inventors: Matthew Sharifi, Victor Carbune
  • Publication number: 20220392449
    Abstract: Implementations set forth herein relate to an automated assistant that provides a response for certain user queries based on a level of interaction of the user with respect to the automated assistant. Interaction can be characterized by sensor data, which can be processed using one or more trained machine learning models in order to identify parameters for generating a response. In this way, the response can be limited to preserve computational resources and/or ensure that the response is more readily understood given the amount of interaction exhibited by the user. In some instances, a response that embodies information that is supplemental, to an otherwise suitable response, can be provided when a user is exhibiting a particular level of interaction. In other instances, such supplemental information can be withheld when the user is not exhibiting that particular level of interaction, at least in order to preserve computational resources.
    Type: Application
    Filed: August 22, 2022
    Publication date: December 8, 2022
    Inventors: Victor Carbune, Matthew Sharifi
  • Patent number: 11521597
    Abstract: Implementations can receive audio data corresponding to a spoken utterance of a user, process the audio data to generate a plurality of speech hypotheses, determine an action to be performed by an automated assistant based on the speech hypotheses, and cause the computing device to render an indication of the action. In response to the computing device rendering the indication, implementations can receive additional audio data corresponding to an additional spoken utterance of the user, process the additional audio data to determine that a portion of the spoken utterance is similar to an additional portion of the additional spoken utterance, supplant the action with an alternate action, and cause the automated assistant to initiate performance of the alternate action. Some implementations can determine whether to render the indication of the action based on a confidence level associated with the action.
    Type: Grant
    Filed: September 3, 2020
    Date of Patent: December 6, 2022
    Assignee: GOOGLE LLC
    Inventors: Matthew Sharifi, Victor Carbune
  • Patent number: 11521618
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for collaboration between multiple voice controlled devices are disclosed. In one aspect, a method includes the actions of identifying, by a first computing device, a second computing device that is configured to respond to a particular, predefined hotword; receiving audio data that corresponds to an utterance; receiving a transcription of additional audio data outputted by the second computing device in response to the utterance; based on the transcription of the additional audio data and based on the utterance, generating a transcription that corresponds to a response to the additional audio data; and providing, for output, the transcription that corresponds to the response.
    Type: Grant
    Filed: December 17, 2019
    Date of Patent: December 6, 2022
    Assignee: GOOGLE LLC
    Inventors: Victor Carbune, Pedro Gonnet Anders, Thomas Deselaers, Sandro Feuz
  • Patent number: 11521600
    Abstract: Techniques are described herein for enabling an automated assistant to adjust its behavior depending on a detected vocabulary level or other vocal characteristics of an input utterance provided to an automated assistant. The estimated vocabulary level or other vocal characteristics may be used to influence various aspects of a data processing pipeline employed by the automated assistant. In some implementations, one or more tolerance thresholds associated with, for example, grammatical tolerances or vocabulary tolerances, may be adjusted based on the estimated vocabulary level or vocal characteristics of the input utterance.
    Type: Grant
    Filed: May 26, 2020
    Date of Patent: December 6, 2022
    Assignee: GOOGLE LLC
    Inventors: Pedro Gonnet Anders, Victor Carbune, Daniel Keysers, Thomas Deselaers, Sandro Feuz
  • Publication number: 20220382802
    Abstract: Techniques are described herein for providing smart suggestions for image zoom regions. A method includes: receiving a search query; performing a search using the search query to identify search results that include image search results including a plurality of images that are responsive to the search query; for a given image of the plurality of images included in the image search results, determining at least one zoom region in the given image; and providing the search results including the image search results, including providing the given image and an indication of the at least one zoom region in the given image.
    Type: Application
    Filed: June 1, 2021
    Publication date: December 1, 2022
    Inventors: Matthew Sharifi, Victor Carbune
  • Patent number: 11514109
    Abstract: Implementations can identify a given assistant device from among a plurality of assistant devices in an ecosystem, obtain device-specific signal(s) that are generated by the given assistant device, process the device-specific signal(s) to generate candidate semantic label(s) for the given assistant device, select a given semantic label for the given semantic device from among the candidate semantic label(s), and assigning, in a device topology representation of the ecosystem, the given semantic label to the given assistant device. Implementations can optionally receive a spoken utterance that includes a query or command at the assistant device(s), determine a semantic property of the query or command matches the given semantic label to the given assistant device, and cause the given assistant device to satisfy the query or command.
    Type: Grant
    Filed: October 29, 2020
    Date of Patent: November 29, 2022
    Assignee: GOOGLE LLC
    Inventors: Matthew Sharifi, Victor Carbune
  • Publication number: 20220366911
    Abstract: Implementations described herein relate to an application and/or automated assistant that can identify arrangement operations to perform for arranging text during speech-to-text operations—without a user having to expressly identify the arrangement operations. In some instances, a user that is dictating a document (e.g., an email, a text message, etc.) can provide a spoken utterance to an application in order to incorporate textual content. However, in some of these instances, certain corresponding arrangements are needed for the textual content in the document. The textual content that is derived from the spoken utterance can be arranged by the application based on an intent, vocalization features, and/or contextual features associated with the spoken utterance and/or a type of the application associated with the document, without the user expressly identifying the corresponding arrangements.
    Type: Application
    Filed: June 3, 2021
    Publication date: November 17, 2022
    Inventors: Victor Carbune, Krishna Sapkota, Behshad Behzadi, Julia Proskurnia, Jacopo Sannazzaro Natta, Justin Lu, Magali Boizot-Roche, Márius Sajgalík, Nicolo D'Ercole, Zaheed Sabur, Luv Kothari