Patents by Inventor Victor Carbune
Victor Carbune has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 12217751Abstract: A method includes instructing an always-on first processor to operate in a follow-on query detection mode, and while the always-on first processor operates in the follow-on query detection mode: receiving follow-on audio data captured by the assistant-enabled device; determining, using a voice activity detection (VAD) model executing on the always-on first processor, whether or not the VAD model detects voice activity in the follow-on audio data; performing, using a speaker identification (SID) model executing on the always-on first processor, speaker verification on the follow-on audio data to determine whether the follow-on audio data includes an utterance spoken by the same user. The method also includes initiating a wake-up process on a second processor to determine whether the utterance includes a follow-on query.Type: GrantFiled: December 15, 2021Date of Patent: February 4, 2025Assignee: Google LLCInventors: Victor Carbune, Matthew Sharifi
-
Patent number: 12217736Abstract: Implementations can detect respective audio data that captures an acoustic event at multiple assistant devices in an ecosystem that includes a plurality of assistant devices, process the respective audio data locally at each of the multiple assistant devices to generate respective measures that are associated with the acoustic event using respective event detection models, process the respective measures to determine whether the detected acoustic event is an actual acoustic event, and cause an action associated with the actional acoustic event to be performed in response to determining that the detected acoustic event is the actual acoustic event. In some implementations, the multiple assistant devices that detected the respective audio data are anticipated to detect the respective audio data that captures the actual acoustic event based on a plurality of historical acoustic events being detected at each of the multiple assistant devices.Type: GrantFiled: September 13, 2023Date of Patent: February 4, 2025Assignee: GOOGLE LLCInventors: Matthew Sharifi, Victor Carbune
-
Publication number: 20250028770Abstract: Implementations described herein relate to pairing a location-based automated assistant with a user device. The user device can include, for example, a headphones apparatus and/or a device that is paired with the headphones apparatus. The user device provides an indication that it is present at a location that is associated with a location-based automated assistant. A trust measure is determined that is indicative of trust between the user device and the location-based automated assistant. User information is provided by the user device to the location-based automated assistant. The location-based automated assistant determines response data to provide, via one or more speakers associated with the user device, that is specific to the location and further based on the user information.Type: ApplicationFiled: October 7, 2024Publication date: January 23, 2025Inventors: Victor Carbune, Matthew Sharifi
-
Publication number: 20250029607Abstract: Implementations relate to an automated assistant that can respond to communications received via a third party application and/or other third party communication modality. The automated assistant can determine that the user is participating in multiple different conversations via multiple different third party communication services. In some implementations, conversations can be processed to identify particular features of the conversations. When the automated assistant is invoked to provide input to a conversation, the automated assistant can compare the input to the identified conversation features in order to select the particular conversation that is most relevant to the input. In this way, the automated assistant can assist with any of multiple disparate conversations that are each occurring via a different third party application.Type: ApplicationFiled: October 7, 2024Publication date: January 23, 2025Inventors: Victor Carbune, Matthew Sharifi
-
Patent number: 12202125Abstract: Implementations set forth herein relate to a robotic computing device that can perform certain operations, such as communicating between users in a common space, according to certain preferences of the users. When interacting with a particular user, the robotic computing device can perform an operation at a preferred location relative to the particular user based on an express or implied preference of that particular user. For instance, certain types of operations can be performed at a first location within a room, and other types of operations can be performed at a second location within the room. When an operation involves following or guiding a user, parameters for driving the robotic computing device can be selected based on preferences of the user and/or a context in which the robotic computing device is interacting with the user (e.g., whether or not the context indicates some amount of urgency).Type: GrantFiled: November 23, 2021Date of Patent: January 21, 2025Assignee: GOOGLE LLCInventors: Victor Carbune, Matthew Sharifi
-
Patent number: 12205591Abstract: A method includes detecting multiple users, receiving a first query issued by a first user, the first query including a command for a digital assistant to perform a first action, and enabling a round robin mode to control performance of actions commanded by queries. The method also includes, while performing the first action, receiving audio data corresponding to a second query including a command to perform a second action, performing speaker identification on the audio data, determining that the second query was spoken by the first user, preventing performing the second action, and prompting at least another user to issue a query. The method further includes receiving a third query issued by a second user, the third query including a command for the digital assistant to perform a third action, and when the digital assistant completes performing the first action, executing performance of the third action.Type: GrantFiled: October 6, 2022Date of Patent: January 21, 2025Assignee: Google LLCInventors: Matthew Sharifi, Victor Carbune
-
Publication number: 20250022464Abstract: Techniques are described herein for warm word arbitration between automated assistant devices. A method includes: determining that warm word arbitration is to be initiated between a first assistant device and one or more additional assistant devices, including a second assistant device; broadcasting, by the first assistant device, to the one or more additional assistant devices, an active set of warm words for the first assistant device; for each of the one or more additional assistant devices, receiving, from the additional assistant device, an active set of warm words for the additional assistant device; identifying a matching warm word included in the active set of warm words for the first assistant device and included in the active set of warm words for the second assistant device; and enabling or disabling detection of the matching warm word by the first assistant device, in response to identifying the matching warm word.Type: ApplicationFiled: September 27, 2024Publication date: January 16, 2025Inventors: Matthew Sharifi, Victor Carbune
-
Patent number: 12197818Abstract: Implementations relate to an automated assistant that can automate repeatedly performed procedures. The automation can involve communicating with different users, organizations, and/or other automated assistants. The automated assistant, with prior permission from respective user(s), can detect repeated performance of a particular series of manually initiated computational actions. Based on this determination, the automated assistant can determine automated assistant computational action(s) that can be performed by the automated assistant in order to reduce latency in performing a procedure, reduce quantity and/or size of transmissions in performing the procedure, and/or reduce an amount of client device resources required for performing the procedure. Such actions can include communicating with an additional automated assistant that may be associated with another user and/or organization.Type: GrantFiled: September 1, 2023Date of Patent: January 14, 2025Assignee: GOOGLE LLCInventors: Matthew Sharifi, Victor Carbune
-
Publication number: 20250014574Abstract: Implementations set forth herein relate to an automated assistant that can jointly process spoken utterances that can include at least one corrective utterance and one utterance that may have been misinterpreted by the automated assistant. In this way, the automated assistant does not need to exclusively rely on the corrective utterance to determine a request of a user but, rather, can utilize data capturing any prior spoken utterance to more accurately determine the request. The automated assistant can process utterances jointly after determining that they are acoustically similar and/or received within a threshold duration of time. The joint processing can involve generating a single input, or multiple inputs, to characterize the series of utterances and processing the single input, or multiple inputs, with one or more neural network models and/or machine learning models. The automated assistant can then proactively correct any previous responses to misinterpreted utterances.Type: ApplicationFiled: July 3, 2023Publication date: January 9, 2025Inventors: Matthew Sharifi, Victor Carbune
-
Publication number: 20250012589Abstract: To present a navigation directions preview, a server device receives a request for navigation directions from a starting location to a destination location and generates a set of navigation directions in response to the request. The set of navigation directions includes a set of route segments for traversing from the starting location to the destination location. The server device selects a subset of the route segments based on characteristics of each route segment in the set of route segments. For each selected route segment, the server device provides a preview of the route segment to be displayed on a client device. The preview of the route segment includes panoramic street level imagery depicting the route segment.Type: ApplicationFiled: September 17, 2024Publication date: January 9, 2025Inventors: Victor Carbune, Matthew Sharifi
-
Patent number: 12190892Abstract: Some implementations relate to performing speech biasing, NLU biasing, and/or other biasing based on historical assistant interaction(s). It can be determined, for one or more given historical interactions of a given user, whether to affect future biasing for (1) the given user account, (2) additional user account(s), and/or (3) the shared assistant device as a whole. Some implementations disclosed herein additionally and/or alternatively relate to: determining, based on utterance(s) of a given user to a shared assistant device, an association of first data and second data; storing the association as accessible to a given user account of the given user; and determining whether to store the association as also accessible by additional user account(s) and/or the shared assistant device.Type: GrantFiled: October 18, 2023Date of Patent: January 7, 2025Assignee: GOOGLE LLCInventors: Matthew Sharifi, Victor Carbune
-
Patent number: 12192022Abstract: Systems and methods for video conference content auto-retrieval and focus based on learned relevance is provided. A method may include determining collaborative documents each associated with at least one user participating in a video conference, and providing one or more first inputs to a machine learning model, the one or more first inputs identifying the plurality of collaborative documents, and comprising an indication of a request to identify a collaborative document that is most relevant to a first subject discussed during the video conference at a first point in time. The method may further include determining, based on one or more first outputs of the machine learning model, the collaborative document that is the most relevant to the first subject discussed during the video conference at the first point in time, and providing the determined collaborative document for presentation on one or more client devices while the first subject is being discussed during the video conference.Type: GrantFiled: February 29, 2024Date of Patent: January 7, 2025Assignee: Google LLCInventors: Victor Carbune, Daniel Keysers, Thomas Deselaers
-
Patent number: 12188782Abstract: To provide dynamic generation and suggestion of map tiles, a server device receives from a user device a request for map data for a particular geographic region. The server device obtains a set of user contextual data and a set of candidate map tiles associated with the particular geographic region. The server device then selects one or more of the set of candidate map tiles based on the set of user contextual data, and transmits the one or more selected map tile to the user device for display.Type: GrantFiled: December 19, 2019Date of Patent: January 7, 2025Assignee: GOOGLE LLCInventors: Victor Carbune, Kevin Allekotte
-
Publication number: 20250006184Abstract: Implementations described herein include detecting a stream of audio data that captures a spoken utterance of the user and that captures ambient noise occurring within a threshold time period of the spoken utterance being spoken by the user. Implementations further include processing a portion of the audio data that includes the ambient noise to determine ambient noise classification(s), processing a portion of the audio data that includes the spoken utterance to generate a transcription, processing both the transcription and the ambient noise classification(s) with a machine learning model to generate a user intent and parameter(s) for the user intent, and performing one or more automated assistant actions based on the user intent and using the parameter(s).Type: ApplicationFiled: September 16, 2024Publication date: January 2, 2025Inventors: Victor Carbune, Matthew Sharifi
-
Patent number: 12183348Abstract: Implementations are directed to biasing speaker authentication on a per-user basis and on a device-by-device basis and/or contextual feature(s) basis. In some of those implementations, in performing speaker authentication based on a spoken utterance, different biasing parameters are determined for each of multiple different registered users of an assistant device at which the spoken utterance was detected. In those implementations, each of the biasing parameters can be used to make it more likely or less likely (in dependence of the biasing parameter) that a corresponding registered user will be verified using the speaker authentication. Through utilization of biasing parameter(s) in performing speaker authentication, accuracy and/or robustness of speaker authentication can be increased.Type: GrantFiled: October 23, 2023Date of Patent: December 31, 2024Assignee: GOOGLE LLCInventors: Matthew Sharifi, Victor Carbune
-
Patent number: 12175981Abstract: Implementations are directed to dynamically adapting which assistant on-device model(s) are locally stored at assistant devices of an assistant device group and/or dynamically adapting the assistant processing role(s) of the assistant device(s) of the assistant device group. In some of those implementations, the corresponding on-device model(s) and/or corresponding processing role(s), for each of the assistant devices of the group, is determined based on collectively considering individual processing capabilities of the assistant devices of the group. Implementations are additionally or alternatively directed to cooperatively utilizing assistant devices of a group, and their associated post-adaptation on-device model(s) and/or post-adaptation processing role(s), in cooperatively processing assistant requests that are directed to any one of the assistant devices of the group.Type: GrantFiled: July 28, 2023Date of Patent: December 24, 2024Assignee: GOOGLE LLCInventors: Matthew Sharifi, Victor Carbune
-
Publication number: 20240420698Abstract: Systems and methods for determining whether to combine responses from multiple automated assistants. An automated assistant may be invoked by a user utterance, followed by a query, which is provided to a plurality of automated assistants. A first response is received from a first automated assistant and a second response is received from a second automated assistant. Based on similarity between the responses, a primary automated assistant determines whether to combine the responses into a combined response. Once the combined response has been generated, one or more actions are performed in response to the combined response.Type: ApplicationFiled: August 30, 2024Publication date: December 19, 2024Inventors: Matthew Sharifi, Victor Carbune
-
Publication number: 20240420699Abstract: Systems and methods described herein relate to determining whether to incorporate recognized text, that corresponds to a spoken utterance of a user of a client device, into a transcription displayed at the client device, or to cause an assistant command, that is associated with the transcription and that is based on the recognized text, to be performed by an automated assistant implemented by the client device. The spoken utterance is received during a dictation session between the user and the automated assistant. Implementations can process, using automatic speech recognition model(s), audio data that captures the spoken utterance to generate the recognized text. Further, implementations can determine whether to incorporate the recognized text into the transcription or cause the assistant command to be performed based on touch input being directed to the transcription, a state of the transcription, and/or audio-based characteristic(s) of the spoken utterance.Type: ApplicationFiled: August 26, 2024Publication date: December 19, 2024Inventors: Victor Carbune, Alvin Abdagic, Behshad Behzadi, Jacopo Sannazzaro Natta, Julia Proskurnia, Krzysztof Andrzej Goj, Srikanth Pandiri, Viesturs Zarins, Nicolo D'Ercole, Zaheed Sabur, Luv Kothari
-
Patent number: 12164556Abstract: Techniques are described herein for providing smart suggestions for image zoom regions. A method includes: receiving a search query; performing a search using the search query to identify search results that include image search results including a plurality of images that are responsive to the search query; for a given image of the plurality of images included in the image search results, determining at least one zoom region in the given image; and providing the search results including the image search results, including providing the given image and an indication of the at least one zoom region in the given image.Type: GrantFiled: June 1, 2021Date of Patent: December 10, 2024Assignee: GOOGLE LLCInventors: Matthew Sharifi, Victor Carbune
-
Patent number: 12164572Abstract: Implementations can identify a given assistant device from among a plurality of assistant devices in an ecosystem, obtain device-specific signal(s) that are generated by the given assistant device, process the device-specific signal(s) to generate candidate semantic label(s) for the given assistant device, select a given semantic label for the given semantic device from among the candidate semantic label(s), and assigning, in a device topology representation of the ecosystem, the given semantic label to the given assistant device. Implementations can optionally receive a spoken utterance that includes a query or command at the assistant device(s), determine a semantic property of the query or command matches the given semantic label to the given assistant device, and cause the given assistant device to satisfy the query or command.Type: GrantFiled: December 6, 2023Date of Patent: December 10, 2024Assignee: GOOGLE LLCInventors: Matthew Sharifi, Victor Carbune