Patents by Inventor Kenneth Mixter
Kenneth Mixter has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20220277727Abstract: A method for proactive notifications in a voice interface device includes: receiving a first user voice request for an action with an future performance time; assigning the first user voice request to a voice assistant service for performance; subsequent to the receiving, receiving a second user voice request and in response to the second user voice request initiating a conversation with the user; and during the conversation: receiving a notification from the voice assistant service of performance of the action; triggering a first audible announcement to the user to indicate a transition from the conversation and interrupting the conversation; triggering a second audible announcement to the user to indicate performance of the action; and triggering a third audible announcement to the user to indicate a transition back to the conversation and rejoining the conversation.Type: ApplicationFiled: May 16, 2022Publication date: September 1, 2022Inventors: Kenneth Mixter, Daniel Colish, Tuan Nguyen
-
Patent number: 11392213Abstract: Techniques are described herein for reducing false positives in vision sensor-equipped assistant devices. In various implementations, initial image frame(s) may be obtained from vision sensor(s) of an assistant device and analyzed to classify a particular region of the initial image frames as being likely to contain visual noise. Subsequent image frame(s) obtained from the vision sensor(s) may then be analyzed to detect actionable user-provided visual cue(s), in a manner that reduces or eliminates false positives. In some implementations, no analysis may be performed on the particular region of the subsequent image frame(s). Additionally or alternatively, in some implementations, a first candidate visual cue detected within the particular region may be weighted less heavily than a second candidate visual cue detected elsewhere in the one or more subsequent image frames. An automated assistant may then take responsive action based on the detected actionable visual cue(s).Type: GrantFiled: April 13, 2021Date of Patent: July 19, 2022Assignee: GOOGLE LLCInventors: Kenneth Mixter, Yuan Yuan, Tuan Nguyen
-
Patent number: 11355116Abstract: An electronic device configures a device-agnostic voice assistant library for execution on the electronic device based on the electronic device having a first device type. The electronic device also selects an implementation for the voice assistant library. After the configuring, the electronic device receives a verbal input from a user. It extracts request information from the verbal input by processing the verbal input using the voice assistant library executing on the device. It transmits a request to a remote system, the request including the extracted request information. The electronic device receives a response to the request. The response is generated by the remote system in accordance with the extracted request information. The electronic device performs an operation in accordance with the response by one or more voice processing modules of the configured voice assistant library.Type: GrantFiled: May 29, 2020Date of Patent: June 7, 2022Assignee: Google LLCInventors: Kenneth Mixter, Raunaq Shah
-
Patent number: 11335319Abstract: A method for proactive notifications in a voice interface device includes: receiving a first user voice request for an action with an future performance time; assigning the first user voice request to a voice assistant service for performance; subsequent to the receiving, receiving a second user voice request and in response to the second user voice request initiating a conversation with the user; and during the conversation: receiving a notification from the voice assistant service of performance of the action; triggering a first audible announcement to the user to indicate a transition from the conversation and interrupting the conversation; triggering a second audible announcement to the user to indicate performance of the action; and triggering a third audible announcement to the user to indicate a transition back to the conversation and rejoining the conversation.Type: GrantFiled: June 5, 2020Date of Patent: May 17, 2022Assignee: Google LLCInventors: Kenneth Mixter, Daniel Colish, Tuan Nguyen
-
Patent number: 11237630Abstract: Invoking one or more previously dormant functions of an automated assistant in response to detecting, based on processing of vision data from one or more vision components: (1) a particular gesture (e.g., of one or more “invocation gestures”) of a user; and/or (2) detecting that a gaze of the user is directed at an assistant device that provides an automated assistant interface (graphical and/or audible) of the automated assistant. For example, the previously dormant function(s) can be invoked in response to detecting the particular gesture, detecting that the gaze of the user is directed at an assistant device for at least a threshold amount of time, and optionally that the particular gesture and the directed gaze of the user co-occur or occur within a threshold temporal proximity of one another.Type: GrantFiled: December 3, 2020Date of Patent: February 1, 2022Assignee: Google LLCInventors: Yuan Yuan, Kenneth Mixter, Tuan Nguyen
-
Publication number: 20210400348Abstract: Systems and techniques are provided for automation and recommendation based on device control protocols. HDMI-CEC device control data may be received from a connected electronic device that may be connected to an electronic display device. The HDMI-CEC device control data may be based on a HDMI-CEC device control signal from the electronic display device. Signals may be received from sensor devices located in a same structure as the electronic display device. The HDMI-CEC device control data and signals from the sensor devices may be input to the machine learning system. The machine learning system may generate a control signal for a device. The control signal may be sent to the device to be implemented by the device.Type: ApplicationFiled: June 17, 2020Publication date: December 23, 2021Inventors: Kenneth Mixter, Marci Meingast, Jennifer Yick
-
Publication number: 20210334070Abstract: The various implementations described herein include methods, devices, and systems for attending to a presenting user. In one aspect, a method is performed at an electronic device that includes an image sensor, microphones, a display, processor(s), and memory. The device (1) obtains audio signals by concurrently receiving audio data at each microphone; (2) determines based on the obtained audio signals that a person is speaking in a vicinity of the device; (3) obtains video data from the image sensor; (4) determines via the video data that the person is not within a field of view of the image sensor; (5) reorients the electronic device based on differences in the received audio data; (6) after reorienting the electronic device, obtains second video data from the image sensor and determines that the person is within the field of view; and (7) attends to the person by directing the display toward the person.Type: ApplicationFiled: July 8, 2021Publication date: October 28, 2021Inventors: Yuan Yuan, Johan Schalkwyk, Kenneth Mixter
-
Publication number: 20210249015Abstract: The various implementations described herein include methods and systems for determining device leadership among voice interface devices. In one aspect, a method is performed at a first electronic device of a plurality of electronic devices, each having microphones, a speaker, processors, and memory storing programs for execution by the processors. The first device detects a voice input. It determines a device state and a relevance of the voice input. It identifies a subset of electronic devices from the plurality to which the voice input is relevant. In accordance with a determination that the subset includes the first device, the first device determines a first score of a criterion associated with the voice input and receives second scores of the criterion from other devices in the subset. In accordance with a determination that the first score is higher than the second scores, the first device responds to the detected input.Type: ApplicationFiled: April 27, 2021Publication date: August 12, 2021Inventors: Kenneth Mixter, Diego Melendo Casado, Alexander Houston Gruenstein, Terry Tai, Christopher Thaddeus Hughes, Matthew Nirvan Sharifi
-
Patent number: 11086597Abstract: The various implementations described herein include methods, devices, and systems for attending to a presenting user. In one aspect, a method is performed at an electronic device that includes an image sensor, microphones, a display, processor(s), and memory. The device (1) obtains audio signals by concurrently receiving audio data at each microphone; (2) determines based on the obtained audio signals that a person is speaking in a vicinity of the device; (3) obtains video data from the image sensor; (4) determines via the video data that the person is not within a field of view of the image sensor; (5) reorients the electronic device based on differences in the received audio data; (6) after reorienting the electronic device, obtains second video data from the image sensor and determines that the person is within the field of view; and (7) attends to the person by directing the display toward the person.Type: GrantFiled: August 14, 2018Date of Patent: August 10, 2021Assignee: GOOGLE LLCInventors: Yuan Yuan, Johan Schalkwyk, Kenneth Mixter
-
Publication number: 20210232231Abstract: Techniques are described herein for reducing false positives in vision sensor-equipped assistant devices. In various implementations, initial image frame(s) may be obtained from vision sensor(s) of an assistant device and analyzed to classify a particular region of the initial image frames as being likely to contain visual noise. Subsequent image frame(s) obtained from the vision sensor(s) may then be analyzed to detect actionable user-provided visual cue(s), in a manner that reduces or eliminates false positives. In some implementations, no analysis may be performed on the particular region of the subsequent image frame(s). Additionally or alternatively, in some implementations, a first candidate visual cue detected within the particular region may be weighted less heavily than a second candidate visual cue detected elsewhere in the one or more subsequent image frames. An automated assistant may then take responsive action based on the detected actionable visual cue(s).Type: ApplicationFiled: April 13, 2021Publication date: July 29, 2021Inventors: Kenneth Mixter, Yuan Yuan, Tuan Nguyen
-
Publication number: 20210225387Abstract: A method at an electronic device with one or more microphones and a speaker, the electronic device configured to be responsive to any of a plurality of affordances including a voice-based affordance, includes determining background noise of an environment associated with the electronic device, and before detecting the voice-based affordance: determining whether the background noise would interfere with recognition of the hotword in voice inputs detected by the electronic device, and if so, indicating to a user to use an affordance other than the voice-based affordance.Type: ApplicationFiled: April 8, 2021Publication date: July 22, 2021Inventor: Kenneth Mixter
-
Publication number: 20210201927Abstract: A method at an electronic device with one or more microphones and a speaker, the electronic device configured to be responsive to any of a plurality of affordances including a voice-based affordance, includes determining background noise of an environment associated with the electronic device, and before detecting the voice-based affordance: determining whether the background noise would interfere with recognition of the hotword in voice inputs detected by the electronic device, and if so, indicating to a user to use an affordance other than the voice-based affordance.Type: ApplicationFiled: August 17, 2020Publication date: July 1, 2021Inventor: Kenneth Mixter
-
Patent number: 11023051Abstract: Techniques are described herein for reducing false positives in vision sensor-equipped assistant devices. In various implementations, initial image frame(s) may be obtained from vision sensor(s) of an assistant device and analyzed to classify a particular region of the initial image frames as being likely to contain visual noise. Subsequent image frame(s) obtained from the vision sensor(s) may then be analyzed to detect actionable user-provided visual cue(s), in a manner that reduces or eliminates false positives. In some implementations, no analysis may be performed on the particular region of the subsequent image frame(s). Additionally or alternatively, in some implementations, a first candidate visual cue detected within the particular region may be weighted less heavily than a second candidate visual cue detected elsewhere in the one or more subsequent image frames. An automated assistant may then take responsive action based on the detected actionable visual cue(s).Type: GrantFiled: May 4, 2018Date of Patent: June 1, 2021Assignee: GOOGLE LLCInventors: Kenneth Mixter, Yuan Yuan, Tuan Nguyen
-
Patent number: 11024311Abstract: The various implementations described herein include methods and systems for determining device leadership among voice interface devices. In one aspect, a method is performed at a first electronic device of a plurality of electronic devices, each having microphones, a speaker, processors, and memory storing programs for execution by the processors. The first device detects a voice input. It determines a device state and a relevance of the voice input. It identifies a subset of electronic devices from the plurality to which the voice input is relevant. In accordance with a determination that the subset includes the first device, the first device determines a first score of a criterion associated with the voice input and receives second scores of the criterion from other devices in the subset. In accordance with a determination that the first score is higher than the second scores, the first device responds to the detected input.Type: GrantFiled: February 10, 2020Date of Patent: June 1, 2021Assignee: GOOGLE LLCInventors: Kenneth Mixter, Diego Melendo Casado, Alexander Houston Gruenstein, Terry Tai, Christopher Thaddeus Hughes, Matthew Nirvan Sharifi
-
Publication number: 20210117720Abstract: Methods, apparatus, systems, and computer-readable media are provided for generating and/or adapting automated assistant content according to a distance of a user relative to an automated assistant interface that renders the automated assistant content. For instance, the automated assistant can provide data for a client device to render. The client device can request additional data when the user relocates closer to, or further from, the client device. In some implementations, a request for additional data can identify a distance between the user and the client device. In this way, the additional data can be generated or selected according to the distance in the request. Other implementations can allow an automated assistant to determine an active user from a group of users in an environment, and determine a distance between the active user and the client device in order that any rendered content can be tailored for the active user.Type: ApplicationFiled: December 28, 2020Publication date: April 22, 2021Inventors: Tuan Nguyen, Kenneth Mixter, Yuan Yuan
-
Publication number: 20210089125Abstract: Invoking one or more previously dormant functions of an automated assistant in response to detecting, based on processing of vision data from one or more vision components: (1) a particular gesture (e.g., of one or more “invocation gestures”) of a user; and/or (2) detecting that a gaze of the user is directed at an assistant device that provides an automated assistant interface (graphical and/or audible) of the automated assistant. For example, the previously dormant function(s) can be invoked in response to detecting the particular gesture, detecting that the gaze of the user is directed at an assistant device for at least a threshold amount of time, and optionally that the particular gesture and the directed gaze of the user co-occur or occur within a threshold temporal proximity of one another.Type: ApplicationFiled: December 3, 2020Publication date: March 25, 2021Inventors: Yuan Yuan, Kenneth Mixter, Tuan Nguyen
-
Publication number: 20210082258Abstract: A method at an electronic device with an array of indicator lights includes: obtaining first visual output instructions stored at the electronic device, where the first visual output instructions control operation of the array of indicator lights based on operating state of the electronic device; receiving a voice input; obtaining from a remote system a response to the voice input and second visual output instructions, where the second visual output instructions are provided by the remote system along with the response in accordance with a determination that the voice input satisfies one or more criteria; executing the response; and displaying visual output on the array of indicator lights in accordance with the second visual output instructions, where otherwise in absence of the second visual output instructions the electronic device displays visual output on the array of indicator lights in accordance with the first visual output instructions.Type: ApplicationFiled: November 25, 2020Publication date: March 18, 2021Inventors: Kenneth Mixter, Yuan Yuan
-
Publication number: 20210012779Abstract: A method at an electronic device with one or more microphones and a speaker includes receiving a first voice input; comparing the first voice input to one or more voice models; based on the comparing, determining whether the first voice input corresponds to any of a plurality of occupants, and according to the determination, authenticating an occupant and presenting a response, or restricting functionality of the electronic device.Type: ApplicationFiled: August 17, 2020Publication date: January 14, 2021Inventors: Kenneth Mixter, Diego Melendo Casado, Bibo Xu
-
Patent number: 10890969Abstract: Invoking one or more previously dormant functions of an automated assistant in response to detecting, based on processing of vision data from one or more vision components: (1) a particular gesture (e.g., of one or more “invocation gestures”) of a user; and/or (2) detecting that a gaze of the user is directed at an assistant device that provides an automated assistant interface (graphical and/or audible) of the automated assistant. For example, the previously dormant function(s) can be invoked in response to detecting the particular gesture, detecting that the gaze of the user is directed at an assistant device for at least a threshold amount of time, and optionally that the particular gesture and the directed gaze of the user co-occur or occur within a threshold temporal proximity of one another.Type: GrantFiled: May 4, 2018Date of Patent: January 12, 2021Assignee: GOOGLE LLCInventors: Yuan Yuan, Kenneth Mixter, Tuan Nguyen
-
Publication number: 20210005202Abstract: A first electronic device of a local group of connected electronic devices receives a first voice command including a request for a first operation, assigns a first target device from among a local group of connected electronic devices as an in-focus device for performing the first operation, causes the first operation to be performed by the first target device via operation of a server-implemented common network service, receives a second voice command including a request for a second operation, and based on a determination that the second voice command does not include an explicit designation of a second target device and a determination that the second operation can be performed by the first target device, assigning the first target device as the in-focus device for performing the second operation.Type: ApplicationFiled: September 21, 2020Publication date: January 7, 2021Inventors: KENNETH MIXTER, TOMER SHEKEL, TUAN ANH NGUYEN