Patents by Inventor Kenneth Mixter

Kenneth Mixter has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20220277727
    Abstract: A method for proactive notifications in a voice interface device includes: receiving a first user voice request for an action with an future performance time; assigning the first user voice request to a voice assistant service for performance; subsequent to the receiving, receiving a second user voice request and in response to the second user voice request initiating a conversation with the user; and during the conversation: receiving a notification from the voice assistant service of performance of the action; triggering a first audible announcement to the user to indicate a transition from the conversation and interrupting the conversation; triggering a second audible announcement to the user to indicate performance of the action; and triggering a third audible announcement to the user to indicate a transition back to the conversation and rejoining the conversation.
    Type: Application
    Filed: May 16, 2022
    Publication date: September 1, 2022
    Inventors: Kenneth Mixter, Daniel Colish, Tuan Nguyen
  • Patent number: 11392213
    Abstract: Techniques are described herein for reducing false positives in vision sensor-equipped assistant devices. In various implementations, initial image frame(s) may be obtained from vision sensor(s) of an assistant device and analyzed to classify a particular region of the initial image frames as being likely to contain visual noise. Subsequent image frame(s) obtained from the vision sensor(s) may then be analyzed to detect actionable user-provided visual cue(s), in a manner that reduces or eliminates false positives. In some implementations, no analysis may be performed on the particular region of the subsequent image frame(s). Additionally or alternatively, in some implementations, a first candidate visual cue detected within the particular region may be weighted less heavily than a second candidate visual cue detected elsewhere in the one or more subsequent image frames. An automated assistant may then take responsive action based on the detected actionable visual cue(s).
    Type: Grant
    Filed: April 13, 2021
    Date of Patent: July 19, 2022
    Assignee: GOOGLE LLC
    Inventors: Kenneth Mixter, Yuan Yuan, Tuan Nguyen
  • Patent number: 11355116
    Abstract: An electronic device configures a device-agnostic voice assistant library for execution on the electronic device based on the electronic device having a first device type. The electronic device also selects an implementation for the voice assistant library. After the configuring, the electronic device receives a verbal input from a user. It extracts request information from the verbal input by processing the verbal input using the voice assistant library executing on the device. It transmits a request to a remote system, the request including the extracted request information. The electronic device receives a response to the request. The response is generated by the remote system in accordance with the extracted request information. The electronic device performs an operation in accordance with the response by one or more voice processing modules of the configured voice assistant library.
    Type: Grant
    Filed: May 29, 2020
    Date of Patent: June 7, 2022
    Assignee: Google LLC
    Inventors: Kenneth Mixter, Raunaq Shah
  • Patent number: 11335319
    Abstract: A method for proactive notifications in a voice interface device includes: receiving a first user voice request for an action with an future performance time; assigning the first user voice request to a voice assistant service for performance; subsequent to the receiving, receiving a second user voice request and in response to the second user voice request initiating a conversation with the user; and during the conversation: receiving a notification from the voice assistant service of performance of the action; triggering a first audible announcement to the user to indicate a transition from the conversation and interrupting the conversation; triggering a second audible announcement to the user to indicate performance of the action; and triggering a third audible announcement to the user to indicate a transition back to the conversation and rejoining the conversation.
    Type: Grant
    Filed: June 5, 2020
    Date of Patent: May 17, 2022
    Assignee: Google LLC
    Inventors: Kenneth Mixter, Daniel Colish, Tuan Nguyen
  • Patent number: 11237630
    Abstract: Invoking one or more previously dormant functions of an automated assistant in response to detecting, based on processing of vision data from one or more vision components: (1) a particular gesture (e.g., of one or more “invocation gestures”) of a user; and/or (2) detecting that a gaze of the user is directed at an assistant device that provides an automated assistant interface (graphical and/or audible) of the automated assistant. For example, the previously dormant function(s) can be invoked in response to detecting the particular gesture, detecting that the gaze of the user is directed at an assistant device for at least a threshold amount of time, and optionally that the particular gesture and the directed gaze of the user co-occur or occur within a threshold temporal proximity of one another.
    Type: Grant
    Filed: December 3, 2020
    Date of Patent: February 1, 2022
    Assignee: Google LLC
    Inventors: Yuan Yuan, Kenneth Mixter, Tuan Nguyen
  • Publication number: 20210400348
    Abstract: Systems and techniques are provided for automation and recommendation based on device control protocols. HDMI-CEC device control data may be received from a connected electronic device that may be connected to an electronic display device. The HDMI-CEC device control data may be based on a HDMI-CEC device control signal from the electronic display device. Signals may be received from sensor devices located in a same structure as the electronic display device. The HDMI-CEC device control data and signals from the sensor devices may be input to the machine learning system. The machine learning system may generate a control signal for a device. The control signal may be sent to the device to be implemented by the device.
    Type: Application
    Filed: June 17, 2020
    Publication date: December 23, 2021
    Inventors: Kenneth Mixter, Marci Meingast, Jennifer Yick
  • Publication number: 20210334070
    Abstract: The various implementations described herein include methods, devices, and systems for attending to a presenting user. In one aspect, a method is performed at an electronic device that includes an image sensor, microphones, a display, processor(s), and memory. The device (1) obtains audio signals by concurrently receiving audio data at each microphone; (2) determines based on the obtained audio signals that a person is speaking in a vicinity of the device; (3) obtains video data from the image sensor; (4) determines via the video data that the person is not within a field of view of the image sensor; (5) reorients the electronic device based on differences in the received audio data; (6) after reorienting the electronic device, obtains second video data from the image sensor and determines that the person is within the field of view; and (7) attends to the person by directing the display toward the person.
    Type: Application
    Filed: July 8, 2021
    Publication date: October 28, 2021
    Inventors: Yuan Yuan, Johan Schalkwyk, Kenneth Mixter
  • Publication number: 20210249015
    Abstract: The various implementations described herein include methods and systems for determining device leadership among voice interface devices. In one aspect, a method is performed at a first electronic device of a plurality of electronic devices, each having microphones, a speaker, processors, and memory storing programs for execution by the processors. The first device detects a voice input. It determines a device state and a relevance of the voice input. It identifies a subset of electronic devices from the plurality to which the voice input is relevant. In accordance with a determination that the subset includes the first device, the first device determines a first score of a criterion associated with the voice input and receives second scores of the criterion from other devices in the subset. In accordance with a determination that the first score is higher than the second scores, the first device responds to the detected input.
    Type: Application
    Filed: April 27, 2021
    Publication date: August 12, 2021
    Inventors: Kenneth Mixter, Diego Melendo Casado, Alexander Houston Gruenstein, Terry Tai, Christopher Thaddeus Hughes, Matthew Nirvan Sharifi
  • Patent number: 11086597
    Abstract: The various implementations described herein include methods, devices, and systems for attending to a presenting user. In one aspect, a method is performed at an electronic device that includes an image sensor, microphones, a display, processor(s), and memory. The device (1) obtains audio signals by concurrently receiving audio data at each microphone; (2) determines based on the obtained audio signals that a person is speaking in a vicinity of the device; (3) obtains video data from the image sensor; (4) determines via the video data that the person is not within a field of view of the image sensor; (5) reorients the electronic device based on differences in the received audio data; (6) after reorienting the electronic device, obtains second video data from the image sensor and determines that the person is within the field of view; and (7) attends to the person by directing the display toward the person.
    Type: Grant
    Filed: August 14, 2018
    Date of Patent: August 10, 2021
    Assignee: GOOGLE LLC
    Inventors: Yuan Yuan, Johan Schalkwyk, Kenneth Mixter
  • Publication number: 20210232231
    Abstract: Techniques are described herein for reducing false positives in vision sensor-equipped assistant devices. In various implementations, initial image frame(s) may be obtained from vision sensor(s) of an assistant device and analyzed to classify a particular region of the initial image frames as being likely to contain visual noise. Subsequent image frame(s) obtained from the vision sensor(s) may then be analyzed to detect actionable user-provided visual cue(s), in a manner that reduces or eliminates false positives. In some implementations, no analysis may be performed on the particular region of the subsequent image frame(s). Additionally or alternatively, in some implementations, a first candidate visual cue detected within the particular region may be weighted less heavily than a second candidate visual cue detected elsewhere in the one or more subsequent image frames. An automated assistant may then take responsive action based on the detected actionable visual cue(s).
    Type: Application
    Filed: April 13, 2021
    Publication date: July 29, 2021
    Inventors: Kenneth Mixter, Yuan Yuan, Tuan Nguyen
  • Publication number: 20210225387
    Abstract: A method at an electronic device with one or more microphones and a speaker, the electronic device configured to be responsive to any of a plurality of affordances including a voice-based affordance, includes determining background noise of an environment associated with the electronic device, and before detecting the voice-based affordance: determining whether the background noise would interfere with recognition of the hotword in voice inputs detected by the electronic device, and if so, indicating to a user to use an affordance other than the voice-based affordance.
    Type: Application
    Filed: April 8, 2021
    Publication date: July 22, 2021
    Inventor: Kenneth Mixter
  • Publication number: 20210201927
    Abstract: A method at an electronic device with one or more microphones and a speaker, the electronic device configured to be responsive to any of a plurality of affordances including a voice-based affordance, includes determining background noise of an environment associated with the electronic device, and before detecting the voice-based affordance: determining whether the background noise would interfere with recognition of the hotword in voice inputs detected by the electronic device, and if so, indicating to a user to use an affordance other than the voice-based affordance.
    Type: Application
    Filed: August 17, 2020
    Publication date: July 1, 2021
    Inventor: Kenneth Mixter
  • Patent number: 11023051
    Abstract: Techniques are described herein for reducing false positives in vision sensor-equipped assistant devices. In various implementations, initial image frame(s) may be obtained from vision sensor(s) of an assistant device and analyzed to classify a particular region of the initial image frames as being likely to contain visual noise. Subsequent image frame(s) obtained from the vision sensor(s) may then be analyzed to detect actionable user-provided visual cue(s), in a manner that reduces or eliminates false positives. In some implementations, no analysis may be performed on the particular region of the subsequent image frame(s). Additionally or alternatively, in some implementations, a first candidate visual cue detected within the particular region may be weighted less heavily than a second candidate visual cue detected elsewhere in the one or more subsequent image frames. An automated assistant may then take responsive action based on the detected actionable visual cue(s).
    Type: Grant
    Filed: May 4, 2018
    Date of Patent: June 1, 2021
    Assignee: GOOGLE LLC
    Inventors: Kenneth Mixter, Yuan Yuan, Tuan Nguyen
  • Patent number: 11024311
    Abstract: The various implementations described herein include methods and systems for determining device leadership among voice interface devices. In one aspect, a method is performed at a first electronic device of a plurality of electronic devices, each having microphones, a speaker, processors, and memory storing programs for execution by the processors. The first device detects a voice input. It determines a device state and a relevance of the voice input. It identifies a subset of electronic devices from the plurality to which the voice input is relevant. In accordance with a determination that the subset includes the first device, the first device determines a first score of a criterion associated with the voice input and receives second scores of the criterion from other devices in the subset. In accordance with a determination that the first score is higher than the second scores, the first device responds to the detected input.
    Type: Grant
    Filed: February 10, 2020
    Date of Patent: June 1, 2021
    Assignee: GOOGLE LLC
    Inventors: Kenneth Mixter, Diego Melendo Casado, Alexander Houston Gruenstein, Terry Tai, Christopher Thaddeus Hughes, Matthew Nirvan Sharifi
  • Publication number: 20210117720
    Abstract: Methods, apparatus, systems, and computer-readable media are provided for generating and/or adapting automated assistant content according to a distance of a user relative to an automated assistant interface that renders the automated assistant content. For instance, the automated assistant can provide data for a client device to render. The client device can request additional data when the user relocates closer to, or further from, the client device. In some implementations, a request for additional data can identify a distance between the user and the client device. In this way, the additional data can be generated or selected according to the distance in the request. Other implementations can allow an automated assistant to determine an active user from a group of users in an environment, and determine a distance between the active user and the client device in order that any rendered content can be tailored for the active user.
    Type: Application
    Filed: December 28, 2020
    Publication date: April 22, 2021
    Inventors: Tuan Nguyen, Kenneth Mixter, Yuan Yuan
  • Publication number: 20210089125
    Abstract: Invoking one or more previously dormant functions of an automated assistant in response to detecting, based on processing of vision data from one or more vision components: (1) a particular gesture (e.g., of one or more “invocation gestures”) of a user; and/or (2) detecting that a gaze of the user is directed at an assistant device that provides an automated assistant interface (graphical and/or audible) of the automated assistant. For example, the previously dormant function(s) can be invoked in response to detecting the particular gesture, detecting that the gaze of the user is directed at an assistant device for at least a threshold amount of time, and optionally that the particular gesture and the directed gaze of the user co-occur or occur within a threshold temporal proximity of one another.
    Type: Application
    Filed: December 3, 2020
    Publication date: March 25, 2021
    Inventors: Yuan Yuan, Kenneth Mixter, Tuan Nguyen
  • Publication number: 20210082258
    Abstract: A method at an electronic device with an array of indicator lights includes: obtaining first visual output instructions stored at the electronic device, where the first visual output instructions control operation of the array of indicator lights based on operating state of the electronic device; receiving a voice input; obtaining from a remote system a response to the voice input and second visual output instructions, where the second visual output instructions are provided by the remote system along with the response in accordance with a determination that the voice input satisfies one or more criteria; executing the response; and displaying visual output on the array of indicator lights in accordance with the second visual output instructions, where otherwise in absence of the second visual output instructions the electronic device displays visual output on the array of indicator lights in accordance with the first visual output instructions.
    Type: Application
    Filed: November 25, 2020
    Publication date: March 18, 2021
    Inventors: Kenneth Mixter, Yuan Yuan
  • Publication number: 20210012779
    Abstract: A method at an electronic device with one or more microphones and a speaker includes receiving a first voice input; comparing the first voice input to one or more voice models; based on the comparing, determining whether the first voice input corresponds to any of a plurality of occupants, and according to the determination, authenticating an occupant and presenting a response, or restricting functionality of the electronic device.
    Type: Application
    Filed: August 17, 2020
    Publication date: January 14, 2021
    Inventors: Kenneth Mixter, Diego Melendo Casado, Bibo Xu
  • Patent number: 10890969
    Abstract: Invoking one or more previously dormant functions of an automated assistant in response to detecting, based on processing of vision data from one or more vision components: (1) a particular gesture (e.g., of one or more “invocation gestures”) of a user; and/or (2) detecting that a gaze of the user is directed at an assistant device that provides an automated assistant interface (graphical and/or audible) of the automated assistant. For example, the previously dormant function(s) can be invoked in response to detecting the particular gesture, detecting that the gaze of the user is directed at an assistant device for at least a threshold amount of time, and optionally that the particular gesture and the directed gaze of the user co-occur or occur within a threshold temporal proximity of one another.
    Type: Grant
    Filed: May 4, 2018
    Date of Patent: January 12, 2021
    Assignee: GOOGLE LLC
    Inventors: Yuan Yuan, Kenneth Mixter, Tuan Nguyen
  • Publication number: 20210005202
    Abstract: A first electronic device of a local group of connected electronic devices receives a first voice command including a request for a first operation, assigns a first target device from among a local group of connected electronic devices as an in-focus device for performing the first operation, causes the first operation to be performed by the first target device via operation of a server-implemented common network service, receives a second voice command including a request for a second operation, and based on a determination that the second voice command does not include an explicit designation of a second target device and a determination that the second operation can be performed by the first target device, assigning the first target device as the in-focus device for performing the second operation.
    Type: Application
    Filed: September 21, 2020
    Publication date: January 7, 2021
    Inventors: KENNETH MIXTER, TOMER SHEKEL, TUAN ANH NGUYEN