Patents Examined by Oluwadamilola M. Ogunbiyi
  • Patent number: 11967425
    Abstract: A system and method for voice control of operating room electrical equipment. The system comprises an electrosurgical generator a controller with a memory, a graphical user interface controlled by said controller, a power module, a field programmable gate array, and a voice recognition module connected to said field programmable gate array, a data storage connected to said controller in said electrosurgical generator; and electrical operating room equipment connected to said voice recognition module, wherein said electrical operating room equipment is configured to receive and decrypt encrypted commands from said voice recognition module. The electrical operating room equipment may said electrosurgical generator or a robotic surgical system or other electrical equipment in an operating room. The connection between said electrical operating room equipment and said voice recognition module may be wireless.
    Type: Grant
    Filed: September 10, 2021
    Date of Patent: April 23, 2024
    Assignee: Jerome Canady Research Institute
    Inventors: Jerome Canady, Changhan Jun, Taisen Zhuang
  • Patent number: 11948582
    Abstract: A system includes an interactive voice recognition (IVR) module arranged to perform a speech conversation with a first user and receive a first user identifier, where the speech conversation has an interaction context based on a subject matter of the speech conversation. The system includes a datastore arranged to store a group of active words associated with the interaction context and store first user voiceprints derived from pre-captured audio of the first user, where each active word is selected based on one or more selection criterion derived from conversations of a population of users. An automated speech recognition (ASR) module is arranged to perform speech recognition of the first user audio provided during the speech conversation. A voice biometric (VB) module is arranged to generate captured voiceprints and determine a similarity score based on comparisons of captured voiceprints with first user voiceprints.
    Type: Grant
    Filed: March 25, 2019
    Date of Patent: April 2, 2024
    Assignee: OMILIA NATURAL LANGUAGE SOLUTIONS LTD.
    Inventors: Themos Stafylakis, Petr Mizera, Dimitris Vassos
  • Patent number: 11949635
    Abstract: Method and system to control a conversational bot uses a directed acyclic graph to specify a desired conversation flow. A graph node has synthetic conversation transcripts annotated with events, wherein an event in a synthetic conversation transcript has preconfigured event expressions that represent ways in which dialogue at the node can unfold. During an on-going conversation with an actor, the system provides a data model uniquely associated with the conversation and that specifies a linear sequence of observations. The data model includes events representing semantically-related conversation fragments located in annotated historical conversation transcripts. In response to receipt of an input in association with a current graph node, the system determines whether the input extends an event in the synthetic conversation transcript associated with the node. If so, a response that continues a current conversation flow in the graph is provided.
    Type: Grant
    Filed: February 14, 2022
    Date of Patent: April 2, 2024
    Assignee: Drift.com, Inc.
    Inventors: Jeffrey D. Orkin, Luke W. Van Seters, Joseph Sorbonne Demple, Jason D. Crouse
  • Patent number: 11942085
    Abstract: Techniques for naming devices via voice commands are described herein. For instance, a user may issue a voice command to a voice-controlled device stating, “you are the kitchen device”. Thereafter, the device may respond to voice commands directed, by name, to this device. For instance, the user may issue a voice command requesting to “play music on my kitchen device”. Given that the user has configured the device to respond to this name, the device may respond to the command by outputting the requested music.
    Type: Grant
    Filed: October 30, 2020
    Date of Patent: March 26, 2024
    Assignee: AMAZON TECHNOLOGIES, INC.
    Inventors: Rohan Mutagi, Isaac Michael Taylor
  • Patent number: 11934785
    Abstract: Methods and systems for performing multi-task learning of query intent and named entities are provided. One method includes receiving a query comprising query text. The method further includes providing the query text to a neural network model implemented on a computing system, the neural network model having a plurality of layers, wherein at least one layer comprises a plurality of loss functions including a named entity tag learning loss function and an intent classification loss function. The method also includes obtaining, from the neural network model, an identification of a named entity and a query intent derived from the query text. A query response may be formulated based, at least in part, on the named entity and query intent.
    Type: Grant
    Filed: October 25, 2021
    Date of Patent: March 19, 2024
    Assignee: Target Brands, Inc.
    Inventors: Shalin Shah, Ryan Siskind
  • Patent number: 11935548
    Abstract: A multi-channel signal encoding method includes obtaining a multi-channel signal of a current frame; determining an initial multi-channel parameter of the current frame; determining a difference parameter based on the initial multi-channel parameter of the current frame and multi-channel parameters of previous K frames of the current frame, where the difference parameter represents a difference between the initial multi-channel parameter of the current frame and the multi-channel parameters of the previous K frames, and K is an integer greater than or equal to one; determining a multi-channel parameter of the current frame based on the difference parameter and a characteristic parameter of the current frame; and encoding the multi-channel signal based on the multi-channel parameter of the current frame.
    Type: Grant
    Filed: August 20, 2021
    Date of Patent: March 19, 2024
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Zexin Liu, Xingtao Zhang, Haiting Li, Lei Miao
  • Patent number: 11915712
    Abstract: An audio encoder for encoding an audio signal includes: a first encoding processor for encoding a first audio signal portion in a frequency domain, wherein the first encoding processor includes: a time frequency converter for converting the first audio signal portion into a frequency domain representation having spectral lines up to a maximum frequency of the first audio signal portion; a spectral encoder for encoding the frequency domain representation; a second encoding processor for encoding a second different audio signal portion in the time domain; a cross-processor for calculating, from the encoded spectral representation of the first audio signal portion, initialization data of the second encoding processor, so that the second encoding processing is initialized to encode the second audio signal portion immediately following the first audio signal portion in time in the audio signal; a controller configured for analyzing the audio signal and for determining, which portion of the audio signal is the firs
    Type: Grant
    Filed: November 1, 2021
    Date of Patent: February 27, 2024
    Assignee: Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V.
    Inventors: Sascha Disch, Martin Dietz, Markus Multrus, Guillaume Fuchs, Emmanuel Ravelli, Matthias Neusinger, Markus Schnell, Benjamin Schubert, Bernhard Grill
  • Patent number: 11908474
    Abstract: [Problem] Provided is a system that can objectively evaluate a person who makes a presentation (presenter) [Solution] A presentation evaluation system 1 includes: a voice analysis unit 3 that analyzes a content of a conversation, a presentation material related information storage unit 5 that stores information related to a presentation material, a keyword storage unit 7 that stores information related to a keyword in each page of the presentation material, a related term storage unit 9 that stores a related term of each keyword, and an evaluation unit 11 that evaluates the content of the conversation analyzed by the voice analysis unit 3 or a person who had the conversation.
    Type: Grant
    Filed: December 28, 2021
    Date of Patent: February 20, 2024
    Assignee: Interactive Solutions Corp.
    Inventor: Kiyoshi Sekine
  • Patent number: 11907298
    Abstract: According to some disclosed embodiments an action is performed by an electronic social agent. The electronic social agent collects a first dataset indicating the user's state, the user's environment state, and a first user response to the performed action. Then, it is determined whether it is desirable to collect a second response from the user and, if so, it is further determined whether to generate a question to be presented to the user based on an analysis of a first dataset and the first user response. Then, an optimal time for presenting the question to the user is determined. A question that is based on the collected data and the first user response is generated by the electronic social agent for actively collecting an additional user response. Then, based on the collected additional user response, the decision-making model of the electronic social agent is updated and improved.
    Type: Grant
    Filed: February 3, 2021
    Date of Patent: February 20, 2024
    Assignee: INTUITION ROBOTICS, LTD.
    Inventors: Shay Zweig, Eldar Ron, Alex Keagel, Itai Mendelsohn, Roy Amir, Dor Skuler
  • Patent number: 11900923
    Abstract: Systems and processes for operating an intelligent automated assistant are provided. In one example process, a speech input is received from a user. In response to determining that the speech input corresponds to a user intent of obtaining information associated with a user experience of the user, one or more parameters referencing a user experience of the user are identified. Metadata associated with the referenced user experience is obtained from an experiential data structure. Based on the metadata, one or more media items associated with the referenced are retrieved based on the metadata. The one or more media items associated with the referenced user experience are output together.
    Type: Grant
    Filed: September 7, 2021
    Date of Patent: February 13, 2024
    Assignee: Apple Inc.
    Inventors: Marcos Regis Vescovi, Eric M. G. Circlaeys, Richard Warren, Jeffrey Traer Bernstein, Matthaeus Krenn
  • Patent number: 11900947
    Abstract: Embodiments of the present invention provide methods and systems for performing automatic diarisation of sound recordings including speech from one more speakers. The automatic diarisation has a development or training phase and a utilisation or evaluation phase. In the development or training phase background models and hyperparameters are generated from already annotated sound recordings. These models and hyperparameters are applied during the evaluation or utilisation phase to diarise new or not previously diarised or annotated recordings.
    Type: Grant
    Filed: February 24, 2021
    Date of Patent: February 13, 2024
    Assignee: FTR LABS PTY LTD
    Inventors: Houman Ghaemmaghami, Shahram Kalantari, David Dean, Subramanian Sridharan
  • Patent number: 11881212
    Abstract: [Problem] To provide a system capable of correcting an error in speech recognition by easier work than in the prior art. [Solution] A conversation analysis system comprising: a speech analysis unit 3 for analyzing contents contained in a conversation; a topics grasping unit 5 for grasping topics of the conversation; a topics-related correction terms storage unit 7 for, with respect to the respective topics, storing pronunciations to be speech-recognized or speech-recognized terms, and candidates for correction terms related to the pronunciations to be speech-recognized or the speech-recognized terms; and a conversation correction unit 9 for correcting the conversation analyzed by the speech analysis unit 3.
    Type: Grant
    Filed: January 19, 2021
    Date of Patent: January 23, 2024
    Assignee: Interactive Solutions Corp.
    Inventor: Kiyoshi Sekine
  • Patent number: 11875126
    Abstract: A method for training a dialog generation model includes: acquiring a dialog data set, and encoding a post in the dialog data set by using an encoder in the dialog generation model to obtain an encoded representation of the post; fusing, by using a decoder in the dialog generation model, the encoded representation of the post and knowledge information corresponding to the post that is obtained from a knowledge base question answering model through transfer learning to obtain a predicted response corresponding to the post; determining a value of a loss function of the dialog generation model based on the predicted response and the annotated response that correspond to the post; and updating a model parameter of the dialog generation model based on the value of the loss function. A method, an apparatus, a device, and a computer storage medium for generating a dialog are also provided.
    Type: Grant
    Filed: July 6, 2021
    Date of Patent: January 16, 2024
    Assignees: SHENZHEN INSTITUTES OF ADVANCED TECHNOLOGY, CHINESE ACADEMY OF SCIENCES, TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Min Yang, Wei Bi, Xiao Jiang Liu, Lei Chen, Ting Ting Huang
  • Patent number: 11850059
    Abstract: Disclosed is a method of identifying, by at least one processor of a device, a cognitive function state of a user. The method may include inputting user information, at least one voice data, and speech time information related to the at least one voice data to a cognitive function state identification model, and identifying a cognitive function state of the user based on first output data that is output by the cognitive function state identification model.
    Type: Grant
    Filed: May 23, 2023
    Date of Patent: December 26, 2023
    Assignee: HAII CORP.
    Inventors: Jee Hang Lee, Ho Yung Kim, Dong Han Kim, Hye Bin Hwang, Chan Yeong Park, Ji An Choi, Hyun Jeong Ko, Su Yeon Park, Byung Hun Yun
  • Patent number: 11854539
    Abstract: Systems and processes for operating an intelligent automated assistant are provided. In one example process, a speech input is received from a user. In response to determining that the speech input corresponds to a user intent of obtaining information associated with a user experience of the user, one or more parameters referencing a user experience of the user are identified. Metadata associated with the referenced user experience is obtained from an experiential data structure. Based on the metadata, one or more media items associated with the referenced are retrieved based on the metadata. The one or more media items associated with the referenced user experience are output together.
    Type: Grant
    Filed: August 11, 2020
    Date of Patent: December 26, 2023
    Assignee: Apple Inc.
    Inventors: Marcos Regis Vescovi, Eric M. G. Circlaeys, Richard Warren, Jeffrey Traer Bernstein, Matthaeus Krenn
  • Patent number: 11836451
    Abstract: A method for maintaining a dialogue state associated with a dialogue between a user and a digital system includes receiving, by a dialogue state tracker associated with the digital system, a representation of a user communication, updating, by the dialogue state tracker, the dialogue state and providing a system response based on the updated dialogue state. The dialogue state is updated by evaluating, based on the representation of the user communication, a plurality of member scores corresponding to a plurality of ontology members of an ontology set, and selecting, based on the plurality of member scores, zero or more of the plurality of ontology members to add to or remove from the dialogue state.
    Type: Grant
    Filed: February 19, 2021
    Date of Patent: December 5, 2023
    Assignee: salesforce.com, inc.
    Inventors: Victor Zhong, Caiming Xiong
  • Patent number: 11837214
    Abstract: Various embodiments of the present disclosure evaluate transcription accuracy. In some implementations, the system normalizes a first transcription of an audio file and a baseline transcription of the audio file. The baseline transcription can be used as an accurate transcription of the audio file. The system can further determine an error rate of the first transcription by aligning each portion of the first transcription with the portion of the baseline transcription, and assigning a label to each portion based on a comparison of the portion of the first transcription with the portion of the baseline transcription.
    Type: Grant
    Filed: October 29, 2020
    Date of Patent: December 5, 2023
    Assignee: United Services Automobile Association (USAA)
    Inventors: Michael J. Szentes, Carlos Chavez, Robert E. Lewis, Nicholas S. Walker
  • Patent number: 11830499
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for training a dialog system using user feedback. According to one implementation, a method includes receiving, by a dialog engine and from a computing device, a voice input; determining, by the dialog engine, that the received voice input is classified as feedback to an answer of a question; identifying, by the dialog engine, a predetermined feedback score associated with the feedback; and adjusting a confidence score associated with the question and the answer based on the predetermined feedback score.
    Type: Grant
    Filed: February 28, 2022
    Date of Patent: November 28, 2023
    Assignee: Google LLC
    Inventors: Gabriel Taubman, Andrew W. Hogue, John J. Lee
  • Patent number: 11798543
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for suppressing hotword triggers when detecting a hotword in recorded media are disclosed. In one aspect, a method includes the actions of receiving, by a computing device, audio corresponding to playback of an item of media content. The actions further include determining, by the computing device, that the audio includes an utterance of a predefined hotword and that the audio includes an audio watermark. The actions further include analyzing, by the computing device, the audio watermark. The actions further include based on analyzing the audio watermark, determining, by the computing device, whether to perform speech recognition on a portion of the audio following the predefined hotword.
    Type: Grant
    Filed: January 13, 2022
    Date of Patent: October 24, 2023
    Assignee: Google LLC
    Inventor: Ricardo Antonio Garcia
  • Patent number: 11776531
    Abstract: Methods, systems, and apparatus for performing speech recognition. In some implementations, acoustic data representing an utterance is obtained. The acoustic data corresponds to time steps in a series of time steps. One or more computers process scores indicative of the acoustic data using a recurrent neural network to generate a sequence of outputs. The sequence of outputs indicates a likely output label from among a predetermined set of output labels. The predetermined set of output labels includes output labels that respectively correspond to different linguistic units and to a placeholder label that does not represent a classification of acoustic data. The recurrent neural network is configured to use an output label indicated for a previous time step to determine an output label for the current time step. The generated sequence of outputs is processed to generate a transcription of the utterance, and the transcription of the utterance is provided.
    Type: Grant
    Filed: May 28, 2020
    Date of Patent: October 3, 2023
    Assignee: Google LLC
    Inventors: Hasim Sak, Sean Matthew Shannon