Patents Examined by Neeraj Sharma
  • Patent number: 11978455
    Abstract: The present disclosure provides various embodiments of methods for intelligent active speaker identification and information handling systems (IHSs) utilizing such methods. In general, the methods disclosed herein may be used to accurately identify an active speaker in a communication session with an application or an IHS, regardless of whether the active speaker is alone, in a group environment, or using someone else's system or login to participate in the communication session. The methods disclosed herein may use voice processing technology and one or more voice identification databases (VIDs) to identify the active speaker in a communication session. In some embodiments, the disclosed methods may display the identity of the active speaker to other users or participants in the same communication session. In other embodiments, the disclosed methods may dynamically switch between user profiles or accounts during the communication session based on the identity of the active speaker.
    Type: Grant
    Filed: March 7, 2022
    Date of Patent: May 7, 2024
    Inventors: Douglas J. Peeler, Srinivas Kamepalli
  • Patent number: 11967324
    Abstract: The method S200 can include: at an aircraft, receiving an audio utterance from air traffic control S210, converting the audio utterance to text, determining commands from the text using a question-and-answer model S240, and optionally controlling the aircraft based on the commands S250. The method functions to automatically interpret flight commands from the air traffic control (ATC) stream.
    Type: Grant
    Filed: October 28, 2022
    Date of Patent: April 23, 2024
    Assignee: Merlin Labs, Inc.
    Inventors: Michael Pust, Joseph Bondaryk, Matthew George
  • Patent number: 11954452
    Abstract: A translation method includes: selecting a source word from a source sentence; generating mapping information including location information of the selected source word mapped to the selected source word in the source sentence; and correcting a target word, which is generated by translating the source sentence, based on location information of a feature value of the target word and the mapping information.
    Type: Grant
    Filed: July 15, 2022
    Date of Patent: April 9, 2024
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Jihyun Lee, Hwidong Na, Hoshik Lee
  • Patent number: 11955126
    Abstract: Systems, methods and software are disclosed for processing requests from users of an infotainment system. The method includes receiving a request from a user of the infotainment system. The method includes determining a domain for the received request based on information contained in the received request. The domain specifies one or more categories for the request. The method includes routing the received request to a virtual assistant assigned to handle requests for the determined domain. The virtual assistant is one of a plurality of virtual assistants respectively assigned to handle requests for a plurality of respectively assigned domains. The method includes transmitting a response to the request to the user.
    Type: Grant
    Filed: September 29, 2021
    Date of Patent: April 9, 2024
    Assignee: MITSUBISHI ELECTRIC AUTOMOTIVE AMERICA, INC.
    Inventors: Jacek Spiewla, Sorin M. Panainte
  • Patent number: 11955120
    Abstract: The disclosed computer-implemented method may include receiving input voice data synchronous with a visual state of a user interface of the third-party application, generating multiple sentence alternatives for the received input voice data, identifying a best sentence of the multiple sentence alternatives, executing a dialog script for the third-party application using the best sentence, the dialog script generating a response to the received voice data comprising output voice data and a corresponding visual response, and providing the visual response and the output voice data to the third-party application, the third-party application playing the output voice data synchronous with updating the user interface based on the visual response. Various other methods, systems, and computer-readable media are also disclosed.
    Type: Grant
    Filed: January 23, 2021
    Date of Patent: April 9, 2024
    Assignee: Alan AI, Inc.
    Inventors: Andrey Ryabov, Ramu V. Sunkara
  • Patent number: 11935539
    Abstract: A voice support server is used to provide voice control functionality to a third party application that does not natively support voice control functions. The voice support server implements a domain specific to the third party application that maintains a domain-specific language model (DLM) reflecting the functionality of the third party application. The DLM comprises a plurality of intent patterns corresponding to different commands and their possible variations that may be issued by the user, and maps each intent pattern to a corresponding action to be performed by the third party application. Received audio data is analyzed to determine one or more user utterances, which are transcribed and compared to the intent patterns of the DLM to identify an intent corresponding to the user utterance. The voice control module may then transmit instructions to the third party application to perform the action corresponding to the identified intent.
    Type: Grant
    Filed: January 24, 2020
    Date of Patent: March 19, 2024
    Assignee: Alan AI, Inc.
    Inventors: Andrey Ryabov, Anna Miroshnichenko, Evgeny Yusov, Alex Sotnikov
  • Patent number: 11935544
    Abstract: Implementations set forth herein relate to management of casting requests and user inputs at a rechargeable device, which provides access to an automated assistant and is capable of rendering data that is cast from a separate device. Casting requests can be handled by the rechargeable device despite a device SoC of the rechargeable device operating in a sleep mode. Furthermore, spoken utterances provided by a user for invoking the automated assistant can also be adaptively managed by the rechargeable device in order mitigate idle power consumption by the device SoC. Such spoken utterances can be initially processed by a digital signal processor (DSP), and, based on one or more features (e.g., voice characteristic, conformity to a particular invocation phrase, etc.) of the spoken utterance, the device SoC can be initialized for an amount of time that is selected based on the features of the spoken utterance.
    Type: Grant
    Filed: September 13, 2021
    Date of Patent: March 19, 2024
    Assignee: GOOGLE LLC
    Inventors: Andrei Pascovici, Victor Lin, Jianghai Zhu, Paul Gyugyi, Shlomi Regev
  • Patent number: 11922942
    Abstract: Devices and techniques are generally described for generating response templates for natural language processing. In various examples, a first knowledge graph comprising a plurality of entities may be received. First text data may be received for a first response template, the first text data defining a natural language input configured to invoke the first response template. A response definition may be received for the first response template, the response definition defining a response associated with the first response template. Natural language input data may be received. A determination may be made that the natural language input data corresponds to the natural language input configured to invoke the first response template. The first response template may be configured to generate natural language output data.
    Type: Grant
    Filed: June 4, 2020
    Date of Patent: March 5, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Emre Can Kilinc, Thomas Reno, John Zucchi, Joshua Kaplan
  • Patent number: 11922943
    Abstract: In general, this disclosure describes techniques for generating and evaluating automatic transcripts of audio recordings containing human speech. In some examples, a computing system is configured to: generate transcripts of a plurality of audio recordings; determine an error rate for each transcript by comparing the transcript to a reference transcript of the audio recording; receive, for each transcript, a subjective ranking selected from a plurality of subjective rank categories; determine, based on the error rates and subjective rankings, objective rank categories defined by error-rate ranges; and assign an objective ranking to a new machine-generated transcript of a new audio recording, based on the objective rank categories and an error rate of the new machine-generated transcript.
    Type: Grant
    Filed: January 26, 2021
    Date of Patent: March 5, 2024
    Assignee: Wells Fargo Bank, N.A.
    Inventors: Yong Yi Bay, Yang Angelina Yang, Menglin Cao
  • Patent number: 11922344
    Abstract: The exemplary embodiments described herein are related to techniques for automatically generating narratives about data based on communication goal data structures that are associated with configurable content blocks. The use of such communication goal data structures facilitates modes of operation whereby narratives can be generated in real-time and/or interactive manners.
    Type: Grant
    Filed: November 8, 2022
    Date of Patent: March 5, 2024
    Assignee: Narrative Science LLC
    Inventors: Lawrence Birnbaum, Kristian J. Hammond, Nathan Drew Nichols, Andrew R. Paley
  • Patent number: 11922095
    Abstract: A system may use multiple speech interface devices to interact with a user by speech. All or a portion of the speech interface devices may detect a user utterance and may initiate speech processing to determine a meaning or intent of the utterance. Within the speech processing, arbitration is employed to select one of the multiple speech interface devices to respond to the user utterance. Arbitration may be based in part on metadata that directly or indirectly indicates the proximity of the user to the devices, and the device that is deemed to be nearest the user may be selected to respond to the user utterance.
    Type: Grant
    Filed: January 22, 2018
    Date of Patent: March 5, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: James David Meyers, Shah Samir Pravinchandra, Yue Liu, Arlen Dean, Daniel Miller, Arindam Mandal
  • Patent number: 11922946
    Abstract: Systems and methods are disclosed for determining textual transcription from minute facial skin movements. In one implementation, a system may include at least one coherent light source, at least one sensor configured to receive light reflections from the at least one coherent light source; and a processor configured to control the at least one coherent light source to illuminate a region of a face of a user. The processor may receive from the at least one sensor, reflection signals indicative of coherent light reflected from the face in a time interval. The reflection signals may be analyzed to determine minute facial skin movements in the time interval. Then, based on the determined minute facial skin movements in the time interval, the processor may determine a sequence of words associated with the minute facial skin movements, and output a textual transcription corresponding with the determined sequence of words.
    Type: Grant
    Filed: March 10, 2023
    Date of Patent: March 5, 2024
    Assignee: Q (Cue) Ltd.
    Inventors: Aviad Maizels, Avi Barliya, Yonatan Wexler
  • Patent number: 11908479
    Abstract: In one example, a method includes method comprising: receiving audio data generated by a microphone of a current computing device; identifying, based on the audio data, one or more computing devices that each emitted a respective audio signal in response to speech reception being activated at the current computing device; and selecting either the current computing device or a particular computing device from the identified one or more computing devices to satisfy a spoken utterance determined based on the audio data.
    Type: Grant
    Filed: July 1, 2022
    Date of Patent: February 20, 2024
    Assignee: GOOGLE LLC
    Inventor: Jian Wei Leong
  • Patent number: 11893998
    Abstract: Implementations set forth herein relate to management of casting requests and user inputs at a rechargeable device, which provides access to an automated assistant and is capable of rendering data that is cast from a separate device. Casting requests can be handled by the rechargeable device despite a device SoC of the rechargeable device operating in a sleep mode. Furthermore, spoken utterances provided by a user for invoking the automated assistant can also be adaptively managed by the rechargeable device in order mitigate idle power consumption by the device SoC. Such spoken utterances can be initially processed by a digital signal processor (DSP), and, based on one or more features (e.g., voice characteristic, conformity to a particular invocation phrase, etc.) of the spoken utterance, the device SoC can be initialized for an amount of time that is selected based on the features of the spoken utterance.
    Type: Grant
    Filed: September 13, 2021
    Date of Patent: February 6, 2024
    Assignee: GOOGLE LLC
    Inventors: Andrei Pascovici, Victor Lin, Jianghai Zhu, Paul Gyugyi, Shlomi Regev
  • Patent number: 11893979
    Abstract: Systems and methods for e-commerce systems using natural language understanding are described. A computing device is configured receive a user utterance and identify at least one semantic element within the user utterance. An intent associated with the at least one semantic element is identified and an intent flow associated with the identified intent is executed. The intent flow includes a set of tasks executed in a predetermined order. A system utterance is generated by instantiating a response template selected from a plurality of response templates associated with the executed intent.
    Type: Grant
    Filed: December 21, 2021
    Date of Patent: February 6, 2024
    Assignee: Walmart Apollo, LLC
    Inventors: Snehasish Mukherjee, Shankara Bhargava Subramanya
  • Patent number: 11893986
    Abstract: Some embodiments provide systems and methods that can alleviate the burdens on professional advisors of preparing snapshots and answering client questions, particularly questions that do not require professional judgment or strategy. Such embodiments can automate creation of snapshots of account data collected from multiple institutions, suggest questions that are relevant to a user's accounts, and provide programmatic responses to factual or definitional questions pertaining to those accounts.
    Type: Grant
    Filed: January 22, 2021
    Date of Patent: February 6, 2024
    Assignee: Eyeballs Financial, LLC
    Inventors: Mitchell Steven Morrison, Brian Francis Fitzgerald, Rebecca Ann Morrison
  • Patent number: 11886773
    Abstract: Systems and methods to support an application that is voice-controlled are disclosed. Exemplary implementations may establish a synchronous two-way communication session between a first and a second user; present application information to the first and second user; detect first voice-control information from the first user and second voice-control information from the second user; control operations of the application based on the detected voice-control information; generate and/or determine results of the application based on the detected voice-control information; present the results to the users, and/or other steps.
    Type: Grant
    Filed: March 28, 2023
    Date of Patent: January 30, 2024
    Assignee: Volley Inc.
    Inventors: Maxwell Child, James Wilsterman
  • Patent number: 11887614
    Abstract: Provided are a device and a method for transmitting and receiving voice data in a wireless communication system. A method for operating a transmission terminal for transmitting a voice signal comprises the steps of: generating sampling and bitrate request information including sampling information for determining a sampling rate of the voice signal and bitrate information for determining a bitrate of the voice signal, and transmitting the generated sampling and bitrate request information to a reception terminal; receiving, from the reception terminal, combined determination information obtained by at least one combination of the sampling rate determined on the basis of the sampling information and the bitrate determined on the basis of the bitrate information; and compressing the voice signal according to the received combined determination information, and transmitting the compressed voice signal to the reception terminal.
    Type: Grant
    Filed: June 4, 2021
    Date of Patent: January 30, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventor: Kyung-hun Jung
  • Patent number: 11874877
    Abstract: A method uses natural language for visual analysis of a dataset, and is performed at a computer having a display, processors, and memory. The computer displays a data visualization based on a dataset retrieved from a database and receives a first user input to specify a natural language command related to the displayed data visualization. Based on the displayed data visualization and the natural language command, the computer identifies an initial range of values for a first data field in the dataset and displays an editable user interface control showing the initial range of values. The computer receives a second user input, at the user interface control, specifying a modified range of values for the first data field and retrieves an updated dataset from the database using the modified range of values for the first data field. The computer then updates the displayed data visualization using the updated dataset.
    Type: Grant
    Filed: February 8, 2022
    Date of Patent: January 16, 2024
    Assignee: Tableau Software, Inc.
    Inventors: Vidya Raghavan Setlur, Sarah E. Battersby, Melanie Karla Tory, Richard Carl Gossweiler, III, Angel Xuan Chang, Isaac James Dykeman, MD Enamul Hoque Prince
  • Patent number: 11837234
    Abstract: The present disclosure provides a subtitle generation method and apparatus, a device, and a storage medium, and the method includes: in response to a subtitle generation triggering operation directed against at least one audio track in a target audio-video file, performing speech recognition on audio data on each audio track respectively to obtain text fragments corresponding to each audio track; and generating subtitles of the target audio-video file based on the text fragments corresponding to each audio track. Compared with a method of performing overall speech recognition on audio data on all audio tracks, in the present disclosure, independent speech recognition is performed on the audio data on each audio track, and thus, the influences of the audio tracks on each other are avoided, so that more accurate speech recognition results can be obtained, thereby improving the accuracy of subtitles generated based on the speech recognition results.
    Type: Grant
    Filed: December 22, 2022
    Date of Patent: December 5, 2023
    Assignee: BEIJING BYTEDANCE NETWORK TECHNOLOGY CO., LTD.
    Inventors: Yan Zeng, Weiyi Chang, Pingfei Fu, Qifan Zheng, Zhaoqin Lin