Patents Examined by Marcus T. Riley
  • Patent number: 11977848
    Abstract: Computer-based natural language understanding of input and output for a computer interlocutor is improved using a method of classifying conversation segments from transcribed conversations. The improvement includes one or more methods of splitting transcribed conversations into groups related to a conversation ontology using metadata; identifying dominant paths of conversational behavior by counting the frequency of occurrences of the behavior for a given path; creating a conversation model comprising conversation behaviors, metadata, and dominant paths; and using the conversation model to assign a probability score for a matched input to the computer interlocutor or a generated output from the computer interlocutor.
    Type: Grant
    Filed: April 14, 2023
    Date of Patent: May 7, 2024
    Assignee: Discourse.AI, Inc.
    Inventor: Jonathan E. Eisenzopf
  • Patent number: 11978454
    Abstract: A system for performing automated speech recognition (ASR) on audio data includes a queue manager to receive a request to perform ASR on audio data, add the request to a queue of incoming requests, and determine a queue depth representing a number of requests in the queue at a given time. The system also includes a load supervisor to receive the request and the queue depth from the queue manager and assign a service level for the request based on the queue depth. In addition, the system includes a speech-to-text converter to receive the assigned service level for the request from the load supervisor, select an ASR model for the request based on the received service level, receive the audio data associated with the request, and perform ASR on the audio data using the selected ASR model.
    Type: Grant
    Filed: September 16, 2021
    Date of Patent: May 7, 2024
    Assignee: SOUNDHOUND AI IP, LLC
    Inventors: Timothy P. Stonehocker, Zizu Gowayyed, Matthias Eichstaedt, Seyed Majid Emami, Evelyn Jiang, Ryan Berryhill, Mathieu Ramona, Neil Veira
  • Patent number: 11977839
    Abstract: A rule integration device includes determination means for converting, to natural language sentences, each of multiple management rules used by a management device for managing a management target, and determining whether or not the multiple management rules are combinable based on grammar relating to the converted natural language sentences; and combination means for generating a post-combination rule by combining the multiple management rules that have been determined to be combinable by the determination means.
    Type: Grant
    Filed: March 11, 2020
    Date of Patent: May 7, 2024
    Assignee: NEC CORPORATION
    Inventor: Toshimune Ebata
  • Patent number: 11972751
    Abstract: Disclosed are a method and an apparatus for detecting a voice end point by using acoustic and language modeling information to accomplish strong voice recognition. A voice end point detection method according to an embodiment may comprise the steps of: inputting an acoustic feature vector sequence extracted from a microphone input signal into an acoustic embedding extraction unit, a phonemic embedding extraction unit, and a decoder embedding extraction unit, which are based on a recurrent neural network (RNN); combining acoustic embedding, phonemic embedding, and decoder embedding to configure a feature vector by the acoustic embedding extraction unit, the phonemic embedding extraction unit, and the decoder embedding extraction unit; and inputting the combined feature vector into a deep neural network (DNN)-based classifier to detect a voice end point.
    Type: Grant
    Filed: June 29, 2020
    Date of Patent: April 30, 2024
    Assignee: IUCF-HYU (INDUSTRY-UNIVERSITY COOPERATION FOUNDATION HANYANG UNIVERSITY)
    Inventors: Joon-Hyuk Chang, Inyoung Hwang
  • Patent number: 11966713
    Abstract: In various embodiments, a computer-implemented language identification and communication system is provided. The system includes an application engine configured for processing data associated with multiple access devices of a population of users who are users seeking human language translation services and users providing human language translation services. A geolocation module is provided for locating a position of various users, such as the different locations of human language translators. The application engine is further programmed for receiving translator selections from user access devices and establishing communication connections between or among different user access devices.
    Type: Grant
    Filed: September 16, 2021
    Date of Patent: April 23, 2024
    Assignee: Zoose Language Technologies LLC
    Inventors: Patrick S. Allocco, Shalini Kadavill
  • Patent number: 11961520
    Abstract: Disclosed are a voice recognition system and a display device using the same. The disclosed voice recognition system includes a plate structure, a vibration sensor, and a voice recognition device. The plate structure vibrates based on propagation of a voice wave generated from a user, and the vibration sensor is provided in contact with the plate structure to detect the vibration of the plate structure. The voice recognition device recognizes voice of the user by receiving a signal output from the vibration sensor.
    Type: Grant
    Filed: December 16, 2022
    Date of Patent: April 16, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Cheheung Kim, Jaehyung Jang, Hyeokki Hong
  • Patent number: 11948573
    Abstract: A method for automatically rectifying in real-time anomalies in natural language processing systems. The method can include determining an output corresponding to a user request from a user device for a user based on a new request template or machine learning. The method further can include retrieving one or more entity rules corresponding to entity data of the user request. The method also can include overwriting entity information of the entity data corresponding to the one or more entity rules. Additionally, the method can include outputting the output. Furthermore, the method can include transmitting, to the user device, a response to the user. Other embodiments are disclosed.
    Type: Grant
    Filed: October 31, 2022
    Date of Patent: April 2, 2024
    Assignee: WALMART APOLLO, LLC
    Inventors: Snehasish Mukherjee, Haoxuan Chen, Phani Ram Sayapaneni, Shankara Bhargava Subramanya
  • Patent number: 11942087
    Abstract: A device performs a method for using image data to aid voice recognition. The method includes the device capturing image data of a vicinity of the device and adjusting, based on the image data, a set of parameters for voice recognition performed by the device. The set of parameters for the device performing voice recognition include, but are not limited to: a trigger threshold of a trigger for voice recognition; a set of beamforming parameters; a database for voice recognition; and/or an algorithm for voice recognition. The algorithm may include using noise suppression or using acoustic beamforming.
    Type: Grant
    Filed: January 13, 2021
    Date of Patent: March 26, 2024
    Assignee: Google Technology Holdings LLC
    Inventors: Robert A. Zurek, Adrian M. Schuster, Fu-Lin Shau, Jincheng Wu
  • Patent number: 11922939
    Abstract: A system and method are disclosed for ignoring a wakeword received at a speech-enabled listening device when it is determined the wakeword is reproduced audio from an audio-playing device. Determination can be by detecting audio distortions, by an ignore flag sent locally between an audio-playing device and speech-enabled device, by and ignore flag sent from a server, by comparison of received audio played audio to a wakeword within an audio-playing device or a speech-enabled device, and other means.
    Type: Grant
    Filed: May 4, 2022
    Date of Patent: March 5, 2024
    Assignee: SoundHound AI IP, LLC
    Inventors: Hsuan Yang, Qindí Zhãng, Warren S. Heit
  • Patent number: 11914658
    Abstract: The invention relates to a multi-node word segmentation system and method for keyword search. A word distribution proxy server in that system is communication connection with each word segmentation node server respectively, receive a large number of words, and distributes each word to a corresponding word segmentation node server according to a preset word segmentation rule; the client proxy server is communication connection with each word segmentation node server respectively, receives terms to be searched uploaded by a user terminal, and distributes the terms to be searched to the corresponding word segmentation node servers according to a preset word segmentation rule; the segmentation node server inquires the words matched with the terms to be searched, and returns the words obtained by inquiry to the client proxy server; and that client proxy server sends the query words to a corresponding user terminal.
    Type: Grant
    Filed: May 15, 2020
    Date of Patent: February 27, 2024
    Assignee: SHENZHEN SEKORM COMPONENT NETWORK CO., LTD
    Inventor: Lianglei Shang
  • Patent number: 11914923
    Abstract: Techniques for computer system-based conversations are described. In an example, a system receives, from a first device, first data corresponding to a first interaction in a conversation that requests a function. The system causes the first device to output a first response to the first interaction. Prior to an execution of the function, the system determines that the conversation is to be paused and causes the first device to output a first indication that the conversation is paused. Upon determining that the conversation is to be resumed, the system causes a second device to output a second indication that the conversation is resumed. The second device can be the same or different from the first device. The system receives, from the second device, second data corresponding to a second interaction in the conversation and causes the execution of the function based at least in part on the second data.
    Type: Grant
    Filed: December 10, 2021
    Date of Patent: February 27, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Shiveesh Fotedar, Saurabh Rathi, Steven Bishop
  • Patent number: 11906320
    Abstract: The present disclosure provides a method and an apparatus for managing navigation broadcast, and a device, related to an Intelligent transportation technology field. A specific implementation solution includes: obtaining a geographical identifier of a user; obtaining a statement-conversion template set corresponding to the geographical identifier based on the geographical identifier of the user; converting a standard navigation broadcast statement based on the statement-conversion template set to generate a geographical navigation broadcast statement; and performing navigation broadcast based on the geographical navigation broadcast statement. Thereby, the navigation broadcasts are matched with respective regions, users in different regions are provided with diversified and personalized navigation broadcasts.
    Type: Grant
    Filed: April 7, 2020
    Date of Patent: February 20, 2024
    Assignee: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD.
    Inventors: Ran Ji, Jizhou Huang, Ying Li, Yongzhi Ji, Lei Jia
  • Patent number: 11907667
    Abstract: A system for assisting sharing of information includes circuitry to: input a plurality of sentences each representing a statement made by one of a plurality of users, the sentence being generated by speaking or writing during a meeting or by extracting from at least one of meeting data, email data, electronic file data, and chat data at any time; determine a statement type of the statement represented by each one of the plurality of sentences, the statement type being one of a plurality of statement types previously determined; select, from among the plurality of sentences being input, one or more sentences each representing a statement of a specific statement type of the plurality of types; and output a list of the selected one or more sentences as key statements of the plurality of sentences.
    Type: Grant
    Filed: August 11, 2022
    Date of Patent: February 20, 2024
    Assignee: RICOH COMPANY, LTD.
    Inventor: Tomohiro Shima
  • Patent number: 11908481
    Abstract: Provided is a method for encoding live-streaming data, including: acquiring first state information associated with a current data frame; generating backup state information by backing up the first state information; generating a first encoded data frame by encoding the current data frame based on a first bit rate and the first state information; generating reset state information by resetting the updated first state information based on the backup state information; generating a second encoded data frame by encoding the current data frame based on a second bit rate and the reset state information; and generating a first target data frame corresponding to the current data frame based on the first encoded data frame and the second encoded data frame.
    Type: Grant
    Filed: January 24, 2022
    Date of Patent: February 20, 2024
    Assignee: BEIJING DAJIA INTERNET INFORMATION TECHNOLOGY CO., LTD.
    Inventors: Wenhao Xing, Chen Zhang
  • Patent number: 11908453
    Abstract: A method and a system for training a machine-learning algorithm (MLA) to determine a user class of a user of an electronic device are provided. The method comprises: receiving a training audio signal representative of a training user utterance; soliciting, by the processor, a plurality of assessor-generated labels for the training audio signal, the given one of the plurality of assessor-generated labels being indicative of whether the training user is perceived to be one of a first class and a second class; generating an amalgamated assessor-generated label for the training audio signal, the amalgamated assessor-generated label being indicative of a label distribution of the plurality of assessor-generated labels between the first class and the second class; generating a training set of data including the training audio signal and the amalgamated assessor-generated to train the MLA to determine the user class of the user producing an in-use user utterance.
    Type: Grant
    Filed: August 23, 2021
    Date of Patent: February 20, 2024
    Assignee: Direct Cursus Technology L.L.C
    Inventors: Vladimir Andreevich Aliev, Stepan Aleksandrovich Kargaltsev, Artem Valerevich Babenko
  • Patent number: 11900062
    Abstract: Described are methods and systems are for generating dynamic conversational queries. For example, as opposed to being a simply reactive system, the methods and systems herein provide means for actively determining a user's intent and generating a dynamic query based on the determined user intent. Moreover, these methods and systems generate these queries in a conversational environment.
    Type: Grant
    Filed: October 1, 2021
    Date of Patent: February 13, 2024
    Assignee: Capital One Services, LLC
    Inventors: Minh Le, Arturo Hernandez Zeledon, Md Arafat Hossain Khan
  • Patent number: 11900518
    Abstract: A method of producing an avatar video, the method comprising the steps of: providing a reference image of a person's face; providing a plurality of characteristic features representative of a facial model X0 of the person's face, the characteristic features defining a facial pose dependent on the person speaking; providing a target phrase to be rendered over a predetermined time period during the avatar video and providing a plurality of time intervals t within the predetermined time period; generating, for each of said times intervals t, speech features from the target phrase, to provide a sequence of speech features; and generating, using the plurality of characteristic features and sequence of speech features, a sequence of facial models Xt for each of said time intervals t.
    Type: Grant
    Filed: September 17, 2019
    Date of Patent: February 13, 2024
    Assignee: VirtTari Limited
    Inventors: Peter Alistair Brady, Hayden Allen-Vercoe, Sathish Sankarpandi, Ethan Dickson
  • Patent number: 11893996
    Abstract: Techniques for generating a personalization identifier that is usable by a skill to customize output of supplemental content to a user, without the skill being able to determine an identity of the user based on the personalization identifier, are described. A personalization identifier may be generated to be specific to a skill, such that different skills receive different personalization identifiers with respect to the same user. The personalization identifier may be generated by performing a one-way hash of a skill identifier, and a user profile identifier and/or a device identifier. User-perceived latency may be reduced by generating the personalization identifier at least partially in parallel to performing ASR processing and/or NLU processing.
    Type: Grant
    Filed: June 29, 2021
    Date of Patent: February 6, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Mark Conrad Kockerbeck, Song Chen, Aditi Srinivasan, Ryan Idrogo-Lam, Jilani Zeribi, John Botros
  • Patent number: 11893989
    Abstract: A system and method for controlling an electronic eyewear device using voice commands receives audio data from a microphone, processes the audio data to identify a wake word, and upon identification of a wake word, processes the audio data to identify at least one action keyword in the audio data. The audio data is provided to one of a plurality of controllers associated with different action keywords or sets of action keywords to implement an action. For example, the audio data may be provided to a settings controller to adjust settings of the electronic eyewear device when the action keyword is indicative of a request to adjust a setting of the electronic eyewear device or to a navigation controller to navigate to the system information of the electronic eyewear device when the action keyword is indicative of a request to navigate to system information of the electronic eyewear device.
    Type: Grant
    Filed: July 13, 2021
    Date of Patent: February 6, 2024
    Assignee: Snap Inc.
    Inventor: Piotr Gurgul
  • Patent number: 11887602
    Abstract: Techniques for performing audio-based device location determinations are described. A system may send, to a first device, a command to output audio requesting a location of the first device be determined. A second device may receive the audio and send, to the system, data representing the second device received the audio, where the received data includes spectral energy data representing a spectral energy of the audio as received by the second device. The system may, using the spectral energy data, determine attenuation data representing an attenuation experienced by the audio as it traveled from the first device to the second device. The system may generate, based on the attenuation data, spatial relationship data representing a spatial relationship between the first device and the second device, where the spatial relationship data is usable to determine a device for outputting a response to a subsequently received user input.
    Type: Grant
    Filed: December 10, 2021
    Date of Patent: January 30, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Brendon Jude Wilson, Henry Michael D Souza, Cindy Angie Hou, Christopher Evans, Sumit Garg, Ravina Chopra