Patents Examined by Vu B. Hang
  • Patent number: 12223971
    Abstract: The purpose of the present invention is to reduce distortion a frequency band component encoded with a small number of bits in a time domain and improve quality. An audio decoding device (10) decodes an encoded audio signal and outputs the audio signal. A decoding unit (10a) decodes an encoded sequence containing an encoded audio signal and obtains a decoded signal. A selective temporal envelope shaping unit (10b) shapes a temporal envelope of a decoded signal in the frequency band on the basis of decoding related information concerning decoding of the encoded sequence.
    Type: Grant
    Filed: July 27, 2022
    Date of Patent: February 11, 2025
    Assignee: NTT DOCOMO, INC
    Inventors: Kei Kikuiri, Atsushi Yamaguchi
  • Patent number: 12223964
    Abstract: The technology described in this document can be embodied in a computer-implemented method that includes presenting a first user-interface that includes a user-selectable menu of multiple virtual personal assistant (VPA) service providers, and receiving a user-selection identifying a particular VPA service provider. The method also includes receiving a first signal representing input speech data, and processing the first signal to generate a first electronic file that includes at least a portion of the input speech data, the first electronic file being generated in accordance with a specification of the particular VPA service provider. The method further includes transmitting the first electronic file to one or more remote computing devices associated with the particular VPA service provider, receiving at least a second electronic file including a response to the input speech data, and causing an acoustic transducer to generate an acoustic output based on the second electronic file.
    Type: Grant
    Filed: January 31, 2022
    Date of Patent: February 11, 2025
    Assignee: Bose Corporation
    Inventors: Naganagouda B. Patil, Andre Todman, Bernice A. Cramer
  • Patent number: 12217171
    Abstract: Engagement signals may be generated and analyzed based on user interactions with documents, particularly in a collaboration environment. The user interactions may generate raw collaboration signals that may be received and processed into cleaned collaboration signals. For example, noise may be removed from the raw collaboration signals to generate the cleaned collaboration signals. The cleaned collaboration signals may be grouped into engagement signals, where each engagement signal represents an individual event or engagement event of the user with the document. The grouping may be based on boundary signals, time frames, and/or any other reasonable limiting element. Each of the engagement signals may be classified into one of several engagement types based on the cleaned collaboration signals in the engagement signal. The engagement signals may then be analyzed to make determinations, recommendations, or the like regarding one or more users of the document, the document content, or the like.
    Type: Grant
    Filed: April 30, 2021
    Date of Patent: February 4, 2025
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Theo Lorrain-Hale, William D. Tierney, Feng Liu, Douglas Lane Milvaney, Manon Knoertzer
  • Patent number: 12211504
    Abstract: Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for displaying augmented reality content. The program and method provide for receiving, by a device, speech input to select augmented reality content for display; determining at least one keyword included in the speech input; identifying, from plural augmented reality content items, an augmented reality content item corresponding to the at least one keyword; and displaying the augmented reality content item with an image captured by a camera of the device.
    Type: Grant
    Filed: October 17, 2023
    Date of Patent: January 28, 2025
    Assignee: Snap Inc.
    Inventors: Joseph Timothy Fortier, Celia Nicole Mourkogiannis, Evan Spiegel, Kaveh Anvaripour
  • Patent number: 12211496
    Abstract: A processor-implemented utterance time estimation method includes: determining a plurality of attention weight matrices using an attention-based sequence-to-sequence model; selecting an attention weight matrix from the plurality of attention weight matrices; and estimating an utterance time corresponding to an output sequence based on the selected attention weight matrix.
    Type: Grant
    Filed: October 7, 2020
    Date of Patent: January 28, 2025
    Assignee: Samsung Electronics Co., Ltd.
    Inventor: Min-Joong Lee
  • Patent number: 12190875
    Abstract: Systems and methods for preemptive wakeword detection are disclosed. For example, a first part of a wakeword is detected from audio data representing a user utterance. When this occurs, on-device speech processing is initiated prior to when the entire wakeword is detected. When the entire wakeword is detected, results from the on-device speech processing and/or the audio data is sent to a speech processing system to determine a responsive action to be performed by the device. When the entire wakeword is not detected, on-device processing is canceled and the device refrains from sending the audio data to the speech processing system.
    Type: Grant
    Filed: September 30, 2021
    Date of Patent: January 7, 2025
    Assignee: Amazon Technologies, Inc.
    Inventors: Eli Joshua Fidler, Aaron Challenner, Zoe Adams, Sree Hari Krishnan Parthasarathi, Gengshen Fu
  • Patent number: 12190882
    Abstract: The present disclosure relates to a method and an apparatus for generating general voice commands, and the method includes: obtaining View tree content of a display interface of an application; traversing information nodes in the View tree content, and configuring different voice commands for different information nodes based on attributes of the information nodes; and aggregating all voice commands in the display interface, and mixing and filtering the commands to form a final voice command set.
    Type: Grant
    Filed: November 30, 2021
    Date of Patent: January 7, 2025
    Assignee: HANGZHOU LINGBAN TECHNOLOGY CO. LTD.
    Inventor: Weiming Liu
  • Patent number: 12190874
    Abstract: A voice control system for ophthalmologic laser treatment systems sets parameters for delivering laser energy based on voice commands and prevents potentially harmful parameters due to operator mistakes and misunderstood voice commands by providing incremental parameter adjustment and restricting the amount by which the parameters can be adjusted for each executed voice command. Valid voice commands include indications of which parameter to set, a value for the parameter, and whether to increase or decrease the value of the parameter. In one example, parameter values can only be increased or decreased by a certain percentage with respect to the current value. In another example, the parameters are adjusted by selecting the next highest or lowest value with respect to the current parameter value from a predetermined sequence of possible values for particular parameters. Voice control functionality can also be deactivated under certain conditions such as when it is determined that a parameter was not set.
    Type: Grant
    Filed: August 5, 2021
    Date of Patent: January 7, 2025
    Assignee: NORLASE APS
    Inventors: Greg Fava, Peter Skovgaard
  • Patent number: 12184937
    Abstract: Apparatuses and methods related to a voice recognition system, a voice recognition server and a control method of a display apparatus, are provided. More particularly, apparatuses and methods relate to a voice recognition system which performs a voice recognition function by using at least one of a current usage status with respect to the display apparatus and a function that is currently performed by the display apparatus. A voice recognition system includes: a voice receiver which receives a voice command; and a controller which determines at least one from among a current usage status with respect to a display apparatus and a function currently performed by the display apparatus, determines an operation corresponding to the received voice command by using at least one from among the determined current usage status and the function currently performed by the display apparatus, and performs the determined operation.
    Type: Grant
    Filed: June 9, 2022
    Date of Patent: December 31, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Jeong-Min Sim, Do-Wan Kim
  • Patent number: 12183362
    Abstract: A speech recognition method. The method includes: performing speech activity detection on speech data to obtain multiple speech segments; determining, for each of the speech segments, a number of speakers involved in the each of the speech segments; for each of at least one of the speech segments with the determined number greater than 1: performing speech separation on the each of at least one of the speech segments to obtain multiple audio segments; performing speech recognition on each of the audio segments to obtain respective first speech recognition results for the audio segments; performing feature extraction on each of the audio segments to obtain respective voiceprint feature vectors; and performing clustering on the audio segments with respect to the speakers to obtain a clustering result; and obtaining a second speech recognition result for the speech data based on the clustering result and the respective first speech recognition results.
    Type: Grant
    Filed: April 11, 2024
    Date of Patent: December 31, 2024
    Assignee: MASHANG CONSUMER FINANCE CO., LTD.
    Inventors: Qinglin Meng, Bin Yang, Ning Jiang, Haiying Wu, Quan Lu, Min Liu
  • Patent number: 12183339
    Abstract: The present disclosure relates to detecting the use of fake voice command to activate microphones of smart devices. In one embodiment, sound characteristics associated with an audio signal from a microphone of smart device may be compared with other microphones of the smart device in order to detect fake voice commands. In another embodiment, sound characteristics associated with the audio signal from the microphone may be compared with a threshold range of stored sound characteristics in order to detect fake voice commands. In some embodiments, a controller may triangulate a position associated with a source of a sound in order to detect a fake voice command. In a further embodiment, a controller may verify that a user or associated electronic device are near a smart device to authorize a voice command.
    Type: Grant
    Filed: May 31, 2022
    Date of Patent: December 31, 2024
    Assignee: United Services Automobile Association (USAA)
    Inventors: Carlos Jp Chavez, Sacha Melquiades De'Angeli, Oscar Guerra, David M. Jones, Jr., Gregory Brian Meyer, Christopher Russell, Arthur Quentin Smith
  • Patent number: 12175964
    Abstract: A computer-implemented method is provided. The method including receiving speech audio of dictation associated with a user ID, deriving acoustic features from the speech audio, storing the derived acoustic features in a user profile associated with the user ID, receiving a request for acoustic features through an application programming interface (API), the request including the user ID, and sending the derived acoustic features through the API.
    Type: Grant
    Filed: May 19, 2021
    Date of Patent: December 24, 2024
    Inventors: Kiran Garaga Lokeswarappa, Joel Gedalius, Bernard Mont-Reynaud, Jun Huang
  • Patent number: 12170085
    Abstract: Systems and methods are presented herein for increasing user engagement with an interface by suggesting commands or queries for the user. A plurality of content items available for consumption are identified and metadata for each of the plurality of content items is retrieved. One or more candidate voice commands are generated based on a plurality of voice command templates based on a target verb and a subset of the metadata corresponding to the plurality of the content items available for consumption. A recall score is generated for each candidate voice command based at least in part on a detection of phonetic features that match between clauses of each candidate voice command. At least the candidate voice command with the highest recall score is selected and output using a suggestion system.
    Type: Grant
    Filed: July 21, 2021
    Date of Patent: December 17, 2024
    Assignee: Adeia Guides Inc.
    Inventors: Ankur Anil Aher, Jeffry Copps Robert Jose
  • Patent number: 12154566
    Abstract: A method, computer system, and computer readable medium are provided for activating speech recognition based on keyword spotting (KWS). Waveform data corresponding to one or more speakers is received. One or more direction features are extracted from the received waveform data. One or more keywords are determined from the received waveform data based on the one or more extracted features. Speech recognition is activated based on detecting the determined keyword.
    Type: Grant
    Filed: June 3, 2022
    Date of Patent: November 26, 2024
    Assignee: TENCENT AMERICA LLC
    Inventors: Meng Yu, Dong Yu
  • Patent number: 12154546
    Abstract: A method and system for acoustic model conditioning on non-phoneme information features for optimized automatic speech recognition is provided. The method includes using an encoder model to encode sound embedding from a known key phrase of speech and conditioning an acoustic model with the sound embedding to optimize its performance in inferring the probabilities of phonemes in the speech. The sound embedding can comprise non-phoneme information related to the key phrase and the following utterance. Further, the encoder model and the acoustic model can be neural networks that are jointly trained with audio data.
    Type: Grant
    Filed: July 6, 2023
    Date of Patent: November 26, 2024
    Assignee: SoundHound AI IP, LLC.
    Inventors: Zizu Gowayyed, Keyvan Mohajer
  • Patent number: 12142371
    Abstract: In some aspects, a multi-turn conversational system includes: an artificial intelligence to provide a conversation interface configured to execute multiple turns of human-like conversation with a user and a control logic, in communication with the conversation interface, and configured to generate one or more control signals based on evaluating multiple turns of upstream human-like conversation between the conversation interface and the user. The control signals contribute in part to construction of multiple turns of downstream human-like conversation between the conversation interface and the user.
    Type: Grant
    Filed: February 29, 2024
    Date of Patent: November 12, 2024
    Inventors: Munjal Shah, Vishal Parikh, Meenesh Bhimani, Subhabrata Mukherjee, Alex Miller, Saad Godil, Debajyoti Datta, Paul Gamble, Rae Lasko
  • Patent number: 12135942
    Abstract: The present disclosure describes a conversation facilitation system for facilitating conversation-based social interactions to improve senior health, one or more operations and functions being efficiently achieved via this system comprise: receiving a dialog act of a conversation, applying natural language understanding (NLU) processing on the dialog act, computing a conversation metric, and generating a result of the conversation to conclude the conversation based on the conversation metric.
    Type: Grant
    Filed: October 23, 2023
    Date of Patent: November 5, 2024
    Assignee: CLEARCARE, INC.
    Inventors: Geoffrey Nudd, David Cristman, John Taylor, Jonathan J. Hull
  • Patent number: 12135736
    Abstract: Questions play a central role in assessment of a candidate's expertise during an interview or examination. However, generating such questions from input text documents manually needs specialized expertise and experience. Further, techniques that are available for automated question generation require input sentence as well as an answer phrase in that sentence to generate question. This in-turn requires large training datasets consisting tuples of input sentence answer-phrase and the corresponding question. Additionally, training datasets are available are for general purpose text, but not for technical text. Present application provides systems and methods for generating technical questions from technical documents. The system extracts meta information and linguistic information of text data present in technical documents. The system then identifies relationships that exist in provided text data. The system further creates one or more graphs based on the identified relationships.
    Type: Grant
    Filed: August 26, 2022
    Date of Patent: November 5, 2024
    Assignee: TATA CONSULTANCY SERVICES LIMITED
    Inventors: Sangameshwar Suryakant Patil, Samiran Pal, Avinash Kumar Singh, Soham Datta, Girish Keshav Palshikar, Indrajit Bhattacharya, Harsimran Bedi, Yash Agrawal, Vasudeva Varma Kalidindi
  • Patent number: 12125498
    Abstract: According to various embodiments, an electronic device may include: a microphone; an audio connector; a wireless communication circuit; a processor operatively connected to the microphone, the audio connector, and the wireless communication circuit; and a memory operatively connected to the processor, wherein the memory may store instructions that, when executed, cause the processor to: receive a first audio signal through the microphone, the audio connector, or the wireless communication circuit, extract audio feature information from the first audio signal, and recognize a speech section in a second audio signal, received after the first audio signal through the microphone, the audio connector, or the wireless communication circuit, using the audio feature information.
    Type: Grant
    Filed: January 7, 2022
    Date of Patent: October 22, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Seungbeom Ryu, Sungjae Park, Hyuk Oh, Myeungyong Choi, Junkwon Choi
  • Patent number: 12125496
    Abstract: The disclosed technology relates to methods, voice enhancement systems, and non-transitory computer readable media for real-time voice enhancement. In some examples, input audio data including foreground speech content, non-content elements, and speech characteristics is fragmented into input speech frames. The input speech frames are converted to low-dimensional representations of the input speech frames. One or more of the fragmentation or the conversion is based on an application of a first trained neural network to the input audio data. The low-dimensional representations of the input speech frames omit one or more of the non-content elements. A second trained neural network is applied to the low-dimensional representations of the input speech frames to generate target speech frames. The target speech frames are combined to generate output audio data. The output audio data further includes one or more portions of the foreground speech content and one or more of the speech characteristics.
    Type: Grant
    Filed: April 24, 2024
    Date of Patent: October 22, 2024
    Assignee: SANAS.AI INC.
    Inventors: Shawn Zhang, Lukas Pfeifenberger, Jason Wu, Piotr Dura, David Braude, Bajibabu Bollepalli, Alvaro Escudero, Gokce Keskin, Ankita Jha, Maxim Serebryakov