Patents Examined by Leonard Saint-Cyr
  • Patent number: 12217764
    Abstract: An encoding method includes dividing a to-be-encoded time-domain signal into a low band signal and a high band signal, performing encoding on the low band signal to obtain a low frequency encoding parameter, performing encoding on the high band signal to obtain a high frequency encoding parameter, obtaining a synthesized high band signal, performing short-time post-filtering processing on the synthesized high band signal to obtain a short-time filtering signal, and calculating a high frequency gain based on the high band signal and the short-time filtering signal.
    Type: Grant
    Filed: November 30, 2023
    Date of Patent: February 4, 2025
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Bin Wang, Zexin Liu, Lei Miao
  • Patent number: 12216995
    Abstract: A device and method for providing recommended words for a character input by a user are provided. The method by which the device provides recommended words includes: receiving an input for inputting a character in a character input window; recommending at least one pseudo-morpheme including the input character by analyzing the input character; recommending at least one extended word including a selected pseudo-morpheme in response to receiving an input for selecting one of the at least one pseudo-morpheme; and displaying a selected extended word in response to receiving an input for selecting one of the at least one extended word.
    Type: Grant
    Filed: October 18, 2022
    Date of Patent: February 4, 2025
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Hee-jun Song, Jung-wook Kim
  • Patent number: 12211483
    Abstract: The present application provides a robot that a user feels more strongly to be a living being. A robot includes a speech generating unit that generates speech and a speech output unit that outputs the generated speech. The robot outputs speech generated by the robot rather than outputting speech prepared in advance. Because of this, speech that is in accordance with sensor information can be generated and output, or speech unique to the robot can be generated and output.
    Type: Grant
    Filed: February 26, 2021
    Date of Patent: January 28, 2025
    Assignee: GROOVE X, INC.
    Inventors: Kaname Hayashi, John Belmonte, Atsuya Kose, Masaya Matsuura, Hideya Minamiji, Taiji Fukaya, Takuma Miyamoto
  • Patent number: 12205599
    Abstract: Provided is an encoding apparatus for integrally encoding and decoding a speech signal and a audio signal, and may include: an input signal analyzer to analyze a characteristic of an input signal; a stereo encoder to down mix the input signal to a mono signal when the input signal is a stereo signal, and to extract stereo sound image information; a frequency band expander to expand a frequency band of the input signal; a sampling rate converter to convert a sampling rate; a speech signal encoder to encode the input signal using a speech encoding module when the input signal is a speech characteristics signal; a audio signal encoder to encode the input signal using a audio encoding module when the input signal is a audio characteristic signal; and a bitstream generator to generate a bitstream.
    Type: Grant
    Filed: June 21, 2023
    Date of Patent: January 21, 2025
    Assignees: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE, KWANGWOON UNIVERSITY INDUSTRY-ACADEMIC COLLABORATION FOUNDATION
    Inventors: Tae Jin Lee, Seung-Kwon Baek, Min Je Kim, Dae Young Jang, Jeongil Seo, Kyeongok Kang, Jin-Woo Hong, Hochong Park, Young-Cheol Park
  • Patent number: 12205587
    Abstract: An embodiment of a summarization application divides collected conversation data into media and text components. The application implements respective machine learning mechanisms to enhance modeling operations of the text and media components to identify key elements from the conversation. The application generates a headline banner from a group of key elements based on an analysis involving first predetermined criteria. The application also combines additional key elements to the group of key elements to form a second group of key elements. The application generates a summary from the second group of key elements based on a second analysis involving second predetermined criteria. The application presents, via a display, the headline banner according to a first output of the first key element analysis and the summary according to a second output of the second key element analysis.
    Type: Grant
    Filed: April 6, 2022
    Date of Patent: January 21, 2025
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Trudy L. Hewitt, Liam S. Harpur, Jonathan D. Dunne, Kelley Anders
  • Patent number: 12204532
    Abstract: A device configured to perform data analytics comprising a memory and a processor may be configured to perform the techniques described in this disclosure. The memory may store multi-dimensional data. The processor may receive a sequence of inputs defining a recipe for analyzing the multi-dimensional data according to a language sub-surface specifying a natural language containment hierarchy defining a grammar for a natural language as a hierarchical arrangement of a plurality of language sub-surfaces. The processor may also receive data indicative of a summarized narration of the recipe and parameterize a field of the summarized narration to insert a user adjustable parameter that enables manipulation of the underlying recipe and obtain a parameterized summary that includes the user adjustable parameter. The processor may next present, via a first user interface, the parameterized summary.
    Type: Grant
    Filed: December 9, 2021
    Date of Patent: January 21, 2025
    Assignee: DataChat.ai
    Inventors: Jignesh Patel, Robert Konrad Claus, Amos Kendall, Rogers Jeffrey Leo John, Ushmal Ramesh, Jiatong Li
  • Patent number: 12198699
    Abstract: Methods and systems are provided for displaying transcriptions of radio communication transcription for an aircraft. The method comprises capturing audio signals of radio communication traffic to and from the aircraft. The captured audio signals are preprocessed to divide the signals into independent spoken utterances. Each spoken utterance is transcribed using a speech recognition decoder that utilizes an air traffic control (ATC) speech recognition model and classification data is extracted from the transcription of each spoken utterance. The transcription of each spoken utterance is logged with reference to the classification data and a textual display of the transcription is provided to a crew member of the aircraft.
    Type: Grant
    Filed: June 6, 2022
    Date of Patent: January 14, 2025
    Assignee: HONEYWELL INTERNATIONAL INC.
    Inventors: Chaya Garg, Vasantha Paulraj, Robert De Mers, Roger Burgin, Jitender Kumar Agarwal, Mahesh Kumar Sampath, Mohan M. Thippeswamy, Naveen Venkatesh Prasad Nama, Rahul Pradhan, Nitish Sharma
  • Patent number: 12190061
    Abstract: Systems and methods for topic modeling are described. The systems and methods include encoding words of a document using an embedding matrix to obtain word embeddings for the document. The words of the document comprise a subset of words in a vocabulary, and the embedding matrix is trained as part of a topic attention network based on a plurality of topics. The systems and methods further include encoding a topic-word distribution matrix using the embedding matrix to obtain a topic embedding matrix. The topic-word distribution matrix represents relationships between the plurality of topics and the words of the vocabulary. The systems and methods further include computing a topic context matrix based on the topic embedding matrix and the word embeddings and identifying a topic for the document based on the topic context matrix.
    Type: Grant
    Filed: December 17, 2021
    Date of Patent: January 7, 2025
    Assignee: ADOBE INC.
    Inventors: Shashank Shailabh, Madhur Panwar, Milan Aggarwal, Pinkesh Badjatiya, Simra Shahid, Nikaash Puri, S Sejal Naidu, Sharat Chandra Racha, Balaji Krishnamurthy, Ganesh Karbhari Palwe
  • Patent number: 12190063
    Abstract: A machine learning model analyzes text describing an item to determine portions of the text that correspond to multiple characteristics of the item. A first set of embeddings that represent the text describing the item is determined. A second set of embeddings that represent text indicating the characteristics is determined. The second set of embeddings includes a token for each characteristic that is used to indicate tokens that are associated with a particular characteristic. The first set of embeddings and portions of the second set of embeddings for a particular characteristic are used to determine a set of interaction embeddings for each characteristic by determining an element-wise product. These interaction embeddings are analyzed to determine label predictions indicating text that is associated with each characteristic. Text for multiple characteristics may therefore be identified using a single pass rather than multiple passes.
    Type: Grant
    Filed: June 1, 2022
    Date of Patent: January 7, 2025
    Assignee: AMAZON TECHNOLOGIES, INC.
    Inventors: Anubhav Shrimal, Avi Jain, Kartik Mehta, Promod Yenigalla
  • Patent number: 12190866
    Abstract: Methods and systems for manual-based image review can involve associating an image with a group of keyword utterances, the image displayable in a display screen of a computing device, the group of keyword utterances including different keyword utterances. A prompt for the user to utter a keyword utterance can be displayed in a first area of the image in the display screen and another prompt for the user to utter another keyword utterance can be displayed in another area of the image in the display screen. Audio of the keyword utterances displayed in the display screen can be captured and processed by natural language processing (NLP), when uttered by the user. The utterances can be displayed respectively as text in the first area and the other area of the image in response to processing by NLP of the audio. Thus, instead of users typing in the results and changing their focus between screen and keyboard, for example, the user can speak to the results, which increases the throughput of the results.
    Type: Grant
    Filed: March 29, 2022
    Date of Patent: January 7, 2025
    Assignee: Conduent Business Services, LLC
    Inventor: Lee D Roche
  • Patent number: 12183335
    Abstract: Provided is an information processing system including: a voice information acquisition unit that acquires voice information including an utterance made by a person; a status acquisition unit that acquires status information related to status of the person; and a support information generation unit that generates support information used for supporting operation of the person based on the voice information and the status information.
    Type: Grant
    Filed: July 5, 2019
    Date of Patent: December 31, 2024
    Assignee: NEC CORPORATION
    Inventor: Masamichi Tanabe
  • Patent number: 12183343
    Abstract: Embodiments include methods, devices, systems, and non-transitory process-readable storage media for voice-activated message filtering rule generation. Some embodiments may include receiving a spoken command from a communication device, parsing the spoken command to identify an element of the spoken command, generating a message rule based on the identified element of the spoken command, determining whether the generated message rule has been met, and sending a message to the communication device in response to determining that the message rule has been met.
    Type: Grant
    Filed: June 23, 2023
    Date of Patent: December 31, 2024
    Assignee: Stryker Corporation
    Inventors: Sridhar Acharya, Arun Mirchandani
  • Patent number: 12183355
    Abstract: Many portable playback devices cannot decode and playback encoded audio content having wide bandwidth and wide dynamic range with consistent loudness and intelligibility unless the encoded audio content has been prepared specially for these devices. This problem can be overcome by including with the encoded content some metadata that specifies a suitable dynamic range compression profile by either absolute values or differential values relative to another known compression profile. A playback device may also adaptively apply gain and limiting to the playback audio. Implementations in encoders, in transcoders and in decoders are disclosed.
    Type: Grant
    Filed: March 28, 2024
    Date of Patent: December 31, 2024
    Assignees: Dolby Laboratories Licensing Corporation, DOLBY INTERNATIONAL AB
    Inventors: Jeffrey Riedmiller, Harald Mundt, Michael Schug, Martin Wolters
  • Patent number: 12182524
    Abstract: Systems and methods for natural language processing are described. One or more aspects of a method, apparatus, and non-transitory computer readable medium include receiving a text phrase; encoding the text phrase using an encoder to obtain a hidden representation of the text phrase, wherein the encoder is trained during a first training phrase using self-supervised learning based on a first contrastive loss and during a second training phrase using supervised learning based on a second contrastive learning loss; identifying an intent of the text phrase from a predetermined set of intent labels using a classification network, wherein the classification network is jointly trained with the encoder in the second training phase; and generating a response to the text phrase based on the intent.
    Type: Grant
    Filed: November 4, 2021
    Date of Patent: December 31, 2024
    Assignee: ADOBE INC.
    Inventors: Jianguo Zhang, Trung Huu Bui, Seunghyun Yoon, Xiang Chen, Quan Hung Tran, Walter W. Chang
  • Patent number: 12175982
    Abstract: Systems and methods are provided in which a speaker profile-data inquiry is transmitted to a mobile device associated with a first speaker. In response to the speaker-profile-data inquiry, speaker profile data associated with the first speaker is received. Audio data representing a voice input is received. The first speaker is identified as providing the voice input, the identification being based on a comparison of characteristics of the received audio data with the speaker profile data of a plurality of speakers for whom speaker profile data is stored. An instruction, which includes a speaker-relative signifier, is determined from the received audio data, and determining the instruction includes determining a referent of the speaker-relative signifier based on the first speaker profile data. An action indicated by the instruction is performed.
    Type: Grant
    Filed: February 7, 2022
    Date of Patent: December 24, 2024
    Assignee: DRNC HOLDINGS, INC.
    Inventor: Keith Edwards
  • Patent number: 12175997
    Abstract: Provided is a first acoustic information acquisition unit configured to acquire a first acoustic information obtained by receiving a sound wave emitted from a first sound source by a wearable device worn by a user, a second acoustic information acquisition unit configured to acquire a second acoustic information obtained by receiving a sound wave emitted from a second sound source that is different from the first sound source by the wearable device, and a third acoustic information acquisition unit configured to acquire a third acoustic information used for biometric matching of the user based on the first acoustic information and the second acoustic information.
    Type: Grant
    Filed: January 7, 2020
    Date of Patent: December 24, 2024
    Assignee: NEC CORPORATION
    Inventors: Koji Okabe, Takayuki Arakawa, Takafumi Koshinaka
  • Patent number: 12175995
    Abstract: There is provided servers and methods of generating a waveform based on a spectrogram and a noise input. The method includes acquiring a trained flow-based vocoder including invertible blocks, and an untrained feed-forward vocoder including non-invertible blocks, which form a student-teacher network. The method includes executing a training process in the student-teacher network during which the server generates (i) a teacher waveform by the trained flow-based vocoder using a first spectrogram and a first noise input, (ii) a student waveform by the untrained feed-forward vocoder using the first spectrogram and the first noise input, and (iii) a loss value for the given training iteration using the teacher waveform and the student waveform. The server then trains the untrained feed-forward vocoder to generate the waveform. The trained feed-forward vocoder in then used lieu of the trained flow-based vocoder for generating waveforms based on spectrograms and noise inputs.
    Type: Grant
    Filed: May 31, 2022
    Date of Patent: December 24, 2024
    Assignee: Y.E. Hub Armenia LLC
    Inventors: Vladimir Vladimirovich Kirichenko, Aleksandr Aleksandrovich Molchanov, Dmitry Mikhailovich Chernenkov, Artem Valerevich Babenko, Vladimir Andreevich Aliev, Dmitry Aleksandrovich Baranchuk
  • Patent number: 12159115
    Abstract: Examples described herein generate training data for machine learning (ML) for natural language (NL) processing (such as semantic parsing for translating NL). A formula tree is generated based on sampling both a formula grammar and NL templates. Using the formula tree, an ML training data instance pair is generated comprising a formula example and an NL example. A context example may also be used during instantiation of the formula tree. An ML model is trained with training data including the ML training data instance pair, and ML output is generated from NL input. The ML output includes, for example, a machine-interpretable formula, a database querying language command, or a general programming language instruction. Some examples support context-free grammar, probabilistic context-free grammar, and/or non-context-free production rules.
    Type: Grant
    Filed: October 19, 2021
    Date of Patent: December 3, 2024
    Assignee: Microsoft Technology Licensing, LLC.
    Inventors: Zeqi Lin, Yu Hu, Haiyuan Cao, Yi Liu, Jian-Guang Lou, Kuralmani Elango, PalaniRaj Kaliyaperumal, Weizhu Chen, Kunal Mukerjee
  • Patent number: 12154570
    Abstract: Various embodiments contemplate systems and methods for performing automatic speech recognition (ASR) and natural language understanding (NLU) that enable high accuracy recognition and understanding of freely spoken utterances which may contain proper names and similar entities. The proper name entities may contain or be comprised wholly of words that are not present in the vocabularies of these systems as normally constituted. Recognition of the other words in the utterances in question, e.g. words that are not part of the proper name entities, may occur at regular, high recognition accuracy. Various embodiments provide as output not only accurately transcribed running text of the complete utterance, but also a symbolic representation of the meaning of the input, including appropriate symbolic representations of proper name entities, adequate to allow a computer system to respond appropriately to the spoken request without further analysis of the user's input.
    Type: Grant
    Filed: June 6, 2023
    Date of Patent: November 26, 2024
    Assignee: PROMPTU SYSTEMS CORPORATION
    Inventor: Harry William Printz
  • Patent number: 12153894
    Abstract: A generation apparatus 100 includes: an argumentative scheme adding unit 10 which adds an argumentative scheme with respect to pair data constituted by an input utterance and a counter utterance 121 that voices a negative opinion with respect to the input utterance and which generates argumentative scheme-added pair data 122; a generation model learning unit 20 which learns a generation model for generating a counter utterance from an input utterance in consideration of the argumentative scheme by using the argumentative scheme-added pair data 122 as learning data and which generates a learned counter utterance generation model 123; and a counter utterance generating unit 30 which acquires an input utterance of a user and a designated argumentative scheme and which outputs a counter utterance using the counter utterance generation model 123.
    Type: Grant
    Filed: December 11, 2019
    Date of Patent: November 26, 2024
    Assignee: NIPPON TELEGRAPH AND TELEPHONE CORPORATION
    Inventors: Ko Mitsuda, Ryuichiro Higashinaka, Yushi Aono