Patents Examined by Vu B. Hang
  • Patent number: 11379509
    Abstract: A system is configured to: (a) ingest a set of codes from a code reference, the code reference including words and non-words; (b) associate the set of codes to a set of representations, forming descriptive links between the set of codes and the set of representations; (c) ingest an unstructured document having a general outline, the unstructured document including (i) at least one of the words in the code reference, (ii) at least one code in the set of codes, or (iii) both (i) and (ii); (d) determine an ontology from content of the unstructured document with an unsupervised learning algorithm; and (e) store the determined ontology in a database, the ontology having an acyclic directed graph structure.
    Type: Grant
    Filed: October 29, 2019
    Date of Patent: July 5, 2022
    Assignee: UST Global (Singapore) Pte. Ltd.
    Inventor: Adnan Masood
  • Patent number: 11380326
    Abstract: A speech recognition method includes receiving a first multi-channel audio signal; obtaining at least one of a speech signal characteristic or a noise signal characteristic for at least one frequency band of frequency bands corresponding to channel audio signals included in the first multi-channel audio signal; generating a signal with an enhanced speech component by performing beamforming on the first multi-channel audio signal based on the speech signal characteristic, a speech signal characteristic obtained for a previous frame that was obtained before a certain time that the first multi-channel audio signal was obtained, and the noise signal characteristic; determining whether the enhanced speech component includes a wake word; and based on determining that the enhanced speech component includes the wake word: activating a speech recognition operation based on the signal with the enhanced speech component.
    Type: Grant
    Filed: May 15, 2020
    Date of Patent: July 5, 2022
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Changwoo Han, Minkyu Shin, Jonguk Yoo, Dokyun Lee, Kangseok Choi, Jaewon Lee, Hyeontaek Lim
  • Patent number: 11373661
    Abstract: The present document relates to a method of layered encoding of a frame of a compressed higher-order Ambisonics, HOA, representation of a sound or sound field. The compressed HOA representation comprises a plurality of transport signals. The method comprises assigning the plurality of transport signals to a plurality of hierarchical layers, the plurality of layers including a base layer and one or more hierarchical enhancement layers, generating, for each layer, a respective HOA extension payload including side information for parametrically enhancing a reconstructed HOA representation obtainable from the transport signals assigned to the respective layer and any layers lower than the respective layer, assigning the generated HOA extension payloads to their respective layers, and signaling the generated HOA extension payloads in an output bitstream.
    Type: Grant
    Filed: July 10, 2020
    Date of Patent: June 28, 2022
    Assignee: Dolby International AB
    Inventors: Sven Kordon, Alexander Krueger
  • Patent number: 11373660
    Abstract: The present document relates to a method of layered encoding of a compressed sound representation of a sound or sound field. The compressed sound representation comprises a basic compressed sound representation comprising a plurality of components, basic side information for decoding the basic compressed sound representation to a basic reconstructed sound representation of the sound or sound field, and enhancement side information including parameters for improving the basic reconstructed sound representation.
    Type: Grant
    Filed: July 1, 2020
    Date of Patent: June 28, 2022
    Assignee: Dolby International AB
    Inventors: Sven Kordon, Alexander Krueger
  • Patent number: 11372926
    Abstract: A remote personal assistant server/service is configured with a plurality of different state machines for supporting asynchronous communications with the front-end interfaces of personal assistants and for supporting the interfacing with a plurality of proprietary back-end intelligence bots, which are independently controlled and separate from the remote personal assistant service/server.
    Type: Grant
    Filed: July 22, 2019
    Date of Patent: June 28, 2022
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Gerardo Andres Lecaros Easton, Minko Stefanov Minkov, Matthew Vernon Johnson
  • Patent number: 11361761
    Abstract: A system, method, and computer program product for determining statement attributions. The system includes at least one processing component, at least one memory component, a feature extractor, a model generator, a model database, and an attribution selector. The method includes receiving a statement, generating at least one pattern that defines a grammatical feature of the statement, and generating a statement model from the at least one pattern. The method also includes determining a similarity value for the statement model and at least one reference model.
    Type: Grant
    Filed: October 16, 2019
    Date of Patent: June 14, 2022
    Assignee: International Business Machines Corporation
    Inventors: Stephen Arthur Boxwell, Keith Gregory Frost, Kyle Matthew Brake, Stanley John Vernier
  • Patent number: 11348583
    Abstract: The present disclosure discloses a data processing method and apparatus for an intelligent device, and a storage medium, which relates to a field of artificial intelligence technologies. The method includes: extracting key voice information from collected user voice information; in a non-wireless fidelity (WiFi) network environment, transmitting the key voice information to a mobile terminal, so that the mobile terminal transmits the key voice information to a server, and receives a processing result fed back by the server after the server processes the key voice information; and obtaining the processing result from the mobile terminal to display the processing result.
    Type: Grant
    Filed: June 21, 2020
    Date of Patent: May 31, 2022
    Assignees: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD., SHANGHAI XIAODU TECHNOLOGY CO. LTD.
    Inventors: Yang Liu, Xi Xi, Long Quan
  • Patent number: 11335333
    Abstract: A method includes obtaining audio data for a long-form utterance and segmenting the audio data for the long-form utterance into a plurality of overlapping segments. The method also includes, for each overlapping segment of the plurality of overlapping segments: providing features indicative of acoustic characteristics of the long-form utterance represented by the corresponding overlapping segment as input to an encoder neural network; processing an output of the encoder neural network using an attender neural network to generate a context vector; and generating word elements using the context vector and a decoder neural network. The method also includes generating a transcription for the long-form utterance by merging the word elements from the plurality of overlapping segments and providing the transcription as an output of the automated speech recognition system.
    Type: Grant
    Filed: December 17, 2019
    Date of Patent: May 17, 2022
    Assignee: Google LLC
    Inventors: Wei Han, Chung-Cheng Chiu, Yu Zhang, Yonghui Wu, Patrick Nguyen, Sergey Kishchenko
  • Patent number: 11328707
    Abstract: Methods and apparatuses are described for building a knowledge base taxonomy from structured or unstructured computer text for use in automated user interactions. A server computing device receives one or more of structured text or unstructured text corresponding to historical user interaction data from a database. The server computing device extracts one or more terms from the received text that are most relevant to a subject matter domain. The server computing device organizes the extracted one or more terms into a taxonomy data structure.
    Type: Grant
    Filed: May 11, 2020
    Date of Patent: May 10, 2022
    Assignee: FMR LLC
    Inventors: Pranab Mohanty, Bianca De Oliveira Pereira, Cecile Robin, Tobias Daudert, John McCrae, Paul Buitelaar
  • Patent number: 11310374
    Abstract: This system includes: an apparatus; a first server; and a second server connectable to the apparatus and the first server. The first server includes a voice receiver that receives a voice from a user, a command transmitter that transmits a command based on the voice to the second server, and an output part that, responsive to a response signal received from the second server, provides an output based on the response signal. The second server includes a command processor that transmits a response signal to the first server responsive to a first command received from the first server, and that acquires a status of the apparatus, and a status transmitter that transmits a response signal based on the acquired status of the apparatus to the first server responsive to a second command received from the first server.
    Type: Grant
    Filed: September 27, 2019
    Date of Patent: April 19, 2022
    Assignee: SHARP KABUSHIKI KAISHA
    Inventor: Katsuyoshi Kondoh
  • Patent number: 11295753
    Abstract: When the noise in an audio signal made up of both speech and noise is suppressed, the quality of the speech in the audio signal is usually degraded. The speech obtained from a noise-suppressed audio signal is improved by determining linear predictive coding (LPC) characteristics of the audio signal without or prior to noise suppression and by determining the LPC characteristics of the noise-suppressed audio. The convolution of those differing characteristics provides an improved-quality speech signal, with the original noise level reduced or suppressed.
    Type: Grant
    Filed: March 3, 2015
    Date of Patent: April 5, 2022
    Assignee: Continental Automotive Systems, Inc.
    Inventor: Bijal Joshi
  • Patent number: 11295743
    Abstract: This disclosure proposes systems and methods enabling on-device/hybrid processing of speech requests using a hub device. The hub device is capable of receiving audio data from surrounding devices and performing speech processing on the audio data to improve latency and/or provide functionality to other devices within a private network. The hub device may receive multiple requests corresponding to different utterances. If the hub device receives a second utterance while processing a first utterance, the hub device may send an error notification, process the first utterance and the second utterance sequentially, suspend processing of the first utterance to process the second utterance first, send the second utterance to another hub device or remote system, or suspend processing of the first utterance and send the first utterance to the remote system in order to process the second utterance.
    Type: Grant
    Filed: May 26, 2020
    Date of Patent: April 5, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Fabian Andreas Bumberger, Sabria Farheen, Maciej Makowski, Eli Joshua Fidler, Sasitheran Shanmugarajah
  • Patent number: 11257493
    Abstract: Systems and methods for processing speech are described. In certain examples, image data is used to generate visual feature tensors and audio data is used to generate audio feature tensors. The visual feature tensors and the audio feature tensors are used by a linguistic model to determine linguistic features that are usable to parse an utterance of a user. The generation of the feature tensors may be jointly configured with the linguistic model. Systems may be provided in a client-server architecture.
    Type: Grant
    Filed: July 11, 2019
    Date of Patent: February 22, 2022
    Assignee: SoundHound, Inc.
    Inventors: Cristina Vasconcelos, Zili Li
  • Patent number: 11250871
    Abstract: In an acoustic signal separation device (1), a determination unit (6) determines whether or not components from a plurality of sound sources are mixed in each of acoustic signals of respective components regenerated by a signal regeneration unit (5), and when it is determined that a plurality of components is mixed, a series of processes by a feature value extraction unit (2), a data estimation unit (3), a data classification unit (4), and a signal regeneration unit (5) is repeatedly executed until acoustic signals of the components of the respective sound sources are regenerated.
    Type: Grant
    Filed: January 15, 2018
    Date of Patent: February 15, 2022
    Assignee: MITSUBISHI ELECTRIC CORPORATION
    Inventors: Tatsuhiko Saito, Keigo Kawashima, Jun Ishii, Yohei Okato
  • Patent number: 11250859
    Abstract: The technology described in this document can be embodied in a computer-implemented method that includes presenting a first user-interface that includes a user-selectable menu of multiple virtual personal assistant (VPA) service providers, and receiving a user-selection identifying a particular VPA service provider. The method also includes receiving a first signal representing input speech data, and processing the first signal to generate a first electronic file that includes at least a portion of the input speech data, the first electronic file being generated in accordance with a specification of the particular VPA service provider. The method further includes transmitting the first electronic file to one or more remote computing devices associated with the particular VPA service provider, receiving at least a second electronic file including a response to the input speech data, and causing an acoustic transducer to generate an acoustic output based on the second electronic file.
    Type: Grant
    Filed: June 15, 2020
    Date of Patent: February 15, 2022
    Assignee: Bose Corporation
    Inventors: Naganagouda B. Patil, Andre Todman, Bernice A. Cramer
  • Patent number: 11238242
    Abstract: Some implementations are directed to translating chatspeak to a normalized form, where the chatspeak is included in natural language input formulated by a user via a user interface input device of a computing device—such as input provided by the user to an automated assistant. The normalized form of the chatspeak may be utilized by the automated assistant in determining reply content that is responsive to the natural language input, and that reply content may be presented to the user via one or more user interface output devices of the computing device of the user. Some implementations are additionally and/or alternatively directed to providing, for presentation to a user, natural language output that includes chatspeak in lieu of a normalized form of the chatspeak, based at least in part on a “chatspeak measure” that is determined based on past usage of chatspeak by the user and/or by additional users.
    Type: Grant
    Filed: March 21, 2019
    Date of Patent: February 1, 2022
    Assignee: Google LLC
    Inventors: Wan Fen Nicole Quah, Bryan Horling, Maryam Garrett, Brian Roark, Richard Sproat
  • Patent number: 11232799
    Abstract: Techniques for speech recognition are described. For example, a speech recognition service is to receive a request to perform speech recognition on speech data from a chatbot using a particular speech recognition service; determine a group of hosts to route the speech data to, the group of hosts to host a plurality of speech recognition services including the particular speech recognition service; determine a path to the determined group of hosts using a set of one or more rules; determine a particular host of the group of hosts to perform speech recognition on the speech data, the particular host having the speech recognition service in memory to process the request and being preferred for performing the speech recognition on the speech data; route the speech data to the particular host; perform speech recognition on the speech data using the particular host; and provide a text result of the speech recognition.
    Type: Grant
    Filed: October 31, 2018
    Date of Patent: January 25, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Apoorv Birthare, John Baker, Kranthi Kumar Boyapati, Krishna Chaitanya Gourishetti, Enrico Sartorello
  • Patent number: 11227589
    Abstract: Systems and processes for operating an intelligent automated assistant to perform intelligent list reading are provided. In one example process, a spoken user request associated with a plurality of data items is received. The process determines whether a degree of specificity of the spoken user request is less than a threshold level. In response to determining that a degree of specificity of the spoken user request is less than a threshold level, one or more attributes related to the spoken user request are determined. The one or more attributes are not defined in the spoken user request. Additionally, a list of data items based on the spoken user request and the one or more attributes is obtained. A spoken response comprising a subset of the list of data items is generated and the spoken response is provided.
    Type: Grant
    Filed: January 16, 2019
    Date of Patent: January 18, 2022
    Assignee: Apple Inc.
    Inventors: Susan L. Booker, Murali Krishnan, Aimee Piercy, Garrett L. Weinberg
  • Patent number: 11227596
    Abstract: A laundry scheduling device according to an embodiment of the present invention includes an input interface that receives speech including a plurality of words respectively representing a plurality of laundry items from a use, and a processor that acquires one or more features represented by each of the plurality of words by inputting speech data corresponding to the received speech to a learning model and performing word embedding, and generates a laundry schedule of the plurality of laundry items based on the one or more features represented by each of the plurality of words.
    Type: Grant
    Filed: October 11, 2019
    Date of Patent: January 18, 2022
    Assignee: LG ELECTRONICS INC.
    Inventors: Taeho Lee, Jaehong Kim, Hyoeun Kim, Hangil Jeong, Jongwoo Han, Doyoung Lee, Hyejeong Jeon, Dami Choe
  • Patent number: 11227624
    Abstract: This disclosure relates generally to speech signal processing, and more particularly to a method and system for processing speech signal for emotion identification. The system processes a speech signal collected as input, during which a plurality of differential features corresponding to a plurality of frames of the speech signal are extracted. Further, the differential features are compared with an emotion recognition model to identify at least one emotion matching the speech signal, and then the at least one emotion is associated with the speech signal.
    Type: Grant
    Filed: March 9, 2020
    Date of Patent: January 18, 2022
    Assignee: Tata Consultancy Services Limited
    Inventors: Gauri Ashutosh Deshpande, Sachin Patel, Mayuri Duggirala, Venkata Subramanian Viraraghavan