Patents Examined by Edgar X Guerra-Erazo
  • Patent number: 10297256
    Abstract: Disclosed herein are example techniques for voice detection by multiple NMDs. An example implementation may involve one or more servers receiving, via a network interface, data representing multiple audio recordings of a voice input spoken by a given user, each audio recording recorded by a respective NMD of the multiple NMDs, wherein the voice input comprises a detected wake-word. Based on respective sound pressure levels of the multiple audio recordings of the voice input, the servers (i) select a particular NMD of the multiple NMDs and (ii) forego selection of other NMDs of the multiple NMDs. The servers send, via the network interface to the particular NMD, data representing a playback command that corresponds to a voice command in the voice input represented in the multiple audio recordings, wherein the data representing the playback command causes the particular NMD to play back audio content according to the playback command.
    Type: Grant
    Filed: December 10, 2018
    Date of Patent: May 21, 2019
    Assignee: Sonos, Inc.
    Inventors: Jonathon Reilly, Gregory Burlingame, Christopher Butts, Romi Kadri, Jonathan P. Lang
  • Patent number: 10291758
    Abstract: The present disclosure provides an intelligent voice reminder system, a server and a method thereof. The intelligent voice reminder system includes a terminal device, a player and a server. The server is connected with the player and the terminal device. The server includes a storage, a processor, a transceiver and a communication bus. The storage is configured to store data from the terminal device. The transceiver is configured to receive a voice input and output the voice input to the processor. The processor is configured to obtain a calendar reminder event or a phonebook from the storage based on the received voice input. The transceiver outputs the calendar reminder event or a phonebook to the player for playing.
    Type: Grant
    Filed: August 5, 2016
    Date of Patent: May 14, 2019
    Assignee: HUIZHOU TCL MOBILE COMMUNICATION CO., LTD
    Inventor: Yuxin Xu
  • Patent number: 10275444
    Abstract: Aspects of the subject disclosure may include, for example, a computer that performs a statistical natural language processing analysis on a plurality of text documents to determine a plurality of topics, creates a proper subset of topics from the plurality of topics, based on user input, maps one or more topics in the proper subset of topics to each document in the plurality of text documents, thereby creating a plurality of topic-document pairs, identifies n-dimensions of bias for each topic-document pair from the text, creates clusters of topics from the proper subset of topics, and generates presentable content depicting each cluster of the clusters of topics according to a corresponding image configuration. The topics and n-dimensions of bias data can be further analyzed with co-collected structured data for statistical relationships.
    Type: Grant
    Filed: July 15, 2016
    Date of Patent: April 30, 2019
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Pamela Bogdan, Gary Gressel, Gary Reser, Alex Rubarkh, Kenneth Shirley
  • Patent number: 10275462
    Abstract: A request to translate an input collection of collocated separate strings from a source language to a target language is received. The input collection is associated with a combination of meanings for the collocated separate strings within a context of the input collection. A search for the collocated separate strings is performed in a database storing a number of collection units of separate strings in the source language and a correspondingly translated number of collection units of translated separate strings in the target language. An output collection unit of separate strings in the target language is determined as a translated collection of separate strings of the input collection based on evaluating the input collection as a unit and evaluating the number of collection units of separate strings in the source language and the correspondingly translated number of collection units of translated separate strings in the target language at the database.
    Type: Grant
    Filed: September 18, 2017
    Date of Patent: April 30, 2019
    Assignee: SAP SE
    Inventor: Jens Scharnbacher
  • Patent number: 10269348
    Abstract: A vehicle based system and method for receiving voice inputs and determining whether to perform a voice recognition analysis using in-vehicle resources or resources external to the vehicle.
    Type: Grant
    Filed: February 28, 2017
    Date of Patent: April 23, 2019
    Assignee: Honda Motor Co., Ltd.
    Inventors: Ritchie Huang, Pedram Vaghefinazari, Stuart Yamamoto
  • Patent number: 10268671
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for generating parse trees for input text segments. One of the methods includes obtaining an input text segment comprising a plurality of inputs arranged according to an input order; processing the inputs in the input text segment using an encoder long short term memory (LSTM) neural network to generate a respective encoder hidden state for each input in the input text segment; and processing the respective encoder hidden states for the inputs in the input text segment using an attention-based decoder LSTM neural network to generate a linearized representation of a parse tree for the input text segment.
    Type: Grant
    Filed: December 30, 2016
    Date of Patent: April 23, 2019
    Assignee: Google LLC
    Inventors: Lukasz Mieczyslaw Kaiser, Oriol Vinyals
  • Patent number: 10262659
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for receiving audio data corresponding to an utterance, determining that the audio data corresponds to a hotword, generating a hotword audio fingerprint of the audio data that is determined to correspond to the hotword, comparing the hotword audio fingerprint to one or more stored audio fingerprints of audio data that was previously determined to correspond to the hotword, detecting whether the hotword audio fingerprint matches a stored audio fingerprint of audio data that was previously determined to correspond to the hotword based on whether the comparison indicates a similarity between the hotword audio fingerprint and one of the one or more stored audio fingerprints that satisfies a predetermined threshold, and in response to detecting that the hotword audio fingerprint matches a stored audio fingerprint, disabling access to a computing device into which the utterance was spoken.
    Type: Grant
    Filed: March 1, 2018
    Date of Patent: April 16, 2019
    Assignee: Google LLC
    Inventors: Matthew Sharifi, Jakob Nicolaus Foerster
  • Patent number: 10255269
    Abstract: Long short term memory units that accept a non-predefined number of inputs are used to provide natural language relation extraction over a user-specified range on content. Content written for human consumption is parsed with distant supervision in segments (e.g., sentences, paragraphs, chapters) to determine relationships between various words within and between those segments.
    Type: Grant
    Filed: December 30, 2016
    Date of Patent: April 9, 2019
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Christopher Brian Quirk, Kristina Nikolova Toutanova, Wen-tau Yih, Hoifung Poon, Nanyun Peng
  • Patent number: 10249293
    Abstract: A device may receive audio data based on a capturing of sounds associated with a structure. The device may obtain a model associated with the structure. The model may have been trained to receive the audio data as input, determine a score that identifies a likelihood that a sound is present in the audio data, and identify the sound based on the score. The device may determine at least one parameter associated with the sound. The device may generate a metric based on the at least one parameter associated with the sound, and perform an action based on the metric.
    Type: Grant
    Filed: June 11, 2018
    Date of Patent: April 2, 2019
    Assignee: Capital One Services, LLC
    Inventors: Michael Mossoba, Joshua Edwards, Abdelkadar M'Hamed Benkreira, Austen Novis, Sophie Bermudez
  • Patent number: 10235999
    Abstract: Among other things, requests are received from voice assistant devices expressed in accordance with different corresponding protocols of one or more voice assistant frameworks. Each of the requests represents a voiced input by a user to the corresponding voice assistant device. The received requests are re-expressed in accordance with a common request protocol. Based on the received requests, responses to the requests are expressed in accordance with a common response protocol. Each of the responses is re-expressed according to a protocol of the framework with respect to which the corresponding request was expressed. The responses are sent to the voice assistant devices for presentation to the users.
    Type: Grant
    Filed: June 5, 2018
    Date of Patent: March 19, 2019
    Assignee: Voicify, LLC
    Inventors: Robert T. Naughton, Nicholas G. Laidlaw, Alexander M. Dunn, Jeffrey K. McMahon
  • Patent number: 10229675
    Abstract: This document generally describes systems and methods for dynamically adapting speech recognition for individual voice queries of a user using class-based language models. The method may include receiving a voice query from a user that includes audio data corresponding to an utterance of the user, and context data associated with the user. One or more class models are then generated that collectively identify a first set of terms determined based on the context data, and a respective class to which the respective term is assigned for each respective term in the first set of terms. A language model that includes a residual unigram may then be accessed and processed for each respective class to insert a respective class symbol at each instance of the residual unigram that occurs within the language model. A transcription of the utterance of the user is then generated using the modified language model.
    Type: Grant
    Filed: December 30, 2016
    Date of Patent: March 12, 2019
    Assignee: Google LLC
    Inventors: Justin Max Scheiner, Petar Aleksic
  • Patent number: 10224058
    Abstract: This specification describes computer-implemented methods and systems. One method includes receiving, by a neural network of a speech recognition system, first data representing a first raw audio signal and second data representing a second raw audio signal. The first raw audio signal and the second raw audio signal describe audio occurring at a same period of time. The method further includes generating, by a spatial filtering layer of the neural network, a spatial filtered output using the first data and the second data, and generating, by a spectral filtering layer of the neural network, a spectral filtered output using the spatial filtered output. Generating the spectral filtered output comprises processing frequency-domain data representing the spatial filtered output. The method still further includes processing, by one or more additional layers of the neural network, the spectral filtered output to predict sub-word units encoded in both the first raw audio signal and the second raw audio signal.
    Type: Grant
    Filed: November 14, 2016
    Date of Patent: March 5, 2019
    Assignee: Google LLC
    Inventors: Ehsan Variani, Kevin William Wilson, Ron J. Weiss, Tara N. Sainath, Arun Narayanan
  • Patent number: 10223067
    Abstract: An environmental context of a user may be taken into account to enhance the user's communication throughput. An “environmental context” can include spatial surroundings of a user, device, and/or sensor of the device and more broadly to denote the context of the user in a multiplicity of environments such as, for example, the surroundings of a user, a digital environment such as the user or other individuals' interactions with or made near a device, etc. The techniques can include obtaining contextual data to provide context-predicted suggestions of words and/or phrases that a user can select to be output on the user's behalf. In some examples, the techniques can also use contextual data to weight, sort, rank, and/or filter word and/or phrase suggestions.
    Type: Grant
    Filed: July 15, 2016
    Date of Patent: March 5, 2019
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Meredith Morris, Shaun K. Kane
  • Patent number: 10216724
    Abstract: Performing semantic analysis on a user-generated text string includes training a neural network model with a plurality of known text strings to obtain a first distributed vector representation of the known text strings and a second distributed vector representation of a plurality of words in the known text strings, computing a relevance matrix of the first and second distributed representations based on a cosine distance between each of the plurality of words and the plurality of known text strings, and performing a latent dirichlet allocation (LDA) operation using the relevance matrix as an input to obtain a distribution of topics associated with the plurality of known text strings.
    Type: Grant
    Filed: April 7, 2017
    Date of Patent: February 26, 2019
    Assignee: Conduent Business Services, LLC
    Inventors: Manjira Sinha, Tridib Mukherjee, Preethy Varma, Satarupa Guha
  • Patent number: 10210867
    Abstract: Techniques are disclosed for adjusting user experience of a software application based on paralinguistic information. One embodiment presented herein includes a computer-implemented method for adjusting a user experience of a software application. The method comprises receiving, at a computing device, an audio stream comprising audio of a user. The method further comprises analyzing the audio stream for paralinguistic information to determine an attribute of the user. The method further comprises identifying content of the audio stream. The method further comprises determining one or more actions based on the content of the audio stream. The method further comprises selecting at least one of the one or more actions based on the attribute of the user.
    Type: Grant
    Filed: October 11, 2017
    Date of Patent: February 19, 2019
    Assignee: INTUIT INC.
    Inventors: Raymond Chan, Igor A. Podgorny, Benjamin Indyk
  • Patent number: 10210885
    Abstract: A speech-based system utilizes a speech interface device located in the home of a user. The system may interact with different users based on different user profiles. The system may include messaging services that generate and/or provide messages to the user through the speech interface device. The speech interface device may have indicators that are capable of being illuminated in different colors. To notify a user regarding the currently active user profile, each user profile is associated with a different color and the color of the active profile is displayed on the speech interface device when the user is interacting with the system. To notify the user regarding awaiting messages, different types of messages are associated with different colors and the colors of the message types of waiting messages are displayed on the speech interface whenever the user is not interacting with the system.
    Type: Grant
    Filed: May 20, 2014
    Date of Patent: February 19, 2019
    Assignee: Amazon Technologies, Inc.
    Inventor: Ty Loren Carlson
  • Patent number: 10192556
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media for learning pronunciations from acoustic sequences. One method includes receiving an acoustic sequence, the acoustic sequence representing an utterance, and the acoustic sequence comprising a sequence of multiple frames of acoustic data at each of a plurality of time steps; stacking one or more frames of acoustic data to generate a sequence of modified frames of acoustic data; processing the sequence of modified frames of acoustic data through an acoustic modeling neural network comprising one or more recurrent neural network (RNN) layers and a final CTC output layer to generate a neural network output, wherein processing the sequence of modified frames of acoustic data comprises: subsampling the modified frames of acoustic data; and processing each subsampled modified frame of acoustic data through the acoustic modeling neural network.
    Type: Grant
    Filed: November 13, 2017
    Date of Patent: January 29, 2019
    Assignee: Google LLC
    Inventors: Hasim Sak, Andrew W. Senior
  • Patent number: 10185711
    Abstract: The subject matter of this specification can be embodied in, among other things, a method that includes receiving two or more data sets each representing speech of a corresponding individual attending an internet-based social networking video conference session, decoding the received data sets to produce corresponding text for each individual attending the internet-based social networking video conference, and detecting characteristics of the session from a coalesced transcript produced from the decoded text of the attending individuals for providing context to the internet-based social networking video conference session.
    Type: Grant
    Filed: July 5, 2016
    Date of Patent: January 22, 2019
    Assignee: Google LLC
    Inventors: Glen Shires, Sterling Swigart, Jonathan Zolla, Jason J. Gauci
  • Patent number: 10185714
    Abstract: A terminology marker system integrates a terminology analytical component for quantifying the amount of linguistic noise found in the translation output as measured against a dictionary; further, correlating the noise measured on a continuous basis enables the analytical component to build terminology predictive models used in a feedback loop to upstream components of the supply chain to improve future translation of new content. The system also provides a smart terminology assessment component for assessing linguistic assets and improving the quality of those assets to assist in translation. The system also provides a smart terminology evaluation component that is able to analyze MT output to make smart decisions on reducing the amount of post editing corrections needed for delivering a persistent level of translation quality.
    Type: Grant
    Filed: July 6, 2017
    Date of Patent: January 22, 2019
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Christophe D. Chenon, Marc P. Drapeau, Francis X. Rojas
  • Patent number: 10170128
    Abstract: A method and an apparatus for processing a temporal envelope of an audio signal, and an encoder are disclosed. When multiple temporal envelopes are solved, continuity of signal energy can be well maintained, and in addition, complexity of calculating a temporal envelope is reduced. The method includes: obtaining a high-band signal of the current frame audio signal according to the received current frame audio signal; dividing the high-band signal of the current frame signal into M subframes according to a predetermined temporal envelope quantity M, where M is an integer that is greater than or equal to 2; calculating a temporal envelope of each of the subframes; performing windowing on the first subframe of the M subframes and the last subframe of the M subframes by using an asymmetric window function; and performing windowing on a subframe except the first subframe and the last subframe of the M subframes.
    Type: Grant
    Filed: September 19, 2017
    Date of Patent: January 1, 2019
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Zexin Liu, Lei Miao