Patents Examined by Shaun Roberts
  • Patent number: 10497368
    Abstract: Apparatuses, methods, systems, and program products are disclosed for transmitting audio to an identified recipient. A method includes detecting, by a processor, audio input at a first information handling device. The audio input is intended for a recipient. The method includes deriving an identity of the intended recipient of the audio input based on the audio input. The method includes transmitting the audio input to a second information handling device that is associated with the intended recipient.
    Type: Grant
    Filed: August 15, 2017
    Date of Patent: December 3, 2019
    Assignee: Lenovo (Singapore) PTE. LTD.
    Inventors: Amy Leigh Rose, John Scott Crowe, Gary David Cudak, Jennifer Lee-Baron, Nathan J. Peterson
  • Patent number: 10497371
    Abstract: A system, method and computer-readable storage devices are disclosed for multi-modal interactions with a system via a long-touch gesture on a touch-sensitive display. A system operating per this disclosure can receive a multi-modal input comprising speech and a touch on a display, wherein the speech comprises a pronoun. When the touch on the display has a duration longer than a threshold duration, the system can identify an object within a threshold distance of the touch, associate the object with the pronoun in the speech, to yield an association, and perform an action based on the speech and the association.
    Type: Grant
    Filed: April 29, 2019
    Date of Patent: December 3, 2019
    Inventors: Brant J. Vasilieff, Patrick Ehlen, Michael J. Johnston
  • Patent number: 10490198
    Abstract: A sensor device may include a computing device in communication with multiple microphones. A neural network executing on the computing device may receive audio signals from each microphone. One microphone signal may serve as a reference signal. The neural network may extract differences in signal characteristics of the other microphone signals as compared to the reference signal. The neural network may combine these signal differences into a lossy compressed signal. The sensor device may transmit the lossy compressed signal and the lossless reference signal to a remote neural network executing in a cloud computing environment for decompression and sound recognition analysis.
    Type: Grant
    Filed: December 18, 2017
    Date of Patent: November 26, 2019
    Assignee: GOOGLE LLC
    Inventors: Chanwoo Kim, Rajeev Conrad Nongpiur, Tara Sainath
  • Patent number: 10474756
    Abstract: Systems and methods for using autoencoders for training natural language classifiers. An example method comprises: producing, by a computer system, a plurality of feature vectors, wherein each feature vector represents a natural language text of a text corpus, wherein the text corpus comprises a first plurality of annotated natural language texts and a second plurality of un-annotated natural language texts; training, using the plurality of feature vectors, an autoencoder represented by an artificial neural network; producing, by the autoencoder, an output of the hidden layer, by processing a training data set comprising the first plurality of annotated natural language texts; and training, using the training data set, a text classifier that accepts an input vector comprising the output of the hidden layer and yields a degree of association, with a certain text category, of a natural language text utilized to produce the output of the hidden layer.
    Type: Grant
    Filed: December 22, 2017
    Date of Patent: November 12, 2019
    Assignee: ABBYY Production LLC
    Inventors: Konstantin Vladimirovich Anisimovich, Evgenii Mikhailovich Indenbom, Ivan Ivanovich Ivashnev
  • Patent number: 10468016
    Abstract: Disclosed herein is a system for compensating for dialects and accents comprising an automatic speech recognition system comprising an automatic speech recognition device that is operative to receive an utterance in an acoustic format from a user with a user interface; a speech to text conversion engine that is operative to receive the utterance from the automatic speech recognition device and to prepare a textual statement of the utterance; and a correction database that is operative to store textual statements of all utterances; where the correction database is operative to secure a corrected transcript of the textual statement of the utterance from the speech to text conversion engine and adds it to the corrections database if the corrected transcript of the textual statement of the utterance is not available.
    Type: Grant
    Filed: November 24, 2015
    Date of Patent: November 5, 2019
    Inventors: David Jaramillo, Neil Katz, Robert Smart, Viney A. Ugave
  • Patent number: 10468031
    Abstract: An approach is provided that receives an audio stream and utilizes a voice activation detection (VAD) process to create a digital audio stream of voices from at least two different speakers. An automatic speech recognition (ASR) process is applied to the digital stream with the ASR process resulting in the spoken words to which a speaker turn detection (STD) process is applied to identify a number of speaker segments with each speaker segment ending at a word boundary. The STD process analyzes a number of speaker segments using a language model that determines when speaker changes occur. A speaker clustering algorithm is then applied to the speaker segments to associate one of the speakers with each of the speaker segments.
    Type: Grant
    Filed: November 21, 2017
    Date of Patent: November 5, 2019
    Assignee: International Business Machines Corporation
    Inventors: Kenneth W. Church, Dimitrios B. Dimitriadis, Petr Fousek, Miroslav Novak, George A. Saon
  • Patent number: 10460721
    Abstract: A dialogue act estimation method, in a dialogue act estimation apparatus, includes acquiring first training data indicating, in a mutually associated manner, text data of a first sentence that can be a current uttered sentence, and text data of a second sentence that can be an uttered sentence immediately previous to the first sentence. The method also includes speaker change information indicating whether a speaker of the first sentence is the same as a speaker of the second sentence, and dialogue act information indicating a class of the first sentence. The method further includes learning an association between the current uttered sentence and the dialogue act information by applying the first training data to a model, and storing a result of the learning as learning result information in a memory.
    Type: Grant
    Filed: June 7, 2017
    Date of Patent: October 29, 2019
    Inventor: Takashi Ushio
  • Patent number: 10452674
    Abstract: A device receives, from a virtual assistant device, a first user input associated with a first account of a user, and causes a natural language processing analysis to be performed on the first user input to identify first information, the first account, and a first operation to be performed in association with first information in the first account. The device identifies a first data management platform, associated with the first account, that is configured to maintain the first information in a first data structure associated with the first data management platform, and determines that the first data management platform is a first type of data management platform based on the first data structure. The device causes the first operation to be performed using a RPA, that uses a user interface of the first data management platform, based on the first data management platform being the first type of data management platform.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: October 22, 2019
    Assignee: Accenture Global Solutions Limited
    Inventors: Gaurav Diwan, Tracy Ann Goguen
  • Patent number: 10446136
    Abstract: A system and method for accent invariant speech recognition comprising: maintaining a database scoring a set of language units in a given language, and for each of the language units, scoring audio samples of pronunciation variations of the language unit pronounced by a plurality of speakers; extracting and storing m the database a feature vector for locating each of the audio samples in a feature space; identifying pronunciation variation distances, which are distances between locations of audio samples of the same language unit in the feature space, and inter-unit distances, which are distances between locations of audio samples of different language units in the feature space; calculating a transformation applicable on the feature space to reduce the pronunciation variation distances relative to the inter-unit distances; and based on the calculated transformation, training a processor to classify as a same language unit pronunciation variations of the same language unit.
    Type: Grant
    Filed: May 11, 2017
    Date of Patent: October 15, 2019
    Inventors: Ron Fridental, Ilya Blayvas, Pavel Nosko
  • Patent number: 10430522
    Abstract: An adaptive localization system translates and displays translated content to a user, for example through a website or application using the adaptive localization system. A user can view, receive, or otherwise interact with the translated content, which can be differently translated based on desired language, geographic location, an intended user, or other relevant characteristics of the viewing user. The adaptive localization engine can translate the inherent meaning of content rather than, for example, creating an exact grammatical or “word-for-word” translation of individual words or phrases in the content. The adaptive localization engine displays alternate variations of the same translation of content to different users and based on user response to the alternate translations, determines the accuracy or correctness of a certain translations of content and modifies future translations accordingly.
    Type: Grant
    Filed: March 14, 2017
    Date of Patent: October 1, 2019
    Assignee: Qordoba, Inc.
    Inventors: Waseem Alshikh, May Habib
  • Patent number: 10430896
    Abstract: There is provided an information processing server including a specification unit configured to specify an individual registered in a predetermined database on the basis of identification information sent from a near-field communication device and collected voice information.
    Type: Grant
    Filed: July 28, 2017
    Date of Patent: October 1, 2019
    Inventor: Kazuyoshi Horie
  • Patent number: 10417335
    Abstract: Various aspects of the subject technology relate to systems, methods, and machine-readable media for automated quantitative assessment of text complexity. A system may include processing at least one body of text in a text-based query using a natural language processing engine. The processed text may include sub-blocks of text in a predetermined sequence size such as an n-gram. The system may compare reference bases to the processed text, where each reference base is associated with a different natural language. The system determines which of the reference bases has a highest number of matching words within the body of text, and thereby identifies the reference base as the source language of the supplied text. The system then determines an average complexity score for n-gram using a quantitative assessment engine. The system then applies a readability score to the body of text based on the average complexity scores of the n-grams.
    Type: Grant
    Filed: October 10, 2017
    Date of Patent: September 17, 2019
    Assignee: Colossio, Inc.
    Inventor: Joseph A. Jaroch
  • Patent number: 10418026
    Abstract: Systems and methods are described for processing and interpreting audible commands spoken in one or more languages. Speech recognition systems disclosed herein may be used as a stand-alone speech recognition system or comprise a portion of another content consumption system. A requesting user may provide audio input (e.g., command data) to the speech recognition system via a computing device to request an entertainment system to perform one or more operational commands. The speech recognition system may analyze the audio input across a variety of linguistic models, and may parse the audio input to identify a plurality of phrases and corresponding action classifiers. In some embodiments, the speech recognition system may utilize the action classifiers and other information to determine the one or more identified phrases that appropriately match the desired intent and operational command associated with the user's spoken command.
    Type: Grant
    Filed: July 15, 2016
    Date of Patent: September 17, 2019
    Assignee: Comcast Cable Communications, LLC
    Inventors: George Thomas Des Jardins, Vikrant Sagar
  • Patent number: 10403278
    Abstract: Systems and processes for operating an intelligent automated assistant to provide media items based on phonetic matching techniques are provided. An example method includes receiving a speech input from a user and determining whether the speech input includes a user request for a media item. The method further includes, in accordance with a determination that the speech input includes a user request for obtaining a media item, determining a candidate media item from a plurality of media items. The method further includes determining, based on a difference between a phonetic representation of the candidate media item and a phonetic representation of the speech input, whether the candidate media item is to be provided to the user. The method further includes, in accordance with a determination that the candidate media item is to be provided to the user, providing the candidate media item to the user.
    Type: Grant
    Filed: September 13, 2017
    Date of Patent: September 3, 2019
    Assignee: Apple Inc.
    Inventors: Adrian Skilling, Melvyn J. Hunt, Gunnar Evermann
  • Patent number: 10403291
    Abstract: Methods, systems, apparatus, including computer programs encoded on computer storage medium, to facilitate language independent-speaker verification. In one aspect, a method includes actions of receiving, by a user device, audio data representing an utterance of a user. Other actions may include providing, to a neural network stored on the user device, input data derived from the audio data and a language identifier. The neural network may be trained using speech data representing speech in different languages or dialects. The method may include additional actions of generating, based on output of the neural network, a speaker representation and determining, based on the speaker representation and a second representation, that the utterance is an utterance of the user. The method may provide the user with access to the user device based on determining that the utterance is an utterance of the user.
    Type: Grant
    Filed: June 1, 2018
    Date of Patent: September 3, 2019
    Assignee: Google LLC
    Inventors: Ignacio Lopez Moreno, Li Wan, Quan Wang
  • Patent number: 10390160
    Abstract: An apparatus and method for verifying voice messages generated by notification devices in an emergency address system includes one or more verification devices and a validation module, which can be part of a control panel or a connected services system. The verification devices can be mobile computing devices or permanently installed devices associated with each notification device in the emergency address system. The verification devices include microphones, network interfaces, and controllers executing speech to text conversion processes. During testing, the notification devices play voice messages and the verification devices detect the messages, convert the messages to text and send the text-converted messages to the validation module, which validates the text-converted messages against the intended messages for each notification device and confirms that the messages were played in the correct locations.
    Type: Grant
    Filed: June 12, 2017
    Date of Patent: August 20, 2019
    Assignee: Tyco Fire & Security GmbH
    Inventor: Joseph Piccolo, III
  • Patent number: 10373614
    Abstract: In one example, an assistant support server may maintain a web portal to crowdsource responses to a user input. The assistant support server may maintain a web portal accessible by a developer device. The assistant support server may store an assistant rule based on a developer input associating an input word set describing a hypothetical user input with a deep link for a website. The assistant support server may receive in the web portal the developer input. The assistant support server may direct a smart assistant module executed by a user device to connect to the deep link in response to receiving a user input from the smart assistant module matching the input word set.
    Type: Grant
    Filed: December 8, 2016
    Date of Patent: August 6, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Kiril Seksenov, Avishek Mazumder, Kevin Hill, Aditya Pruthi
  • Patent number: 10373604
    Abstract: An acoustic model is adapted, relating acoustic units to speech vectors. The acoustic model comprises a set of acoustic model parameters related to a given speech factor. The acoustic model parameters enable the acoustic model to output speech vectors with different values of the speech factor. The method comprises inputting a sample of speech which is corrupted by noise; determining values of the set of acoustic model parameters which enable the acoustic model to output speech with said first value of the speech factor; and employing said determined values of the set of speech factor parameters in said acoustic model. The acoustic model parameters are obtained by obtaining corrupted speech factor parameters using the sample of speech, and mapping the corrupted speech factor parameters to clean acoustic model parameters using noise characterization paramaters characterizing the noise.
    Type: Grant
    Filed: February 2, 2017
    Date of Patent: August 6, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Javier Latorre-Martinez, Vincent Ping Leung Wan, Kayoko Yanagisawa
  • Patent number: 10366725
    Abstract: Systems and methods are provided to implement and facilitate cross-fading, interstitials and other effects/processing of two or more media elements in a personalized media delivery service. Effects or crossfade processing can occur on the broadcast, publisher or server-side, but can still be personalized to a specific user, in a manner that minimizes processing on the downstream side or client device. The cross-fade can be implemented after decoding, processing, re-encoding, and rechunking the relevant chunks of each component clip. Alternatively, the cross-fade or other effect can be implemented on the relevant chunks in the compressed domain, thus obviating any loss of quality by re-encoding. A large scale personalized content delivery service can limit the processing to essentially the first and last chunks of any file, there being no need to process the full clip.
    Type: Grant
    Filed: September 18, 2017
    Date of Patent: July 30, 2019
    Assignee: Sirius XM Radio Inc.
    Inventors: Raymond Lowe, Christopher Ward
  • Patent number: 10346544
    Abstract: Approaches presented herein enable assignment of translated work to an agent in a support environment based on a confidence factor that measures accuracy of translation and an agent's language skill. Specifically, agent proficiencies in a set of natural languages are measured and scored. An incoming communication is translated into one or more natural languages and each language translation is assigned a translation score based on a confidence of translation. The skill score and translation score are utilized to calculate a confidence factor for each language. In one approach, the communication is assigned to an agent that has a confidence factor greater than a predetermined threshold confidence factor. In another approach, the communication is only assigned if a rule optimizing agent availability and risk of constrained resources is satisfied.
    Type: Grant
    Filed: November 20, 2017
    Date of Patent: July 9, 2019
    Assignee: International Business Machines Corporation
    Inventors: Gary R. Brophy, Dennis D. Koski, Todd A. Mueller, Jeffrey A. Schmidt