Patents Examined by Susan McFadden
  • Patent number: 9690782
    Abstract: The technology described herein employs techniques for overlaying (superimposing) translated text on top of (over) scanned text in realtime. The technology recognizes text in an image and identifies a first language in which the text is written. The technology translates the text into a second language and generates an overlay in the second language. The overlay is then inserted into the display over the text in the image.
    Type: Grant
    Filed: May 2, 2015
    Date of Patent: June 27, 2017
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Magdalena Vukosavljevic, Ruchita Bhargava, Matthew S. Ashman, Ivan Stojiljkovic, Pavle Josipović, Djordje Nijemcevic, Milan Vugdelija
  • Patent number: 9685158
    Abstract: Systems and methods to process and/or present information relating to voice messages for a user that are received from other persons. In one embodiment, a method implemented in a data processing system includes: receiving first data associated with prior communications or activities for the user on a mobile device; receiving a voice message for the user; transcribing the voice message using the first data to provide a transcribed message; and sending the transcribed message to the mobile device for display to the user. In one embodiment, options are displayed for selection by the user to correct a misspelled word in the transcribed message.
    Type: Grant
    Filed: February 27, 2015
    Date of Patent: June 20, 2017
    Assignee: YAHOO! INC.
    Inventor: Jeffrey Bonforte
  • Patent number: 9685153
    Abstract: In some embodiments, the recognition results produced by a speech processing system (which may include a top recognition result and one or more alternative recognition results) based on an analysis of a speech input, are evaluated for indications of potential significant errors. In some embodiments, the recognition results may be evaluated to determine whether a meaning of any of the alternative recognition results differs from a meaning of the top recognition result in a manner that is significant for the domain. In some embodiments, one or more of the recognition results may be evaluated to determine whether the result(s) include one or more words or phrases that, when included in a result, would change a meaning of the result in a manner that would be significant for the domain.
    Type: Grant
    Filed: May 15, 2015
    Date of Patent: June 20, 2017
    Assignee: Nuance Communications, Inc.
    Inventors: William F. Ganong, III, Raghu Vemula, Robert Fleming
  • Patent number: 9679582
    Abstract: A user device provides dynamic speech processing services during variable network connectivity with a network server. The user device includes a monitor that monitors a level of network connectivity between the user device and the network server. A user device speech processor processes speech data and is initiated based on a determination that the level of network connectivity between the user device and the network server is impaired. The monitor determines when the level of network connectivity between the user device and the network server is no longer impaired.
    Type: Grant
    Filed: July 20, 2015
    Date of Patent: June 13, 2017
    Assignee: Nuance Communications, Inc.
    Inventor: Horst Schroeter
  • Patent number: 9674353
    Abstract: A third party verification system, method, and computer readable medium comprising a response unit in communication with at least a first party, said response unit configured to automatically elicit information in the form of a dual tone multi-frequency (DTMF) response from the at least first party providing evidence of consent granted by the at least first party to enter into at least one transaction with at least a second party, and a print server in communication with said response unit, said print server configured to receive and store said information to provide a record for third party verification that the at least first party consented to enter into the at least one transaction with the at least one second party.
    Type: Grant
    Filed: September 23, 2014
    Date of Patent: June 6, 2017
    Assignee: Open Invention Network, LLC
    Inventor: Jonathan P. McIntosh
  • Patent number: 9672810
    Abstract: A method in a computing device for decoding a weighted finite state transducer (WFST) for automatic speech recognition is described. The method includes sorting a set of one or more WFST arcs based on their arc weight in ascending order. The method further includes iterating through each arc in the sorted set of arcs according to the ascending order until the score of the generated token corresponding to an arc exceeds a score threshold. The method further includes discarding any remaining arcs in the set of arcs that have yet to be considered.
    Type: Grant
    Filed: September 26, 2014
    Date of Patent: June 6, 2017
    Assignee: Intel Corporation
    Inventors: Joachim Hofer, Georg Stemmer
  • Patent number: 9668115
    Abstract: A system, method, and computer-readable storage device for sending a spoken message as a text message. The method includes receiving from a subscriber a spoken message and spoken disambiguating information indicating how to deliver the spoken message to a recipient. The method further includes converting the spoken message to text and communicating the text to the recipient address. The method can also include determining a recipient address for the recipient based on spoken disambiguating information.
    Type: Grant
    Filed: November 2, 2015
    Date of Patent: May 30, 2017
    Assignee: Genesis Star Management Limited
    Inventor: Sangar Dowlatkhah
  • Patent number: 9652787
    Abstract: A system comprising a computer-readable storage medium storing at least one program and a computer-implemented method for creating messages using generative grammar models is presented. Consistent with some embodiments, the method may include receiving a request to generate a message, which in an example embodiment is to be published to a social network platform. In response to receiving the request, a generative grammar model defining the structure of the message is accessed. The generative grammar model may include a number of blanks and may specify a source along with a grammatical constraint for a term to populate each blank. The method may further include generating the message in accordance with the generative grammar model, and causing the generated message to be published.
    Type: Grant
    Filed: September 29, 2014
    Date of Patent: May 16, 2017
    Assignee: eBay Inc.
    Inventors: Elizabeth Churchill, Atish Das Sarma, Corinne Elizabeth Sherman, Gyanit Singh
  • Patent number: 9646630
    Abstract: An apparatus, system, and computer readable media for data pre-processing and processing for voice recognition are described herein. The apparatus includes logic to pre-process multi-channel audio data and logic to resolve a source location. The apparatus also includes logic to perform wide range adaptive beam forming, and logic to perform full voice recognition.
    Type: Grant
    Filed: August 30, 2013
    Date of Patent: May 9, 2017
    Assignee: Intel Corporation
    Inventor: Gangatharan Jothiswaran
  • Patent number: 9646002
    Abstract: There is provided a method that includes displaying, on a display, a viewing pane of available video contents including a first video content, receiving a selection of the first video content from the available video contents, transmitting a language selection and the selection of the first video content to a server, receiving a language content corresponding to the language selection and the selection of the first video content from the server, and displaying, on the display, the first video content in synchronization with playing the language content.
    Type: Grant
    Filed: December 15, 2014
    Date of Patent: May 9, 2017
    Assignee: Disney Enterprises, Inc.
    Inventors: Artin Nazarian, Greg Head, Paul Marz
  • Patent number: 9639149
    Abstract: A method and system is provided that controls an external output function of a mobile device according to control interactions received via the microphone. The method includes, activating a microphone according to preset optional information when the mobile device enters an external output mode, performing an external output operation in the external output mode, detecting an interaction based on sound information in the external output mode, and controlling the external output according to the interaction.
    Type: Grant
    Filed: December 29, 2014
    Date of Patent: May 2, 2017
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Hee Woon Kim, Si Hak Jang
  • Patent number: 9633004
    Abstract: Systems and processes for operating a virtual assistant programmed to refer to shared domain concepts using concept nodes are provided. In some examples, to process a textual representation of user speech using an active ontology having these concept nodes, a primary user intent can be determined from the textual representation of user speech. Concepts referred to by the primary user intent can be identified, and substrings of the textual representation of user speech corresponding to the concepts can be identified. Secondary user intents for the substrings can be determined and a task flow based on the primary user intent and the secondary user intents can be generated and performed.
    Type: Grant
    Filed: September 29, 2014
    Date of Patent: April 25, 2017
    Assignee: Apple Inc.
    Inventors: Richard D. Giuli, Nicholas K. Treadgold
  • Patent number: 9626431
    Abstract: Various embodiments for adjusting a search result user interface when a foreign language search query is detected. A search query associated with a search request is obtained. A language of the search query is identified. An search result user interface is generated using a foreign language template when the language varies from an expected language. The search result user interface includes results based on the foreign language search query.
    Type: Grant
    Filed: January 27, 2015
    Date of Patent: April 18, 2017
    Assignee: Amazon Technologies, Inc.
    Inventors: Rahul H. Bhagat, Michael Brundage, Daniel R. Parshall
  • Patent number: 9620105
    Abstract: Systems and processes for analyzing audio input for efficient speech and music recognition are provided. In one example process, an audio input can be received. A determination can be made as to whether the audio input includes music. In addition, a determination can be made as to whether the audio input includes speech. In response to determining that the audio input includes music, an acoustic fingerprint representing a portion of the audio input that includes music is generated. In response to determining that the audio input includes speech rather than music, an end-point of a speech utterance of the audio input is identified.
    Type: Grant
    Filed: September 29, 2014
    Date of Patent: April 11, 2017
    Assignee: Apple Inc.
    Inventor: Henry Mason
  • Patent number: 9620127
    Abstract: Systems, computer-implemented methods, and tangible computer-readable media are provided for voice authentication. The method includes receiving a speech sample from a user through an Internet browser for authentication as part of a request for a restricted-access resource, pre-authenticating the speech sample, transmitting an authentication request to an authentication server, which compares the speech sample to a previously established speech profile associated with the user, and providing access to the restricted access resource based on the response.
    Type: Grant
    Filed: November 10, 2014
    Date of Patent: April 11, 2017
    Assignee: Nuance Communications, Inc.
    Inventor: Saurabh Kumar
  • Patent number: 9607609
    Abstract: Disclosed are embodiments for use in an articulatory-based text-to-speech conversion system configured to establish an articulatory speech synthesis model of a person's voice based on facial characteristics defining exteriorly visible articulatory speech synthesis model parameters of the person's voice and on a predefined articulatory speech synthesis model selected from among stores of predefined models.
    Type: Grant
    Filed: September 25, 2014
    Date of Patent: March 28, 2017
    Assignee: INTEL CORPORATION
    Inventors: Shamim Begum, Alexander A. Oganezov
  • Patent number: 9601125
    Abstract: A method includes receiving a first value of a mixing factor. The first value corresponds to a first portion of an audio signal received at an audio encoder. The method includes receiving a second value of the mixing factor. The second value corresponds to a second portion of the audio signal. The method also includes generating a third value of the mixing factor at least partially based on the first value and the second value and mixing an excitation signal with modulated noise based on the third value. Another method includes determining a first set of spectral frequency values corresponding to an audio signal and determining a second set of spectral frequency values that approximates the first set of spectral frequency values. A gain value corresponding to at least a portion of the audio signal is adjusted based on a difference between the first set and the second set.
    Type: Grant
    Filed: August 28, 2013
    Date of Patent: March 21, 2017
    Assignee: QUALCOMM Incorporated
    Inventors: Venkatraman Srinivasa Atti, Venkatesh Krishnan
  • Patent number: 9601109
    Abstract: A method for training a deep neural network, comprises receiving and formatting speech data for the training, preconditioning a system of equations to be used for analyzing the speech data in connection with the training by using a non-fixed point quasi-Newton preconditioning scheme, and employing flexible Krylov subspace solvers in response to variations in the preconditioning scheme for different iterations of the training.
    Type: Grant
    Filed: September 29, 2014
    Date of Patent: March 21, 2017
    Assignee: International Business Machines Corporation
    Inventors: Lior Horesh, Brian E. D. Kingsbury, Tara N. Sainath
  • Patent number: 9570056
    Abstract: An audio data synthesis method including a time of a plurality of audio data is adjusted without using a device which can acquire the standard time. Specifically, audio data is obtained based on synchronized recording of the first and second recorders without using standard time. A time difference is calculated between an own terminal and another terminal, based on the time at which output of a first sound from the audio output module is started, a time at which input of a sound corresponding to the audio data to the audio input module is started, a time indicated by the first information, and a time indicated by the second information. Second and third audio data is synthesized after a time difference between the second and third audio data based on the third sound which is input to the audio input module is adjusted, based on the time difference.
    Type: Grant
    Filed: September 26, 2014
    Date of Patent: February 14, 2017
    Assignee: OLYMPUS CORPORATION
    Inventor: Ryuichi Kiyoshige
  • Patent number: 9570065
    Abstract: Techniques for performing multi-style speech synthesis. The techniques include using at least one computer hardware processor to perform: obtaining input comprising text and an identification of a first speaking style to use in rendering the text as speech; identifying a plurality of speech segments for use in rendering the text as speech, the identified plurality of speech segments comprising a first speech segment having the first speaking style and a second speech segment having a second speaking style different from the first speaking style; and rendering the text as speech having the first speaking style, at least in part, by using the identified plurality of speech segments.
    Type: Grant
    Filed: September 29, 2014
    Date of Patent: February 14, 2017
    Assignee: Nuance Communications, Inc.
    Inventor: Vincent Pollet