Patents Examined by Michael Colucci
  • Patent number: 10074376
    Abstract: A technology of accurately coding and decoding coefficients which are convertible into linear prediction coefficients even for a frame in which the spectrum variation is great while suppressing an increase in the code amount as a whole is provided. A coding device includes: a first coding unit that obtains a first code by coding coefficients which are convertible into linear prediction coefficients of more than one order; and a second coding unit that obtains a second code by coding at least quantization errors of the first coding unit if (A-1) an index Q commensurate with how high the peak-to-valley height of a spectral envelope is, the spectral envelope corresponding to the coefficients which are convertible into the linear prediction coefficients of more than one order, is larger than or equal to a predetermined threshold value Th1 and/or (B-1) an index Q? commensurate with how short the peak-to-valley height of the spectral envelope is, is smaller than or equal to a predetermined threshold value Th1?.
    Type: Grant
    Filed: March 16, 2015
    Date of Patent: September 11, 2018
    Assignee: NIPPON TELEGRAPH AND TELEPHONE CORPORATION
    Inventors: Takehiro Moriya, Yutaka Kamamoto, Noboru Harada
  • Patent number: 10071590
    Abstract: A greeting card having an audio message recording and playback device permits recording of personalized audio messages to be played upon opening of the greeting card. The recording device is operable in either a trial mode or a use mode. In the trial mode, which would be applicable when the card is displayed in a store, a potential purchaser may experience the functionality of the card by recording their own test message. The test message is played back initially for the potential purchaser but is not subsequently played back to be later heard by other potential purchasers. In the use mode, which the card may be switched to after purchase by removal of a trial mode panel from the greeting card, a user recorded message is played repeatedly upon subsequent openings of the card. The user recorded message may be followed by a prerecorded recording, such as a song. Additional prerecorded messages, such as voice prompts with instructions for recording a message, may also be included.
    Type: Grant
    Filed: April 11, 2008
    Date of Patent: September 11, 2018
    Assignee: HALLMARK CARDS, INCORPORATED
    Inventors: Timothy J. Lien, Randy S. Knipp, John B. Watkins
  • Patent number: 10068577
    Abstract: A method of encoding adaptive audio, comprising receiving N objects and associated spatial metadata that describes the continuing motion of these objects, and partitioning the audio into segments based on the spatial metadata. The method encodes adaptive audio having objects and channel beds by capturing a continuing motion of a number N objects in a time-varying matrix trajectory comprising a sequence of matrices, coding coefficients of the time-varying matrix trajectory in spatial metadata to be transmitted via a high-definition audio format for rendering the adaptive audio through a number M output channels, and segmenting the sequence of matrices into a plurality of sub-segments based on the spatial metadata, wherein the plurality of sub-segments are configured to facilitate coding of one or more characteristics of the adaptive audio.
    Type: Grant
    Filed: April 23, 2015
    Date of Patent: September 4, 2018
    Assignee: Dolby Laboratories Licensing Corporation
    Inventors: Vinay Melkote, Malcolm James Law, Roy M. Fejgin
  • Patent number: 10062390
    Abstract: A decoder for generating a frequency enhanced audio signal, includes: a feature extractor for extracting a feature from a core signal; a side information extractor for extracting a selection side information associated with the core signal; a parameter generator for generating a parametric representation for estimating a spectral range of the frequency enhanced audio signal not defined by the core signal, wherein the parameter generator is configured to provide a number of parametric representation alternatives in response to the feature, and wherein the parameter generator is configured to select one of the parametric representation alternatives as the parametric representation in response to the selection side information; and a signal estimator for estimating the frequency enhanced audio signal using the parametric representation selected.
    Type: Grant
    Filed: August 3, 2017
    Date of Patent: August 28, 2018
    Assignee: Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.
    Inventors: Frederik Nagel, Sascha Disch, Andreas Niedermeier
  • Patent number: 10062378
    Abstract: A computer-implemented method and an apparatus are provided. The method includes obtaining, by a processor, a frequency spectrum of an audio signal data. The method further includes extracting, by the processor, periodic indications from the frequency spectrum. The method also includes inputting, by the processor, the periodic indications and components of the frequency spectrum into a neural network. The method additionally includes estimating, by the processor, sound identification information from the neural network.
    Type: Grant
    Filed: February 24, 2017
    Date of Patent: August 28, 2018
    Assignee: International Business Machines Corporation
    Inventors: Takashi Fukuda, Osamu Ichikawa, Bhuvana Ramabhadran
  • Patent number: 10056080
    Abstract: A system receives candidate strings from a speech recognition engine. Where the speech recognition indicates success, the candidate string may be reported or otherwise used. Otherwise, candidate strings may be processed by comparing collapsed forms of the candidate strings to collapsed forms of a user's contacts. The collapsed form may replace contiguous consonants with a 1 and contiguous vowels with a 0. If a candidate string has a collapsed form matching a contact's collapsed form, that contact may be used. Where multiple candidate strings match the collapsed forms of multiple contacts, the contacts may be filtered according to textual similarity to the candidate strings, proximity to a typical usage location for the contacts, and a typical usage time for the contacts.
    Type: Grant
    Filed: October 18, 2016
    Date of Patent: August 21, 2018
    Assignee: FORD GLOBAL TECHNOLOGIES, LLC
    Inventors: Shant Tokatyan, Shounak Athavale
  • Patent number: 10056077
    Abstract: Speech recorded by an audio capture facility of a music facility is processed by a speech recognition facility to generate results that are provided to the music facility. When information related to a music application running on the music facility are provided to the speech recognition facility, the results generated are based at least in part on the application related information. The speech recognition facility uses an unstructured language model for generating results. The user of the music facility may optionally be allowed to edit the results being provided to the music facility. The speech recognition facility may also adapt speech recognition based on usage of the results.
    Type: Grant
    Filed: August 1, 2008
    Date of Patent: August 21, 2018
    Assignee: Nuance Communications, Inc.
    Inventors: Joseph P. Cerra, John N. Nguyen, Michael S. Phillips, Han Shu
  • Patent number: 10056083
    Abstract: The disclosed embodiments illustrate method and system of processing multimedia content to generate a text transcript. The method includes segmenting each of a set of text frames to determine spatial regions. The method further includes extracting one or more keywords from each of the determined spatial regions. The method further includes determining the first set of keywords from the extracted one or more keywords based on filtering of one or more off-topic keywords from the extracted one or more keywords. The method further includes extracting a second set of keywords based on the determined first set of keywords. The method further includes generating a graph between each of a first set of keywords and one or more of a second set of keywords. The method further includes dynamically generating the text transcript of audio content in the multimedia content based on the generated graph.
    Type: Grant
    Filed: October 18, 2016
    Date of Patent: August 21, 2018
    Inventors: Sumit Negi, Sonal S Patil, Arijit Biswas, Ankit Gandhi, Om D Deshmukh
  • Patent number: 10048934
    Abstract: A method and system are provided. The method includes receiving, by a microphone and camera, user utterances indicative of user commands and associated user gestures for the user utterances. The method further includes parsing, by a hardware-based recognizer, sample utterances and the user utterances into verb parts and noun parts. The method also includes recognizing, by a hardware-based recognizer, the user utterances and the associated user gestures based on the sample utterances and descriptions of associated supporting gestures for the sample utterances. The recognizing step includes comparing the verb parts and the noun parts from the user utterances individually and as pairs to the verb parts and the noun parts of the sample utterances. The method additionally includes selectively performing a given one of the user commands responsive to a recognition result.
    Type: Grant
    Filed: February 16, 2015
    Date of Patent: August 14, 2018
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Jonathan Lenchner, Vinay Venkataraman
  • Patent number: 10049101
    Abstract: The present invention discloses a method and system for processing semantic fragments. Some embodiments of the present invention provides a method for processing semantic fragments. The method comprises: obtaining a plurality of groups of semantic fragments, the plurality of groups of semantic fragments at least including a first group of semantic fragments generated from a first data processing flow and a second group of semantic fragments generated from a second data processing flow, the first data processing flow being different from the second data processing flow; and merging the first group of semantic fragment and the second group of semantic fragment based on semantic equivalence. A corresponding system is also disclosed.
    Type: Grant
    Filed: August 6, 2015
    Date of Patent: August 14, 2018
    Assignee: International Business Machines Corporation
    Inventors: Wei Hua Duan, Jia Ji, Jiang Lu, Wei Jie Wang, Qiang Xu, Liang Xue
  • Patent number: 10048935
    Abstract: A method and system are provided. The method includes receiving, by a microphone and camera, user utterances indicative of user commands and associated user gestures for the user utterances. The method further includes parsing, by a hardware-based recognizer, sample utterances and the user utterances into verb parts and noun parts. The method also includes recognizing, by a hardware-based recognizer, the user utterances and the associated user gestures based on the sample utterances and descriptions of associated supporting gestures for the sample utterances. The recognizing step includes comparing the verb parts and the noun parts from the user utterances individually and as pairs to the verb parts and the noun parts of the sample utterances. The method additionally includes selectively performing a given one of the user commands responsive to a recognition result.
    Type: Grant
    Filed: June 24, 2015
    Date of Patent: August 14, 2018
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Jonathan Lenchner, Vinay Venkataraman
  • Patent number: 10037458
    Abstract: A sign language recognizer is configured to detect interest points in an extracted sign language feature, wherein the interest points are localized in space and time in each image acquired from a plurality of frames of a sign language video; apply a filter to determine one or more extrema of a central region of the interest points; associate features with each interest point using a neighboring pixel function; cluster a group of extracted sign language features from the images based on a similarity between the extracted sign language features; represent each image by a histogram of visual words corresponding to the respective image to generate a code book; train a classifier to classify each extracted sign language feature using the code book; detect a posture in each frame of the sign language video using the trained classifier; and construct a sign gesture based on the detected postures.
    Type: Grant
    Filed: May 2, 2017
    Date of Patent: July 31, 2018
    Assignee: King Fahd University of Petroleum and Minerals
    Inventors: Sabri A. Mahmoud, Ala Addin Sidig
  • Patent number: 10032461
    Abstract: An apparatus includes microphone receivers configured to receive microphone signals from a plurality of microphones. A comparator configured to determine a speech similarity indication indicative of a similarity between the microphone signal and non-reverberant speech for each microphone signal. The determination is in response to a comparison of a property derived from the microphone signal to a reference property for non-reverberant speech. In some embodiments, the comparator is configured to determine the similarity indication by comparing to reference properties for speech samples of a set of non-reverberant speech samples. A generator is configured to generate a speech signal by combining the microphone signals in response to the similarity indications. The apparatus may be distributed over a plurality of devices each containing a microphone, and the approach may determine the most suited microphone for generating the speech signal.
    Type: Grant
    Filed: February 18, 2014
    Date of Patent: July 24, 2018
    Assignee: KONINKLIJKE PHILIPS N.V.
    Inventor: Sriram Srinivasan
  • Patent number: 10013991
    Abstract: The present invention proposes a new method for improving the performance of a real-valued filterbank based spectral envelope adjuster. By adaptively locking the gain values for adjacent channels dependent on the sign of the channels, as defined in the application, reduced aliasing is achieved. Furthermore, the grouping of the channels during gain-calculation, gives an improved energy estimate of the real valued subband signals in the filterbank.
    Type: Grant
    Filed: October 31, 2017
    Date of Patent: July 3, 2018
    Assignee: Dolby International AB
    Inventors: Kristofer Kjoerling, Lars Villemoes
  • Patent number: 10013418
    Abstract: There are included an input unit for inputting an input sentence, and an output unit for outputting an output sentence obtained by translating the input sentence into a translation language. The translation language is set based on located language information and position information of a translation device. The located language information includes a predetermined location of each of a plurality of speakers and a used language of each of the plurality of speakers. Accordingly, the translation language, which is a translation target, may be set from a plurality of languages while reducing the operation burden on a user.
    Type: Grant
    Filed: October 20, 2016
    Date of Patent: July 3, 2018
    Assignee: PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD.
    Inventor: Hikaru Usami
  • Patent number: 10013972
    Abstract: An electronic challenge system is used to control access to resources by using a spoken test to identify an origin of a voice. The test is based on a series of questions posed during an interactive dialog session with the entity attempting access.
    Type: Grant
    Filed: July 19, 2016
    Date of Patent: July 3, 2018
    Assignee: J. Nicholas and Kristin Gross Trust U/A/D Apr. 13, 2010
    Inventor: John Nicholas Gross
  • Patent number: 10009474
    Abstract: A method and apparatus of sharing documents during a conference call data is disclosed. One example method may include initiating a document sharing operation during a conference call conducted between at least two participants communicating during the conference call. The method may also include transferring the document from one of the two participants to another of the two participants, and recording at least one action performed to the document by the participants during the conference call.
    Type: Grant
    Filed: March 21, 2017
    Date of Patent: June 26, 2018
    Assignee: West Corporation
    Inventors: Mark J. Pettay, Hendryanto Rilantono, Myron P. Sojka
  • Patent number: 9997172
    Abstract: A system, method and computer program product are described for voice activity detection (VAD) within a digitally encoded bitstream. A parameter extraction module is configured to extract parameters from a sequence of coded frames from a digitally encoded bitstream containing speech. A VAD classifier is configured to operate with input of the digitally encoded bitstream to evaluate each coded frame based on bitstream coding parameter classification features to output a VAD decision indicative of whether or not speech is present in one or more of the coded frames.
    Type: Grant
    Filed: December 2, 2013
    Date of Patent: June 12, 2018
    Assignee: Nuance Communications, Inc.
    Inventors: Daniel A. Barreda, Jose E. G. Lainez, Dushyant Sharma, Patrick Naylor
  • Patent number: 9990929
    Abstract: The present invention proposes a new method for improving the performance of a real-valued filterbank based spectral envelope adjuster. By adaptively locking the gain values for adjacent channels dependent on the sign of the channels, as defined in the application, reduced aliasing is achieved. Furthermore, the grouping of the channels during gain-calculation, gives an improved energy estimate of the real valued subband signals in the filterbank.
    Type: Grant
    Filed: October 31, 2017
    Date of Patent: June 5, 2018
    Assignee: Dolby International AB
    Inventors: Kristofer Kjoerling, Lars Villemoes
  • Patent number: 9971757
    Abstract: The present disclosure relates to a syntax parsing apparatus based on syntax preprocessing and a method thereof. In specific, the present disclosure parses syntaxes that can be parsed by rules and patterns without ambiguity by syntax parsing preprocessing, draws all possible syntax parsing results by applying syntax rules based on a result of syntax parsing preprocessing in which ambiguity is partially resolved, and resolves structural ambiguity by applying a statistic syntax parsing model learned from a syntax tree attachment learning corpus so as to reduce ambiguity in rule-based syntax parsing and to resolve ambiguity by a statistics-based scheme so that parsing correctness and processing efficiency in a syntax parsing method can be enhanced.
    Type: Grant
    Filed: February 28, 2017
    Date of Patent: May 15, 2018
    Assignee: SK PLANET CO., LTD.
    Inventors: Young Sook Hwang, Chang Hao Yin