Preliminary Matching Patents (Class 704/252)
  • Patent number: 10468136
    Abstract: Disclosed are embodiments of method and system to predict health condition of a human subject. The method comprises receiving historical human-subject related data including records corresponding to multiple data views. The method estimates one or more latent variables based on: a first value indicative of count of records in a cluster, a second value indicative of count of records, and a third value indicative of a parameter utilizable to predict a fourth value. The fourth value corresponds to selection probability of a D-vine pair copula family, of a D-vine mixture model, utilizable to model a cluster. The method generates the D-vine mixture model based on the estimated one or more latent variables. The method further comprises receiving multi-view data of a second human subject and predicting health condition of the second human subject based on the multi-view data using a classifier trained based on the estimated latent variables.
    Type: Grant
    Filed: August 29, 2016
    Date of Patent: November 5, 2019
    Assignee: CONDUENT BUSINESS SERVICES, LLC
    Inventors: Lavanya Sita Tekumalla, Vaibhav Rajan
  • Patent number: 10459008
    Abstract: A method is disclosed for triggering upon signal events occurring in frequency domain signals. The method includes repeatedly sampling a time-varying signal and generating a plurality of digital frequency domain spectrums based on the samples of the time-varying signal. A frequency domain bitmap for the time-varying signal is repeatedly updated via application of the digital frequency domain spectrums. The method further includes selecting a portion of the frequency domain bitmap, determining a signal occupancy in the selected portion, and triggering a capture of the time-varying signal based on and in response to the occupancy determination for the selected portion of the bitmap.
    Type: Grant
    Filed: October 14, 2014
    Date of Patent: October 29, 2019
    Assignee: Tektronix, Inc.
    Inventors: Robert E. Tracy, Kathryn A. Engholm, Alfred K. Hillman, Jr.
  • Patent number: 10460074
    Abstract: Disclosed are embodiments of methods and systems for predicting a health condition of a first human subject. The method comprises receiving a measure of one or more physiological parameters associated with the first human subject. The method estimates one or more latent variables based on a first count indicative of a number of the plurality of d-vines, a second count indicative of a number of the one or more records, a first value that is representative of a number of the one or more records clustered into a d-vine from the plurality of d-vines, and a second value that is representative of a parameter utilizable to predict a third value. The method generates the plurality of d-vines based on the estimated one or more latent variables. The method predicts health condition of the first human subject by utilizing a trained classifier based on the estimated one or more latent variables.
    Type: Grant
    Filed: April 5, 2016
    Date of Patent: October 29, 2019
    Assignee: CONDUENT BUSINESS SERVICES, LLC
    Inventors: Lavanya Sita Tekumalla, Vaibhav Rajan
  • Patent number: 10448898
    Abstract: Disclosed are embodiments of methods and systems for predicting a health condition of a first human subject. The method comprises extracting a historical data including physiological parameters of second human subjects. Thereafter, a first distribution of a first physiological parameter is determined based on a marginal cumulative distribution of a rank transformed historical data. Further, a second distribution of a second physiological parameter is determined based on the first distribution and a first conditional cumulative distribution of the rank transformed historical data. Further, a latent variable is determined based on the first and the second distributions. Thereafter, one or more parameters of at least one bivariate distribution, corresponding to a D-vine copula, are estimated based on the latent variable. Further, a classifier is trained based on the D-vine copula. The classifier is utilizable to predict the health condition of the first human subject based on his/her physiological parameters.
    Type: Grant
    Filed: July 14, 2015
    Date of Patent: October 22, 2019
    Assignee: CONDUENT BUSINESS SERVICES, LLC
    Inventors: Lavanya Sita Tekumalla, Vaibhav Rajan
  • Patent number: 10360903
    Abstract: According to one embodiment, an apparatus includes a storage unit, a first acquisition unit, a second acquisition unit, an analyzer, and a recognition unit. The storage unit stores first situation information about a situation assumed in advance, a first representation representing a meaning of a sentence assumed, intention information representing an intention to be estimated, and a first value representing a degree of application of the first representation to the first situation information and the intention information. The first acquisition unit acquires a natural sentence. The second acquisition unit acquires second situation information about a situation when acquiring the natural sentence. The analyzer analyzes the natural sentence and generates a second representation representing a meaning of the natural sentence. The recognition unit obtains an estimated value based on the first value associated with the first situation information and the first representation.
    Type: Grant
    Filed: February 15, 2017
    Date of Patent: July 23, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Hiromi Wakaki, Kenji Iwata, Masayuki Okamoto
  • Patent number: 10354308
    Abstract: Offer listings can be classified as accessory offers or product offers using a classification operation performed on a corpus of offers. Data from the classification operation can be used to classify received queries as either product or accessory, and to classify results as products or accessories for purposes of presenting a relevant list of results to a user.
    Type: Grant
    Filed: July 2, 2015
    Date of Patent: July 16, 2019
    Assignee: GOOGLE LLC
    Inventors: Srinath Sridhar, Ashutosh Garg, Kedar Dhamdhere, Varun Kacholia
  • Patent number: 10339918
    Abstract: An embodiment of a speech endpoint detector apparatus may include a speech detector to detect a presence of speech in an electronic speech signal, a pause duration measurer communicatively coupled to the speech detector to measure a duration of a pause following a period of detected speech, an end of utterance detector communicatively coupled to the pause duration measurer to detect if the pause measured following the period of detected speech is greater than a pause threshold corresponding to an end of an utterance, and a pause threshold adjuster to adaptively adjust the pause threshold corresponding to an end of an utterance based on stored pause information. Other embodiments are disclosed and claimed.
    Type: Grant
    Filed: September 27, 2016
    Date of Patent: July 2, 2019
    Assignee: Intel IP Corporation
    Inventors: Joachim Hofer, Munir Georges
  • Patent number: 10337878
    Abstract: An in-vehicle terminal sends a spoken voice as a voice signal to a relay server, and the relay server includes a voice recognition unit which converts the received voice signal into a string, and a control unit which searches for information stored in a main database or a temporary storage database by using the string and sends a search result to the in-vehicle terminal, and, upon searching for information stored in the main database, stores the search result in the temporary storage database. Upon receiving a voice signal, when the search result is stored in the temporary storage database, the control unit searches for information stored in the temporary storage database by using the string converted from the received voice signal, and, when the search result is not stored in the temporary storage database, the control unit searches for information stored in the main database by using the string.
    Type: Grant
    Filed: October 2, 2015
    Date of Patent: July 2, 2019
    Assignee: CLARION CO., LTD.
    Inventors: Susumu Kojima, Takashi Yamaguchi, Hideki Takano, Yasushi Nagai
  • Patent number: 10268989
    Abstract: Example medical device data platforms are disclosed herein. In an example, the platform may include at least one integration device to access information originating from a plurality of implantable medical devices manufactured by a plurality of manufacturers and implanted in a plurality of patients. The system may also include an information processor to process the accessed information to generate at least one of patient-oriented information and provider-oriented information. The system may also include at least one communication device providing at least one of a patient portal and a provider portal to provide the patient-oriented information and the provider-oriented information, respectively.
    Type: Grant
    Filed: April 20, 2016
    Date of Patent: April 23, 2019
    Assignee: Murj, Inc.
    Inventors: Richard Todd Butka, Christopher Steven Irving, Patrick Beaulieu
  • Patent number: 10224026
    Abstract: An electronic device comprising circuitry configured to record sensor data that is obtained from data sources and to retrieve information from the recorded sensor data using concepts that are defined by a user.
    Type: Grant
    Filed: March 2, 2017
    Date of Patent: March 5, 2019
    Assignee: SONY CORPORATION
    Inventors: Aurel Bordewieck, Fabien Cardinaux, Wilhelm Hagg, Thomas Kemp, Stefan Uhlich, Fritz Hohl
  • Patent number: 10224030
    Abstract: In speech processing systems personalization is added in the Natural Language Understanding (NLU) processor by incorporating external knowledge sources of user information to improve entity recognition performance of the speech processing system. Personalization in the NLU is effected by incorporating one or more dictionaries of entries, or gazetteers, with information personal to a respective user, that provide the user's information to permit disambiguation of semantic interpretation for input utterances to improve quality of speech processing results.
    Type: Grant
    Filed: March 14, 2013
    Date of Patent: March 5, 2019
    Assignee: AMAZON TECHNOLOGIES, INC.
    Inventors: Imre Attila Kiss, Arthur Richard Toth, Lambert Mathias
  • Patent number: 10152899
    Abstract: A training tool, method and a system for measuring crew member communication skills are disclosed, wherein an audio data processing terminal interfaced with a crew training apparatus, typically a crew-operated vehicle simulator. Audio data corresponding to a conversation between at least two crew members is recording during a training session and stored. Respective audio data of each crew member is extracted from the stored audio data, and a series of measures for at least one prosodic parameter in each respective audio data extracted is computed. A correlation coefficient of the series of measures is then computed, wherein the correlation coefficient is indicative of a level of prosodic accommodation between the at least two crew members. Specific communication skills in addition to prosodic accommodation performance can the be determined inferred.
    Type: Grant
    Filed: July 31, 2014
    Date of Patent: December 11, 2018
    Assignee: Crewfactors Limited
    Inventors: Brian Vaughan, Celine De Looze
  • Patent number: 10134386
    Abstract: Systems and methods for identifying content corresponding to a language are provided. Language spoken by a first user based on verbal input received from the first user is automatically determined with voice recognition circuitry. A database of content sources is cross-referenced to identify a content source associated with a language field value that corresponds to the determined language spoken by the first user. The language field in the database identifies the language that the associated content source transmits content to a plurality of users. A representation of the identified content source is generated for display to the first user.
    Type: Grant
    Filed: July 21, 2015
    Date of Patent: November 20, 2018
    Assignee: Rovi Guides, Inc.
    Inventor: Shuchita Mehra
  • Patent number: 10108608
    Abstract: A dialog state tracking system. One aspect of the system is the use of multiple utterance decoders and/or multiple spoken language understanding (SLU) engines generating competing results that improve the likelihood that the correct dialog state is available to the system and provide additional features for scoring dialog state hypotheses. An additional aspect is training a SLU engine and a dialog state scorer/ranker DSR engine using different subsets from a single annotated training data set. A further aspect is training multiple SLU/DSR engine pairs from inverted subsets of the annotated training data set. Another aspect is web-style dialog state ranking based on dialog state features using discriminative models with automatically generated feature conjunctions. Yet another aspect is using multiple parameter sets with each ranking engine and averaging the rankings. Each aspect independently improves dialog state tracking accuracy and may be combined in various combinations for greater improvement.
    Type: Grant
    Filed: June 12, 2014
    Date of Patent: October 23, 2018
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Jason D. Williams, Geoffrey G. Zweig
  • Patent number: 10043516
    Abstract: Systems and processes for operating an automated assistant are disclosed. In one example process, an electronic device provides an audio output via a speaker of the electronic device. While providing the audio output, the electronic device receives, via a microphone of the electronic device, a natural language speech input. The electronic device derives a representation of user intent based on the natural language speech input and the audio output, identifies a task based on the derived user intent; and performs the identified task.
    Type: Grant
    Filed: December 20, 2016
    Date of Patent: August 7, 2018
    Assignee: Apple Inc.
    Inventors: Harry J. Saddler, Aimee T. Piercy, Garrett L. Weinberg, Susan L. Booker
  • Patent number: 9911409
    Abstract: A speech recognition apparatus includes a processor configured to recognize a user's speech using any one or combination of two or more of an acoustic model, a pronunciation dictionary including primitive words, and a language model including primitive words; and correct word spacing in a result of speech recognition based on a word-spacing model.
    Type: Grant
    Filed: July 21, 2016
    Date of Patent: March 6, 2018
    Assignee: Samsung Electronics Co., Ltd.
    Inventor: Seokjin Hong
  • Patent number: 9805715
    Abstract: A method of recognizing speech commands includes generating a background acoustic model for a sound using a first sound sample, the background acoustic model characterized by a first precision metric. A foreground acoustic model is generated for the sound using a second sound sample, the foreground acoustic model characterized by a second precision metric. A third sound sample is received and decoded by assigning a weight to the third sound sample corresponding to a probability that the sound sample originated in a foreground using the foreground acoustic model and the background acoustic model. The method further includes determining if the weight meets predefined criteria for assigning the third sound sample to the foreground and, when the weight meets the predefined criteria, interpreting the third sound sample as a portion of a speech command. Otherwise, recognition of the third sound sample as a portion of a speech command is forgone.
    Type: Grant
    Filed: December 13, 2013
    Date of Patent: October 31, 2017
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Shuai Yue, Li Lu, Xiang Zhang, Dadong Xie, Haibo Liu, Bo Chen, Jian Liu
  • Patent number: 9784765
    Abstract: A system and method are provided for graphically actuating a trigger in a test and measurement device. The method includes displaying a visual representation of signal properties for one or more time-varying signals. A graphical user input is received, in which a portion of the visual representation is designated. The method further includes configuring a trigger of the test and measurement device in response to the graphical user input, by setting a value for a trigger parameter of the trigger. The set value for the trigger parameters varies with and is dependent upon the particular portion of the visual representation that is designated by the graphical user input. The trigger is then employed in connection with subsequent monitoring of signals within the test and measurement device.
    Type: Grant
    Filed: November 3, 2009
    Date of Patent: October 10, 2017
    Assignee: Tektronix, Inc.
    Inventors: Kathryn A. Engholm, Cecilia A. Case
  • Patent number: 9703394
    Abstract: In some examples, a method includes outputting a graphical keyboard (120) for display and responsive to receiving an indication of a first input (124), determining a new character string that is not included in a language model. The method may include adding the new character string to the language model and associating a likelihood value with the new character string. The method may include, responsive to receiving an indication of a second input, predicting the new character string, and responsive to receiving an indication of a third input that rejects the new character string, decreasing the likelihood value associated with the new character string.
    Type: Grant
    Filed: October 1, 2015
    Date of Patent: July 11, 2017
    Assignee: Google Inc.
    Inventors: Yu Ouyang, Shumin Zhai
  • Patent number: 9704482
    Abstract: A method for spoken term detection, comprising generating a time-marked word list, wherein the time-marked word list is an output of an automatic speech recognition system, generating an index from the time-marked word list, wherein generating the index comprises creating a word loop weighted finite state transducer for each utterance, i, receiving a plurality of keyword queries, and searching the index for a plurality of keyword hits.
    Type: Grant
    Filed: March 11, 2015
    Date of Patent: July 11, 2017
    Assignee: International Business Machines Corporation
    Inventors: Brian E. D. Kingsbury, Lidia Mangu, Michael A. Picheny, George A. Saon
  • Patent number: 9697830
    Abstract: A method for spoken term detection, comprising generating a time-marked word list, wherein the time-marked word list is an output of an automatic speech recognition system, generating an index from the time-marked word list, wherein generating the index comprises creating a word loop weighted finite state transducer for each utterance, i, receiving a plurality of keyword queries, and searching the index for a plurality of keyword hits.
    Type: Grant
    Filed: June 25, 2015
    Date of Patent: July 4, 2017
    Assignee: International Business Machines Corporation
    Inventors: Brian E. D. Kingsbury, Lidia Mangu, Michael A. Picheny, George A. Saon
  • Patent number: 9672201
    Abstract: Systems, methods and apparatus for learning parsing rules and argument identification from crowdsourcing of proposed command inputs. Crowdsourcing techniques to generate rules for parsing input sentences. A parse is used to determine whether the input sentence invokes a specific action, and if so, what arguments are to be passed to the invocation of the action.
    Type: Grant
    Filed: April 27, 2016
    Date of Patent: June 6, 2017
    Assignee: Google Inc.
    Inventors: Jakob D. Uszkoreit, Percy Liang
  • Patent number: 9620109
    Abstract: A server and a guide sentence generating method are provided. The method includes receiving user speech, analyzing the user speech, determining a category of the user speech from among a plurality of categories, storing the user speech in the determined category, determining a usage frequency and a popularity of each of the plurality of categories, selecting a category from among the plurality of categories based on the usage frequency and the popularity, and generating a guide sentence corresponding to the selected category.
    Type: Grant
    Filed: February 18, 2015
    Date of Patent: April 11, 2017
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: In-jee Song, Ji-hye Chung
  • Patent number: 9542936
    Abstract: A method including: receiving, on a computer system, a text search query, the query including one or more query words; generating, on the computer system, for each query word in the query, one or more anchor segments within a plurality of speech recognition processed audio files, the one or more anchor segments identifying possible locations containing the query word; post-processing, on the computer system, the one or more anchor segments, the post-processing including: expanding the one or more anchor segments; sorting the one or more anchor segments; and merging overlapping ones of the one or more anchor segments; and searching, on the computer system, the post-processed one or more anchor segments for instances of at least one of the one or more query words using a constrained grammar.
    Type: Grant
    Filed: May 2, 2013
    Date of Patent: January 10, 2017
    Assignee: Genesys Telecommunications Laboratories, Inc.
    Inventors: Amir Lev-Tov, Avi Faizakof, Yochai Konig
  • Patent number: 9508339
    Abstract: A method for updating language understanding classifier models includes receiving via one or more microphones of a computing device, a digital voice input from a user of the computing device. Natural language processing using the digital voice input is used to determine a user voice request. Upon determining the user voice request does not match at least one of a plurality of pre-defined voice commands in a schema definition of a digital personal assistant, a GUI of an end-user labeling tool is used to receive a user selection of at least one of the following: at least one intent of a plurality of available intents and/or at least one slot for the at least one intent. A labeled data set is generated by pairing the user voice request and the user selection, and is used to update a language understanding classifier.
    Type: Grant
    Filed: January 30, 2015
    Date of Patent: November 29, 2016
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Vishwac Sena Kannan, Aleksandar Uzelac, Daniel J. Hwang
  • Patent number: 9436382
    Abstract: Natural language image editing techniques are described. In one or more implementations, a natural language input is converted from audio data using a speech-to-text engine. A gesture is recognized from one or more touch inputs detected using one or more touch sensors. Performance is then initiated of an operation identified from a combination of the natural language input and the recognized gesture.
    Type: Grant
    Filed: November 21, 2012
    Date of Patent: September 6, 2016
    Assignee: Adobe Systems Incorporated
    Inventors: Gregg D. Wilensky, Walter W. Chang, Lubomira A. Dontcheva, Gierad P. Laput, Aseem O. Agarwala
  • Patent number: 9275139
    Abstract: System and method to search audio data, including: receiving audio data representing speech; receiving a search query related to the audio data; compiling, by use of a processor, the search query into a hierarchy of scored speech recognition sub-searches; searching, by use of a processor, the audio data for speech identified by one or more of the sub-searches to produce hits; and combining, by use of a processor, the hits by use of at least one combination function to provide a composite search score of the audio data. The combination function may include an at-least-M-of-N function that produces a high score when at least M of N function inputs exceed a predetermined threshold value. The composite search score employ a soft time window such as a spline function.
    Type: Grant
    Filed: September 27, 2012
    Date of Patent: March 1, 2016
    Assignee: Aurix Limited
    Inventor: Keith Michael Ponting
  • Patent number: 9251141
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for training an entity identification model. In one aspect, a method includes obtaining a plurality of complete sentences that each include entity text that references a first entity; for each complete sentence in the plurality of complete sentences: providing a first portion of the complete sentence as input to an entity identification model that determines a predicted entity for the first portion of the complete sentence, the first portion being less than all of the complete sentence; comparing the predicted entity to the first entity; and updating the entity identification model based on the comparison of the predicted entity to the first entity.
    Type: Grant
    Filed: May 12, 2014
    Date of Patent: February 2, 2016
    Assignee: Google Inc.
    Inventors: Maxim Gubin, Sangsoo Sung, Krishna Bharat, Kenneth W. Dauber
  • Patent number: 9117452
    Abstract: A language processing system identifies, from log data, command inputs that parsed to a parsing rule associated with an action. If the command input has a signal indicative of user satisfaction, where the signal is derived from data that is not generated from performance of the action (e.g., user interactions with data provided in response to the performance of another, different action; resources identified in response to the performance of another, different action having a high quality score; etc.), then exception data is generated for the parsing rule. The exception data specifies the particular instance of the sentence parsed by the parsing rule, and precludes invocation of the action associated with the rule.
    Type: Grant
    Filed: June 25, 2013
    Date of Patent: August 25, 2015
    Assignee: Google Inc.
    Inventors: Jakob D. Uszkoreit, Percy Liang, Daniel M. Bikel
  • Patent number: 9037470
    Abstract: Apparatus and methods are provided for using automatic speech recognition to analyze a voice interaction and verify compliance of an agent reading a script to a client during the voice interaction. In one aspect of the invention, a communications system includes a user interface, a communications network, and a call center having an automatic speech recognition component. In other aspects of the invention, a script compliance method includes the steps of conducting a voice interaction between an agent and a client and evaluating the voice interaction with an automatic speech recognition component adapted to analyze the voice interaction and determine whether the agent has adequately followed the script. In yet still further aspects of the invention, the duration of a given interaction can be analyzed, either apart from or in combination with the script compliance analysis above, to seek to identify instances of agent non-compliance, of fraud, or of quality-analysis issues.
    Type: Grant
    Filed: June 25, 2014
    Date of Patent: May 19, 2015
    Assignee: West Business Solutions, LLC
    Inventors: Mark J. Pettay, Fonda J. Narke
  • Patent number: 9026447
    Abstract: A first communication path for receiving a communication is established. The communication includes speech, which is processed. A speech pattern is identified as including a voice-command. A portion of the speech pattern is determined as including the voice-command. That portion of the speech pattern is separated from the speech pattern and compared with a second speech pattern. If the two speech patterns match or resemble each other, the portion of the speech pattern is accepted as the voice-command. An operation corresponding to the voice-command is determined and performed. The operation may perform an operation on a remote device, forward the voice-command to a remote device, or notify a user. The operation may create a second communication path that may allow a headset to join in a communication between another headset and a communication device, several headsets to communicate with each other, or a headset to communicate with several communication devices.
    Type: Grant
    Filed: January 25, 2008
    Date of Patent: May 5, 2015
    Assignee: CenturyLink Intellectual Property LLC
    Inventors: Erik Geldbach, Kelsyn D. Rooks, Sr., Shane M. Smith, Mark Wilmoth
  • Patent number: 9026431
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for semantic parsing with multiple parsers. One of the methods includes obtaining one or more transcribed prompt n-grams from a speech to text recognizer, providing the transcribed prompt n-grams to a first semantic parser that executes on the user device and accesses a first knowledge base for results responsive to the spoken prompt, providing the transcribed prompt n-grams to a second semantic parser that accesses a second knowledge base for results responsive to the spoken prompt, the first knowledge base including first data not included in the second knowledge base, receiving a result responsive to the spoken prompt from the first semantic parser or the second semantic parser, wherein the result is selected from the knowledge base associated with the semantic parser that provided the result to the user device, and performing an operation based on the result.
    Type: Grant
    Filed: July 30, 2013
    Date of Patent: May 5, 2015
    Assignee: Google Inc.
    Inventors: Pedro J. Moreno Mengibar, Diego Melendo Casado, Fadi Biadsy
  • Patent number: 9015043
    Abstract: A computer-implemented method includes receiving an electronic representation of one or more human voices, recognizing words in a first portion of the electronic representation of the one or more human voices, and sending suggested search terms to a display device for display to a user in a text format. The suggested search terms are based on the recognized words in the first portion of the electronic representation of the one or more human voices. A search query is received from the user, which includes one or more of the suggested search terms that were displayed to the user.
    Type: Grant
    Filed: October 1, 2010
    Date of Patent: April 21, 2015
    Assignee: Google Inc.
    Inventor: Scott Jenson
  • Patent number: 9009025
    Abstract: In some implementations, a digital work provider may provide language model information related to a plurality of different contexts, such as a plurality of different digital works. For example, the language model information may include language model difference information identifying a plurality of sequences of one or more words in a digital work that have probabilities of occurrence that differ from probabilities of occurrence in a base language model by a threshold amount. The language model difference information corresponding to a particular context may be used in conjunction with the base language model to recognize an utterance made by a user of a user device. In some examples, the recognition is performed on the user device. In other examples, the utterance and associated context information are sent over a network to a recognition computing device that performs the recognition.
    Type: Grant
    Filed: December 27, 2011
    Date of Patent: April 14, 2015
    Assignee: Amazon Technologies, Inc.
    Inventor: Brandon W. Porter
  • Patent number: 9002710
    Abstract: The invention involves the loading and unloading of dynamic section grammars and language models in a speech recognition system. The values of the sections of the structured document are either determined in advance from a collection of documents of the same domain, document type, and speaker; or collected incrementally from documents of the same domain, document type, and speaker; or added incrementally to an already existing set of values. Speech recognition in the context of the given field is constrained to the contents of these dynamic values. If speech recognition fails or produces a poor match within this grammar or section language model, speech recognition against a larger, more general vocabulary that is not constrained to the given section is performed.
    Type: Grant
    Filed: September 12, 2012
    Date of Patent: April 7, 2015
    Assignee: Nuance Communications, Inc.
    Inventors: Alwin B. Carus, Larissa Lapshina, Raghu Vemula
  • Patent number: 8996368
    Abstract: A feature transform for speech recognition is described. An input speech utterance is processed to produce a sequence of representative speech vectors. A time-synchronous speech recognition pass is performed using a decoding search to determine a recognition output corresponding to the speech input. The decoding search includes, for each speech vector after some first threshold number of speech vectors, estimating a feature transform based on the preceding speech vectors in the utterance and partial decoding results of the decoding search. The current speech vector is then adjusted based on the current feature transform, and the adjusted speech vector is used in a current frame of the decoding search.
    Type: Grant
    Filed: February 22, 2010
    Date of Patent: March 31, 2015
    Assignee: Nuance Communications, Inc.
    Inventor: Daniel Willett
  • Patent number: 8990070
    Abstract: A method, system and computer program product for building an expression, including utilizing any formal grammar of a context-free language, displaying an expression on a computer display via a graphical user interface, replacing at least one non-terminal display object within the displayed expression with any of at least one non-terminal display object and at least one terminal display object, and repeating the replacing step a plurality of times for a plurality of non-terminal display objects until no non-terminal display objects remain in the displayed expression, wherein the non-terminal display objects correspond to non-terminal elements within the grammar, and wherein the terminal display objects correspond to terminal elements within the grammar.
    Type: Grant
    Filed: November 18, 2011
    Date of Patent: March 24, 2015
    Assignee: International Business Machines Corporation
    Inventors: Yigal S. Dayan, Gil Fuchs, Josemina M. Magdalen
  • Patent number: 8990071
    Abstract: A method for managing an interaction of a calling party to a communication partner is provided. The method includes automatically determining if the communication partner expects DTMF input. The method also includes translating speech input to one or more DTMF tones and communicating the one or more DTMF tones to the communication partner, if the communication partner expects DTMF input.
    Type: Grant
    Filed: March 29, 2010
    Date of Patent: March 24, 2015
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Yun-Cheng Ju, Stefanie Tomko, Frank Liu, Ivan Tashev
  • Patent number: 8990086
    Abstract: A recognition confidence measurement method, medium and system which can more accurately determine whether an input speech signal is an in-vocabulary, by extracting an optimum number of candidates that match a phone string extracted from the input speech signal and estimating a lexical distance between the extracted candidates is provided. A recognition confidence measurement method includes: extracting a phoneme string from a feature vector of an input speech signal; extracting candidates by matching the extracted phoneme string and phoneme strings of vocabularies registered in a predetermined dictionary and; estimating a lexical distance between the extracted candidates; and determining whether the input speech signal is an in-vocabulary, based on the lexical distance.
    Type: Grant
    Filed: July 31, 2006
    Date of Patent: March 24, 2015
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Sang-Bae Jeong, Nam Hoon Kim, Ick Sang Han, In Jeong Choi, Gil Jin Jang, Jae-Hoon Jeong
  • Patent number: 8983841
    Abstract: A network communication node includes an audio outputter that outputs an audible representation of data to be provided to a requester. The network communication node also includes a processor that determines a categorization of the data to be provided to the requester and that varies a pause between segments of the audible representation of the data in accordance with the categorization of the data to be provided to the requester.
    Type: Grant
    Filed: July 15, 2008
    Date of Patent: March 17, 2015
    Assignee: AT&T Intellectual Property, I, L.P.
    Inventors: Gregory Pulz, Steven Lewis, Charles Rajnai
  • Patent number: 8977549
    Abstract: A natural language business system and method is developed to understand the underlying meaning of a person's speech, such as during a transaction with the business system. The system includes a speech recognition engine, and action classification engine, and a control module. The control module causes the system to execute an inventive method wherein the speech recognition and action classification models may be recursively optimized on an unisolated performance metric that is pertinent to the overall performance of the natural language business system, as opposed to the isolated model-specific criteria previously employed.
    Type: Grant
    Filed: September 26, 2013
    Date of Patent: March 10, 2015
    Assignee: Nuance Communications, Inc.
    Inventors: Sabine V. Deligne, Yuqing Gao, Vaibhava Goel, Hong-Kwang Kuo, Cheng Wu
  • Patent number: 8977547
    Abstract: A voice recognition system includes: a voice input unit 11 for inputting a voice uttered a plurality of times; a registering voice data storage unit 12 for storing voice data uttered the plurality of times and input into the voice input unit 11; an utterance stability verification unit 13 for determining a similarity between the voice data uttered the plurality of times that are read from the registering voice data storage unit 12, and determining that registration of the voice data is acceptable when the similarity is greater than a threshold Tl; and a standard pattern creation unit 14 for creating a standard pattern by using the voice data where the utterance stability verification unit 13 determines that registration is acceptable.
    Type: Grant
    Filed: October 8, 2009
    Date of Patent: March 10, 2015
    Assignee: Mitsubishi Electric Corporation
    Inventors: Michihiro Yamazaki, Jun Ishii, Hiroki Sakashita, Kazuyuki Nogi
  • Patent number: 8938382
    Abstract: An item of information (212) is transmitted to a distal computer (220), translated to a different sense modality and/or language (222), and in substantially real time, and the translation (222) is transmitted back to the location (211) from which the item was sent. The device sending the item is preferably a wireless device, and more preferably a cellular or other telephone (210). The device receiving the translation is also preferably a wireless device, and more preferably a cellular or other telephone, and may advantageously be the same device as the sending device. The item of information (212) preferably comprises a sentence of human of speech having at least ten words, and the translation is a written expression of the sentence. All of the steps of transmitting the item of information, executing the program code, and transmitting the translated information preferably occurs in less than 60 seconds of elapsed time.
    Type: Grant
    Filed: March 21, 2012
    Date of Patent: January 20, 2015
    Assignee: Ulloa Research Limited Liability Company
    Inventor: Robert D. Fish
  • Patent number: 8938388
    Abstract: Maintaining and supplying a plurality of speech models is provided. A plurality of speech models and metadata for each speech model are stored. A query for a speech model is received from a source. The query includes one or more conditions. The speech model with metadata most closely matching the supplied one or more conditions is determined. The determined speech model is provided to the source. A refined speech model is received from the source, and the refined speech model is stored.
    Type: Grant
    Filed: July 9, 2012
    Date of Patent: January 20, 2015
    Assignee: International Business Machines Corporation
    Inventors: Bin Jia, Ying Liu, E. Feng Lu, Jia Wu, Zhen Zhang
  • Patent number: 8930191
    Abstract: Methods, systems, and computer readable storage medium related to operating an intelligent digital assistant are disclosed. A user request is received, the user request including at least a speech input received from a user. In response to the user request, (1) an echo of the speech input based on a textual interpretation of the speech input, and (2) a paraphrase of the user request based at least in part on a respective semantic interpretation of the speech input are presented to the user.
    Type: Grant
    Filed: March 4, 2013
    Date of Patent: January 6, 2015
    Assignee: Apple Inc.
    Inventors: Thomas Robert Gruber, Harry Joseph Saddler, Adam John Cheyer, Dag Kittlaus, Christopher Dean Brigham, Richard Donald Giuli, Didier Rene Guzzoni, Marcello Bastea-Forte
  • Patent number: 8930187
    Abstract: An apparatus for utilizing textual data and acoustic data corresponding to speech data to detect sentiment may include a processor and memory storing executable computer code causing the apparatus to at least perform operations including evaluating textual data and acoustic data corresponding to voice data associated with captured speech content. The computer program code may further cause the apparatus to analyze the textual data and the acoustic data to detect whether the textual data or the acoustic data includes one or more words indicating at least one sentiment of a user that spoke the speech content. The computer program code may further cause the apparatus to assign at least one predefined sentiment to at least one of the words in response to detecting that the word(s) indicates the sentiment of the user. Corresponding methods and computer program products are also provided.
    Type: Grant
    Filed: January 3, 2012
    Date of Patent: January 6, 2015
    Assignee: Nokia Corporation
    Inventors: Imre Attila Kiss, Joseph Polifroni, Francois Mairesse, Mark Adler
  • Patent number: 8924197
    Abstract: Disclosed are systems, methods, and computer readable media for converting a natural language query into a logical query. The method embodiment comprises receiving a natural language query and converting the natural language query using an extensible engine to generate a logical query, the extensible engine being linked to the toolkit and knowledge base. In one embodiment, a natural language query can be processed in a domain independent method to generate a logical query.
    Type: Grant
    Filed: October 30, 2007
    Date of Patent: December 30, 2014
    Assignee: Semantifi, Inc.
    Inventors: Sreenivasa Rao Pragada, Viswanath Dasari, Abhijit A Patil
  • Patent number: 8924214
    Abstract: A method for detecting and recognizing speech is provided that remotely detects body motions from a speaker during vocalization with one or more radar sensors. Specifically, the radar sensors include a transmit aperture that transmits one or more waveforms towards the speaker, and each of the waveforms has a distinct wavelength. A receiver aperture is configured to receive the scattered radio frequency energy from the speaker. Doppler signals correlated with the speaker vocalization are extracted with a receiver. Digital signal processors are configured to develop feature vectors utilizing the vocalization Doppler signals, and words associated with the feature vectors are recognized with a word classifier.
    Type: Grant
    Filed: June 7, 2011
    Date of Patent: December 30, 2014
    Assignee: The United States of America, as represented by the Secretary of the Navy
    Inventors: Jefferson M Willey, Todd Stephenson, Hugh Faust, James P. Hansen, George J Linde, Carol Chang, Justin Nevitt, James A Ballas, Thomas Herne Crystal, Vincent Michael Stanford, Jean W. De Graaf
  • Patent number: 8918320
    Abstract: An apparatus for generating a review based in part on detected sentiment may include a processor and memory storing executable computer code causing the apparatus to at least perform operations including determining a location(s) of the apparatus and a time(s) that the location(s) was determined responsive to capturing voice data of speech content associated with spoken reviews of entities. The computer program code may further cause the apparatus to analyze textual and acoustic data corresponding to the voice data to detect whether the textual or acoustic data includes words indicating a sentiment(s) of a user speaking the speech content. The computer program code may further cause the apparatus to generate a review of an entity corresponding to a spoken review(s) based on assigning a predefined sentiment to a word(s) responsive to detecting that the word indicates the sentiment of the user. Corresponding methods and computer program products are also provided.
    Type: Grant
    Filed: January 3, 2012
    Date of Patent: December 23, 2014
    Assignee: Nokia Corporation
    Inventors: Mark Adler, Imre Attila Kiss, Francois Mairesse, Joseph Polifroni
  • Patent number: 8918316
    Abstract: The content of a media program is recognized by analyzing its audio content to extract therefrom prescribed features, which are compared to a database of features associated with identified content. The identity of the content within the database that has features that most closely match the features of the media program being played is supplied as the identity of the program being played. The features are extracted from a frequency domain version of the media program by a) filtering the coefficients to reduce their number, e.g., using triangular filters; b) grouping a number of consecutive outputs of triangular filters into segments; and c) selecting those segments that meet prescribed criteria, such as those segments that have the largest minimum segment energy with prescribed constraints that prevent the segments from being too close to each other. The triangular filters may be log-spaced and their output may be normalized.
    Type: Grant
    Filed: July 29, 2003
    Date of Patent: December 23, 2014
    Assignee: Alcatel Lucent
    Inventors: Jan I Ben, Christopher J Burges, Madjid Sam Mousavi, Craig R. Nohl