Similarity Patents (Class 704/239)
  • Patent number: 11138978
    Abstract: A method and system of automatically identifying topics of a conversation are provided. An electronic data package comprising a sequence of utterances between conversation entities is received by a computing device. Each utterance is classified to a corresponding social action. One or more utterances in the sequence are grouped into a segment based on a deep learning model. A similarity of topics between adjacent segments is determined. Upon determining that the similarity is above a predetermined threshold, the adjacent segments are grouped together. A transcript of the conversation including the grouping of the adjacent segments is stored in a memory.
    Type: Grant
    Filed: July 24, 2019
    Date of Patent: October 5, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Margaret Helen Szymanski, Lei Huang, Robert John Moore, Raphael Arar, Shun Jiang, Guangjie Ren, Eric Liu, Pawan Chowdhary, Chung-hao Tan, Sunhwan Lee
  • Patent number: 11043214
    Abstract: Described herein is a system for rescoring automatic speech recognition hypotheses for conversational devices that have multi-turn dialogs with a user. The system leverages dialog context by incorporating data related to past user utterances and data related to the system generated response corresponding to the past user utterance. Incorporation of this data improves recognition of a particular user utterance within the dialog.
    Type: Grant
    Filed: November 29, 2018
    Date of Patent: June 22, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Behnam Hedayatnia, Anirudh Raju, Ankur Gandhe, Chandra Prakash Khatri, Ariya Rastrow, Anushree Venkatesh, Arindam Mandal, Raefer Christopher Gabriel, Ahmad Shikib Mehri
  • Patent number: 11024297
    Abstract: A method for using speech disfluencies detected in speech input to assist in interpreting the input is provided. The method includes providing access to a set of content items with metadata describing the content items, and receiving a speech input intended to identify a desired content item. The method further includes detecting a speech disfluency in the speech input and determining a measure of confidence of a user in a portion of the speech input following the speech disfluency. If the confidence measure is lower than a threshold value, the method includes determining an alternative query input based on replacing the portion of the speech input following the speech disfluency with another word or phrase. The method further includes selecting content items based on comparing the speech input, the alternative query input (when the confidence measure is low), and the metadata associated with the content items.
    Type: Grant
    Filed: October 25, 2018
    Date of Patent: June 1, 2021
    Assignee: Veveo, Inc.
    Inventors: Murali Aravamudan, Daren Gill, Sashikumar Venkataraman, Vineet Agarwal, Ganesh Ramamoorthy
  • Patent number: 10991365
    Abstract: A method of enhancing an automated speech recognition confidence classifier includes receiving a set of baseline confidence features from one or more decoded words, deriving word embedding confidence features from the baseline confidence features, joining the baseline confidence features with word embedding confidence features to create a feature vector, and executing the confidence classifier to generate a confidence score, wherein the confidence classifier is trained with a set of training examples having labeled features corresponding to the feature vector.
    Type: Grant
    Filed: April 8, 2019
    Date of Patent: April 27, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Kshitiz Kumar, Anastasios Anastasakos, Yifan Gong
  • Patent number: 10978058
    Abstract: Disclosed are an electronic apparatus, a control method thereof, and a computer program product for the same, the electronic apparatus including: a receiver comprising receiving circuitry configured to receive a sound; and a processor configured to: identify with a given sensitivity whether a characteristic of a received sound corresponds to a voice command of a user in response to the sound being received through the receiver, identify the voice command based on identifying that the characteristic of the received sound corresponds to the voice command, and perform an operation corresponding to the identified voice command, and change the sensitivity based on identifying that the characteristic of the received sound does not correspond to the voice command. Thus, the electronic apparatus performs the optimum and/or improved audio process by properly controlling the sensitivity based on the circumstances.
    Type: Grant
    Filed: December 3, 2018
    Date of Patent: April 13, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Jonguk Yoo, Kihoon Shin
  • Patent number: 10964316
    Abstract: One non-limiting embodiment provides a method, including: receiving, from a user, user input comprising a trigger event; identifying, using at least one processor, active media content; and performing, based upon the trigger event, an action with respect to the active media content. This embodiment is intended to be non-limiting and other embodiments are contemplated, disclosed, and discussed.
    Type: Grant
    Filed: August 9, 2017
    Date of Patent: March 30, 2021
    Assignee: Lenovo (Singapore) Pte. Ltd.
    Inventors: Roderick Echols, Ryan Charles Knudson, Timothy Winthrop Kingsbury, Jonathan Gaither Knox
  • Patent number: 10943580
    Abstract: Methods and systems for phonological clustering are disclosed. A method includes: segmenting, by a computing device, a sentence into a plurality of tokens; determining, by the computing device, a plurality of phoneme variants corresponding to the plurality of tokens; clustering, by the computing device, the plurality of phoneme variants; creating, by the computing device, an initial vectorization of the plurality of phoneme variants based on the clustering; embedding, by the computing device, the initial vectorization of the plurality of phoneme variants into a deep learning model; and determining, by the computing device, a radial set of phoneme variants using the deep learning model.
    Type: Grant
    Filed: May 11, 2018
    Date of Patent: March 9, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Craig M. Trim, John M. Ganci, Jr., James E. Bostick, Carlos A. Fonseca
  • Patent number: 10929448
    Abstract: A computer-implemented method determines a category of a request provided by a user by means of a user device. The user device includes connection means and means for receiving a request description relating to said request from said user. The method includes receiving, from the user, the request description, by means of the device, and uploading the request description to a server. The server has access to a database which includes a number of previously categorized requests each including a category and a vocabulary, which includes a number of word vector representations. The method further includes identifying, by the server, a number of component words belonging to a natural language text string included in the request description; obtaining, for at least one of the component words, an associated word vector representation from the vocabulary, and determining a request vector, based on at least one obtained word vector representation.
    Type: Grant
    Filed: August 10, 2018
    Date of Patent: February 23, 2021
    Assignee: KBC GROEP NV
    Inventors: Hans Verstraete, Hans Verstraete, Pieter Van Hertum, Rahul Maheshwari, Jeroen D'Haen, Michaël Mariën, Barak Chizi, Frank Fripon, Sven Evens
  • Patent number: 10923111
    Abstract: A system configured to recognize text represented by speech may determine that a first portion of audio data corresponds to speech from a first speaker and that a second portion of audio data corresponds to speech from the first speaker and a second speaker. Features of the first portion are compared to features of the second portion to determine a similarity therebetween. Based on this similarity, speech from the first speaker is distinguished from speech from the second speaker and text corresponding to speech from the first speaker is determined.
    Type: Grant
    Filed: March 28, 2019
    Date of Patent: February 16, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Xing Fan, I-Fan Chen, Yuzong Liu, Bjorn Hoffmeister, Yiming Wang, Tongfei Chen
  • Patent number: 10915435
    Abstract: Methods and systems for a deep learning based problem advisor are disclosed. A method includes: obtaining, by a computing device, a log file including events generated during execution of a software application; determining, by the computing device, at least one possible cause for a problem in the software application using the obtained log file and a knowledge base including calling paths for each of a plurality of methods in source code of the software application; for each of the at least one possible cause for the problem, the computing device simulating user actions in the software application; and determining, by the computing device, a root cause based on the simulating user actions in the software application.
    Type: Grant
    Filed: November 28, 2018
    Date of Patent: February 9, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Jian Zhang, Yi Bin Wang, Wu Weilin, Mu Dan Cao, Dan Tan
  • Patent number: 10896671
    Abstract: A command-processing server provides natural language services to applications. More specifically, the command-processing server receives natural language inputs from users for use in applications such as virtual assistants. Some user inputs create user-defined rules that consist of trigger conditions and of corresponding actions that are executed when the triggers fire. The command-processing server stores the rules received from a user in association with the specific user. The command-processing server also identifies rules that can be generalized across users and promoted into generic rules applicable to many or all users. The generic rules may or may not have an associated context constraining their application.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: January 19, 2021
    Assignee: SoundHound, Inc.
    Inventors: Keyvan Mohajer, Christopher S. Wilson, Bernard Mont-Reynaud, Robert MacRae
  • Patent number: 10818193
    Abstract: A communications training system is provided having a user interface, a computer-based simulator and a performance measurement database. The user interface is configured to receive a speech communication input from the user based on a training content and the computer-based simulator is configured to transform the speech communication to a text data whereby the text data can be aligned to performance measurement database values to determine a performance measure of the speech communication. The format of the text data and the performance measurement database values enable the speech communication to be aligned with predefined performance measurement database values representing expected speech communications for that training content.
    Type: Grant
    Filed: February 20, 2017
    Date of Patent: October 27, 2020
    Assignee: Aptima, Inc.
    Inventors: Kevin Sullivan, Matthew Roberts, Michael Knapp, Brian Riordan
  • Patent number: 10657327
    Abstract: Mechanisms are provided for clarifying homophone usage in natural language content. The mechanisms analyze natural language content to identify a homophone instance in the natural language content, the homophone instance being a first term having a first definition and a first pronunciation for which there is a second term having the first pronunciation and a second definition different from the first definition. The mechanisms, in response to identifying the homophone instance, analyze the natural language content to identify a third term that is a synonym for the second term. The third term has a third definition that is nearly the same as the second definition. The mechanisms, in response to the natural language content comprising the third term, perform a clarifying operation to modify the natural language content to clarify the homophone instance and generate a modified natural language content.
    Type: Grant
    Filed: August 1, 2017
    Date of Patent: May 19, 2020
    Assignee: International Business Machines Corporation
    Inventors: Kelley L. Anders, Paul R. Bastide, Stacy M. Cannon, Trudy L. Hewitt
  • Patent number: 10628567
    Abstract: Methods, computing systems and computer program products implement embodiments of the present invention that include defining a verification string including a sequence of verification characters and a delimiter character between each sequential pair of the verification characters, the delimiter character being different from the verification characters. The verification string to a user, and upon receiving, from the user, a series of verification vocal inputs in response to presenting the verification string, a set of verification features from each of the verification vocal inputs are computed so as to generate sets of verification features. A one-to-one correspondence is established between each of the verification vocal inputs and each of the verification characters, and the user is authenticated based on the verification vocal inputs and their corresponding sets of verification features.
    Type: Grant
    Filed: September 5, 2016
    Date of Patent: April 21, 2020
    Assignee: International Business Machines Corporation
    Inventor: Hagai Aronowitz
  • Patent number: 10559305
    Abstract: [Object] To provide an information processing system and an information processing method capable of auditing the utterance data of an agent more flexibly. [Solution] An information processing system including: a storage section that stores utterance data of an agent; a communication section that receives request information transmitted from a client terminal and requesting utterance data of a specific agent from a user; and a control section that, when the request information is received through the communication section, replies to the client terminal with corresponding utterance data, and in accordance with feedback from the user with respect to the utterance data, updates an utterance probability level expressing a probability that the specific agent will utter utterance content indicated by the utterance data, and records the updated utterance probability level in association with the specific agent and the utterance content in the storage section.
    Type: Grant
    Filed: February 2, 2017
    Date of Patent: February 11, 2020
    Assignee: SONY CORPORATION
    Inventor: Akihiro Komori
  • Patent number: 10540963
    Abstract: A computer-implemented method for generating an input for a classifier. The method includes obtaining n-best hypotheses which is an output of an automatic speech recognition (ASR) for an utterance, combining the n-best hypotheses horizontally in a predetermined order with a separator between each pair of hypotheses, and outputting the combined n-best hypotheses as a single text input to a classifier.
    Type: Grant
    Filed: February 2, 2017
    Date of Patent: January 21, 2020
    Assignee: International Business Machines Corporation
    Inventors: Nobuyasu Itoh, Gakuto Kurata, Ryuki Tachibana
  • Patent number: 10387805
    Abstract: The present invention provides a method for ranking an incoming news feed comprising a header. The method comprising the steps of: receiving the incoming news feed with headers, extracting the incoming news feed's header, performing part-of-speech-tagging of the extracted header's words and associating to each of the header's words a code characterizing its grammatical function, generating the list of the incoming header's words codes, associating the generated list to the incoming news feed, as its pattern and computing a score for the incoming news feed according to predefined rules defining the score based on the its pattern.
    Type: Grant
    Filed: July 16, 2014
    Date of Patent: August 20, 2019
    Assignee: DEEP IT LTD
    Inventors: Eliezer Katz, Ofer Weintraub
  • Patent number: 10212181
    Abstract: A method comprises creating a word vector from a message, wherein the word vector comprises an entry for each word of a plurality of words, and wherein each word of the plurality of words is assigned a weight. The method further comprises calculating a value for the word vector based on each entry of the word vector and the weights assigned to the plurality of words, and identifying that the message belongs to a first group by comparing the value for the word vector to a threshold. The word vector comprises an entry for each word of a plurality of words, and wherein each word of the plurality of words is assigned a weight.
    Type: Grant
    Filed: November 18, 2016
    Date of Patent: February 19, 2019
    Assignee: Bank of America Corporation
    Inventors: Pinak Chakraborty, Vidhu Beohar, Chetan Phanse
  • Patent number: 10102732
    Abstract: A danger monitoring system is disclosed. A danger monitoring device comprises a microphone configured to continuously digitize environmental sound, a first memory, a first processor configured to determine whether a stored interval meets a threshold criteria for a dangerous event, and a first network interface configured to send a danger observation data to a server. The danger monitoring server comprises a second memory, a second processor configured to verify the dangerous event digitized by the danger monitoring device and determine an event location of the verified dangerous event, and a second network interface configured to send a danger alert. A danger mitigation device comprises a third network interface configured to receive the danger alert, a GPS receiver, a screen, a third memory comprising map data, and a third processor configured to render a map indicating at least a current location of the danger mitigation device, and the event location.
    Type: Grant
    Filed: October 27, 2017
    Date of Patent: October 16, 2018
    Assignee: INFINITE DESIGNS, LLC
    Inventor: Adam Gersten
  • Patent number: 10089061
    Abstract: According to one embodiment, an electronic device includes a memory and a hardware processor. The hardware processor is in communication with the memory. The hardware processor is configured to obtain a sound file including sound data and attached data, determine a type of meeting of the sound file classified based on an utterance state of the sound data, and display the sound file based on at least one of the sound data and the attached data such that the type of meeting is visually recognizable.
    Type: Grant
    Filed: February 16, 2016
    Date of Patent: October 2, 2018
    Assignee: KABUSHIKI KAISHA TOSHIBA
    Inventor: Yusaku Kikugawa
  • Patent number: 10062385
    Abstract: A system and method for selecting a speech-to-text engine are disclosed. The method includes selecting, by an engine selection component, at least two speech-to-text engines to decode a portion of computer-readable speech data. The portion of speech data can be decoded simultaneously by the selected speech-to-text engines for a designated length of time. In some embodiments portions of the speech data can be simultaneously decoded with selected speech-to-text engines at periodic intervals. An accuracy of decoding can be determined for each selected speech-to-text engine by an accuracy testing component. Additionally, the relative accuracies and speeds of the selected speech-to-text engines can be compared by an output comparison component. The engine selection component can then select the most accurate speech-to-text engine accurate to decode a next portion of speech data. Further, the engine selection module may select a speech-to-text engine that meets or exceeds a speed and/or accuracy threshold.
    Type: Grant
    Filed: September 30, 2016
    Date of Patent: August 28, 2018
    Assignee: International Business Machines Corporation
    Inventors: Alexander Cook, Manuel Orozco, Christopher R. Sabotta, John M. Santosuosso
  • Patent number: 10049672
    Abstract: The subject matter of this specification can be embodied in, among other things, a method that includes receiving an audio signal and initiating speech recognition tasks by a plurality of speech recognition systems (SRS's). Each SRS is configured to generate a recognition result specifying possible speech included in the audio signal and a confidence value indicating a confidence in a correctness of the speech result. The method also includes completing a portion of the speech recognition tasks including generating one or more recognition results and one or more confidence values for the one or more recognition results, determining whether the one or more confidence values meets a confidence threshold, aborting a remaining portion of the speech recognition tasks for SRS's that have not generated a recognition result, and outputting a final recognition result based on at least one of the generated one or more speech results.
    Type: Grant
    Filed: June 2, 2016
    Date of Patent: August 14, 2018
    Assignee: Google LLC
    Inventors: Brian Patrick Strope, Francoise Beaufays, Olivier Siohan
  • Patent number: 9984689
    Abstract: Disclosed is an apparatus and method for correcting pronunciation by contextual recognition. The apparatus may include an interface configured to receive, from a speech recognition server, first text data obtained by converting speech data to a text, and a processor configured to extract a keyword from the received first text data, calculate a suitability of a word in the first text data in association with the extracted keyword, and update the first text data to second text data by replacing, with an alternative word, a word in the first text data having a suitability less than a preset reference value.
    Type: Grant
    Filed: November 10, 2016
    Date of Patent: May 29, 2018
    Inventor: Sung Hyuk Kim
  • Patent number: 9953640
    Abstract: Method and systems are provided for interpreting speech data. A method and system for recognizing speech involving a filter module to generate a set of processed audio data based on raw audio data; a translation module to provide a set of translation results for the raw audio data; and a decision module to select the text data that represents the raw audio data. A method for minimizing noise in audio signals received by a microphone array is also described. A method and system of automatic entry of data into one or more data fields involving receiving a processed audio data; and operating a processing module to: search in a trigger dictionary for a field identifier that corresponds to the trigger identifier; identify a data field associated with a data field identifier corresponding to the field identifier; and providing content data associated with the trigger identifier to the identified data field.
    Type: Grant
    Filed: June 5, 2015
    Date of Patent: April 24, 2018
    Assignee: INTERDEV TECHNOLOGIES INC.
    Inventors: Janet M. Rice, Peng Liang, Terence W. Kuehn
  • Patent number: 9940318
    Abstract: Methods, apparatus, systems, and computer-readable media are provided for generating and applying outgoing communication templates. In various implementations a corpus of outgoing communications sent by a user may be grouped into a plurality of clusters based on one or more attributes of a context of the user. One or more segments of each outgoing communication of a particular cluster may be classified as fixed in response to a determination that a count of occurrences of the one or more segments across the particular cluster satisfies a criterion. One or more remaining segments of each communication of the particular cluster may or may not be classified as transient. Based on sequences of classified segments associated with each communication of the particular cluster, an outgoing communication template may be generated to automatically populate at least a portion of a draft outgoing communication being prepared by the user.
    Type: Grant
    Filed: January 1, 2016
    Date of Patent: April 10, 2018
    Assignee: GOOGLE LLC
    Inventors: Balint Miklos, Julia Proskurnia, Luis Garcia Pueyo, Marc-Allen Cartright, Tobias Kaufmann, Ivo Krka
  • Patent number: 9928851
    Abstract: A voice verifying system, which comprises: a microphone, which is always turned on to output at least one input audio signal; a speech determining device, for determining if the input audio signal is valid or not according to a reference value, wherein the speech determining device passes the input audio signal if the input audio signal is valid; and a verifying module, for verifying a speech signal generated from the input audio signal and for outputting a device activating signal to activate a target device if the speech signal matches a predetermined rule; and a reference value generating device, for generating the reference value according to speech signal information from the verifying module.
    Type: Grant
    Filed: September 12, 2013
    Date of Patent: March 27, 2018
    Assignee: MEDIATEK INC.
    Inventors: Liang-Che Sun, Yiou-Wen Cheng, Ting-Yuan Chiu
  • Patent number: 9830039
    Abstract: A wizard control panel may be used by a human wizard to adjust the operation of a Natural Language (NL) conversational system during a real-time dialog flow. Input to the wizard control panel is detected and used to interrupt/change an automatic operation of one or more of the NL conversational system components used during the flow. For example, the wizard control panel may be used to adjust results determined by an Automated Speech Recognition (ASR) component, a Natural Language Understanding (NLU) component, a Dialog Manager (DM) component, and a Natural Language Generation (NLG) before the results are used to perform an automatic operation within the flow. A timeout may also be set such that when the timeout expires, the conversational system performs an automated operation by using the results shown in the wizard control panel (edited/not edited).
    Type: Grant
    Filed: March 4, 2013
    Date of Patent: November 28, 2017
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Lisa Stifelman, Dilek Hakkani-Tur, Larry Heck, Madhusudan Chinthakunta
  • Patent number: 9799333
    Abstract: A system and method are provided for performing speech processing. A system includes an audio detection system configured to receive a signal including speech and a memory having stored therein a database of keyword models forming an ensemble of filters associated with each keyword in the database. A processor is configured to receive the signal including speech from the audio detection system, decompose the signal including speech into a sparse set of phonetic impulses, and access the database of keywords and convolve the sparse set of phonetic impulses with the ensemble of filters. The processor is further configured to identify keywords within the signal including speech based a result of the convolution and control operation the electronic system based on the keywords identified.
    Type: Grant
    Filed: August 31, 2015
    Date of Patent: October 24, 2017
    Assignee: The Johns Hopkins University
    Inventors: Keith Kintzley, Aren Jansen, Hynek Hermansky, Kenneth Church
  • Patent number: 9684433
    Abstract: A method includes identifying individuals that are affiliated with a user. The method includes incorporating trusted devices associated with the identified individuals into an event monitor network that is configured to monitor for an occurrence of a monitored event. The method includes identifying a particular input that suggests the occurrence of the monitored event. The method includes communicating to the trusted devices, an input sample that is used for recognition of the particular input from general input that is measured by sensors of the trusted devices. The method includes receiving from at least one of the trusted devices, an event message that indicates the particular input is observed by at least one of the sensors. In response to the event message, the method includes communicating to a user interface of a user device associated with the user, an alarm message that indicates the occurrence of the monitored event.
    Type: Grant
    Filed: December 30, 2014
    Date of Patent: June 20, 2017
    Assignee: EBAY INC.
    Inventor: Kamal Zamer
  • Patent number: 9685153
    Abstract: In some embodiments, the recognition results produced by a speech processing system (which may include a top recognition result and one or more alternative recognition results) based on an analysis of a speech input, are evaluated for indications of potential significant errors. In some embodiments, the recognition results may be evaluated to determine whether a meaning of any of the alternative recognition results differs from a meaning of the top recognition result in a manner that is significant for the domain. In some embodiments, one or more of the recognition results may be evaluated to determine whether the result(s) include one or more words or phrases that, when included in a result, would change a meaning of the result in a manner that would be significant for the domain.
    Type: Grant
    Filed: May 15, 2015
    Date of Patent: June 20, 2017
    Assignee: Nuance Communications, Inc.
    Inventors: William F. Ganong, III, Raghu Vemula, Robert Fleming
  • Patent number: 9558335
    Abstract: A method includes receiving, from a user via an electronic device, input representing a password to be utilized for an account; automatically determining, utilizing a processor, a complexity value for the input password; automatically determining, based on the determined complexity value, security settings for the account; receiving, from a user via an electronic device, input representing an attempt to login to the account, the input representing an attempt to login to the account including an attempted password; automatically determining that the attempted password does not match the password to be utilized for the account; and determining a course of action to take in response to the determination that the attempted password does not match the password to be utilized for the account, the course of action being determined based at least in part on the automatically determined security settings for the account.
    Type: Grant
    Filed: December 14, 2015
    Date of Patent: January 31, 2017
    Assignee: ALLSCRIPTS SOFTWARE, LLC
    Inventors: David Thomas Windell, Todd Michael Eischeid, Scott David Bower
  • Patent number: 9476718
    Abstract: A vehicle navigation system may send and receive communications, such as text messages. Speech recognition may generate a text message without affecting a driver's control of the vehicle. A user may audibly control the navigation system and generate a text message through a speech recognition element. A microphone may record a user's voice, which is then transformed into a text message for transmission. The message may be recorded sentence-by-sentence, word-by-word, or letter-by-letter. The recorded text message may be visually or audibly presented to the user before transmission.
    Type: Grant
    Filed: July 10, 2007
    Date of Patent: October 25, 2016
    Assignee: HARMAN BECKER AUTOMOTIVE SYSTEMS GMBH
    Inventor: Mirko Herforth
  • Patent number: 9477753
    Abstract: Systems and methods for processing a query include determining a plurality of sets of match candidates for a query using a processor, each of the plurality of sets of match candidates being independently determined from a plurality of diverse word lattice generation components of different type. The plurality of sets of match candidates is merged by generating a first score for each match candidate to provide a merged set of match candidates. A second score is computed for each match candidate of the merged set based upon features of that match candidate. The first score and the second score are combined to provide a final set of match candidates as matches to the query.
    Type: Grant
    Filed: March 12, 2013
    Date of Patent: October 25, 2016
    Assignee: International Business Machines Corporation
    Inventors: Brian E. D. Kingsbury, Hong-Kwang Jeff Kuo, Lidia Luminita Mangu, Hagen Soltau
  • Patent number: 9449126
    Abstract: A user interface for presenting a set of related pages of an electronic content work for view at the same time. The pages are sized according to a target format for presentation of the electronic content work, and may also be formatted according to user-defined font and zoom criteria. Each of the related pages comprises a media object, for example a markup language object. Responsive to user manipulation of presentation criterion for the set of related pages, the set of related pages is reformatted and so presented in near real-time. In some instances, a user may manipulate controls of the user interface to isolate a content object included within the set of related pages, have information regarding that content object presented and even edit the content object in-line with the present view.
    Type: Grant
    Filed: June 1, 2012
    Date of Patent: September 20, 2016
    Assignee: Inkling Systems, Inc.
    Inventors: Thomas Charles Genoni, Peter S. Cho, Norris Hung, Eric Todd Lovett, Huan Zhao
  • Patent number: 9438578
    Abstract: A biometric authentication system is disclosed that provides authentication capability using biometric data in connection with a challenge for parties engaging in digital communications such as digital text-oriented, interactive digital communications. End-user systems may be coupled to devices that include biometric data capture devices such as retina scanners, fingerprint recorders, cameras, microphones, ear scanners, DNA profilers, etc., so that biometric data of a communicating party may be captured and used for authentication purposes.
    Type: Grant
    Filed: August 17, 2013
    Date of Patent: September 6, 2016
    Assignee: AT&T INTELLECTUAL PROPERTY II, L.P.
    Inventors: Pradeep K. Bansal, Lee Begeja, Carroll W. Creswell, Jeffrey Farah, Benjamin J. Stern, Jay Wilpon
  • Patent number: 9437187
    Abstract: A search string acquiring unit acquires a search string. A converting unit converts the search string into a phoneme sequence. A time length deriving unit derives the spoken time length of the voice corresponding to the search string. A zone designating unit designates a likelihood acquisition zone in a target voice signal. A likelihood acquiring device acquires a likelihood indicating how likely the likelihood acquisition interval is an interval in which voice corresponding to the search string is spoken. A repeating unit changes the likelihood acquisition zone designated by the zone designating unit, and repeats the process of the zone designating unit and the likelihood acquiring device. An identifying unit identifies, from the target voice signal, estimated intervals for which the voice corresponding to the search string is estimated to be spoken, on the basis of the likelihoods acquired for each of the likelihood acquisition zones.
    Type: Grant
    Filed: January 23, 2015
    Date of Patent: September 6, 2016
    Assignee: CASIO COMPUTER CO., LTD.
    Inventor: Hiroyasu Ide
  • Patent number: 9431007
    Abstract: In a voice search device, a processor acquires a search word, converts the search word into a phoneme sequence, acquires, for each frame, an output probability of a feature quantity of a target voice signal being output from each phoneme included in the phoneme sequence, and executes relative calculation of the output probability acquired from each phoneme, based on an output probability acquired from another phoneme included in the phoneme sequence. In addition, the processor successively designates likelihood acquisition zones, acquires a likelihood indicating how likely a designated likelihood acquisition zone is a zone in which voice corresponding to the search word is spoken, and identifies from the target voice signal an estimated zone for which the voice corresponding to the search word is estimated to be spoken, based on the acquired likelihood.
    Type: Grant
    Filed: January 15, 2015
    Date of Patent: August 30, 2016
    Assignee: CASIO COMPUTER CO., LTD.
    Inventor: Hiroki Tomita
  • Patent number: 9330656
    Abstract: A speech dialogue system generates a response sentence in a way to improve the efficiency of the dialogue with the user, based on a result of estimation on an attribute of a proper name in an utterance of a user. The system includes a database attribute estimation unit to estimate the attribute of the input proper name by utilizing a database, and a web attribute estimation unit to estimate an attribute of an input proper name by utilizing information on the web. A reliability integration unit calculates integrated reliability of estimation for each of possible attributes obtained from the estimation by the units, by integrating first reliability of the estimation. A response generation unit generates a response sentence to an input utterance based on the integrated reliabilities of the possible attributes.
    Type: Grant
    Filed: February 26, 2014
    Date of Patent: May 3, 2016
    Assignee: HONDA MOTOR CO., LTD.
    Inventors: Mikio Nakano, Kazunori Komatani, Tsugumi Otsuka
  • Patent number: 9330358
    Abstract: A system and method include comparing a context to cases stored in a case base, where the cases include Boolean and non-Boolean independent weight variables and a domain-specific dependency variable. The case and context independent weight variables are normalized and a normalized weight vector is determined for the case base. A match between the received context and each case of the case base is determined using the normalized context and case variables and the normalized weight vector. A skew value is determined for each category of domain specific dependency variables and the category of domain specific dependency variables having the minimal skew value is selected. The dependency variable associated with the selected category is then displayed to a user.
    Type: Grant
    Filed: September 26, 2013
    Date of Patent: May 3, 2016
    Assignee: THE UNITED STATES OF AMERICA AS REPRESENTED BY THE SECRETARY OF THE NAVY
    Inventor: Stuart H. Rubin
  • Patent number: 9311932
    Abstract: A method, system, and computer program product for adaptive pause detection in speech recognition are provided in the illustrative embodiments. A speech stream comprising audio signal of a speech is received. A first point in the speech stream is marked with a beginning time stamp. After the first point, a pause is detected in the speech stream. The pause is of a duration at least equal to a pause duration threshold. A second point after the pause in the speech stream is marked with an ending time stamp. A portion of the speech stream between the beginning and the ending time stamps forms a first speech segment. A speech rate of the first speech segment is computed using a number of words in the first speech segment, the beginning time stamp, and the ending time stamp. The pause duration is adjusted according to the first speech segment's speech rate.
    Type: Grant
    Filed: January 23, 2014
    Date of Patent: April 12, 2016
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventor: William S. Carter
  • Patent number: 9262693
    Abstract: An object detection apparatus includes a storage section storing a plurality of selection patterns as combinations of one of a plurality of recognition dictionaries and one of a plurality of image recognition algorithms, a specifying means for specifying at least one of a distance from a position at which an input image is taken and a target corresponding to the detection object within the input image and a state of light of the input image, a selection means for selecting one from the plurality of the selection patterns based on at least one of the distance and the state of the light specified by the specifying means, and a detection means for detecting the detection object within the input image by performing an image recognition process using the image recognition dictionary and the image recognition algorithm included in the selection pattern selected by the selection means.
    Type: Grant
    Filed: April 16, 2014
    Date of Patent: February 16, 2016
    Assignee: DENSO CORPORATION
    Inventor: Yasunori Kamiya
  • Patent number: 9197416
    Abstract: In a verification apparatus, a biometric information acquisition unit acquires a plurality of biometric information pieces from an object. A first verification unit calculates, as a verification score, the similarity between the biometric information piece and a verification information piece, and compares the calculated verification score with a first determination value to determine whether the biometric information piece matches the verification information piece. When the verification fails, a second verification unit performs verification on the plurality of biometric information pieces having a predetermined relationship, using the verification information piece and a second determination value which defines a less stringent criterion than the first determination value.
    Type: Grant
    Filed: August 8, 2013
    Date of Patent: November 24, 2015
    Assignee: FUJITSU FRONTECH LIMITED
    Inventor: Shinichi Eguchi
  • Patent number: 9147395
    Abstract: The present disclosure relates to a mobile terminal and a voice recognition method thereof. The voice recognition method may include receiving a user's voice; providing the received voice to a first voice recognition engine provided in the server and a second voice recognition engine provided in the mobile terminal; acquiring first voice recognition data as a result of recognizing the received voice by the first voice recognition engine; acquiring second voice recognition data as a result of recognizing the received voice by the second voice recognition engine; estimating a function corresponding to the user's intention based on at least one of the first and the second voice recognition data; calculating a similarity between the first and the second voice recognition data when personal information is required for the estimated function; and selecting either one of the first and the second voice recognition data based on the calculated similarity.
    Type: Grant
    Filed: June 21, 2013
    Date of Patent: September 29, 2015
    Assignee: LG ELECTRONICS INC.
    Inventors: Juhee Kim, Hyunseob Lee, Joonyup Lee, Jungkyu Choi
  • Patent number: 9142212
    Abstract: An apparatus and method for recognizing a voice command for use in an interactive voice user interface are provided. The apparatus includes a command intention belief generation unit that is configured to recognize a first voice command and that may generate one or more command intention beliefs for the first voice command. The apparatus also includes a command intention belief update unit that is configured to update each of the command intention beliefs based on a system response to the first voice command and a second voice commands. The apparatus also includes a command intention belief selection unit that is configured to select one of the updated command intention beliefs for the first voice command. The apparatus also includes an operation signal output unit that is configured to select a final command intention from the selected updated command intention belief and to output an operation signal based on the selected final command intention.
    Type: Grant
    Filed: April 26, 2011
    Date of Patent: September 22, 2015
    Inventors: Chi-Youn Park, Byung-Kwan Kwak, Jeong-Su Kim, Jeong-Mi Cho
  • Patent number: 9064493
    Abstract: In some embodiments, the recognition results produced by a speech processing system (which may include a top recognition result and one or more alternative recognition results) based on an analysis of a speech input, are evaluated for indications of potential significant errors. In some embodiments, the recognition results may be evaluated to determine whether a meaning of any of the alternative recognition results differs from a meaning of the top recognition result in a manner that is significant for the domain. In some embodiments, one or more of the recognition results may be evaluated to determine whether the result(s) include one or more words or phrases that, when included in a result, would change a meaning of the result in a manner that would be significant for the domain.
    Type: Grant
    Filed: July 9, 2012
    Date of Patent: June 23, 2015
    Assignee: Nuance Communications, Inc.
    Inventors: William F. Ganong, III, Raghu Vemula, Robert Fleming
  • Patent number: 9043207
    Abstract: The present invention relates to a method for speaker recognition, comprising the steps of obtaining and storing speaker information for at least one target speaker; obtaining a plurality of speech samples from a plurality of telephone calls from at least one unknown speaker; classifying the speech samples according to the at least one unknown speaker thereby providing speaker-dependent classes of speech samples; extracting speaker information for the speech samples of each of the speaker-dependent classes of speech samples; combining the extracted speaker information for each of the speaker-dependent classes of speech samples; comparing the combined extracted speaker information for each of the speaker-dependent classes of speech samples with the stored speaker information for the at least one target speaker to obtain at least one comparison result; and determining whether one of the at least one unknown speakers is identical with the at least one target speaker based on the at least one comparison result.
    Type: Grant
    Filed: November 12, 2009
    Date of Patent: May 26, 2015
    Assignee: Agnitio S.L.
    Inventors: Johan Nikolaas Langehoven Brummer, Luis Buera Rodriguez, Marta Garcia Gomar
  • Publication number: 20150127342
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for performing speaker identification. In some implementations, an utterance vector that is derived from an utterance is obtained. Hash values are determined for the utterance vector according to multiple different hash functions. A set of speaker vectors from a plurality of hash tables is determined using the hash values, where each speaker vector was derived from one or more utterances of a respective speaker. The speaker vectors in the set are compared with the utterance vector. A speaker vector is selected based on comparing the speaker vectors in the set with the utterance vector.
    Type: Application
    Filed: October 24, 2014
    Publication date: May 7, 2015
    Inventors: Matthew Sharifi, Ignacio Lopez Moreno, Ludwig Schmidt
  • Patent number: 9026437
    Abstract: A location determination system includes a first mobile terminal and a second mobile terminal. The first mobile terminal includes a first processor to acquire a first sound signal, analyze the first sound signal to obtain a first analysis result, and transmit the first analysis result. The second mobile terminal includes a second processor to acquire a second sound signal, analyze the second sound signal to obtain a second analysis result, receive the first analysis result from the first mobile terminal, compare the second analysis result with the first analysis result to obtain a comparison result, and determine whether the first mobile terminal locates in an area in which the second mobile terminal locates, based on the comparison result.
    Type: Grant
    Filed: March 26, 2012
    Date of Patent: May 5, 2015
    Assignee: Fujitsu Limited
    Inventor: Eiji Hasegawa
  • Patent number: 9020816
    Abstract: A method, system and apparatus are shown for identifying non-language speech sounds in a speech or audio signal. An audio signal is segmented and feature vectors are extracted from the segments of the audio signal. The segment is classified using a hidden Markov model (HMM) that has been trained on sequences of these feature vectors. Post-processing components can be utilized to enhance classification. An embodiment is described in which the hidden Markov model is used to classify a segment as a language speech sound or one of a variety of non-language speech sounds. Another embodiment is described in which the hidden Markov model is trained using discriminative learning.
    Type: Grant
    Filed: August 13, 2009
    Date of Patent: April 28, 2015
    Assignee: 21CT, Inc.
    Inventor: Matthew McClain
  • Patent number: 9015045
    Abstract: A method for refining a search is provided. Embodiments may include receiving a first speech signal corresponding to a first utterance and receiving a second speech signal corresponding to a second utterance, wherein the second utterance is a refinement to the first utterance. Embodiments may also include identifying information associated with the first speech signal as first speech signal information and identifying information associated with the second speech signal as second speech signal information. Embodiments may also include determining a first quantity of search results based upon the first speech signal information and determining a second quantity of search results based upon the second speech signal information.
    Type: Grant
    Filed: March 11, 2013
    Date of Patent: April 21, 2015
    Assignee: Nuance Communications, Inc.
    Inventor: Jean-Francois Lavallee