Clustering Patents (Class 704/245)
  • Patent number: 11328044
    Abstract: A dynamic recognition method includes, when the terminal device detects that the user is in a first distance range, obtaining, by the terminal device, first feature information of the user. The method further includes performing first identity authentication on the first feature information of the user, where the first feature information includes facial feature information, voice feature information, or behavioral feature information. The method further includes increasing, by the terminal device, a level of a default threshold of second identity authentication when the first identity authentication succeeds. The method further includes, when the terminal device detects that the user is in a second distance range, obtaining, by the terminal device, second feature information of the user, and performing second identity authentication on the second feature information of the user based on the default threshold whose level is increased.
    Type: Grant
    Filed: May 27, 2017
    Date of Patent: May 10, 2022
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Chi Wah Sun, Po Chin Yu
  • Patent number: 11295756
    Abstract: A system for ontology-aware sound classification. The system includes an electronic processor that is configured to create a first graph based on relationships between fine audio classification labels and create a second graph based on relationships between coarse audio classification labels. The electronic processor is also configured to receive an audio clip including one or more sounds, execute a first graph convolutional network with the first graph as input, and execute a second graph convolutional network with the second graph as input. Using the outputs of the first graph convolutional network and the second graph convolutional network, the electronic processor is configured to determine one or more coarse labels, one or more fine labels, or both to classify the one or more sounds in the audio clip.
    Type: Grant
    Filed: December 27, 2019
    Date of Patent: April 5, 2022
    Assignee: Robert Bosch GmbH
    Inventors: Shabnam Ghaffarzadegan, Zhe Feng, Yiwei Sun
  • Patent number: 11252169
    Abstract: A Cyber-Physical System (“CPS”) may have monitoring nodes that generate a series of current monitoring node values representing current operation of the CPS. A normal space data source may store, for each monitoring node, a series of normal monitoring node values representing normal operation of the CPS. An abnormal data generation platform may utilize information in the normal space data source and a generative model to create generated abnormal to represent abnormal operation of the CPS. An abnormality detection model creation computer may receive the normal monitoring node values (and generate normal feature vectors) and automatically calculate and output an abnormality detection model including information about a decision boundary created via supervised learning based on the normal feature vectors and the generated abnormal data.
    Type: Grant
    Filed: April 3, 2019
    Date of Patent: February 15, 2022
    Assignee: GENERAL ELECTRIC COMPANY
    Inventors: Weizhong Yan, Masoud Abbaszadeh
  • Patent number: 11210473
    Abstract: Techniques for identifying vocabulary associated with semantic objects used for generating natural language text with a natural language generation (NLG) system, the semantic objects including a first semantic object having a first set of ordered attributes. The techniques include: obtaining text segments; identifying, from among the text segments and using at least one first machine learning classifier, groups of text segments corresponding to respective semantic objects in the plurality of semantic objects; identifying, from the groups of text segments and using at least one second machine learning classifier, a plurality of vocabularies for the plurality of semantic objects; and generating natural language text using the NLG system, the plurality of vocabularies, and the plurality of semantic objects; and outputting the generated natural language text.
    Type: Grant
    Filed: March 12, 2020
    Date of Patent: December 28, 2021
    Assignee: YSEOP SA
    Inventors: Dominique Mariko, Yagmur Ozturk, Hugues Sézille de Mazancourt
  • Patent number: 11163297
    Abstract: One embodiment provides a method, including: obtaining historical information for equipment having at least one control, wherein the historical information indicates a setting for the at least one control during operation of the equipment and identifies operating performance of the equipment corresponding to the indicated setting; receiving a goal for the equipment, wherein the goal is related to a desired operating performance of the equipment; identifying, a plurality of sets of contiguous good reference segments, wherein a contiguous set of good reference segments comprises a plurality of operating time segments where the desired operating performance goal was achieved for a predetermined of time; identifying, a subset of sets comprising reference segments that are achievable from a current operating state of the equipment; selecting, a reference segment that is attainable based upon exogenous factors related to an operating environment of the equipment; and providing a recommendation to an operator of the
    Type: Grant
    Filed: September 11, 2018
    Date of Patent: November 2, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Pankaj S. Dayama, Prabuchandran Krithivasan Jayachandran, Nitin Singh, Vinayaka Pandit
  • Patent number: 11158325
    Abstract: A biometric system is tested to see whether a proposed use matches a configuration of the system. An enrolment input is received from an enrolling user, and compared with a system configuration model to obtain a configuration matching score value. The enrollment is then controlled based on a result of comparing the received enrollment input with the system configuration model. In the case of a voice biometric system, when a test input is received from a speaker, it is determined whether audio conditions applying to the test input correspond to system configuration conditions. Verification is performed by comparing the test input with a model of the speech of an enrolled user to generate a verification score for use in deciding whether to accept or reject the speaker, depending on whether it is determined that audio conditions applying to the test input correspond to the system configuration conditions.
    Type: Grant
    Filed: October 24, 2019
    Date of Patent: October 26, 2021
    Assignee: Cirrus Logic, Inc.
    Inventors: David Martínez González, Carlos Vaquero Avilés-Casco, Ana Mantecon
  • Patent number: 11055319
    Abstract: A method is described of identifying time-series signals that contain information useful for predicting impending event messages relating to one or more of safety, maintenance, and system operation information before they occur. The method includes loading a plurality of time-series signals with assigned signal name and associated time-series data into a machine-readable storage medium and grouping the plurality of time-series signals based on textual similarity of the corresponding signal names into a signal cluster.
    Type: Grant
    Filed: March 29, 2018
    Date of Patent: July 6, 2021
    Assignee: Hamilton Sundstrand Corporation
    Inventors: Joseph J. Ensberg, Chetan Prabhu, Marlee Ann Stevenson, Kamron Saniee
  • Patent number: 11024291
    Abstract: In an embodiment, the disclosed technologies include automatically recognizing speech content of an audio stream that may contain multiple different classes of speech content, by receiving, by an audio capture device, an audio stream; outputting, by one or more classifiers, in response to an inputting to the one or more classifiers of digital data that has been extracted from the audio stream, score data; where a score of the score data indicates a likelihood that a particular time segment of the audio stream contains speech of a particular class; where the one or more classifiers use one or more machine-learned models that have been trained to recognize audio of one or more particular classes to determine the score data; using a sliding time window process, selecting particular scores from the score data; using the selected particular scores, determining and outputting one or more decisions as to whether one or more particular time segments of the audio stream contain speech of one or more particular classes
    Type: Grant
    Filed: March 27, 2019
    Date of Patent: June 1, 2021
    Assignee: SRI INTERNATIONAL
    Inventors: Diego Castan Lavilla, Harry Bratt, Mitchell Leigh McLaren
  • Patent number: 11023676
    Abstract: Systems and methods for efficiently detecting and coordinating step changes, trends, cycles, and bursts affecting lexical items within data streams are provided. Data streams can be sourced from documents that can optionally be labeled with metadata. Changes can be grouped across lexical and/or metavalue vocabularies to summarize the changes that are synchronous in time. The methods described herein can be applied either retrospectively to a corpus of data or in a streaming mode.
    Type: Grant
    Filed: March 18, 2016
    Date of Patent: June 1, 2021
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Jeremy Wright, Alicia Abella, John Grothendieck
  • Patent number: 10969775
    Abstract: A building management system includes connected equipment configured to measure a plurality of monitored variables and a predictive diagnostics system configured to receive the monitored variables from the connected equipment; generate a probability distribution of the plurality of monitored variables; determine a boundary for the probability distribution using a supervised machine learning technique to separate normal conditions from faulty conditions indicated by the plurality of monitored variables; separate the faulty conditions into sub-patterns using an unsupervised machine learning technique to generate a fault prediction model, each sub-pattern corresponding with a fault, and each fault associated with a fault diagnosis; receive a current set of the monitored variables from the connected equipment; determine whether the current set of monitored variables correspond with one of the sub-patterns of the fault prediction model to facilitate predicting whether a corresponding fault will occur; and determin
    Type: Grant
    Filed: June 21, 2018
    Date of Patent: April 6, 2021
    Assignee: Johnson Controls Technology Company
    Inventors: Sumant S. Khalate, Tushar Shripad Joshi, Dishant Mittal
  • Patent number: 10965435
    Abstract: The disclosure relates to a transmission device, comprising: a processor configured: to generate a multicarrier signal based on a combination of data symbols and reference symbols, wherein the multicarrier signal comprises a first plurality of inband subcarriers and a second plurality of out-of band (OOB) subcarriers, and to precode the multicarrier signal based on a mapping function with respect to the first plurality of inband subcarriers and the second plurality of out-of band subcarriers, wherein the mapping function is configured to mitigate the OOB subcarriers.
    Type: Grant
    Filed: May 16, 2019
    Date of Patent: March 30, 2021
    Assignee: Huawei Technologies Duesseldorf GmbH
    Inventors: Mohamed Ibrahim, Wen Xu
  • Patent number: 10936965
    Abstract: A method executable via operation of configured processing circuitry may include constructing a mutual information graph for categorical data with respect to observed attributes of a plurality of entities described in terms of respective ones of the observed attributes by the categorical data, determining a clique tree correlating attributes having at least a threshold level of mutual dependence among the observed attributes, and determining a normality rating for an entity relative to the plurality of entities based on the clique tree.
    Type: Grant
    Filed: October 5, 2017
    Date of Patent: March 2, 2021
    Assignee: The John Hopkins University
    Inventor: Cetin Savkli
  • Patent number: 10929514
    Abstract: A user registration method and a device for a smart robot. The method comprises: conducting a voice dialogue with a new user to be registered, acquiring a user name of the user from the voice dialogue, and simultaneously collecting biological characteristic information that can uniquely identify the user; wherein the biological characteristic information comprises at least two different types of biological characteristic information, judging whether at least one type of the biological characteristic information satisfies a corresponding preset registration condition, and if yes, using the biological characteristic information that satisfies the preset registration condition as a characteristic template, establishing a correspondence relation between the characteristic template and the user name, and saving the correspondence relation, to complete the user registration.
    Type: Grant
    Filed: August 14, 2017
    Date of Patent: February 23, 2021
    Assignee: Goertek Inc.
    Inventors: Cui Liu, Honglong Ma, Chuan Chen
  • Patent number: 10872597
    Abstract: A speech synthesis dictionary delivery device that delivers a dictionary for performing speech synthesis to terminals, comprises a storage device for speech synthesis dictionary database that stores a first dictionary which includes an acoustic model of a speaker and is associated with identification information of the speaker, that stores a second dictionary which includes an acoustic model generated using voice data of a plurality of speakers, and that stores parameter sets of the speakers to be used with the second dictionary and which are associated with identification information of the speakers, a processor that determines one of the first dictionary and the second dictionary, which should be used in the terminal for a specified speaker, and an input output interface (I/F) that receives the identification information of a speaker transmitted from the terminal and then delivers at least one of a first dictionary, the second dictionary, and a parameter set of the second dictionary, on the basis of the rec
    Type: Grant
    Filed: August 8, 2018
    Date of Patent: December 22, 2020
    Assignees: Kabushiki Kaisha Toshiba, Toshiba Digital Solutions Cornoration
    Inventors: Kouichirou Mori, Gou Hirabayashi, Masahiro Morita, Yamato Ohtani
  • Patent number: 10847138
    Abstract: Systems and methods are disclosed for generating internal state representations of a neural network during processing and using the internal state representations for classification or search. In some embodiments, the internal state representations are generated from the output activation functions of a subset of nodes of the neural network. The internal state representations may be used for classification by training a classification model using internal state representations and corresponding classifications. The internal state representations may be used for search, by producing a search feature from an search input and comparing the search feature with one or more feature representations to find the feature representation with the highest degree of similarity.
    Type: Grant
    Filed: May 21, 2019
    Date of Patent: November 24, 2020
    Assignee: Deepgram, Inc.
    Inventors: Jeff Ward, Adam Sypniewski, Scott Stephenson
  • Patent number: 10824657
    Abstract: To provide a system capable of appropriately proposing a search term candidate for each page of a document. Provided is a search document information storage device comprising: a vocabulary extraction means 3; a keyword storage means 5; a keyword extraction means 7; a topic term storage means 9; a topic term extraction means 11; a search term candidate extraction means 13; a search term candidate display means 17; a search term input means 19; and a document search information storage means 21.
    Type: Grant
    Filed: May 7, 2018
    Date of Patent: November 3, 2020
    Assignee: Interactive Solutions Inc.
    Inventor: Kiyoshi Sekine
  • Patent number: 10789958
    Abstract: Methods, computer program products, and systems are presented. The methods include, for instance: obtaining the media file with a speech and identifying speakers on clusters separated by disfluencies and change of speakers. Clusters are re-segmented rearranged during diarization. Speaker identifications for the clusters in the media file is produced.
    Type: Grant
    Filed: September 30, 2019
    Date of Patent: September 29, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Aaron K. Baughman, Stephen C. Hammer
  • Patent number: 10607111
    Abstract: Described is a system for classifying novel objects in imagery. In operation, the system extracts salient patches from a plurality of unannotated images using a multi-layer network. Activations of the multi-layer network are clustered into key attribute, with the key attributes being displayed to a user on a display, thereby prompting the user to annotate the key attributes with class label. An attribute database is then generated based on user prompted annotations of the key attributes. A test image can then be passed through the system, allowing the system to classify at least one object in the test image by identifying an object class in the attribute database. Finally, a device can be caused to operate or maneuver based on the classification of the at least one object in the test image.
    Type: Grant
    Filed: February 4, 2019
    Date of Patent: March 31, 2020
    Assignee: HRL Laboratories, LLC
    Inventors: Soheil Kolouri, Charles E. Martin, Kyungnam Kim, Heiko Hoffmann
  • Patent number: 10592611
    Abstract: Embodiments of the present invention provide a system for automatically extracting conversational structure from a voice record based on lexical and acoustic features. The system also aggregates business-relevant statistics and entities from a collection of spoken conversations. The system may infer a coarse-level conversational structure based on fine-level activities identified from extracted acoustic features. The system improves significantly over previous systems by extracting structure based on lexical and acoustic features. This enables extracting conversational structure on a larger scale and finer level of detail than previous systems, and can feed an analytics and business intelligence platform, e.g. for customer service phone calls. During operation, the system obtains a voice record. The system then extracts a lexical feature using automatic speech recognition (ASR). The system extracts an acoustic feature.
    Type: Grant
    Filed: October 24, 2016
    Date of Patent: March 17, 2020
    Assignee: Conduent Business Services, LLC
    Inventors: Jesse Vig, Harish Arsikere, Margaret H. Szymanski, Luke R. Plurkowski, Kyle D. Dent, Daniel G. Bobrow, Daniel Davies, Eric Saund
  • Patent number: 10559303
    Abstract: The method comprises receive first audio comprising speech from a user of a computing device, detecting an end of speech in the first audio, generating an ASR result based, at least in part, on a portion of the first audio prior to the detected end of speech, determining whether a valid action can be performed by a speech-enabled application installed on the computing device using the ASR result, and processing second audio when it is determined that a valid action cannot be performed by the speech-enabled application using the ASR result.
    Type: Grant
    Filed: May 23, 2016
    Date of Patent: February 11, 2020
    Assignee: Nuance Communications, Inc.
    Inventor: Mark Fanty
  • Patent number: 10559311
    Abstract: Methods, computer program products, and systems are presented. The methods include, for instance: obtaining the media file with a speech and identifying speakers on clusters separated by disfluencies and change of speakers. Clusters are re-segmented rearranged during diarization. Speaker identifications for the clusters in the media file is produced.
    Type: Grant
    Filed: March 31, 2017
    Date of Patent: February 11, 2020
    Assignee: International Business Machines Corporation
    Inventors: Aaron K. Baughman, Stephen C. Hammer
  • Patent number: 10553206
    Abstract: According to one embodiment, a voice keyword detection apparatus includes a memory and a circuit coupled with the memory. The circuit calculates a first score for a first sub-keyword and a second score for a second sub-keyword. The circuit detects the first and second sub-keywords based on the first and second scores. The circuit determines, when the first sub-keyword is detected from one or more first frames, to accept the first sub-keyword. The circuit determines, when the second sub-keyword is detected from one or more second frames, whether to accept the second sub-keyword based on a start time and/or an end time of the one or more first frames and a start time and/or an end time of the one or more second frames.
    Type: Grant
    Filed: August 30, 2017
    Date of Patent: February 4, 2020
    Assignee: KABUSHIKI KAISHA TOSHIBA
    Inventor: Hiroshi Fujimura
  • Patent number: 10535339
    Abstract: According to an embodiment, a speech recognition result output device includes a storage and processing circuitry. The storage is configured to store a language model for speech recognition. The processing circuitry is coupled to the storage and configured to acquire a phonetic sequence, convert the phonetic sequence into a phonetic sequence feature vector, convert the phonetic sequence feature vector into graphemes using the language model, and output the graphemes.
    Type: Grant
    Filed: June 15, 2016
    Date of Patent: January 14, 2020
    Assignee: KABUSHIKI KAISHA TOSHIBA
    Inventor: Hiroshi Fujimura
  • Patent number: 10496930
    Abstract: With reference to information storing a co-occurrence probability of each of plural words in association with each of distribution-destinations, the apparatus extracts, from a message to be distributed, an unknown-word that is not included in the plural words, where the co-occurrence probability indicates a probability that each word is included in a message to be distributed to each distribution-destination.
    Type: Grant
    Filed: September 11, 2017
    Date of Patent: December 3, 2019
    Assignee: FUJITSU LIMITED
    Inventors: Yukihiro Watanabe, Ken Yokoyama, Masahiro Asaoka, Hiroshi Otsuka, Reiko Kondo
  • Patent number: 10459980
    Abstract: A display system for an issue comprises an input unit, a display unit and an processing unit. The input unit receives an initial keyword corresponding to an issue. The display unit displays at least a derivative issue generated from the issue during a time period according to time-based characteristics. The processing unit coupled to the input unit and the display unit obtains tags of subject contents of web pages, and obtains a present keywords group according to co-occurrence correlation of the tags. The processing unit analyzes the correlation between the present keywords calculated based on social voice, analyzing overlap rate for the present keywords compared with the initial keywords, and compares correlation between the present keywords with correlation between the initial keywords calculated based on social voice, in order to determine whether at least one of the derivative issue is generated.
    Type: Grant
    Filed: April 20, 2016
    Date of Patent: October 29, 2019
    Assignee: Institute For Information Industry
    Inventors: Tai-Ta Kuo, Ping-I Chen
  • Patent number: 10402742
    Abstract: A method includes accessing a first sensor log and a corresponding first reference log. Each of the first sensor log and the first reference log includes a series of measured values of a parameter according to a first time series. The method also includes accessing a second sensor log and a corresponding second reference log. Each of the second sensor log and the second reference log includes a series of measured values of a parameter according to a second time series. The method also includes dynamically time warping the first reference log and/or second reference log by a first transformation between the first time series and a common time-frame and/or a second transformation between the second time series and the common time-frame. The method also includes generating first and second warped sensor logs by applying the or each transformation to the corresponding ones of the first and second sensor logs.
    Type: Grant
    Filed: December 12, 2017
    Date of Patent: September 3, 2019
    Assignee: Palantir Technologies Inc.
    Inventors: Ezra Spiro, Andre Frederico Cavalheiro Menck, Peter Maag, Thomas Powell
  • Patent number: 10380997
    Abstract: Systems and methods are disclosed for generating internal state representations of a neural network during processing and using the internal state representations for classification or search. In some embodiments, the internal state representations are generated from the output activation functions of a subset of nodes of the neural network. The internal state representations may be used for classification by training a classification model using internal state representations and corresponding classifications. The internal state representations may be used for search, by producing a search feature from an search input and comparing the search feature with one or more feature representations to find the feature representation with the highest degree of similarity.
    Type: Grant
    Filed: August 22, 2018
    Date of Patent: August 13, 2019
    Assignee: Deepgram, Inc.
    Inventors: Jeff Ward, Adam Sypniewski, Scott Stephenson
  • Patent number: 10325602
    Abstract: Systems, methods, devices, and other techniques for training and using a speaker verification neural network. A computing device may receive data that characterizes a first utterance. The computing device provides the data that characterizes the utterance to a speaker verification neural network. Subsequently, the computing device obtains, from the speaker verification neural network, a speaker representation that indicates speaking characteristics of a speaker of the first utterance. The computing device determines whether the first utterance is classified as an utterance of a registered user of the computing device. In response to determining that the first utterance is classified as an utterance of the registered user of the computing device, the device may perform an action for the registered user of the computing device.
    Type: Grant
    Filed: August 2, 2017
    Date of Patent: June 18, 2019
    Assignee: Google LLC
    Inventors: Hasim Sak, Ignacio Lopez Moreno, Alan Sean Papir, Li Wan, Quan Wang
  • Patent number: 10269356
    Abstract: There is provided a system comprising a microphone, configured to receive an input speech from an individual, an analog-to-digital (A/D) converter to convert the input speech to digital form and generate a digitized speech, a memory storing an executable code and an age estimation database, a hardware processor executing the executable code to receive the digitized speech, identify a plurality of boundaries in the digitized speech delineating a plurality of phonemes in the digitized speech, extract a plurality of formant-based feature vectors from each phoneme in the digitized speech based on at least one of a formant position, a formant bandwidth, and a formant dispersion, compare the plurality of formant-based feature vectors with age determinant formant-based feature vectors of the age estimation database, determine the age of the individual when the comparison finds a match in the age estimation database, and communicate an age-appropriate response to the individual.
    Type: Grant
    Filed: August 22, 2016
    Date of Patent: April 23, 2019
    Assignee: Disney Enterprises, Inc.
    Inventors: Rita Singh, Jill Fain Lehman
  • Patent number: 10249294
    Abstract: A speech recognition method capable of automatic generation of phones according to the present invention includes: unsupervisedly learning a feature vector of speech data; generating a phone set by clustering acoustic features selected based on an unsupervised learning result; allocating a sequence of phones to the speech data on the basis of the generated phone set; and generating an acoustic model on the basis of the sequence of phones and the speech data to which the sequence of phones is allocated.
    Type: Grant
    Filed: July 11, 2017
    Date of Patent: April 2, 2019
    Assignee: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Dong Hyun Kim, Young Jik Lee, Sang Hun Kim, Seung Hi Kim, Min Kyu Lee, Mu Yeol Choi
  • Patent number: 10199036
    Abstract: A network device for implementing voice input comprises an input-obtaining module for obtaining voice input information, a sequence-determining module for determining an input character sequence corresponding to the voice input information based on a voice recognition model, an accuracy-determining module for determining appearance-probability information corresponding to word segments in the input character sequence so as to obtain accuracy information of the word segments, and a transmitting module for transmitting, to a user device, the input character sequence and the accuracy information of the word segments corresponding to the voice input information.
    Type: Grant
    Filed: December 17, 2013
    Date of Patent: February 5, 2019
    Assignee: Baidu Online Network Technology (Beijing) Co., LTD.
    Inventors: Yangyang Lu, Lei Jia
  • Patent number: 10147438
    Abstract: Embodiments of the invention include method, systems and computer program products for role modeling. Aspects of the invention include receiving, by a processor, audio data, wherein the audio data includes a plurality of audio conversation for one or more speakers. The one or more segments for each of the plurality of audio conversations are partitioned. A speaker is associated with each of the one or more segments. The one or more segments for each of the plurality of audio conversations are labeled with roles utilizing a speaker recognition engine. Speakers are clustered based at least in part on a number of times the speakers are present in an audio conversation.
    Type: Grant
    Filed: March 2, 2017
    Date of Patent: December 4, 2018
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Kenneth W. Church, Jason W. Pelecanos, Josef Vopicka, Weizhong Zhu
  • Patent number: 10141009
    Abstract: Methods, systems, and apparatuses for audio event detection, where the determination of a type of sound data is made at the cluster level rather than at the frame level. The techniques provided are thus more robust to the local behavior of features of an audio signal or audio recording. The audio event detection is performed by using Gaussian mixture models (GMMs) to classify each cluster or by extracting an i-vector from each cluster. Each cluster may be classified based on an i-vector classification using a support vector machine or probabilistic linear discriminant analysis. The audio event detection significantly reduces potential smoothing error and avoids any dependency on accurate window-size tuning. Segmentation may be performed using a generalized likelihood ratio and a Bayesian information criterion, and the segments may be clustered using hierarchical agglomerative clustering. Audio frames may be clustered using K-means and GMMs.
    Type: Grant
    Filed: May 31, 2017
    Date of Patent: November 27, 2018
    Assignee: Pindrop Security, Inc.
    Inventors: Elie Khoury, Matthew Garland
  • Patent number: 10121466
    Abstract: Speech recognition systems that use voice templates may create (or update) voice templates for a particular user by training (or re-training). If a training results in a vocabulary with similar voice templates, then the speech recognition system's performance may suffer. The present invention provides embraces methods for training a speech recognition system to prevent voice template similarity. In these methods, a trained word's voice template may be evaluated for similarity to other vocabulary templates prior to enrolling the voice template into the vocabulary. If template similarity is found, then a user may be prompted to retrain the system using an alternate word. Alternatively, the user may be prompted to retrain the system with the word spoken more clearly. This dynamic enrollment training analysis insures that all templates in the vocabulary are distinct.
    Type: Grant
    Filed: February 11, 2015
    Date of Patent: November 6, 2018
    Assignee: Hand Held Products, Inc.
    Inventor: John Pecorari
  • Patent number: 10109280
    Abstract: In a method of diarization of audio data, audio data is segmented into a plurality of utterances. Each utterance is represented as an utterance model representative of a plurality of feature vectors. The utterance models are clustered. A plurality of speaker models are constructed from the clustered utterance models. A hidden Markov model is constructed of the plurality of speaker models. A sequence of identified speaker models is decoded.
    Type: Grant
    Filed: December 12, 2017
    Date of Patent: October 23, 2018
    Assignee: VERINT SYSTEMS LTD.
    Inventors: Oana Sidi, Ron Wein
  • Patent number: 9947314
    Abstract: Software that trains an artificial neural network for generating vector representations for natural language text, by performing the following steps: (i) receiving, by one or more processors, a set of natural language text; (ii) generating, by one or more processors, a set of first metadata for the set of natural language text, where the first metadata is generated using supervised learning method(s); (iii) generating, by one or more processors, a set of second metadata for the set of natural language text, where the second metadata is generated using unsupervised learning method(s); and (iv) training, by one or more processors, an artificial neural network adapted to generate vector representations for natural language text, where the training is based, at least in part, on the received natural language text, the generated set of first metadata, and the generated set of second metadata.
    Type: Grant
    Filed: February 21, 2017
    Date of Patent: April 17, 2018
    Assignee: International Business Machines Corporation
    Inventors: Liangliang Cao, James J. Fan, Chang Wang, Bing Xiang, Bowen Zhou
  • Patent number: 9875743
    Abstract: Disclosed herein are methods of diarizing audio data using first-pass blind diarization and second-pass blind diarization that generate speaker statistical models, wherein the first pass-blind diarization is on a per-frame basis and the second pass-blind diarization is on a per-word basis, and methods of creating acoustic signatures for a common speaker based only on the statistical models of the speakers in each audio session.
    Type: Grant
    Filed: January 26, 2016
    Date of Patent: January 23, 2018
    Assignee: VERINT SYSTEMS LTD.
    Inventors: Alex Gorodetski, Ido Shapira, Ron Wein, Oana Sidi
  • Patent number: 9860669
    Abstract: An audio apparatus includes a receiver configured to receive audio data and audio transducer position data for a plurality of audio transducers; and a renderer configured to render the audio data by generating audio transducer drive signals for the audio transducers from the audio data. Further, a clusterer is configured to cluster the audio transducers into a set of clusters in response to the audio transducer position data and to distances between audio transducers in accordance with a distance metric. A render controller is configured to adapt the rendering in response to the clustering. The apparatus is configured to select array processing techniques for specific subsets that contain audio transducers that are sufficiently close and allow automatic adaptation to audio transducer configurations thereby, e.g., allowing a user increased flexibility in positioning loudspeakers.
    Type: Grant
    Filed: May 6, 2014
    Date of Patent: January 2, 2018
    Assignee: KONINKLIJKE PHILIPS N.V.
    Inventors: Werner Paulus Josephus De Bruijn, Arnoldus Werner Johannes Oomen, Aki Sakari Haermae
  • Patent number: 9837068
    Abstract: A method for verifying at least one sound sample to be used in generating a sound detection model in an electronic device includes receiving a first sound sample; extracting a first acoustic feature from the first sound sample; receiving a second sound sample; extracting a second acoustic feature from the second sound sample; and determining whether the second acoustic feature is similar to the first acoustic feature.
    Type: Grant
    Filed: April 8, 2015
    Date of Patent: December 5, 2017
    Assignee: QUALCOMM Incorporated
    Inventors: Sunkuk Moon, Minho Jin, Haiying Xia, Hesu Huang, Warren Frederick Dale
  • Patent number: 9830931
    Abstract: One embodiment of the present invention sets forth a technique for determining a set of sound parameters associated with a sound type. The technique includes receiving, via a network and from each a first plurality of remote computing devices, an audio recording of a first sound type and a descriptor associated with the first sound type. The technique further includes processing the audio recordings via a processor to determine a first set of sound parameters associated with the first sound type. The technique further includes receiving a request associated with the descriptor from at least one remote computing device and, in response, transmitting the first set of sound parameters associated with the first sound type to the at least one remote computing device.
    Type: Grant
    Filed: December 31, 2015
    Date of Patent: November 28, 2017
    Assignee: HARMAN INTERNATIONAL INDUSTRIES, INCORPORATED
    Inventors: Donald Joseph Butts, Brandon Stacey
  • Patent number: 9714884
    Abstract: A statistical basic classification model of acoustic features generated for at least one reference object is automatically adapted by a data processing unit based on acoustic features of a noise generated by an object to be investigated to obtain an individually adapted statistical classification model. The data processing unit then classifies the state of the noise-generating object based on the individually adapted statistical classification model.
    Type: Grant
    Filed: April 29, 2009
    Date of Patent: July 25, 2017
    Assignee: SIEMENS AKTIENGESELLSCHAFT
    Inventors: Joachim Hofer, Lutz Leutelt
  • Patent number: 9691391
    Abstract: Systems and methods to perform speaker clustering determine which audio segments appear to include sound generated by the same speaker. Speaker clustering is based on creating a graph in which a node represents an audio segment and an edge between two nodes represents a relationship and/or correspondence that reflects a probability, likelihood, or other indication that the two nodes represent audio segments of the same speaker. This graph is analyzed to detect individual communities of nodes that associate to an individual speaker.
    Type: Grant
    Filed: May 21, 2015
    Date of Patent: June 27, 2017
    Assignee: KnuEdge Incorporated
    Inventor: Rodney Gateau
  • Patent number: 9672814
    Abstract: Software that trains an artificial neural network for generating vector representations for natural language text, by performing the following steps: (i) receiving, by one or more processors, a set of natural language text; (ii) generating, by one or more processors, a set of first metadata for the set of natural language text, where the first metadata is generated using supervised learning method(s); (iii) generating, by one or more processors, a set of second metadata for the set of natural language text, where the second metadata is generated using unsupervised learning method(s); and (iv) training, by one or more processors, an artificial neural network adapted to generate vector representations for natural language text, where the training is based, at least in part, on the received natural language text, the generated set of first metadata, and the generated set of second metadata.
    Type: Grant
    Filed: May 8, 2015
    Date of Patent: June 6, 2017
    Assignee: International Business Machines Corporation
    Inventors: Liangliang Cao, James J. Fan, Chang Wang, Bing Xiang, Bowen Zhou
  • Patent number: 9666192
    Abstract: Methods and apparatus for reducing latency in speech recognition applications. The method comprises receive first audio comprising speech from a user of a computing device, detecting an end of speech in the first audio, generating an ASR result based, at least in part, on a portion of the first audio prior to the detected end of speech, determining whether a valid action can be performed by a speech-enabled application installed on the computing device using the ASR result, and processing second audio when it is determined that a valid action cannot be performed by the speech-enabled application using the ASR result.
    Type: Grant
    Filed: May 26, 2015
    Date of Patent: May 30, 2017
    Assignee: Nuance Communications, Inc.
    Inventor: Mark Fanty
  • Patent number: 9659560
    Abstract: Software that trains an artificial neural network for generating vector representations for natural language text, by performing the following steps: (i) receiving, by one or more processors, a set of natural language text; (ii) generating, by one or more processors, a set of first metadata for the set of natural language text, where the first metadata is generated using supervised learning method(s); (iii) generating, by one or more processors, a set of second metadata for the set of natural language text, where the second metadata is generated using unsupervised learning method(s); and (iv) training, by one or more processors, an artificial neural network adapted to generate vector representations for natural language text, where the training is based, at least in part, on the received natural language text, the generated set of first metadata, and the generated set of second metadata.
    Type: Grant
    Filed: September 30, 2015
    Date of Patent: May 23, 2017
    Assignee: International Business Machines Corporation
    Inventors: Liangliang Cao, James J. Fan, Chang Wang, Bing Xiang, Bowen Zhou
  • Patent number: 9641968
    Abstract: A system for sharing moment experiences is described. A system receives moment data from an input to a mobile device. The system receives geographic location information, time information, and contextual information that is local to the mobile device. The system creates a message about the moment data based on the geographic location information, the time information, and the contextual information. The system outputs the moment data with the message.
    Type: Grant
    Filed: May 15, 2015
    Date of Patent: May 2, 2017
    Assignee: Krumbs, Inc.
    Inventors: Neilesh Jain, Ramesh Jain, Pinaki Sinha
  • Patent number: 9620148
    Abstract: Systems, vehicles, and methods for limiting speech-based access to an audio metadata database are described herein. Audio metadata databases described herein include a plurality of audio metadata entries. Each audio metadata entry includes metadata information associated with at least one audio file. Embodiments described herein determine when a size of the audio metadata database reaches a threshold size, and limit which of the plurality of audio metadata entries may be accessed in response to the speech input signal when the size of the audio metadata database reaches the threshold size.
    Type: Grant
    Filed: July 1, 2013
    Date of Patent: April 11, 2017
    Assignee: Toyota Motor Engineering & Manufacturing North America, Inc.
    Inventor: Eric Randell Schmidt
  • Patent number: 9595260
    Abstract: A modeling device comprises a front end which receives enrollment speech data from each target speaker, a reference anchor set generation unit which generates a reference anchor set using the enrollment speech data based on an anchor space, and a voice print generation unit which generates voice prints based on the reference anchor set and the enrollment speech data. By taking the enrollment speech and speaker adaptation technique into account, anchor models with a smaller size can be generated, so reliable and robust speaker recognition with a smaller size reference anchor set is possible.
    Type: Grant
    Filed: December 10, 2010
    Date of Patent: March 14, 2017
    Assignee: PANASONIC INTELLECTUAL PROPERTY CORPORATION OF AMERICA
    Inventors: Haifeng Shen, Long Ma, Bingqi Zhang
  • Patent number: 9576582
    Abstract: Disclosed herein are systems, computer-implemented methods, and computer-readable storage media for recognizing speech by adapting automatic speech recognition pronunciation by acoustic model restructuring. The method identifies an acoustic model and a matching pronouncing dictionary trained on typical native speech in a target dialect. The method collects speech from a new speaker resulting in collected speech and transcribes the collected speech to generate a lattice of plausible phonemes. Then the method creates a custom speech model for representing each phoneme used in the pronouncing dictionary by a weighted sum of acoustic models for all the plausible phonemes, wherein the pronouncing dictionary does not change, but the model of the acoustic space for each phoneme in the dictionary becomes a weighted sum of the acoustic models of phonemes of the typical native speech. Finally the method includes recognizing via a processor additional speech from the target speaker using the custom speech model.
    Type: Grant
    Filed: February 23, 2016
    Date of Patent: February 21, 2017
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Andrej Ljolje, Alistair D. Conkie, Ann K. Syrdal
  • Patent number: 9524291
    Abstract: Techniques involving visual display of information related to matching user utterances against graph patterns are described. In one or more implementations, an utterance of a user is obtained that has been indicated as corresponding to a graph pattern through linguistic analysis. The utterance is displayed in a user interface as a representation of the graph pattern.
    Type: Grant
    Filed: October 6, 2010
    Date of Patent: December 20, 2016
    Assignee: Virtuoz SA
    Inventors: Dan Teodosiu, Elizabeth Ireland Powers, Pierre Serge Vincent LeRoy, Sebastien Jean-Marie Christian Saunier