Preliminary Matching Patents (Class 704/247)
  • Patent number: 10262660
    Abstract: A method includes detecting an event published to a workflow activity by a voice based dialog view, wherein the event indicates a state of asset retrieval, navigating to a built-in asset retrieval work activity, retrieving an asset, and dismissing the workflow activity to revert to a workflow activity associated with the voice based dialog view.
    Type: Grant
    Filed: January 5, 2016
    Date of Patent: April 16, 2019
    Assignee: HAND HELD PRODUCTS, INC.
    Inventors: Shawn Zabel, Jeffrey Pike, Brian Bender, Dennis Doubleday, Mark Murawski
  • Patent number: 10249306
    Abstract: A speaker identification device includes: a primary speaker identification unit that computes, for each pre-stored registered speaker, a score that indicates the similarity between input speech and speech of the registered speakers; a similar speaker selection unit that selects a plurality of the registered speakers as similar speakers according to the height of the scores thereof; a learning unit that creates a classifier for each similar speaker by sorting the speech of a certain similar speaker among the similar speakers as a positive instance and the speech of the other similar speakers as negative instances; and a secondary speaker identification unit that computes, for each classifier, a score of the classifier with respect to the input speech, and outputs an identification result.
    Type: Grant
    Filed: January 16, 2014
    Date of Patent: April 2, 2019
    Assignee: NEC CORPORATION
    Inventors: Masahiro Tani, Takafumi Koshinaka, Yoshifumi Onishi, Shigeru Sawada
  • Patent number: 10140992
    Abstract: Systems, computer-implemented methods, and tangible computer-readable media are provided for voice authentication. The method includes receiving, on a mobile device, a speech sample from a user as part of a request for a restricted-access resource separate from the mobile device. When the user has previously established an identity with the mobile device, the method includes transmitting the speech sample along with the request to an authentication server which compares the speech sample to a previously established speech profile associated with the user and providing access to the restricted-access resource based on a response to the request from the authentication server if the speech sample from the user matches the speech profile on the authentication server with a minimum certainty threshold.
    Type: Grant
    Filed: April 6, 2017
    Date of Patent: November 27, 2018
    Assignee: NUANCE COMMUNICATIONS, INC.
    Inventor: Saurabh Kumar
  • Patent number: 10068445
    Abstract: Systems and methods of a security system are provided, including detecting, by a sensor, a sound event, and selecting, by a processor coupled to the sensor, at least a portion of sound data captured by the sensor that corresponds to at least one sound feature of the detected sound event. The systems and methods include classifying the at least one sound feature into one or more sound categories, and determining, by a processor, based upon a database of home-specific sound data, whether the at least one sound feature is a human-generated sound. A notification can be transmitted to a computing device according to the sound event.
    Type: Grant
    Filed: June 24, 2015
    Date of Patent: September 4, 2018
    Assignee: Google LLC
    Inventors: Rajeev Conrad Nongpiur, Michael Dixon
  • Patent number: 9691410
    Abstract: The present invention relates to a frequency band extending device and method, an encoding device and method, a decoding device and method, and a program, whereby music signals can be played with higher sound quality due to the extension of frequency bands. A bandpass filter 13 divides an input signal into multiple sub-band signals, a feature amount calculating circuit 14 calculates feature amount using at least one of the multiple divided sub-band signals and the input signal, a high frequency sub-band power estimating circuit 15 calculates an estimated value of a high frequency sub-band power based on the calculated feature amount, a high frequency signal generating circuit 16 generates a high frequency signal component based on the multiple sub-band signals divided by the bandpass filter 13, and the estimated value of the high frequency sub-band power calculated by the high frequency sub-band power estimating circuit 15.
    Type: Grant
    Filed: September 30, 2015
    Date of Patent: June 27, 2017
    Assignee: Sony Corporation
    Inventors: Yuki Yamamoto, Toru Chinen, Hiroyuki Honma, Yuhki Mitsufuji
  • Patent number: 9672841
    Abstract: The present document relates to a voice activity detection (VAD) method and methods used for voice activity detection and apparatus thereof, the VAD method includes: obtaining sub-band signals and spectrum amplitudes of a current frame; computing values of a energy feature and a spectral centroid feature of the current frame according to the sub-band signals; computing a signal to noise ratio parameter of the current frame according to a background noise energy estimated from a previous frame, an energy of SNR sub-bands and a energy feature of the current frame; computing a VAD decision result according to a tonality signal flag, a signal to noise ratio parameter, a spectral centroid feature, and a frame energy feature. The methods and apparatus of the present document can improve the accuracy of non-stationary noise (such as office noise) and music detection.
    Type: Grant
    Filed: June 30, 2015
    Date of Patent: June 6, 2017
    Assignee: ZTE Corporation
    Inventors: Dongping Jiang, Hao Yuan, Changbao Zhu
  • Patent number: 9584562
    Abstract: An embodiment of a method and apparatus for provisioning of a communication device includes receiving a registration request from a first communication device. The registration request includes an address associated with the first communication device. The method further includes registering the first communication device in response to receiving the registration request, placing a call request to the first communication device, and establishing a call session with the first communication device. The method further includes prompting a user of the first communication device for a user identifier, and receiving a user identifier from the user of the first communication device. The method still further includes sending one or more configuration parameters associated with the user identifier to the first communication device. The one or more configuration parameters are operable to configure the first communication device.
    Type: Grant
    Filed: November 17, 2014
    Date of Patent: February 28, 2017
    Assignee: CenturyLink Intellectual Property LLC
    Inventors: Mike A. Roberts, Shekhar Gupta, Jim Kevin Edwards
  • Patent number: 9489950
    Abstract: Embodiments of systems and methods for speaker verification are provided. In various embodiments, a method includes receiving an utterance from a speaker and determining a text-independent speaker verification score and a text-dependent speaker verification score in response to the utterance. Various embodiments include a system for speaker verification, the system comprising an audio receiving device for receiving an utterance from a speaker and converting the utterance to an utterance signal, and a processor coupled to the audio receiving device for determining speaker verification in response to the utterance signal, wherein the processor determines speaker verification in response to a UBM-independent speaker-normalized score.
    Type: Grant
    Filed: May 23, 2013
    Date of Patent: November 8, 2016
    Assignee: Agency for Science, Technology and Research
    Inventors: Anthony Larcher, Kong Aik Lee, Bin Ma, Thai Ngoc Thuy Huong
  • Patent number: 9396730
    Abstract: Systems and methods for determining an identity of an individual are provided. Audio may be received that includes a key phrase spoken by the individual, and the key phrase may include an identifier spoken by the individual. A key phrase voice print and key phrase text corresponding to the audio may be obtained. The key phrase text may include text corresponding to the identifier spoken by the individual. Voice prints may be retrieved based on the text corresponding to the identifier, and the voice prints may be provided to a voice biometric engine for comparison to the key phrase voice print. The individual may be authenticated based on a comparison of the key phrase voice print to the voice prints. The identifier may include a first name and a last name of the individual.
    Type: Grant
    Filed: September 30, 2013
    Date of Patent: July 19, 2016
    Assignee: Bank of America Corporation
    Inventors: David Karpey, Mark Pender
  • Patent number: 9251808
    Abstract: According to one embodiment, a speaker clustering apparatus includes a clustering unit, an extraction unit, and an error detection unit. The clustering unit is configured to extract acoustic features for speakers from an acoustic signal, and to cluster utterances included in the acoustic signal into the speakers by using the acoustic features. The extraction unit is configured to acquire character strings representing contents of the utterances, and to extract linguistic features of the speakers by using the character strings. The error detection unit is configured to decide that, when one of the character strings does not fit with a linguistic feature of a speaker into which an utterance of the one is clustered, the utterance is erroneously clustered by the clustering unit.
    Type: Grant
    Filed: March 6, 2012
    Date of Patent: February 2, 2016
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Tomoo Ikeda, Manabu Nagao, Osamu Nishiyama, Hirokazu Suzuki, Koji Ueno, Nobuhiro Shimogori
  • Patent number: 9237232
    Abstract: Systems and methods for analyzing digital recordings of the human voice in order to find characteristics unique to an individual. A biometrics engine may use an analytics service in a contact center to supply audio streams based on configured rules and providers for biometric detection. The analytics service may provide call audio data and attributes to connected engines based on a provider-set of selection rules. The connected providers send call audio data and attributes through the analytics service. The engines are notified when a new call is available for processing and can then retrieve chunks of audio data and call attributes by polling an analytics service interface. A mathematical model of the human vocal tract in the call audio data is created and/or matched against existing models. The result is analogous to a fingerprint, i.e., a pattern unique to an individual to within some level of probability.
    Type: Grant
    Filed: March 14, 2014
    Date of Patent: January 12, 2016
    Assignee: VERINT AMERICAS INC.
    Inventors: Jamie Richard Williams, Robert John Barnes, Ian Linsdell, Scott M. Bluman
  • Patent number: 9123182
    Abstract: Computer program products, methods, systems, etc. for generating an animated preview of a number of images are disclosed. A selection of a group of images is received. A set of digital images from the group of images are identified as being representative of the group. At least some portion of the identified set of representative digital images from the group is then used to create an animated image. The animated image serves as a preview of the group of images, such that, when a user browses the images and sees the preview associated with a corresponding folder or directory, the user is able to quickly and easily associate the images in the group with a particular event and identify contents of the folder or directory.
    Type: Grant
    Filed: November 5, 2010
    Date of Patent: September 1, 2015
    Assignee: Adobe Systems Incorporated
    Inventor: Sanjeev Kumar Biswas
  • Patent number: 9064494
    Abstract: The present invention relates to a method of and a device (30) for automatic recognition of given keywords and/or terms within voice data (25) of a talk between at least two participants, said voice data (25) being continuously compared during said talk with said given keywords and/or terms (26) with regard to the occurrence of said given keywords and/or terms (26) within said voice data (25). In order to provide a solution which guarantees that topics which should be part of a talk are actually dealt with during said talk, the method is characterized in that a visualized representation (48) of the results (45) of said comparison is presented to at least one participant during the talk.
    Type: Grant
    Filed: September 13, 2010
    Date of Patent: June 23, 2015
    Assignee: VODAFONE GMBH
    Inventor: Stefan Holtel
  • Patent number: 9043210
    Abstract: A biometric voice command and control switching device has a microphone assembly for receiving a currently spoken challenge utterance and a reference utterance, and a voice processing circuit for creating electronic signals indicative thereof. The device further includes a memory for storing the electronic signals, and a processor for comparing the electronic signals to determine if there is a match. If there is a match, an interface circuit enables the operable control of the controlled device.
    Type: Grant
    Filed: October 2, 2012
    Date of Patent: May 26, 2015
    Assignee: Voice Security Systems, Inc.
    Inventors: Sherrie Adcock, Kent Robinson, Brad Clements, Mark Keith Brockelsby, William Keith Brockelsby
  • Patent number: 9037470
    Abstract: Apparatus and methods are provided for using automatic speech recognition to analyze a voice interaction and verify compliance of an agent reading a script to a client during the voice interaction. In one aspect of the invention, a communications system includes a user interface, a communications network, and a call center having an automatic speech recognition component. In other aspects of the invention, a script compliance method includes the steps of conducting a voice interaction between an agent and a client and evaluating the voice interaction with an automatic speech recognition component adapted to analyze the voice interaction and determine whether the agent has adequately followed the script. In yet still further aspects of the invention, the duration of a given interaction can be analyzed, either apart from or in combination with the script compliance analysis above, to seek to identify instances of agent non-compliance, of fraud, or of quality-analysis issues.
    Type: Grant
    Filed: June 25, 2014
    Date of Patent: May 19, 2015
    Assignee: West Business Solutions, LLC
    Inventors: Mark J. Pettay, Fonda J. Narke
  • Patent number: 9026447
    Abstract: A first communication path for receiving a communication is established. The communication includes speech, which is processed. A speech pattern is identified as including a voice-command. A portion of the speech pattern is determined as including the voice-command. That portion of the speech pattern is separated from the speech pattern and compared with a second speech pattern. If the two speech patterns match or resemble each other, the portion of the speech pattern is accepted as the voice-command. An operation corresponding to the voice-command is determined and performed. The operation may perform an operation on a remote device, forward the voice-command to a remote device, or notify a user. The operation may create a second communication path that may allow a headset to join in a communication between another headset and a communication device, several headsets to communicate with each other, or a headset to communicate with several communication devices.
    Type: Grant
    Filed: January 25, 2008
    Date of Patent: May 5, 2015
    Assignee: CenturyLink Intellectual Property LLC
    Inventors: Erik Geldbach, Kelsyn D. Rooks, Sr., Shane M. Smith, Mark Wilmoth
  • Patent number: 9015043
    Abstract: A computer-implemented method includes receiving an electronic representation of one or more human voices, recognizing words in a first portion of the electronic representation of the one or more human voices, and sending suggested search terms to a display device for display to a user in a text format. The suggested search terms are based on the recognized words in the first portion of the electronic representation of the one or more human voices. A search query is received from the user, which includes one or more of the suggested search terms that were displayed to the user.
    Type: Grant
    Filed: October 1, 2010
    Date of Patent: April 21, 2015
    Assignee: Google Inc.
    Inventor: Scott Jenson
  • Patent number: 9009025
    Abstract: In some implementations, a digital work provider may provide language model information related to a plurality of different contexts, such as a plurality of different digital works. For example, the language model information may include language model difference information identifying a plurality of sequences of one or more words in a digital work that have probabilities of occurrence that differ from probabilities of occurrence in a base language model by a threshold amount. The language model difference information corresponding to a particular context may be used in conjunction with the base language model to recognize an utterance made by a user of a user device. In some examples, the recognition is performed on the user device. In other examples, the utterance and associated context information are sent over a network to a recognition computing device that performs the recognition.
    Type: Grant
    Filed: December 27, 2011
    Date of Patent: April 14, 2015
    Assignee: Amazon Technologies, Inc.
    Inventor: Brandon W. Porter
  • Patent number: 9002710
    Abstract: The invention involves the loading and unloading of dynamic section grammars and language models in a speech recognition system. The values of the sections of the structured document are either determined in advance from a collection of documents of the same domain, document type, and speaker; or collected incrementally from documents of the same domain, document type, and speaker; or added incrementally to an already existing set of values. Speech recognition in the context of the given field is constrained to the contents of these dynamic values. If speech recognition fails or produces a poor match within this grammar or section language model, speech recognition against a larger, more general vocabulary that is not constrained to the given section is performed.
    Type: Grant
    Filed: September 12, 2012
    Date of Patent: April 7, 2015
    Assignee: Nuance Communications, Inc.
    Inventors: Alwin B. Carus, Larissa Lapshina, Raghu Vemula
  • Patent number: 8996368
    Abstract: A feature transform for speech recognition is described. An input speech utterance is processed to produce a sequence of representative speech vectors. A time-synchronous speech recognition pass is performed using a decoding search to determine a recognition output corresponding to the speech input. The decoding search includes, for each speech vector after some first threshold number of speech vectors, estimating a feature transform based on the preceding speech vectors in the utterance and partial decoding results of the decoding search. The current speech vector is then adjusted based on the current feature transform, and the adjusted speech vector is used in a current frame of the decoding search.
    Type: Grant
    Filed: February 22, 2010
    Date of Patent: March 31, 2015
    Assignee: Nuance Communications, Inc.
    Inventor: Daniel Willett
  • Patent number: 8996387
    Abstract: For clearing transaction data selected for a processing, there is generated in a portable data carrier (1) a transaction acoustic signal (003; 103; 203) (S007; S107; S207) upon whose acoustic reproduction by an end device (10) at least transaction data selected for the processing are reproduced superimposed acoustically with a melody specific to a user of the data carrier (1) (S009; S109; S209). The generated transaction acoustic signal (003; 103; 203) is electronically transferred to an end device (10) (S108; S208), which processes the selected transaction data (S011; S121; S216) only when the user of the data carrier (1) confirms vis-à-vis the end device (10) an at least partial match both of the acoustically reproduced melody with the user-specific melody and of the acoustically reproduced transaction data with the selected transaction data (S010; S110, S116; S210).
    Type: Grant
    Filed: September 8, 2009
    Date of Patent: March 31, 2015
    Assignee: Giesecke & Devrient GmbH
    Inventors: Thomas Stocker, Michael Baldischweiler
  • Patent number: 8990071
    Abstract: A method for managing an interaction of a calling party to a communication partner is provided. The method includes automatically determining if the communication partner expects DTMF input. The method also includes translating speech input to one or more DTMF tones and communicating the one or more DTMF tones to the communication partner, if the communication partner expects DTMF input.
    Type: Grant
    Filed: March 29, 2010
    Date of Patent: March 24, 2015
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Yun-Cheng Ju, Stefanie Tomko, Frank Liu, Ivan Tashev
  • Patent number: 8983837
    Abstract: A computerized alert mode management method of a communication device, the communication device includes a sound capture unit. Vocal sounds of the environment around the communication device are extracted at regular intervals using the sound capture unit. Voice characteristic information of the captured vocal sounds is extracted using a speech recognition method and/or a voice recognition method. The communication device is controlled to work at one of a plurality of predetermined alert modes according to the extracted voice characteristic information.
    Type: Grant
    Filed: December 6, 2012
    Date of Patent: March 17, 2015
    Assignee: Hon Hai Precision Industry Co., Ltd.
    Inventor: Tsung-Jen Chuang
  • Patent number: 8977547
    Abstract: A voice recognition system includes: a voice input unit 11 for inputting a voice uttered a plurality of times; a registering voice data storage unit 12 for storing voice data uttered the plurality of times and input into the voice input unit 11; an utterance stability verification unit 13 for determining a similarity between the voice data uttered the plurality of times that are read from the registering voice data storage unit 12, and determining that registration of the voice data is acceptable when the similarity is greater than a threshold Tl; and a standard pattern creation unit 14 for creating a standard pattern by using the voice data where the utterance stability verification unit 13 determines that registration is acceptable.
    Type: Grant
    Filed: October 8, 2009
    Date of Patent: March 10, 2015
    Assignee: Mitsubishi Electric Corporation
    Inventors: Michihiro Yamazaki, Jun Ishii, Hiroki Sakashita, Kazuyuki Nogi
  • Patent number: 8977549
    Abstract: A natural language business system and method is developed to understand the underlying meaning of a person's speech, such as during a transaction with the business system. The system includes a speech recognition engine, and action classification engine, and a control module. The control module causes the system to execute an inventive method wherein the speech recognition and action classification models may be recursively optimized on an unisolated performance metric that is pertinent to the overall performance of the natural language business system, as opposed to the isolated model-specific criteria previously employed.
    Type: Grant
    Filed: September 26, 2013
    Date of Patent: March 10, 2015
    Assignee: Nuance Communications, Inc.
    Inventors: Sabine V. Deligne, Yuqing Gao, Vaibhava Goel, Hong-Kwang Kuo, Cheng Wu
  • Patent number: 8976943
    Abstract: Provided is a method and a telephone-based system with voice-verification capabilities that enable a user to safely and securely conduct transactions with his or her online financial transaction program account over the phone in a convenient and user-friendly fashion, without having to depend on an internet connection.
    Type: Grant
    Filed: September 25, 2012
    Date of Patent: March 10, 2015
    Assignee: Ebay Inc.
    Inventor: Will Tonini
  • Publication number: 20150046162
    Abstract: Device, system, and method of liveness detection using voice biometrics. For example, a method comprises: generating a first matching score based on a comparison between: (a) a voice-print from a first text-dependent audio sample received at an enrollment stage, and (b) a second text-dependent audio sample received at an authentication stage; generating a second matching score based on a text-independent audio sample; and generating a liveness score by taking into account at least the first matching score and the second matching score.
    Type: Application
    Filed: October 27, 2014
    Publication date: February 12, 2015
    Applicant: Nuance Communications, Inc.
    Inventors: ALMOG ALEY-RAZ, Nir Moshe Krause, Michael Itzhak Salmon, Ran Yehoshua Gazit
  • Patent number: 8948350
    Abstract: Disclosed is a secure telephone call management system for authenticating users of a telephone system in an institutional facility. Authentication of the users is accomplished by using a personal identification number, preferably in conjunction with speaker independent voice recognition and speaker dependent voice identification. When a user first enters the system, the user speaks his or her name which is used as a sample voice print. During each subsequent use of the system, the user is required to speak his or her name. Voice identification software is used to verify that the provided speech matches the sample voice print. The secure system includes accounting software to limit access based on funds in a user's account or other related limitations. Management software implements widespread or local changes to the system and can modify or set any number of user account parameters.
    Type: Grant
    Filed: July 11, 2008
    Date of Patent: February 3, 2015
    Assignee: Global Tel*Link Corporation
    Inventor: Stephen L. Hodge
  • Patent number: 8938382
    Abstract: An item of information (212) is transmitted to a distal computer (220), translated to a different sense modality and/or language (222), and in substantially real time, and the translation (222) is transmitted back to the location (211) from which the item was sent. The device sending the item is preferably a wireless device, and more preferably a cellular or other telephone (210). The device receiving the translation is also preferably a wireless device, and more preferably a cellular or other telephone, and may advantageously be the same device as the sending device. The item of information (212) preferably comprises a sentence of human of speech having at least ten words, and the translation is a written expression of the sentence. All of the steps of transmitting the item of information, executing the program code, and transmitting the translated information preferably occurs in less than 60 seconds of elapsed time.
    Type: Grant
    Filed: March 21, 2012
    Date of Patent: January 20, 2015
    Assignee: Ulloa Research Limited Liability Company
    Inventor: Robert D. Fish
  • Patent number: 8938388
    Abstract: Maintaining and supplying a plurality of speech models is provided. A plurality of speech models and metadata for each speech model are stored. A query for a speech model is received from a source. The query includes one or more conditions. The speech model with metadata most closely matching the supplied one or more conditions is determined. The determined speech model is provided to the source. A refined speech model is received from the source, and the refined speech model is stored.
    Type: Grant
    Filed: July 9, 2012
    Date of Patent: January 20, 2015
    Assignee: International Business Machines Corporation
    Inventors: Bin Jia, Ying Liu, E. Feng Lu, Jia Wu, Zhen Zhang
  • Patent number: 8930191
    Abstract: Methods, systems, and computer readable storage medium related to operating an intelligent digital assistant are disclosed. A user request is received, the user request including at least a speech input received from a user. In response to the user request, (1) an echo of the speech input based on a textual interpretation of the speech input, and (2) a paraphrase of the user request based at least in part on a respective semantic interpretation of the speech input are presented to the user.
    Type: Grant
    Filed: March 4, 2013
    Date of Patent: January 6, 2015
    Assignee: Apple Inc.
    Inventors: Thomas Robert Gruber, Harry Joseph Saddler, Adam John Cheyer, Dag Kittlaus, Christopher Dean Brigham, Richard Donald Giuli, Didier Rene Guzzoni, Marcello Bastea-Forte
  • Patent number: 8930187
    Abstract: An apparatus for utilizing textual data and acoustic data corresponding to speech data to detect sentiment may include a processor and memory storing executable computer code causing the apparatus to at least perform operations including evaluating textual data and acoustic data corresponding to voice data associated with captured speech content. The computer program code may further cause the apparatus to analyze the textual data and the acoustic data to detect whether the textual data or the acoustic data includes one or more words indicating at least one sentiment of a user that spoke the speech content. The computer program code may further cause the apparatus to assign at least one predefined sentiment to at least one of the words in response to detecting that the word(s) indicates the sentiment of the user. Corresponding methods and computer program products are also provided.
    Type: Grant
    Filed: January 3, 2012
    Date of Patent: January 6, 2015
    Assignee: Nokia Corporation
    Inventors: Imre Attila Kiss, Joseph Polifroni, Francois Mairesse, Mark Adler
  • Patent number: 8924197
    Abstract: Disclosed are systems, methods, and computer readable media for converting a natural language query into a logical query. The method embodiment comprises receiving a natural language query and converting the natural language query using an extensible engine to generate a logical query, the extensible engine being linked to the toolkit and knowledge base. In one embodiment, a natural language query can be processed in a domain independent method to generate a logical query.
    Type: Grant
    Filed: October 30, 2007
    Date of Patent: December 30, 2014
    Assignee: Semantifi, Inc.
    Inventors: Sreenivasa Rao Pragada, Viswanath Dasari, Abhijit A Patil
  • Patent number: 8924214
    Abstract: A method for detecting and recognizing speech is provided that remotely detects body motions from a speaker during vocalization with one or more radar sensors. Specifically, the radar sensors include a transmit aperture that transmits one or more waveforms towards the speaker, and each of the waveforms has a distinct wavelength. A receiver aperture is configured to receive the scattered radio frequency energy from the speaker. Doppler signals correlated with the speaker vocalization are extracted with a receiver. Digital signal processors are configured to develop feature vectors utilizing the vocalization Doppler signals, and words associated with the feature vectors are recognized with a word classifier.
    Type: Grant
    Filed: June 7, 2011
    Date of Patent: December 30, 2014
    Assignee: The United States of America, as represented by the Secretary of the Navy
    Inventors: Jefferson M Willey, Todd Stephenson, Hugh Faust, James P. Hansen, George J Linde, Carol Chang, Justin Nevitt, James A Ballas, Thomas Herne Crystal, Vincent Michael Stanford, Jean W. De Graaf
  • Patent number: 8918406
    Abstract: A method of processing content files may include receiving the content file, employing processing circuitry to determine an identity score of a source of a portion of at least a portion the content file, to determine a word score based for the content file and to determine a metadata score for the content file, determining a composite priority score based on the identity score, the word score and the metadata score, and associating the composite priority score with the content file for electronic provision of the content file together with the composite priority score to a human analyst.
    Type: Grant
    Filed: December 14, 2012
    Date of Patent: December 23, 2014
    Assignee: Second Wind Consulting LLC
    Inventor: Donna Rober
  • Patent number: 8918316
    Abstract: The content of a media program is recognized by analyzing its audio content to extract therefrom prescribed features, which are compared to a database of features associated with identified content. The identity of the content within the database that has features that most closely match the features of the media program being played is supplied as the identity of the program being played. The features are extracted from a frequency domain version of the media program by a) filtering the coefficients to reduce their number, e.g., using triangular filters; b) grouping a number of consecutive outputs of triangular filters into segments; and c) selecting those segments that meet prescribed criteria, such as those segments that have the largest minimum segment energy with prescribed constraints that prevent the segments from being too close to each other. The triangular filters may be log-spaced and their output may be normalized.
    Type: Grant
    Filed: July 29, 2003
    Date of Patent: December 23, 2014
    Assignee: Alcatel Lucent
    Inventors: Jan I Ben, Christopher J Burges, Madjid Sam Mousavi, Craig R. Nohl
  • Patent number: 8918320
    Abstract: An apparatus for generating a review based in part on detected sentiment may include a processor and memory storing executable computer code causing the apparatus to at least perform operations including determining a location(s) of the apparatus and a time(s) that the location(s) was determined responsive to capturing voice data of speech content associated with spoken reviews of entities. The computer program code may further cause the apparatus to analyze textual and acoustic data corresponding to the voice data to detect whether the textual or acoustic data includes words indicating a sentiment(s) of a user speaking the speech content. The computer program code may further cause the apparatus to generate a review of an entity corresponding to a spoken review(s) based on assigning a predefined sentiment to a word(s) responsive to detecting that the word indicates the sentiment of the user. Corresponding methods and computer program products are also provided.
    Type: Grant
    Filed: January 3, 2012
    Date of Patent: December 23, 2014
    Assignee: Nokia Corporation
    Inventors: Mark Adler, Imre Attila Kiss, Francois Mairesse, Joseph Polifroni
  • Patent number: 8914290
    Abstract: Method and apparatus that dynamically adjusts operational parameters of a text-to-speech engine in a speech-based system. A voice engine or other application of a device provides a mechanism to alter the adjustable operational parameters of the text-to-speech engine. In response to one or more environmental conditions, the adjustable operational parameters of the text-to-speech engine are modified to increase the intelligibility of synthesized speech.
    Type: Grant
    Filed: May 18, 2012
    Date of Patent: December 16, 2014
    Assignee: Vocollect, Inc.
    Inventors: James Hendrickson, Debra Drylie Scott, Duane Littleton, John Pecorari, Arkadiusz Slusarczyk
  • Patent number: 8903727
    Abstract: A machine, system and method for user-guided teaching and modifications of voice commands and actions to be executed by a conversational learning system. The machine includes a system bus for communicating data and control signals received from the conversational learning system to a computer system, a vehicle data and control bus for connecting devices and sensors in the machine, a bridge module for connecting the vehicle data and control bus to the system bus, machine subsystems coupled to the vehicle data and control bus having a respective user interface for receiving a voice command or input signal from a user, a memory coupled to the system bus for storing action command sequences learned for a new voice command and a processing unit coupled to the system bus for automatically executing the action command sequences learned when the new voice command is spoken.
    Type: Grant
    Filed: March 6, 2013
    Date of Patent: December 2, 2014
    Assignee: Nuance Communications, Inc.
    Inventors: Liam David Comerford, Mahesh Viswanathan
  • Patent number: 8892446
    Abstract: Methods, systems, and computer readable storage medium related to operating an intelligent digital assistant are disclosed. A user request is received, the user request including at least a speech input received from the user. The user request is processed to obtain a representation of user intent, where the representation of user intent associates the user request with a task flow operationalizing a requested task, and the task flow is operable to invoke a plurality of services each supporting functions according to a respective plurality of service parameters. Based on the representation of user intent, one or more relevant task parameters are identified from a plurality of task parameters of the task flow. A subset of the plurality of services are selectively invoked during execution of the task flow, where the selectively invoked subset of the plurality of services support functions according to the identified one or more relevant task parameters.
    Type: Grant
    Filed: December 21, 2012
    Date of Patent: November 18, 2014
    Assignee: Apple Inc.
    Inventors: Adam John Cheyer, Didier Rene Guzzoni, Thomas Robert Gruber, Christopher Dean Brigham
  • Patent number: 8874442
    Abstract: Device, system, and method of liveness detection using voice biometrics. For example, a method comprises: generating a first matching score based on a comparison between: (a) a voice-print from a first text-dependent audio sample received at an enrollment stage, and (b) a second text-dependent audio sample received at an authentication stage; generating a second matching score based on a text-independent audio sample; and generating a liveness score by taking into account at least the first matching score and the second matching score.
    Type: Grant
    Filed: April 17, 2013
    Date of Patent: October 28, 2014
    Assignee: Nuance Communications, Inc.
    Inventors: Almog Aley-Raz, Nir Moshe Krause, Michael Itzhak Salmon, Ran Yehoshua Gazit
  • Patent number: 8874440
    Abstract: A speech detection apparatus and method are provided. The speech detection apparatus and method determine whether a frame is speech or not using feature information extracted from an input signal. The speech detection apparatus may estimate a situation related to an input frame and determine which feature information is required for speech detection for the input frame in the estimated situation. The speech detection apparatus may detect a speech signal using dynamic feature information that may be more suitable to the situation of a particular frame, instead of using the same feature information for each and every frame.
    Type: Grant
    Filed: April 16, 2010
    Date of Patent: October 28, 2014
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Chi-youn Park, Nam-hoon Kim, Jeong-mi Cho
  • Patent number: 8831942
    Abstract: A method is provided for identifying a gender of a speaker. The method steps include obtaining speech data of the speaker, extracting vowel-like speech frames from the speech data, analyzing the vowel-like speech frames to generate a feature vector having pitch values corresponding to the vowel-like frames, analyzing the pitch values to generate a most frequent pitch value, determining, in response to the most frequent pitch value being between a first pre-determined threshold and a second pre-determined threshold, an output of a male Gaussian Mixture Model (GMM) and an output of a female GMM using the pitch values as inputs to the male GMM and the female GMM, and identifying the gender of the speaker by comparing the output of the male GMM and the output of the female GMM based on a pre-determined criterion.
    Type: Grant
    Filed: March 19, 2010
    Date of Patent: September 9, 2014
    Assignee: Narus, Inc.
    Inventor: Antonio Nucci
  • Patent number: 8825482
    Abstract: Consumer electronic devices have been developed with enormous information processing capabilities, high quality audio and video outputs, large amounts of memory, and may also include wired and/or wireless networking capabilities. Additionally, relatively unsophisticated and inexpensive sensors, such as microphones, video camera, GPS or other position sensors, when coupled with devices having these enhanced capabilities, can be used to detect subtle features about users and their environments. A variety of audio, video, simulation and user interface paradigms have been developed to utilize the enhanced capabilities of these devices. These paradigms can be used separately or together in any combination. One paradigm automatically creating user identities using speaker identification. Another paradigm includes a control button with 3-axis pressure sensitivity for use with game controllers and other input devices.
    Type: Grant
    Filed: September 15, 2006
    Date of Patent: September 2, 2014
    Assignee: Sony Computer Entertainment Inc.
    Inventors: Gustavo Hernandez-Abrego, Xavier Menendez-Pidal, Steven Osman, Ruxin Chen, Rishi Deshpande, Care Michaud-Wideman, Richard Marks, Eric Larsen, Xiaodong Mao
  • Patent number: 8818816
    Abstract: A voice recognition device includes a voice input unit 11 for inputting a voice of an uttered button name to convert the voice into an electric signal, a voice recognition processing unit 12 for performing a voice recognition process according to a sound signal sent thereto, as the electric signal, from the voice input unit, a button candidate detecting unit 13 for detecting, as a button candidate, a button having a button name which partially matches a voice recognition result acquired by the voice recognition processing unit, a display control unit 15 for, when a plurality of candidate buttons are detected by the button candidate detecting unit, producing a screen showing a state in which at least one of the plurality of button candidates is selected, and a display unit 16 for displaying the screen produced by the display control unit.
    Type: Grant
    Filed: April 23, 2009
    Date of Patent: August 26, 2014
    Assignee: Mitsubishi Electric Corporation
    Inventors: Yuzuru Inoue, Takayoshi Chikuri, Yuki Furumoto
  • Patent number: 8818807
    Abstract: This invention describes methods for implementing human speech recognition. The methods described here are of using sub-events that are sounds between spaces (typically a fully spoken word) that is then compared with a library of sub-events. All sub-events are packaged with it's own speech recognition function as individual units. This invention illustrates how this model can be used as a Large Vocabulary Speech Recognition System.
    Type: Grant
    Filed: May 24, 2010
    Date of Patent: August 26, 2014
    Inventor: Darrell Poirier
  • Patent number: 8812326
    Abstract: A computer-driven device assists a user in self-regulating speech control of the device. The device processes an input signal representing human speech to compute acoustic signal quality indicators indicating conditions likely to be problematic to speech recognition, and advises the user of those conditions.
    Type: Grant
    Filed: August 6, 2013
    Date of Patent: August 19, 2014
    Assignee: Promptu Systems Corporation
    Inventors: Naren Chittar, Vikas Gulati, Matthew Pratt, Harry Printz
  • Patent number: 8812318
    Abstract: One-to-many comparisons of callers' voice prints with known voice prints to identify any matches between them. When a customer communicates with a particular entity, such as a customer service center, the system makes a recording of the real-time call including both the customer's and agent's voices. The system segments the recording to extract at least a portion of the customer's voice to create a customer voice print, and it formats the segmented voice print for network transmission to a server. The server compares the customer's voice print with multiple known voice prints to determine any matches, meaning that the customer's voice print and one of the known voice prints are likely from the same person. The identification of any matches can be used for a variety of purposes, such as determining whether to authorize a transaction requested by the customer.
    Type: Grant
    Filed: February 6, 2012
    Date of Patent: August 19, 2014
    Assignee: III Holdings 1, LLC
    Inventors: Vicki Broman, Vernon Marshall, Seshasayee Bellamkonda, Marcel Leyva, Cynthia Hanson
  • Patent number: 8804973
    Abstract: In an example signal clustering apparatus, a feature of a signal is divided into segments. A first feature vector of each segment is calculated, the first feature vector having has a plurality of elements corresponding to each reference model. A value of an element attenuates when a feature of the segment shifts from a center of a distribution of the reference model corresponding to the element. A similarity between two reference models is calculated. A second feature vector of each segment is calculated, the second feature vector having a plurality of elements corresponding to each reference model. A value of an element is a weighted sum and segments of second feature vectors of which the plurality of elements are similar values are clustered to one class.
    Type: Grant
    Filed: March 19, 2012
    Date of Patent: August 12, 2014
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Makoto Hirohata, Kazunori Imoto, Hisashi Aoki
  • Patent number: 8805685
    Abstract: Disclosed herein are systems, methods, and tangible computer readable-media for detecting synthetic speaker verification. The method comprises receiving a plurality of speech samples of the same word or phrase for verification, comparing each of the plurality of speech samples to each other, denying verification if the plurality of speech samples demonstrate little variance over time or are the same, and verifying the plurality of speech samples if the plurality of speech samples demonstrates sufficient variance over time. One embodiment further adds that each of the plurality of speech samples is collected at different times or in different contexts. In other embodiments, variance is based on a pre-determined threshold or the threshold for variance is adjusted based on a need for authentication certainty. In another embodiment, if the initial comparison is inconclusive, additional speech samples are received.
    Type: Grant
    Filed: August 5, 2013
    Date of Patent: August 12, 2014
    Assignee: AT&T Intellectual Property I, L.P.
    Inventor: Horst J. Schroeter