Systems Using Speaker Recognizers (epo) Patents (Class 704/E17.003)
  • Patent number: 11824642
    Abstract: Systems and methods are provided for use in provisioning a biometric image template to a card device. One example portable communication device includes a processor, a wireless communication interface coupled to the processor, and an input device coupled to the processor. The processor is configured to transmit to a card device, via the wireless communication interface, a certificate associated with the portable communication device, whereby the portable communication device is authenticated by the card device, receive from the card device, via the wireless communication interface, a certificate associated with the card device, and authenticate the card device based on the received certificate. After the card device is authenticated, the processor is configured to capture, via the at least one input device, a biometric image of the user and provision, via the wireless communication interface, the captured biometric image to the card device.
    Type: Grant
    Filed: January 31, 2022
    Date of Patent: November 21, 2023
    Assignee: MASTERCARD INTERNATIONAL INCORPORATED
    Inventors: Ashfaq Kamal, Sumeet Bhatt, Robert D. Reany
  • Patent number: 11783841
    Abstract: A method and system for secure speaker authentication between a caller device and a first device using an authentication server are provided. The system comprises extracting features into a feature matrix from an incoming audio call; generating a partial i-vector, wherein the partial i-vector includes a first low-order statistic; sending the partial i-vector to the authentication server; and receiving from the authentication server a match score generated based on a full i-vector and another i-vector being stored on the authentication server, wherein the full i-vector is generated from the partial i-vector.
    Type: Grant
    Filed: March 15, 2021
    Date of Patent: October 10, 2023
    Assignee: ILLUMA LABS INC.
    Inventor: Milind Borkar
  • Patent number: 11751260
    Abstract: An apparatus and method for establishing a pairing between a mobile device and a smart wearable device is provided. The method includes generating and assigning a unique code for each of the detected plurality of smart wearable devices respectively. The unique code is shared with all the connected smart devices/devices in the communicative range of the mobile device. The smart devices (having a display) receive the corresponding unique codes and display the code on their respective displays. A request to establish a pairing connection is generated in response to an activity of a user in respect of the at least one smart wearable device and a communication is established between the at least one smart wearable device and the mobile device displaying the same unique code.
    Type: Grant
    Filed: June 25, 2018
    Date of Patent: September 5, 2023
    Inventors: Sudheendra Shantharam, Prashant Adkoli
  • Patent number: 11699445
    Abstract: A system and method for improving T-matrix training for speaker recognition, comprising receiving an audio input, divisible into a plurality of audio frames including at least an audio sample of a human speaker; generating for each audio frame a feature vector; generating for a first plurality of feature vectors centered statistics of at least a zero order and a first order; generating a first i-vector, the first i-vector representing the human speaker; and generating an optimized T-matrix training sequence computation, based on at least the first i-vector.
    Type: Grant
    Filed: March 15, 2021
    Date of Patent: July 11, 2023
    Assignee: ILLUMA LABS INC.
    Inventor: Milind Borkar
  • Patent number: 11521622
    Abstract: A system and method for efficient universal background model (UBM) training for speaker recognition, including: receiving an audio input, divisible into a plurality of audio frames, wherein at least a first audio frame of the plurality of audio frames includes an audio sample having a length above a first threshold extracting at least one identifying feature from the first audio frame and generating a feature vector based on the at least one identifying feature; generating an optimized training sequence computation based on the feature vector and a Gaussian Mixture Model (GMM), wherein the GMM is associated with a plurality of components, wherein each of the plurality of components is defined by a covariance matrix, a mean vector, and a weight vector; and updating any of the associated components of the GMM based on the generated optimized training sequence computation.
    Type: Grant
    Filed: October 27, 2020
    Date of Patent: December 6, 2022
    Assignee: ILLUMA Labs Inc.
    Inventor: Milind Borkar
  • Patent number: 10789449
    Abstract: An electronic device for distinguishing between fingerprint feature points and non-fingerprint feature points and a method for the same are provided. A plurality of feature points are extracted from a fingerprint input image, and non-fingerprint feature points are distinguished from fingerprint feature points and deleted. Accordingly, the electronic device for distinguishing between fingerprint feature points and non-fingerprint feature points and the method for the same can prevent non-fingerprint feature points from existing in a fingerprint registration template of a user, so as to protect fingerprint recognition from being affected by non-fingerprint feature points, thereby increasing safety and accuracy of identity authentication.
    Type: Grant
    Filed: November 19, 2018
    Date of Patent: September 29, 2020
    Assignee: EGIS TECHNOLOGY INC.
    Inventors: Yuan-Lin Chiang, Yu-Chun Cheng
  • Patent number: 10685075
    Abstract: An input audio signal is monitored at an electronic digital assistant of an electronic computing device. The assistant determines that the audio signal includes audio from multiple speaking parties, identifies first and second parties of the multiple parties and roles of the parties, accesses a database of multi-party interaction templates and selects, as a function of the roles of the parties, a particular multi-party interaction template that matches the roles, and then generates, based on a detected content of the audio signal and the particular multi-party interaction template, a database query to retrieve information responsive to or supplemental to content of the audio signal. The assistant then causes the query to be performed results of the query provided to the first party, the second party, a dispatcher or administrator associated with the parties, or a computer-aided-dispatch system associated with the parties.
    Type: Grant
    Filed: April 11, 2018
    Date of Patent: June 16, 2020
    Assignee: MOTOROLA SOLUTIONS, INC.
    Inventors: Alejandro G. Blanco, Hao Zeng, Eric Johnson
  • Publication number: 20140118472
    Abstract: In one embodiment, a method includes receiving requests to join a conference from a plurality of user devices proximate a first endpoint. The requests include a username. The method also includes receiving an audio signal for the conference from the first endpoint. The first endpoint is operable to capture audio proximate the first endpoint. The method also includes transmitting the audio signal to a second endpoint, remote from the first endpoint. The method also includes identifying, by a processor, an active speaker proximate the first endpoint based on information received from the plurality of user devices.
    Type: Application
    Filed: October 31, 2012
    Publication date: May 1, 2014
    Inventors: Yanghua Liu, Weidong Chen, Biren Gandhi, Raghurama Bhat, Joseph Fouad Khouri, John Joseph Houston, Brian Thomas Toombs
  • Publication number: 20140046664
    Abstract: Methods and apparatuses for secure device pairing are disclosed. In one example, a user voice is received simultaneously at a first device and a second device to pair the devices.
    Type: Application
    Filed: August 13, 2012
    Publication date: February 13, 2014
    Applicant: PLANTRONICS, INC.
    Inventors: Shantanu Sarkar, Cary Bran, Joe Burton, Joseph Stachula
  • Publication number: 20140039897
    Abstract: A system for automated adaptation and improvement of speaker authentication in a voice biometric system environment, comprising a speech sample collector, a target selector, a voice analyzer, a voice data modifier, and a call flow creator. The speech sample collector retrieves speech samples from a database of enrolled participants in a speaker authentication system. The target selector selects target users that will be used to test the speaker authentication system. The voice analyzer extracts a speech component data set from each of the speech samples. The call flow creator creates a plurality of call flows for testing the speaker authentication system, each call flow being either an impostor call flow or a legitimate call flow. The call flows created by the call flow creator are used to test the speaker authentication system.
    Type: Application
    Filed: August 6, 2012
    Publication date: February 6, 2014
    Inventor: Alok Kulkarni
  • Publication number: 20140039892
    Abstract: In one embodiment, a human interactive proof portal 140 may use a biometric input to determine whether a user is a standard user or a malicious actor. The human interactive proof portal 140 may receive an access request 302 for an online data service 122 from a user device 110. The human interactive proof portal 140 may send a proof challenge 304 to the user device 110 for presentation to a user. The human interactive proof portal 140 may receive from the user device 110 a proof response 306 having a biometric metadata description 430 based on a biometric input from the user.
    Type: Application
    Filed: August 2, 2012
    Publication date: February 6, 2014
    Applicant: Microsoft Corporation
    Inventors: Chad Mills, Robert Sim, Scott Laufer, Sung Chung
  • Publication number: 20130289991
    Abstract: According to a present invention embodiment, a system utilizes a voice tag to automatically tag one or more entities within a social media environment, and comprises a computer system including at least one processor. The system analyzes the voice tag to identify one or more entities, where the voice tag includes voice signals providing information pertaining to one or more entities. One or more characteristics of each identified entity are determined based on the information within the voice tag. One or more entities appropriate for tagging within the social media environment are determined based on the characteristics and user settings within the social media environment of the identified entities, and automatically tagged. Embodiments of the present invention further include a method and computer program product for utilizing a voice tag to automatically tag one or more entities within a social media environment in substantially the same manner described above.
    Type: Application
    Filed: April 30, 2012
    Publication date: October 31, 2013
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Bhavani K. Eshwar, Martin A. Oberhofer, Sushain Pandit
  • Publication number: 20130166300
    Abstract: An electronic device includes a voice recognition analyzing module, a manipulation identification module, and a manipulating module. The voice recognition analyzing module is configured to recognize and analyze a voice of a user. The manipulation identification module is configured to, using the analyzed voice, identify an object on a screen and identify a requested manipulation associated with the object. The manipulating module is configured to perform the requested manipulation.
    Type: Application
    Filed: September 12, 2012
    Publication date: June 27, 2013
    Applicant: KABUSHIKI KAISHA TOSHIBA
    Inventors: Sachie Yokoyama, Hideki Tsutsui
  • Publication number: 20130144619
    Abstract: Techniques for ability enhancement are described. Some embodiments provide an ability enhancement facilitator system (“AEFS”) configured to enhance voice conferencing among multiple speakers. In one embodiment, the AEFS receives data that represents utterances of multiple speakers who are engaging in a voice conference with one another. The AEFS then determines speaker-related information, such as by identifying a current speaker, locating an information item (e.g., an email message, document) associated with the speaker, or the like. The AEFS then informs a user of the speaker-related information, such as by presenting the speaker-related information on a display of a conferencing device associated with the user.
    Type: Application
    Filed: January 23, 2012
    Publication date: June 6, 2013
    Inventors: Richard T. Lord, Robert W. Lord, Nathan P. Myhrvold, Clarence T. Tegreene, Roderick A. Hyde, Lowell L. Wood, JR., Muriel Y. Ishikawa, Victoria Y.H. Wood, Charles Whitmer, Paramvir Bahl, Doughlas C. Burger, Ranveer Chandra, William H. Gates, III, Paul Holman, Jordin T. Kare, Craig J. Mundie, Tim Paek, Desney S. Tan, Lin Zhong, Matthew G. Dyor
  • Publication number: 20120271633
    Abstract: The present invention provides an interactive device which allows quick utterance recognition results and sequential output thereof and which diminishes a recognition rate decrease even if user's utterance is divided by a short interval into frames for quick decision. The interactive device: sets a recognition section for voice recognition; performs voice recognition for the recognition section; when the voice recognition includes a key phrase, determines response actions corresponding thereto; and executes the response actions. The interactive device repeatedly updates the set recognition terminal point to a frame which is the predetermined time length ahead of the set recognition terminal point to set a plurality of recognition sections. The interactive device performs voice recognition for each recognition section.
    Type: Application
    Filed: April 19, 2012
    Publication date: October 25, 2012
    Applicant: HONDA MOTOR CO., LTD.
    Inventors: Yuichi Yoshida, Taku Osada
  • Patent number: 8286071
    Abstract: A computer program product, for automatically editing a medical record transcription, resides on a computer-readable medium and includes computer-readable instructions for causing a computer to obtain a first medical transcription of a dictation, the dictation being from medical personnel and concerning a patient, analyze the first medical transcription for presence of a first trigger phrase associated with a first standard text block, determine that the first trigger phrase is present in the first medical transcription if an actual phrase in the first medical transcription corresponds with the first trigger phrase, and insert the first standard text block into the first medical transcription.
    Type: Grant
    Filed: June 29, 2006
    Date of Patent: October 9, 2012
    Assignee: eScription, Inc.
    Inventors: Roger S. Zimmerman, Paul Egerman, Robert G. Titemore, George Zavaliagkos
  • Publication number: 20120232903
    Abstract: The invention relates to a kitchen and/or domestic appliance comprising input means, which are connected to a voice-recognition system, for acoustic operator commands. The invention is characterised in that means for executing command-dependent actions are provided and that the voice-recognition system is used to identify and check the authorisation of a user.
    Type: Application
    Filed: March 15, 2012
    Publication date: September 13, 2012
    Applicant: ELECTROLUX PROFESSIONAL SPA
    Inventors: Claudio Cenedese, Dragan Raus, Omero Tuzzi, Maurizio Ugel, Ennio Pippia
  • Publication number: 20120209608
    Abstract: A mobile communication terminal apparatus and method are capable of recognizing an input voice of a user and executing an application related to the recognized voice. The apparatus includes a voice input unit to receive a first input voice; a voice recognition unit to acquire first voice instruction information based on the first input voice; a voice control table acquiring unit to acquire a first voice control table comprising the first voice instruction information and first icon position information; and an application execution unit to execute a first application based on the first icon position information included in the first voice control table. The method for registering voice instruction information includes acquiring voice instruction information for a selected application; acquiring execution information of the selected application; generating a voice control table comprising the execution information, and the voice instruction information; and storing the voice control table.
    Type: Application
    Filed: September 29, 2011
    Publication date: August 16, 2012
    Applicant: PANTECH CO., LTD.
    Inventor: Chang-Dae LEE
  • Publication number: 20120101822
    Abstract: A biometric speaker-identification apparatus is disclosed that generates ordered speaker-identity candidates for a probe based on prototypes. Probe match scores are clustered, and templates that correspond to clusters having top M probe match scores are compared with the prototypes to obtain template-prototype match scores. The probe is also compared with the prototypes, and those templates corresponding to template-prototype match scores that are nearest to probe-prototype match scores are selected as speaker-identity candidates. The speaker-identity candidates are ordered based on their similarity to the probe.
    Type: Application
    Filed: October 25, 2010
    Publication date: April 26, 2012
    Applicant: LOCKHEED MARTIN CORPORATION
    Inventor: Jonathan J. DINERSTEIN
  • Publication number: 20120072218
    Abstract: Disclosed are systems, methods, and computer readable media for tracking a person of interest. The method embodiment comprises identifying a person of interest, capturing a voiceprint of the person of interest, comparing a received voiceprint of a caller with the voiceprint of the person of interest, and tracking the caller if the voiceprint of the caller is a substantial match to the voiceprint of the person of interest.
    Type: Application
    Filed: November 28, 2011
    Publication date: March 22, 2012
    Applicant: AT&T Intellectual Property II, L.P.
    Inventors: Gustavo De Los Reyes, Sanjay Macwan
  • Publication number: 20120051525
    Abstract: A method includes registering a voice of a party in order to provide voice verification for communications with an entity. A call is received from a party at a voice response system. The party is prompted for information and verbal communication spoken by the party is captured. A voice model associated with the party is created by processing the captured verbal communication spoken by the party and is stored. The identity of the party is verified and a previously stored voice model of the party, registered during a previous call from the party, is updated. The creation of the voice model is imperceptible to the party.
    Type: Application
    Filed: November 9, 2011
    Publication date: March 1, 2012
    Applicant: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventor: Mazin GILBERT
  • Publication number: 20120010884
    Abstract: Systems and methods are disclosed for displaying electronic multimedia content to a user. One computer-implemented method for manipulating electronic multimedia content includes generating, using a processor, a speech model and at least one speaker model of an individual speaker. The method further includes receiving electronic media content over a network; extracting an audio track from the electronic media content; and detecting speech segments within the electronic media content based on the speech model. The method further includes detecting a speaker segment within the electronic media content and calculating a probability of the detected speaker segment involving the individual speaker based on the at least one speaker model.
    Type: Application
    Filed: June 9, 2011
    Publication date: January 12, 2012
    Inventors: Peter F. Kocks, Guoning Hu, Ping-Hao Wu
  • Publication number: 20110307256
    Abstract: A system enables voice authentication via a network. The system may include an intelligent voice response engine operatively coupled to the network for receiving transaction or access requests from a plurality of telecommunications devices over the network. A speech recognition and verification services engine may be operatively coupled to the network and a database may be operatively coupled to the speech recognition and verification services engine for storing user voice print profiles. The speech recognition and verification services engine may receive a speaker verification call from the intelligent voice response engine and perform speaker verification on the received speaker verification call based on the stored user voice print profiles. The speech recognition and verification services engine may generate a verification score based upon results of the speaker verification.
    Type: Application
    Filed: August 19, 2011
    Publication date: December 15, 2011
    Applicant: VERIZON BUSINESS GLOBAL LLC
    Inventor: Paul T. SCHULTZ
  • Publication number: 20110231310
    Abstract: Various systems and methods are described for conducting and staging money transfer transactions from a vehicle. The systems and methods may include using a wireless transmitter on board the vehicle for wirelessly communicating with a money transfer system. Also, a user-input device integrated with the vehicle's navigation system, entertainment system, and/or climate system may be used. Using the wireless transmitter, the user-input device may interface with the money transfer system. Also, one or more identity authentication devices may be integrated with the vehicle that collect information to authenticate a user's identity.
    Type: Application
    Filed: March 18, 2010
    Publication date: September 22, 2011
    Applicant: The Western Union Company
    Inventors: Melanie Roberts, Dianna Lyons
  • Publication number: 20110205331
    Abstract: When a communication terminal determines that restriction processing is to be applied based on a sound input through the communication terminal or a user instruction, the communication terminal restricts output of conference information received from a counterpart terminal.
    Type: Application
    Filed: February 24, 2011
    Publication date: August 25, 2011
    Inventor: Yoshinaga KATO
  • Publication number: 20110208524
    Abstract: This is directed to processing voice inputs received by an electronic device. In particular, this is directed to receiving a voice input and identifying the user providing the voice input. The voice input can be processed using a subset of words from a library used to identify the words or phrases of the voice input. The particular subset can be selected such that voice inputs provided by the user are more likely to include words from the subset. The subset of the library can be selected using any suitable approach, including for example based on the user's interests and words that relate to those interests. For example, the subset can include one or more words related to media items selected by the user for storage on the electronic device, names of the user's contacts, applications or processes used by the user, or any other words relating to the user's interactions with the device.
    Type: Application
    Filed: February 25, 2010
    Publication date: August 25, 2011
    Applicant: Apple Inc.
    Inventor: Allen P. Haughay
  • Publication number: 20110193726
    Abstract: An emotive advisory system for use by one or more occupants of an automotive vehicle includes a computer. The computer is configured to receive input indicative of an operating state of the vehicle. The computer determines at least one of a need to provide time information to an occupant based on the operating state of the vehicle and an occupant request to provide time information to the occupant. The computer generates data representing an avatar having an appearance, and data representing a spoken statement for the avatar. The spoken statement provides time information to the occupant in spoken dialog based on at least one of the need and the request. The computer outputs data representing the avatar for visual display, and data representing the statement for the avatar for audio play.
    Type: Application
    Filed: February 9, 2010
    Publication date: August 11, 2011
    Applicant: FORD GLOBAL TECHNOLOGIES, LLC
    Inventors: Steven Joseph Szwabowski, Perry Robinson MacNeille
  • Publication number: 20110184735
    Abstract: Embodiments are disclosed that relate to the use of identity information to help avoid the occurrence of false positive speech recognition events in a speech recognition system. One embodiment provides a method comprising receiving speech recognition data comprising a recognized speech segment, acoustic locational data related to a location of origin of the recognized speech segment as determined via signals from the microphone array, and confidence data comprising a recognition confidence value, and also receiving image data comprising visual locational information related to a location of each person in an image. The acoustic locational data is compared to the visual locational data to determine whether the recognized speech segment originated from a person in the field of view of the image sensor, and the confidence data is adjusted depending on this determination.
    Type: Application
    Filed: January 22, 2010
    Publication date: July 28, 2011
    Applicant: MICROSOFT CORPORATION
    Inventors: Jason Flaks, Dax Hawkins, Christian Klein, Mitchell Stephen Dernis, Tommer Leyvand, Ali M. Vassigh, Duncan McKay
  • Publication number: 20110150270
    Abstract: System, methods, and apparatuses. A method includes receiving a voice input from a user, the voice input corresponding to printed information on a mail piece. The method includes performing a voice recognition process on the voice input to produce a voice address result, the voice recognition process using voice attributes from a database, and performing an optical character recognition process on an image of the printed information to produce recognized text and a confidence value. The method includes storing updated voice attributes corresponding to the voice input and recognized text in the database when the confidence value meets a first threshold, and combining the recognized text and the voice address result to produce a combined OCR result. The method includes sending the combined OCR result to a sorting system that sorts the mail piece according to the combined OCR result.
    Type: Application
    Filed: December 14, 2010
    Publication date: June 23, 2011
    Inventors: Michael D. Carpenter, Dale E. Redford
  • Publication number: 20110154385
    Abstract: An application for a television has a detector capable of determining the identity and/or presence of at least one viewer in a viewing area of the television. In response to viewers entering and leaving the viewing area of the television, the television adjusts its operation based upon settings for the currently present viewers (e.g., enables channels, content, etc).
    Type: Application
    Filed: December 22, 2009
    Publication date: June 23, 2011
    Applicant: VIZIO, INC.
    Inventors: William Pat Price, Peter Jay Schwartz, Marcus P. Apitz
  • Publication number: 20110112838
    Abstract: A system and method are provided to authenticate a voice in a frequency domain. A voice in the time domain is transformed to a signal in the frequency domain. The first harmonic is set to a predetermined frequency and the other harmonic components are equalized. Similarly, the amplitude of the first harmonic is set to a predetermined amplitude, and the harmonic components are also equalized. The voice signal is then filtered. The amplitudes of each of the harmonic components are then digitized into bits to form at least part of a voice ID. In another system and method, a voice is authenticated in a time domain. The initial rise time, initial fall time, second rise time, second fall time and final oscillation time are digitized into bits to form at least part of a voice ID. The voice IDs are used to authenticate a user's voice.
    Type: Application
    Filed: November 10, 2009
    Publication date: May 12, 2011
    Applicant: Research In Motion Limited
    Inventor: Sasan Adibi
  • Publication number: 20110077941
    Abstract: Techniques for assigning a spoken tag in a telecom web platform are provided. The techniques include receiving a spoken tag, comparing the spoken tag to a set of one or more template tags, if the spoken tag is a match to a template tag, assigning the spoken tag and updating frequency of the tag in the set of one or more template tags, and if the spoken tag is not a match to a template tag, assigning the spoken tag and registering the spoken tag as a new tag in the set of one or more template tags.
    Type: Application
    Filed: September 30, 2009
    Publication date: March 31, 2011
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Kuntal Dey, Anupam Jain, Arun Kumar, Natwar Modani, Amit Anil Nanavati, Nitendra Rajput
  • Publication number: 20110058028
    Abstract: An information processing apparatus includes: a storage unit which stores person relationship information representing the relationship between multiple people as a subject in a storage medium; an acquisition unit which acquires image data generated by imaging people as a subject; a detection unit which detects each person in an image based on image data acquired by the acquisition unit; a specification unit which specifies each person detected from the image by the detection unit; and a determination unit which determines the relationship between the multiple people detected from the image by the detection unit, wherein, when at least one person from among the multiple people detected from the image by the detection unit is specified and another person is unable to be specified, the specification unit specifies another person on the basis of the relationship between the multiple people determined by the determination unit and the person relationship information stored in the storage medium.
    Type: Application
    Filed: September 1, 2010
    Publication date: March 10, 2011
    Applicant: Sony Corporation
    Inventor: Yasuo Sakai
  • Publication number: 20110035217
    Abstract: A system and method for detecting a refrain in an audio file having vocal components. The method and system includes generating a phonetic transcription of a portion of the audio file, analyzing the phonetic transcription and identifying a vocal segment in the generated phonetic transcription that is repeated frequently. The method and system further relate to the speech-driven selection based on similarity of detected refrain and user input.
    Type: Application
    Filed: October 19, 2010
    Publication date: February 10, 2011
    Applicant: Harman International Industries, Incorporated
    Inventors: Franz S. Gerl, Daniel Willett, Raymond Brueckner
  • Publication number: 20100291901
    Abstract: Voice recording used for authentication is transmitted to the voice portal of the service provider not by way of the voice channel, but rather by way of a data channel. In this connection, the voice recording is sent not synchronous to speech and subject to loss, but rather asynchronously and loss-free, in an extra data package (for example advantageously as a Multimedia Messaging Service (MMS) data package). For this purpose, the resources that are available as standard items in most mobile phone terminals, such as digital voice recording and MMS transmission, can be utilized. Preferably, the subscriber's identification module or SIM card in the mobile phone terminal can store and implement the corresponding control program.
    Type: Application
    Filed: July 23, 2010
    Publication date: November 18, 2010
    Applicant: VoiceCash IP GmbH
    Inventor: Hermann Geupel
  • Publication number: 20100286983
    Abstract: An operation control apparatus and method of controlling a plurality of operationally connected voice recognition-enabled systems, each having reciprocal control operational states corresponding to an enabled/disabled state.
    Type: Application
    Filed: September 14, 2009
    Publication date: November 11, 2010
    Inventor: Chung Bum Cho
  • Publication number: 20100223057
    Abstract: System and process for audio authentication of an individual or speaker including a processor for decomposing an audio signal received at the sensor into vectors representative of the speaker to be authenticated for transforming the super-vector V of the speaker resulting from the concatenation of the vectors associated with the said speaker into binary data 1001100 . . . 0 taking as an input the mean super-vector M resulting from the mean super-vector, and comparing the super-vector V of the speaker with the mean super-vector M, the said binary data thus obtained being transmitted to a module for extracting the speaker authentication taking as an input the public keys Kpub(1) in order to authenticate the speaker and/or to generate a cryptographic key associated with the speaker.
    Type: Application
    Filed: December 22, 2009
    Publication date: September 2, 2010
    Applicant: Thales
    Inventors: François Capman, Sandra Marcello, Jean Martinelli
  • Publication number: 20100145504
    Abstract: An address label rework station according to the invention includes a conveyor which can simultaneously transport a series of spaced parcels along a conveyor path in a substantially horizontal stream. The station includes a camera, A work space, a microphone is usable by an operator in the work space, a targeting device directing a human operator in the work space. Station also includes a sensor system, a labeler and a printer to print shipping information. The station also includes a computer configured to receive and recognize image and voice data, generate a label with a recognized shipping address, and control the conveyer to apply a new label to the parcel by the labeler at the position selected using the targeting device.
    Type: Application
    Filed: December 3, 2009
    Publication date: June 10, 2010
    Inventors: Dale E. Redford, Michael D. Carpenter, James M. Pippin
  • Publication number: 20100102941
    Abstract: The invention relates to a mobile communication device and an input device provided for operating said communication device by means of which input operations necessary for using the communication device may be carried out. The aim of the invention is to provide solutions by means of which it is possible, in particular for relatively compact mobile electronic devices, in particular communication devices and digital cameras to carry out said input operations in a manner which is particularly easy to coordinate for the user.
    Type: Application
    Filed: March 26, 2008
    Publication date: April 29, 2010
    Inventors: Wolfgang Richter, Roland Aubauer
  • Publication number: 20100017209
    Abstract: The present invention provides a random voiceprint certification system comprises a training system, a random cipher generator, and a testing system, which is employed to process training or testing operation for the input raw voice data. In training voice, the training system obtains an appointment voiceprint feature model parameter groups from the input raw voice data. From the appointment voiceprint feature model parameter groups several voiceprint characteristic units are obtained and at least one reference voiceprint password, which is for the testing system to carry out the voice testing operation is built. In processing testing voice, the random cipher generator generates randomly at least one reference voiceprint password from the voiceprint characteristic units of the appointment voiceprint feature model parameter groups to build the random voiceprint cipher lock. The present invention generates randomly one or several reference voiceprint passwords.
    Type: Application
    Filed: December 6, 2007
    Publication date: January 21, 2010
    Inventors: Kun-Lang Yu, Yen-Chieh Ouyang
  • Publication number: 20100004979
    Abstract: A method and system to provide Broadcasting networks revenue by Monetizing “shout outs and holler outs”. Consumer Participants Register with Internet domain network and then issue shout outs on participating broadcasting (radio, TV . . . etc) networks. The consumers are then paid a fee for mentioning an advertiser during the shout out, if conducted in correct format.
    Type: Application
    Filed: June 27, 2009
    Publication date: January 7, 2010
    Applicants: Anyspace Systems Inc.
    Inventor: Tyrone S. Brunson
  • Publication number: 20090290698
    Abstract: A method of transmitting voice data in a communication network and a device for transmitting such voice data, as well as a method of receiving voice data in a communication network and a device for receiving such data are provided. The voice data is comprised in a data packet transmitted by the transmitting device to the receiving device.
    Type: Application
    Filed: May 23, 2008
    Publication date: November 26, 2009
    Applicant: SONY ERICSSON MOBILE COMMUNICATIONS AB
    Inventors: Jonas LUNDGREN, Mikael SALMEN, Christian EHRENBORG
  • Publication number: 20090276218
    Abstract: A method for optimizing message transmission and decoding comprises: reading data from a memory of an originating device, the data comprising information regarding the originating device; encoding the data by converting the data to a subset of words having a ranked recognition accuracy higher than the remainder of words; transmitting the encoded data from the originating device to a receiving system audibly as words via a telephone connection; utilizing a voice recognition software to recognize the words; decoding the words back to the data; and taking a predetermined action based on the data.
    Type: Application
    Filed: April 28, 2009
    Publication date: November 5, 2009
    Inventors: Patrick Alan Hussey, Maryellen Abreu
  • Publication number: 20090276223
    Abstract: An administration method and system. The method includes receiving by a computing system, a telephone call from an administrator. The computing system presents an audible menu associated with a plurality of computers to the administrator. The computing system receives from the administrator, an audible selection for a computer from the audible menu. The computing system receives from the administrator, an audible verbal command for performing a maintenance operation on the computer. The computing system executes the maintenance operation on the computer. The computing system receives from the computer, confirmation data indicating that the maintenance operation has been completed. The computing system converts the confirmation data into an audible verbal message. The computing system transmits the second audible verbal message to the administrator.
    Type: Application
    Filed: May 1, 2008
    Publication date: November 5, 2009
    Inventors: Peeyush Jaiswal, Naveen Narayan
  • Publication number: 20090248414
    Abstract: An apparatus includes unit acquiring speaker information including a first duration of a speaker and a name specified by name specifying information used to indicate a name, and acquiring the first duration as a first period, unit acquiring a second period including an utterance, unit extracting, if the second period is included in the first period, a first amount that characterizes a speaker, and associating the first amount with a name corresponding to the first period, unit creating speaker models from amounts, unit acquiring, from the content information, a third duration as an duration to be recognized, unit extracting, if the second period is included in the third period, a second amount that characterizes a speaker, unit calculating degrees of similarity between amounts of speaker models and the second amount, and unit recognizing a name of a speaker model which satisfies a set condition of the degrees as a performer.
    Type: Application
    Filed: March 23, 2009
    Publication date: October 1, 2009
    Applicant: KABUSHIKI KAISHA TOSHIBA
    Inventors: Taishi Shimomori, Tatsuya Uehara
  • Publication number: 20090217324
    Abstract: A method, system and program product for customizing presentation of television content to a specific viewer. The method includes obtaining an Internet Protocol address for a set-top box provided for an interactive television system viewed by a viewer for establishing a physical location of the television system, authenticating, using a voice recognition system installed on the set-top box provided, an identity of the viewer within a viewing area of the interactive television system, recording television viewing habits of the viewer, selecting a customized television content based on the viewer identity authenticated and the viewing habits recorded for the viewer and delivering the customized television content selected to the viewer, based on the viewer's physical location.
    Type: Application
    Filed: February 26, 2008
    Publication date: August 27, 2009
    Applicant: International Business Machines Corporation
    Inventor: Michael L. Massimi
  • Publication number: 20090147995
    Abstract: An information processing apparatus includes information input units which inputs observation information in a real space; an event detection unit which generates event information including estimated position and identification information on users existing in the actual space through analysis of the input information; and an information integration processing unit which sets hypothesis probability distribution data regarding user position and user identification information and generates analysis information including the user position information through hypothesis update and sorting out based on the event information, in which the event detection unit detects a face area from an image frame input from an image information input unit, extracts face attribute information from the face area, and calculates and outputs a face attribute score corresponding to the extracted face attribute information to the information integration processing unit, and the information integration processing unit applies the face
    Type: Application
    Filed: December 5, 2008
    Publication date: June 11, 2009
    Inventors: Tsutomu SAWADA, Takeshi Ohashi
  • Publication number: 20090112589
    Abstract: A multi-party communication enhancer includes an audio data input adapted to receive voice data associated with a plurality of communication participants. A participant identifier included in the multi-party communication enhancer is adapted to distinguish the voice of a number of communication participants as represented within the received voice data. A cue generator, also included in the multi-party communication enhancer, is operable to generate a cue for each distinguished voice, with the generated cue being outputted in association with the corresponding distinguished voice.
    Type: Application
    Filed: October 30, 2007
    Publication date: April 30, 2009
    Inventors: Per Olof Hiselius, Jonas Magnus Andersson
  • Patent number: RE42868
    Abstract: A method and apparatus accesses a database where entries are linked to at least two sets of patterns. One or more patterns of a first set of patterns are recognized within a received signal. The recognized patterns are used to identify entries and compile a list of patterns in a second set of patterns to which those entries are also linked. The list is then used to recognize a second received signal. The received signals may, for example, be voice signals or signals indicating the origin or destination of the received signals.
    Type: Grant
    Filed: October 25, 1995
    Date of Patent: October 25, 2011
    Assignee: Cisco Technology, Inc.
    Inventors: David J. Attwater, Steven J. Whittaker, Francis J. Scahill, Alison D. Simons
  • Patent number: RE45262
    Abstract: A navigation system and method involving wireless communications technology and speech processing technology is presented. In accordance with an embodiment of the invention, the navigation system includes a subscriber unit communicating with a service provider. The subscriber unit includes a global positioning system mechanism to determine subscriber position information and a speech processing mechanism to receive destination information spoken by a subscriber. The subscriber unit transmits the subscriber position and destination information to the service provider, which gathers navigation information, including a map and a route from the subscriber position to the specified destination. The service provider transmits the navigation information to the subscriber unit. The subscriber unit conveys the received navigation information to the subscriber via an output mechanism, such as a speech synthesis unit or a graphical display.
    Type: Grant
    Filed: December 2, 2004
    Date of Patent: November 25, 2014
    Assignee: Intel Corporation
    Inventor: Christopher R. Wiener