Recognition Of Special Voice Characteristics, E.g., For Use In A Lie Detector; Recognition Of Animal Voices, Etc. (epo) Patents (Class 704/E17.002)
  • Patent number: 11938880
    Abstract: Systems and methods for detecting low impact collisions for a vehicle (100). The system includes at least one sensor (99, 110, 111, 115, 120-123, 125-136, 140, 141) and an electronic controller (150). The electronic controller (150) is configured to receive sensor data from the sensor (99, 110, 111, 115, 120-123, 125-136, 140, 141) and determine one or more features of the sensor data received from the at least one sensor. The electronic controller (150) is further configured to determine if a collision has occurred based upon the one or more features of the sensor data, and take at least one action in response to determining that the collision has occurred.
    Type: Grant
    Filed: October 31, 2019
    Date of Patent: March 26, 2024
    Assignee: Robert Bosch GmbH
    Inventors: Robert Jones, Michael Vincent Masserant, Rameez Ahmad, Ulrich Christian Michelfeit, Dean Eiger
  • Patent number: 11861417
    Abstract: A peripheral information acquisition unit (121) acquires information relating to a first application (13a) running in a terminal, information relating to control of the terminal, or information that can be acquired from a sensor included in the terminal, as peripheral information. The peripheral information acquired by the peripheral information acquisition unit (121) is accumulated in a peripheral information accumulation unit (122). A dialogue interface unit (11) accepts input of information from a user and outputs information to the user. When the peripheral information accumulated in the peripheral information accumulation unit (122) and information input to the dialogue interface unit (11) satisfy a predetermined condition, a scenario control unit (123) causes the dialogue interface unit (11) to output information relating to execution of a second application (14) that is associated with the condition in advance.
    Type: Grant
    Filed: October 9, 2019
    Date of Patent: January 2, 2024
    Assignee: Nippon Telegraph and Telephone Corporation
    Inventors: Makoto Komiyama, Takeshi Masuda, Akira Kataoka, Masashi Tadokoro, Hidetaka Koya
  • Patent number: 11854574
    Abstract: A method of operating a chatbot based on a companion animal's emotion by using a user terminal, according to an embodiment of the present disclosure includes receiving voice information and activity amount information from an TOT device when receiving a chatting value from a user; calculating a basic emotion variable, a situation variable, and a behavior pattern variable based on the voice information and the activity amount information; and searching for an answer value corresponding to the chatting value under conditions of the basic emotion variable, the situation variable, and the behavior pattern variable to output the searched answer value, wherein the TOT device includes a microphone and a behavior sensor and generates the voice information and the activity amount information by detecting a crying sound and a behavior of the companion animal wearing the TOT device.
    Type: Grant
    Filed: August 27, 2021
    Date of Patent: December 26, 2023
    Assignee: PETPULS LAB INC.
    Inventor: Yun Ock Jang
  • Patent number: 11741560
    Abstract: Described are medias, systems, and computer-implemented methods to detect improper residency status by performing a data mining task to data source to detect one or more improper residency indicia, applying a machine learning algorithm to identify an initial candidate, calculating a probability that the initial candidate has an improper residency status, and validating the detection. Further described are medias, systems, and computer-implemented methods to detect improper occupancy tax status. Further described are medias, systems, and computer-implemented methods to detect an improper homeowner exemption.
    Type: Grant
    Filed: September 9, 2020
    Date of Patent: August 29, 2023
    Assignee: DECKARD TECHNOLOGIES, INC.
    Inventors: Gregory G. Rose, Nickolas Del Pego, Craig Brown
  • Patent number: 11620978
    Abstract: An automatic interpretation method performed by a correspondent terminal communicating with an utterer terminal includes receiving, by a communication unit, voice feature information about an utterer and an automatic translation result, obtained by automatically translating a voice uttered in a source language by the utterer in a target language, from the utterer terminal and performing, by a sound synthesizer, voice synthesis on the basis of the automatic translation result and the voice feature information to output a personalized synthesis voice as an automatic interpretation result. The voice feature information about the utterer includes a hidden variable including a first additional voice result and a voice feature parameter and a second additional voice feature, which are extracted from a voice of the utterer.
    Type: Grant
    Filed: August 11, 2020
    Date of Patent: April 4, 2023
    Assignee: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Seung Yun, Sang Hun Kim, Min Kyu Lee
  • Patent number: 11582336
    Abstract: A system and method for authenticating a caller may include receiving an incoming call from the caller, determining a gender of the caller, and selecting, based on the determined gender, to search for the caller in one of: a watchlist of untrustworthy female callers, and a watchlist of untrustworthy male callers.
    Type: Grant
    Filed: August 4, 2021
    Date of Patent: February 14, 2023
    Assignee: Nice Ltd.
    Inventors: Guy Earman, Matan Keret, Roman Frenkel
  • Patent number: 11410675
    Abstract: An animal audio signature may be collected by a solar powered sound collection device. The solar powered collection device may use a supercapacitor to store power. The animal audio signature may be compared to a database of known animal audio signatures. The database may contain one or more identities for each of the known animal audio signatures. A known animal audio signature that matches the collected animal audio signature may be identified. An identity associated with the known animal audio signature may be transmitted to a data repository over a 5G wireless network.
    Type: Grant
    Filed: July 24, 2020
    Date of Patent: August 9, 2022
    Assignee: International Business Machines Corporation
    Inventors: Jeb R. Linton, Jonathan Samn, Poojitha Bikki, Naeem Altaf
  • Patent number: 11373219
    Abstract: A computerized matching system enable users to utilize video clips to promote and/or find a person/people, place, or consumer item. The system is a recommendation service, embodied preferably in a mobile app, suggesting potential matches for its users. These suggestions are based upon the information and data entered by the users in their user profiles, as well as their Ideal Match Criteria (i.e. what it is they are seeking). The suggestions are also based upon algorithms that analyze and learn from available rating history data to make predictions on potential matches, utilizing rating trends of that user (i.e. the prior ratings entered by that user during their usage of the system while viewing other videos) as well as the rating trends of other users who have rated the same videos in a similar manner (herein referred to as Similar Rating Groups, or SRGs).
    Type: Grant
    Filed: August 12, 2016
    Date of Patent: June 28, 2022
    Inventor: Eric Koenig
  • Patent number: 10950222
    Abstract: The present disclosure is directed to a system for generating a personality assessment that uses multimodal behavioral signal processing technology and machine learning prediction technology. This system takes a video as input, processes it through an artificial intelligence software built for extracting hundreds of behavioral features, and consequently generates an accurate and reliable personality assessment with its machine-learning predictive software. The personality assessment is based on the five-factor model (FFM), also known as the big 5 personality traits.
    Type: Grant
    Filed: April 24, 2019
    Date of Patent: March 16, 2021
    Assignee: YOBS TECHNOLOGIES, INC.
    Inventors: Raphael Louis Jean Danilo, Prathmesh A. Gat, Aamir K. Goriawala
  • Patent number: 10908677
    Abstract: A system in a vehicle includes one or more sensors configured to obtain input from an occupant of the vehicle. The system also includes a controller in communication with the one or more sensors, wherein the controller is configured to, receive the input from the one or more sensors, determine a mood of the occupant utilizing at least the input, and output a solution to a vehicle display in response to the mood of the occupant.
    Type: Grant
    Filed: March 25, 2019
    Date of Patent: February 2, 2021
    Assignee: DENSO INTERNATIONAL AMERICA, INC.
    Inventors: Vikas Upmanue, Yu Zhang, Bilal Alasry, Te-Ping Kang, Jordan Necovski, Sean Bleicher, Doua Vang, Eleanor Duke, Nicholaus Spunar
  • Patent number: 10636418
    Abstract: Methods, apparatus, and computer readable media are described related to automated assistants that proactively incorporate, into human-to-computer dialog sessions, unsolicited content of potential interest to a user. In various implementations, in an existing human-to-computer dialog session between a user and an automated assistant, it may be determined that the automated assistant has responded to all natural language input received from the user. Based on characteristic(s) of the user, information of potential interest to the user or action(s) of potential interest to the user may be identified. Unsolicited content indicative of the information of potential interest to the user or the action(s) may be generated and incorporated by the automated assistant into the existing human-to-computer dialog session.
    Type: Grant
    Filed: March 22, 2017
    Date of Patent: April 28, 2020
    Assignee: GOOGLE LLC
    Inventors: Ibrahim Badr, Zaheed Sabur, Vladimir Vuskovic, Adrian Zumbrunnen, Lucas Mirelmann
  • Patent number: 10592733
    Abstract: Systems and methods are provided providing a spoken dialog system. Output is provided from a spoken dialog system that determines audio responses to a person based on recognized speech content from the person during a conversation between the person and the spoken dialog system. Video data associated with the person interacting with the spoken dialog system is received. A video engagement metric is derived from the video data, where the video engagement metric indicates a level of the person's engagement with the spoken dialog system.
    Type: Grant
    Filed: May 19, 2017
    Date of Patent: March 17, 2020
    Assignee: Educational Testing Service
    Inventors: Vikram Ramanarayanan, David Suendermann-Oeft, Patrick Lange, Alexei V. Ivanov, Keelan Evanini, Yao Qian, Eugene Tsuprun, Hillary R. Molloy
  • Patent number: 10566010
    Abstract: Methods, systems, and related products that provide emotion-sensitive responses to user's commands and other utterances received at an utterance-based user interface. Acknowledgements of user's utterances are adapted to the user and/or the user device, and emotions detected in the user's utterance that have been mapped from one or more emotion features extracted from the utterance. In some examples, extraction of a user's changing emotion during a sequence of interactions is used to generate a response to a user's uttered command. In some examples, emotion processing and command processing of natural utterances are performed asynchronously.
    Type: Grant
    Filed: April 20, 2018
    Date of Patent: February 18, 2020
    Assignee: SPOTIFY AB
    Inventors: Daniel Bromand, David Gustafsson, Richard Mitic, Sarah Mennicken
  • Patent number: 10089298
    Abstract: At least one computer-mediated communication produced by or received by an author is collected and parsed to identify categories of information within it. The categories of information are processed with at least one analysis to quantify at least one type of information in each category. A first output communication is generated regarding the at least one computer-mediated communication, describing the psychological state, attitudes or characteristics of the author of the communication. A second output communication is generated when a difference between the quantification of at least one type of information for at least one category and a reference for the at least one category is detected involving a psychological state, attitude or characteristic of the author to which a responsive action should be taken.
    Type: Grant
    Filed: February 21, 2018
    Date of Patent: October 2, 2018
    Assignee: Stroz Friedberg LLC
    Inventor: Eric D. Shaw
  • Patent number: 10052769
    Abstract: A humanoid robot is provided, the robot being capable of holding a dialog with at least one user, the dialog using two modes of voice recognition, one open and the other closed, the closed mode being defined by a concept characterizing a dialog sequence. The dialog may also be influenced by events that are neither speech nor a text. The robot is capable of executing behaviors and generating expressions and emotions. It has the advantage of considerably reducing programming time and latency of execution of dialog sequences, providing a fluency and naturalness close to human dialogs.
    Type: Grant
    Filed: April 3, 2013
    Date of Patent: August 21, 2018
    Assignee: SOFTBANK ROBOTICS EUROPE
    Inventors: David Houssin, Gwennael Gate
  • Patent number: 9953078
    Abstract: An emotion estimator of a dialogue control device estimates a dialogue partner's emotion based on the dialogue partner's facial expression or voice prosody. A dialogue controller controls the dialogue with the dialogue partner using a topic selected based on whether the emotion estimated by the emotion estimator is good or bad. Specifically, the dialogue controller controls the dialogue with the dialogue partner based on a topic map created by a topic mapper and including a topic that the dialogue partner likes.
    Type: Grant
    Filed: June 30, 2016
    Date of Patent: April 24, 2018
    Assignee: CASIO COMPUTER CO., LTD.
    Inventor: Yoshihiro Kawamura
  • Patent number: 9807559
    Abstract: Systems, methods, apparatuses, and computer program products are described for implementing a digital personal assistant. The digital personal assistant is capable of determining that a user has asked a question or made a statement that is intended to engage with a persona of the digital personal assistant. In response to determining that the user has asked such a question or made such a statement, the digital personal assistant provides a response thereto by displaying or playing back a multimedia object associated with a popular culture reference within or by a user interface of the digital personal assistant. Additionally or alternatively, in response to determining that the user has asked such a question or made such a statement, the digital personal assistant provides the response thereto by generating or playing back speech that comprises an impersonation of a voice of a person associated with the popular culture reference.
    Type: Grant
    Filed: June 25, 2014
    Date of Patent: October 31, 2017
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Lee Dicks Clark, Deborah B. Harrison, Susan Hendrich, David Gardner, Sogol Malekzadeh, Catherine L. Maritan, Melissa Lim, Mary P. Czerwinski, Ran Gilad-Bachrach
  • Patent number: 8407055
    Abstract: An information processing apparatus includes an obtaining unit that obtains meta-information concerning content; a predicting unit that predicts an emotion of a user who is viewing the content from the meta-information obtained by the obtaining unit; and a recognizing unit that recognizes an emotion of the user using the emotion predicted by the predicting unit and user information acquired from the user.
    Type: Grant
    Filed: August 3, 2006
    Date of Patent: March 26, 2013
    Assignee: Sony Corporation
    Inventors: Yasuharu Asano, Noriyuki Yamamoto
  • Publication number: 20120232899
    Abstract: A system and method for identification of a speaker by phonograms of oral speech is disclosed. Similarity between a first phonogram of the speaker and a second, or sample, phonogram is evaluated by matching formant frequencies in referential utterances of a speech signal, where the utterances for comparison are selected from the first phonogram and the second phonogram. Referential utterances of speech signals are selected from the first phonogram and the second phonogram, where the referential utterances include formant paths of at least three formant frequencies. The selected referential utterances including at least two identical formant frequencies are compared therebetween. Similarity of the compared referential utterances from matching other formant frequencies is evaluated, where similarity of the phonograms is determined from evaluation of similarity of all the compared referential utterances.
    Type: Application
    Filed: March 23, 2012
    Publication date: September 13, 2012
    Applicant: Obschestvo s orgranichennoi otvetstvennost'yu "Centr Rechevyh Technologij"
    Inventor: Sergey Lvovich Koval
  • Publication number: 20120221336
    Abstract: A computer implemented method, data processing system, apparatus and computer program product for determining current behavioral, psychological and speech styles characteristics of a speaker in a given situation and context, through analysis of current speech utterances of the speaker. The analysis calculates different prosodic parameters of the speech utterances, consisting of unique secondary derivatives of the primary pitch and amplitude speech parameters, and compares these parameters with pre-obtained reference speech data, indicative of various behavioral, psychological and speech styles characteristics. The method includes the formation of the classification speech parameters reference database, as well as the analysis of the speaker's speech utterances in order to determine the current behavioral, psychological and speech styles characteristics of the speaker in the given situation.
    Type: Application
    Filed: May 7, 2012
    Publication date: August 30, 2012
    Applicant: VOICESENSE LTD.
    Inventors: Yoav DEGANI, Yishai ZAMIR
  • Publication number: 20120134478
    Abstract: One-to-many comparisons of callers' voice prints with known voice prints to identify any matches between them. When a customer communicates with a particular entity, such as a customer service center, the system makes a recording of the real-time call including both the customers and agent's voices. The system segments the recording to extract at least a portion of the customer's voice to create a customer voice print, and it formats the segmented voice print for network transmission to a server. The server compares the customer's voice print with multiple known voice prints to determine any matches, meaning that the customer's voice print and one of the known voice prints are likely from the same person. The identification of any matches can be used for a variety of purposes, such as determining whether to authorize a transaction requested by the customer.
    Type: Application
    Filed: February 6, 2012
    Publication date: May 31, 2012
    Applicant: American Express Travel Related Services Company, Inc.
    Inventors: Vicki Broman, Vernon Marshall, Seshasayee Bellamkonda, Marcel Leyva, Cynthia Hanson
  • Publication number: 20120130714
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media relating to speaker verification. In one aspect, a system receives a first user identity from a second user, and, based on the identity, accesses voice characteristics. The system randomly generates a challenge sentence according to a rule and/or grammar, based on the voice characteristics, and prompts the second user to speak the challenge sentence. The system verifies that the second user is the first user if the spoken challenge sentence matches the voice characteristics. In an enrollment aspect, the system constructs an enrollment phrase that covers a minimum threshold of unique speech sounds based on speaker-distinctive phonemes, phoneme clusters, and prosody. Then user utters the enrollment phrase and extracts voice characteristics for the user from the uttered enrollment phrase.
    Type: Application
    Filed: November 24, 2010
    Publication date: May 24, 2012
    Applicant: AT&T Intellectual Property I, L.P.
    Inventors: Ilija Zeljkovic, Taniya Mishra, Amanda Stent, Ann K. Syrdal, Jay Wilpon
  • Patent number: 8179284
    Abstract: The present invention relates to an alarm management system intended to be carried onboard an aircraft. More precisely, the invention is aimed at improving the certainty level relating to the integrity of the announcements of faults or information intended for the pilot and based on voice syntheses. For this purpose, the present invention proposes a device and a method for detecting the digital origin of an analog signal providing a validity signal (VAL) enabling the voice announcements made to the pilot to be rendered secure.
    Type: Grant
    Filed: June 25, 2009
    Date of Patent: May 15, 2012
    Assignee: Thales
    Inventors: Christian Pitot, Jacques Phelippeau, Philippe Bieth
  • Publication number: 20120116763
    Abstract: A voice data analyzing device comprises speaker model deriving means which derives speaker models as models each specifying character of voice of each speaker from voice data including a plurality of utterances to each of which a speaker label as information for identifying a speaker has been assigned and speaker co-occurrence model deriving means which derives a speaker co-occurrence model as a model representing the strength of co-occurrence relationship among the speakers from session data obtained by segmenting the voice data in units of sequences of conversation by use of the speaker models derived by the speaker model deriving means.
    Type: Application
    Filed: June 3, 2010
    Publication date: May 10, 2012
    Applicant: NEC CORPORATION
    Inventor: Takafumi Koshinaka
  • Publication number: 20110276323
    Abstract: The illustrative embodiments described herein provide systems and methods for authenticating a speaker. In one embodiment, a method includes receiving reference speech input including a reference passphrase to form a reference recording, and receiving test speech input including a test passphrase to form a test recording. The method includes determining whether the test passphrase matches the reference passphrase, and determining whether one or more voice features of the speaker of the test passphrase matches one or more voice features of the speaker of the reference passphrase. The method authenticates the speaker of the test speech input in response to determining that the reference passphrase matches the test passphrase and that one or more voice features of the speaker of the test passphrase matches one or more voice features of the speaker of the reference passphrase.
    Type: Application
    Filed: May 6, 2010
    Publication date: November 10, 2011
    Applicant: Senam Consulting, Inc.
    Inventor: Serge Olegovich Seyfetdinov
  • Publication number: 20110258548
    Abstract: A method for creating a multimedia presentation, the method including determining the physical presence of at least one individual, selecting at least one multimedia object associated with the at least one individual whose physical presence has been determined, creating a multimedia presentation using the at least one multimedia object associated with the at least one individual, determining the physical presence of additional individuals other than the at least one individual, selecting at least one multimedia object associated with each of the additional individuals whose physical presence has been determined, updating the multimedia presentation using the at least one multimedia object associated with each of the additional individuals, and displaying the multimedia presentation.
    Type: Application
    Filed: April 16, 2010
    Publication date: October 20, 2011
    Applicant: CANON KABUSHIKI KAISHA
    Inventors: Nabil Mohammed Abujbara, Richard Kirkor Yardumian
  • Publication number: 20110246198
    Abstract: The present invention refers to a method for verifying the identity of a speaker based on the speakers voice comprising the steps of: a) receiving a voice utterance; b) using biometric voice data to verify (10) that the speakers voice corresponds to the speaker the identity of which is to be verified based on the received voice utterance; and c) verifying (12, 13) that the received voice utterance is not falsified, preferably after having verified the speakers voice; d) accepting (16) the speakers identity to be verified in case that both verification steps give a positive result and not accepting (15) the speakers identity to be verified if any of the verification steps give a negative result. The invention further refers to a corresponding computer readable medium and a computer.
    Type: Application
    Filed: December 10, 2008
    Publication date: October 6, 2011
    Inventors: Marta Sánchez Asenjo, Alfredo Gutiérrez Navarro, Alberto Martin De Los Santos De Las Heras, Marta Garcia Gomar
  • Publication number: 20110153317
    Abstract: An apparatus for wireless communications includes a processing system. The processing system is configured to receive an input sound stream of a user, split the input sound stream into a plurality of frames, classify each of the frames as one selected from the group consisting of a non-speech frame and a speech frame, determine a pitch of each of the frames in a subset of the speech frames, and identify a gender of the user from the determined pitch. To determine the pitch, the processing system is configured to filter the speech frames to compute an error signal, compute an autocorrelation of the error signal, find a maximum autocorrelation value, and set the pitch to an index of the maximum autocorrelation value.
    Type: Application
    Filed: December 23, 2009
    Publication date: June 23, 2011
    Applicant: QUALCOMM INCORPORATED
    Inventors: Yinian Mao, Gene Marsh
  • Patent number: 7940914
    Abstract: A system, method and article of manufacture are provided for detecting emotion using statistics. First, a database is provided. The database has statistics including human associations of voice parameters with emotions. Next, a voice signal is received. At least one feature is extracted from the voice signal. Then the extracted voice feature is compared to the voice parameters in the database. An emotion is selected from the database based on the comparison of the extracted voice feature to the voice parameters and is then output.
    Type: Grant
    Filed: April 10, 2001
    Date of Patent: May 10, 2011
    Assignee: Accenture Global Services Limited
    Inventor: Valery A. Petrushin
  • Publication number: 20100312559
    Abstract: A method of playing pictures comprises the steps of: receiving (11) a voice message; extracting (12) a key feature from the voice message; selecting (13) pictures by matching the key feature with pre-stored picture information; generating (14) a picture-voice sequence by integrating the selected pictures and the voice message; and playing (15) the picture-voice sequence. An electronic apparatus comprises a processing unit for implementing the different steps of the method.
    Type: Application
    Filed: December 11, 2008
    Publication date: December 9, 2010
    Applicant: Koninklijke Philips Electronics N.V.
    Inventors: Sheng Jin, Xin Chen, Yang Peng, Ningjiang Chen, Yunji Xia
  • Publication number: 20100211387
    Abstract: Computer implemented speech processing is disclosed. First and second voice segments are extracted from first and second microphone signals originating from first and second microphones. The first and second voice segments correspond to a voice sound originating from a common source. An estimated source location is generated based on a relative energy of the first and second voice segments and/or a correlation of the first and second voice segments. A determination whether the voice segment is desired or undesired may be made based on the estimated source location.
    Type: Application
    Filed: February 2, 2010
    Publication date: August 19, 2010
    Applicant: Sony Computer Entertainment Inc.
    Inventor: Ruxin Chen
  • Publication number: 20100198598
    Abstract: A method for recognizing a speaker of an utterance in a speech recognition system is disclosed. A likelihood score for each of a plurality of speaker models for different speakers is determined. The likelihood score indicating how well the speaker model corresponds to the utterance. For each of the plurality of speaker models, a probability that the utterance originates from that speaker is determined. The probability is determined based on the likelihood score for the speaker model and requires the estimation of a distribution of likelihood scores expected based at least in part on the training state of the speaker.
    Type: Application
    Filed: February 4, 2010
    Publication date: August 5, 2010
    Applicant: NUANCE COMMUNICATIONS, INC.
    Inventors: Tobias Herbig, Franz Gerl
  • Publication number: 20100191519
    Abstract: A runtime framework and authoring tool are provided for enabling linguistic experts to author text normalization maps and grammar libraries without requiring high level of technical or programming skills. Authors define or select terminals, map the terminals, and define rules for the mapping. The tool enables an author to validate their work, by executing the map in the same way the recognition engine does, causing consistency in results from authoring to user operations. The runtime is used by the speech engines and by the tools to provide consistent normalization for supported scenarios.
    Type: Application
    Filed: January 28, 2009
    Publication date: July 29, 2010
    Applicant: Microsoft Corporation
    Inventors: Rachel I. Morton, Nicholas J. Gedge, Heiko W. Rahmel
  • Publication number: 20100164685
    Abstract: A system, apparatus, method and article to establish device connections are described. The apparatus may include an input device to receive identification information for a selected target device. The apparatus may also include a connection module operative to send a connection request message including the identification information to one or more target devices and receive a connection response message from the selected target device. The connection response message includes confirmation that the identification information corresponds to local identification information stored on the selected target device. The apparatus may also include a communication module operative to establish a wireless connection between the mobile computing device and the selected target device based on the corresponding identification information. Other embodiments are described and claimed.
    Type: Application
    Filed: December 31, 2008
    Publication date: July 1, 2010
    Inventors: Trevor Pering, Roy Want, Kenton Lyons, Shivani A. Sud, Barbara Rosario
  • Publication number: 20100153110
    Abstract: A voice recognition system and method of a mobile communication device. The mobile communication device has a storage device which stores voice templates and characteristics of each of the voice templates. The method recognizes voice data form a sound input, calculates a similarity ratio between characteristics of the sound input and the characteristics of each of the voice templates, and sorts the voice templates according to the similarity ratio in a list of text symbols representing the voice templates. The list of text symbols can be selected as a proper text input by a user.
    Type: Application
    Filed: August 26, 2009
    Publication date: June 17, 2010
    Applicant: CHI MEI COMMUNICATION SYSTEMS, INC.
    Inventor: TANG-YU CHANG
  • Publication number: 20100145695
    Abstract: An apparatus for context awareness includes: a voice-based recognition unit that recognizes a user's emotional state on the basis of a voice signal; a motion-based recognition unit that recognizes the user's emotional state on the basis of a motion; a position recognition unit that recognizes a location where the user is positioned; and a mergence-recognition unit that recognizes a user's context by analyzing the recognition results of the voice-based recognition unit, the motion-based recognition unit, and the position recognition unit. Accordingly, it is possible to rapidly and accurately accidents or dangerous contexts caused to a user.
    Type: Application
    Filed: December 7, 2009
    Publication date: June 10, 2010
    Applicant: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Younggiu JUNG, Munsung HAN, Dongwan RYOO, Changseok BAE, Jintae KIM
  • Publication number: 20100138223
    Abstract: An object of the present invention is to allow classification of sequentially input speech signals with good accuracy based on similarity of speakers and environments by using a realistic memory use amount, a realistic processing speed, and an on-line operation. A speech classification probability calculation means 103 calculates a probability (probability of classification into each cluster) that a latest one of the speech signals (speech data) belongs to each cluster based on a generative model which is a probability model. A parameter updating means 107 successively estimates parameters that define the generative model based on the probability of classification of the speech data into each cluster calculated by the speech classification probability calculation means 103 (in FIG. 1).
    Type: Application
    Filed: March 13, 2008
    Publication date: June 3, 2010
    Inventor: Takafumi Koshinaka
  • Publication number: 20100085415
    Abstract: A method for efficiently determining and displaying pertinent information determined from multiple input and calculated parameters associated with a videoconference call. The method for efficiently determining and displaying this personal information is performed using input from the user at an endpoint and calculated information throughout the videoconference to present personal information, about the currently speaking person, to all participants. Videoconferencing systems are typically used by multiple people at multiple locations. The method of this disclosure allows for more user interaction and knowledge transfer amongst the participants. By sharing information between the different locations participants are more aware of who is speaking at any given time and the importance to be applied to what that particular person is saying.
    Type: Application
    Filed: October 2, 2008
    Publication date: April 8, 2010
    Applicant: POLYCOM, INC
    Inventor: MOHAMMED RAHMAN
  • Publication number: 20090254342
    Abstract: A method for detecting barge-in in a speech dialogue system comprising determining whether a speech prompt is output by the speech dialogue system, and detecting whether speech activity is present in an input signal based on a time-varying sensitivity threshold of a speech activity detector and/or based on speaker information, where the sensitivity threshold is increased if output of a speech prompt is determined and decreased if no output of a speech prompt is determined. If speech activity is detected in the input signal, the speech prompt may be interrupted or faded out. A speech dialogue system configured to detect barge-in is also disclosed.
    Type: Application
    Filed: March 31, 2009
    Publication date: October 8, 2009
    Applicant: Harman Becker Automotive Systems GmbH
    Inventors: Markus Buck, Franz Gerl, Tim Haulick, Tobias Herbig, Gerhard Uwe Schmidt, Matthias Schulz
  • Publication number: 20090198492
    Abstract: An adaptive noise modeling speech recognition system improves speech recognition by modifying an activation of the system's grammar rules or models based on detected noise characteristics. An adaptive noise modeling speech recognition system includes a sensor that receives acoustic data having a speech component and a noise component. A processor analyzes the acoustic data and generates a noise indicator that identifies a characteristic of the noise component. An integrating decision logic processes the noise indicator and generates a noise model activation data structure that includes data that may be used by a speech recognition engine to adjust the activation of associated grammar rules or models.
    Type: Application
    Filed: January 31, 2008
    Publication date: August 6, 2009
    Inventor: Rod Rempel
  • Publication number: 20090100454
    Abstract: Methods, devices, systems and tools are presented that allow the summarization of text, audio, and audiovisual presentations, such as movies, into less lengthy forms. High-content media files are shortened in a manner that preserves important details, by splitting the files into segments, rating the segments, and reassembling preferred segments into a final abridged piece. Summarization of media can be customized by user selection of criteria, and opens new possibilities for delivering entertainment, news, and information in the form of dense, information-rich content that can be viewed by means of broadcast or cable distribution, “on-demand” distribution, internet and cell phone digital video streaming, or can be downloaded onto an iPod™ and other portable video playback devices.
    Type: Application
    Filed: April 23, 2007
    Publication date: April 16, 2009
    Inventor: Frank Elmo Weber
  • Publication number: 20090080623
    Abstract: A method of providing voice metrics over an established telephone call between a user and a subscriber can include receiving voice information from the user over the call and determining biometric information from the voice information for the user. The method further can include encoding the biometric metric information and sending the biometric information to the subscriber over the call.
    Type: Application
    Filed: December 8, 2008
    Publication date: March 26, 2009
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Thomas E. Creamer, Peeyush Jaiswal, Victor S. Moore
  • Publication number: 20090043578
    Abstract: Disclosed are arrangements that provide security for items to which access is restricted by providing a single layer of security requiring a biometric signature (101) for access. A memory-space reduction approach reduces the search time to identify a matching biometric signature, this based on an auxiliary input (112) that is vocally uttered by a user seeking access to the controlled item (811). The disclosed arrangements can also use the auxiliary codes as control signals to expand the utility of the disclosed voice controlled memory partitioning arrangements.
    Type: Application
    Filed: December 13, 2005
    Publication date: February 12, 2009
    Inventor: Christopher John Burke
  • Patent number: 7457753
    Abstract: A system for remote assessment of a user is disclosed. The system comprises application software resident on a server and arranged to interact across a network with a user operating a client device to obtain one or more sample signals of the user's speech. A datastore is arranged to store the user speech samples in association with details of the user. A feature extraction engine is arranged to extract one or more first features from respective speech samples. A comparator is arranged to compare the first features extracted from a speech sample with second features extracted from one or more reference samples and to provide a measure of any differences between the first and second features for assessment of the user.
    Type: Grant
    Filed: June 29, 2005
    Date of Patent: November 25, 2008
    Assignee: University College Dublin National University of Ireland
    Inventors: Rosalyn Moran, Richard Reilly, Philip De Chazal, Brian O'Mullane, Peter Lacy
  • Publication number: 20080177661
    Abstract: This specification describes technologies relating to a phone-based payment system for transferring funds between payers and payees, and methods of providing such a system. In general, one aspect is implemented as a method of electronic payment that includes receiving a payer identifier from a payer, and the payer identifier is selected from a group of a registered phone number and a registered business server identifier. The method also includes identifying the payer as an authorized user based on the received payer identifier. The method further includes authorizing a payment transfer from a bank account of the payer to a bank account of a payee if the identified payer is an authorized user. Other implementations of this aspect include corresponding systems, apparatus, and computer program products.
    Type: Application
    Filed: January 22, 2007
    Publication date: July 24, 2008
    Inventor: Divya Mehra
  • Publication number: 20080147396
    Abstract: A speech recognition method is provided.
    Type: Application
    Filed: July 3, 2007
    Publication date: June 19, 2008
    Applicant: DELTA ELECTRONICS, INC.
    Inventor: Jui-Chang Wang
  • Publication number: 20080120105
    Abstract: Methods and apparatus to operate an audience metering device with voice commands are described herein. An example method to identify audience members based on voice, includes: obtaining an audio input signal including a program audio signal and a human voice signal; receiving an audio line signal from an audio output line of a monitored media device; processing the audio line signal with a filter having adaptive weights to generate a delayed and attenuated line signal; subtracting the delayed and attenuated line signal from the audio input signal to develop a residual audio signal; identifying a person that spoke to create the human voice signal based on the residual audio signal; and logging an identity of the person as an audience member.
    Type: Application
    Filed: February 1, 2008
    Publication date: May 22, 2008
    Inventor: VENUGOPAL SRINIVASAN