Transformation Of Speech Into A Nonaudible Representation, E.g., Speech Visualization, Speech Processing For Tactile Aids, Etc. (epo) Patents (Class 704/E21.019)
  • Publication number: 20090276222
    Abstract: A method and system for improving the context and accuracy of speech and video analytics searches by incorporating one or more inputs and defining and applying a plurality of rules for the different stages of said speech and video analytics system searches.
    Type: Application
    Filed: May 1, 2009
    Publication date: November 5, 2009
    Applicant: CYBERTECH INFORMATION INTERNATIONAL, INC.
    Inventor: Raman Ramesh
  • Publication number: 20090271182
    Abstract: In accordance with the present invention, computer implemented methods and systems are provided for representing and modeling the temporal structure of audio signals. In response to receiving a signal, a time-to-frequency domain transformation on at least a portion of the received signal to generate a frequency domain representation is performed. The time-to-frequency domain transformation converts the signal from a time domain representation to the frequency domain representation. A frequency domain linear prediction (FDLP) is performed on the frequency domain representation to estimate a temporal envelope of the frequency domain representation. Based on the temporal envelope, one or more speech features are generated.
    Type: Application
    Filed: February 12, 2009
    Publication date: October 29, 2009
    Applicant: THE TRUSTEES OF COLUMBIA UNIVERSITY IN THE CITY OF NEW YORK
    Inventors: Marios Athineos, Daniel P.W. Ellis
  • Publication number: 20090259475
    Abstract: A text edit apparatus which presents, based on language analysis information regarding a text, a portion of the text where voice quality may change when the text is read aloud has advantages of predicting likelihood of the voice quality change and judging whether or not the voice quality change will occur.
    Type: Application
    Filed: June 5, 2006
    Publication date: October 15, 2009
    Inventors: Katsuyoshi Yamagami, Yumiko Kato, Shinobu Adachi
  • Publication number: 20090206993
    Abstract: A method (700) and system (900) for authenticating a user is provided. The method can include receiving one or more spoken utterances from a user (702), recognizing a phrase corresponding to one or more spoken utterances (704), identifying a biometric voice print of the user from one or more spoken utterances of the phrase (706), determining a device identifier associated with the device (708), and authenticating the user based on the phrase, the biometric voice print, and the device identifier (710). A location of the handset or the user can be employed as criteria for granting access to one or more resources (712).
    Type: Application
    Filed: April 13, 2009
    Publication date: August 20, 2009
    Applicant: PORTICUS TECHNOLOGY, INC.
    Inventors: GERMANO DI MAMBRO, Bernardas Salna
  • Publication number: 20090192801
    Abstract: A method for controlling an electronic device with voice commands using a mobile phone (1) includes: storing voice characteristics of voice command settings of a user and corresponding remote control signals of the voice command settings in a memory (12); recording voice command input of the user through a microphone (11) in the mobile phone; processing the voice command input, and detecting voice characteristics of the voice command input; determining whether the voice characteristics of the voice command input match voice characteristics of one of the plurality of voice command settings stored in the memory; and sending the remote control signal corresponding to the voice command setting that matches the voice command input to a wireless receiver (13) of the electronic device through a wireless transmitter (21) of the mobile phone, so as to control the electronic device remotely according the remote control signal.
    Type: Application
    Filed: June 13, 2008
    Publication date: July 30, 2009
    Applicant: CHI MEI COMMUNICATION SYSTEMS, INC.
    Inventor: MENG-CHIEH CHOU
  • Publication number: 20090157410
    Abstract: Disclosed is a speech translating system for translating speech from a first language to a language selected from a set of second languages. The system includes an input unit, a processor, and an output unit. The input unit is capable of receiving the speech in the first language. The processor is operatively coupled to the input unit and is capable of converting the speech in the first language to the speech in the selected language. The output unit is operatively coupled to the processor. The output unit is capable of outputting the speech in the selected language.
    Type: Application
    Filed: December 13, 2007
    Publication date: June 18, 2009
    Inventor: James Donohoe
  • Publication number: 20090157393
    Abstract: An encoding device (200) includes an MDCT unit (202) that transforms an input signal in a time domain into a frequency spectrum including a lower frequency spectrum, a BWE encoding unit (204) that generates extension data which specifies a higher frequency spectrum at a higher frequency than the lower frequency spectrum, and an encoded data stream generating unit (205) that encodes to output the lower frequency spectrum obtained by the MDCT unit (202) and the extension data obtained by the BWE encoding unit (204). The BWE encoding unit (204) generates as the extension data (i) a first parameter which specifies a lower subband which is to be copied as the higher frequency spectrum from among a plurality of the lower subbands which form the lower frequency spectrum obtained by the MDCT unit (202) and (ii) a second parameter which specifies a gain of the lower subband after being copied.
    Type: Application
    Filed: February 12, 2009
    Publication date: June 18, 2009
    Inventors: Mineo TSUSHIMA, Takeshi NORIMATSU, Kosuke NISHIO, Naoya TANAKA
  • Publication number: 20090150160
    Abstract: Embodiments of the present invention improve methods of performing speech recognition using human gestures. In one embodiment, the present invention includes a speech recognition method comprising detecting a gesture, selecting a first recognition set based on the gesture, receiving a speech input signal, and recognizing the speech input signal in the context of the first recognition set.
    Type: Application
    Filed: September 25, 2008
    Publication date: June 11, 2009
    Applicant: Sensory, Incorporated
    Inventor: Todd F. Mozer
  • Publication number: 20090007346
    Abstract: A method for controlling an information display using an avatar of a washing machine is disclosed, which displays all information associated with usage- and control-information of the washing machine via the avatar, and allows a user of the washing machine to easily recognize the usage- and control-information of the washing machine, resulting in increased information transmission characteristics of the washing machine. Basic usages of the washing machine, and various methods for displaying operation- and control-states of the washing machine will be indicated by at least one dynamic character, such that a user of the washing machine can easily recognize necessary information of the washing machine, resulting in not only increased information transmission characteristics of the washing machine but also emphasized entertainment elements required by modern consumers of the washing machine.
    Type: Application
    Filed: June 27, 2006
    Publication date: January 8, 2009
    Applicant: LG ELECTRONICS INC.
    Inventors: Mi Kyung Ha, Gyeong Ho Moon, Sang Su Lee
  • Publication number: 20080312930
    Abstract: According to MPEG-4's TTS architecture, facial animation can be driven by two streams simultaneously—text, and Facial Animation Parameters. In this architecture, text input is sent to a Text-To-Speech converter at a decoder that drives the mouth shapes of the face. Facial Animation Parameters are sent from an encoder to the face over the communication channel. The present invention includes codes (known as bookmarks) in the text string transmitted to the Text-to-Speech converter, which bookmarks are placed between words as well as inside them. According to the present invention, the bookmarks carry an encoder time stamp. Due to the nature of text-to-speech conversion, the encoder time stamp does not relate to real-world time, and should be interpreted as a counter. In addition, the Facial Animation Parameter stream carries the same encoder time stamp found in the bookmark of the text.
    Type: Application
    Filed: August 18, 2008
    Publication date: December 18, 2008
    Applicant: AT&T Corp.
    Inventors: Andrea Basso, Mark Charles Beutnagel, Joern Ostermann
  • Publication number: 20080294431
    Abstract: Displays a character string representing content of speech in synchronization with reproduction of the speech. An apparatus includes: a unit for obtaining scenario data representing the speech; a unit for dividing textual data resulting from recognition of the speech to generate pieces of recognition pieces of recognition data; a unit for detecting in the scenario data a character matching each character contained in each piece of recognition data for which no matching character string has been detected to detect in the scenario data a character string that matches the piece of recognition data; and a unit for setting the display timing of displaying each of character strings contained in the scenario data to the timing at which speech recognized as the piece of recognition data that matches the character string is reproduced.
    Type: Application
    Filed: June 2, 2008
    Publication date: November 27, 2008
    Inventors: Kohtaroh Miyamoto, Midori Shoji
  • Publication number: 20080253613
    Abstract: A method for facilitating cooperation between humans and remote vehicles comprises creating image data, detecting humans within the image data, extracting gesture information from the image data, mapping the gesture information to a remote vehicle behavior, and activating the remote vehicle behavior. Alternatively, voice commands can by used to activate the remote vehicle behavior.
    Type: Application
    Filed: April 11, 2008
    Publication date: October 16, 2008
    Inventors: Christopher Vernon Jones, Odest Chadwicke Jenkins, Matthew M. Loper
  • Publication number: 20080235031
    Abstract: An interface apparatus according to an embodiment of the invention includes: an operation detecting section configured to detect a device operation; a status detecting section configured to detect a status change or status continuance of a device or in the vicinity of the device; an operation history accumulating section configured to accumulate a operation detection result and a status detection result in association with each other; an operation history matching section configured to match a status detection result for a newly detected against accumulated status detection results, and select a device operation that corresponds to the status detection result for the newly detected; and an utterance section configured to utter as sound a word corresponding to the selected device operation.
    Type: Application
    Filed: March 13, 2008
    Publication date: September 25, 2008
    Applicant: KABUSHIKI KAISHA TOSHIBA
    Inventor: Daisuke Yamamoto
  • Publication number: 20080228497
    Abstract: The invention describes a method for communication by means of a communication device (DS), in which synthesized speech (ss) is output from the communication device (DS), and in which light signals (ls) are output simultaneously with the synthesized speech (ss) in accordance with the semantic content of the synthesized speech (ss). Furthermore, an appropriate communication device (DS) is described.
    Type: Application
    Filed: July 3, 2006
    Publication date: September 18, 2008
    Applicant: KONINKLIJKE PHILIPS ELECTRONICS, N.V.
    Inventors: Thomas Portele, Holger R. Scholl
  • Publication number: 20080147413
    Abstract: This invention generally relates to system, methods and computer program code for editing or modifying speech affect. A speech affect processing system to enable a user to edit an affect content of a speech signal, the system comprising: input to receive speech analysis data from a speech processing system said speech analysis data, comprising a set of parameters representing said speech signal; a user input to receive user input data defining one or more affect-related operations to be performed on said speech signal; and an affect modification system coupled to said user input and to said speech processing system to modify said parameters in accordance with said one or more affect-related operations and further comprising a speech reconstruction system to reconstruct an affect modified speech signal from said modified parameters; and an output coupled to said affect modification system to output said affect modified speech signal.
    Type: Application
    Filed: October 18, 2007
    Publication date: June 19, 2008
    Inventor: Tal Sobol-Shikler
  • Publication number: 20080140383
    Abstract: A system of communicating over cellular telephones using symbols in place of language parts. The symbols are downloaded as part of the lookup table. A user can enter a message in text, and that message is converted into the symbols. The symbols can then be sent to a different interactive device, which can read those symbols, and convert them into text in a different language. The user sees the symbols and begins to learn meanings of the symbols. In addition, the symbols are language independent.
    Type: Application
    Filed: November 29, 2007
    Publication date: June 12, 2008
    Applicant: UBIQUITY HOLDINGS
    Inventors: Kenneth S. Bailey, Christopher Carmichael
  • Publication number: 20080103772
    Abstract: When characters are provided, probabilities for each character are multiplied by the most heavily weighted word that could possibly be derived from the characters provided. The products are compared, and the greatest product is predicted to show the path to the word that the user presumably is entering. Second, third, etc. predictions are made such that the assumption of the first character and other characters can change once the first word prediction has been culled from the remaining possibilities.
    Type: Application
    Filed: October 31, 2007
    Publication date: May 1, 2008
    Inventor: Duncan Bates
  • Publication number: 20080091434
    Abstract: Systems and methods for training voice activation control of electronic equipment are disclosed. One example method includes receiving a selection corresponding to at least one command used to control the electronic equipment. The method further includes instructing a user to speak, and responsive to the instruction, receiving a digitized speech stream. The method further includes segmenting the speech stream into speech segments, storing at least one of the speech segments as an entry in a dictionary, and associating the dictionary entry with the selected command.
    Type: Application
    Filed: December 14, 2007
    Publication date: April 17, 2008
    Applicant: SCIENTIFIC ATLANTA
    Inventors: Arturo Rodriguez, David Sedacca, Albert Garcia
  • Publication number: 20080091433
    Abstract: A system has a speech control interface operable to provide a speech sequence to which a user response is required before system functionality is operational. A user response can be provided via a spoken command. The system is operable, in response to interruption of the speech sequence to permit a non-verbal user response using an alternative interface.
    Type: Application
    Filed: May 15, 2007
    Publication date: April 17, 2008
    Applicant: DENSO CORPORATION
    Inventor: Kurt Dusterhoff