Speech Patents (Class 434/185)
  • Patent number: 8301447
    Abstract: The present invention relates to creating a phonetic index of phonemes from an audio segment that includes speech content from multiple sources. The phonemes in the phonetic index are directly or indirectly associated with the corresponding source of the speech from which the phonemes were derived. By associating the phonemes with a corresponding source, the phonetic index of speech content from multiple sources may be searched based on phonetic content as well as the corresponding source.
    Type: Grant
    Filed: October 10, 2008
    Date of Patent: October 30, 2012
    Assignee: Avaya Inc.
    Inventors: John H. Yoakum, Stephen Whynot
  • Publication number: 20120264091
    Abstract: One example embodiment of the present disclosure includes method and system to improve communication comprising an assembly fixedly positioned during use in close proximity to a user's ear. The assembly includes an accelerometer to detect the initiation and duration of the user's speech and an output presentation system. The output presentation system comprises a non-occlusive ear fitting that presents unintelligible noise that is unrelated to the sound-frequency or intonation of the user's current speech. The unintelligible noise is presented to the patient at a level less than 85 dB. The system further comprises a control arrangement to maintain presentation of the noise substantially throughout the detected duration of the user's speech, but substantially not at other times.
    Type: Application
    Filed: February 16, 2012
    Publication date: October 18, 2012
    Applicant: Purdue Research Foundation
    Inventors: Jessica E. Huber, Scott Kepner, Derek Tully, Barbara S. Tully, James Thomas Jones, Kirk Solon Foster
  • Patent number: 8272874
    Abstract: Speech data from the operation of a speech recognition application is recorded over the course of one or more language learning sessions. The operation of the speech recognition application during each language learning sessions corresponds to a user speaking, and the speech recognition application generating text data. The text data may a recognition of what the user spoke. The speech data may comprise the text data, and confidence values that are an indication of an accuracy of the recognition. The speech data from each language learning session may be analyzed to determine an overall performance level of the user.
    Type: Grant
    Filed: November 22, 2004
    Date of Patent: September 25, 2012
    Assignee: BravoBrava L.L.C.
    Inventors: Luc Julia, Jerome Dubreuil, Jehan Bing
  • Publication number: 20120237906
    Abstract: By performing a comparison of words spoken by a speaker and defined material which is presented to the speaker, information can be determined which allows for the convenient control of the presentation of material and external devices. A comparison of a speaker's words with defined material can be beneficially used as an input for controlling the operation of an exercise apparatus, video games, material presented to an audience, and the presentation of the material itself. Similar feedback loops can also be used with measurement and stimulation of neurophysiologic states, to make the activity of reading more enjoyable and convenient, or for other purposes.
    Type: Application
    Filed: February 19, 2009
    Publication date: September 20, 2012
    Inventors: Andrew B. Glass, Henry Van Styn, Coleman Kane
  • Patent number: 8271281
    Abstract: Techniques for assessing pronunciation abilities of a user are provided. The techniques include recording a sentence spoken by a user, performing a classification of the spoken sentence, wherein the classification is performed with respect to at least one N-ordered class, and wherein the spoken sentence is represented by a set of at least one acoustic feature extracted from the spoken sentence, and determining a score based on the classification, wherein the score is used to determine an optimal set of at least one question to assess pronunciation ability of the user without human intervention.
    Type: Grant
    Filed: June 27, 2008
    Date of Patent: September 18, 2012
    Assignee: Nuance Communications, Inc.
    Inventors: Jayadeva, Sachindra Joshi, Himanshu Pant, Ashish Verma
  • Patent number: 8258947
    Abstract: Embodiments of the present invention provide a method, system and computer program product for translation verification of source strings for controls in a target application graphical user interface (GUI). In an embodiment of the invention, a method for translation verification of source strings for controls in a target application GUI can include loading a target GUI for an application under test in a functional testing tool executing in memory by a processor of a computing system, retrieving different translated source strings in a target spoken language for respectively different control elements of the target GUI and, determining a score for each one of the translated source strings. Thereafter, an alert can be provided in the functional testing tool for each translated source string corresponding to a determined score failing to meet a threshold value, such as a score that falls below a threshold value, or a score that exceeds a threshold value.
    Type: Grant
    Filed: September 29, 2009
    Date of Patent: September 4, 2012
    Assignee: International Business Machines Corporation
    Inventors: Jennifer G. Becker, Kenneth Lee McClamroch, VinodKumar Raghavan, Peter Sun
  • Publication number: 20120219932
    Abstract: A system and method of speech instruction including generating computer audible speech from any text that is to be spoken by a human speaker, recording the human speech of the text data that was repeated by the human speaker based on the computer generated speech, evaluating the human speech by an automated speech recognition engine to determine a quality of the human speech, and providing feedback to the speaker.
    Type: Application
    Filed: February 27, 2012
    Publication date: August 30, 2012
    Inventors: Eyal ESHED, Ariel Velikovsky, Sherrie Ellen Shammass
  • Publication number: 20120214141
    Abstract: A method of teaching pronunciation is provided which includes communicating by a voice portal server to a user a model word and detecting a response by the user to the voice portal server. The method also includes comparing the response word to the model word and determining a confidence level based on the comparison of the response word to the model word. The method further includes comparing an acceptance limit to the confidence level and confirming a correct pronunciation of the model word if the confidence level one of equals and exceeds the acceptance limit.
    Type: Application
    Filed: May 2, 2012
    Publication date: August 23, 2012
    Inventors: Madhuri RAYA, Karsten Funk, Sharmila Ravula, Yao Meng
  • Patent number: 8239184
    Abstract: The teachings described herein generally relate to a system and method for multilingual teaching of numeric or language skills through an electronic translation of a source phrase to a destination language selected from multiple languages. The electronic translation can occur as a spoken translation, can be in real-time, and can mimic the voice of the user of the system.
    Type: Grant
    Filed: March 13, 2007
    Date of Patent: August 7, 2012
    Assignee: NewTalk, Inc.
    Inventors: Bruce W. Nash, Craig A. Robinson, Martha P. Robinson, Robert H. Clemons
  • Patent number: 8226416
    Abstract: The present invention is a method and apparatus for reading education. In one embodiment, a method for recognizing an utterance spoken by a reader, includes receiving text to be read by the reader, generating a grammar for speech recognition, in accordance with the text, receiving the utterance, interpreting the utterance in accordance with the grammar, and outputting feedback indicative of reader performance.
    Type: Grant
    Filed: December 7, 2007
    Date of Patent: July 24, 2012
    Assignee: SRI International
    Inventors: Victor Abrash, Douglas Bercow
  • Patent number: 8221126
    Abstract: Speech data from the operation of a speech recognition application is recorded over the course of one or more language learning sessions. The operation of the speech recognition application during each language learning sessions corresponds to a user speaking, and the speech recognition application generating text data. The text data may be recognition of what the user spoke. The speech data may comprise the text data, and confidence values that are an indication of an accuracy of the recognition. The speech data from each language learning session may be analyzed to determine an overall performance level of the user.
    Type: Grant
    Filed: November 22, 2004
    Date of Patent: July 17, 2012
    Assignee: BravoBrava L.L.C.
    Inventors: Luc Julia, Jerome Dubreuil, Jehan Bing
  • Patent number: 8210851
    Abstract: A method on a computing device for enhancing the memory and cognitive ability of an older adult by requiring the adult to differentiate between rapidly presented stimuli. The method utilizes a sequence of phonemes from a confusable pair which are systematically manipulated to make discrimination between the phonemes less difficult or more difficult based on the success of the adult, such as processing the consonant and vowel portions of the phonemes by emphasizing the portions, stretching the portions, and/or separating the consonant and vowel portions by time intervals. As the adult improves in auditory processing, the discriminations are made progressively more difficult by reducing the amount of processing to that of normal speech.
    Type: Grant
    Filed: August 15, 2006
    Date of Patent: July 3, 2012
    Assignee: Posit Science Corporation
    Inventors: Travis W. Wade, Joseph L. Hardy
  • Publication number: 20120164612
    Abstract: Speech errors for a learner of a language (e.g., an English language learner) are identified automatically based on aggregated characteristics of that learner's speech.
    Type: Application
    Filed: December 28, 2011
    Publication date: June 28, 2012
    Applicant: EnglishCentral, Inc.
    Inventors: Laurence Gillick, Alan Schwartz, Jean-Manuel Van Thong, Peter Wolf, Don McAllaster
  • Publication number: 20120156660
    Abstract: A dialogue system include learning initiation unit which receives conversation education domain and target completion condition inconversation education domain and receives user's utterance, voice recognition unit which converts user's utterance into utterance text based on utterance information, language understanding unit which determines user's dialogue act based on converted utterance text and generates logical expression using slot expression corresponding to determined dialogue act and slot expression defined in conversation education domain, dialogue/progress management unit which determines utterance vertex with logical expression similar to that of utterance patterns of plurality of utterance vertices connected to system's final utterance vertex in dynamic dialogue graph and determines utterance vertices connected to determined utterance vertex as next utterance, system dialogue generation unit which retrieves utterance patterns connected to utterance vertex corresponding to determined next utterance
    Type: Application
    Filed: December 15, 2011
    Publication date: June 21, 2012
    Applicant: Electronics and Telecommunications Research Institute
    Inventors: Oh Woog KWON, Sung Kwon CHOI, Ki Young LEE, Yoon Hyung ROH, Young Kil KIM, Eun Jin PARK, Yun JIN, Chang Hyun KIM, Young Ae SEO, Seong YANG, II, Jin Xia HUANG, Jong Hun SHIN, Yun Keun LEE, Sang Kyu PARK
  • Patent number: 8175882
    Abstract: A method for task execution improvement, the method includes: generating a baseline model for executing a task; recording a user executing a task; comparing the baseline model to the user's execution of the task; and providing feedback to the user based on the differences in the user's execution and the baseline model.
    Type: Grant
    Filed: January 25, 2008
    Date of Patent: May 8, 2012
    Assignee: International Business Machines Corporation
    Inventors: Sara H. Basson, Dimitiri Kanevsky, Edward E. Kelley, Bhuvana Ramabhadran
  • Publication number: 20120107778
    Abstract: The present invention is an interactive display board and method for teaching speech and communication skills through the interactive display board to an individual with language disabilities. A target response formed of sound units is requested from the individual. A plurality of visual indicators, such as lights, corresponds to each sound unit of the target response. The visual indicators are presented, and each sound unit is associated with a corresponding visual indicator. The association is a different sensory interaction, such a pressing a button, activating a switch, or pointing. The display board includes visual indicators with pressure-activated switches for the association of the sound units and visual indicators. The target response is the produced verbally by the individual concurrent with repeating the interaction with the visual indicators by the individual.
    Type: Application
    Filed: October 31, 2011
    Publication date: May 3, 2012
    Inventor: Patricia M. SCHUETTE
  • Patent number: 8157566
    Abstract: An adjustable hierarchical scoring method is disclosed. A hierarchical scoring structure for speech-sound data is generated. Historical data of speech-sound scoring for the speech-sound data is retrieved from a speech-sound practice database while a professional scoring result for the hierarchical scoring structure of the speech-sound data is being collected. A weight-adjusting operation is performed according to the historical data of speech-sound scoring and the professional scoring result to calculate optimum adjusting weights for each layer of the hierarchical scoring structure. Retrieved speech-sound data of a student is scored using a speech-sound scoring system and a weighted average operation is performed to the score of the speech-sound data of the student according to a corresponding adjusting weight, thereby generating the final score of the speech-sound data of the student.
    Type: Grant
    Filed: December 29, 2008
    Date of Patent: April 17, 2012
    Assignee: Institute For Information Industry
    Inventors: Te-Lu Tsai, Jyh-Shing Jang, Yu-Shiang Hung, Tzu-Chieh Tien
  • Publication number: 20120070810
    Abstract: Disclosed is a computer-based aid for teaching language for children with speech disorders. The teaching aid comprises a plurality of categorized, discrete multimedia files, each of which comprises discrete synchronized audio and video content pertaining to a subject matter, and a means for aggregating two or more discrete multimedia files into a single aggregated multimedia file comprising aggregated synchronized audio and video content comprising an aggregation of the individual discrete content of the of the two or more discrete multimedia files. The aggregated content is to be learnt by a learner.
    Type: Application
    Filed: May 24, 2011
    Publication date: March 22, 2012
    Inventor: Laura Marie Kasbar
  • Patent number: 8140341
    Abstract: The present invention relates to methods for editing timed and annotated data, the method comprising the steps of acquiring a multimedia data stream, segmenting the multimedia stream into respective audio/video audio data streams, wherein the playback times of the audio/video data streams are synchronized. Playback time annotation indicators are associated with the time synchronized audio/video data streams, wherein discrete playback time annotation indicators of the video data stream segments correlate with discrete playback time annotation indicators of the audio data stream segments, and a transcript of the audio data stream is created.
    Type: Grant
    Filed: January 19, 2007
    Date of Patent: March 20, 2012
    Assignee: International Business Machines Corporation
    Inventors: Alexander Faisman, Grabarnik Genady, Dimitri Kanevsky, Larisa Shwartz
  • Patent number: 8109765
    Abstract: Methods and related computer program products, systems, and devices for providing intelligent feedback to a user based on audio input associated with a user reading a passage are disclosed. The method can include assessing a level of fluency of a user's reading of the sequence of words using speech recognition technology to compare the audio input with an expected sequence of words and providing feedback to the user related to the level of fluency for a word.
    Type: Grant
    Filed: September 10, 2004
    Date of Patent: February 7, 2012
    Assignee: Scientific Learning Corporation
    Inventors: Valerie L. Beattie, Marilyn Jager Adams, Michael Barrow
  • Publication number: 20120021390
    Abstract: A method, system and apparatus for developing concepts, language and speech in language learning disabled and generally learning disabled subjects. The approach attempts to build associations between the various implementations of language, namely, visual, oral, aural and written language within the subject. The technique utilizes the subject's main strengths, often the visual sense and develops language by building on that strength, gradually progressing to spoken and heard language. Graphically rich content is sued to convey the concepts to the subject. The disclosed techniques may be implemented using a computer program.
    Type: Application
    Filed: July 26, 2011
    Publication date: January 26, 2012
    Inventor: Enda Patrick Dodd
  • Patent number: 8103503
    Abstract: Systems and methods for processing a user speech input to determine whether the user has correctly read a target sentence string are provided. One disclosed method may include receiving a sentence array including component words of the target sentence string and processing the sentence array to generate a symbolic representation of the target sentence string. The symbolic representation may include a subset of words selected from the component words of the target sentence string, having fewer words than the sentence array. The method may include processing user speech input to recognize in the user speech input each of the words in the subset of words in the symbolic representation of the target sentence string. The method may further include, upon recognizing the subset of words, making a determination that the user has correctly read the target sentence string.
    Type: Grant
    Filed: November 1, 2007
    Date of Patent: January 24, 2012
    Assignee: Microsoft Corporation
    Inventor: Duncan
  • Publication number: 20110318719
    Abstract: Acoustical voice-feedback systems include headsets connected through a connector or central hub for vocalizing and sound formation assistance. A user's voice is conveyed directly into his or her own ear as well as the ears of each connected user. A central hub allows selective connection of one or more headsets to each other, or allows an instructor to selectively connect an instructor headset to one or more instructee headsets.
    Type: Application
    Filed: March 2, 2010
    Publication date: December 29, 2011
    Applicant: HAREBRAIN, INC.
    Inventors: Steven Swain, Jeffrey Waffensmith
  • Publication number: 20110294097
    Abstract: A training game that requires a player to role-play a conversation in which the player is judged according to the player's demonstration of a Hidden Skill. Skills relevant to successful performance in a particular type of conversation (for example, a sales call) are identified. The player is aware that these are the skills the player is expected to learn and to demonstrate. The player is judged, however, as to the player's performance on a subset of the skills. The player does not know which of the skills form the subset on which the player will be judged. By hiding the specific skill being evaluated from the role player, the game encourages the player to perform all necessary conversation skills during the role-play. Only then is the player guaranteed to execute the Hidden Skill. The game thereby trains players in all aspects of the relevant conversation skills.
    Type: Application
    Filed: August 12, 2011
    Publication date: December 1, 2011
    Applicant: You Make the Call!, Inc.
    Inventor: Laura Montocchio
  • Patent number: 8068107
    Abstract: In a multimedia presentation, having speech and graphic contributions, a list of graphic objects is provided. Each graphic is associated to a graphic file capable of being executed by a computer to display a corresponding graphic contribution on a screen. A speech file comprising a sequence of phrases is also created, each phrase comprising a speech contribution explaining at least one graphic contribution associated to a respective graphic object. Then, an arrangement string is created obtained as a sequence of a first graphic object and a respective first phrase, and then a second graphic object and a respective second phrase, and so on up to completion of all graphic objects and phrases of said list and of said speech file respectively. A processing speed for displaying the graphic objects is chosen.
    Type: Grant
    Filed: November 22, 2004
    Date of Patent: November 29, 2011
    Inventor: Mario Pirchio
  • Patent number: 8057234
    Abstract: The increasing globalization of the world necessitates further strengthening of foreign language learning policies. The brain's foreign language recognition activity is enhanced and effective foreign language learning is enabled by extracting signals of prescribed bands from a speech signal in a foreign language using a first bandpass filter section having two or more bandpass filters, extracting the envelopes of each frequency band signal using envelope extraction sections having envelope extractors, applying a noise source signal to a second bandpass filter section having two or more bandpass filters and extracting noise signals corresponding to the prescribed bands, multiplying the outputs of the first bandpass filter section and the second bandpass filter section in multiplication sections, summing up the outputs of the multiplication sections in an addition section to produce a Noise-Vocoded Speech Sound signal, and presenting the Noise-Vocoded Speech Sound signals for listening.
    Type: Grant
    Filed: April 27, 2006
    Date of Patent: November 15, 2011
    Inventors: Hiroshi Rikimaru, Shinichi Sakamoto, Takeshi Nakaichi
  • Publication number: 20110256513
    Abstract: Auditory training systems are disclosed and include at least one speech perception activity, and at least one audio signal, at least one video signal or a combination thereof, wherein the system is user interactive, performance adaptive or a combination thereof.
    Type: Application
    Filed: March 3, 2011
    Publication date: October 20, 2011
    Inventors: Harry Levitt, Christopher Oden
  • Patent number: 8033831
    Abstract: Speech data from the operation of a speech recognition application is recorded over the course of one or more language learning sessions. The operation of the speech recognition application during each language learning sessions corresponds to a user speaking, and the speech recognition application generating text data. The text data may correspond to a recognition of what the user spoke. The speech data may comprise the text data, and confidence values that are an indication of an accuracy of the recognition. The speech data from each language learning session may be analyzed to determine an overall performance level of the user.
    Type: Grant
    Filed: November 22, 2004
    Date of Patent: October 11, 2011
    Assignee: Bravobrava L.L.C.
    Inventors: Luc Julia, Jerome Dubreuil, Jehan Bing
  • Patent number: 8036896
    Abstract: A server for providing language and literacy tutoring information to a plurality of user devices connected to a communications network, comprising a network adapter for connection to the network; a content database for providing learning content to devices via the network adaptor and the network; a plurality of speech recognition models stored in the server; a processor for processing speech data and session control data generated by a user and sent to the server by the network, the processor evaluating which of the speech recognition models provides most accurate speech recognition results; and a performance evaluator for evaluating speech produced by the user using the speech recognition model that produces the most accurate results. A system, including user devices. A method for operating the system, and a program storage medium having computer code thereon for implementing the method and system.
    Type: Grant
    Filed: April 18, 2006
    Date of Patent: October 11, 2011
    Assignee: Nuance Communications, Inc.
    Inventors: Hugh William Adams, Jr., Peter Gustav Fairweather, Yael Ravin
  • Patent number: 8002551
    Abstract: A student providing a multi-word response in a computerized language teaching system provides a manual input concurrently with each responsive word. For example, he might enter a keystroke correspondent to the first letter of each word. When using the teaching computer silently, a student will typically “speak” each word mentally as he enters a keystroke, so the limited experience is almost as effective as speaking out loud. When a student types one or more keystrokes concurrently with each word that he speaks, the computer will be able to detect when a student is responding with a correct word, but merely mispronouncing it. Also, since the computer will receive a keystroke as the student starts each new word, it is better able to distinguish the boundaries between words and recognize them more reliably.
    Type: Grant
    Filed: December 5, 2008
    Date of Patent: August 23, 2011
    Assignee: Rosetta Stone, Ltd.
    Inventors: Gregory Keim, Jack August Marmorstein, Ronald Bryce Inouye, John Fairfield, Michael Scott Fulkerson
  • Publication number: 20110123966
    Abstract: An apparatus for developing, promoting and/or augmenting communication skills includes an apparatus base dimensioned for positioning relative to a support. The apparatus base defines longitudinal, vertical and orthogonal axes, and a reference plane inclusive of the longitudinal and orthogonal axes. The apparatus base includes a top support surface arranged at a predetermined angle greater than 0 degrees relative to the reference plane. The top support surface has an article retaining element associated therewith. At least one icon containing article having communication media adapted for conveying information is releasably mounted to the article retaining element of the top support surface. The top support surface may be arranged at an angle ranging from about 10 degrees to about 60 degrees relative to the reference plane. In one embodiment, the top support surface is arranged at an angle of about 45 degrees relative to the reference plane.
    Type: Application
    Filed: November 25, 2009
    Publication date: May 26, 2011
    Inventor: Patricia Anne Toback
  • Publication number: 20110104647
    Abstract: A system for conditioning a child to learn any language without an accent, contains a memory and a processor configured by the memory to perform the steps of: separating from a full International Phonetic Alphabet (IPA) set, a native language IPA, resulting in a derivative set, wherein the native language is at least one native language of the child; sequencing the derivative set according to aspects of the derivative set, wherein aspects are selected from the group including consonants, vowels, and tones, resulting in a sequenced derivative set; selecting words in foreign languages that use at least one aspect of the sequenced derivative set each; embedding the selected words in an easy to follow format, in the native language of the child; and playing the easy to follow format with the embedded words according to the prior determined sequence of the sequenced derivative set.
    Type: Application
    Filed: October 29, 2010
    Publication date: May 5, 2011
    Inventor: Gadi BenMark Markovitch
  • Publication number: 20110059423
    Abstract: Presented are a system and method for providing a graphical user interface (GUI) based modular platform having educational content. The method includes providing an interactive GUI on a computing device accessible by a user, receiving a first indication of a language being studied, displaying a GUI layer presenting a selection of level, unit, activity, and/or lesson, receiving a lesson selection, and computing a rating or score of the user's performance for the lesson. The system includes a control module, a language parser accessing content within repositories and providing language specific content, a syntax rule base providing language specific rules to the language parser, a lesson format base providing a lesson style to the control module, and a speech recognition system evaluating a user's utterance for pronunciation and sentence structure accuracy.
    Type: Application
    Filed: September 3, 2010
    Publication date: March 10, 2011
    Inventors: Naomi Kadar, Nadav Kadar, Vincent Thomas J. Rowe, Avraham Kadar
  • Publication number: 20110021320
    Abstract: The present invention relates to a device for rhythm pacing using a real time recording of stepping (or speech motor production) with a radiowave (speed of light) connection to a head-mounted device. The rhythm is initiated with the patient's initial performance and increased by increments to entrain motor skill. By automating the pulsing in a fashion that reflects real time stepping, trial and error can be avoided, yielding efficient learning and gait improvement.
    Type: Application
    Filed: January 6, 2009
    Publication date: January 27, 2011
    Inventor: Martin L. Lenhardt
  • Publication number: 20110014595
    Abstract: The present invention relates a partner assisted voice output communication method to aid individuals of all ages with overcoming verbal language expression difficulties. This method is conducted in conjunction with an interactive voice output device as a teaching tool which allows a child/learner—by pressing buttons/symbols provided on an overlay—to see and hear words in average daily activities and selected activities. The learner's partner, teacher, or guide uses this tool to model single words in context or combines 2 or more words while speaking in a natural voice, increasing functional, effective, interactive verbal communication. Strategies and supplemental training aids for using the method and device to engage the language learner/student are disclosed.
    Type: Application
    Filed: July 20, 2010
    Publication date: January 20, 2011
    Inventor: Sydney Birr
  • Patent number: 7862418
    Abstract: A gaming system and related methods comprising a gaming device and a voucher mechanism in communication with the gaming device. The voucher mechanism is configured to present various types of vouchers to the player depending on a game outcome. The types of vouchers include, without limitation, a jackpot voucher, a merchandise voucher, a free play voucher, a mystery voucher, a competition entry voucher, and a restricted machine play voucher.
    Type: Grant
    Filed: November 20, 2002
    Date of Patent: January 4, 2011
    Assignee: Bally Gaming, Inc.
    Inventors: Robert A. Luciano, Jr., Warren R. White, Russ T. Bradford
  • Publication number: 20100304343
    Abstract: A system is provided for acquiring increased proficiency in a new or target language by enabling frequent repeated hearing and mimicking of discrete audio sound bites in the target language, wherein the sound bites and their frequency of repetition may be easily selected by the user before or during use. The sound bites may be played singly or in sequential groups or user defined groups, and automatically repeated as necessary or desired to achieve audio imprinting of the sound bites by simply listening and repeating the sounds without substantial intellectual effort or attempts at memorization. The system enables user selected or sequential hearing and mimicking of specific sound bites or groups of sound bites repeatedly in each use session in the context of meaning, without requiring dedicated language learning time, and thereby helps reproduce the natural process of initial native language acquisition which occurs without substantial intellectual effort and memorization.
    Type: Application
    Filed: June 2, 2010
    Publication date: December 2, 2010
    Inventor: Louis R. Bucalo
  • Patent number: 7818179
    Abstract: Devices and methods are provided for automatically analyzing a user's conversation or verbal presentation and providing feedback to the user, in real-time and/or off-line, to make the user aware of improper speech habits and thus enable the user to acquire better language habits while speaking. For example, a device for providing user awareness of speech habits may be a portable device, which can be worn or carried by the user, has various functions for monitoring and analyzing a speech session (e.g. conversation, verbal presentation) of a user and providing notification to the user of improper speech or language habits that are identified during the speech session.
    Type: Grant
    Filed: November 12, 2004
    Date of Patent: October 19, 2010
    Assignee: International Business Machines Corporation
    Inventors: Sophia Krasikov, John F. Morar, Senthilnathan Velayudham
  • Publication number: 20100248195
    Abstract: Disclosed is a method and complementary apparatus for teaching speech sound production of a target sound that is positioned within the consonant blend of a word in which the consonant blend is split at a precise point in order to focus the student's attention on the target sound to teach its proper pronunciation and in which the split consonant blend is distinguishably demarcated within the word and the word is presented to the student in and on a variety of suitable media.
    Type: Application
    Filed: March 24, 2010
    Publication date: September 30, 2010
    Inventor: Cheri Fletcher
  • Publication number: 20100233662
    Abstract: A therapeutic method for developing the ability of subjects with autism spectrum disorders to produce and perceive spoken language, including sequentially modeling a set of words, the speaking of which involves making a first and a second articulatory gesture, as pictures corresponding to such words are displayed, in order to induce the subject to attempt to say the modeled words, until the subject is able to produce the constrictions of the oral-pharyngeal cavity associated with both of said articulatory gestures of the words together with vibration of the vocal folds, and the subject is able to produce such words intelligibly. The subject's ability is incrementally expanded using sets of words involving the making of other articulatory gestures until the subject is able to intelligibly produce words involving substantially all of the articulatory gestures used in the language of interest. Positive visual reinforcement is given to the subject for each word that the subject is able to produce intelligibly.
    Type: Application
    Filed: March 11, 2009
    Publication date: September 16, 2010
    Applicant: THE SPEECH INSTITUTE, LLC
    Inventor: Maureen Casper
  • Patent number: 7797146
    Abstract: A method of simulating interactive communication between a user and a human subject. The method comprises: assigning at least one phrase to a stored content sequence, wherein the content sequence comprises a content clip of the subject; parsing the at least one phrase to produce at least one phonetic clone; associating the at least one phonetic clone with the stored content sequence; receiving an utterance from the user; matching the utterance to the at least one phonetic clone; and displaying the stored content sequence associated with the at least one phonetic clone.
    Type: Grant
    Filed: May 13, 2003
    Date of Patent: September 14, 2010
    Assignee: Interactive Drama, Inc.
    Inventors: William G. Harless, Michael G. Harless, Marcia A. Zier
  • Patent number: 7778834
    Abstract: The present disclosure presents a useful metric for assessing the relative difficulty which non-native speakers face in pronouncing a given utterance and a method and systems for using such a metric in the evaluation and assessment of the utterances of non-native speakers. In an embodiment, the metric may be based on both known sources of difficulty for language learners and a corpus-based measure of cross-language sound differences. The method may be applied to speakers who primarily speak a first language speaking utterances in any non-native second language.
    Type: Grant
    Filed: August 11, 2008
    Date of Patent: August 17, 2010
    Assignee: Educational Testing Service
    Inventors: Derrick Higgins, Klaus Zechner, Yoko Futagi, Rene Lawless
  • Publication number: 20100151427
    Abstract: An adjustable hierarchical scoring method is disclosed. A hierarchical scoring structure for speech-sound data is generated. Historical data of speech-sound scoring for the speech-sound data is retrieved from a speech-sound practice database while a professional scoring result for the hierarchical scoring structure of the speech-sound data is being collected. A weight-adjusting operation is performed according to the historical data of speech-sound scoring and the professional scoring result to calculate optimum adjusting weights for each layer of the hierarchical scoring structure. Retrieved speech-sound data of a student is scored using a speech-sound scoring system and a weighted average operation is performed to the score of the speech-sound data of the student according to a corresponding adjusting weight, thereby generating the final score of the speech-sound data of the student.
    Type: Application
    Filed: December 29, 2008
    Publication date: June 17, 2010
    Applicant: INSTITUTE FOR INFORMATION INDUSTRY
    Inventors: Te-Lu Tsai, Jyh-Shing Jang, Yu-Shiang Hung, Tzu-Chieh Tien
  • Publication number: 20100143874
    Abstract: A student providing a multi-word response in a computerized language teaching system provides a manual input concurrently with each responsive word. For example, he might enter a keystroke correspondent to the first letter of each word. When using the teaching computer silently, a student will typically “speak” each word mentally as he enters a keystroke, so the limited experience is almost as effective as speaking out loud. When a student types one or more keystrokes concurrently with each word that he speaks, the computer will be able to detect when a student is responding with a correct word, but merely mispronouncing it. Also, since the computer will receive a keystroke as the student starts each new word, it is better able to distinguish the boundaries between words and recognize them more reliably.
    Type: Application
    Filed: December 5, 2008
    Publication date: June 10, 2010
    Inventors: Gregory Keim, Jack August Marmorstein, Ronald Bryce Inouye, John Fairfield, Michael Scott Fulkerson
  • Patent number: 7729907
    Abstract: Preparing for the full-fledged aged society, measures to prevent senility are required. Senility is prevented by extracting signals of prescribed bands from a speech signal using a first bandpass filter section having a plurality of bandpass filters, extracting the envelopes of each frequency band signal using an envelope extraction section having envelope extractors, applying a noise source signal to a second bandpass filter section having a plurality of bandpass filters and extracting noise signals corresponding to the prescribed bands, multiplying the outputs from the first bandpass filter section and the second bandpass filter section in a multiplication section, summing up the outputs from the multiplication section in an addition section to produce a Noise-Vocoded Speech Sound signal, and presenting the Noise-Vocoded Speech Sound signal for listening.
    Type: Grant
    Filed: February 21, 2005
    Date of Patent: June 1, 2010
    Assignees: Rion Co., Ltd.
    Inventor: Hiroshi Rikimaru
  • Publication number: 20100120002
    Abstract: Disclosed is a system and method for conversation practice in simulated situations. The system comprises situational conversation teaching material, an audio processing module and a conversation processing module. The teaching material consists of multi-flow conversation paths and conversational sentences with a plurality of replaceable vocabulary items. According to different contents of the situational conversation teaching material and biased data of the teaching material, the audio processing module dynamically adjusts speech recognition model and recognizes the inputted audio signal of the learners to determine the information on the recognition results. The conversation processing module determines the information in response to the learners, based on the information on the recognition results, the situational conversation teaching material and biased data of the teaching material.
    Type: Application
    Filed: August 27, 2009
    Publication date: May 13, 2010
    Inventors: Chieh-Chih Chang, Sen-Chia Chang, Chung-Jen Chiu, Jian-Yung Hung, Lin-Chi Huang
  • Publication number: 20100105015
    Abstract: A system for decoding foreign accents includes a visual display monitor, an audio display device, and a processor. The processor is electrically connected to, and configured for communication with, the visual display monitor and the audio display device. The processor is further configured to exert a measure of control over the visual display monitor and the audio display device to display aural and visual content. The processor is still further configured to train a user of the system to understand the nature of a mispronunciation of a native language word by a foreign language speaker, and to comprehend a response thereto. The processor is yet still further configured to validate the user's understanding and comprehension by allowing the user to determine whether his response to a mispronunciation was accurately performed following an aural display of the mispronunciation through the audio display device.
    Type: Application
    Filed: October 15, 2009
    Publication date: April 29, 2010
    Inventors: Judy Ravin, Corissa Niemann
  • Publication number: 20100092933
    Abstract: The interactive phoneme video game of the present invention includes an action game section where a user can collect a phoneme item. At least one phoneme item is associated with a phoneme from a word. A word and phoneme pairing challenge encourages a user to make use of phoneme information gained during the action game. A phoneme sound is preferably played during interaction with the video game such as when a phoneme item is collected or during navigation of a phoneme keypad.
    Type: Application
    Filed: October 15, 2008
    Publication date: April 15, 2010
    Inventors: William Kuchera, Martin Fletcher, Alan Aldworth
  • Publication number: 20100062403
    Abstract: Systems, methods, and other embodiments associated with producing an immersive training content module (ITCM) are described. One example system includes a capture logic to acquire information from which the ITCM may be produced. An ITCM may include a set of nodes, a set of measures, a logic to control transitions between nodes during a training session, and a logic to establish values for measures during the training sessions. Therefore, the example system may also include an assessment definition logic to define a set of measures to be included in the ITCM and an interaction logic to define a set of interactions to be included in the ITCM. The ITCM may be written to a computer-readable medium.
    Type: Application
    Filed: December 21, 2007
    Publication date: March 11, 2010
    Applicant: CASE WESTERN RESERVE UNIVERSITY
    Inventors: Stacy L. Williams, Marc Buchner
  • Patent number: 7676372
    Abstract: A speech transformation apparatus comprises a microphone 21 for detecting speech and generating a speech signal; a signal processor 22 for performing a speech recognition process using the speech signal; a speech information generator for transforming the recognition result responsive to the physical state of the user, the operating conditions, and/or the purpose for using the apparatus; and a display unit 26 and loudspeaker 25 for generating a control signal for outputting a raw recognition result and/or a transformed recognition result. In a speech transformation apparatus thus constituted, speech enunciated by a spoken-language-impaired individual can be transformed and presented to the user, and sounds from outside sources can also be transformed and presented to the user.
    Type: Grant
    Filed: February 16, 2000
    Date of Patent: March 9, 2010
    Assignee: Yugen Kaisha GM&M
    Inventor: Toshihiko Oba