Speech Patents (Class 434/185)
-
Patent number: 8301447Abstract: The present invention relates to creating a phonetic index of phonemes from an audio segment that includes speech content from multiple sources. The phonemes in the phonetic index are directly or indirectly associated with the corresponding source of the speech from which the phonemes were derived. By associating the phonemes with a corresponding source, the phonetic index of speech content from multiple sources may be searched based on phonetic content as well as the corresponding source.Type: GrantFiled: October 10, 2008Date of Patent: October 30, 2012Assignee: Avaya Inc.Inventors: John H. Yoakum, Stephen Whynot
-
Publication number: 20120264091Abstract: One example embodiment of the present disclosure includes method and system to improve communication comprising an assembly fixedly positioned during use in close proximity to a user's ear. The assembly includes an accelerometer to detect the initiation and duration of the user's speech and an output presentation system. The output presentation system comprises a non-occlusive ear fitting that presents unintelligible noise that is unrelated to the sound-frequency or intonation of the user's current speech. The unintelligible noise is presented to the patient at a level less than 85 dB. The system further comprises a control arrangement to maintain presentation of the noise substantially throughout the detected duration of the user's speech, but substantially not at other times.Type: ApplicationFiled: February 16, 2012Publication date: October 18, 2012Applicant: Purdue Research FoundationInventors: Jessica E. Huber, Scott Kepner, Derek Tully, Barbara S. Tully, James Thomas Jones, Kirk Solon Foster
-
Patent number: 8272874Abstract: Speech data from the operation of a speech recognition application is recorded over the course of one or more language learning sessions. The operation of the speech recognition application during each language learning sessions corresponds to a user speaking, and the speech recognition application generating text data. The text data may a recognition of what the user spoke. The speech data may comprise the text data, and confidence values that are an indication of an accuracy of the recognition. The speech data from each language learning session may be analyzed to determine an overall performance level of the user.Type: GrantFiled: November 22, 2004Date of Patent: September 25, 2012Assignee: BravoBrava L.L.C.Inventors: Luc Julia, Jerome Dubreuil, Jehan Bing
-
Publication number: 20120237906Abstract: By performing a comparison of words spoken by a speaker and defined material which is presented to the speaker, information can be determined which allows for the convenient control of the presentation of material and external devices. A comparison of a speaker's words with defined material can be beneficially used as an input for controlling the operation of an exercise apparatus, video games, material presented to an audience, and the presentation of the material itself. Similar feedback loops can also be used with measurement and stimulation of neurophysiologic states, to make the activity of reading more enjoyable and convenient, or for other purposes.Type: ApplicationFiled: February 19, 2009Publication date: September 20, 2012Inventors: Andrew B. Glass, Henry Van Styn, Coleman Kane
-
Patent number: 8271281Abstract: Techniques for assessing pronunciation abilities of a user are provided. The techniques include recording a sentence spoken by a user, performing a classification of the spoken sentence, wherein the classification is performed with respect to at least one N-ordered class, and wherein the spoken sentence is represented by a set of at least one acoustic feature extracted from the spoken sentence, and determining a score based on the classification, wherein the score is used to determine an optimal set of at least one question to assess pronunciation ability of the user without human intervention.Type: GrantFiled: June 27, 2008Date of Patent: September 18, 2012Assignee: Nuance Communications, Inc.Inventors: Jayadeva, Sachindra Joshi, Himanshu Pant, Ashish Verma
-
Patent number: 8258947Abstract: Embodiments of the present invention provide a method, system and computer program product for translation verification of source strings for controls in a target application graphical user interface (GUI). In an embodiment of the invention, a method for translation verification of source strings for controls in a target application GUI can include loading a target GUI for an application under test in a functional testing tool executing in memory by a processor of a computing system, retrieving different translated source strings in a target spoken language for respectively different control elements of the target GUI and, determining a score for each one of the translated source strings. Thereafter, an alert can be provided in the functional testing tool for each translated source string corresponding to a determined score failing to meet a threshold value, such as a score that falls below a threshold value, or a score that exceeds a threshold value.Type: GrantFiled: September 29, 2009Date of Patent: September 4, 2012Assignee: International Business Machines CorporationInventors: Jennifer G. Becker, Kenneth Lee McClamroch, VinodKumar Raghavan, Peter Sun
-
Publication number: 20120219932Abstract: A system and method of speech instruction including generating computer audible speech from any text that is to be spoken by a human speaker, recording the human speech of the text data that was repeated by the human speaker based on the computer generated speech, evaluating the human speech by an automated speech recognition engine to determine a quality of the human speech, and providing feedback to the speaker.Type: ApplicationFiled: February 27, 2012Publication date: August 30, 2012Inventors: Eyal ESHED, Ariel Velikovsky, Sherrie Ellen Shammass
-
Publication number: 20120214141Abstract: A method of teaching pronunciation is provided which includes communicating by a voice portal server to a user a model word and detecting a response by the user to the voice portal server. The method also includes comparing the response word to the model word and determining a confidence level based on the comparison of the response word to the model word. The method further includes comparing an acceptance limit to the confidence level and confirming a correct pronunciation of the model word if the confidence level one of equals and exceeds the acceptance limit.Type: ApplicationFiled: May 2, 2012Publication date: August 23, 2012Inventors: Madhuri RAYA, Karsten Funk, Sharmila Ravula, Yao Meng
-
Patent number: 8239184Abstract: The teachings described herein generally relate to a system and method for multilingual teaching of numeric or language skills through an electronic translation of a source phrase to a destination language selected from multiple languages. The electronic translation can occur as a spoken translation, can be in real-time, and can mimic the voice of the user of the system.Type: GrantFiled: March 13, 2007Date of Patent: August 7, 2012Assignee: NewTalk, Inc.Inventors: Bruce W. Nash, Craig A. Robinson, Martha P. Robinson, Robert H. Clemons
-
Patent number: 8226416Abstract: The present invention is a method and apparatus for reading education. In one embodiment, a method for recognizing an utterance spoken by a reader, includes receiving text to be read by the reader, generating a grammar for speech recognition, in accordance with the text, receiving the utterance, interpreting the utterance in accordance with the grammar, and outputting feedback indicative of reader performance.Type: GrantFiled: December 7, 2007Date of Patent: July 24, 2012Assignee: SRI InternationalInventors: Victor Abrash, Douglas Bercow
-
Patent number: 8221126Abstract: Speech data from the operation of a speech recognition application is recorded over the course of one or more language learning sessions. The operation of the speech recognition application during each language learning sessions corresponds to a user speaking, and the speech recognition application generating text data. The text data may be recognition of what the user spoke. The speech data may comprise the text data, and confidence values that are an indication of an accuracy of the recognition. The speech data from each language learning session may be analyzed to determine an overall performance level of the user.Type: GrantFiled: November 22, 2004Date of Patent: July 17, 2012Assignee: BravoBrava L.L.C.Inventors: Luc Julia, Jerome Dubreuil, Jehan Bing
-
Patent number: 8210851Abstract: A method on a computing device for enhancing the memory and cognitive ability of an older adult by requiring the adult to differentiate between rapidly presented stimuli. The method utilizes a sequence of phonemes from a confusable pair which are systematically manipulated to make discrimination between the phonemes less difficult or more difficult based on the success of the adult, such as processing the consonant and vowel portions of the phonemes by emphasizing the portions, stretching the portions, and/or separating the consonant and vowel portions by time intervals. As the adult improves in auditory processing, the discriminations are made progressively more difficult by reducing the amount of processing to that of normal speech.Type: GrantFiled: August 15, 2006Date of Patent: July 3, 2012Assignee: Posit Science CorporationInventors: Travis W. Wade, Joseph L. Hardy
-
Publication number: 20120164612Abstract: Speech errors for a learner of a language (e.g., an English language learner) are identified automatically based on aggregated characteristics of that learner's speech.Type: ApplicationFiled: December 28, 2011Publication date: June 28, 2012Applicant: EnglishCentral, Inc.Inventors: Laurence Gillick, Alan Schwartz, Jean-Manuel Van Thong, Peter Wolf, Don McAllaster
-
Publication number: 20120156660Abstract: A dialogue system include learning initiation unit which receives conversation education domain and target completion condition inconversation education domain and receives user's utterance, voice recognition unit which converts user's utterance into utterance text based on utterance information, language understanding unit which determines user's dialogue act based on converted utterance text and generates logical expression using slot expression corresponding to determined dialogue act and slot expression defined in conversation education domain, dialogue/progress management unit which determines utterance vertex with logical expression similar to that of utterance patterns of plurality of utterance vertices connected to system's final utterance vertex in dynamic dialogue graph and determines utterance vertices connected to determined utterance vertex as next utterance, system dialogue generation unit which retrieves utterance patterns connected to utterance vertex corresponding to determined next utteranceType: ApplicationFiled: December 15, 2011Publication date: June 21, 2012Applicant: Electronics and Telecommunications Research InstituteInventors: Oh Woog KWON, Sung Kwon CHOI, Ki Young LEE, Yoon Hyung ROH, Young Kil KIM, Eun Jin PARK, Yun JIN, Chang Hyun KIM, Young Ae SEO, Seong YANG, II, Jin Xia HUANG, Jong Hun SHIN, Yun Keun LEE, Sang Kyu PARK
-
Patent number: 8175882Abstract: A method for task execution improvement, the method includes: generating a baseline model for executing a task; recording a user executing a task; comparing the baseline model to the user's execution of the task; and providing feedback to the user based on the differences in the user's execution and the baseline model.Type: GrantFiled: January 25, 2008Date of Patent: May 8, 2012Assignee: International Business Machines CorporationInventors: Sara H. Basson, Dimitiri Kanevsky, Edward E. Kelley, Bhuvana Ramabhadran
-
Publication number: 20120107778Abstract: The present invention is an interactive display board and method for teaching speech and communication skills through the interactive display board to an individual with language disabilities. A target response formed of sound units is requested from the individual. A plurality of visual indicators, such as lights, corresponds to each sound unit of the target response. The visual indicators are presented, and each sound unit is associated with a corresponding visual indicator. The association is a different sensory interaction, such a pressing a button, activating a switch, or pointing. The display board includes visual indicators with pressure-activated switches for the association of the sound units and visual indicators. The target response is the produced verbally by the individual concurrent with repeating the interaction with the visual indicators by the individual.Type: ApplicationFiled: October 31, 2011Publication date: May 3, 2012Inventor: Patricia M. SCHUETTE
-
Patent number: 8157566Abstract: An adjustable hierarchical scoring method is disclosed. A hierarchical scoring structure for speech-sound data is generated. Historical data of speech-sound scoring for the speech-sound data is retrieved from a speech-sound practice database while a professional scoring result for the hierarchical scoring structure of the speech-sound data is being collected. A weight-adjusting operation is performed according to the historical data of speech-sound scoring and the professional scoring result to calculate optimum adjusting weights for each layer of the hierarchical scoring structure. Retrieved speech-sound data of a student is scored using a speech-sound scoring system and a weighted average operation is performed to the score of the speech-sound data of the student according to a corresponding adjusting weight, thereby generating the final score of the speech-sound data of the student.Type: GrantFiled: December 29, 2008Date of Patent: April 17, 2012Assignee: Institute For Information IndustryInventors: Te-Lu Tsai, Jyh-Shing Jang, Yu-Shiang Hung, Tzu-Chieh Tien
-
Publication number: 20120070810Abstract: Disclosed is a computer-based aid for teaching language for children with speech disorders. The teaching aid comprises a plurality of categorized, discrete multimedia files, each of which comprises discrete synchronized audio and video content pertaining to a subject matter, and a means for aggregating two or more discrete multimedia files into a single aggregated multimedia file comprising aggregated synchronized audio and video content comprising an aggregation of the individual discrete content of the of the two or more discrete multimedia files. The aggregated content is to be learnt by a learner.Type: ApplicationFiled: May 24, 2011Publication date: March 22, 2012Inventor: Laura Marie Kasbar
-
Patent number: 8140341Abstract: The present invention relates to methods for editing timed and annotated data, the method comprising the steps of acquiring a multimedia data stream, segmenting the multimedia stream into respective audio/video audio data streams, wherein the playback times of the audio/video data streams are synchronized. Playback time annotation indicators are associated with the time synchronized audio/video data streams, wherein discrete playback time annotation indicators of the video data stream segments correlate with discrete playback time annotation indicators of the audio data stream segments, and a transcript of the audio data stream is created.Type: GrantFiled: January 19, 2007Date of Patent: March 20, 2012Assignee: International Business Machines CorporationInventors: Alexander Faisman, Grabarnik Genady, Dimitri Kanevsky, Larisa Shwartz
-
Patent number: 8109765Abstract: Methods and related computer program products, systems, and devices for providing intelligent feedback to a user based on audio input associated with a user reading a passage are disclosed. The method can include assessing a level of fluency of a user's reading of the sequence of words using speech recognition technology to compare the audio input with an expected sequence of words and providing feedback to the user related to the level of fluency for a word.Type: GrantFiled: September 10, 2004Date of Patent: February 7, 2012Assignee: Scientific Learning CorporationInventors: Valerie L. Beattie, Marilyn Jager Adams, Michael Barrow
-
Publication number: 20120021390Abstract: A method, system and apparatus for developing concepts, language and speech in language learning disabled and generally learning disabled subjects. The approach attempts to build associations between the various implementations of language, namely, visual, oral, aural and written language within the subject. The technique utilizes the subject's main strengths, often the visual sense and develops language by building on that strength, gradually progressing to spoken and heard language. Graphically rich content is sued to convey the concepts to the subject. The disclosed techniques may be implemented using a computer program.Type: ApplicationFiled: July 26, 2011Publication date: January 26, 2012Inventor: Enda Patrick Dodd
-
Patent number: 8103503Abstract: Systems and methods for processing a user speech input to determine whether the user has correctly read a target sentence string are provided. One disclosed method may include receiving a sentence array including component words of the target sentence string and processing the sentence array to generate a symbolic representation of the target sentence string. The symbolic representation may include a subset of words selected from the component words of the target sentence string, having fewer words than the sentence array. The method may include processing user speech input to recognize in the user speech input each of the words in the subset of words in the symbolic representation of the target sentence string. The method may further include, upon recognizing the subset of words, making a determination that the user has correctly read the target sentence string.Type: GrantFiled: November 1, 2007Date of Patent: January 24, 2012Assignee: Microsoft CorporationInventor: Duncan
-
Publication number: 20110318719Abstract: Acoustical voice-feedback systems include headsets connected through a connector or central hub for vocalizing and sound formation assistance. A user's voice is conveyed directly into his or her own ear as well as the ears of each connected user. A central hub allows selective connection of one or more headsets to each other, or allows an instructor to selectively connect an instructor headset to one or more instructee headsets.Type: ApplicationFiled: March 2, 2010Publication date: December 29, 2011Applicant: HAREBRAIN, INC.Inventors: Steven Swain, Jeffrey Waffensmith
-
Publication number: 20110294097Abstract: A training game that requires a player to role-play a conversation in which the player is judged according to the player's demonstration of a Hidden Skill. Skills relevant to successful performance in a particular type of conversation (for example, a sales call) are identified. The player is aware that these are the skills the player is expected to learn and to demonstrate. The player is judged, however, as to the player's performance on a subset of the skills. The player does not know which of the skills form the subset on which the player will be judged. By hiding the specific skill being evaluated from the role player, the game encourages the player to perform all necessary conversation skills during the role-play. Only then is the player guaranteed to execute the Hidden Skill. The game thereby trains players in all aspects of the relevant conversation skills.Type: ApplicationFiled: August 12, 2011Publication date: December 1, 2011Applicant: You Make the Call!, Inc.Inventor: Laura Montocchio
-
Patent number: 8068107Abstract: In a multimedia presentation, having speech and graphic contributions, a list of graphic objects is provided. Each graphic is associated to a graphic file capable of being executed by a computer to display a corresponding graphic contribution on a screen. A speech file comprising a sequence of phrases is also created, each phrase comprising a speech contribution explaining at least one graphic contribution associated to a respective graphic object. Then, an arrangement string is created obtained as a sequence of a first graphic object and a respective first phrase, and then a second graphic object and a respective second phrase, and so on up to completion of all graphic objects and phrases of said list and of said speech file respectively. A processing speed for displaying the graphic objects is chosen.Type: GrantFiled: November 22, 2004Date of Patent: November 29, 2011Inventor: Mario Pirchio
-
Patent number: 8057234Abstract: The increasing globalization of the world necessitates further strengthening of foreign language learning policies. The brain's foreign language recognition activity is enhanced and effective foreign language learning is enabled by extracting signals of prescribed bands from a speech signal in a foreign language using a first bandpass filter section having two or more bandpass filters, extracting the envelopes of each frequency band signal using envelope extraction sections having envelope extractors, applying a noise source signal to a second bandpass filter section having two or more bandpass filters and extracting noise signals corresponding to the prescribed bands, multiplying the outputs of the first bandpass filter section and the second bandpass filter section in multiplication sections, summing up the outputs of the multiplication sections in an addition section to produce a Noise-Vocoded Speech Sound signal, and presenting the Noise-Vocoded Speech Sound signals for listening.Type: GrantFiled: April 27, 2006Date of Patent: November 15, 2011Inventors: Hiroshi Rikimaru, Shinichi Sakamoto, Takeshi Nakaichi
-
Publication number: 20110256513Abstract: Auditory training systems are disclosed and include at least one speech perception activity, and at least one audio signal, at least one video signal or a combination thereof, wherein the system is user interactive, performance adaptive or a combination thereof.Type: ApplicationFiled: March 3, 2011Publication date: October 20, 2011Inventors: Harry Levitt, Christopher Oden
-
Patent number: 8033831Abstract: Speech data from the operation of a speech recognition application is recorded over the course of one or more language learning sessions. The operation of the speech recognition application during each language learning sessions corresponds to a user speaking, and the speech recognition application generating text data. The text data may correspond to a recognition of what the user spoke. The speech data may comprise the text data, and confidence values that are an indication of an accuracy of the recognition. The speech data from each language learning session may be analyzed to determine an overall performance level of the user.Type: GrantFiled: November 22, 2004Date of Patent: October 11, 2011Assignee: Bravobrava L.L.C.Inventors: Luc Julia, Jerome Dubreuil, Jehan Bing
-
Patent number: 8036896Abstract: A server for providing language and literacy tutoring information to a plurality of user devices connected to a communications network, comprising a network adapter for connection to the network; a content database for providing learning content to devices via the network adaptor and the network; a plurality of speech recognition models stored in the server; a processor for processing speech data and session control data generated by a user and sent to the server by the network, the processor evaluating which of the speech recognition models provides most accurate speech recognition results; and a performance evaluator for evaluating speech produced by the user using the speech recognition model that produces the most accurate results. A system, including user devices. A method for operating the system, and a program storage medium having computer code thereon for implementing the method and system.Type: GrantFiled: April 18, 2006Date of Patent: October 11, 2011Assignee: Nuance Communications, Inc.Inventors: Hugh William Adams, Jr., Peter Gustav Fairweather, Yael Ravin
-
Patent number: 8002551Abstract: A student providing a multi-word response in a computerized language teaching system provides a manual input concurrently with each responsive word. For example, he might enter a keystroke correspondent to the first letter of each word. When using the teaching computer silently, a student will typically “speak” each word mentally as he enters a keystroke, so the limited experience is almost as effective as speaking out loud. When a student types one or more keystrokes concurrently with each word that he speaks, the computer will be able to detect when a student is responding with a correct word, but merely mispronouncing it. Also, since the computer will receive a keystroke as the student starts each new word, it is better able to distinguish the boundaries between words and recognize them more reliably.Type: GrantFiled: December 5, 2008Date of Patent: August 23, 2011Assignee: Rosetta Stone, Ltd.Inventors: Gregory Keim, Jack August Marmorstein, Ronald Bryce Inouye, John Fairfield, Michael Scott Fulkerson
-
Publication number: 20110123966Abstract: An apparatus for developing, promoting and/or augmenting communication skills includes an apparatus base dimensioned for positioning relative to a support. The apparatus base defines longitudinal, vertical and orthogonal axes, and a reference plane inclusive of the longitudinal and orthogonal axes. The apparatus base includes a top support surface arranged at a predetermined angle greater than 0 degrees relative to the reference plane. The top support surface has an article retaining element associated therewith. At least one icon containing article having communication media adapted for conveying information is releasably mounted to the article retaining element of the top support surface. The top support surface may be arranged at an angle ranging from about 10 degrees to about 60 degrees relative to the reference plane. In one embodiment, the top support surface is arranged at an angle of about 45 degrees relative to the reference plane.Type: ApplicationFiled: November 25, 2009Publication date: May 26, 2011Inventor: Patricia Anne Toback
-
Publication number: 20110104647Abstract: A system for conditioning a child to learn any language without an accent, contains a memory and a processor configured by the memory to perform the steps of: separating from a full International Phonetic Alphabet (IPA) set, a native language IPA, resulting in a derivative set, wherein the native language is at least one native language of the child; sequencing the derivative set according to aspects of the derivative set, wherein aspects are selected from the group including consonants, vowels, and tones, resulting in a sequenced derivative set; selecting words in foreign languages that use at least one aspect of the sequenced derivative set each; embedding the selected words in an easy to follow format, in the native language of the child; and playing the easy to follow format with the embedded words according to the prior determined sequence of the sequenced derivative set.Type: ApplicationFiled: October 29, 2010Publication date: May 5, 2011Inventor: Gadi BenMark Markovitch
-
Publication number: 20110059423Abstract: Presented are a system and method for providing a graphical user interface (GUI) based modular platform having educational content. The method includes providing an interactive GUI on a computing device accessible by a user, receiving a first indication of a language being studied, displaying a GUI layer presenting a selection of level, unit, activity, and/or lesson, receiving a lesson selection, and computing a rating or score of the user's performance for the lesson. The system includes a control module, a language parser accessing content within repositories and providing language specific content, a syntax rule base providing language specific rules to the language parser, a lesson format base providing a lesson style to the control module, and a speech recognition system evaluating a user's utterance for pronunciation and sentence structure accuracy.Type: ApplicationFiled: September 3, 2010Publication date: March 10, 2011Inventors: Naomi Kadar, Nadav Kadar, Vincent Thomas J. Rowe, Avraham Kadar
-
Publication number: 20110021320Abstract: The present invention relates to a device for rhythm pacing using a real time recording of stepping (or speech motor production) with a radiowave (speed of light) connection to a head-mounted device. The rhythm is initiated with the patient's initial performance and increased by increments to entrain motor skill. By automating the pulsing in a fashion that reflects real time stepping, trial and error can be avoided, yielding efficient learning and gait improvement.Type: ApplicationFiled: January 6, 2009Publication date: January 27, 2011Inventor: Martin L. Lenhardt
-
Publication number: 20110014595Abstract: The present invention relates a partner assisted voice output communication method to aid individuals of all ages with overcoming verbal language expression difficulties. This method is conducted in conjunction with an interactive voice output device as a teaching tool which allows a child/learner—by pressing buttons/symbols provided on an overlay—to see and hear words in average daily activities and selected activities. The learner's partner, teacher, or guide uses this tool to model single words in context or combines 2 or more words while speaking in a natural voice, increasing functional, effective, interactive verbal communication. Strategies and supplemental training aids for using the method and device to engage the language learner/student are disclosed.Type: ApplicationFiled: July 20, 2010Publication date: January 20, 2011Inventor: Sydney Birr
-
Patent number: 7862418Abstract: A gaming system and related methods comprising a gaming device and a voucher mechanism in communication with the gaming device. The voucher mechanism is configured to present various types of vouchers to the player depending on a game outcome. The types of vouchers include, without limitation, a jackpot voucher, a merchandise voucher, a free play voucher, a mystery voucher, a competition entry voucher, and a restricted machine play voucher.Type: GrantFiled: November 20, 2002Date of Patent: January 4, 2011Assignee: Bally Gaming, Inc.Inventors: Robert A. Luciano, Jr., Warren R. White, Russ T. Bradford
-
Publication number: 20100304343Abstract: A system is provided for acquiring increased proficiency in a new or target language by enabling frequent repeated hearing and mimicking of discrete audio sound bites in the target language, wherein the sound bites and their frequency of repetition may be easily selected by the user before or during use. The sound bites may be played singly or in sequential groups or user defined groups, and automatically repeated as necessary or desired to achieve audio imprinting of the sound bites by simply listening and repeating the sounds without substantial intellectual effort or attempts at memorization. The system enables user selected or sequential hearing and mimicking of specific sound bites or groups of sound bites repeatedly in each use session in the context of meaning, without requiring dedicated language learning time, and thereby helps reproduce the natural process of initial native language acquisition which occurs without substantial intellectual effort and memorization.Type: ApplicationFiled: June 2, 2010Publication date: December 2, 2010Inventor: Louis R. Bucalo
-
Patent number: 7818179Abstract: Devices and methods are provided for automatically analyzing a user's conversation or verbal presentation and providing feedback to the user, in real-time and/or off-line, to make the user aware of improper speech habits and thus enable the user to acquire better language habits while speaking. For example, a device for providing user awareness of speech habits may be a portable device, which can be worn or carried by the user, has various functions for monitoring and analyzing a speech session (e.g. conversation, verbal presentation) of a user and providing notification to the user of improper speech or language habits that are identified during the speech session.Type: GrantFiled: November 12, 2004Date of Patent: October 19, 2010Assignee: International Business Machines CorporationInventors: Sophia Krasikov, John F. Morar, Senthilnathan Velayudham
-
Publication number: 20100248195Abstract: Disclosed is a method and complementary apparatus for teaching speech sound production of a target sound that is positioned within the consonant blend of a word in which the consonant blend is split at a precise point in order to focus the student's attention on the target sound to teach its proper pronunciation and in which the split consonant blend is distinguishably demarcated within the word and the word is presented to the student in and on a variety of suitable media.Type: ApplicationFiled: March 24, 2010Publication date: September 30, 2010Inventor: Cheri Fletcher
-
Publication number: 20100233662Abstract: A therapeutic method for developing the ability of subjects with autism spectrum disorders to produce and perceive spoken language, including sequentially modeling a set of words, the speaking of which involves making a first and a second articulatory gesture, as pictures corresponding to such words are displayed, in order to induce the subject to attempt to say the modeled words, until the subject is able to produce the constrictions of the oral-pharyngeal cavity associated with both of said articulatory gestures of the words together with vibration of the vocal folds, and the subject is able to produce such words intelligibly. The subject's ability is incrementally expanded using sets of words involving the making of other articulatory gestures until the subject is able to intelligibly produce words involving substantially all of the articulatory gestures used in the language of interest. Positive visual reinforcement is given to the subject for each word that the subject is able to produce intelligibly.Type: ApplicationFiled: March 11, 2009Publication date: September 16, 2010Applicant: THE SPEECH INSTITUTE, LLCInventor: Maureen Casper
-
Patent number: 7797146Abstract: A method of simulating interactive communication between a user and a human subject. The method comprises: assigning at least one phrase to a stored content sequence, wherein the content sequence comprises a content clip of the subject; parsing the at least one phrase to produce at least one phonetic clone; associating the at least one phonetic clone with the stored content sequence; receiving an utterance from the user; matching the utterance to the at least one phonetic clone; and displaying the stored content sequence associated with the at least one phonetic clone.Type: GrantFiled: May 13, 2003Date of Patent: September 14, 2010Assignee: Interactive Drama, Inc.Inventors: William G. Harless, Michael G. Harless, Marcia A. Zier
-
Patent number: 7778834Abstract: The present disclosure presents a useful metric for assessing the relative difficulty which non-native speakers face in pronouncing a given utterance and a method and systems for using such a metric in the evaluation and assessment of the utterances of non-native speakers. In an embodiment, the metric may be based on both known sources of difficulty for language learners and a corpus-based measure of cross-language sound differences. The method may be applied to speakers who primarily speak a first language speaking utterances in any non-native second language.Type: GrantFiled: August 11, 2008Date of Patent: August 17, 2010Assignee: Educational Testing ServiceInventors: Derrick Higgins, Klaus Zechner, Yoko Futagi, Rene Lawless
-
Publication number: 20100151427Abstract: An adjustable hierarchical scoring method is disclosed. A hierarchical scoring structure for speech-sound data is generated. Historical data of speech-sound scoring for the speech-sound data is retrieved from a speech-sound practice database while a professional scoring result for the hierarchical scoring structure of the speech-sound data is being collected. A weight-adjusting operation is performed according to the historical data of speech-sound scoring and the professional scoring result to calculate optimum adjusting weights for each layer of the hierarchical scoring structure. Retrieved speech-sound data of a student is scored using a speech-sound scoring system and a weighted average operation is performed to the score of the speech-sound data of the student according to a corresponding adjusting weight, thereby generating the final score of the speech-sound data of the student.Type: ApplicationFiled: December 29, 2008Publication date: June 17, 2010Applicant: INSTITUTE FOR INFORMATION INDUSTRYInventors: Te-Lu Tsai, Jyh-Shing Jang, Yu-Shiang Hung, Tzu-Chieh Tien
-
Publication number: 20100143874Abstract: A student providing a multi-word response in a computerized language teaching system provides a manual input concurrently with each responsive word. For example, he might enter a keystroke correspondent to the first letter of each word. When using the teaching computer silently, a student will typically “speak” each word mentally as he enters a keystroke, so the limited experience is almost as effective as speaking out loud. When a student types one or more keystrokes concurrently with each word that he speaks, the computer will be able to detect when a student is responding with a correct word, but merely mispronouncing it. Also, since the computer will receive a keystroke as the student starts each new word, it is better able to distinguish the boundaries between words and recognize them more reliably.Type: ApplicationFiled: December 5, 2008Publication date: June 10, 2010Inventors: Gregory Keim, Jack August Marmorstein, Ronald Bryce Inouye, John Fairfield, Michael Scott Fulkerson
-
Patent number: 7729907Abstract: Preparing for the full-fledged aged society, measures to prevent senility are required. Senility is prevented by extracting signals of prescribed bands from a speech signal using a first bandpass filter section having a plurality of bandpass filters, extracting the envelopes of each frequency band signal using an envelope extraction section having envelope extractors, applying a noise source signal to a second bandpass filter section having a plurality of bandpass filters and extracting noise signals corresponding to the prescribed bands, multiplying the outputs from the first bandpass filter section and the second bandpass filter section in a multiplication section, summing up the outputs from the multiplication section in an addition section to produce a Noise-Vocoded Speech Sound signal, and presenting the Noise-Vocoded Speech Sound signal for listening.Type: GrantFiled: February 21, 2005Date of Patent: June 1, 2010Assignees: Rion Co., Ltd.Inventor: Hiroshi Rikimaru
-
Publication number: 20100120002Abstract: Disclosed is a system and method for conversation practice in simulated situations. The system comprises situational conversation teaching material, an audio processing module and a conversation processing module. The teaching material consists of multi-flow conversation paths and conversational sentences with a plurality of replaceable vocabulary items. According to different contents of the situational conversation teaching material and biased data of the teaching material, the audio processing module dynamically adjusts speech recognition model and recognizes the inputted audio signal of the learners to determine the information on the recognition results. The conversation processing module determines the information in response to the learners, based on the information on the recognition results, the situational conversation teaching material and biased data of the teaching material.Type: ApplicationFiled: August 27, 2009Publication date: May 13, 2010Inventors: Chieh-Chih Chang, Sen-Chia Chang, Chung-Jen Chiu, Jian-Yung Hung, Lin-Chi Huang
-
Publication number: 20100105015Abstract: A system for decoding foreign accents includes a visual display monitor, an audio display device, and a processor. The processor is electrically connected to, and configured for communication with, the visual display monitor and the audio display device. The processor is further configured to exert a measure of control over the visual display monitor and the audio display device to display aural and visual content. The processor is still further configured to train a user of the system to understand the nature of a mispronunciation of a native language word by a foreign language speaker, and to comprehend a response thereto. The processor is yet still further configured to validate the user's understanding and comprehension by allowing the user to determine whether his response to a mispronunciation was accurately performed following an aural display of the mispronunciation through the audio display device.Type: ApplicationFiled: October 15, 2009Publication date: April 29, 2010Inventors: Judy Ravin, Corissa Niemann
-
Publication number: 20100092933Abstract: The interactive phoneme video game of the present invention includes an action game section where a user can collect a phoneme item. At least one phoneme item is associated with a phoneme from a word. A word and phoneme pairing challenge encourages a user to make use of phoneme information gained during the action game. A phoneme sound is preferably played during interaction with the video game such as when a phoneme item is collected or during navigation of a phoneme keypad.Type: ApplicationFiled: October 15, 2008Publication date: April 15, 2010Inventors: William Kuchera, Martin Fletcher, Alan Aldworth
-
Publication number: 20100062403Abstract: Systems, methods, and other embodiments associated with producing an immersive training content module (ITCM) are described. One example system includes a capture logic to acquire information from which the ITCM may be produced. An ITCM may include a set of nodes, a set of measures, a logic to control transitions between nodes during a training session, and a logic to establish values for measures during the training sessions. Therefore, the example system may also include an assessment definition logic to define a set of measures to be included in the ITCM and an interaction logic to define a set of interactions to be included in the ITCM. The ITCM may be written to a computer-readable medium.Type: ApplicationFiled: December 21, 2007Publication date: March 11, 2010Applicant: CASE WESTERN RESERVE UNIVERSITYInventors: Stacy L. Williams, Marc Buchner
-
Patent number: 7676372Abstract: A speech transformation apparatus comprises a microphone 21 for detecting speech and generating a speech signal; a signal processor 22 for performing a speech recognition process using the speech signal; a speech information generator for transforming the recognition result responsive to the physical state of the user, the operating conditions, and/or the purpose for using the apparatus; and a display unit 26 and loudspeaker 25 for generating a control signal for outputting a raw recognition result and/or a transformed recognition result. In a speech transformation apparatus thus constituted, speech enunciated by a spoken-language-impaired individual can be transformed and presented to the user, and sounds from outside sources can also be transformed and presented to the user.Type: GrantFiled: February 16, 2000Date of Patent: March 9, 2010Assignee: Yugen Kaisha GM&MInventor: Toshihiko Oba