Segmentation Or Word Limit Detection (epo) Patents (Class 704/E15.005)
-
Publication number: 20100256978Abstract: A method for performing speech recognition relating to an object for the purpose of affecting automatic processing of the object by a processing system. The object carries information with at least a character string of processing information. The character string spoken by an operator is processed by way of a speech recognition procedure to generate a first result. Based on the need for more information of an element of the first result additional processing data is requested. An operator's response generates a second result. The first result is then modified to achieve consistency with the operator's response.Type: ApplicationFiled: April 6, 2010Publication date: October 7, 2010Applicant: SIEMENS AKTIENGESELLSCHAFTInventor: Walter Rosenbaum
-
Publication number: 20100250251Abstract: Architecture that suppresses the unexpected appearance of words by applying appropriate restrictions to long-term and short-term memory. The quickness of adaptation is also realized by leveraging the restriction. The architecture includes a history component for processing user input history for conversion of a phonetic string by a conversion process that output conversion results, and an adaptation component for adapting the conversion process to the user input history based on restriction(s) applied to short-term memory that impacts word appearances during the conversion process. The architecture performs probability boosting based on context-dependent probability differences (short-term memory), and dynamic linear-interpolation between long-term memory and baseline language model based on frequency of preceding context of word (long-term memory).Type: ApplicationFiled: March 30, 2009Publication date: September 30, 2010Applicant: Microsoft CorporationInventors: Katsutoshi Ohtsuki, Takashi Umeoka
-
Publication number: 20100229116Abstract: A control apparatus includes a voice recognition unit for recognizing user utterance to output a recognized word, a function storage unit for determining and storing a desired function that corresponds to the recognized word, a detector for detecting a preset user operation, a button display unit for displaying on a screen a shortcut button that instructs execution of the desired function stored in the storage unit when the detector detects the preset user operation, and a control unit for controlling execution of the desired function when the shortcut button is operated. By storing the desired function in association with the recognized word and by detecting user instruction, the control apparatus displays a shortcut button for a necessary function only.Type: ApplicationFiled: March 4, 2010Publication date: September 9, 2010Applicant: DENSO CORPORATIONInventors: Fumihiko Murase, Ichiro Akahori, Shinji Niwa
-
Publication number: 20100223056Abstract: A method, apparatus, and system are described for a continuous speech recognition engine that includes a fine speech recognizer model, a coarse sound representation generator, and a coarse match generator. The fine speech recognizer model receives a time coded sequence of sound feature frames, applies a speech recognition process to the sound feature frames and determines at least a best guess at each recognizable word that corresponds to the sound feature frames. The coarse sound representation generator generates a coarse sound representation of the recognized word. The coarse match generator determines a likelihood of the coarse sound representation actually being the recognized word based on comparing the coarse sound representation of the recognized word to a database containing the known sound of that recognized word and assigns the likelihood as a robust confidence level parameter to that recognized word.Type: ApplicationFiled: February 27, 2009Publication date: September 2, 2010Applicant: AUTONOMY CORPORATION LTD.Inventor: Mahapathy Kadirkamanathan
-
Publication number: 20100217597Abstract: Systems and methods for using an annotation guide to label utterances and speech data with a call type are disclosed. A method embodiment monitors labelers of speech data by presenting via a processor a test utterance to a labeler, receiving input from the labeler that selects a particular call type from a list of call types and determining via the processor if the labeler labeled the test utterance correctly. Based on the determining step, the method performs at least one of the following: revising the annotation guide, retraining the labeler or altering the test utterance.Type: ApplicationFiled: May 3, 2010Publication date: August 26, 2010Applicant: AT&T Intellectual Property II, LP via a transfer from AT&T Corp.Inventors: Lee Begeja, Richard Vandervoort Cox, Harris Drucker, David Crawford Gibbon, Allen Louis Gorin, Patrick Guy Haffner, Steven H. Lewis, Zhu Liu, Mazin G. Rahim, Bernard S. Renger, Behzad Shahraray
-
Publication number: 20100217598Abstract: A speech recognition system in which, even when the user makes an utterance including a word that satisfies a predetermined condition such as an unknown word, such a fact can be presented to the user, and the user can confirm the fact easily, is provided.Type: ApplicationFiled: February 22, 2007Publication date: August 26, 2010Applicant: NEC CORPORATIONInventor: Fumihiro Adachi
-
Publication number: 20100211390Abstract: The present invention relates to a method of generating a candidate list from a list of entries in accordance with a string of subword units corresponding to a speech input in a speech recognition system, the list of entries including plural list entries each comprising at least one fragment having one or more subword units. For each list entry, the fragments of the list entry are compared with the string of subword units. A matching score for each of the compared fragments based on the comparison is determined. The matching score for a fragment is further based on a comparison of at least one other fragment of the same list entry with the string of subword units. A total score for each list entry is determined based on the matching scores for the compared fragments of the respective list entry. A candidate list with the best matching entries from the list of entries based on the total scores of the list entries is generated.Type: ApplicationFiled: February 16, 2010Publication date: August 19, 2010Applicant: NUANCE COMMUNICATIONS, INC.Inventors: Christian Hillebrecht, Markus Schwarz
-
Publication number: 20100198583Abstract: The present invention relates to an indicating method for speech recognition system, comprising a multimedia electronic product and a speech recognition device. The steps of this method include: users enter voice commands into a voice input unit and convert these commands into speech signals, which are acquired and stored by a recording unit, converted by a microprocessor into a volume indicating oscillogram, and then displayed by a display module. At the same time, compliance with speech recognition conditions will be decided in that process.Type: ApplicationFiled: February 4, 2009Publication date: August 5, 2010Applicant: AIBELIVE CO., LTD.Inventors: Chen-Wei Su, Chun-Ping Fang, Min-Ching Wu
-
Publication number: 20100185448Abstract: In embodiments of the present invention improved capabilities are described for interacting with a mobile communication facility comprising receiving a switch activation from a user to initiate a speech recognition recording session, wherein the speech recognition recording session comprises a voice command from the user followed by the speech to be recognized from the user; recording the speech recognition recording session using a mobile communication facility resident capture facility; recognizing at least a portion of the voice command as an indication that user speech for recognition will begin following the end of the at least a portion of the voice command; recognizing the recorded speech using a speech recognition facility to produce an external output; and using the selected output to perform a function on the mobile communication facility.Type: ApplicationFiled: January 21, 2010Publication date: July 22, 2010Inventor: William S. Meisel
-
Publication number: 20100185446Abstract: It is provided a speech recognition system installed in a terminal coupled to a server via a network. The terminal holds map data including a landmark. The speech recognition system manages recognition data including a word corresponding to a name of the landmark, and sends update area information and updated time to the server. The server generates, when recognition data of the area of the update area information sent from the terminal has been changed after updated time, difference data between latest recognition data and recognition data of the update area information at a time of the updated time, and sends the generated difference data and map data of the update area information to the terminal. The terminal updates the map data based on the map data sent from the server. The speech recognition system updates the recognition data managed by the terminal based on the difference data.Type: ApplicationFiled: January 20, 2010Publication date: July 22, 2010Inventors: Takeshi HOMMA, Hiroaki Kokubo, Akinori Asahara, Hisashi Takahashi
-
Publication number: 20100161334Abstract: An utterance verification method for an isolated word N-best speech recognition result includes: calculating log likelihoods of a context-dependent phoneme and an anti-phoneme model based on an N-best speech recognition result for an input utterance; measuring a confidence score of an N-best speech-recognized word using the log likelihoods; calculating distance between phonemes for the N-best speech-recognized word; comparing the confidence score with a threshold and the distance with a predetermined mean of distances; and accepting the N-best speech-recognized word when the compared results for the confidence score and the distance correspond to acceptance.Type: ApplicationFiled: August 4, 2009Publication date: June 24, 2010Applicant: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTEInventors: Jeom Ja Kang, Yunkeun Lee, Jeon Gue Park, Ho-Young Jung, Hyung-Bae Jeon, Hoon Chung, Sung Joo Lee, Euisok Chung, Ji Hyun Wang, Byung Ok Kang, Ki-young Park, Jong Jin Kim
-
Publication number: 20100161335Abstract: A method and apparatus for detecting use of an utterance. A voice session including voice signals generated during a conversation between a first participant and a second participant is monitored by a speech analytics processor. The speech analytics processor detects the use of an utterance. A speech recognition processor channel selected from a pool of speech recognition processor channels and is coupled to the voice session. The speech recognition processor provided speech recognition services to a voice-enabled application. The speech recognition processor channel is then decoupled from the voice session. The speech analytics processor continues to monitor the conversation for subsequent use of the utterance.Type: ApplicationFiled: December 22, 2008Publication date: June 24, 2010Applicant: NORTEL NETWORKS LIMITEDInventor: Stephen Whynot
-
Publication number: 20100154015Abstract: A metadata search apparatus using speech recognition includes a metadata processor for processing contents metadata to obtain allomorph of target vocabulary required for speech recognition and search; a metadata storage unit for storing the contents metadata; a speech recognizer for performing speech recognition on speech data uttered by a user by searching the allomorph of the target vocabulary; a query language processor for extracting a keyword from the vocabulary speech-recognized by the speech recognizer; and a search processor for searching the metadata storage unit to extract the contents metadata corresponding to the keyword. An IPTV receiving apparatus employs the metadata search apparatus to provide IPTV services through the functions of speech recognition.Type: ApplicationFiled: May 7, 2009Publication date: June 17, 2010Applicant: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTEInventors: Byung Ok KANG, Eui Sok CHUNG, Ji Hyun WANG, Yun Keun LEE, Jeom Ja KANG, Jong Jin KIM, Ki-young PARK, Jeon Gue PARK, Sung Joo LEE, Hyung-Bae JEON, Ho-Young JUNG, Hoon CHUNG
-
Publication number: 20100135527Abstract: An image recognition algorithm includes a keypoints-based comparison and a region-based color comparison. A method of identifying a target image using the algorithm includes: receiving an input at a processing device, the input including data related to the target image; performing a retrieving step including retrieving an image from an image database, and, until the image is either accepted or rejected, designating the image as a candidate image; performing an image recognition step including using the processing device to perform an image recognition algorithm on the target and candidate images in order to obtain an image recognition algorithm output; and performing a comparison step including: if the image recognition algorithm output is within a pre-selected range, accepting the candidate image as the target image; and if the image recognition algorithm output is not within the pre-selected range, rejecting the candidate image and repeating the retrieving, image recognition, and comparison steps.Type: ApplicationFiled: December 2, 2008Publication date: June 3, 2010Inventors: Yi Wu, Horst Haussecker, Igor Kozintsev
-
Publication number: 20100128985Abstract: Method for online character recognition of Arabic text, the method including receiving handwritten Arabic text from a user in the form of handwriting strokes, sampling the handwriting strokes to acquire a sequence of two dimensional point representations thereof, with associated temporal data, geometrically pre processing and extracting features on the point representations, detecting delayed strokes and word parts in the pre processed point representations, projecting the delayed strokes onto the body of the word parts, constructing feature vector representations for each word part, thereby generating an observation sequence, and determining the word with maximum probability given the observation sequence, resulting in a list of word probabilities.Type: ApplicationFiled: July 26, 2007Publication date: May 27, 2010Applicant: BGN TECHNOLOGIES LTD.Inventors: Jihad El-Sana, Fadi Biadsy
-
Publication number: 20100121643Abstract: The technology disclosed relates to a system and method for fast, accurate and parallelizable speech search, called Crystal Decoder. It is particularly useful for search applications, as opposed to dictation. It can achieve both speed and accuracy, without sacrificing one for the other. It can search different variations of records in the reference database without a significant increase in elapsed processing time. Even the main decoding part can be parallelized as the number of words increase to maintain a fast response time.Type: ApplicationFiled: November 2, 2009Publication date: May 13, 2010Applicant: Melodis CorporationInventors: Keyvan Mohajer, Seyed Majid Emami, Jon Grossman, Joe Kyaw Soe Aung, Sina Sohangir
-
Publication number: 20100100382Abstract: The disclosure describes a speech detection system for detecting one or more desired speech segments in an audio stream. The speech detection system includes an audio stream input and a speech detection technique. The speech detection technique may be performed in various ways, such as using pattern matching and/or signal processing. The pattern matching implementation may extract features representing types of sounds as in phrases, words, syllables, phonemes and so on. The signal processing implementation may extract spectrally-localized frequency-based features, amplitude-based features, and combinations of the frequency-based and amplitude-based features. Metrics may be obtained and used to determine a desired word in the audio stream. In addition, a keypad stream having keypad entries may be used in determining the desired word.Type: ApplicationFiled: October 16, 2009Publication date: April 22, 2010Inventors: Ashwin P Rao, Gregory M. Aronov, Marat V. Garafutdinov
-
Publication number: 20100100384Abstract: A language processing system may determine a display form of a spoken word by analyzing the spoken form using a language model that includes dictionary entries for display forms of homonyms. The homonyms may include trade names as well as given names and other phrases. The language processing system may receive spoken language and produce a display form of the language while displaying the proper form of the homonym. Such a system may be used in search systems where audio input is converted to a graphical display of a portion of the spoken input.Type: ApplicationFiled: October 21, 2008Publication date: April 22, 2010Applicant: Microsoft CorporationInventors: Yun-Cheng Ju, Julian J. Odell
-
Publication number: 20100076751Abstract: A voice recognition system used for onboard equipment having a genre database (DB) that stores search target vocabularies in accordance with respective genres. It has a mike 1 for outputting speech sounds as spoken data; a first voice recognition dictionary 2a for recognizing words of search target genres in the genre DB; a second voice recognition dictionary 2b for recognizing words outside the search target genres; a voice recognition unit 3 for recognizing the speech sounds by collating the spoken data delivered from the mike with the vocabularies contained in the first and second voice recognition dictionaries; an interactive control unit 4 for outputting, when a word delivered from the voice recognition unit as a recognition result is a word obtained using the second voice recognition dictionary, a message so stating as presentation information; and a presentation unit 5 for presenting the presentation information to an outside.Type: ApplicationFiled: October 14, 2007Publication date: March 25, 2010Inventor: Takayoshi Chikuri
-
Publication number: 20100076764Abstract: A method of dialing phone numbers using an in-vehicle speech recognition system includes receiving speech input at a vehicle, separating the speech input into a word segment and a digit segment, identifying the letters in a word segment, converting the letters in the word segment to digits, and operating an alphanumeric keypad based on the digit speech segment and the converted word segment.Type: ApplicationFiled: September 19, 2008Publication date: March 25, 2010Applicant: General Motors CorporationInventor: Rathinavelu Chengalvarayan
-
Publication number: 20100063818Abstract: This invention is directed to providing voice feedback to a user of an electronic device. Because each electronic device display may include several speakable elements (i.e., elements for which voice feedback is provided), the elements may be ordered. To do so, the electronic device may associate a tier with the display of each speakable element. The electronic device may then provide voice feedback for displayed speakable elements based on the associated tier. To reduce the complexity in designing the voice feedback system, the voice feedback features may be integrated in a Model View Controller (MVC) design used for displaying content to a user. For example, the model and view of the MVC design may include additional variables associated with speakable properties.Type: ApplicationFiled: September 5, 2008Publication date: March 11, 2010Applicant: Apple Inc.Inventors: James Eric Mason, Jesse Boettcher
-
Publication number: 20100045801Abstract: Disclosed are a word-image input system and a word-image input method, and more particularly, to a system for inputting an word-image into an information device and a method thereof that can enable the information device to read the word-image formed by placing a printed matter printed with words having a complicated language system at a proper location and pho-tographing the printed matter at optimized illuminance and distance. According to the present invention, it is possible to prevent the recognition rate for the word-image from being lowered due to shaking, spreading phenomenon and unstable illuminance when the word-image is recognized by the information device provided with the word recognition application.Type: ApplicationFiled: November 29, 2007Publication date: February 25, 2010Applicant: UNICHAL Inc.Inventors: Sang Woo Lee, Choung Suk Oh, Min Uk Kang
-
Publication number: 20100049517Abstract: An automatic answering device and an automatic answering method for automatically answering to a user utterance are configured: to prepare a conversation scenario that is a set of input sentences and replay sentences, the input sentences each corresponding to a user utterance assumed to be uttered by a user, the reply sentences each being an automatic reply to the inputted sentence; to accept a user utterance; to determine the reply sentence to the accepted user utterance on the basis of the conversation scenario; and to present the determined reply sentence to the user. Data of the conversation scenario have a data structure that enables the inputted sentences and the reply sentences to be expressed in a state transition diagram in which each of the inputted sentences is defined as a morphism and the reply sentence corresponding to the inputted sentence is defined as an object.Type: ApplicationFiled: August 17, 2009Publication date: February 25, 2010Applicant: Aruze Corp.Inventors: Shengyang Huang, Hiroshi Katukura
-
Publication number: 20100049518Abstract: A system for providing consistency between the pronunciation of a word by a user and a confirmation pronunciation issued by a voice server (1), said voice server comprising both a voice recognition system (20) suitable for recognizing the pronunciation of the word by the user and for associating a text label therewith, and a speech synthesis system (30) suitable for issuing said confirmation pronunciation on the basis of said text label. The text label is a phonetic text label (21) constructed by concatenating the phonemes of the pronunciation as recognized by the voice recognition system (20).Type: ApplicationFiled: March 29, 2007Publication date: February 25, 2010Applicant: France TelecomInventor: Laurence Ferrieux
-
Publication number: 20100040207Abstract: A system includes an acoustic input engine configured to accept a speech input, to recognize phonemes of the speech input, and to create word strings based on the recognized phonemes. The system includes a semantic engine coupled to the acoustic engine and operable to identify actions and to identify objects by parsing the word strings. The system also includes an action-object pairing system to identify a dominant entry from the identified actions and the identified objects, to select a complement to the dominant entry from the identified actions and the identified objects, and to form an action-object pair that includes the dominant entry and the complement. The system further includes an action-object routing table operable to provide a routing destination based on the action-object pair. The system also includes a call routing module to route a call to the routing destination.Type: ApplicationFiled: October 22, 2009Publication date: February 18, 2010Applicant: AT&T INTELLECTUAL PROPERTY I, L.P.Inventors: Robert R. Bushey, Michael Sabourin, Carl Potvin, Benjamin Anthony Knott, John Mills Martin
-
Publication number: 20100017210Abstract: A system for searching stored audio data is described. The system includes a memory configured to store audio data received from a radio receiver and a processing circuit. The processing circuit is configured to receive a search pattern, search the stored audio data for the search pattern, and provide audio data based on the search.Type: ApplicationFiled: January 7, 2005Publication date: January 21, 2010Inventors: David A. Blaker, Thomas R. Olson, Brian L. Douthitt
-
Publication number: 20100004931Abstract: An apparatus is provided for speech utterance verification. The apparatus is configured to compare a first prosody component from a recorded speech with a second prosody component for a reference speech. The apparatus determines a prosodic verification evaluation for the recorded speech utterance in dependence of the comparison.Type: ApplicationFiled: September 15, 2006Publication date: January 7, 2010Inventors: Bin Ma, Haizhou Li, Minghui Dong
-
Publication number: 20090326945Abstract: An apparatus may include a processor configured to receive vocabulary entry data. The processor may be further configured to determine a class for the received vocabulary entry data. The processor may be additionally configured to identify one or more languages for the vocabulary entry data based upon the determined class. The processor may also be configured to generate a phoneme sequence for the vocabulary entry data for each identified language. Corresponding methods and computer program products are also provided.Type: ApplicationFiled: June 26, 2008Publication date: December 31, 2009Inventor: Jilei Tian
-
Publication number: 20090319272Abstract: A method for voice ordering utilizing catalog taxonomies and hierarchical categorization relationships in product information management (PIM) systems includes: prompting a user with a query to input speech into a speech recognition engine; translating the inputted speech into a series of words; querying a product information management component (PIM) based on the series of words; wherein the querying is performed as a matching algorithm against PIM category and attribute keywords; returning coded results to a voice synthesizer to produce at least one of: a voice response, and a text response to the user; and wherein in the coded results indicate one or more of: a not found message for zero matches, a confirmation of a suitable single match, a request for additional information in the event one or more of the following occurs: more than one matching item, category, and item attribute was found in the PIM.Type: ApplicationFiled: June 30, 2008Publication date: December 24, 2009Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Beatrice Coulomb, Dimitris Gibault, Audrey Occello, Christophe Palaysi
-
Publication number: 20090313019Abstract: An emotion recognition apparatus is capable of performing accurate and stable speech-based emotion recognition, irrespective of individual, regional, and language differences of prosodic information.Type: ApplicationFiled: May 21, 2007Publication date: December 17, 2009Inventors: Yumiko Kato, Takahiro Kamai, Yoshihisa Nakatoh, Yoshifumi Hirose
-
Publication number: 20090305203Abstract: A pronunciation diagnosis device according to the present invention diagnoses the pronunciation of a speaker using articulatory attribute data including articulatory attribute values corresponding to an articulatory attribute of a desirable pronunciation for each phoneme in each audio language system, the articulatory attribute including any one condition of the tongue in the oral cavity, the lips, the vocal cord, the uvula, the nasal cavity, the teeth, and the jaws, or a combination including at least one of the conditions of the articulatory organs; the way of applying force in the conditions of articulatory organs; and a combination of breathing conditions; extracting an acoustic feature from an audio signal generated by a speaker, the acoustic feature being a frequency feature quantity, a sound volume, and a duration time, a rate of change or change pattern thereof, and at least one combination thereof; estimating an attribute value associated with the articulatory attribute on the basis of the extractedType: ApplicationFiled: September 29, 2006Publication date: December 10, 2009Inventors: Machi Okumura, Hiroaki Kojima, Hiroshi Omura
-
Publication number: 20090306983Abstract: Systems and methods for enabling user access and update of personal health records stored in a health data store via voice inputs are provided. The system may include a computer program having a recognizer module configured to process structured word data of a user voice input received from a voice platform, to produce a set of tagged structured word data based on a healthcare-specific glossary. The computer program may further include a health data store interface configured to apply a rule set to the tagged structured word data to produce a query to the health data store and receive a response from the health data store based on the query, and a grammar generator configured to generate a reply sentence based on the response received from the health data store and pass the reply sentence to the voice platform to be played as a voice reply to the user.Type: ApplicationFiled: June 9, 2008Publication date: December 10, 2009Applicant: Microsoft CorporationInventor: Vaibhav Bhandari
-
Publication number: 20090299733Abstract: A method for creating and editing an XML-based speech synthesis document for input to a text-to-speech engine is provided. The method includes recording voice utterances of a user reading a pre-selected text and parsing the recorded voice utterances into individual words and periods of silence. The method also includes recording a synthesized speech output generated by a text-to-speech engine, the synthesized speech output being an audible rendering of the pre-selected text, and parsing the synthesized speech output into individual words and periods of silence. The method further includes annotating the XML-based speech synthesis document based upon a comparison of the recorded voice utterances and the recorded synthesized speech output.Type: ApplicationFiled: June 3, 2008Publication date: December 3, 2009Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Ciprian Agapi, Oswaldo Gago, Maria Elena Smith, Roberto Vila
-
Publication number: 20090292541Abstract: Methods and apparatus for the enhancement of speech to text engines, by providing indications to the correctness of the found words, based on additional sources besides the internal indication provided by the STT engine. The enhanced indications comprise sources of data such as acoustic features, CTI features, phonetic search and others. The apparatus and methods also enable the detection of important or significant keywords found in audio files, thus enabling more efficient usages, such as further processing or transfer of interactions to relevant agents, escalation of issues, or the like. The methods and apparatus employ a training phase in which word model and key phrase model are generated for determining an enhanced correctness indication for a word and an enhanced importance indication for a key phrase, based on the additional features.Type: ApplicationFiled: May 25, 2008Publication date: November 26, 2009Applicant: Nice Systems Ltd.Inventors: Ezra Daya, Oren Pereg, Yuval Lubowich, Moshe Wasserblat
-
Publication number: 20090292540Abstract: A method including displaying content on a display of a device, receiving a speech input designating a segment of the content to be excerpted and transferring the excerpted content to a predetermined location for storage and retrieval.Type: ApplicationFiled: May 22, 2008Publication date: November 26, 2009Applicant: NOKIA CORPORATIONInventors: Huanglingzi Liu, Yue Zhong Tang, Yu Zhang
-
Publication number: 20090292531Abstract: Streaming voice signals, such as might be received at a contact center or similar operation, are analyzed to detect the occurrence of one or more unprompted, predetermined utterances. The predetermined utterances preferably constitute a vocabulary of words and/or phrases having particular meaning within the context in which they are uttered. Detection of one or more of the predetermined utterances during a call causes a determination of response-determinative significance of the detected utterance(s). Based on the response-determinative significance of the detected utterance(s), a responsive action may be further determined. Additionally, long term storage of the call corresponding to the detected utterance may also be initiated. Conversely, calls in which no predetermined utterances are detected may be deleted from short term storage.Type: ApplicationFiled: May 22, 2009Publication date: November 26, 2009Applicant: ACCENTURE GLOBAL SERVICES GMBHInventors: Thomas J. Ryan, Biji K. Janan
-
Publication number: 20090281807Abstract: A voice quality conversion device converts voice quality of an input speech using information of the speech.Type: ApplicationFiled: May 8, 2008Publication date: November 12, 2009Inventors: Yoshifumi Hirose, Takahiro Kamai, Yumiko Kato
-
Publication number: 20090276219Abstract: In the present invention, a voice input system and a voice input method are provided. The voice input method includes the steps of: (A) initiating a speech recognition process by a first input associated with a first parameter of a first speech recognition subject; (B) providing a voice and a searching space constructed by a speech recognition model associated with the first speech recognition subject; (C) obtaining a sub-searching space from the searching space based on the first parameter; (D) searching at least one candidate item associated with the voice from the sub-searching space; and (E) showing the at least one candidate item.Type: ApplicationFiled: April 29, 2009Publication date: November 5, 2009Applicant: DELTA ELECTRONICS, INC.Inventors: Keng-Hung Yeh, Liang-Sheng Huang, Chao-Jen Huang, Jia-Lin Shen
-
Publication number: 20090276223Abstract: An administration method and system. The method includes receiving by a computing system, a telephone call from an administrator. The computing system presents an audible menu associated with a plurality of computers to the administrator. The computing system receives from the administrator, an audible selection for a computer from the audible menu. The computing system receives from the administrator, an audible verbal command for performing a maintenance operation on the computer. The computing system executes the maintenance operation on the computer. The computing system receives from the computer, confirmation data indicating that the maintenance operation has been completed. The computing system converts the confirmation data into an audible verbal message. The computing system transmits the second audible verbal message to the administrator.Type: ApplicationFiled: May 1, 2008Publication date: November 5, 2009Inventors: Peeyush Jaiswal, Naveen Narayan
-
Publication number: 20090276220Abstract: A system evaluates a hands free communication system. The system automatically selects a consonant-vowel-consonant (CVC), vowel-consonant-vowel (VCV), or other combination of sounds from an intelligent database. The selection is transmitted with another communication stream that temporally overlaps the selection. The quality of the communication system is evaluated through an automatic speech recognition engine. The evaluation occurs at a location remote from the transmitted selection.Type: ApplicationFiled: April 29, 2009Publication date: November 5, 2009Inventors: Shreyas Paranjpe, Mark Fallat
-
Publication number: 20090254344Abstract: A spoken language understanding method and system are provided. The method includes classifying a set of labeled candidate utterances based on a previously trained classifier, generating classification types for each candidate utterance, receiving confidence scores for the classification types from the trained classifier, sorting the classified utterances based on an analysis of the confidence score of each candidate utterance compared to a respective label of the candidate utterance, and rechecking candidate utterances according to the analysis. The system includes modules configured to control a processor in the system to perform the steps of the method.Type: ApplicationFiled: June 16, 2009Publication date: October 8, 2009Applicant: AT&T Corp.Inventors: Dilek Z. Hakkani-Tur, Mazin G. Rahim, Gokhan Tur
-
Publication number: 20090254542Abstract: Search methods and systems are provided. First, a voice is received. The voice is recognized using voice recognition technology to obtain at least two keywords, and a specific keyword combination is generated according to the keywords. A specific requirement among a plurality of requirements is associated according to the specific keyword combination, wherein each requirement corresponds to at least one of a plurality of keyword combinations. Then, according to the specific requirement, a database is searched to obtain at least candidate data conforming to the specific requirement.Type: ApplicationFiled: January 12, 2009Publication date: October 8, 2009Applicant: MITAC INTERNATIONAL CORP.Inventor: Fang-Wei Kang
-
Publication number: 20090240500Abstract: A speech recognition apparatus includes a storage unit which store vocabularies, each of vocabularies including plural word body data, each of the word body data obtained by removing a specific word head from a word or sentence, and store at least one word head portion including labeled nodes to express at least one common word head common to at least two of the vocabularies, an instruction receiving unit which receive an instruction of a target vocabulary and an instruction of a operation, a grammar network generating unit which generate, when adding is instructed, a grammar network containing the word head portion, the target vocabulary and connection information indicating that each of the word body data contained in the target vocabulary is connected to a specific one of the labeled nodes contained in the word head portion, and a speech recognition unit which execute speech recognition using the generated grammar network.Type: ApplicationFiled: March 19, 2009Publication date: September 24, 2009Inventors: Mitsuyoshi TACHIMORI, Shinichi TANAKA
-
Publication number: 20090234650Abstract: A spoken dialogue apparatus which retrieves desired data from a data storage unit storing data in association with attribute values concerning a plurality of attributes of the data, includes a calculation unit configured to calculate estimated numbers of data as costs concerning a first response and a second response based on the retrieval condition and the retrieval result, the data being included in retrieval result after narrowed retrieval performed in accordance with an input speech corresponding to the first response asking the user to input an attribute value of an attribute which has not been input and corresponding to the second response ascertaining the truth of an attribute value of an attribute which has been input, and a selection unit configured to select a response having the lowest cost and present the response to the user.Type: ApplicationFiled: March 11, 2009Publication date: September 17, 2009Inventor: Takehide YANO
-
Publication number: 20090234651Abstract: A speech recognition system includes a mobile device and a remote server. The mobile device receives the speech from the user and extracts the features and phonemes from the speech. Selected phonemes and measures of uncertainty are transmitted to the server, which processes the phonemes for speech understanding and transmits a text of the speech (or the context or understanding of the speech) back to the mobile device.Type: ApplicationFiled: March 12, 2009Publication date: September 17, 2009Inventors: Otman A. Basir, William Ben Miners
-
Publication number: 20090234847Abstract: Provided is an information retrieval method including: retrieving, by a computer, a name including input characters from a database for storing the name, an attribute word associated with the name, and a degree of relevance between the name and the attribute word; outputting the retrieved name as a candidate name; and extracting an attribute word associated with the candidate name, the extracting including: calculating a degree of independency indicating a degree of difference between the extracted attribute words, a degree of coverage indicating an extent to which the combination of the extracted attribute words covers the candidate names, and a degree of equality of a number of corresponding candidate names for each attribute word; and calculating a score of the combination of the attribute words based on at least one of the independency, the coverage and the equality to output the combinations of the attribute words to an output unit.Type: ApplicationFiled: February 10, 2009Publication date: September 17, 2009Inventors: Takeshi HOMMA, Hiroaki Kokubo, Hisashi Takahashi
-
Publication number: 20090222267Abstract: The invention relates to a task classification system (900) that interacts with a user. The task classification system (900) may include a recognizer (920) that may recognize symbols in the user's input communication, and a natural language understanding unit (900) that may determine whether the user's input communication can be understood. If the user's input communication can be understood, the natural language understanding unit (930) may generate understanding data. The system may also include a communicative goal generator that may generate communicative goals based on the symbols recognized by the recognizer (920) and understanding data from the natural language understanding unit (930). The generated communicative goals may be related to information needed to be obtained from the user.Type: ApplicationFiled: February 26, 2009Publication date: September 3, 2009Applicant: AT&T Corp.Inventors: Marilyn A. Walker, Owen Christopher Rambow, Monica Rogati
-
Publication number: 20090216525Abstract: A system and method for homonym treatment in a speech recognition system and method are provided. The system and method for homonym treatment in a speech recognition system may be used in a mobile wireless communication devices that are voice operated after their initial activation.Type: ApplicationFiled: February 20, 2009Publication date: August 27, 2009Applicant: Vocera Communications, Inc.Inventor: Robert E. Shostak
-
Publication number: 20090203427Abstract: A slot machine 1 of the present invention makes a control so as to: sequentially store the number of game values consumed per unit game; sequentially store the number of game values given per unit game; calculating a difference between the total number of game values given and the total number of game values consumed, as a self game value difference; transmitting the self game value difference to outside; receiving someone's game value difference from outside; when the self game value difference and the someone's game value difference are in a predetermined relationship, voice-outputting, by the conversation controller 91, an answer at volume corresponding to the predetermined relationship from the speaker 23 in response to a voice input through the microphone 90; and delete the stored numbers of game values given and consumed, under a predetermined condition.Type: ApplicationFiled: January 21, 2009Publication date: August 13, 2009Applicant: ARUZE GAMING AMERICA, INC.Inventor: Kazuo OKADA
-
Publication number: 20090199235Abstract: Embodiments of a computer-implemented advertisement display system are disclosed. In one embodiment, the system includes a speech recognition component that processes a video clip and produces a corresponding collection of speech recognition data indicative of an audio portion of the video clip. The system also includes a collection of advertising material. An advertisement selection component selects an advertisement from the collection of advertising material based on the corresponding collection of speech recognition data. The system also includes a display. An advertisement presentation component displays an indication of the selected advertisement on the display during a simultaneous display of the video clip.Type: ApplicationFiled: June 6, 2008Publication date: August 6, 2009Applicant: MICROSOFT CORPORATIONInventors: Arungunram C. Surendran, Patrick Nguyen, Milind V. Mahajan