Segmentation Or Word Limit Detection (epo) Patents (Class 704/E15.005)
  • Publication number: 20100256978
    Abstract: A method for performing speech recognition relating to an object for the purpose of affecting automatic processing of the object by a processing system. The object carries information with at least a character string of processing information. The character string spoken by an operator is processed by way of a speech recognition procedure to generate a first result. Based on the need for more information of an element of the first result additional processing data is requested. An operator's response generates a second result. The first result is then modified to achieve consistency with the operator's response.
    Type: Application
    Filed: April 6, 2010
    Publication date: October 7, 2010
    Applicant: SIEMENS AKTIENGESELLSCHAFT
    Inventor: Walter Rosenbaum
  • Publication number: 20100250251
    Abstract: Architecture that suppresses the unexpected appearance of words by applying appropriate restrictions to long-term and short-term memory. The quickness of adaptation is also realized by leveraging the restriction. The architecture includes a history component for processing user input history for conversion of a phonetic string by a conversion process that output conversion results, and an adaptation component for adapting the conversion process to the user input history based on restriction(s) applied to short-term memory that impacts word appearances during the conversion process. The architecture performs probability boosting based on context-dependent probability differences (short-term memory), and dynamic linear-interpolation between long-term memory and baseline language model based on frequency of preceding context of word (long-term memory).
    Type: Application
    Filed: March 30, 2009
    Publication date: September 30, 2010
    Applicant: Microsoft Corporation
    Inventors: Katsutoshi Ohtsuki, Takashi Umeoka
  • Publication number: 20100229116
    Abstract: A control apparatus includes a voice recognition unit for recognizing user utterance to output a recognized word, a function storage unit for determining and storing a desired function that corresponds to the recognized word, a detector for detecting a preset user operation, a button display unit for displaying on a screen a shortcut button that instructs execution of the desired function stored in the storage unit when the detector detects the preset user operation, and a control unit for controlling execution of the desired function when the shortcut button is operated. By storing the desired function in association with the recognized word and by detecting user instruction, the control apparatus displays a shortcut button for a necessary function only.
    Type: Application
    Filed: March 4, 2010
    Publication date: September 9, 2010
    Applicant: DENSO CORPORATION
    Inventors: Fumihiko Murase, Ichiro Akahori, Shinji Niwa
  • Publication number: 20100223056
    Abstract: A method, apparatus, and system are described for a continuous speech recognition engine that includes a fine speech recognizer model, a coarse sound representation generator, and a coarse match generator. The fine speech recognizer model receives a time coded sequence of sound feature frames, applies a speech recognition process to the sound feature frames and determines at least a best guess at each recognizable word that corresponds to the sound feature frames. The coarse sound representation generator generates a coarse sound representation of the recognized word. The coarse match generator determines a likelihood of the coarse sound representation actually being the recognized word based on comparing the coarse sound representation of the recognized word to a database containing the known sound of that recognized word and assigns the likelihood as a robust confidence level parameter to that recognized word.
    Type: Application
    Filed: February 27, 2009
    Publication date: September 2, 2010
    Applicant: AUTONOMY CORPORATION LTD.
    Inventor: Mahapathy Kadirkamanathan
  • Publication number: 20100217597
    Abstract: Systems and methods for using an annotation guide to label utterances and speech data with a call type are disclosed. A method embodiment monitors labelers of speech data by presenting via a processor a test utterance to a labeler, receiving input from the labeler that selects a particular call type from a list of call types and determining via the processor if the labeler labeled the test utterance correctly. Based on the determining step, the method performs at least one of the following: revising the annotation guide, retraining the labeler or altering the test utterance.
    Type: Application
    Filed: May 3, 2010
    Publication date: August 26, 2010
    Applicant: AT&T Intellectual Property II, LP via a transfer from AT&T Corp.
    Inventors: Lee Begeja, Richard Vandervoort Cox, Harris Drucker, David Crawford Gibbon, Allen Louis Gorin, Patrick Guy Haffner, Steven H. Lewis, Zhu Liu, Mazin G. Rahim, Bernard S. Renger, Behzad Shahraray
  • Publication number: 20100217598
    Abstract: A speech recognition system in which, even when the user makes an utterance including a word that satisfies a predetermined condition such as an unknown word, such a fact can be presented to the user, and the user can confirm the fact easily, is provided.
    Type: Application
    Filed: February 22, 2007
    Publication date: August 26, 2010
    Applicant: NEC CORPORATION
    Inventor: Fumihiro Adachi
  • Publication number: 20100211390
    Abstract: The present invention relates to a method of generating a candidate list from a list of entries in accordance with a string of subword units corresponding to a speech input in a speech recognition system, the list of entries including plural list entries each comprising at least one fragment having one or more subword units. For each list entry, the fragments of the list entry are compared with the string of subword units. A matching score for each of the compared fragments based on the comparison is determined. The matching score for a fragment is further based on a comparison of at least one other fragment of the same list entry with the string of subword units. A total score for each list entry is determined based on the matching scores for the compared fragments of the respective list entry. A candidate list with the best matching entries from the list of entries based on the total scores of the list entries is generated.
    Type: Application
    Filed: February 16, 2010
    Publication date: August 19, 2010
    Applicant: NUANCE COMMUNICATIONS, INC.
    Inventors: Christian Hillebrecht, Markus Schwarz
  • Publication number: 20100198583
    Abstract: The present invention relates to an indicating method for speech recognition system, comprising a multimedia electronic product and a speech recognition device. The steps of this method include: users enter voice commands into a voice input unit and convert these commands into speech signals, which are acquired and stored by a recording unit, converted by a microprocessor into a volume indicating oscillogram, and then displayed by a display module. At the same time, compliance with speech recognition conditions will be decided in that process.
    Type: Application
    Filed: February 4, 2009
    Publication date: August 5, 2010
    Applicant: AIBELIVE CO., LTD.
    Inventors: Chen-Wei Su, Chun-Ping Fang, Min-Ching Wu
  • Publication number: 20100185448
    Abstract: In embodiments of the present invention improved capabilities are described for interacting with a mobile communication facility comprising receiving a switch activation from a user to initiate a speech recognition recording session, wherein the speech recognition recording session comprises a voice command from the user followed by the speech to be recognized from the user; recording the speech recognition recording session using a mobile communication facility resident capture facility; recognizing at least a portion of the voice command as an indication that user speech for recognition will begin following the end of the at least a portion of the voice command; recognizing the recorded speech using a speech recognition facility to produce an external output; and using the selected output to perform a function on the mobile communication facility.
    Type: Application
    Filed: January 21, 2010
    Publication date: July 22, 2010
    Inventor: William S. Meisel
  • Publication number: 20100185446
    Abstract: It is provided a speech recognition system installed in a terminal coupled to a server via a network. The terminal holds map data including a landmark. The speech recognition system manages recognition data including a word corresponding to a name of the landmark, and sends update area information and updated time to the server. The server generates, when recognition data of the area of the update area information sent from the terminal has been changed after updated time, difference data between latest recognition data and recognition data of the update area information at a time of the updated time, and sends the generated difference data and map data of the update area information to the terminal. The terminal updates the map data based on the map data sent from the server. The speech recognition system updates the recognition data managed by the terminal based on the difference data.
    Type: Application
    Filed: January 20, 2010
    Publication date: July 22, 2010
    Inventors: Takeshi HOMMA, Hiroaki Kokubo, Akinori Asahara, Hisashi Takahashi
  • Publication number: 20100161334
    Abstract: An utterance verification method for an isolated word N-best speech recognition result includes: calculating log likelihoods of a context-dependent phoneme and an anti-phoneme model based on an N-best speech recognition result for an input utterance; measuring a confidence score of an N-best speech-recognized word using the log likelihoods; calculating distance between phonemes for the N-best speech-recognized word; comparing the confidence score with a threshold and the distance with a predetermined mean of distances; and accepting the N-best speech-recognized word when the compared results for the confidence score and the distance correspond to acceptance.
    Type: Application
    Filed: August 4, 2009
    Publication date: June 24, 2010
    Applicant: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Jeom Ja Kang, Yunkeun Lee, Jeon Gue Park, Ho-Young Jung, Hyung-Bae Jeon, Hoon Chung, Sung Joo Lee, Euisok Chung, Ji Hyun Wang, Byung Ok Kang, Ki-young Park, Jong Jin Kim
  • Publication number: 20100161335
    Abstract: A method and apparatus for detecting use of an utterance. A voice session including voice signals generated during a conversation between a first participant and a second participant is monitored by a speech analytics processor. The speech analytics processor detects the use of an utterance. A speech recognition processor channel selected from a pool of speech recognition processor channels and is coupled to the voice session. The speech recognition processor provided speech recognition services to a voice-enabled application. The speech recognition processor channel is then decoupled from the voice session. The speech analytics processor continues to monitor the conversation for subsequent use of the utterance.
    Type: Application
    Filed: December 22, 2008
    Publication date: June 24, 2010
    Applicant: NORTEL NETWORKS LIMITED
    Inventor: Stephen Whynot
  • Publication number: 20100154015
    Abstract: A metadata search apparatus using speech recognition includes a metadata processor for processing contents metadata to obtain allomorph of target vocabulary required for speech recognition and search; a metadata storage unit for storing the contents metadata; a speech recognizer for performing speech recognition on speech data uttered by a user by searching the allomorph of the target vocabulary; a query language processor for extracting a keyword from the vocabulary speech-recognized by the speech recognizer; and a search processor for searching the metadata storage unit to extract the contents metadata corresponding to the keyword. An IPTV receiving apparatus employs the metadata search apparatus to provide IPTV services through the functions of speech recognition.
    Type: Application
    Filed: May 7, 2009
    Publication date: June 17, 2010
    Applicant: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Byung Ok KANG, Eui Sok CHUNG, Ji Hyun WANG, Yun Keun LEE, Jeom Ja KANG, Jong Jin KIM, Ki-young PARK, Jeon Gue PARK, Sung Joo LEE, Hyung-Bae JEON, Ho-Young JUNG, Hoon CHUNG
  • Publication number: 20100135527
    Abstract: An image recognition algorithm includes a keypoints-based comparison and a region-based color comparison. A method of identifying a target image using the algorithm includes: receiving an input at a processing device, the input including data related to the target image; performing a retrieving step including retrieving an image from an image database, and, until the image is either accepted or rejected, designating the image as a candidate image; performing an image recognition step including using the processing device to perform an image recognition algorithm on the target and candidate images in order to obtain an image recognition algorithm output; and performing a comparison step including: if the image recognition algorithm output is within a pre-selected range, accepting the candidate image as the target image; and if the image recognition algorithm output is not within the pre-selected range, rejecting the candidate image and repeating the retrieving, image recognition, and comparison steps.
    Type: Application
    Filed: December 2, 2008
    Publication date: June 3, 2010
    Inventors: Yi Wu, Horst Haussecker, Igor Kozintsev
  • Publication number: 20100128985
    Abstract: Method for online character recognition of Arabic text, the method including receiving handwritten Arabic text from a user in the form of handwriting strokes, sampling the handwriting strokes to acquire a sequence of two dimensional point representations thereof, with associated temporal data, geometrically pre processing and extracting features on the point representations, detecting delayed strokes and word parts in the pre processed point representations, projecting the delayed strokes onto the body of the word parts, constructing feature vector representations for each word part, thereby generating an observation sequence, and determining the word with maximum probability given the observation sequence, resulting in a list of word probabilities.
    Type: Application
    Filed: July 26, 2007
    Publication date: May 27, 2010
    Applicant: BGN TECHNOLOGIES LTD.
    Inventors: Jihad El-Sana, Fadi Biadsy
  • Publication number: 20100121643
    Abstract: The technology disclosed relates to a system and method for fast, accurate and parallelizable speech search, called Crystal Decoder. It is particularly useful for search applications, as opposed to dictation. It can achieve both speed and accuracy, without sacrificing one for the other. It can search different variations of records in the reference database without a significant increase in elapsed processing time. Even the main decoding part can be parallelized as the number of words increase to maintain a fast response time.
    Type: Application
    Filed: November 2, 2009
    Publication date: May 13, 2010
    Applicant: Melodis Corporation
    Inventors: Keyvan Mohajer, Seyed Majid Emami, Jon Grossman, Joe Kyaw Soe Aung, Sina Sohangir
  • Publication number: 20100100382
    Abstract: The disclosure describes a speech detection system for detecting one or more desired speech segments in an audio stream. The speech detection system includes an audio stream input and a speech detection technique. The speech detection technique may be performed in various ways, such as using pattern matching and/or signal processing. The pattern matching implementation may extract features representing types of sounds as in phrases, words, syllables, phonemes and so on. The signal processing implementation may extract spectrally-localized frequency-based features, amplitude-based features, and combinations of the frequency-based and amplitude-based features. Metrics may be obtained and used to determine a desired word in the audio stream. In addition, a keypad stream having keypad entries may be used in determining the desired word.
    Type: Application
    Filed: October 16, 2009
    Publication date: April 22, 2010
    Inventors: Ashwin P Rao, Gregory M. Aronov, Marat V. Garafutdinov
  • Publication number: 20100100384
    Abstract: A language processing system may determine a display form of a spoken word by analyzing the spoken form using a language model that includes dictionary entries for display forms of homonyms. The homonyms may include trade names as well as given names and other phrases. The language processing system may receive spoken language and produce a display form of the language while displaying the proper form of the homonym. Such a system may be used in search systems where audio input is converted to a graphical display of a portion of the spoken input.
    Type: Application
    Filed: October 21, 2008
    Publication date: April 22, 2010
    Applicant: Microsoft Corporation
    Inventors: Yun-Cheng Ju, Julian J. Odell
  • Publication number: 20100076751
    Abstract: A voice recognition system used for onboard equipment having a genre database (DB) that stores search target vocabularies in accordance with respective genres. It has a mike 1 for outputting speech sounds as spoken data; a first voice recognition dictionary 2a for recognizing words of search target genres in the genre DB; a second voice recognition dictionary 2b for recognizing words outside the search target genres; a voice recognition unit 3 for recognizing the speech sounds by collating the spoken data delivered from the mike with the vocabularies contained in the first and second voice recognition dictionaries; an interactive control unit 4 for outputting, when a word delivered from the voice recognition unit as a recognition result is a word obtained using the second voice recognition dictionary, a message so stating as presentation information; and a presentation unit 5 for presenting the presentation information to an outside.
    Type: Application
    Filed: October 14, 2007
    Publication date: March 25, 2010
    Inventor: Takayoshi Chikuri
  • Publication number: 20100076764
    Abstract: A method of dialing phone numbers using an in-vehicle speech recognition system includes receiving speech input at a vehicle, separating the speech input into a word segment and a digit segment, identifying the letters in a word segment, converting the letters in the word segment to digits, and operating an alphanumeric keypad based on the digit speech segment and the converted word segment.
    Type: Application
    Filed: September 19, 2008
    Publication date: March 25, 2010
    Applicant: General Motors Corporation
    Inventor: Rathinavelu Chengalvarayan
  • Publication number: 20100063818
    Abstract: This invention is directed to providing voice feedback to a user of an electronic device. Because each electronic device display may include several speakable elements (i.e., elements for which voice feedback is provided), the elements may be ordered. To do so, the electronic device may associate a tier with the display of each speakable element. The electronic device may then provide voice feedback for displayed speakable elements based on the associated tier. To reduce the complexity in designing the voice feedback system, the voice feedback features may be integrated in a Model View Controller (MVC) design used for displaying content to a user. For example, the model and view of the MVC design may include additional variables associated with speakable properties.
    Type: Application
    Filed: September 5, 2008
    Publication date: March 11, 2010
    Applicant: Apple Inc.
    Inventors: James Eric Mason, Jesse Boettcher
  • Publication number: 20100045801
    Abstract: Disclosed are a word-image input system and a word-image input method, and more particularly, to a system for inputting an word-image into an information device and a method thereof that can enable the information device to read the word-image formed by placing a printed matter printed with words having a complicated language system at a proper location and pho-tographing the printed matter at optimized illuminance and distance. According to the present invention, it is possible to prevent the recognition rate for the word-image from being lowered due to shaking, spreading phenomenon and unstable illuminance when the word-image is recognized by the information device provided with the word recognition application.
    Type: Application
    Filed: November 29, 2007
    Publication date: February 25, 2010
    Applicant: UNICHAL Inc.
    Inventors: Sang Woo Lee, Choung Suk Oh, Min Uk Kang
  • Publication number: 20100049517
    Abstract: An automatic answering device and an automatic answering method for automatically answering to a user utterance are configured: to prepare a conversation scenario that is a set of input sentences and replay sentences, the input sentences each corresponding to a user utterance assumed to be uttered by a user, the reply sentences each being an automatic reply to the inputted sentence; to accept a user utterance; to determine the reply sentence to the accepted user utterance on the basis of the conversation scenario; and to present the determined reply sentence to the user. Data of the conversation scenario have a data structure that enables the inputted sentences and the reply sentences to be expressed in a state transition diagram in which each of the inputted sentences is defined as a morphism and the reply sentence corresponding to the inputted sentence is defined as an object.
    Type: Application
    Filed: August 17, 2009
    Publication date: February 25, 2010
    Applicant: Aruze Corp.
    Inventors: Shengyang Huang, Hiroshi Katukura
  • Publication number: 20100049518
    Abstract: A system for providing consistency between the pronunciation of a word by a user and a confirmation pronunciation issued by a voice server (1), said voice server comprising both a voice recognition system (20) suitable for recognizing the pronunciation of the word by the user and for associating a text label therewith, and a speech synthesis system (30) suitable for issuing said confirmation pronunciation on the basis of said text label. The text label is a phonetic text label (21) constructed by concatenating the phonemes of the pronunciation as recognized by the voice recognition system (20).
    Type: Application
    Filed: March 29, 2007
    Publication date: February 25, 2010
    Applicant: France Telecom
    Inventor: Laurence Ferrieux
  • Publication number: 20100040207
    Abstract: A system includes an acoustic input engine configured to accept a speech input, to recognize phonemes of the speech input, and to create word strings based on the recognized phonemes. The system includes a semantic engine coupled to the acoustic engine and operable to identify actions and to identify objects by parsing the word strings. The system also includes an action-object pairing system to identify a dominant entry from the identified actions and the identified objects, to select a complement to the dominant entry from the identified actions and the identified objects, and to form an action-object pair that includes the dominant entry and the complement. The system further includes an action-object routing table operable to provide a routing destination based on the action-object pair. The system also includes a call routing module to route a call to the routing destination.
    Type: Application
    Filed: October 22, 2009
    Publication date: February 18, 2010
    Applicant: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Robert R. Bushey, Michael Sabourin, Carl Potvin, Benjamin Anthony Knott, John Mills Martin
  • Publication number: 20100017210
    Abstract: A system for searching stored audio data is described. The system includes a memory configured to store audio data received from a radio receiver and a processing circuit. The processing circuit is configured to receive a search pattern, search the stored audio data for the search pattern, and provide audio data based on the search.
    Type: Application
    Filed: January 7, 2005
    Publication date: January 21, 2010
    Inventors: David A. Blaker, Thomas R. Olson, Brian L. Douthitt
  • Publication number: 20100004931
    Abstract: An apparatus is provided for speech utterance verification. The apparatus is configured to compare a first prosody component from a recorded speech with a second prosody component for a reference speech. The apparatus determines a prosodic verification evaluation for the recorded speech utterance in dependence of the comparison.
    Type: Application
    Filed: September 15, 2006
    Publication date: January 7, 2010
    Inventors: Bin Ma, Haizhou Li, Minghui Dong
  • Publication number: 20090326945
    Abstract: An apparatus may include a processor configured to receive vocabulary entry data. The processor may be further configured to determine a class for the received vocabulary entry data. The processor may be additionally configured to identify one or more languages for the vocabulary entry data based upon the determined class. The processor may also be configured to generate a phoneme sequence for the vocabulary entry data for each identified language. Corresponding methods and computer program products are also provided.
    Type: Application
    Filed: June 26, 2008
    Publication date: December 31, 2009
    Inventor: Jilei Tian
  • Publication number: 20090319272
    Abstract: A method for voice ordering utilizing catalog taxonomies and hierarchical categorization relationships in product information management (PIM) systems includes: prompting a user with a query to input speech into a speech recognition engine; translating the inputted speech into a series of words; querying a product information management component (PIM) based on the series of words; wherein the querying is performed as a matching algorithm against PIM category and attribute keywords; returning coded results to a voice synthesizer to produce at least one of: a voice response, and a text response to the user; and wherein in the coded results indicate one or more of: a not found message for zero matches, a confirmation of a suitable single match, a request for additional information in the event one or more of the following occurs: more than one matching item, category, and item attribute was found in the PIM.
    Type: Application
    Filed: June 30, 2008
    Publication date: December 24, 2009
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Beatrice Coulomb, Dimitris Gibault, Audrey Occello, Christophe Palaysi
  • Publication number: 20090313019
    Abstract: An emotion recognition apparatus is capable of performing accurate and stable speech-based emotion recognition, irrespective of individual, regional, and language differences of prosodic information.
    Type: Application
    Filed: May 21, 2007
    Publication date: December 17, 2009
    Inventors: Yumiko Kato, Takahiro Kamai, Yoshihisa Nakatoh, Yoshifumi Hirose
  • Publication number: 20090305203
    Abstract: A pronunciation diagnosis device according to the present invention diagnoses the pronunciation of a speaker using articulatory attribute data including articulatory attribute values corresponding to an articulatory attribute of a desirable pronunciation for each phoneme in each audio language system, the articulatory attribute including any one condition of the tongue in the oral cavity, the lips, the vocal cord, the uvula, the nasal cavity, the teeth, and the jaws, or a combination including at least one of the conditions of the articulatory organs; the way of applying force in the conditions of articulatory organs; and a combination of breathing conditions; extracting an acoustic feature from an audio signal generated by a speaker, the acoustic feature being a frequency feature quantity, a sound volume, and a duration time, a rate of change or change pattern thereof, and at least one combination thereof; estimating an attribute value associated with the articulatory attribute on the basis of the extracted
    Type: Application
    Filed: September 29, 2006
    Publication date: December 10, 2009
    Inventors: Machi Okumura, Hiroaki Kojima, Hiroshi Omura
  • Publication number: 20090306983
    Abstract: Systems and methods for enabling user access and update of personal health records stored in a health data store via voice inputs are provided. The system may include a computer program having a recognizer module configured to process structured word data of a user voice input received from a voice platform, to produce a set of tagged structured word data based on a healthcare-specific glossary. The computer program may further include a health data store interface configured to apply a rule set to the tagged structured word data to produce a query to the health data store and receive a response from the health data store based on the query, and a grammar generator configured to generate a reply sentence based on the response received from the health data store and pass the reply sentence to the voice platform to be played as a voice reply to the user.
    Type: Application
    Filed: June 9, 2008
    Publication date: December 10, 2009
    Applicant: Microsoft Corporation
    Inventor: Vaibhav Bhandari
  • Publication number: 20090299733
    Abstract: A method for creating and editing an XML-based speech synthesis document for input to a text-to-speech engine is provided. The method includes recording voice utterances of a user reading a pre-selected text and parsing the recorded voice utterances into individual words and periods of silence. The method also includes recording a synthesized speech output generated by a text-to-speech engine, the synthesized speech output being an audible rendering of the pre-selected text, and parsing the synthesized speech output into individual words and periods of silence. The method further includes annotating the XML-based speech synthesis document based upon a comparison of the recorded voice utterances and the recorded synthesized speech output.
    Type: Application
    Filed: June 3, 2008
    Publication date: December 3, 2009
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Ciprian Agapi, Oswaldo Gago, Maria Elena Smith, Roberto Vila
  • Publication number: 20090292541
    Abstract: Methods and apparatus for the enhancement of speech to text engines, by providing indications to the correctness of the found words, based on additional sources besides the internal indication provided by the STT engine. The enhanced indications comprise sources of data such as acoustic features, CTI features, phonetic search and others. The apparatus and methods also enable the detection of important or significant keywords found in audio files, thus enabling more efficient usages, such as further processing or transfer of interactions to relevant agents, escalation of issues, or the like. The methods and apparatus employ a training phase in which word model and key phrase model are generated for determining an enhanced correctness indication for a word and an enhanced importance indication for a key phrase, based on the additional features.
    Type: Application
    Filed: May 25, 2008
    Publication date: November 26, 2009
    Applicant: Nice Systems Ltd.
    Inventors: Ezra Daya, Oren Pereg, Yuval Lubowich, Moshe Wasserblat
  • Publication number: 20090292540
    Abstract: A method including displaying content on a display of a device, receiving a speech input designating a segment of the content to be excerpted and transferring the excerpted content to a predetermined location for storage and retrieval.
    Type: Application
    Filed: May 22, 2008
    Publication date: November 26, 2009
    Applicant: NOKIA CORPORATION
    Inventors: Huanglingzi Liu, Yue Zhong Tang, Yu Zhang
  • Publication number: 20090292531
    Abstract: Streaming voice signals, such as might be received at a contact center or similar operation, are analyzed to detect the occurrence of one or more unprompted, predetermined utterances. The predetermined utterances preferably constitute a vocabulary of words and/or phrases having particular meaning within the context in which they are uttered. Detection of one or more of the predetermined utterances during a call causes a determination of response-determinative significance of the detected utterance(s). Based on the response-determinative significance of the detected utterance(s), a responsive action may be further determined. Additionally, long term storage of the call corresponding to the detected utterance may also be initiated. Conversely, calls in which no predetermined utterances are detected may be deleted from short term storage.
    Type: Application
    Filed: May 22, 2009
    Publication date: November 26, 2009
    Applicant: ACCENTURE GLOBAL SERVICES GMBH
    Inventors: Thomas J. Ryan, Biji K. Janan
  • Publication number: 20090281807
    Abstract: A voice quality conversion device converts voice quality of an input speech using information of the speech.
    Type: Application
    Filed: May 8, 2008
    Publication date: November 12, 2009
    Inventors: Yoshifumi Hirose, Takahiro Kamai, Yumiko Kato
  • Publication number: 20090276219
    Abstract: In the present invention, a voice input system and a voice input method are provided. The voice input method includes the steps of: (A) initiating a speech recognition process by a first input associated with a first parameter of a first speech recognition subject; (B) providing a voice and a searching space constructed by a speech recognition model associated with the first speech recognition subject; (C) obtaining a sub-searching space from the searching space based on the first parameter; (D) searching at least one candidate item associated with the voice from the sub-searching space; and (E) showing the at least one candidate item.
    Type: Application
    Filed: April 29, 2009
    Publication date: November 5, 2009
    Applicant: DELTA ELECTRONICS, INC.
    Inventors: Keng-Hung Yeh, Liang-Sheng Huang, Chao-Jen Huang, Jia-Lin Shen
  • Publication number: 20090276223
    Abstract: An administration method and system. The method includes receiving by a computing system, a telephone call from an administrator. The computing system presents an audible menu associated with a plurality of computers to the administrator. The computing system receives from the administrator, an audible selection for a computer from the audible menu. The computing system receives from the administrator, an audible verbal command for performing a maintenance operation on the computer. The computing system executes the maintenance operation on the computer. The computing system receives from the computer, confirmation data indicating that the maintenance operation has been completed. The computing system converts the confirmation data into an audible verbal message. The computing system transmits the second audible verbal message to the administrator.
    Type: Application
    Filed: May 1, 2008
    Publication date: November 5, 2009
    Inventors: Peeyush Jaiswal, Naveen Narayan
  • Publication number: 20090276220
    Abstract: A system evaluates a hands free communication system. The system automatically selects a consonant-vowel-consonant (CVC), vowel-consonant-vowel (VCV), or other combination of sounds from an intelligent database. The selection is transmitted with another communication stream that temporally overlaps the selection. The quality of the communication system is evaluated through an automatic speech recognition engine. The evaluation occurs at a location remote from the transmitted selection.
    Type: Application
    Filed: April 29, 2009
    Publication date: November 5, 2009
    Inventors: Shreyas Paranjpe, Mark Fallat
  • Publication number: 20090254344
    Abstract: A spoken language understanding method and system are provided. The method includes classifying a set of labeled candidate utterances based on a previously trained classifier, generating classification types for each candidate utterance, receiving confidence scores for the classification types from the trained classifier, sorting the classified utterances based on an analysis of the confidence score of each candidate utterance compared to a respective label of the candidate utterance, and rechecking candidate utterances according to the analysis. The system includes modules configured to control a processor in the system to perform the steps of the method.
    Type: Application
    Filed: June 16, 2009
    Publication date: October 8, 2009
    Applicant: AT&T Corp.
    Inventors: Dilek Z. Hakkani-Tur, Mazin G. Rahim, Gokhan Tur
  • Publication number: 20090254542
    Abstract: Search methods and systems are provided. First, a voice is received. The voice is recognized using voice recognition technology to obtain at least two keywords, and a specific keyword combination is generated according to the keywords. A specific requirement among a plurality of requirements is associated according to the specific keyword combination, wherein each requirement corresponds to at least one of a plurality of keyword combinations. Then, according to the specific requirement, a database is searched to obtain at least candidate data conforming to the specific requirement.
    Type: Application
    Filed: January 12, 2009
    Publication date: October 8, 2009
    Applicant: MITAC INTERNATIONAL CORP.
    Inventor: Fang-Wei Kang
  • Publication number: 20090240500
    Abstract: A speech recognition apparatus includes a storage unit which store vocabularies, each of vocabularies including plural word body data, each of the word body data obtained by removing a specific word head from a word or sentence, and store at least one word head portion including labeled nodes to express at least one common word head common to at least two of the vocabularies, an instruction receiving unit which receive an instruction of a target vocabulary and an instruction of a operation, a grammar network generating unit which generate, when adding is instructed, a grammar network containing the word head portion, the target vocabulary and connection information indicating that each of the word body data contained in the target vocabulary is connected to a specific one of the labeled nodes contained in the word head portion, and a speech recognition unit which execute speech recognition using the generated grammar network.
    Type: Application
    Filed: March 19, 2009
    Publication date: September 24, 2009
    Inventors: Mitsuyoshi TACHIMORI, Shinichi TANAKA
  • Publication number: 20090234650
    Abstract: A spoken dialogue apparatus which retrieves desired data from a data storage unit storing data in association with attribute values concerning a plurality of attributes of the data, includes a calculation unit configured to calculate estimated numbers of data as costs concerning a first response and a second response based on the retrieval condition and the retrieval result, the data being included in retrieval result after narrowed retrieval performed in accordance with an input speech corresponding to the first response asking the user to input an attribute value of an attribute which has not been input and corresponding to the second response ascertaining the truth of an attribute value of an attribute which has been input, and a selection unit configured to select a response having the lowest cost and present the response to the user.
    Type: Application
    Filed: March 11, 2009
    Publication date: September 17, 2009
    Inventor: Takehide YANO
  • Publication number: 20090234651
    Abstract: A speech recognition system includes a mobile device and a remote server. The mobile device receives the speech from the user and extracts the features and phonemes from the speech. Selected phonemes and measures of uncertainty are transmitted to the server, which processes the phonemes for speech understanding and transmits a text of the speech (or the context or understanding of the speech) back to the mobile device.
    Type: Application
    Filed: March 12, 2009
    Publication date: September 17, 2009
    Inventors: Otman A. Basir, William Ben Miners
  • Publication number: 20090234847
    Abstract: Provided is an information retrieval method including: retrieving, by a computer, a name including input characters from a database for storing the name, an attribute word associated with the name, and a degree of relevance between the name and the attribute word; outputting the retrieved name as a candidate name; and extracting an attribute word associated with the candidate name, the extracting including: calculating a degree of independency indicating a degree of difference between the extracted attribute words, a degree of coverage indicating an extent to which the combination of the extracted attribute words covers the candidate names, and a degree of equality of a number of corresponding candidate names for each attribute word; and calculating a score of the combination of the attribute words based on at least one of the independency, the coverage and the equality to output the combinations of the attribute words to an output unit.
    Type: Application
    Filed: February 10, 2009
    Publication date: September 17, 2009
    Inventors: Takeshi HOMMA, Hiroaki Kokubo, Hisashi Takahashi
  • Publication number: 20090222267
    Abstract: The invention relates to a task classification system (900) that interacts with a user. The task classification system (900) may include a recognizer (920) that may recognize symbols in the user's input communication, and a natural language understanding unit (900) that may determine whether the user's input communication can be understood. If the user's input communication can be understood, the natural language understanding unit (930) may generate understanding data. The system may also include a communicative goal generator that may generate communicative goals based on the symbols recognized by the recognizer (920) and understanding data from the natural language understanding unit (930). The generated communicative goals may be related to information needed to be obtained from the user.
    Type: Application
    Filed: February 26, 2009
    Publication date: September 3, 2009
    Applicant: AT&T Corp.
    Inventors: Marilyn A. Walker, Owen Christopher Rambow, Monica Rogati
  • Publication number: 20090216525
    Abstract: A system and method for homonym treatment in a speech recognition system and method are provided. The system and method for homonym treatment in a speech recognition system may be used in a mobile wireless communication devices that are voice operated after their initial activation.
    Type: Application
    Filed: February 20, 2009
    Publication date: August 27, 2009
    Applicant: Vocera Communications, Inc.
    Inventor: Robert E. Shostak
  • Publication number: 20090203427
    Abstract: A slot machine 1 of the present invention makes a control so as to: sequentially store the number of game values consumed per unit game; sequentially store the number of game values given per unit game; calculating a difference between the total number of game values given and the total number of game values consumed, as a self game value difference; transmitting the self game value difference to outside; receiving someone's game value difference from outside; when the self game value difference and the someone's game value difference are in a predetermined relationship, voice-outputting, by the conversation controller 91, an answer at volume corresponding to the predetermined relationship from the speaker 23 in response to a voice input through the microphone 90; and delete the stored numbers of game values given and consumed, under a predetermined condition.
    Type: Application
    Filed: January 21, 2009
    Publication date: August 13, 2009
    Applicant: ARUZE GAMING AMERICA, INC.
    Inventor: Kazuo OKADA
  • Publication number: 20090199235
    Abstract: Embodiments of a computer-implemented advertisement display system are disclosed. In one embodiment, the system includes a speech recognition component that processes a video clip and produces a corresponding collection of speech recognition data indicative of an audio portion of the video clip. The system also includes a collection of advertising material. An advertisement selection component selects an advertisement from the collection of advertising material based on the corresponding collection of speech recognition data. The system also includes a display. An advertisement presentation component displays an indication of the selected advertisement on the display during a simultaneous display of the video clip.
    Type: Application
    Filed: June 6, 2008
    Publication date: August 6, 2009
    Applicant: MICROSOFT CORPORATION
    Inventors: Arungunram C. Surendran, Patrick Nguyen, Milind V. Mahajan