Miscellaneous Analysis Or Detection Of Speech Characteristics (epo) Patents (Class 704/E11.001)
  • Publication number: 20130054247
    Abstract: Methods and arrangements for facilitating tangible interactions in voice applications. At least two tangible objects are provided, along with a measurement interface. The at least two tangible objects are disposed to each be displaceable with respect to one another and with respect to the measurement interface. The measurement interface is communicatively connected with a voice application. At least one of the two tangible objects is displaced with respect to the measurement interface, and the displacement of at least one of the at least two tangible objects is converted to input for the voice application.
    Type: Application
    Filed: August 31, 2011
    Publication date: February 28, 2013
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Nitendra Rajput, Shrey Sahay, Saurabh Srivastava, Kundan Shrivastava
  • Publication number: 20130054227
    Abstract: A phonetic symbol system formed by phonetic symbols using letters of English Alphabet is described. The cases or the styles of the letters do not affect the sounds of the phonetic symbols. The phonetic symbols are systematically and logically defined. The phonetic symbol system can be used where language is involved. The phonetic symbol system provides convenient ways to represent languages. In some embodiments, the phonetic symbol system provides ways to represent English language.
    Type: Application
    Filed: August 28, 2011
    Publication date: February 28, 2013
    Inventors: George Mizhen Wang, Lili Ni Wang
  • Publication number: 20130046543
    Abstract: Interactive voice response (IVR) systems and methods for delivery of healthcare services (e.g., by one or more medical professionals, such as, for example, in a hospital or clinic). In some embodiments, the present systems can be configured to: prompt one or more users for a plurality of voice inputs with information associated with at least one of a patient and a user; and determine whether each of the plurality of voice inputs is consistent with records related to the patient or the one or more users. In some embodiments, the present systems can be configured to: during performance of a procedure on a patient, prompt one or more users to provide a plurality of voice inputs with information related to progress of the procedure or characteristics of the patient; and/or prompt the user to perform each of a plurality of steps of the procedure.
    Type: Application
    Filed: July 20, 2012
    Publication date: February 21, 2013
    Inventors: Judy Kitchens, Frank Mazza
  • Publication number: 20130046544
    Abstract: A system and method for entering text from a user includes a programmed processor that receives inputs from the user and disambiguates the inputs to present word choices corresponding to the text. In one embodiment, inputs are received in two or more modalities and are analyzed to present the word choices. In another embodiment, a keyboard is divided into zones each of which represents two more input characters. A sequence of zones selected by the user is analyzed to present word choices corresponding to the zone selected.
    Type: Application
    Filed: March 14, 2011
    Publication date: February 21, 2013
    Applicant: Nuance Communications, Inc.
    Inventors: David Kay, Pim Van Meurs, Erland Unruh, Santosh Sharan, Gordon Kessler, Mayank Agarwal, Greg Sigel
  • Publication number: 20130044231
    Abstract: An exemplary electronic device is connected with an earphone. The earphone includes a first storage unit storing information as to functions of the earphone. The electronic device includes a second storage unit storing a function information table recording information as to user-controllable functions of the electronic device, function units corresponding to the user-controllable functions, an identifying module, and a control module. The identifying module retrieves the information as to functions stored in the first storage unit, and determines whether one or more of the controllable functions of the electronic device are controllable by the earphone. The control module activates all of the function units corresponding to the controllable functions of the electronic device which are controllable by the earphone, and controls one or more of the activated function units according to one or more control signals transmitted from the earphone to the electronic device. A related method is also provided.
    Type: Application
    Filed: October 28, 2011
    Publication date: February 21, 2013
    Applicants: HON HAI PRECISION INDUSTRY CO., LTD., FU TAI HUA INDUSTRY (SHENZHEN) CO., LTD.
    Inventors: TSUNG-JEN CHUANG, SHIH-FANG WONG, WEN-WU WANG, WEN-DONG LUO, BIN LIU
  • Publication number: 20130041653
    Abstract: In one embodiment, a system includes one or more computing systems that implement a social networking environment and is operable to parse users' actions that include free form text to determine and store objects and affinities contained in the text string through natural-language processing. The method comprises accessing a text string, identifying objects and affinity declarations via natural-language processing, assessing the combination of objects and context data to determine an instance of a broader concept, and determining an affinity coefficient through a natural-language processing dictionary. Once a database of stored instances and affinities has been generated and stored, it may be leveraged to push suggestions to members of the social network to enhance their social networking experience.
    Type: Application
    Filed: August 12, 2011
    Publication date: February 14, 2013
    Inventor: Erick Tseng
  • Publication number: 20130041670
    Abstract: In an interactive computer controlled display system with speech command input recognition and visual feedback including means for predetermining a plurality of speech commands for respectively initiating each of a corresponding plurality of system actions in combination with means for providing for each of the plurality of speech commands an associated set of speech terms, each term having relevance to its associated command Also included are means responsive to a detected speech term having relevance to one of the speech commands for displaying a relevant command. The system preferably may display basic speech commands simultaneously along with relevant commands. The means for providing the associated set of speech terms may comprise a stored relevance table of universal speech input commands and universal computer operation terms conventionally associated with system actions initiated by the input commands, and means for relating operation terms of the system with terms in the relevance table.
    Type: Application
    Filed: August 14, 2012
    Publication date: February 14, 2013
    Applicant: Nuance Communications, Inc.
    Inventors: Scott Anthony Morgan, David John Roberts, Craig Ardner Swearingen, Alan Richard Tannenbaum, Anthony Christopher Courtney Temple
  • Publication number: 20130039498
    Abstract: An annoyance judgment system includes: a biological signal measurement section configured to measure an electroencephalogram signal of a user; a database retaining a plurality of monosyllabic speech sounds such that, for each speech sound, the speech sound and a reference latency of an electroencephalogram negative component corresponding to the speech sound are retained in association; a presented-speech sound determination section configured to deteremine a monosyllabic speech sound to be presented by referring to the database; and an annoyance judgment section configured to judge annoyance of the output speech sound by comparing a peak latency of a negative component of the electroencephalogram signal in a range from 50 ms to 350 ms from a starting point, the a starting point being a point in time at which the determined speech sound is presented, against the reference latency corresponding to the determined speech sound that is retained in the database.
    Type: Application
    Filed: September 28, 2012
    Publication date: February 14, 2013
    Applicant: PANASONIC CORPORATION
    Inventor: Panasonic Corporation
  • Publication number: 20130041661
    Abstract: A device may include a communication interface configured to receive audio signals associated with audible communications from a user; an output device; and logic. The logic may be configured to determine one or more audio qualities associated with the audio signals, map the one or more audio qualities to at least one value, generate audio-related information based on the mapping, and provide, via the output device during the audible communications, the audio-related information to the user.
    Type: Application
    Filed: August 8, 2011
    Publication date: February 14, 2013
    Applicants: CELLCO PARTNERSHIP, VERIZON NEW JERSEY INC.
    Inventors: Woo Beum Lee, Arvind Basra
  • Publication number: 20130035933
    Abstract: Likelihood calculation means extracts audio features expressing features of a voice signal and a non-voice signal from an acquired audio signal, and calculates likelihood expressing probability that the voice signal is included in the audio signal using the audio features. Spectral feature extraction means performs a frequency analysis to the audio signal to extract a spectral feature. Using the spectral feature, first basis matrix producing means produces a first basis matrix expressing the feature of the non-voice signal. Second basis matrix producing means specifies a component having a high association with the voice signal in the first basis matrix using the likelihood, and excludes the component to produce a second basis matrix. Spectral feature estimation means estimates a spectral feature of the voice signal or a spectral feature of the non-voice signal by performing nonnegative matrix factorization to the spectral feature using the second basis matrix.
    Type: Application
    Filed: March 15, 2012
    Publication date: February 7, 2013
    Inventor: Makoto HIROHATA
  • Publication number: 20130030793
    Abstract: Potential linguistic errors within a sequence of words of a sentence are identified based on analysis of a configurable sliding window. The analysis is performed based on an assumption that if a sequence of words occurs frequently enough within a large, well-formed corpus, its joint probability for occurring in a sentence is very likely to be greater than the same words randomly ordered.
    Type: Application
    Filed: July 28, 2011
    Publication date: January 31, 2013
    Applicant: MICROSOFT CORPORATION
    Inventors: Yizheng Cai, Kevin Roland Powell, Ravi Chandru Shahani, Lei Wang
  • Publication number: 20130030792
    Abstract: A method, an apparatus and an article of manufacture for customizing a natural language processing engine. The method includes enabling selection of one or more parameters of a desired natural language processing task, the one or more parameters intended for use by a trained and an untrained user, mapping the one or more selected parameters to a collection of one or more intervals of an input parameter to an optimization algorithm, and applying the optimization algorithm with the collection of one or more intervals of an input parameter to a model used by a natural language processing engine to produce a customized model.
    Type: Application
    Filed: July 26, 2011
    Publication date: January 31, 2013
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Bing Zhao, Vittorio Castelli
  • Publication number: 20130030812
    Abstract: Provided is an emotion information generating apparatus that is capable of recognizing a user's emotional state for each function of a terminal. The emotion information generating apparatus detects a user's emotional state and maps the user's emotional state to a function of the terminal, thus creating emotion information.
    Type: Application
    Filed: January 6, 2012
    Publication date: January 31, 2013
    Inventor: Hyun-Jun Kim
  • Publication number: 20130013299
    Abstract: A system for developing, deploying and maintaining a voice application over a communications network to one or more recipients has a voice application server connected to a data network for storing and serving voice applications, a network communications server connected to the data network and to the communications network for routing the voice applications to their intended recipients, a computer station connected to the data network having control access to at least the voice application server, and a software application running on the computer station for creating applications and managing their states. The system is characterized in that a developer operating the software application from the computer station creates voice applications through object modeling and linking, stores them for deployment in the application server, and manages deployment and state of deployed applications including scheduled deployment and repeat deployments in terms of intended recipients.
    Type: Application
    Filed: September 14, 2012
    Publication date: January 10, 2013
    Applicant: Apptera, Inc.
    Inventors: Michael S. Yuen, Leo Chiu
  • Publication number: 20130013300
    Abstract: A band broadening apparatus includes a processor configured to analyze a fundamental frequency based on an input signal bandlimited to a first band, generate a signal that includes a second band different from the first band based on the input signal, control a frequency response of the second band based on the fundamental frequency, reflect the frequency response of the second band on the signal that includes the second band and generate a frequency-response-adjusted signal that includes the second band, and synthesize the input signal and the frequency-response-adjusted signal.
    Type: Application
    Filed: September 14, 2012
    Publication date: January 10, 2013
    Applicant: FUJITSU LIMITED
    Inventors: Takeshi OTANI, Taro TOGAWA, Masanao SUZUKI, Shusaku ITO
  • Publication number: 20130013317
    Abstract: In one embodiment, the present disclosure is a method and apparatus for navigation of a dialogue system. In one embodiment, a method for facilitating navigation of a menu of a dialogue system includes encoding data including information for navigating the menu in a machine-readable data structure and outputting the machine-readable data structure.
    Type: Application
    Filed: September 14, 2012
    Publication date: January 10, 2013
    Applicant: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventor: BRENT HARVEY
  • Publication number: 20130006611
    Abstract: One embodiment provides a system for extracting shadow entities from emails. During operation, the system receives a number of document corpora. The system then calculates word-collocation statistics associated with different n-gram sizes for the document corpora. Next, the system receives an email and identifies shadow entities in the email based on the calculated word-collocation statistics for the document corpora.
    Type: Application
    Filed: June 30, 2011
    Publication date: January 3, 2013
    Applicant: PALO ALTO RESEARCH CENTER INCORPORATED
    Inventors: Oliver Brdiczka, Petro Hizalev
  • Publication number: 20130006641
    Abstract: A method, system and computer program product for generating answers to questions. In one embodiment, the method comprises receiving an input query, decomposing the input query into a plurality of different subqueries, and conducting a search in one or more data sources to identify at least one candidate answer to each of the subqueries. A ranking function is applied to each of the candidate answers to determine a ranking for each of these candidate answers; and for each of the subqueries, one of the candidate answers to the subquery is selected based on this ranking. A logical synthesis component is applied to synthesize a candidate answer for the input query from the selected the candidate answers to the subqueries. In one embodiment, the procedure applied by the logical synthesis component to synthesize the candidate answer for the input query is determined from the input query.
    Type: Application
    Filed: September 14, 2012
    Publication date: January 3, 2013
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Eric W. Brown, Jennifer Chu-Carroll, David A. Ferrucci, Adam P. Lally, James W. Murdock, IV, John M. Prager
  • Publication number: 20130006609
    Abstract: The embodiments provide for automatic incremental learning of programming language grammar. A corpus (i.e., a text file of software code written in a particular programming language) is parsed based on a set of grammar rules. An unparsed statement from the corpus is identified along with a section thereof, which did not match any of the grammar rules in the set. A subset of the set of grammar rules at fault for the parsing failure is identified. Groups of new grammar rules are developed such that each group comprises at least one new grammar rule, such that each group can parse the unparsed statement, and such that each new grammar rule is a modification of grammar rule(s) in the subset. One specific group can then be selected for possible incorporation into the set of grammar rules. Optionally, before a specific group is selected, the groups can be heuristically pruned and/or ranked.
    Type: Application
    Filed: June 28, 2011
    Publication date: January 3, 2013
    Applicant: International Business Machines Corporation
    Inventors: Pankaj Dhoolia, Mangala G. Nanda, Krishna Nandivada Venkata, Diptikalyan Saha
  • Publication number: 20130006625
    Abstract: A system, method, and computer program product for automatically analyzing multimedia data audio content are disclosed. Embodiments receive multimedia data, detect portions having specified audio features, and output a corresponding subset of the multimedia data and generated metadata. Audio content features including voices, non-voice sounds, and closed captioning, from downloaded or streaming movies or video clips are identified as a human probably would do, but in essentially real time. Particular speakers and the most meaningful content sounds and words and corresponding time-stamps are recognized via database comparison, and may be presented in order of match probability. Embodiments responsively pre-fetch related data, recognize locations, and provide related advertisements. The content features may be also sent to search engines so that further related content may be identified. User feedback and verification may improve the embodiments over time.
    Type: Application
    Filed: June 28, 2011
    Publication date: January 3, 2013
    Applicant: Sony Corporation
    Inventors: Priyan Gunatilake, Djung Nguyen, Abhishek Patil, Dipendu Saha
  • Publication number: 20130007043
    Abstract: Methods and systems for time-synchronous voice annotation of video and audio media enable effective searching of time-based media content. A user record one or more types voice annotation onto corresponding named voice annotation tracks, which are stored within a media object comprising the time-based media and the annotations. The one or more annotation tracks can then be selectively searched for content using speech or text search terms. Various workflows enable voice annotation to be performed using media editing systems, or one or more stand alone voice annotations systems that permit multiple annotators to operate in parallel, generating different kinds of annotations, and returning their annotation tracks to a central location for consolidation.
    Type: Application
    Filed: June 30, 2011
    Publication date: January 3, 2013
    Inventors: Michael E. Phillips, Paul J. Gray
  • Publication number: 20130006616
    Abstract: According to an embodiment, an information retrieving apparatus includes a housing; an input-output unit to perform dialogue processing with a user; a first detecting unit to detect means of transfer which indicates present means of transfer for the user; a second detecting unit to detect a holding status which indicates whether the user is holding the housing; a third detecting unit to detect a talking posture which indicates whether the housing is held near the face of the user; a selecting unit to select, from among a plurality of interaction modes that establish the dialogue processing, an interaction mode according to a combination of the means of transfer, the holding status, and the talking posture; an dialogue manager to control the dialogue processing according to the selected interaction mode; and a information retrieval unit to retrieve information using a keyword that is input during the dialogue processing.
    Type: Application
    Filed: July 2, 2012
    Publication date: January 3, 2013
    Applicant: KABUSHIKI KAISHA TOSHIBA
    Inventors: Hiromi Wakaki, Kazuo Sumita, Hiroko Fujii, Masaru Suzuki, Michiaki Ariga
  • Publication number: 20130006622
    Abstract: A continuous comfort noise is provided that is overlaid for the entire duration of a conference call scenario. The comfort noise may be adapted to match the levels of the actual background noise detected on one or more of the conference call participant's devices on the transmitting end(s) of a conference call as well as the participants' speech levels. The comfort noise may also be adapted to the type of listening device employed on the receiving end of a conference call. The comfort noise level may be customized to an appropriate and comfortable level for the type of listening device being used, and the system may continuously mix the comfort noise with incoming audio signals for the entire duration of a conference call, lowering the comfort noise level gradually during speaking periods for additional user experience improvement.
    Type: Application
    Filed: June 28, 2011
    Publication date: January 3, 2013
    Applicant: MICROSOFT CORPORATION
    Inventors: Hosam Khalil, Xiaoqin Sun, Hong Wang Sodoma, Warren Lam
  • Publication number: 20120330666
    Abstract: A method and system for vocalizing user-selected sporting event scores. A customized spoken score application module can be configured in association with a device. A real-time score can be preselected by a user from an existing sporting event website for automatically vocalizing the score in a multitude of languages utilizing a speech synthesizer and a translation engine. An existing text-to-speech engine can be integrated with the spoken score application module and controlled by the application module to automatically vocalize the preselected scores listed on the sporting event site. The synthetically-voiced, real-time score can be transmitted to the device at a predetermined time interval. Such an approach automatically and instantly pushes the real time vocal alerts thereby permitting the user to continue multitasking without activating the pre-selected vocal alerts.
    Type: Application
    Filed: June 6, 2012
    Publication date: December 27, 2012
    Inventors: Anthony Verna, Luis M. Ortiz
  • Publication number: 20120330642
    Abstract: Translating a phrase from one language into another using an order-based set of declarative rules is disclosed. Information to be communicated as sensory perceptible output is received. An ordered set of rules is applied to generate a representation that expresses the information in a manner that embodies applicable communication system rules of a target symbolic communication system in which the information is to be communicated.
    Type: Application
    Filed: June 21, 2011
    Publication date: December 27, 2012
    Applicant: APPLE INC.
    Inventors: Aaron S. Kalb, Ryan P. Perry, Thomas Matthieu Alsina
  • Publication number: 20120330647
    Abstract: The described implementations relate to natural language processing, and more particularly to training a language prior model using a model structure. The language prior model can be trained using parameterized representations of lexical structures such as training sentences, as well as parameterized representations of lexical units such as words or n-grams. During training, the parameterized representations of the lexical structures and the lexical units can be adjusted using the model structure. When the language prior model is trained, the parameterized representations of the lexical structures can reflect how the lexical units were used in the lexical structures.
    Type: Application
    Filed: June 24, 2011
    Publication date: December 27, 2012
    Applicant: MICROSOFT CORPORATION
    Inventors: Chris J.C. Burges, Andrzej Pastusiak
  • Publication number: 20120330646
    Abstract: A computer system and method where text is recognized from a real world image, and this recognized text is used as input data for a processing program selected by a user. A computer system and method where text is recognized from a real world image, and contextual information is used in conjunction with the text to develop a semantic denotation of the recognized text. The contextual information may include GPS location data. The contextual information may include previous images, captured shortly prior to the image with the recognized text. A computer system and method wherein text is recognized from a real world image, then normalized to be in the plane of the image, then translated and then the translated text is made into an image that is anti-normalized and inserted into the original image (or an image similar to the original image). In this way, the translated text will appear realistically in place of the original untranslated text of the real world image.
    Type: Application
    Filed: June 23, 2011
    Publication date: December 27, 2012
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Rajiv Augusto Santos Galvao de Andrade, Lucas Goncalves Franco, Christopher A. Robbins
  • Publication number: 20120330650
    Abstract: Methods, systems, and computer readable media for fricatives and high frequencies detection are disclosed. According to one method, the method includes receiving a narrowband signal. The method also includes detecting, using one or more autocorrelation coefficients, a high frequency speech component associated with the narrowband signal.
    Type: Application
    Filed: June 21, 2011
    Publication date: December 27, 2012
    Inventors: Emmanuel Rossignol Thepie Fapi, Eric Poulin
  • Publication number: 20120323558
    Abstract: A method for creating a predictive model is disclosed herein, including the steps of determining trends and patterns in electronic data, using at least a first machine language algorithm, refining the determination of the algorithm, searching for social models that describe the identified trends and patterns using at least a second machine language algorithm, verifying causal links, constructing at least one model about human node behavior and interactions, utilizing the social models to do at least one of the following: validate hypotheses, predict future behavior, and examine hypothetical scenarios, and automatically updating predictions when new data is introduced.
    Type: Application
    Filed: September 21, 2011
    Publication date: December 20, 2012
    Applicant: DECISIVE ANALYTICS CORPORATION
    Inventors: JAMES J. NOLAN, MARK E. FRYMIRE, JONATHAN C. DAY, Peter F. David
  • Publication number: 20120323554
    Abstract: A method for tuning translation parameters in statistical machine translation based on ranking of the translation parameters is disclosed. According to one embodiment, the method includes sampling pairs of candidate translation units from a set of candidate translation units corresponding to a source unit, each candidate translation unit corresponding to numeric values assigned to one or more features, receiving an initial weighting value for each feature, comparing the pairs of candidate translation units to produce binary results, and using the binary results to adjust the initial weighting values to produce modified weighting values.
    Type: Application
    Filed: June 15, 2011
    Publication date: December 20, 2012
    Inventors: Mark Hopkins, Jonathan May
  • Publication number: 20120323566
    Abstract: Methods, systems, and media for determining a response to be generated in an environment are provided. The methods, systems, and media monitor the environment for a voice activity of an individual. The voice activity of the individual is detected and analyzed. A content descriptor of the voice activity is determined based on the voice activity of the individual. A demographic descriptor of the individual is determined based on the voice activity of the individual. The content descriptor, the demographic descriptor, and known information are correlated to determine the response to be generated in the environment.
    Type: Application
    Filed: August 13, 2012
    Publication date: December 20, 2012
    Applicant: AT&T Intellectual Property I, L.P.
    Inventors: Michael Johnston, Hisao M. Chang, Harry E. Blanchard, Bernard S. Renger, Linda Roberts
  • Publication number: 20120323579
    Abstract: An audio conference is supplemented based on speaker content. Speaker content from at least one audio conference participant is monitored using a computer with a tangible non-transitory processor and memory. A set of words is selected from the speaker content. The selected set of words is determined to be associated with supplemental media content from at least one external source. The supplemental media content is made available to at least one audience member for the audio conference. The supplemental media content is selectively presented to the at least one audience member.
    Type: Application
    Filed: June 17, 2011
    Publication date: December 20, 2012
    Applicant: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: David C. GIBBON, Lee BEGEJA, Zhu LIU, Bernard S. RENGER, Behzad SHAHRARAY, Eric ZAVESKY
  • Publication number: 20120323556
    Abstract: A system and method for using pinyin and a dynamic memory state for modifying a Hanyu vocabulary test are provided. The system uses a Chinese input method for conducting the Hanyu vocabulary test, and includes an input device, an output device, a computing processing device, and a word database. The computing processing device includes a user login verification unit, a screening and sampling unit, a grading and timing unit, and a recording and analyzing unit. The user uses the input device and output device to operate the system. After the user is verified by the user login verification unit, the screening and sampling unit samples a test bank from the word database based on the test record of the user. The grading and timing unit calculates the answering time and determines if the user has answered the questions correctly. The recording and analyzing unit records and analyzes a test result of the user, and then stores the test result to the word database.
    Type: Application
    Filed: June 15, 2011
    Publication date: December 20, 2012
    Inventors: Ti-Chih CHEN, Meng-Tien Wu
  • Publication number: 20120316883
    Abstract: According to one embodiment of the present invention a computerized voice-analysis device for determining an S, H, G profile is provided (as described herein, such an S, H, G profile relates to the strengths (e.g., relative strengths) of three human instinctive drives). Of note, the present invention may be used for one or more of the following: analyzing a previously recorded voice sample; real-time analysis of voice as it is being spoken; combination voice analysis—that is, a combination of: (a) previously recorded and/or real-time voice; and (b) answers to a questionnaire.
    Type: Application
    Filed: August 20, 2012
    Publication date: December 13, 2012
    Applicant: Exaudios Technologies
    Inventors: Yoram Levanon, Lan Lossos-Shifin
  • Publication number: 20120316882
    Abstract: An adaptive workflow system can be used to implement captioning projects, such as projects for creating captions or subtitles for live and non-live broadcasts. Workers can repeat words spoken during a broadcast program or other program into a voice recognition system, which outputs text that may be used as captions or subtitles. The process of workers repeating these words to create such text can be referred to as respeaking. Respeaking can be used as an effective alternative to more expensive and hard-to-find stenographers for generating captions and subtitles.
    Type: Application
    Filed: June 10, 2011
    Publication date: December 13, 2012
    Inventor: Morgan Fiumi
  • Publication number: 20120316860
    Abstract: A caption translation system is described herein that provides a way to reach a greater world-wide audience when displaying video content by providing dynamically translated captions based on the language the user has selected for their browser. The system provides machine-translated captions to accompany the video content by determining the language the user has selected for their browser or a manual language selection of the user. The system uses the language value to invoke an automated translation application-programming interface that returns translated caption text in the selected language. The system can use one or more well-known caption formats to store the translated captions, so that video playing applications that know how to consume captions can automatically display the translated captions. The video playing application plays back the video file and displays captions in the user's language.
    Type: Application
    Filed: June 8, 2011
    Publication date: December 13, 2012
    Applicant: MICROSOFT CORPORATION
    Inventor: Erik Reitan
  • Publication number: 20120310629
    Abstract: Systems and methods are provided to automatically determine culture-based behavioral tendencies and preferences of individuals in the context of customer service interactions. For example, systems and methods are provided to process natural language dialog input of an individual to detect linguistic features indicative of individualistic and collectivistic behavioral tendencies and predict whether such individual will be cooperative or uncooperative with automated customer service.
    Type: Application
    Filed: August 10, 2012
    Publication date: December 6, 2012
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Osamuyimen T. Stewart, David M. Lubensky, Joyram Chakraborty
  • Publication number: 20120306836
    Abstract: Various apparatus, systems and methods for an electro-luminance display on a garment are described. An apparatus comprises an electro-luminance portion being configured to display a graphic image. The electro-luminance portion is further configured to removably join to an outside surface of the garment. A housing is configured to removably join to a back side of the electro-luminance portion. A processor is contained within the housing and coupled to the electro-luminance portion. The processor is configured to control functions of the electro-luminance portion. A transceiver system comprises a receiver contained within the housing. The receiver is configured to receive GPS location information and communicate the GPS location information to the processor where the processor is operable to control the electro-luminance portion based in part on the GPS location information.
    Type: Application
    Filed: June 1, 2011
    Publication date: December 6, 2012
    Inventor: Farrell Ackerman
  • Publication number: 20120303374
    Abstract: A mobile terminal includes an input unit receiving an input; a data storage unit storing data; a communication unit communicating signals; and a controller. The controller is configured to receive a selection input of a video data, the selection input being processed to select the video data among a plurality of video data stored in the data storage unit; temporarily store a selected portion of the video data for transmission based on a start position and a stop position specifying the selected portion in the video data; automatically attach the selected portion of the video data for transmission to a message without receiving any further user input when the selected portion of the video data is specified; transmit the message with the selected portion of the video data; and delete the selected portion of the video data from the data storage unit when the transmission of the message is completed.
    Type: Application
    Filed: August 2, 2012
    Publication date: November 29, 2012
    Applicant: LG ELECTRONICS INC.
    Inventor: Sang-Cheol SHIN
  • Publication number: 20120299826
    Abstract: A human/machine (HM) interface that enables a human operator to control a corresponding machine using the geometric degrees of freedom of the operator's vocal tract, for example, using the tongue as a virtual joystick. In one embodiment, the HM interface has an acoustic sensor configured to monitor, in real time, the geometry of the operator's vocal tract using acoustic reflectometry. A signal processor analyzes the reflected acoustic signals detected by the acoustic sensor, e.g., using signal-feature selection and quantification, and translates these signals into commands and/or instructions for the machine. Both continuous changes in the machine's operating parameters and discrete changes in the machine's operating configuration and/or state can advantageously be implemented.
    Type: Application
    Filed: May 24, 2011
    Publication date: November 29, 2012
    Applicant: ALCATEL-LUCENT USA INC.
    Inventor: Lothar Benedikt Moeller
  • Publication number: 20120303372
    Abstract: Techniques for enabling a secure transaction with a remote site that uses voice interaction are provided. The techniques include authenticating a remote site to enable a secure transaction, wherein authenticating the remote site comprises using a dynamically generated audio signal.
    Type: Application
    Filed: August 8, 2012
    Publication date: November 29, 2012
    Applicant: International Business Machines Corporation
    Inventor: Anupam Joshi
  • Publication number: 20120296631
    Abstract: Methods, systems, devices, and computer-storage media having computer-usable instructions embodied thereon, for displaying key pinyins are provided. Pinyins are words and/or symbols that represent a pronunciation of a character. The character may be a Chinese character. A single page of a user interface of a device may display dynamic updates of key pinyins that are associated with a desired pinyin and dynamic updates of one or more characters that are associated with a desired pinyin. Additional embodiments may provide the associated pinyins and/or characters in a single page that is scrollable at varying speeds depending on a received gesture.
    Type: Application
    Filed: September 15, 2011
    Publication date: November 22, 2012
    Applicant: MICROSOFT CORPORATION
    Inventors: FANG ZHANG, MINGQIN LI
  • Publication number: 20120296652
    Abstract: A method for obtaining information on an audio video program being presented on a consumer electronics (CE) device includes receiving at the CE device a viewer command to recognize the audio video program being presented on the CE device. The method also includes receiving signals from a microphone representative of audio from the audio video program as sensed by the microphone as the audio is played real time on the CE device. The method then includes executing voice recognition on the signals from the microphone to determine words in the audio from the audio video program as sensed by the microphone. Words are then uploaded to an Internet server, where they are correlated to at least one audio video script. The method then includes receiving back from the Internet server information correlated by the server using the words to the audio video program.
    Type: Application
    Filed: May 18, 2011
    Publication date: November 22, 2012
    Inventors: Seth Hill, Frederick J. Zustak
  • Publication number: 20120296646
    Abstract: Concepts and technologies are described herein for multi-mode text input. In accordance with the concepts and technologies disclosed herein, content is received. The content can include one or more input indicators. The input indicators can indicate that user input can be used in conjunction with consumption or use of the content. The application is configured to analyze the content to determine context associated with the content and/or the client device executing the application. The application also is configured to determine, based upon the content and/or the contextual information, which input device to use to obtain input associated with use or consumption of the content. Input captured with the input device can be converted to text and used during use or consumption of the content.
    Type: Application
    Filed: May 17, 2011
    Publication date: November 22, 2012
    Applicant: Microsoft Corporation
    Inventors: Mohan Varthakavi, Jayaram Nanduri, Nikhil Kothari
  • Publication number: 20120290292
    Abstract: A system to process unstructured data is provided. An example system to process unstructured data comprises a receiver to access a source of unstructured data, an entity type module to determine an entity type, a rules generator to automatically generate a linguistic rule based on the determined entity type, and an entity extractor to obtain an entity from the source of unstructured data, using the linguistic rule. The entity comprises an alpha-numeric string.
    Type: Application
    Filed: July 18, 2011
    Publication date: November 15, 2012
    Applicant: Business Objects Software Ltd.
    Inventors: Jing Liu, Wuzhen Xiong, Edison Tedjing
  • Publication number: 20120290287
    Abstract: A method and mobile device are provided for inputting text. In one implementation, a method is provided for associating a first language with a first text input apparatus of the mobile device. The method further associates a second language, different from the first language, with a second text input apparatus of the mobile device. The method also enables text input from the first text input apparatus in the first language and from the second text input apparatus in the second language.
    Type: Application
    Filed: May 13, 2011
    Publication date: November 15, 2012
    Inventors: Vadim Fux, Gabriel Lee Gilbert Shelley, Michael Elizaroc, Timothy Koo, Parui Nanoa Gilani
  • Publication number: 20120290291
    Abstract: A mobile computing device that operates a method that processes handwritten user input for character matching and predictive word matching. A user inputs handwritten input on a touch-sensitive display using, for example, a stylus. The method determines and displays a set of candidate character matches for the handwritten input. The user then selects a character from the candidate character matches. The method determines and displays a set of candidate predicted word matches based on the user selected character match. The user can then select to input a desired candidate predicted word match.
    Type: Application
    Filed: May 13, 2011
    Publication date: November 15, 2012
    Inventors: Gabriel Lee Gilbert Shelley, Parul Nanda Gilani
  • Publication number: 20120281097
    Abstract: A vehicle media system is disclosed. The system includes a media unit mounted to a headrest of the vehicle. The media unit includes a video camera oriented towards a passenger area of the vehicle and an infrared camera configured to capture infrared images. A display monitor is disposed in the vehicle in a field-of-view of a vehicle operator to display images captured by the video camera.
    Type: Application
    Filed: May 6, 2011
    Publication date: November 8, 2012
    Inventor: David Wood
  • Publication number: 20120284031
    Abstract: A method and device for operating technical equipment, in particular in a motor vehicle. Speech inputs are fed by a speech input unit and manual inputs are fed by means of a manual input unit as operating instructions to a controller by which a command corresponding to the operating instruction is generated and fed to the corresponding technical equipment, which then executes the operating procedure associated with the operating instruction. A basic structure of the command is established by the speech input unit or the manual input unit, and then the basic structure of the command is supplemented by the manual input unit or the speech input unit.
    Type: Application
    Filed: December 9, 2010
    Publication date: November 8, 2012
    Applicant: CONTINENTAL AUTOMOTIVE GMBH
    Inventors: Ronald Hain, Herbert Meier, Nhu Nguyen Thien, Thomas Rosenstock, Alexander Stege
  • Publication number: 20120278083
    Abstract: A voice control device includes a storage module, a voice recording module, and a processing module. The storage module stores a number of computerized voice commands. The voice recording module records audio signals of a user. The processing module processes the recorded voice signals to a machine readable command, determines whether the determined machine readable command matches one stored computerized voice command, and controls the device to execute a function according to the machine readable command if the determined machine readable command matches one stored computerized voice command. The processing module stores the determined machine readable command as a history command. The processing module further obtains all the history commands and determines which function the voice controlled device is to do according to the history commands if the determined machine readable command is partially the same as at least two of the stored computerized voice commands.
    Type: Application
    Filed: August 22, 2011
    Publication date: November 1, 2012
    Applicant: HON HAI PRECISION INDUSTRY CO., LTD.
    Inventors: YING-CHUAN YU, HSING-CHU WU