Audio Input For On-screen Manipulation (e.g., Voice Controlled Gui) Patents (Class 715/728)
  • Patent number: 8144839
    Abstract: The present invention concerns a method for determining a sequence of services linked to a conversation in which a user (12) takes part from a communication system (1), as well as the associated communication system (1). According to the invention, the method includes the following stages: retrieval (E2) of all the expressions used during the conversation; semantic analysis (E31) of the expressions retrieved; selection (E32) of a plurality of services from a list of services previously saved, based on at least said semantic analysis, with each service of said list able to be launched from said communication system (1); contextual analysis (E33) of the expressions used during the conversation; determination of a sequence of said services selected based on the contextual analysis; display (E4) on said communication system (1) of the sequence of services selected.
    Type: Grant
    Filed: September 9, 2009
    Date of Patent: March 27, 2012
    Assignee: Alcatel Lucent
    Inventors: Mathieu Boussard, Alain Pastor
  • Patent number: 8132104
    Abstract: Methods, computer-readable media, and systems for populating an electronic clinical document capable of receiving multiple types of data are provided. In one embodiment, dictation audio input is embedded directly into one or more sections of an electronic clinical document along with other data types, such as structured user input, free-text input, or system-generated input. An electronic clinical document having embedded dictation audio and other data types can receive transcribed text corresponding to the embedded dictation audio.
    Type: Grant
    Filed: February 22, 2007
    Date of Patent: March 6, 2012
    Assignee: Cerner Innovation, Inc.
    Inventors: Michael A. Ash, John Q. Deverter, Pramod Pagadala
  • Patent number: 8102373
    Abstract: A hybrid device having a plurality of task-handlers corresponding to a function key uses a user interface method. The user interface method includes receiving information necessary to change priorities of the task-handlers corresponding to the function key; changing the priorities of the task-handlers based on the received information; and storing information regarding the changed priorities according to a predetermined application type.
    Type: Grant
    Filed: April 22, 2004
    Date of Patent: January 24, 2012
    Assignee: Samsung Electronics Co., Ltd.
    Inventor: Seong-ho Kwon
  • Publication number: 20120011443
    Abstract: A method for speech enabling an application can include the step of specifying a speech input within a speech-enabled markup. The speech-enabled markup can also specify an application operation that is to be executed responsive to the detection of the speech input. After the speech input has been defined within the speech-enabled markup, the application can be instantiated. The specified speech input can then he detected and the application operation can be responsively executed in accordance with the specified speech-enabled markup.
    Type: Application
    Filed: September 20, 2011
    Publication date: January 12, 2012
    Applicant: Nuance Communications, Inc.
    Inventors: Charles W. Cross, Leslie R. Wilson, Steven G. Woodward
  • Patent number: 8095527
    Abstract: The present invention is intended to automatically construct a database of contents data which are distributed over plural reproducing apparatuses and search this database on the basis of user's fragmentary memory. A contents sharing management system practiced as one embodiment of the invention comprises an episode server installed at user's home and plural reproducing apparatuses including a component stereo set, portable player, portable wireless terminal, and MD player, which are interconnected in a wireless manner based on wireless communication technologies such as Bluetooth. The episode server wirelessly connects to the portable player for example to get the episode information stored therein and organizes the retrieved episode information into a database. The episode server also searches the database upon request from the portable player to identify a source apparatus in which desired contents data are stored and supplies the retrieved contents data to the requesting portable player.
    Type: Grant
    Filed: October 7, 2008
    Date of Patent: January 10, 2012
    Assignee: Sony Corporation
    Inventors: Noriyuki Yamamoto, Kazunori Ohmura
  • Patent number: 8091028
    Abstract: To facilitate the use of audio files for annotation purposes, an audio file format, which includes audio data for playback purposes, is augmented with a parallel data channel of line identifiers, or with a map associating time codes for the audio data with line numbers on the original document. The line number-time code information in the audio file is used to navigate within the audio file, and also to associate bookmark links and captured audio annotation files with line numbers of the original text document. An annotation device may provide an output document wherein links to audio and/or text annotation files are embedded at corresponding line numbers. Also, a navigation index may be generated, having links to annotation files and associated document line numbers, as well as bookmark links to selected document line numbers.
    Type: Grant
    Filed: February 27, 2009
    Date of Patent: January 3, 2012
    Assignee: Copernicus Investments, LLC
    Inventors: Steven Spielberg, Samuel Gustman
  • Publication number: 20110320950
    Abstract: Systems and associated methods configured to provide user-driven audio content navigation for the spoken web are described. Embodiments allow users to skim audio for content that seems to be of relevance to the user, similar to visual skimming of standard web pages, and mark point of interest within the audio. Embodiments provide techniques for navigating audio content while interacting with information systems in a client-server environment, where the client device can be a simple, standard telephone.
    Type: Application
    Filed: June 24, 2010
    Publication date: December 29, 2011
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Nitendra Rajput, Om D. Deshmukh
  • Publication number: 20110313768
    Abstract: A multimedia entertainment system combines both gestures and voice commands to provide an enhanced control scheme. A user's body position or motion may be recognized as a gesture, and may be used to provide context to recognize user generated sounds, such as speech input. Likewise, speech input may be recognized as a voice command, and may be used to provide context to recognize a body position or motion as a gesture. Weights may be assigned to the inputs to facilitate processing. When a gesture is recognized, a limited set of voice commands associated with the recognized gesture are loaded for use. Further, additional sets of voice commands may be structured in a hierarchical manner such that speaking a voice command from one set of voice commands leads to the system loading a next set of voice commands.
    Type: Application
    Filed: June 18, 2010
    Publication date: December 22, 2011
    Inventors: Christian Klein, Ali M. Vassigh, Jason S. Flaks, Vanessa Larco, Thomas M. Soemo
  • Patent number: 8077840
    Abstract: One embodiment of a representative system for web integrated interactive voice response includes an interactive voice response system adapted to provide a plurality of voice menus to a user over a telephone and a graphical user interface system adapted to provide a plurality of menus in a graphical format to the user over a network connection. Information provided in the voice menus corresponds to information provided in the menus in the graphical format and is responsive to commands received by the graphical user interface system from the user. Other systems and methods are also provided.
    Type: Grant
    Filed: December 19, 2007
    Date of Patent: December 13, 2011
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Xiaofeng Gao, David Scott, Sam Zellner
  • Publication number: 20110296505
    Abstract: A system and method is disclosed for sensing, storing and using personal trait profile data. Once sensed and stored, this personal trait profile data may be used for a variety of purposes. In one example, a user's personal trait profile data may be accessed and downloaded to different computing systems with which a user may interact so that the different systems may be instantly tuned to the user's personal traits and manner of interaction. In a further example, a user's personal trait profile data may also be used for authentication purposes.
    Type: Application
    Filed: May 28, 2010
    Publication date: December 1, 2011
    Applicant: MICROSOFT CORPORATION
    Inventors: Kathryn Stone Perez, Alex Aben-Athar Kipman, John Clavin, Joseph Molnar, Aaron E. Kornblum
  • Patent number: 8059167
    Abstract: A shooting apparatus includes a shooting section that shoots a subject, an acquiring section that acquires sound of surroundings, and an assigning section that assigns an electronic mark indicating a characteristic feature of the sound to a position in the image with sound corresponding to the sound, in each of a case when a level of the sound has become equal to a first threshold or more for a first period of time or more, and a case when, thereafter, the level of the sound has become less than a second threshold for a second period of time or more.
    Type: Grant
    Filed: August 25, 2008
    Date of Patent: November 15, 2011
    Assignee: Sony Corporation
    Inventors: Mitsutoshi Shinkai, Yoshiaki Shibata
  • Publication number: 20110271194
    Abstract: This specification describes technologies relating to content presentation. In general, one aspect of the subject matter described in this specification can be embodied in methods that include the actions of presenting a content item to a user; receiving a user input indicating a voice interaction; receiving a voice input from the user; transmitting the voice input to a content system; receiving a command responsive to the voice input; and executing, using one or more processors, the command including modifying the content item. Other embodiments of this aspect include corresponding systems, apparatus, and computer program products.
    Type: Application
    Filed: April 29, 2010
    Publication date: November 3, 2011
    Applicant: GOOGLE INC.
    Inventors: Jennifer W. Lin, Ping Wu
  • Publication number: 20110239115
    Abstract: Disclosed are techniques that allow the user of a mobile device to select an avatar within a virtual world presented on the display screen of the mobile device. In some embodiments, a user manipulates a thumbwheel. As the thumbwheel is turned, the avatars on the display screen are highlighted one after another. The user then presses a thumbwheel button to select a desired avatar. Some embodiments allow the user to select more than one avatar at a time. Several highlighting techniques are available. In some embodiments, the user uses speech commands instead of a thumbwheel to highlight the avatars one by one. Speech input is also used to select one or more avatars. Some devices support a touch-screen interface. Embodiments for these devices allow the user to select an avatar by, for example, drawing an arc enclosing the avatar.
    Type: Application
    Filed: March 26, 2010
    Publication date: September 29, 2011
    Applicant: MOTOROLA, INC.
    Inventors: Jay J. Williams, Renxiang Li, Jingjing Meng
  • Patent number: 8024665
    Abstract: A mobile electronic communications device and a method for modifying a scheduling database are provided. The scheduling database is modified via the mobile electronic communications device in communication with a proximate electronic device comprising a display device, the mobile electronic communications device is engaging an oral communications session with a second communications device. Time slot data associated with an available time slot is retrieved from the scheduling database, via the mobile electronic communications device. The time slot data is transmitted to the proximate electronic device causing it to display the time slot data at the display device. Input data representative of a command to update the scheduling database is received at the mobile electronic communications device. An entry of the scheduling database is updated, via the mobile electronic device, based on the time slot data associated with the available time slot displayed at the display device.
    Type: Grant
    Filed: March 11, 2010
    Date of Patent: September 20, 2011
    Assignee: Research In Motion Limited
    Inventors: Jerome Pasquero, Steven Fyke, Michael Joseph DeFazio
  • Patent number: 8015009
    Abstract: A computer system comprising hardware and software elements; the hardware elements including a processor, a display means and a speaker, the software elements comprising a speech synthesizer, a database platform and a software application comprising a methodology of inputting and tabulating visual elements and verbal elements into the database, links for linking the visual elements and verbal elements; operations for manipulating the database and for enunciating the verbal elements as the corresponding visual elements are displayed on the display means.
    Type: Grant
    Filed: May 3, 2006
    Date of Patent: September 6, 2011
    Inventors: Joel Jay Harband, Uziel Yosef Harband
  • Patent number: 8014542
    Abstract: A system, set-top box, and method of providing audio content are disclosed. In a particular embodiment, the set-top box device includes an audio input device that is configured to receive a voice command. The set-top box device also includes a display disposed on a visible side of a set-top box housing. Additionally, the set-top box device includes a media control logic module to retrieve audio content identified by the voice command and data related to the audio content. The media control logic module transmits an audio signal corresponding to the audio content to an audio output device. The data related to the audio content is displayable on the display.
    Type: Grant
    Filed: November 4, 2005
    Date of Patent: September 6, 2011
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Zesen Chen, Peter Chou, Steve Aspell
  • Patent number: 8010885
    Abstract: Differential dynamic content delivery with a presenter alterable session copy of a user profile. Typical embodiments include providing a session document for a presentation, wherein the session document includes a session grammar and a session structured document; providing a session copy of a user profile including a user classification; receiving, from a presenter, a user classification instruction to change a user classification in the session copy of a user profile; changing the user classification in the session copy of a user profile in dependence upon the presenter's instruction; selecting from the session structured document a classified structural element in dependence upon a user classification in the session copy of a user profile of a user in the presentation; and presenting the selected structural element to the user.
    Type: Grant
    Filed: September 29, 2008
    Date of Patent: August 30, 2011
    Assignee: International Business Machines Corporation
    Inventors: William Kress Bodin, Michael John Burkhart, Daniel G. Eisenhauer, Daniel Mark Schumacher, Thomas J. Watson
  • Patent number: 8006186
    Abstract: An editing system is proposed for automatically, or semi-automatically, editing input data to generate output production. The input material is annotated by, or analyzed to derive, a set of media descriptors which describe the input material and which are derived from the input material. The style of editing is controlled using style data which is optionally derived from a user. The input material may include any or more of motion video, still images, music, speech, sound effects, animated graphics and text. The style data and the descriptors are used to generate a set of operations which, when carried out on the input data, generate an edited output production.
    Type: Grant
    Filed: December 22, 2000
    Date of Patent: August 23, 2011
    Assignee: Muvee Technologies Pte. Ltd.
    Inventors: Peter Rowan Kellock, Edward James Altman
  • Publication number: 20110153342
    Abstract: The present invention is a nonprescription medication consumer tool for selecting nonprescription medications which allows an individual to select his or her symptoms and contraindications to determine a suitable nonprescription medication. The symptoms, contraindications and information provided to the individual, such as dosing and warnings, are provided in layperson's terms. The nonprescription medication consumer tool may be available to individuals as a website and/or mobile device application or as part of a vending system or dispensing machine.
    Type: Application
    Filed: December 17, 2009
    Publication date: June 23, 2011
    Inventor: John Rose
  • Publication number: 20110138287
    Abstract: A system is disclosed for navigating the display of content in a windowed computing environment, the system comprising a computing device comprising a voice recognition engine, a first window and a second window, wherein the second window comprises at least one hyperlink linked to additional content. A user command receiver receives a voice command from a user while the user is working in the first window, and in response to the voice command follows the hyperlink in the second window while the user remains in productive control of the first window, wherein following the hyperlink in the second window causes the additional content to be displayed in the second window.
    Type: Application
    Filed: January 25, 2011
    Publication date: June 9, 2011
    Inventor: Ronald Mark Katsuranis
  • Publication number: 20110138286
    Abstract: The invention discloses a method and apparatus for (a) processing a voice input from the user of computer technology, (b) recognizing potential objects of interest, and (c) using electronic displays to present visual artefacts directing user's attention to the spatial locations of the objects of interest. The voice input is matched with attributes of the information objects, which are visually presented to the viewer. If one or several objects match the voice input sufficiently, the system visually marks or highlights the object or objects to help the viewers direct his or her attention to the matching object or objects. The sets of visual objects and their attributes, used in the matching, may be different for different user tasks and types of visually displayed information.
    Type: Application
    Filed: August 7, 2010
    Publication date: June 9, 2011
    Inventors: Viktor Kaptelinin, Elena Oleinik
  • Patent number: 7949529
    Abstract: A mobile system is provided that includes speech-based and non-speech-based interfaces for telematics applications. The mobile system identifies and uses context, prior information, domain knowledge, and user specific profile data to achieve a natural environment for users that submit requests and/or commands in multiple domains. The invention creates, stores and uses extensive personal profile information for each user, thereby improving the reliability of determining the context and presenting the expected results for a particular question or command. The invention may organize domain specific behavior and information into agents, that are distributable or updateable over a wide area network.
    Type: Grant
    Filed: August 29, 2005
    Date of Patent: May 24, 2011
    Assignee: VoiceBox Technologies, Inc.
    Inventors: Chris Weider, Richard Kennewick, Mike Kennewick, Philippe Di Cristo, Robert A. Kennewick, Samuel Menaker, Lynn Elise Armstrong
  • Publication number: 20110119590
    Abstract: A system and method in a personal electronic book system for providing speech-controlled operation thereof. As non-limiting examples, an electronic book reader may comprise one or more modules operable to utilize a default set of speech commands and/or develop a suite of customized speech commands to be utilized for controlling operation of the electronic book reader.
    Type: Application
    Filed: November 17, 2010
    Publication date: May 19, 2011
    Inventor: Nambirajan Seshadri
  • Publication number: 20110115702
    Abstract: A method and system for computer programming using speech and one or two hand gesture input is described. The system generally uses a plurality of microphones and cameras as input devices. A configurable event recognition system is described allowing various software objects in a system to respond to speech and hand gesture and other input. From this input program code is produced that can be compiled at any time. Various speech and hand gesture events invoke functions within programs to modify programs, move text and punctuation in a word processor, manipulate mathematical objects, perform data mining, perform natural language interne search, modify project management tasks and visualizations, perform 3D modeling, web page design and web page data entry, and television and DVR programming.
    Type: Application
    Filed: July 9, 2009
    Publication date: May 19, 2011
    Inventor: David Seaberg
  • Publication number: 20110099476
    Abstract: Disclosed herein are systems and methods for decorating a display environment. In one embodiment, a user may decorate a display environment by making one or more gestures, using voice commands, using a suitable interface device, and/or combinations thereof. A voice command can be detected for user selection of an artistic feature, such as, for example, a color, a texture, an object, and a visual effect for decorating in a display environment. The user can also gesture for selecting a portion of the display environment for decoration. Next, the selected portion of the display environment can be altered based on the selected artistic feature. The user's motions can be reflected in the display environment by an avatar. In addition, a virtual canvas or three-dimensional object can be displayed in the display environment for decoration by the user.
    Type: Application
    Filed: October 23, 2009
    Publication date: April 28, 2011
    Applicant: Microsoft Corporation
    Inventors: Gregory N. Snook, Relja Markovic, Stephen G. Latta, Kevin Geisner, Christopher Vuchetich, Darren Alexander Bennett, Arthur Charles Tomlin, Joel Deaguero, Matt Puls, Matt Coohill, Ryan Hastings, Kate Kolesar, Brian Scott Murphy
  • Publication number: 20110093545
    Abstract: Embodiments are disclosed herein that relate to enabling a vehicle user to obtain, via a voice-activated system, content that the user requests and that is not to be locally stored. For example, one disclosed embodiment comprises, in a vehicle, a voice-activated computing device including a processor, and memory comprising instructions stored thereon that are executable by the processor to receive from a user an input of a voice command requesting an item of content, determine whether the item of content is stored locally, and if the item of content is not stored locally, then send via a network the voice command to a remote service for analysis. The computing device is further configured to receive a response from the remote service comprising information regarding the item of content for presentation to the user, and present the information via an output to the user.
    Type: Application
    Filed: October 21, 2009
    Publication date: April 21, 2011
    Applicant: MICROSOFT CORPORATION
    Inventors: Manuela Papadopol, Gregory P. Baribault
  • Publication number: 20110083075
    Abstract: An emotive advisory system for use by one or more occupants of an automotive vehicle includes a directional speaker array, and a computer. The computer is configured to determine an audio direction, and output data representing an avatar for visual display. The computer is further configured to output data representing a spoken statement for the avatar for audio play from the speaker array such that the audio from the speaker array is directed in the determined audio direction. A visual appearance of the avatar and the spoken statement for the avatar convey a simulated emotional state.
    Type: Application
    Filed: October 2, 2009
    Publication date: April 7, 2011
    Applicant: FORD GLOBAL TECHNOLOGIES, LLC
    Inventors: Perry Robinson MacNeille, Oleg Yurievitch Gusikhin, Kacie Alane Theisen
  • Patent number: 7921364
    Abstract: Controlling a computer user interface with sound including defining ranges of audio frequencies; associating each range with a user interface command; receiving as input from a user a sound representing an instruction to execute a user interface command, the sound characterized by a spectrum of frequencies and energies; identifying the user interface command to be executed in dependence upon the spectrum of frequencies and energies, the defined ranges of audio frequencies, and the user interface commands associated with the ranges; and executing the identified command.
    Type: Grant
    Filed: November 3, 2005
    Date of Patent: April 5, 2011
    Assignee: Nuance Communications, Inc.
    Inventor: Nelson Ramirez
  • Publication number: 20110067099
    Abstract: A method for interpreting messages, user-defined alert conditions, voice commands and performing an action in response is described. A method for annotating media content is described. A method for presenting additional content associated with media content identified based on a fingerprint is described. A method for identifying that an advertisement portion of media content is being played based on a fingerprint derived from the media content is described. A method of one media device recording particular media content automatically in response to another media device recording the particular media content is described. A method of concurrently playing media content on multiple devices is described. A method of publishing information associated with recording of media content is described. A method of deriving fingerprints by media devices that meet an idleness criteria is described. A method of loading, modifying, and displaying a high definition frame from a frame buffer is described.
    Type: Application
    Filed: December 4, 2009
    Publication date: March 17, 2011
    Inventors: James M. Barton, Brian Lanier, Amir H. Gharaat, James Cheng
  • Patent number: 7908565
    Abstract: A system is disclosed for displaying a second window of a second application while a first window of a first application has input focus in a windowed computing environment having a voice recognition engine. The system comprises a retriever for launching the second application, a user command receiver for receiving commands from the voice recognition engine, and an application manager. The application manager responds to a command from the user command receiver by invoking the retriever to launch the second application and display the second window while the first window maintains substantially uninterrupted input focus.
    Type: Grant
    Filed: April 9, 2008
    Date of Patent: March 15, 2011
    Inventor: Ronald M. Katsuranis
  • Publication number: 20110035671
    Abstract: The present invention is intended to share information as to voice operation in an image processing device with a voice operation function with another image processing device, thereby improving operability for using another one. An image processing device allowed to be connected to a network comprising: an operational panel for displaying a menu screen and receiving a manual operation to the menu screen; a speech input part for inputting speech; an operation item specifying part for specifying an operation item to be a target of operation based on a voice word; a voice operation control part for executing a processing corresponding to the specified operation item; a history information generation part for generating a voice operation history information in which the voice word and the specified operation item are associated; and a transmission part for transmitting the generated voice operation history information to another image processing device through the network.
    Type: Application
    Filed: July 23, 2010
    Publication date: February 10, 2011
    Applicant: KONICA MINOLTA BUSINESS TECHNOLOGIES, INC.
    Inventors: Hidetaka IWAI, Kazuo Inui, Nobuhiro Mishima, Kaitaku Ozawa
  • Patent number: 7885816
    Abstract: A method, a system, and an apparatus for efficiently presenting correction options. The present invention is capable of analyzing user voice commands and sorting multiple input requests based on user selection probability to determine whether a confirmation step should be presented and, if so, the manner in which the confirmation step should be presented. In particular, the method requests an information input from the user and then assigns a confidence level to the information input. If the confidence level is LOW, then the system performs an immediate confirmation step. If the confidence level assigned is MEDIUM or HIGH, then the information is placed into a data set that is confirmed in a batch confirmation step. The batch confirmation step presents the captured information to the user for confirmation. If any of the information is incorrect, then the method sorts the information in ascending order by confidence level and creates a menu of items that may be changed. The user then makes the change.
    Type: Grant
    Filed: December 8, 2003
    Date of Patent: February 8, 2011
    Assignee: International Business Machines Corporation
    Inventors: Brent L. Davis, J. Scott Gee, James R. Lewis, Vanessa V. Michelini, Melanie D. Polkosky
  • Patent number: 7882437
    Abstract: Methods, systems, and products are disclosed for creating a voice response grammar in a voice response server including identifying presentation documents for a presentation, each presentation document having a presentation grammar. Typical embodiments include storing each presentation grammar in a voice response grammar on a voice response server. In typical embodiments, identifying presentation documents for a presentation includes creating a data structure representing a presentation and listing at least one presentation document in the data structure representing a presentation. In typical embodiments listing the at least one presentation document includes storing a location of the presentation document in the data structure representing a presentation and storing each presentation grammar includes retrieving a presentation grammar of the presentation document in dependence upon the location of the presentation document.
    Type: Grant
    Filed: June 10, 2008
    Date of Patent: February 1, 2011
    Assignee: Nuance Communications, Inc.
    Inventors: William Kress Bodin, Michael John Burkhart, Daniel G. Eisenhauer, Daniel Mark Schumacher, Thomas J. Watson
  • Patent number: 7877698
    Abstract: A system comprising user interface software configured to provide a plurality of spectral data display and manipulation options on a display displaying spectral data, a user interface processor configured to display a pictorial representation of the spectral data display and manipulation options, and a user interface instruction engine configured to provide a plurality of data display and manipulation instructions to the user interface processor to provide display and manipulation of the displayed spectral data. The plurality of data display and manipulation instructions are received from a control device connected to a computing system and the manipulation of the displayed spectral data is based on a user response to the displayed spectral data manipulation options or a pre-determined selection of the displayed spectral data display manipulation options.
    Type: Grant
    Filed: October 9, 2007
    Date of Patent: January 25, 2011
    Assignee: Meta Geek, LLC
    Inventors: Ryan Woodings, Christian Harrison, Brian Tuttle
  • Patent number: 7877260
    Abstract: The present invention relates to creating a web page and voice browsing of the web page, and more particularly, it improves accessibility for the voice browsing of the web page through a synthetic voice, efficiently with high reliability. A content creation system 20 of the present invention is used for creating a content which may be viewed through the synthetic voice, the system including: a database 22 for storing a structured document; and an information process section 24 for creating a speech node series 18 from the structured document, and calculating a reaching time from starting voice synthesis of the speech node series 18 until each node is outputted as the synthetic voice. The information process section 24 includes a support process section 36 to determine a graphic display corresponding to the reaching time, and to visually display the reaching time to a predetermined node by the voice synthesis on a screen of a display section 26.
    Type: Grant
    Filed: October 20, 2005
    Date of Patent: January 25, 2011
    Assignee: Nuance Communications, Inc.
    Inventors: Hironobu Takagi, Chieko Asakawa
  • Publication number: 20110014952
    Abstract: The user interface for a mobile communication device may be provided based on the current context of a voice session, as recognized by an automated audio recognition engine. In one implementation, the mobile device may transcribe, by an audio recognition engine in the mobile device, audio from a voice session conducted through the mobile device; detect, by the mobile device and based at least on the transcribed audio, changes in context during the voice session that relate to a change in functionality of the user interface of the mobile device; and update, by the mobile device, the user interface in response to the detected change in context.
    Type: Application
    Filed: July 15, 2009
    Publication date: January 20, 2011
    Applicant: Sony Ericsson Mobile Communications AB
    Inventor: Wayne Christopher MINTON
  • Publication number: 20110016397
    Abstract: A method, system, and computer-readable product for positioning a virtual sound capturing device in a graphical user interface (GUI) are disclosed. The method includes displaying a virtual sound capturing device in relation to a virtual sound producing device in a three dimensional interface and in a two dimensional graphical map. Additionally, the method includes adjusting the display of the virtual sound capturing device in relation to the virtual sound producing device in both the three dimensional interface and the two dimensional graphical map in response to commands received from an input device.
    Type: Application
    Filed: July 20, 2009
    Publication date: January 20, 2011
    Applicant: Apple Inc.
    Inventors: Markus Sapp, Kerstin Heitmann, Thorsten Quandt, Manfred Knauff, Marko Junghanns
  • Patent number: 7859689
    Abstract: Methods and apparatus are provided for receiving a PDL File that describes a print job, processing the PDL commands in the file, identifying print settings and printer factors that may affect the entire print job, and reporting such print settings and printer factors. Methods and apparatus are also provided for receiving a PDL File that describes a print job, processing the PDL commands in the file, identifying any objects in the print job that have associated predetermined attributes, and reporting the print objects that have any such predetermined attributes.
    Type: Grant
    Filed: July 31, 2003
    Date of Patent: December 28, 2010
    Assignee: Electronics for Imaging, Inc.
    Inventors: Paul Michel, Jonathan Marsden
  • Publication number: 20100313133
    Abstract: A method is provided for using a wireless controller to interact with a user interface presented on a display. The method includes receiving an audio signal and a position signal from the wireless controller. The audio signal is based on an audio input applied to the wireless controller, while the position signal is based on a position input applied to the wireless controller. The method includes selecting a user interface item displayed on the display, based on the audio signal and the position signal. One or more position signals from the wireless controller may also be received and processed to cause navigation of the user interface to highlight a user interface item for selection.
    Type: Application
    Filed: June 8, 2009
    Publication date: December 9, 2010
    Applicant: MICROSOFT CORPORATION
    Inventors: Adam Green, Robert Matthew Craig, Dennis Tom, Jeffrey Ma, Erik Arthur
  • Patent number: 7831432
    Abstract: Methods, systems, and computer program products are provided for creating an audio menu describing media content of a media player. Embodiments include retrieving metadata describing the media files managed by the media player; converting at least a portion of the metadata to speech; creating one or more media files for the audio menu; and saving the speech in the audio portion of the one or more the media files for the audio menu.
    Type: Grant
    Filed: September 29, 2006
    Date of Patent: November 9, 2010
    Assignee: International Business Machines Corporation
    Inventors: William K. Bodin, David Jaramillo, Jerry W. Redman, Derral C. Thorson
  • Publication number: 20100280829
    Abstract: A system and method are provided for photo management using expression-based voice commands. The method interfaces a photo-image discovery device, having no dedicated display, to a display monitor. Expression-based user voice prompt are received and used to access a photo-image in storage at a storage site. The accessed photo-image is then presented on the display monitor. The photo-image in storage at the storage site can be accessed to perform an operation such as: selecting a storage site, selecting a photo-image, transforming a selected photo-image, converting a file format of a selected photo-image, and selecting a delivery option. In one aspect, a menu of photo-image user prompt options are presented on the display monitor, originating from the photo discovery device, and the expression-based user voice prompts are received in response to the presented menu.
    Type: Application
    Filed: May 5, 2009
    Publication date: November 4, 2010
    Inventors: Paramesh Gopi, Vinay Ravuri, Dimitry Vaysburg, Prodyut Hazarika
  • Patent number: 7826945
    Abstract: An automotive system provides an integrated user interface for control and communication functions in an automobile or other type of vehicle. The user interface supports voice enabled interactions, as well as other modes of interaction, such as manual interactions using controls such as dashboard or steering wheel mounted controls. The system also includes interfaces to devices in the vehicle, such as wireless interfaces to mobile devices that are brought into the vehicle. The system also provides interfaces to information sources such as a remote server, for example, for accessing information.
    Type: Grant
    Filed: July 1, 2005
    Date of Patent: November 2, 2010
    Inventors: You Zhang, Jeffery J. Faneuff, William Hidden, James T. Hotary, Steven C. Lee, Vasu Iyengar
  • Patent number: 7825797
    Abstract: A proximity sensor device and method is provided that facilitates improved usability. Specifically, the proximity sensor device and method provides the ability for a user to easily select the type of adjustment inputted by the proximity sensor device. In one embodiment, the proximity sensor device includes an adjustment region and one or more start tabs adjacent to the adjustment region. The proximity sensor also includes a processor adapted to indicate adjustment of a first type responsive to sensed object motion originating in a first start tab and continuing in the adjustment region. Thus, a user can cause an adjustment of the first type by introducing an object proximate the first start tab, and moving the object from the first start tab and into the adjustment region.
    Type: Grant
    Filed: June 2, 2006
    Date of Patent: November 2, 2010
    Assignee: Synaptics Incorporated
    Inventors: Fidel Zawde, Thuy T. B. Le, John Feland, Ray Alexander Trent, Jr., Mark Huie
  • Publication number: 20100275122
    Abstract: A “Click-Through Controller” uses various mobile electronic devices (e.g., cell phones, media players, digital cameras, etc.) to provide real-time interaction with content (e.g., maps, places, images, documents, etc.) displayed on the device's screen via selection of one or more “overlay menu items” displayed on top of that content. Navigation through displayed contents is provided by recognizing 2D and/or 3D device motions and rotations. This allows users to navigate through the displayed contents by simply moving the mobile device. Overlay menu items activate predefined or user-defined functions to interact with the content that is directly below the selected overlay menu item on the display. In various embodiments, there is a spatial correspondence between the overlay menu items and buttons or keys of the mobile device (e.g., a cell phone dial pad or the like) such that overlay menu items are directly activated by selection of one or more corresponding buttons.
    Type: Application
    Filed: April 27, 2009
    Publication date: October 28, 2010
    Applicant: MICROSOFT CORPORATION
    Inventors: William A. S. Buxton, John SanGiovanni
  • Publication number: 20100251147
    Abstract: A system that incorporates teachings of the present disclosure may include, for example, an intermediary engine having a controller, and a storage medium for storing instructions to be executed by the controller. The instructions, when executed by the controller, can cause the controller to receive a measure of a mood of an individual, invoke an intermediary according to the measure of the mood of the individual, and present the intermediary to an entity requesting to interact with the individual. The measure of the mood can be synthesized from collected information associated with a behavior of the individual, and can have a plurality of dimensions. The mood can also indicate an availability of the individual and a receptiveness of the individual to accept a request to interact with an entity. Other embodiments are disclosed.
    Type: Application
    Filed: March 27, 2009
    Publication date: September 30, 2010
    Applicant: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: JOHN DONOVAN, James Carlton Bedingfield, SR., Barbara Roden
  • Publication number: 20100235744
    Abstract: Systems and/or methods are provided that facilitate surveying media sources in a media entertainment system. An index component can provide an index of a set of available media sources in the media entertainment system. In addition, a selection component can displays the set of available media sources and receive selection requests for a subset of the available media sources. Further, a preview component can display a sample of media generated by the selected subset of media sources prior to active engagement thereof. Moreover, one or more aspects can be implemented in a vehicle entertainment system.
    Type: Application
    Filed: December 12, 2007
    Publication date: September 16, 2010
    Applicant: JOHNSON CONTROLS, INC.
    Inventors: Steven Gene Schultz, Douglas C. Campbell, Eric Sean Deuel, Stephen Todd Sanders
  • Publication number: 20100217458
    Abstract: An interactive information system is indicated for an aircraft to provide passengers with information, which access both an onboard data storage device and an off-board data storage device. A computer uses the onboard and off-board data to compute informative data, which are then visualized to the passenger. The passenger can interactively determine the form and type of data generation and preparation at any time.
    Type: Application
    Filed: June 25, 2008
    Publication date: August 26, 2010
    Applicant: AIRBUS OPERATIONS GMBH
    Inventors: Andy Schweiger, Manfred Sieber, Peter Rodat, Thomas Brüggert
  • Publication number: 20100180202
    Abstract: A mobile telephone (1) includes a multimodal user interface and a rendering unit (10) that can be used to display icons on the display screen (3) of the mobile telephone (1). The rendering unit (10) receives inputs from a number of status factor determiners of the mobile telephone (1), such as an environmental quality assessment unit (7), a data quality of service unit (8), a network signal strength unit (9), an application engine (5), multimodal interface components (11), and an automatic speech recognition unit of a speech engine (22). The rendering unit (10) uses the status information that it receives to select an icon to be displayed to convey information about the current status of the mobile telephone (1) to the user. The icon that is displayed by the rendering unit (10) is in the form of a human face that can show varying expressions and emotions.
    Type: Application
    Filed: July 5, 2006
    Publication date: July 15, 2010
    Applicant: VIDA SOFTWARE S.L.
    Inventor: Rafael Del Valle Lopez
  • Patent number: 7757173
    Abstract: A voice menu system is disclosed. The present invention generally allows for updateable audio menus. Although a device might have some pre-packaged menu components, other menu components can be received from a server. Each menu component, regardless of whether it is original or received from the server, has an associated voiced name. When a user highlights a menu choice, the voiced name can be played. The user then has the option of selecting the menu choice or scrolling to a new menu choice. In this way, a user could navigate the menu without having to actually see a visual display of the menu, which may be especially useful for users unable to see the visual display or users with visual disabilities.
    Type: Grant
    Filed: July 18, 2003
    Date of Patent: July 13, 2010
    Assignee: Apple Inc.
    Inventor: Alexander B. Beaman
  • Publication number: 20100167256
    Abstract: In accordance with a number of embodiments, this document presents an innovative system and method which may be used to input data relating to any number of historical or scientific subjects, store the data in a collaborative format, and output data in any number of static or animated formats [FIG. 2]. In various embodiments, this method may provide a revolutionary means for encoding the entire history of the earth, encoding the entire history of human cultures, and for ensuring that all input data adhere to a universal data format. It provides and specifies a number of innovative and collaborative protocols for input [FIG. 3], storage [FIG. 4], classification [FIGS. 5A-5V], sorting [FIG. 6A], filtering [FIG. 6B], verifying [FIG. 6C], compiling [FIG. 8], updating [FIG. 9], customizing [FIG. 11], and publishing data [FIG. 12]. It may also provide a means for creating a revolutionary format of global historical collaborative animated map [FIGS 10A-10V, FIGS. 11A-11E].
    Type: Application
    Filed: February 12, 2009
    Publication date: July 1, 2010
    Inventor: Douglas Michael Blash