Audio Input For On-screen Manipulation (e.g., Voice Controlled Gui) Patents (Class 715/728)
  • Patent number: 8316302
    Abstract: A method and apparatus is provided for annotating video content with metadata generated using speech recognition technology. The method begins by rendering video content on a display device. A segment of speech is received from a user such that the speech segment annotates a portion of the video content currently being rendered. The speech segment is converted to a text-segment and the text-segment is associated with the rendered portion of the video content. The text segment is stored in a selectively retrievable manner so that it is associated with the rendered portion of the video content.
    Type: Grant
    Filed: May 11, 2007
    Date of Patent: November 20, 2012
    Assignee: General Instrument Corporation
    Inventors: Kevin M. McKoen, Michael A. Grossman
  • Patent number: 8312131
    Abstract: A method for delivering requested media content is provided. A requested media object requested by a user device is received, the requested media object including media of at least a first type. A first new media object is created, the first new media object having at least some of the media of the requested media object of the first type. A second new media object is created, the second new media object having at least some of the media of the requested media object of a second type, if any. At least one user interface mechanism is inserted into the second new media object, the at least one user interface mechanism corresponding to media of the requested media object of the first type included in the first new media object. The second new media object is delivered to the user device, and the first new media object is delivered to a secondary device separate from the user device.
    Type: Grant
    Filed: December 31, 2002
    Date of Patent: November 13, 2012
    Assignee: Motorola Mobility LLC
    Inventor: Venugopal Vasudevan
  • Publication number: 20120278719
    Abstract: A method of providing a list of links on a display apparatus and a display apparatus are provided. The method includes recognizing a voice spoken by a user, searching, among links included in a web page being currently displayed on the display apparatus, for a link including an index which coincides with the voice spoken by the user and generating a list of one or more links, each including the index which coincides with the voice spoken by the user.
    Type: Application
    Filed: April 17, 2012
    Publication date: November 1, 2012
    Applicant: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Byung-jin HWANG, Young-jun RYU, Hye-jeong LEE
  • Publication number: 20120278720
    Abstract: An information processing apparatus includes an imaging unit, an icon display control unit causing a display to display an operation icon, a pickup image display processing unit causing the display to sequentially display an input operation region image constituted by, among pixel regions constituting an image picked up by the imaging unit, a pixel region including at least a portion of a hand of a user, an icon management unit managing event issue definition information, which is a condition for determining that the operation icon has been operated by the user, for each operation icon, an operation determination unit determining whether the user has operated the operation icon based on the input operation region image displayed in the display and the event issue definition information, and a processing execution unit performing predetermined processing corresponding to the operation icon in accordance with a determination result by the operation determination unit.
    Type: Application
    Filed: July 11, 2012
    Publication date: November 1, 2012
    Inventors: Toshiharu YABE, Kenichiro KOBAYASHI
  • Patent number: 8290780
    Abstract: Dynamically extending the speech prompts of a multimodal application including receiving, by the prompt generation engine, a media file having a metadata container; retrieving, by the prompt generation engine from the metadata container, a speech prompt related to content stored in the media file for inclusion in the multimodal application; and modifying, by the prompt generation engine, the multimodal application to include the speech prompt.
    Type: Grant
    Filed: June 24, 2009
    Date of Patent: October 16, 2012
    Assignee: International Business Machines Corporation
    Inventors: Ciprian Agapi, William K. Bodin, Charles W. Cross, Jr.
  • Patent number: 8276076
    Abstract: Visualizing and exploring a music library using metadata, such as genre, sub-genre, artist, and year, is provided. Geometric shapes, such as disks or rectangles, may be divided into sectors representing genre and each sector may be further divided into sub-sectors representing artists associated with each genre. The sector's relative size generally reflects the importance of the corresponding genre within the library. Likewise, the sub-sector's relative size generally reflects the importance of the corresponding artist within the genre which may be determined by the number of media items of the artist. Marks representing each media item may be arranged and displayed within the geometric shape to reflect the mark's corresponding genre, artist, and year. In addition, each mark may reflect an attribute, such as playcount, of the media item and each sector may reflect the mean value of an attribute of all media items within the sector.
    Type: Grant
    Filed: November 16, 2009
    Date of Patent: September 25, 2012
    Assignee: Apple Inc.
    Inventors: Marc Torrens, Patrick Hertzog, Josep-Lluis Arcos
  • Patent number: 8261191
    Abstract: Some embodiments of the invention provide a method that presents a preview of a set of images in a graphical user interface (GUI) of a device. To generate the preview, the method initially selects a subset of images that includes several images in the set but not all the images in the set. After selecting the subset of images, the method concurrently displays the images in the selected subset as the preview of the set of images.
    Type: Grant
    Filed: August 4, 2006
    Date of Patent: September 4, 2012
    Assignee: Apple Inc.
    Inventor: Bas Ording
  • Patent number: 8244541
    Abstract: The present invention relates to creating a web page and voice browsing of the web page, and more particularly, it improves accessibility for the voice browsing of the web page through a synthetic voice, efficiently with high reliability. A content creation system 20 of the present invention is used for creating a content which may be viewed through the synthetic voice, the system including: a database 22 for storing a structured document; and an information process section 24 for creating a speech node series 18 from the structured document, and calculating a reaching time from starting voice synthesis of the speech node series 18 until each node is outputted as the synthetic voice. The information process section 24 includes a support process section 36 to determine a graphic display corresponding to the reaching time, and to visually display the reaching time to a predetermined node by the voice synthesis on a screen of a display section 26.
    Type: Grant
    Filed: July 10, 2008
    Date of Patent: August 14, 2012
    Assignee: Nuance Communications, Inc.
    Inventors: Hironobu Takagi, Chieko Asakawa
  • Patent number: 8234573
    Abstract: A computer program of the type commonly known as a “wizard” is disclosed that initializes user interface software for controlling an audio conferencing device. The wizard allows the desired audio inputs (e.g., microphone, telephones, etc.) and audio outputs (speakers, recording devices, etc.) to be chosen by an audio system administrator. Thereafter, the wizard allows an audio conferencing device (or devices) to be chosen by the administrator, or allows such a device(s) to be optimally chosen dependent upon the chosen inputs and outputs. The wizard then maps the inputs and outputs to the input and output ports on the audio conferencing device. When the administrator finishes the wizard, the wizard computes the mapping parameters and other audio-optimizing parameters for the selected inputs and outputs. These parameters are then loaded into the user interface software automatically.
    Type: Grant
    Filed: May 7, 2009
    Date of Patent: July 31, 2012
    Assignee: Polycom, Inc.
    Inventors: Thomas M. Drewes, James S. Joiner, Michael A. Pocino, Craig H Richardson
  • Patent number: 8230340
    Abstract: A method and apparatus for displaying a content list are provided, which can display various content lists based on information on content being currently played through a user's simple manipulation. The method of displaying a content list includes receiving a user command, and displaying a content list belonging to a category set on the basis of information on content being currently played, in accordance with the received user command.
    Type: Grant
    Filed: June 19, 2008
    Date of Patent: July 24, 2012
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Hyun-Joo Kang, In-Sik Myung, Joo-Kyung Woo, Deok-Won Kim, Ju-Youn Lee
  • Patent number: 8224650
    Abstract: Web server controls are provided for generating client side markups with recognition and/or audible prompting. Three approaches are disclosed for implementation of the controls.
    Type: Grant
    Filed: April 28, 2003
    Date of Patent: July 17, 2012
    Assignee: Microsoft Corporation
    Inventors: Francisco M. Galanes, Hsiao-Wuen Hon, James D. Jacoby, Renaud J. Lecoeuche, Stephen F. Potter, Susan M. Warren
  • Patent number: 8225235
    Abstract: To provide a reproduction apparatus able to easily select a desired content data based on an attribute of the content data by a simple operation from a user and a reproduction method for the same, wherein the reproduction apparatus having: a display displaying the item; a first operation unit instructing a switch of the attribute; a second function unit instructing a selection of a predetermined item on the display; and a processing unit switching a first screen from a screen of a plurality of items so as to display a plurality of items when the first operation key is operated, and switching to a second screen displaying a plurality of item when the second operation key is operated when a plurality of item is displayed on the first screen.
    Type: Grant
    Filed: June 30, 2005
    Date of Patent: July 17, 2012
    Assignee: Sony Corporation
    Inventors: Naoko Takeda, Kissei Matsumoto, Takashi Kumagai, Toshihide Ooba, Hiroshi Iwata, Shingo Yamade
  • Patent number: 8217905
    Abstract: A method and apparatus for a touchscreen based user interface (UI) interaction that controls a volume through circular UI graphics and switches between different application images by dragging an indicator on a touchscreen of a terminal device. The method includes: displaying an application image on a touchscreen; sensing whether the touchscreen is touched on the displayed application image; if the touchscreen is touched, displaying a volume image to control an audio volume at a touch position; and changing a size of the volume image whenever the touchscreen is touched and controlling the audio volume to corresponding to the size of the volume region.
    Type: Grant
    Filed: November 16, 2007
    Date of Patent: July 10, 2012
    Assignee: Samsung Electronics Co., Ltd
    Inventors: Jung-geun Kim, Bo-eun Park, Jong-sung Joo
  • Patent number: 8219913
    Abstract: When it is determined on the basis of the output from a touch sensor of a touch screen 5 that a touch operation is started in the area of one on/off button displayed on a display screen of the touch screen 5, the on/off state of an on/off parameter assigned to the one on/off button is reversed. When it is determined that the touched position has moved to the area of another on/off button with the touch being maintained, an on/off parameter assigned to the another on/off button is set to the same on/off state as that of the on/off parameter assigned to the one on/off button in which the touch operation has been started. Users are allowed to simultaneously set respective on/off states of on/off parameters assigned to on/off buttons by a single touch operation on the touch screen 5.
    Type: Grant
    Filed: August 15, 2006
    Date of Patent: July 10, 2012
    Assignee: Yamaha Corporation
    Inventor: Kotaro Terada
  • Patent number: 8204184
    Abstract: One embodiment of a representative system for web integrated interactive voice response includes an interactive voice response system adapted to provide a plurality of voice menus to a user over a telephone and a graphical user interface system adapted to provide a plurality of menus in a graphical format to the user over a network connection. Information provided in the voice menus corresponds to information provided in the menus in the graphical format and is responsive to commands received by the graphical user interface system from the user. Other systems and methods are also provided.
    Type: Grant
    Filed: December 20, 2007
    Date of Patent: June 19, 2012
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Xiaofeng Gao, David Scott, Sam Zellner
  • Patent number: 8190995
    Abstract: An information processing apparatus and an information selecting method are disclosed which can easily restore an original state after checking the result of executing predetermined processing based on different information. If an operator of an operation unit is pressed in an already processed state where a processing execution unit is executing predetermined processing using already selected information set in a selected information memory, information assigned to the operator is newly set in the selected information memory to execute the predetermined processing with the processing execution unit. After that, if the pressed operator is released, the processing execution unit is restored to the already processed state. With this operation, after checking a new processing state using different information from already selected information by pressing an operator, the already processed state can be restored simply by releasing the pressed operator.
    Type: Grant
    Filed: July 20, 2009
    Date of Patent: May 29, 2012
    Assignee: Alpine Electronics, Inc.
    Inventors: Akihiko Takizawa, Wataru Oikubo, Syu Takahara, Hironaga Ogawa, Hidetoshi Kadoya
  • Publication number: 20120110456
    Abstract: A system and method are disclosed for providing a NUI system including a speech reveal mode where visual objects on a display having an associated voice command are highlighted. This allows a user to quickly and easily identify available voice commands, and also enhances an ability of a user to learn voice commands as there is a direct association between an object and its availability as a voice command.
    Type: Application
    Filed: November 1, 2010
    Publication date: May 3, 2012
    Applicant: MICROSOFT CORPORATION
    Inventors: Vanessa Larco, Alan T. Shen, Michael Han-Young Kim
  • Publication number: 20120110457
    Abstract: Receiving commands from a remote controller and automatically activating display areas for cursor navigation. Content display areas within a display frame respectively correspond to a variety of content items and include a primary display area wherein cursor navigation is activated and secondary display areas wherein cursor navigation is prevented. Remote controller navigational commands, for example, then allow cursor based navigation for the content item currently displayed in the primary display area. A content selection command such as a number key input of the remote controller allows immediate and automatic updating of the primary display area to include a desired content item that is associated to the command (e.g., the particular number).
    Type: Application
    Filed: January 10, 2012
    Publication date: May 3, 2012
    Applicants: Sony Electronics Inc., Sony Corporation
    Inventors: Steven Friedlander, Hyehoon Yi
  • Patent number: 8169630
    Abstract: Methods and apparatus are provided for receiving a PDL File that describes a print job, processing the PDL commands in the file, identifying print settings and printer factors that may affect the entire print job, and reporting such print settings and printer factors. Methods and apparatus are also provided for receiving a PDL File that describes a print job, processing the PDL commands in the file, identifying any objects in the print job that have associated predetermined attributes, and reporting the print objects that have any such predetermined attributes.
    Type: Grant
    Filed: December 16, 2010
    Date of Patent: May 1, 2012
    Assignee: Electronics for Imaging, Inc.
    Inventors: Paul Michel, Jonathan Marsden
  • Patent number: 8171412
    Abstract: A visual presentation system and method for synchronizing presentation data being viewed in a display with speech input. A system is disclosed that includes: a speech recognition system for recognizing speech input; an association system for determining a context of the speech input and matching the context with a relevant portion of the presentation data; and a visual coordination system for coordinating the display of a data item from the presentation data based on a match made by the association system.
    Type: Grant
    Filed: June 1, 2006
    Date of Patent: May 1, 2012
    Assignee: International Business Machines Corporation
    Inventors: Anne R. Sand, Steven M. Miller
  • Patent number: 8165886
    Abstract: A speech processing system which exploits statistical modeling and formal logic to receive and process speech input, which may represent data to be received, such as dictation, or commands to be processed by an operating system, application or process. A command dictionary and dynamic grammars are used in processing speech input to identify, disambiguate and extract commands. The logical processing scheme ensures that putative commands are complete and unambiguous before processing. Context sensitivity may be employed to differentiate data and commands. A multi faceted graphic user interface may be provided for interaction with a user to speech enable interaction with applications and processes that do not necessarily have native support for speech input.
    Type: Grant
    Filed: September 29, 2008
    Date of Patent: April 24, 2012
    Assignee: Great Northern Research LLC
    Inventors: Jean Gagnon, Philippe Roy, Paul J. Lagassey
  • Publication number: 20120096358
    Abstract: Systems and methods are provided for navigating an information hierarchy using a mobile communication device. The method comprises causing a plurality of selectable items to be presented on a display associated with the mobile communication device, in response to receiving, via an audio input device associated with the mobile communication device, a first voice command indicating that one of the plurality of selectable items is to be selected, causing one of the selectable items in the plurality of selectable items to be displayed differently from the other selectable items to thereby form an accentuated selectable item, and, in response to receiving, via the audio input device, a second voice command indicating that the accentuated selectable item is to be selected, causing information associated with the accentuated selectable item to be presented on the display.
    Type: Application
    Filed: December 20, 2011
    Publication date: April 19, 2012
    Applicant: Wounder GmbH., LLC
    Inventor: Melvin L. Barnes, JR.
  • Patent number: 8160866
    Abstract: The present invention can recognize both English and Chinese at the same time. The most important skill is that the features of all English words (without samples) are entirely extracted from the features of Chinese syllables. The invention normalizes the signal waveforms of variable lengths for English words (Chinese syllables) such that the same words (syllables) can have the same features at the same time position. Hence the Bayesian classifier can recognize both the fast and slow utterance of sentences. The invention can improve the feature such that the speech recognition of the unknown English (Chinese) is guaranteed to be correct. Furthermore, since the invention can create the features of English words from the features of Chinese syllables, it can also create the features of other languages from the features of Chinese syllables and hence it can also recognize other languages, such as German, French, Japanese, Korean, Russian, etc.
    Type: Grant
    Filed: October 10, 2008
    Date of Patent: April 17, 2012
    Inventors: Tze Fen Li, Tai-Jan Lee Li, Shih-Tzung Li, Shih-Hon Li, Li-Chuan Liao
  • Publication number: 20120089915
    Abstract: A method and device for temporally sequenced recommendations of activities delivers to users temporally sequenced objects comprising user activities, wherein the delivered objects are selected based, at least in part, on inferences of preferences from usage behaviors. The delivered objects may include activities associated with processor-based devices in addition to human activities. Variations of the system and method include delivering the temporally sequenced objects in accordance with the contents of the objects and user feedback with regard to the objects. Information as to why objects were delivered to users may be provided to the users.
    Type: Application
    Filed: December 8, 2011
    Publication date: April 12, 2012
    Applicant: MANYWORLDS, INC.
    Inventors: Steven Dennis Flinn, Naomi Felina Moneypenny
  • Publication number: 20120089914
    Abstract: User interfaces for navigating structured content. In one example embodiment, a user interface includes a grid, a header row of cells each positioned in a separate column of the grid, a header column of cells each positioned in a separate row of the grid, a plurality of multi-dimensional cells each having a unique position in the grid, and a viewport that displays only a portion of the grid. Upon of reception an indication that the portion of the grid displayed within the viewport should simultaneously scroll both horizontally and vertically, the multi-dimensional cells of the grid are configured to scroll simultaneously within the viewport both horizontally and vertically, and the header row cells and header column cells of the grid are configured to scroll in a synchronous manner so as to remain visible in the viewport and remain aligned with the rows and columns of multi-dimensional cells.
    Type: Application
    Filed: October 12, 2011
    Publication date: April 12, 2012
    Applicant: SURFWAX INC.
    Inventors: Thomas D. Holt, Larry S. Burke
  • Patent number: 8144839
    Abstract: The present invention concerns a method for determining a sequence of services linked to a conversation in which a user (12) takes part from a communication system (1), as well as the associated communication system (1). According to the invention, the method includes the following stages: retrieval (E2) of all the expressions used during the conversation; semantic analysis (E31) of the expressions retrieved; selection (E32) of a plurality of services from a list of services previously saved, based on at least said semantic analysis, with each service of said list able to be launched from said communication system (1); contextual analysis (E33) of the expressions used during the conversation; determination of a sequence of said services selected based on the contextual analysis; display (E4) on said communication system (1) of the sequence of services selected.
    Type: Grant
    Filed: September 9, 2009
    Date of Patent: March 27, 2012
    Assignee: Alcatel Lucent
    Inventors: Mathieu Boussard, Alain Pastor
  • Patent number: 8132104
    Abstract: Methods, computer-readable media, and systems for populating an electronic clinical document capable of receiving multiple types of data are provided. In one embodiment, dictation audio input is embedded directly into one or more sections of an electronic clinical document along with other data types, such as structured user input, free-text input, or system-generated input. An electronic clinical document having embedded dictation audio and other data types can receive transcribed text corresponding to the embedded dictation audio.
    Type: Grant
    Filed: February 22, 2007
    Date of Patent: March 6, 2012
    Assignee: Cerner Innovation, Inc.
    Inventors: Michael A. Ash, John Q. Deverter, Pramod Pagadala
  • Patent number: 8102373
    Abstract: A hybrid device having a plurality of task-handlers corresponding to a function key uses a user interface method. The user interface method includes receiving information necessary to change priorities of the task-handlers corresponding to the function key; changing the priorities of the task-handlers based on the received information; and storing information regarding the changed priorities according to a predetermined application type.
    Type: Grant
    Filed: April 22, 2004
    Date of Patent: January 24, 2012
    Assignee: Samsung Electronics Co., Ltd.
    Inventor: Seong-ho Kwon
  • Publication number: 20120011443
    Abstract: A method for speech enabling an application can include the step of specifying a speech input within a speech-enabled markup. The speech-enabled markup can also specify an application operation that is to be executed responsive to the detection of the speech input. After the speech input has been defined within the speech-enabled markup, the application can be instantiated. The specified speech input can then he detected and the application operation can be responsively executed in accordance with the specified speech-enabled markup.
    Type: Application
    Filed: September 20, 2011
    Publication date: January 12, 2012
    Applicant: Nuance Communications, Inc.
    Inventors: Charles W. Cross, Leslie R. Wilson, Steven G. Woodward
  • Patent number: 8095527
    Abstract: The present invention is intended to automatically construct a database of contents data which are distributed over plural reproducing apparatuses and search this database on the basis of user's fragmentary memory. A contents sharing management system practiced as one embodiment of the invention comprises an episode server installed at user's home and plural reproducing apparatuses including a component stereo set, portable player, portable wireless terminal, and MD player, which are interconnected in a wireless manner based on wireless communication technologies such as Bluetooth. The episode server wirelessly connects to the portable player for example to get the episode information stored therein and organizes the retrieved episode information into a database. The episode server also searches the database upon request from the portable player to identify a source apparatus in which desired contents data are stored and supplies the retrieved contents data to the requesting portable player.
    Type: Grant
    Filed: October 7, 2008
    Date of Patent: January 10, 2012
    Assignee: Sony Corporation
    Inventors: Noriyuki Yamamoto, Kazunori Ohmura
  • Patent number: 8091028
    Abstract: To facilitate the use of audio files for annotation purposes, an audio file format, which includes audio data for playback purposes, is augmented with a parallel data channel of line identifiers, or with a map associating time codes for the audio data with line numbers on the original document. The line number-time code information in the audio file is used to navigate within the audio file, and also to associate bookmark links and captured audio annotation files with line numbers of the original text document. An annotation device may provide an output document wherein links to audio and/or text annotation files are embedded at corresponding line numbers. Also, a navigation index may be generated, having links to annotation files and associated document line numbers, as well as bookmark links to selected document line numbers.
    Type: Grant
    Filed: February 27, 2009
    Date of Patent: January 3, 2012
    Assignee: Copernicus Investments, LLC
    Inventors: Steven Spielberg, Samuel Gustman
  • Publication number: 20110320950
    Abstract: Systems and associated methods configured to provide user-driven audio content navigation for the spoken web are described. Embodiments allow users to skim audio for content that seems to be of relevance to the user, similar to visual skimming of standard web pages, and mark point of interest within the audio. Embodiments provide techniques for navigating audio content while interacting with information systems in a client-server environment, where the client device can be a simple, standard telephone.
    Type: Application
    Filed: June 24, 2010
    Publication date: December 29, 2011
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Nitendra Rajput, Om D. Deshmukh
  • Publication number: 20110313768
    Abstract: A multimedia entertainment system combines both gestures and voice commands to provide an enhanced control scheme. A user's body position or motion may be recognized as a gesture, and may be used to provide context to recognize user generated sounds, such as speech input. Likewise, speech input may be recognized as a voice command, and may be used to provide context to recognize a body position or motion as a gesture. Weights may be assigned to the inputs to facilitate processing. When a gesture is recognized, a limited set of voice commands associated with the recognized gesture are loaded for use. Further, additional sets of voice commands may be structured in a hierarchical manner such that speaking a voice command from one set of voice commands leads to the system loading a next set of voice commands.
    Type: Application
    Filed: June 18, 2010
    Publication date: December 22, 2011
    Inventors: Christian Klein, Ali M. Vassigh, Jason S. Flaks, Vanessa Larco, Thomas M. Soemo
  • Patent number: 8077840
    Abstract: One embodiment of a representative system for web integrated interactive voice response includes an interactive voice response system adapted to provide a plurality of voice menus to a user over a telephone and a graphical user interface system adapted to provide a plurality of menus in a graphical format to the user over a network connection. Information provided in the voice menus corresponds to information provided in the menus in the graphical format and is responsive to commands received by the graphical user interface system from the user. Other systems and methods are also provided.
    Type: Grant
    Filed: December 19, 2007
    Date of Patent: December 13, 2011
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Xiaofeng Gao, David Scott, Sam Zellner
  • Publication number: 20110296505
    Abstract: A system and method is disclosed for sensing, storing and using personal trait profile data. Once sensed and stored, this personal trait profile data may be used for a variety of purposes. In one example, a user's personal trait profile data may be accessed and downloaded to different computing systems with which a user may interact so that the different systems may be instantly tuned to the user's personal traits and manner of interaction. In a further example, a user's personal trait profile data may also be used for authentication purposes.
    Type: Application
    Filed: May 28, 2010
    Publication date: December 1, 2011
    Applicant: MICROSOFT CORPORATION
    Inventors: Kathryn Stone Perez, Alex Aben-Athar Kipman, John Clavin, Joseph Molnar, Aaron E. Kornblum
  • Patent number: 8059167
    Abstract: A shooting apparatus includes a shooting section that shoots a subject, an acquiring section that acquires sound of surroundings, and an assigning section that assigns an electronic mark indicating a characteristic feature of the sound to a position in the image with sound corresponding to the sound, in each of a case when a level of the sound has become equal to a first threshold or more for a first period of time or more, and a case when, thereafter, the level of the sound has become less than a second threshold for a second period of time or more.
    Type: Grant
    Filed: August 25, 2008
    Date of Patent: November 15, 2011
    Assignee: Sony Corporation
    Inventors: Mitsutoshi Shinkai, Yoshiaki Shibata
  • Publication number: 20110271194
    Abstract: This specification describes technologies relating to content presentation. In general, one aspect of the subject matter described in this specification can be embodied in methods that include the actions of presenting a content item to a user; receiving a user input indicating a voice interaction; receiving a voice input from the user; transmitting the voice input to a content system; receiving a command responsive to the voice input; and executing, using one or more processors, the command including modifying the content item. Other embodiments of this aspect include corresponding systems, apparatus, and computer program products.
    Type: Application
    Filed: April 29, 2010
    Publication date: November 3, 2011
    Applicant: GOOGLE INC.
    Inventors: Jennifer W. Lin, Ping Wu
  • Publication number: 20110239115
    Abstract: Disclosed are techniques that allow the user of a mobile device to select an avatar within a virtual world presented on the display screen of the mobile device. In some embodiments, a user manipulates a thumbwheel. As the thumbwheel is turned, the avatars on the display screen are highlighted one after another. The user then presses a thumbwheel button to select a desired avatar. Some embodiments allow the user to select more than one avatar at a time. Several highlighting techniques are available. In some embodiments, the user uses speech commands instead of a thumbwheel to highlight the avatars one by one. Speech input is also used to select one or more avatars. Some devices support a touch-screen interface. Embodiments for these devices allow the user to select an avatar by, for example, drawing an arc enclosing the avatar.
    Type: Application
    Filed: March 26, 2010
    Publication date: September 29, 2011
    Applicant: MOTOROLA, INC.
    Inventors: Jay J. Williams, Renxiang Li, Jingjing Meng
  • Patent number: 8024665
    Abstract: A mobile electronic communications device and a method for modifying a scheduling database are provided. The scheduling database is modified via the mobile electronic communications device in communication with a proximate electronic device comprising a display device, the mobile electronic communications device is engaging an oral communications session with a second communications device. Time slot data associated with an available time slot is retrieved from the scheduling database, via the mobile electronic communications device. The time slot data is transmitted to the proximate electronic device causing it to display the time slot data at the display device. Input data representative of a command to update the scheduling database is received at the mobile electronic communications device. An entry of the scheduling database is updated, via the mobile electronic device, based on the time slot data associated with the available time slot displayed at the display device.
    Type: Grant
    Filed: March 11, 2010
    Date of Patent: September 20, 2011
    Assignee: Research In Motion Limited
    Inventors: Jerome Pasquero, Steven Fyke, Michael Joseph DeFazio
  • Patent number: 8015009
    Abstract: A computer system comprising hardware and software elements; the hardware elements including a processor, a display means and a speaker, the software elements comprising a speech synthesizer, a database platform and a software application comprising a methodology of inputting and tabulating visual elements and verbal elements into the database, links for linking the visual elements and verbal elements; operations for manipulating the database and for enunciating the verbal elements as the corresponding visual elements are displayed on the display means.
    Type: Grant
    Filed: May 3, 2006
    Date of Patent: September 6, 2011
    Inventors: Joel Jay Harband, Uziel Yosef Harband
  • Patent number: 8014542
    Abstract: A system, set-top box, and method of providing audio content are disclosed. In a particular embodiment, the set-top box device includes an audio input device that is configured to receive a voice command. The set-top box device also includes a display disposed on a visible side of a set-top box housing. Additionally, the set-top box device includes a media control logic module to retrieve audio content identified by the voice command and data related to the audio content. The media control logic module transmits an audio signal corresponding to the audio content to an audio output device. The data related to the audio content is displayable on the display.
    Type: Grant
    Filed: November 4, 2005
    Date of Patent: September 6, 2011
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Zesen Chen, Peter Chou, Steve Aspell
  • Patent number: 8010885
    Abstract: Differential dynamic content delivery with a presenter alterable session copy of a user profile. Typical embodiments include providing a session document for a presentation, wherein the session document includes a session grammar and a session structured document; providing a session copy of a user profile including a user classification; receiving, from a presenter, a user classification instruction to change a user classification in the session copy of a user profile; changing the user classification in the session copy of a user profile in dependence upon the presenter's instruction; selecting from the session structured document a classified structural element in dependence upon a user classification in the session copy of a user profile of a user in the presentation; and presenting the selected structural element to the user.
    Type: Grant
    Filed: September 29, 2008
    Date of Patent: August 30, 2011
    Assignee: International Business Machines Corporation
    Inventors: William Kress Bodin, Michael John Burkhart, Daniel G. Eisenhauer, Daniel Mark Schumacher, Thomas J. Watson
  • Patent number: 8006186
    Abstract: An editing system is proposed for automatically, or semi-automatically, editing input data to generate output production. The input material is annotated by, or analyzed to derive, a set of media descriptors which describe the input material and which are derived from the input material. The style of editing is controlled using style data which is optionally derived from a user. The input material may include any or more of motion video, still images, music, speech, sound effects, animated graphics and text. The style data and the descriptors are used to generate a set of operations which, when carried out on the input data, generate an edited output production.
    Type: Grant
    Filed: December 22, 2000
    Date of Patent: August 23, 2011
    Assignee: Muvee Technologies Pte. Ltd.
    Inventors: Peter Rowan Kellock, Edward James Altman
  • Publication number: 20110153342
    Abstract: The present invention is a nonprescription medication consumer tool for selecting nonprescription medications which allows an individual to select his or her symptoms and contraindications to determine a suitable nonprescription medication. The symptoms, contraindications and information provided to the individual, such as dosing and warnings, are provided in layperson's terms. The nonprescription medication consumer tool may be available to individuals as a website and/or mobile device application or as part of a vending system or dispensing machine.
    Type: Application
    Filed: December 17, 2009
    Publication date: June 23, 2011
    Inventor: John Rose
  • Publication number: 20110138287
    Abstract: A system is disclosed for navigating the display of content in a windowed computing environment, the system comprising a computing device comprising a voice recognition engine, a first window and a second window, wherein the second window comprises at least one hyperlink linked to additional content. A user command receiver receives a voice command from a user while the user is working in the first window, and in response to the voice command follows the hyperlink in the second window while the user remains in productive control of the first window, wherein following the hyperlink in the second window causes the additional content to be displayed in the second window.
    Type: Application
    Filed: January 25, 2011
    Publication date: June 9, 2011
    Inventor: Ronald Mark Katsuranis
  • Publication number: 20110138286
    Abstract: The invention discloses a method and apparatus for (a) processing a voice input from the user of computer technology, (b) recognizing potential objects of interest, and (c) using electronic displays to present visual artefacts directing user's attention to the spatial locations of the objects of interest. The voice input is matched with attributes of the information objects, which are visually presented to the viewer. If one or several objects match the voice input sufficiently, the system visually marks or highlights the object or objects to help the viewers direct his or her attention to the matching object or objects. The sets of visual objects and their attributes, used in the matching, may be different for different user tasks and types of visually displayed information.
    Type: Application
    Filed: August 7, 2010
    Publication date: June 9, 2011
    Inventors: Viktor Kaptelinin, Elena Oleinik
  • Patent number: 7949529
    Abstract: A mobile system is provided that includes speech-based and non-speech-based interfaces for telematics applications. The mobile system identifies and uses context, prior information, domain knowledge, and user specific profile data to achieve a natural environment for users that submit requests and/or commands in multiple domains. The invention creates, stores and uses extensive personal profile information for each user, thereby improving the reliability of determining the context and presenting the expected results for a particular question or command. The invention may organize domain specific behavior and information into agents, that are distributable or updateable over a wide area network.
    Type: Grant
    Filed: August 29, 2005
    Date of Patent: May 24, 2011
    Assignee: VoiceBox Technologies, Inc.
    Inventors: Chris Weider, Richard Kennewick, Mike Kennewick, Philippe Di Cristo, Robert A. Kennewick, Samuel Menaker, Lynn Elise Armstrong
  • Publication number: 20110115702
    Abstract: A method and system for computer programming using speech and one or two hand gesture input is described. The system generally uses a plurality of microphones and cameras as input devices. A configurable event recognition system is described allowing various software objects in a system to respond to speech and hand gesture and other input. From this input program code is produced that can be compiled at any time. Various speech and hand gesture events invoke functions within programs to modify programs, move text and punctuation in a word processor, manipulate mathematical objects, perform data mining, perform natural language interne search, modify project management tasks and visualizations, perform 3D modeling, web page design and web page data entry, and television and DVR programming.
    Type: Application
    Filed: July 9, 2009
    Publication date: May 19, 2011
    Inventor: David Seaberg
  • Publication number: 20110119590
    Abstract: A system and method in a personal electronic book system for providing speech-controlled operation thereof. As non-limiting examples, an electronic book reader may comprise one or more modules operable to utilize a default set of speech commands and/or develop a suite of customized speech commands to be utilized for controlling operation of the electronic book reader.
    Type: Application
    Filed: November 17, 2010
    Publication date: May 19, 2011
    Inventor: Nambirajan Seshadri
  • Publication number: 20110099476
    Abstract: Disclosed herein are systems and methods for decorating a display environment. In one embodiment, a user may decorate a display environment by making one or more gestures, using voice commands, using a suitable interface device, and/or combinations thereof. A voice command can be detected for user selection of an artistic feature, such as, for example, a color, a texture, an object, and a visual effect for decorating in a display environment. The user can also gesture for selecting a portion of the display environment for decoration. Next, the selected portion of the display environment can be altered based on the selected artistic feature. The user's motions can be reflected in the display environment by an avatar. In addition, a virtual canvas or three-dimensional object can be displayed in the display environment for decoration by the user.
    Type: Application
    Filed: October 23, 2009
    Publication date: April 28, 2011
    Applicant: Microsoft Corporation
    Inventors: Gregory N. Snook, Relja Markovic, Stephen G. Latta, Kevin Geisner, Christopher Vuchetich, Darren Alexander Bennett, Arthur Charles Tomlin, Joel Deaguero, Matt Puls, Matt Coohill, Ryan Hastings, Kate Kolesar, Brian Scott Murphy