Audio Input For On-screen Manipulation (e.g., Voice Controlled Gui) Patents (Class 715/728)
-
Patent number: 8712685Abstract: An information processing apparatus includes a position detecting unit detecting a position, an explanation information storage unit storing explanation position information and explanation information, a generation unit extracting the explanation information corresponding to a first explanation position from the explanation information storage unit and generating the extracted explanation information with a sound when the first explanation position is detected within a first distance from the detected position, and a group determining unit determining whether or not the first explanation position belongs to the same group as a second explanation position when the second explanation position is detected within the first distance from the detected position after the generation unit starts the generating of the explanation information corresponding to the first explanation position with a sound, wherein the generation unit continues to reproduce the explanation information with a sound when both explanation posType: GrantFiled: August 10, 2012Date of Patent: April 29, 2014Assignee: Fuji Xerox Co., Ltd.Inventors: Hiroyuki Hattori, Kimitake Hasuike, Reina Kikuchi
-
Publication number: 20140112458Abstract: During voice communication between multiple telecommunications devices, a shareable application facilitates concurrent sharing of data and processes between the devices. The application may be configured to monitor the voice communication and execute a predetermined function upon detecting a predetermined condition in the voice communication. The application may further facilitate sharing of functionality and user interface displays during the voice communication. In some implementations, a server computing device on a communications network may facilitate functions of shareable applications on one or more telecommunications devices.Type: ApplicationFiled: December 23, 2013Publication date: April 24, 2014Applicant: T-Mobile USA, Inc.Inventors: Winston Wang, Adam Holt, Jean-Luc Bouthemy, Michael Kemery
-
Publication number: 20140108935Abstract: In one embodiment, a method includes accessing a social graph that includes a plurality of nodes and edges, receiving from a first user a voice message comprising one or more commands, receiving location information associated with the first user, identifying edges and nodes in the social graph based on the location information, where each of the identified edges and nodes corresponds to at least one of the commands of the voice message, and generating new nodes or edges in the social graph based on the identified nodes or identified edges.Type: ApplicationFiled: October 16, 2012Publication date: April 17, 2014Inventors: Jenny Yuen, David Harry Garcia
-
Patent number: 8700594Abstract: Multidimensional search capabilities are enabled on a non-PC (personal computer) device being utilized by a user. An original query submitted by the user via the non-PC device is received. A structured data repository is accessed to extract structured data that is available for the original query, where the extracted structured data represents attributes of the original query. The extracted structured data is provided to the user in the form of a hierarchical menu which allows the user to interactively modify the original query, such modification resulting in a revised query.Type: GrantFiled: May 27, 2011Date of Patent: April 15, 2014Assignee: Microsoft CorporationInventors: Johnson Apacible, Mark Encarnacion, Aleksey Sinyagin
-
Publication number: 20140101553Abstract: A computing device may output a graphical user interface for display at a presence-sensitive screen including an edit region and a graphical keyboard. The computing device may receive an indication of a gesture detected at a location of the presence-sensitive screen within the graphical keyboard. In response, the computing device may output for display at the presence-sensitive screen, a modified graphical user interface including a media insertion user interface with a plurality of media insertion options. The computing device may receive an indication of a selection of at least one media insertion option associated with a media item. The computing device may output for display at the presence-sensitive screen, an updated graphical user interface including the media item within the edit region.Type: ApplicationFiled: October 10, 2012Publication date: April 10, 2014Inventor: Jens Nagel
-
Publication number: 20140096004Abstract: A voice control method and system for browser operations are described. The method comprises the steps of: receiving an inputted voice control command; the command field value found in the predetermined web page template is the template entry of the voice control command wherein the predetermined web page template includes a plurality of template entries and each of the template entries contains an element field, a command field, and an operation field; and searching an element in a current web page wherein the element corresponds to the value of the element field in the template entry such that the element executes the operation corresponding to the operation field. The present method performs the voice control according to the web page content, thus further improving the voice experience effect upon the user.Type: ApplicationFiled: December 5, 2013Publication date: April 3, 2014Applicant: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITEDInventors: Xiaobo ZHOU, Tiange SI, Yuguo LIU
-
Patent number: 8683383Abstract: Unused menu items are automatically removed from a TV menu such as a cross-media bar (XMB).Type: GrantFiled: October 30, 2007Date of Patent: March 25, 2014Assignees: Sony Corporation, Sony Electronics Inc.Inventor: Satoshi Ishiguro
-
Publication number: 20140082501Abstract: A context aware service provision method and apparatus for recognizing the user context and executing an action corresponding to the user context according to a rule defined by the user and feeding back the execution result to the user interactively are provided. The method for providing a context-aware service includes receiving a user input, the user input being at least one of a text input and a speech input, identifying a rule including a condition and an action corresponding to the condition based on the received user input, activating the rule to detect a context which corresponds to the condition of the rule, and executing, when the context is detected, the action corresponding to the condition.Type: ApplicationFiled: September 16, 2013Publication date: March 20, 2014Applicant: Samsung Electronics Co. Ltd.Inventors: Jooyoon BAE, Minjeong KO, Sungsoo KIM, Jinsung KIM, Hwakyung KIM, Jinha JUN
-
Patent number: 8677238Abstract: Methods, systems, and computer programs for controlling a device without requiring visual contact are presented. One method includes an operation for identifying available options for a user according to a state of a program executing on a device, where the available options correspond to possible menu items. Further, the method provides an audible output of the available options and the detection of a touch-based input, where the input is received at the device without providing an identifiable visual cue to the user. Further yet, the method includes operations for determining the selected option from the available options based on the timing of the input relative to the audible output, and for performing the selected option by the computer program. In another method, the options are presented to the user via different vibration patterns or via vibrations of a second device connected to the device.Type: GrantFiled: October 21, 2010Date of Patent: March 18, 2014Assignee: Sony Computer Entertainment Inc.Inventor: Charles Nicholson
-
Publication number: 20140040746Abstract: Methods and apparatus for voice-enabling a web application, wherein the web application includes one or more web pages rendered by a web browser on a computer. At least one information source external to the web application is queried to determine whether information describing a set of one or more supported voice interactions for the web application is available, and in response to determining that the information is available, the information is retrieved from the at least one information source. Voice input for the web application is then enabled based on the retrieved information.Type: ApplicationFiled: August 2, 2012Publication date: February 6, 2014Applicant: Nuance Communications, Inc.Inventors: David E. Reich, Christopher Hardy
-
Publication number: 20140040749Abstract: A method of controlling a graphical user interface (GUI) at a wireless device is disclosed and includes storing a set of audio GUI controls at an interactive voice response server and creating an audio GUI control string that is to be communicated to the wireless device within a voice stream. The audio GUI control string corresponds to a text string that is selectably presentable at the wireless device. Further, the method can include embedding the audio GUI control string within the voice stream. Additionally, the method can include transmitting the voice stream with the embedded audio GUI control string to the wireless device.Type: ApplicationFiled: October 9, 2013Publication date: February 6, 2014Applicant: Qualcomm IncorporatedInventor: BRIAN MINEAR
-
Publication number: 20140040747Abstract: Content items can be viewed on an electronic device based upon a property defined for each of the content items, allowing the user to navigate through the content list and view or select content items. When navigating portions of the list where a selection has been made where no content items are associated with the properties, the result may be the presentation of no data. In order to re-orient users, the selection is modified to display at least one content item.Type: ApplicationFiled: August 2, 2012Publication date: February 6, 2014Applicant: RESEARCH IN MOTION LIMITEDInventor: Dan Zacharias GÄRDENFORS
-
Publication number: 20140040748Abstract: The digital assistant displays a digital assistant object in an object region of a display screen. The digital assistant then obtains at least one information item based on a speech input from a user. Upon determining that the at least one information item can be displayed in its entirety in the display region of the display screen, the digital assistant displays the at least one information item in the display region, where the display region and the object region are not visually distinguishable from one another. Upon determining that the at least one information item cannot be displayed in its entirety in the display region of the video display screen, the digital assistant displays a portion of the at least one information item in the display region, where the display region and the object region are visually distinguishable from one another.Type: ApplicationFiled: October 4, 2013Publication date: February 6, 2014Applicant: Apple Inc.Inventors: Stephen O. Lemay, Alessandro Francesco Sabatelli, Freddy Allen Anzures, Imran Chaudhri, Scott Forstall, Gregory Novick
-
Publication number: 20140040745Abstract: Methods and apparatus for voice-enabling a web application, wherein the web application includes one or more web pages rendered by a web browser on a computer. At least one information source external to the web application is queried to determine whether information describing a set of one or more supported voice interactions for the web application is available, and in response to determining that the information is available, the information is retrieved from the at least one information source. Voice input for the web application is then enabled based on the retrieved information.Type: ApplicationFiled: August 2, 2012Publication date: February 6, 2014Applicant: Nuance Communications, Inc.Inventors: David E. Reich, Christopher Hardy
-
Patent number: 8645840Abstract: Mechanisms are provided herein which accommodate the sharing of a user device among multiple users. In particular, when multiple users are detected as sharing a user device, the user interface of the user device can be automatically altered to facilitate the sharing of the user device. Data associated with each sharing user may also be presented according to each user's presentation preferences and to accommodate presentation needs of each user.Type: GrantFiled: June 7, 2010Date of Patent: February 4, 2014Assignee: Avaya Inc.Inventor: Matthew R. Ellsworth
-
Publication number: 20140033045Abstract: A user interface is provided for one or more users to interact with a computer using gestures coupled with voice to navigate a network that is displayed on the computer screen by the computer application software. The combination of a gesture with a voice command is used improve the reliability of the interpretation of the intent of the user. In addition, the active user who is allowed to control the software is identified through the combined input and the movements of other users are discarded.Type: ApplicationFiled: July 23, 2013Publication date: January 30, 2014Applicant: Global Quality Corp.Inventor: Sudhir Kshirsagar
-
Patent number: 8635532Abstract: Exemplary embodiments of methods to automatically correct timing of recorded audio in GUI are summarized here. One or more controls to adjust resolution of timing and degree of correction for the audio are displayed. The resolution of timing relates to heats on a grid and is affected by the degree of correction. The degree of correction is mapped to a time interval at each beat along the grid. Next, a user manipulation of one or more controls selecting a resolution and a degree of correction is received. Correction of timing is performed according to the selected resolution and degree of correction. Correcting of timing may include aligning a transient of the audio to the beat by compressing or stretching a portion of the audio. Compressing or stretching the portion of the audio depends on a length of the portion relative to a distance between adjacent beats.Type: GrantFiled: October 5, 2009Date of Patent: January 21, 2014Assignee: Apple Inc.Inventors: Gerhard Lengeling, Sol Friedman
-
Patent number: 8635659Abstract: A television and a method are provided. The television includes a display, a first network interface to a packet-based network, a first modular card receptacle coupled to the first network interface, and an audio receiver modular card coupled to the first modular card receptacle, the audio receiver modular card to process audio data, wherein the television is operable to display video content on the display promptly upon receipt of a user request.Type: GrantFiled: June 24, 2005Date of Patent: January 21, 2014Assignee: AT&T Intellectual Property I, L.P.Inventor: Edward Walter
-
Publication number: 20130346867Abstract: Systems and methods for automatically generating a media asset segment based on verbal input are provided. Verbal input is received from a user while a media asset is being presented to the user. The verbal input is processed to extract an instruction and comment information included in the verbal input. The instruction is cross-referenced with a command database to determine whether the instruction corresponds to a segment generation command. In response to determining the instruction corresponds to the segment generation command, a segment that includes a portion of the media asset that was presented to the user when the verbal input was received is generated. The comment information is associated with the generated segment. A message that includes the generated segment and the associated comment information is transmitted to a remote server.Type: ApplicationFiled: June 25, 2012Publication date: December 26, 2013Applicant: UNITED VIDEO PROPERTIES, INC.Inventors: Thomas Steven Woods, William J. Korbecki, Jonathan Arme, Brian Craig Peterson
-
Patent number: 8615316Abstract: An apparatus for processing an audio signal and method thereof are disclosed. The present invention includes receiving the audio signal and preset information; obtaining preset matrix from the preset information, wherein the preset matrix indicates contribution degree of the object to output channel; and adjusting output level of the object by using the preset matrix. Accordingly, without user's setting for each object, if preset metadata to be applied to an audio signal is selected with reference to previously-set preset metadata, levels of objects included in the audio signal can be easily adjusted using preset rendering data corresponding to the selected preset metadata.Type: GrantFiled: January 23, 2009Date of Patent: December 24, 2013Assignee: LG Electronics Inc.Inventors: Hyen O Oh, Yang Won Jung
-
Publication number: 20130339859Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for connecting an interactive wearable device with a network. In one aspect, a method includes loading content from a playlist; recognizing contextual information relating to the content; determining the location of the user; requesting supplemental content via a network based on the contextual information and the location; displaying supplemental information to a user; interacting with the supplemental information at least in part via an interactive headphone.Type: ApplicationFiled: March 13, 2013Publication date: December 19, 2013Applicant: MUZIK LLCInventor: Jason Hardi
-
Publication number: 20130339858Abstract: The technology of the present application provides a method and apparatus to manage speech resources. The method includes detecting a change in a speech application that requires the use of different resources. On detection of the change, the method loads the different resources without the user needing to exit the currently executing speech application. The apparatus provides a switch (which could be a physical or virtual switch) that causes a speech recognition system to identify audio as either commands or text.Type: ApplicationFiled: June 13, 2012Publication date: December 19, 2013Applicant: nVoq IncorporatedInventor: Charles Corfield
-
Patent number: 8612234Abstract: A method is disclosed for applying a multi-state barge-in acoustic model in a spoken dialogue system. The method includes receiving an audio speech input from the user during the presentation of a prompt, accumulating the audio speech input from the user, applying a non-speech component having at least two one-state Hidden Markov Models (HMMs) to the audio speech input from the user, applying a speech component having at least five three-state HMMs to the audio speech input from the user, in which each of the five three-state HMMs represents a different phonetic category, determining whether the audio speech input is a barge-in-speech input from the user, and if the audio speech input is determined to be the barge-in-speech input from the user, terminating the presentation of the prompt.Type: GrantFiled: October 24, 2011Date of Patent: December 17, 2013Assignee: AT&T Intellectual Property I, L.P.Inventor: Andrej Ljolje
-
Publication number: 20130326353Abstract: A wireless communication device with a voice-input and display-touch interface has an interface processor that enables, in part (i) an either display-touch or a voice-input based interface, and in part (ii) only a voice-input based interface for efficiently searching information databases. A sequence of context based search verb and search term is selected via either touch or voce selection and then the human articulated voice query is expanded using a culture and a world intelligence dictionary for conducting more efficient searches though a voice-based input.Type: ApplicationFiled: September 10, 2012Publication date: December 5, 2013Inventor: Tara Chand Singhal
-
Patent number: 8600762Abstract: A method for detecting a character or a word emphasized by a user from a voice inputted in a mobile terminal to refer it as meaningful information for a voice recognition, or emphatically displaying the user-emphasized character or word in a pre-set format when the inputted voice is converted into text, and a mobile terminal implementing the same are disclosed. The mobile terminal includes: a microphone to receive a voice of user; a controller to convert the received voice into corresponding text and detect a character or a word emphatically pronounced by the user from the voice; and a display unit to emphatically display the detected character or word in a pre-set format when the converted text is displayed.Type: GrantFiled: April 17, 2009Date of Patent: December 3, 2013Assignee: LG Electronics Inc.Inventor: Jong-Ho Shin
-
Patent number: 8589160Abstract: Some embodiments disclosed herein store a target application and a dictation application. The target application may be configured to receive input from a user. The dictation application interface may include a full overlay mode option, where in response to selection of the full overlay mode option, the dictation application interface is automatically sized and positioned over the target application interface to fully cover a text area of the target application interface to appear as if the dictation application interface is part of the target application interface. The dictation application may be further configured to receive an audio dictation from the user, convert the audio dictation into text, provide the text in the dictation application interface and in response to receiving a first user command to complete the dictation, automatically copy the text from the dictation application interface and inserting the text into the target application interface.Type: GrantFiled: August 19, 2011Date of Patent: November 19, 2013Assignee: Dolbey & Company, Inc.Inventors: Curtis A. Weeks, Aaron G. Weeks, Stephen E. Barton
-
Patent number: 8572505Abstract: An automatic testing application is provided that can simulate a user interaction with a web application. The automatic testing application can determine that the web application has been completely displayed within a graphical user interface of a web browser using a consistent global page state. The automatic testing application can test that the web application has been correctly displayed by verifying a state of a visualization component of the web application using a communication component. Subsequently, the automatic testing application can transmit an event to the web application that can modify the state of the visualization component using the communication component.Type: GrantFiled: January 31, 2011Date of Patent: October 29, 2013Assignee: Oracle International CorporationInventors: Teck Hua Lee, Gary Look, Hugh Zhang, Diar Ahmed, Prashant Singh
-
Patent number: 8572487Abstract: The present invention is directed to a method and system for applying ratings to digital music and for playing back digital music conforming to a rating level. A method for playing back digital music conforming to a rating level, includes: sending a request for metadata from a music player to a metadata service; receiving metadata sent back from the metadata service in response to the request, the metadata comprising information regarding at least one explicit interval of the track; playing the track on the music player; and selectively applying an edit to each explicit interval of the playing track identified in the metadata, based on a predetermined rating level.Type: GrantFiled: October 9, 2007Date of Patent: October 29, 2013Assignee: International Business Machines CorporationInventor: Lawrence S. Rich
-
Publication number: 20130283168Abstract: A conversation user interface enables users to better understand their interactions with computing devices, particularly when speech input is involved. The conversation user interface conveys a visual representation of a conversation between the computing device, or virtual assistant thereon, and a user. The conversation user interface presents a series of dialog representations that show input from a user (verbal or otherwise) and responses from the device or virtual assistant. Associated with one or more of the dialog representations are one or more graphical elements to convey assumptions made to interpret the user input and derive an associated response. The conversation user interface enables the user to see the assumptions upon which the response was based, and to optionally change the assumption(s). Upon change of an assumption, the conversation GUI is refreshed to present a modified dialog representation of a new response derived from the altered set of assumptions.Type: ApplicationFiled: April 18, 2012Publication date: October 24, 2013Applicant: Next IT CorporationInventors: Fred A. Brown, Tanya M. Miller, Charles C. Wooters, Bryan Michael Culley, Eli D. Snavely
-
Publication number: 20130283169Abstract: Examples of systems and methods for voice-based navigation in one or more virtual areas that define respective persistent virtual communication contexts are described. These examples enable communicants to use voice commands to, for example, search for communication opportunities in the different virtual communication contexts, enter specific ones of the virtual communication contexts, and bring other communicants into specific ones of the virtual communication contexts. In this way, these examples allow communicants to exploit the communication opportunities that are available in virtual areas, even when hands-based or visual methods of interfacing with the virtual areas are not available.Type: ApplicationFiled: April 11, 2013Publication date: October 24, 2013Applicant: Social Communications CompanyInventor: David Van Wie
-
Publication number: 20130283167Abstract: Embodiments relate to systems and methods providing a flip-though format for viewing notification of messages and related items on devices, for example personal mobile devices such as smart phones. According to an embodiment, an unread item most recently received is shown in full screen on the mobile device. While the user is viewing this item, the device will automatically retrieve and load into a cache memory, the next most recently received item. When the user is done viewing the item most recently received, the user can swipe a finger across the touch screen to trigger a page flipping animation and display of the next most recently received item. Embodiments avoid the user having to click back and forth between a list of notifications/links and corresponding notification items.Type: ApplicationFiled: April 18, 2012Publication date: October 24, 2013Applicant: SAP AGInventor: Jian Xu
-
Publication number: 20130275875Abstract: The method includes automatically, without user input and without regard to whether a digital assistant application has been separately invoked by a user, determining that the electronic device is in a vehicle. In some implementations, determining that the electronic device is in a vehicle comprises detecting that the electronic device is in communication with the vehicle (e.g., via a wired or wireless communication techniques and/or protocols). The method also includes, responsive to the determining, invoking a listening mode of a virtual assistant implemented by the electronic device. In some implementations, the method also includes limiting the ability of a user to view visual output presented by the electronic device, provide typed input to the electronic device, and the like.Type: ApplicationFiled: June 8, 2013Publication date: October 17, 2013Inventors: Thomas R. Gruber, Harry J. Saddler, Lia T. Napolitano, Emily Clark Schubert, Brian Conrad Sumner
-
Patent number: 8549402Abstract: A stand-alone desktop application (i.e., client) that resides on an end user's local computer and receives customized music playlists and content via the Internet from a network of distributed servers. The end user may install the application in a variety of ways, including, but not limited to, downloading it from the Internet or installing it from a CD that has been pre-loaded with the client. Songs downloaded through the application may be cached locally. The application forms a playlist based on user-entered preferences, an analysis of music already existing on the computer, and previous user interactions with the application. Similar or related songs, albums, bands, or artists can be displayed graphically in a “song cloud” around a graphical representation of the song being played.Type: GrantFiled: December 29, 2008Date of Patent: October 1, 2013Inventors: Joseph Harold Moore, Stephen Rhett Davis
-
Publication number: 20130246920Abstract: A method of enabling voice input for a graphical user interface (GUI) based application on an electronic device. The method includes: obtaining required properties of one or more user interface objects of the GUI-based application, wherein the one or more user interface objects include one or more input objects; receiving a voice input; extracting from the voice input one or more elements; associating the one or more elements with the one or more input objects; identifying, based on said associating, an input object having a required property which is not satisfied; and outputting, based on the required property, audio output for a prompt for a further voice input.Type: ApplicationFiled: January 29, 2013Publication date: September 19, 2013Applicant: RESEARCH IN MOTION LIMITEDInventors: Gregory Jason FIELDS, Karthigesu NIRANJAN
-
Publication number: 20130239000Abstract: In one embodiment, a method for displaying a user interface on a display of a head worn computer can include displaying a first layer of information in the user interface on a display of the head worn computer. The method can further include receiving a directional input from body movement, eye tracking, or hand gestures. The method can additionally include highlighting an area of the user interface on the display with a second layer of information. The area can be located in the user interface based on the received directional input.Type: ApplicationFiled: March 13, 2013Publication date: September 12, 2013Applicant: Kopin CorporationInventors: Christopher Parkinson, Luke Hopkins, David Niland
-
Patent number: 8533754Abstract: A method, device and/or system for operating or causing a video player embedded into a web page to play a video is disclosed. Embed code is placed into the web page. When the web page is loaded into a web browser, player code is loaded and optionally customized to display one or more videos that can be selected for playback with the player code. When an end user selects the video using the player code, any ad and/or analytics module associated with the video are retrieved. The functionality of the ad and/or analytics module is integrated into the player code to allow tracking of analytics and inserting advertising into the video.Type: GrantFiled: December 30, 2011Date of Patent: September 10, 2013Assignee: Limelight Networks, Inc.Inventor: Paul Cho
-
Patent number: 8527878Abstract: A computer program of the type commonly known as a “wizard” is disclosed that initializes user interface software for controlling an audio conferencing device. The wizard allows the desired audio inputs (e.g., microphone, telephones, etc.) and audio outputs (speakers, recording devices, etc.) to be chosen by an audio system administrator. Thereafter, the wizard allows an audio conferencing device (or devices) to be chosen by the administrator, or allows such a device(s) to be optimally chosen dependent upon the chosen inputs and outputs. The wizard then maps the inputs and outputs to the input and output ports on the audio conferencing device. When the administrator finishes the wizard, the wizard computes the mapping parameters and other audio-optimizing parameters for the selected inputs and outputs. These parameters are then loaded into the user interface software automatically.Type: GrantFiled: September 12, 2011Date of Patent: September 3, 2013Assignee: Polycom, Inc.Inventors: Thomas M. Drewes, James S. Joiner, Michael A. Pocino, Craig H. Richardson
-
Publication number: 20130227418Abstract: Users are enabled to define and modify mappings between (1) gestures and (2) actions performed by one or more computing devices in response to a device detecting performance of a gesture. A generalized gesture-to-action mapping framework allows users to intuitively define and modify such mappings. In response to a device detecting the performance of a particular gesture, one or more devices may cause the execution of one or more device actions based on a set of user-defined gesture-to-action mappings.Type: ApplicationFiled: February 27, 2012Publication date: August 29, 2013Inventors: MARCO DE SA, Elizabeth F. Churchill, R. Preston McAfee, David Ayman Shamma
-
Publication number: 20130227419Abstract: An apparatus to switch an application includes an input unit to receive an input for switching a foreground application, the input including an application distinguishing portion associated with an application switching portion, a control unit to determine an application to be run in the foreground among the applications running in a background, the application distinguishing portion corresponding to the application, and an output unit to output the application in a display as the foreground application.Type: ApplicationFiled: October 23, 2012Publication date: August 29, 2013Applicant: PANTECH CO., LTD.Inventor: Pantech Co.,Ltd.
-
Patent number: 8521534Abstract: A prompt generation engine operates to dynamically extend prompts of a multimodal application. The prompt generation engine receives a media file having a metadata container. The prompt generation engine operates on a multimodal device that supports a voice mode and a non-voice mode for interacting with the multimodal device. The prompt generation engine retrieves from the metadata container a speech prompt related to content stored in the media file for inclusion in the multimodal application. The prompt generation engine modifies the multimodal application to include the speech prompt.Type: GrantFiled: September 12, 2012Date of Patent: August 27, 2013Assignee: Nuance Communications, Inc.Inventors: Ciprian Agapi, William K. Bodin, Charles W. Cross, Jr.
-
Publication number: 20130219277Abstract: A computer readable storage medium stores instructions defining a mobile device browser. The mobile device browser supports direct command inputs and executable instructions to correlate a proxy command to a selected direct command input. The proxy command is alternately expressed as a gesture and a voice command. The selected direct command input is automatically executed by the mobile device browser.Type: ApplicationFiled: February 21, 2012Publication date: August 22, 2013Applicant: MoboTap Inc.Inventors: Yu Wang, Yan Yu, Jia Yuan, Yongzhi Yang, Tiefeng Liu
-
Publication number: 20130205214Abstract: A computerized information apparatus useful for providing information to a user via a display. In one embodiment, the apparatus comprises a processor and network interface and computer readable medium having at least one computer program disposed thereon, the at least one program being configured to receive a speech input from the user, and obtain information relating to the input. In one variant, at least a portion of the information is obtained via the network interface from a remote server. An information and control system for personnel transport devices. In one embodiment, the information and control system is coupled to the elevator system of a building, and includes a touch panel input device, a flat panel display having a touch sensitive screen, and speech recognition and synthesis systems serving each elevator car. The speech recognition and synthesis systems and input device(s) are operatively coupled to a processor and storage devices having a plurality of different types of data stored thereon.Type: ApplicationFiled: December 27, 2012Publication date: August 8, 2013Applicant: WEST VIEW RESEARCH, LLCInventor: WEST VIEW RESEARCH, LLC
-
Publication number: 20130191750Abstract: Apparatus useful for obtaining and displaying information. In one embodiment, the apparatus includes a network interface, display device, and speech recognition apparatus configured to receive user speech input and enable performance of various tasks via a remote entity, such as obtaining desired information relating to directions, sports, finance, weather, or any number of other topics. The downloaded may also, in one variant, be transmitted to a personal user device, such as via a data interface.Type: ApplicationFiled: January 2, 2013Publication date: July 25, 2013Applicant: WEST VIEW RESEARCH, LLCInventor: West View Research, LLC
-
Publication number: 20130185640Abstract: A computerized information and display apparatus useful for providing information to a user via a display. In one embodiment, the apparatus comprises a processor and network interface and computer readable medium having at least one computer program disposed thereon, the at least one program being configured to receive a speech input from the user, and obtain information relating to the input. In one variant, at least a portion of the information is obtained via the network interface from a remote server, and the apparatus includes two components in wireless communication with one another.Type: ApplicationFiled: December 27, 2012Publication date: July 18, 2013Applicant: West View Research, LLCInventor: West View Research, LLC
-
Patent number: 8488014Abstract: The present invention relates to a method to create and reproduce a panoramic sound image. Initially, there is a plurality of images associated with sound sequences, these images are for example taken using a camera, this camera having recorded the ambient sound at the time of taking the photograph. The ambient sound is then filtered to keep only the stationary component. Then, the images are concatenated to form a unique image called “panoramic”, each sound sequence being associated with a part of the panoramic image containing the image associated with this sequence. Finally, the commands are introduced to display the parts of the panoramic image. The command also enables to reproduce continually the stationary component of at least a sound content associated with at least a displayed part of the panoramic image. According to an improvement, the sound sequence is reproduced continually until the introduction of a new command.Type: GrantFiled: May 27, 2008Date of Patent: July 16, 2013Assignee: Thomson LicensingInventors: Louis Chevallier, Lionel Oisel, Jean-Ronan Vigouroux
-
Patent number: 8489992Abstract: In one embodiment, a graphics user interface is provided. The graphics user interface includes a plurality of graphical representations identifying separate audio data, respectively. Each of the plurality of graphical representations is configured in a list to be selected for playback of the respective audio data. A progression icon is displayed in each of the respective graphical representations. Each progression icon illustrates a temporal progression of the playback of the respective audio data.Type: GrantFiled: April 8, 2008Date of Patent: July 16, 2013Assignee: Cisco Technology, Inc.Inventor: Ruben Rohde
-
Patent number: 8478600Abstract: Provided is an input/output apparatus based on voice recognition, and a method thereof. An object of the apparatus is to improve a user interface by making pointing input and command execution such as application program control possible according to a voice command of a user possible based on a voice recognition technology without individual pointing input device such as a mouse and a touch pad, and a method thereof. The apparatus includes: a voice recognizer for recognizing a voice command inputted from outside; a pointing controller for calculating a pointing location on a screen which corresponds to a voice recognition result transmitted from the voice recognizer; a displayer for displaying a screen; and a command controller for processing diverse commands related to a current pointing location.Type: GrantFiled: September 11, 2006Date of Patent: July 2, 2013Assignee: Electronics and Telecommunications Research InstituteInventors: Kwan-Hyun Cho, Mun-Sung Han, Jun-Seok Park, Young-Giu Jung
-
Patent number: 8473857Abstract: A computer navigation system and method has one or more interactive links displayed on a display connected to a computer appliance, and one or more visual linktags associated with individual ones of the one or more interactive links, the linktags displaying one or more numbers, characters or symbols, the system enabled to initiate an interactive link in the display upon user input of one of the numbers, characters or symbols in a linktag.Type: GrantFiled: February 12, 2010Date of Patent: June 25, 2013Assignee: Google Inc.Inventor: Fritz Schneider
-
Patent number: 8464152Abstract: An apparatus and method for providing instructional help, at multiple levels of sophistication, in a learning application includes assistance in the form to at least two optional levels of sophistication. The user can select any level of sophistication according to desire or need. The levels allow for flexibility in learning relative to the person involved, the amount of previous knowledge the person has about the subject matter, the differences in how different people learn. In one embodiment, the assistance is contained in programming on a CD-ROM which is used in an interactive computerized system.Type: GrantFiled: October 31, 2007Date of Patent: June 11, 2013Inventor: Karen A. McKirchy
-
Patent number: RE44326Abstract: A method and system of speech recognition presented by a back channel from multiple user sites within a network supporting cable television and/or video delivery is disclosed.Type: GrantFiled: November 3, 2011Date of Patent: June 25, 2013Assignee: Promptu Systems CorporationInventors: Theodore Calderone, Paul M. Cook, Mark J. Foster