Modification Of At Least One Characteristic Of Speech Waves (epo) Patents (Class 704/E21.001)
  • Publication number: 20140067404
    Abstract: A system and method for selectively applying Intensity Stereo coding to an audio signal is described. The system and method make decisions on whether to apply Intensity Stereo coding to each scale factor band of the audio signal based on (1) the number of bits necessary to encode each scale factor band using Intensity Stereo coding, (2) spatial distortions generated by using Intensity Stereo coding with each scale factor band, and (3) switching distortions for each scale factor band resulting from switching Intensity Stereo coding on or off in relation to a previous scale factor band.
    Type: Application
    Filed: September 4, 2012
    Publication date: March 6, 2014
    Applicant: Apple Inc.
    Inventor: Frank M. Baumgarte
  • Publication number: 20140067403
    Abstract: A method of managing speech interfaces to computer-based services includes beginning a first speech session that is carried out in a vehicle over a short-range wireless connection between a vehicle occupant and a mobile device; detecting an initiation of a second speech session while the first speech session is being carried out; determining an assigned priority level of the first speech session relative to an assigned priority level of the second speech session; and when the assigned priority level of the second speech session has a higher priority than the assigned priority level of the first speech session, carrying out a session-appropriate action on the first speech session.
    Type: Application
    Filed: September 6, 2012
    Publication date: March 6, 2014
    Applicant: GM GLOBAL TECHNOLOGY OPERATIONS LLC
    Inventors: Denis R. Burke, Danilo Gurovich, Daniel E. Rudman, Keith A. Fry, Shane M. McCutchen, Marco T. Carnevale, Mukesh Gupta
  • Publication number: 20140052451
    Abstract: Methods and apparatus for providing a search interface for an electronic device including a tuner configured to tune the electronic device to receive scheduled programming content. A search query is received and one or more data sources including information about media content are searched based, at least in part, on the search query. The results of the search are presented on a user interface using a time-based axis and a time-independent axis.
    Type: Application
    Filed: August 16, 2012
    Publication date: February 20, 2014
    Applicant: Nuance Communications, Inc.
    Inventors: Yuen-Keen Cheong, Steven A. Hatch, Hoi L. Young, Tapio I. Koivuniemi
  • Publication number: 20140052450
    Abstract: Methods and apparatus for providing a search interface for an electronic device including a tuner configured to tune the electronic device to receive scheduled programming content. A search query is received and one or more data sources including information about media content are searched based, at least in part, on the search query. The results of the search are presented on a user interface using a time-based axis and a time-independent axis.
    Type: Application
    Filed: August 16, 2012
    Publication date: February 20, 2014
    Applicant: Nuance Communications, Inc.
    Inventors: Yuen-Keen CHEONG, Steven A. HATCH, Hoi L. YOUNG, Tapio I. KOIVUNIEMI
  • Publication number: 20140052438
    Abstract: In a computer system that permits multiple audio capture applications to get an audio capture feed concurrently, an audio manager manages audio capture and/or audio playback in reaction to trigger events. For example, a trigger event indicates an application has started, stopped or otherwise changed a communication stream, or indicates an application has gained, lost or otherwise changed focus or visibility in a user interface, or indicates a user change. In response to a trigger event, the audio manager applies a set of rules to determine which audio capture application is allowed to get an audio capture feed. Based on the decisions, the audio manager manages the audio capture feed for the applications. The audio manager also sends a notification to each of the audio capture applications that has registered for notifications, so as to indicate whether the application is allowed to get the audio capture feed.
    Type: Application
    Filed: August 20, 2012
    Publication date: February 20, 2014
    Applicant: Microsoft Corporation
    Inventors: Frank Yerrace, Kishore Kotteri, Ryan Beberwyck, Gerrit Swaneveld, John Bregar, Rian Chung
  • Publication number: 20140052453
    Abstract: Methods and apparatus for searching for content to display on a digitally-tunable electronic device configured to display scheduled programming content. The method comprises receiving a search query from a user, and determining, based on the search query, an action the user wants to perform. The method further comprises determining one or more data sources to search based, at least in part, on the action the user wants to perform, and searching based, at least in part, on the search query, the one or more data sources for the content to display on the electronic device.
    Type: Application
    Filed: August 16, 2012
    Publication date: February 20, 2014
    Inventors: Tapio I. Koivuniemi, Tuomas A. Tuononen, Jarkko Koivikko, Teijo J. Kinnunen
  • Publication number: 20140052452
    Abstract: Methods and apparatus for providing a search interface for an electronic device including a tuner configured to tune the electronic device to receive scheduled programming content. A search query is received and one or more data sources including information about media content are searched based, at least in part, on the search query. The results of the search are presented on a user interface using a time-based axis and a time-independent axis.
    Type: Application
    Filed: August 16, 2012
    Publication date: February 20, 2014
    Applicant: Nuance Communications, Inc.
    Inventors: Tapio I. Koivuniemi, Tuomas A. Tuononen, Jarkko Koivikko, Teijo J. Kinnunen
  • Publication number: 20140039892
    Abstract: In one embodiment, a human interactive proof portal 140 may use a biometric input to determine whether a user is a standard user or a malicious actor. The human interactive proof portal 140 may receive an access request 302 for an online data service 122 from a user device 110. The human interactive proof portal 140 may send a proof challenge 304 to the user device 110 for presentation to a user. The human interactive proof portal 140 may receive from the user device 110 a proof response 306 having a biometric metadata description 430 based on a biometric input from the user.
    Type: Application
    Filed: August 2, 2012
    Publication date: February 6, 2014
    Applicant: Microsoft Corporation
    Inventors: Chad Mills, Robert Sim, Scott Laufer, Sung Chung
  • Publication number: 20140039898
    Abstract: Methods and apparatus for voice-enabling a web application, wherein the web application includes one or more web pages rendered by a web browser on a computer. At least one information source external to the web application is queried to determine whether information describing a set of one or more supported voice interactions for the web application is available, and in response to determining that the information is available, the information is retrieved from the at least one information source. Voice input for the web application is then enabled based on the retrieved information.
    Type: Application
    Filed: August 2, 2012
    Publication date: February 6, 2014
    Applicant: Nuance Communications, Inc.
    Inventors: David E. Reich, Christopher Hardy
  • Publication number: 20140039882
    Abstract: The instant application includes computationally-implemented systems and methods that include managing adaptation data, wherein the adaptation data is correlated to at least one aspect of speech of a particular party, facilitating transmission of the adaptation data to a target device, wherein the adaptation data is configured to be applied to the target device to assist in execution of a speech-facilitated transaction, facilitating reception of adaptation result data that is based on at least one aspect of the speech-facilitated transaction between the particular party and the target device, determining whether to modify the adaptation data at least partly based on the adaptation result data, and facilitating transmission of at least a portion of modified adaptation data to a receiving device. In addition to the foregoing, other aspects are described in the claims, drawings, and text.
    Type: Application
    Filed: August 1, 2012
    Publication date: February 6, 2014
    Inventors: Royce A. Levien, Richard T. Lord, Robert W. Lord, Mark A. Malamud
  • Publication number: 20140029701
    Abstract: Systems, methods, and devices synchronize data streams by hashing received data frames to generate a sequence of hash values, comparing the generated hash value sequence to a hash value sequence received in a control stream, and processing data frames when the hash value sequences match. A source device and multiple receiver devices may synchronize audio data encoded in data frames, applying a hash function to each data frame to generate a first sequence of hash values, transmitting the data frames on a first channel and the first sequence of hash values on a control channel, receiving the data frames and the first sequence of hash values in the receiver devices, applying the hash algorithm to received data frames to generate a second sequence of hash values, comparing the first and second sequences of hash values, and processing data frames when the first and second sequences of hash values match.
    Type: Application
    Filed: July 29, 2012
    Publication date: January 30, 2014
    Inventors: Adam E. NEWHAM, Joel Benjamin LINSKY, Rohit SAUHTA, Brian F. MILLER, Kevin Wayne BARTIG
  • Publication number: 20140019139
    Abstract: A blood glucose meter with a simplified programmable voice function, including: a microprocessor; a memory that is both programmable and re-programmable coupled to the microprocessor; and an audio output device coupled to the microprocessor and the memory; wherein a language algorithm and a plurality of language components specific to a language selected by a user are disposed within the memory; and wherein the language algorithm and the plurality of language components are utilized to provide an audio output through the audio output device in the language selected by the user. The language algorithm is operable for determining which language components are utilized to provide the audio output and in what order based on the language selected by the user. Optionally, the audio output is generated by the microprocessor and the memory using a pulse-width modulation scheme and/or the like.
    Type: Application
    Filed: July 12, 2012
    Publication date: January 16, 2014
    Applicant: PRODIGY DIABETES CARE, LLC
    Inventors: Ramzi ABULHAJ, Moo Nam KO, William BAXTER, Amr Yehia Mohamed SHEHAB
  • Publication number: 20140012586
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for determining hotword suitability. In one aspect, a method includes receiving speech data that encodes a candidate hotword spoken by a user, evaluating the speech data or a transcription of the candidate hotword, using one or more predetermined criteria, generating a hotword suitability score for the candidate hotword based on evaluating the speech data or a transcription of the candidate hotword, using one or more predetermined criteria, and providing a representation of the hotword suitability score for display to the user.
    Type: Application
    Filed: August 6, 2012
    Publication date: January 9, 2014
    Applicant: GOOGLE INC.
    Inventors: Andrew E. Rubin, Johan Schalkwyk, Maria Carolina Parada San Martin
  • Publication number: 20130346084
    Abstract: Technologies are described herein for enhancing a user presence status determination. Visual data may be received from a depth camera configured to be arranged within a three-dimensional space. A current user presence status of a user in the three-dimensional space may be determined based on the visual data. A previous user presence status of the user may be transformed to the current user presence status, responsive to determining the current user presence status of the user.
    Type: Application
    Filed: June 22, 2012
    Publication date: December 26, 2013
    Applicant: MICROSOFT CORPORATION
    Inventors: Anne Marie Renée Archambault, Jeffrey Scott Berg, Xiping Zuo, Abhishek Agrawal
  • Publication number: 20130346085
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, and including hardware devices performing a mouth click sound based human-device interaction. In one aspect, receiving at least one mouth click sound signals from a human user, by an acoustic-to-electric sensor of a computing device, and processing the received signals. The received mouth click sound signals may be accompanied by other mouth click sound signals, and other interaction signals.
    Type: Application
    Filed: June 23, 2012
    Publication date: December 26, 2013
    Inventor: Zoltan Stekkelpak
  • Publication number: 20130339007
    Abstract: Embodiments herein include receiving a request to modify an audio characteristic associated with a first user for a voice communication system. One or more suggested modified audio characteristics may be provided for the first user, based on, at least in part, one or more audio preferences established by another user. An input of one or more modified audio characteristics may be received for the first user for the voice communication system. A user-specific audio preference may be associated with the first user for voice communications on the voice communication system, the user-specific audio preference including the one or more modified audio characteristics.
    Type: Application
    Filed: June 18, 2012
    Publication date: December 19, 2013
    Applicant: International Business Machines Corporation
    Inventors: Ruthie D. Lyle, Patrick Joseph O'Sullivan, Lin Sun
  • Publication number: 20130332172
    Abstract: An accessory is configured to receive a request. The accessory transmits information associated with the request to a portable device. An automated assistant application executed by the portable device can interpret the request and provide a report. The portable device can transmit the report to the accessory. The report may include one or more results determined by the automated assistant.
    Type: Application
    Filed: October 1, 2012
    Publication date: December 12, 2013
    Applicant: Apple Inc.
    Inventors: Jude A. Prakash, Shailesh Rathi, Daniel De Rocha Rosario, Sylvain R.Y. Louboutin
  • Publication number: 20130325462
    Abstract: A system and method for assigning one or more tags to an image file. In one aspect, a server computer receives an image file captured by a client device. In one embodiment, the image file includes an audio component embedded therein by the client device, where the audio component was spoken by a user of the client device as a tag of the image file. The server computer determines metadata associated with the image file and identifies a dictionary of potential textual tags from the metadata. The server computer determines a textual tag from the audio component and from the dictionary of potential textual tags. The server computer then associates the textual tag with the image file as additional metadata.
    Type: Application
    Filed: May 31, 2012
    Publication date: December 5, 2013
    Applicant: Yahoo! inc.
    Inventors: Oren Somekh, Nadav Golbandi, Liran Katzir, Ronny Lempel, Yoelle Maarek
  • Publication number: 20130325481
    Abstract: A method of providing navigation on an electronic device when the display screen is locked. The method receives a verbal request to start navigation while the display is locked. The method identifies a route from a current location to a destination based on the received verbal request. While the display screen is locked, the method provides navigational directions on the electronic device from the current location of the electronic device to the destination. Some embodiments provide a method for processing a verbal search request. The method receives a navigation-related verbal search request and prepares a sequential list of the search results based on the received request. The method then provides audible information to present a search result from the sequential list. The method presents the search results in a batch form until the user selects a search result, the user terminates the search, or the search items are exhausted.
    Type: Application
    Filed: September 30, 2012
    Publication date: December 5, 2013
    Applicant: APPLE INC.
    Inventors: Marcel van Os, Sarah G. Barbour, Brady A. Law, Bradford A. Moore
  • Publication number: 20130321390
    Abstract: A system and method are disclosed for augmenting a reading experience in a mixed reality environment. In response to predefined verbal or physical gestures, the mixed reality system is able to answer a user's questions or provide additional information relating to what the user is reading. Responses may be displayed to the user on virtual display slates in a border or around the reading material without obscuring text or interfering with the user's reading experience.
    Type: Application
    Filed: May 31, 2012
    Publication date: December 5, 2013
    Inventors: Stephen G. Latta, Ryan L. Hastings, Cameron G. Brown, Aaron Krauss, Daniel J. McCulloch, Ben J. Sugden
  • Publication number: 20130317828
    Abstract: The effectiveness of targeted content delivery at a multi-user interface can be directly linked to a proper targeting of users. A way of improving targeted content delivery at a multi-user interface can be to determine which users should be targeted based on one or more criteria. The present technology provides various methodologies for selecting one or more users associated with a multi-user interface to receive targeted content. Such users can be selected based on criteria associated with a ranking or priority of the users, criteria associated with an analysis of their interactions with the multi-user interface, criteria based on their most common characteristics, or any combination thereof. The user characteristics associated with such identified used can then be utilized to determine which content should be delivered to the multi-user interface.
    Type: Application
    Filed: May 25, 2012
    Publication date: November 28, 2013
    Applicant: Apple Inc.
    Inventors: Michael Froimowitz Greenzeiger, Mehul K. Sanghavi, Ravindra Phulari
  • Publication number: 20130304479
    Abstract: Methods and systems for determining intent in voice and gesture interfaces are described. An example method includes determining that a gaze direction is in a direction of a gaze target, and determining whether a predetermined time period has elapsed while the gaze direction is in the direction of the gaze target. The method may also include providing an indication that the predetermined time period has elapsed when the predetermined time period has elapsed. According to the method, a voice or gesture command that is received after the predetermined time period has elapsed may be determined to be an input for a computing system. Additional example systems and methods are described herein.
    Type: Application
    Filed: May 8, 2012
    Publication date: November 14, 2013
    Applicant: GOOGLE INC.
    Inventors: Eric Teller, Daniel Aminzade
  • Publication number: 20130297296
    Abstract: Methods and apparatus for signal processing are disclosed. Source separation can be performed to extract source signals from mixtures of source signals by way of independent component analysis. Source direction information is utilized in the separation process, and independent component analysis techniques described herein use multivariate probability density functions to preserve the alignment of frequency bins in the source separation process. It is emphasized that this abstract is provided to comply with the rules requiring an abstract that will allow a searcher or other reader to quickly ascertain the subject matter of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims.
    Type: Application
    Filed: May 4, 2012
    Publication date: November 7, 2013
    Applicant: Sony Computer Entertainment Inc.
    Inventors: Jaekwon Yoo, Ruxin Chen
  • Publication number: 20130297319
    Abstract: A mobile device having at least one microphone sensor and a method for controlling the same are disclosed. The method includes receiving at least two audio signals through the at least one microphone sensor within a predetermined time period, recognizing input directions and voice command from the at least two audio signals sequentially, determining whether the recognized input directions and voice command match to preset input directions and preset voice command mapped to the preset directions, sequentially for the at least two received audio signals, and executing a preset control command, if the recognized input directions and voice command match to the preset input directions and voice command.
    Type: Application
    Filed: July 9, 2012
    Publication date: November 7, 2013
    Inventor: Yongsin KIM
  • Publication number: 20130297298
    Abstract: Methods and apparatus for signal processing are disclosed. Source separation can be performed to extract source signals from mixtures of source signals by way of independent component analysis. Source separation described herein involves mixed multivariate probability density functions that are mixtures of component density functions having different parameters corresponding to frequency components of different sources, different time segments, or some combination thereof.
    Type: Application
    Filed: May 4, 2012
    Publication date: November 7, 2013
    Applicant: Sony Computer Entertainment Inc.
    Inventors: Jaekwon Yoo, Ruxin Chen
  • Publication number: 20130297320
    Abstract: An additive three-dimensional fabrication system includes voice control for user interaction. This voice-controlled interface can enable a variety of voice-controlled functions and operations, while supporting interactions specific to consumer-oriented fabrication processes.
    Type: Application
    Filed: July 23, 2012
    Publication date: November 7, 2013
    Inventors: Anthony James Buser, Nathaniel B. Pettis
  • Publication number: 20130290000
    Abstract: A method is disclosed for controlling a voice-activated device by interpreting a spoken command as a series of voiced and non-voiced intervals. A responsive action is then performed according to the number of voiced intervals in the command. The method is well-suited to applications having a small number of specific voice-activated response functions. Applications using the inventive method offer numerous advantages over traditional speech recognition systems including speaker universality, language independence, no training or calibration needed, implementation with simple microcontrollers, and extremely low cost. For time-critical applications such as pulsers and measurement devices, where fast reaction is crucial to catch a transient event, the method provides near-instantaneous command response, yet versatile voice control.
    Type: Application
    Filed: April 30, 2012
    Publication date: October 31, 2013
    Inventor: David Edward Newman
  • Publication number: 20130282371
    Abstract: A method is disclosed herein for recognizing a repeated utterance in a mobile computing device via a processor. A first utterance is detected being spoken into a first mobile computing device. Likewise, a second utterance is detected being spoken into a second mobile computing device within a predetermined time period. The second utterance substantially matches the first spoken utterance and the first and second mobile computing devices are communicatively coupled to each other. The processor enables capturing, at least temporarily, a matching utterance for performing a subsequent processing function. The performed subsequent processing function is based on a type of captured utterance.
    Type: Application
    Filed: April 20, 2012
    Publication date: October 24, 2013
    Applicant: Motorola Mobility, Inc.
    Inventors: Rachid M Alameh, Jiri Slaby, Hisashi D. Watanabe
  • Publication number: 20130282380
    Abstract: Current human-to-machine interfaces enable users to interact with a company's database and enter into a series of transactions (e.g., purchasing products/services and paying bills). Each transaction may require several operations or stages requiring user input or interaction. Some systems enable a user to enter a voice input parameter providing multiple operations of instruction (e.g., single natural language command). However, users of such a system do not know what types of commands the system is capable of accepting. Embodiments of the present invention facilitate communications for user transactions by determining a user's goal transaction and presenting a visual representation of a voice input parameter for the goal transaction. The use of visual representations notifies the user of the system's capability of accepting single natural language commands and the types of commands the system is capable of accepting, thereby enabling a user to complete a transaction in a shorter period of time.
    Type: Application
    Filed: April 20, 2012
    Publication date: October 24, 2013
    Applicant: Nuance Communications, Inc.
    Inventors: David Andrew Mauro, Simona Gandrabur
  • Publication number: 20130282381
    Abstract: Generally, human-to-machine interfaces are configured to accept speech input from a user. However, such interfaces, e.g., web browsers, must be configured to enable acceptance of speech input from the user. Some interfaces, such as mobile browsers, have less configuration adaptability and are not able to be configured to accept speech input from a user. Embodiments of the present invention speech-enable human-to-machine interfaces by loading content of the human-to-machine interface and adding logic configured to enable speech interaction with the content to the interface. The embodiment then activates speech interaction with the content via the logic for the user. Thus, embodiments of the present invention enable speech interaction with interfaces that are not configured to be adapted to allow speech interaction and are able to enable the speech interaction in a seamless manner.
    Type: Application
    Filed: April 20, 2012
    Publication date: October 24, 2013
    Applicant: Nuance Communications, Inc.
    Inventors: David Andrew Mauro, Henri Bouvier
  • Publication number: 20130262124
    Abstract: System and method to search audio data, including: receiving audio data representing speech; receiving a search query related to the audio data; compiling, by use of a processor, the search query into a hierarchy of scored speech recognition sub-searches; searching, by use of a processor, the audio data for speech identified by one or more of the sub-searches to produce hits; and combining, by use of a processor, the hits by use of at least one combination function to provide a composite search score of the audio data. The combination function may include an at-least-M-of-N function that produces a high score when at least M of N function inputs exceed a predetermined threshold value. The composite search score employ a soft time window such as a spline function.
    Type: Application
    Filed: September 27, 2012
    Publication date: October 3, 2013
    Applicant: AURIX LIMITED
    Inventor: Keith Michael Ponting
  • Publication number: 20130262127
    Abstract: A content processing service may analyze an item of original content and identify several objects, attributes of those objects, and relationships between those objects present in the item of original content. The content processing service may also analyze a source graph, such as a social graph or supplemental graph, and identify several objects, attributes of those objects, and relationships between objects present in the source graph. The content processing service may customize the item of original content by selecting an original object and selecting a source graph object. One or more of the attributes or relationships of the selected original object in the item of original content may be replaced by one or more of the attributes or relationships of the selected source graph object. To customize items of audio content, audio content associated with the source graph object may replace audio content associated with the target graph object.
    Type: Application
    Filed: March 29, 2012
    Publication date: October 3, 2013
    Inventors: Douglas S. Goldstein, Ajay Arora, Douglas Hwang, Guy A. Story, JR., Shirley C. Yang
  • Publication number: 20130238341
    Abstract: An electronic device includes a music play module that plays music and a voice recorder that records ambient voice around the electronic device. The electronic device further includes a music control module that identifies voice characteristics of the ambient voice, and controls the music play module to pause the playing of the music when the voice characteristics of the ambient voice match pre-configured voice reference information.
    Type: Application
    Filed: April 27, 2012
    Publication date: September 12, 2013
    Applicants: HON HAI PRECISION INDUSTRY CO., LTD., FU TAI HUA INDUSTRY (SHENZHEN) CO., LTD.
    Inventor: QIANG YOU
  • Publication number: 20130231927
    Abstract: Implementations of systems, method and devices described herein enable enhancing the intelligibility of a target voice signal included in a noisy audible signal received by a hearing aid device or the like. In particular, in some implementations, systems, methods and devices are operable to generate a machine readable formant based codebook. In some implementations, the method includes determining whether or not a candidate codebook tuple includes a sufficient amount of new information to warrant either adding the candidate codebook tuple to the codebook or using at least a portion of the candidate codebook tuple to update an existing codebook tuple. Additionally and/or alternatively, in some implementations systems, methods and devices are operable to reconstruct a target voice signal by detecting formants in an audible signal, using the detected formants to select codebook tuples, and using the formant information in the selected codebook tuples to reconstruct the target voice signal.
    Type: Application
    Filed: August 20, 2012
    Publication date: September 5, 2013
    Inventors: PIERRE ZAKARAUSKAS, ALEXANDER ESCOTT, CLARENCE S.H. CHU, SHAWN E. STEVENSON
  • Publication number: 20130231930
    Abstract: A computer implemented method and apparatus for automatically filtering an audio input to make a filtered recording comprising: identifying words used in an audio input, determining whether each identified word is contained in a dictionary of banned words, and creating a filtered recording as an audio output, wherein each word identified in the audio input that is found in the dictionary of banned words, is automatically deleted or replaced in the audio output used to make the filtered recording.
    Type: Application
    Filed: March 1, 2012
    Publication date: September 5, 2013
    Applicant: Adobe Systems Inc.
    Inventor: Antonio Sanso
  • Publication number: 20130226588
    Abstract: A method is provided for a simulated conversation by a pre-recorded audio navigator, with particular application to informational and entertainment settings. A monitor may utilize a navigation interface to select pre-recorded responses in the voice of a character represented by a performer. The pre-recorded responses may then be queued and sent to a speaker proximate to the performer. By careful organization of an audio database including audio buckets and script-based navigation with shifts for tailoring to specific guest user profiles and environmental contexts, a convincing and dynamic simulated conversation may be carried out while providing the monitor with a user-friendly navigation interface. Thus, highly specialized training is not necessary and flexible scaling to large-scale deployments is readily supported.
    Type: Application
    Filed: February 28, 2012
    Publication date: August 29, 2013
    Inventors: Holger Irmler, Asa K. Kalama, Raymond J. Scanlon, Brent D. Strong, Cory J. Rouse, Renée M. Johnson, Andrew Stone
  • Publication number: 20130226589
    Abstract: A sound-activated control system includes an audio receiver and a command discriminator. The receiver is configured to receive an audio waveform and to produce a digital audio waveform therefrom. The command discriminator is configured to detect a temporally and/or spectrally compact nonphonetic audio command within the digital audio waveform and to control a voice-activated system an action in response to the nonphonetic command.
    Type: Application
    Filed: February 29, 2012
    Publication date: August 29, 2013
    Applicant: NVIDIA Corporation
    Inventor: Henry P. Largey
  • Publication number: 20130225999
    Abstract: The embodiments of the ultrasound imaging diagnostic apparatus include at least one non-touch input device for receiving a predetermined gesture as an input command. An optional sequence of predetermined gestures is inputted as an operational command and or data to the embodiments of the ultrasound imaging diagnostic apparatus. A gesture is optionally combined with other conventional input modes through devices such as a microphone, a mouse, a keyboard, a button, a panel switch, a touch command screen, a foot switch, a trackball, and the like.
    Type: Application
    Filed: February 29, 2012
    Publication date: August 29, 2013
    Applicants: TOSHIBA MEDICAL SYSTEMS CORPORATION, KABUSHIKI KAISHA TOSHIBA
    Inventors: Zoran BANJANIN, Raymond F. WOODS
  • Publication number: 20130218570
    Abstract: According to one embodiment, in an apparatus for correcting a speech corresponding to a moving image, a separation unit separates at least one audio component from each audio frame of the speech. An estimation unit estimates a scene including a plurality of image frames related in the moving image, based on at least one of a feature of each image frame of the moving image and a feature of the each audio frame. An analysis unit acquires attribute information of the plurality of image frames by analyzing the each image frame. A correction unit determines a correction method of the audio component corresponding to the plurality of image frames, based on the attribute information, and corrects the audio component by the correction method.
    Type: Application
    Filed: August 22, 2012
    Publication date: August 22, 2013
    Applicant: KABUSHIKI KAISHA TOSHIBA
    Inventors: Kazunori IMOTO, Makoto HIROHATA
  • Publication number: 20130218572
    Abstract: A display device with a voice recognition capability may be used to allow a user to speak voice commands for controlling certain features of the display device. As a means for increasing operational efficiency, the display device may utilize a plurality of voice recognition units where each voice recognition unit may be assigned a specific task.
    Type: Application
    Filed: February 17, 2012
    Publication date: August 22, 2013
    Applicant: LG Electronics Inc.
    Inventors: Eunhyung CHO, Joohyeon OH, Howon SON
  • Publication number: 20130201314
    Abstract: A television (TV) includes a display and a processor controlling the display and receiving signals representing human gestures. The processor is programmed to respond to gestures from a first viewer to control the display. Also, the processor is programmed to respond to gestures from a second viewer to control the display only responsive to a determination that the first viewer has both looked toward the second viewer, and that the first viewer has confirmed, as a separate act from looking toward the second viewer, a desire to transfer control of the TV to the second viewer.
    Type: Application
    Filed: February 7, 2012
    Publication date: August 8, 2013
    Inventors: Steven Friedlander, David Thiele
  • Publication number: 20130197903
    Abstract: An exemplary recording method receives the personal information of a speaker transmitted from a RFID tag through a RFID reader. Then the method receives the voice of the speaker through a microphone. The method next receives the personal information of the speaker and the identifier of the audio input device transmitted from the audio input device, and associates the personal information of the speaker with the received identifier of the audio input device. Then, the method receives the voice and the identifier of the audio input device transmitted from the audio input device. The method further converts the received voice to text. The method determines the personal information corresponding to the identifier of the audio input device received with the voice, and associates the converted text with the determined personal information to generate a record.
    Type: Application
    Filed: August 2, 2012
    Publication date: August 1, 2013
    Applicants: HON HAI PRECISION INDUSTRY CO., LTD., HONG FU JIN PRECISION INDUSTRY (ShenZhen) CO., LTD
    Inventor: WEN-HUI ZHANG
  • Publication number: 20130185076
    Abstract: A motion analyzer includes a motion detection unit that detects motion of a part of a body of a subject, a speaking detection unit that detects speaking of the subject, and a determination unit that determines that the subject has performed predetermined motion when motion of a part of the body is detected by the motion detection unit and speaking of the subject is detected by the speaking detection unit.
    Type: Application
    Filed: August 2, 2012
    Publication date: July 18, 2013
    Applicant: FUJI XEROX CO., LTD.
    Inventors: Hirohito YONEYAMA, Kei SHIMOTANI, Kiyoshi IIDA, Yohei NISHINO, Akira FUJII, Haruo HARADA
  • Publication number: 20130173270
    Abstract: An electronic apparatus and a method of controlling the electronic apparatus are provided. The method includes: receiving a voice command; and if the voice command is a first voice start command, changing a mode of the electronic apparatus to a first voice task mode in which the electronic apparatus is controlled according to further voice input, and if the voice command is a second voice start command, changing the mode of the electronic apparatus to a second voice task mode in which the electronic apparatus is controlled according to the further voice input received via an external apparatus which operates with the electronic apparatus. Therefore, providing efficiency and flexibility in controlling the electronic apparatus by using a microphone of the electronic apparatus or a microphone of the external apparatus.
    Type: Application
    Filed: October 10, 2012
    Publication date: July 4, 2013
    Applicant: SAMSUNG ELECTRONICS CO., LTD.
    Inventor: SAMSUNG ELECTRONICS CO., LTD.
  • Publication number: 20130166302
    Abstract: Aspects of customizing digital signage are addressed. For example, an audio feed may be analyzed for keywords occurring in potential customers' speech. These keywords are then employed to customize display screens of a digital display.
    Type: Application
    Filed: December 22, 2011
    Publication date: June 27, 2013
    Applicant: NCR Corporation
    Inventor: Brennan Eul I. Mercado
  • Publication number: 20130166292
    Abstract: A system for accessing content maintains a set of content selections associated with a first user. The system receives first original content from a first content source associated with a first one of the content selections associated with the first user. The system applies, to the first original content, a first rule (such as a parsing rule) that is specific to the first one of the content selections, to produce first derived content. The system changes the state of at least one component of a human-machine dialogue system (such as a text-to-act engine, a dialogue manager, or an act-to-text engine) based on the first derived content. The system may apply a second rule (such as a dialogue rule) to the first derived content to produce rule output and change the state of the human-machine dialogue system based on the rule output.
    Type: Application
    Filed: December 23, 2011
    Publication date: June 27, 2013
    Inventors: James D. Van Sciver, Christopher Bader, Michael Anthony Aliotti, David Carl Bong
  • Publication number: 20130166286
    Abstract: A voice processing apparatus includes: a phase difference calculation unit which calculates for each frequency band a phase difference between first and second frequency signals obtained by applying a time-frequency transform to sounds captured by two voice input units; a detection unit which detects a frequency band for which the percentage of the phase difference falling within a first range that the phase difference can take for a specific sound source direction, the percentage being taken over a predetermined number of frames, does not satisfy a condition corresponding to a sound coming from the direction; a range setting unit which sets, for the detected frequency band, a second range by expanding the first range; and a signal correction unit which makes the amplitude of the first and second frequency signals larger when the phase difference falls within the second range than when the phase difference falls outside the second range.
    Type: Application
    Filed: October 24, 2012
    Publication date: June 27, 2013
    Inventor: Fujitsu Limited
  • Publication number: 20130159002
    Abstract: A system may include a mobile computing device configured to receive voice input; identify, in the voice input, a navigate command including a sequence indication; determine, based on a sequence control map, a control of a user interface corresponding to the sequence indication; and activate the control of the user interface corresponding to the sequence indication.
    Type: Application
    Filed: December 19, 2011
    Publication date: June 20, 2013
    Applicant: Verizon Patent and Licensing Inc.
    Inventors: Giri Ganapathy, Karthikeyan Gurusamy, Balasundaram Nagarajan, Madhu Bommineni, Mallikarjun Marupally
  • Publication number: 20130159003
    Abstract: Disclosed are a method and an apparatus for providing contents about conversation, which collect voice information from conversation between a user and another person, search contents on the basis of the collected voice information, and provide contents about the conversation between the user and the person. The method of providing contents about conversation includes: a voice information collecting step of collecting voice information from conversation between a user and another person; a keyword creating control step of creating search keywords by using the collected voice information; and a contents providing control step of searching contents by using the created search keywords, and providing the searched contents.
    Type: Application
    Filed: August 23, 2012
    Publication date: June 20, 2013
    Applicant: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Hyun KIM, Kang-Woo LEE, Young Ho SUH, Hyoung Sun KIM
  • Publication number: 20130141516
    Abstract: Systems and methods are described that provide both a network-based, e.g., cloud-based, and apparatus based in-call command control for dynamically performing user commands such as voice commands or gesture commands. The result allows for hands free operation of a calling system including one where the user is located separate from a base station. The systems and methods provide for recognizing both voice and gestural commands. The systems and methods further provide for dynamically filtering commands so that commands are not perceived by those participating in the call. Further, the systems and methods provide for notifying the user, who has given the command, regarding the performance of the requested command.
    Type: Application
    Filed: December 6, 2011
    Publication date: June 6, 2013
    Applicant: AT&T INTELLECTUAL PROPERTY I, LP
    Inventor: Christopher Baldwin