Modification Of At Least One Characteristic Of Speech Waves (epo) Patents (Class 704/E21.001)
  • Patent number: 9043733
    Abstract: In one example, a method includes receiving an indication of an input gesture detected at a presence-sensitive input device, where the input gesture includes one or more input points and each input point is detected at a respective location of the presence-sensitive input device. The method may also include determining a focal point of the input gesture, and determining a radius length. The method may also include determining a shape centered at the focal point and having a size determined based on the radius length. The method may also include responding to a change in a geometric property of the shape by scaling information included in a graphical user interface, where the scaling of the information being centered at the focal point.
    Type: Grant
    Filed: March 15, 2013
    Date of Patent: May 26, 2015
    Assignee: Google Inc.
    Inventors: Winson Chung, Adam William Powell, Svetoslav Ganov, Michael Adam Cohen
  • Patent number: 8994522
    Abstract: The described method and system provide for HMI steering for a telematics-equipped vehicle based on likelihood to exceed eye glance guidelines. By determining whether a task is likely to cause the user to exceed eye glance guidelines, alternative HMI processes may be presented to a user to reduce ASGT and EORT and increase compliance with eye glance guidelines. By allowing a user to navigate through long lists of items through vocal input, T9 text input, or heuristic processing rather than through conventional presentation of the full list, a user is much more likely to comply with the eye glance guidelines. This invention is particularly useful in contexts where users may be searching for one item out of a plurality of potential items, for example, within the context of hands-free calling contacts, playing back audio files, or finding points of interest during GPS navigation.
    Type: Grant
    Filed: May 26, 2011
    Date of Patent: March 31, 2015
    Assignees: General Motors LLC, GM Global Technology Operations LLC
    Inventors: Steven C. Tengler, Bijaya Aryal, Scott P. Geisler, Michael A. Wuergler
  • Patent number: 8749405
    Abstract: An object of the invention is also a navigation system having an input device for the input of an input scale value, having a display device for displaying road map information according to a selected display scale value and having a processor device, wherein the number of enterable input scale values is larger than the number of the selectable display scale values.
    Type: Grant
    Filed: March 9, 2012
    Date of Patent: June 10, 2014
    Assignee: Bayerische Motoren Werke Aktiengesellschaft
    Inventors: Karsten Knebel, Liza Hassel, Frank Wolf
  • Publication number: 20140129231
    Abstract: A computer program product comprises computer usable program code for receiving data describing a proposed electronic transaction between first and second communications devices. Additional computer usable program code is provided for generating a first audio signal by sound detected by a first microphone of the first communications device, and for generating a second audio signal by sound detected by a second microphone that is part of the second communications device. Still further computer usable program code provides for authenticating that the first communications device and the second communications device are in the same proximity in response to determining that the first and second audio signals were produced by the same sound event, and for completing the proposed electronic transaction between the first and second communications device in response to authenticating that the first and second communications devices are in close proximity.
    Type: Application
    Filed: November 2, 2012
    Publication date: May 8, 2014
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Dean F. Herring, Ethan G. Holder, Brad M. Johnson, III, Adrian X. Rodriguez, Jeffrey J. Smith
  • Publication number: 20140122085
    Abstract: Embodiments of the present general inventive concept provide a voice controlled vibration data analyzer system, including a vibration sensor to detect vibration data from a machine-under-test, a data acquisition unit to receive the vibration data from the vibration sensor, and a control unit having a user interface to receive manual and audio input from a user, and to communicate information relating to the machine-under-test, the control unit executing commands in response to the manual or audio input to control the data acquisition unit and/or user interface to output an audio or visual message relating to a navigation path of multiple machines to be tested, to collect and process the vibration data, and to receive manual or audio physical observations from the user to characterize collected vibration data.
    Type: Application
    Filed: October 26, 2012
    Publication date: May 1, 2014
    Applicant: Azima Holdings, Inc.
    Inventors: Kenneth Ralph Piety, K. C. Dahl
  • Publication number: 20140114664
    Abstract: Embodiments of methods and systems for dominant speaker identification in video conferencing are described. In one embodiment, the computer-implemented method includes identifying one or more dominant speakers in a video conference. The method may also include generating a list of the one or more dominant speakers. Additionally, the method may include communicating the list of one or more dominant speakers to clients in a video conferencing system. In a further embodiment, the method includes communicating the list of the one or more dominant speakers to a client in response to the client joining the video conference.
    Type: Application
    Filed: October 20, 2012
    Publication date: April 24, 2014
    Applicant: MICROSOFT CORPORATION
    Inventors: Humayun M. Khan, Jiannan Zheng, Timothy M. Moore
  • Publication number: 20140098233
    Abstract: An access control reader enhances audio data captured by a beamforming microphone array. The access control reader determines a direction to a user and then utilizes beamforming in the direction of the user to enhance the user's voice. The user's enhanced voice is then transmitted to security personnel or a control system to validate the user's identity, in one example.
    Type: Application
    Filed: October 5, 2012
    Publication date: April 10, 2014
    Applicant: SENSORMATIC ELECTRONICS, LLC
    Inventors: Walter A. Martin, Martin J. Donaghy
  • Publication number: 20140100853
    Abstract: An interactive voice response system, comprising: a processor configured to control the output of voice prompts for transmission to a user; an alphanumeric string generator controllable by the processor to generate a random or pseudo-random alphanumeric string for outputting by the processor to a user in natural language form; an input module for receiving a user response and configured to recognize alphanumeric characters in the user response and to output a recognized string of one or more alphanumeric characters recognized in the user response; and a validation module.
    Type: Application
    Filed: October 5, 2012
    Publication date: April 10, 2014
    Applicant: TOUCH NETWORKS PTY LTD
    Inventor: Jason Andrew Van
  • Publication number: 20140095153
    Abstract: Methods and apparatus to provide speech privacy are disclosed. An example method includes forming a sampling block based on a first received audio sample, the sampling block representing speech of a user, creating, with a processor, a mask based on the sampling block, the mask to reduce the intelligibility of the speech of the user, wherein the mask is created by converting the sampling block from a time domain to a frequency domain to form a frequency domain sampling block, identifying a first peak within the frequency domain sampling block, demodulating the frequency domain sampling block at the first peak to form a first envelope of the sampling block, distorting the first envelope to form a first distorted envelope, and emitting an acoustic representation of the mask via a speaker.
    Type: Application
    Filed: September 28, 2012
    Publication date: April 3, 2014
    Inventor: Rafael de la Guardia Gonzales
  • Publication number: 20140095166
    Abstract: In a method for deep tagging a recording, a computer records audio comprising speech from one or more people. The computer detects a non-speech sound within the audio. The computer determines that the non-speech sound corresponds to a type of sound, and in response, associates a descriptive term with a time of occurrence of the non-speech sound within the recorded audio to form a searchable tag. The computer stores the searchable tag as metadata of the recorded audio.
    Type: Application
    Filed: September 28, 2012
    Publication date: April 3, 2014
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Denise A. Bell, Lisa Seacat DeLuca, Jana H. Jenkins, Jeffrey A. Kusnitz
  • Publication number: 20140081643
    Abstract: Systems, methods, and non-transitory computer-readable storage media for determining expertise through speech analytics. The system associates speakers with respective segments of an audio conversation to yield associated speaker segments. The system also identifies a number of times a speaker has spoken about a topic in the audio conversation by searching the associated speaker segments for a term associated with the topic. The system then ranks the speaker as an expert in the topic when the number of times the speaker has spoken about the topic in the audio conversation exceeds a threshold. The audio conversation can include a compilation of a plurality of audio conversations. Moreover, the system can tag the associated speaker segments having the term with keyword tags and match a respective segment from the associated speaker segments with the speaker, the respective segment having a keyword tag.
    Type: Application
    Filed: September 14, 2012
    Publication date: March 20, 2014
    Applicant: Avaya Inc.
    Inventors: Ajita JOHN, Michael J. SAMMON, Reinhard KLEMM, Doree Duncan SELIGMANN
  • Publication number: 20140074481
    Abstract: A method is disclosed for identifying a spoken command by detecting intervals of voiced and unvoiced sound, and then comparing the order of voiced and unvoiced sounds to a set of templates. Each template represents one of the predetermined acceptable commands of the application, and is associated with a predetermined action. When the order of voiced and unvoiced intervals in the spoken command matches the order in one of the templates, the associated action is thus selected. Silent intervals in the command may also be included for enhanced recognition. Efficient protocols are disclosed for discriminating voiced and unvoiced sounds, and for detecting the beginning and ending of each sound interval in the command, and for comparing the command sequence to the templates. In a sparse-command application, this method provides fast and robust recognition, and can be implemented with low-cost hardware and extremely minimal software.
    Type: Application
    Filed: September 12, 2012
    Publication date: March 13, 2014
    Inventor: David Edward Newman
  • Publication number: 20140067403
    Abstract: A method of managing speech interfaces to computer-based services includes beginning a first speech session that is carried out in a vehicle over a short-range wireless connection between a vehicle occupant and a mobile device; detecting an initiation of a second speech session while the first speech session is being carried out; determining an assigned priority level of the first speech session relative to an assigned priority level of the second speech session; and when the assigned priority level of the second speech session has a higher priority than the assigned priority level of the first speech session, carrying out a session-appropriate action on the first speech session.
    Type: Application
    Filed: September 6, 2012
    Publication date: March 6, 2014
    Applicant: GM GLOBAL TECHNOLOGY OPERATIONS LLC
    Inventors: Denis R. Burke, Danilo Gurovich, Daniel E. Rudman, Keith A. Fry, Shane M. McCutchen, Marco T. Carnevale, Mukesh Gupta
  • Publication number: 20140067404
    Abstract: A system and method for selectively applying Intensity Stereo coding to an audio signal is described. The system and method make decisions on whether to apply Intensity Stereo coding to each scale factor band of the audio signal based on (1) the number of bits necessary to encode each scale factor band using Intensity Stereo coding, (2) spatial distortions generated by using Intensity Stereo coding with each scale factor band, and (3) switching distortions for each scale factor band resulting from switching Intensity Stereo coding on or off in relation to a previous scale factor band.
    Type: Application
    Filed: September 4, 2012
    Publication date: March 6, 2014
    Applicant: Apple Inc.
    Inventor: Frank M. Baumgarte
  • Publication number: 20140052438
    Abstract: In a computer system that permits multiple audio capture applications to get an audio capture feed concurrently, an audio manager manages audio capture and/or audio playback in reaction to trigger events. For example, a trigger event indicates an application has started, stopped or otherwise changed a communication stream, or indicates an application has gained, lost or otherwise changed focus or visibility in a user interface, or indicates a user change. In response to a trigger event, the audio manager applies a set of rules to determine which audio capture application is allowed to get an audio capture feed. Based on the decisions, the audio manager manages the audio capture feed for the applications. The audio manager also sends a notification to each of the audio capture applications that has registered for notifications, so as to indicate whether the application is allowed to get the audio capture feed.
    Type: Application
    Filed: August 20, 2012
    Publication date: February 20, 2014
    Applicant: Microsoft Corporation
    Inventors: Frank Yerrace, Kishore Kotteri, Ryan Beberwyck, Gerrit Swaneveld, John Bregar, Rian Chung
  • Publication number: 20140052450
    Abstract: Methods and apparatus for providing a search interface for an electronic device including a tuner configured to tune the electronic device to receive scheduled programming content. A search query is received and one or more data sources including information about media content are searched based, at least in part, on the search query. The results of the search are presented on a user interface using a time-based axis and a time-independent axis.
    Type: Application
    Filed: August 16, 2012
    Publication date: February 20, 2014
    Applicant: Nuance Communications, Inc.
    Inventors: Yuen-Keen CHEONG, Steven A. HATCH, Hoi L. YOUNG, Tapio I. KOIVUNIEMI
  • Publication number: 20140052451
    Abstract: Methods and apparatus for providing a search interface for an electronic device including a tuner configured to tune the electronic device to receive scheduled programming content. A search query is received and one or more data sources including information about media content are searched based, at least in part, on the search query. The results of the search are presented on a user interface using a time-based axis and a time-independent axis.
    Type: Application
    Filed: August 16, 2012
    Publication date: February 20, 2014
    Applicant: Nuance Communications, Inc.
    Inventors: Yuen-Keen Cheong, Steven A. Hatch, Hoi L. Young, Tapio I. Koivuniemi
  • Publication number: 20140052452
    Abstract: Methods and apparatus for providing a search interface for an electronic device including a tuner configured to tune the electronic device to receive scheduled programming content. A search query is received and one or more data sources including information about media content are searched based, at least in part, on the search query. The results of the search are presented on a user interface using a time-based axis and a time-independent axis.
    Type: Application
    Filed: August 16, 2012
    Publication date: February 20, 2014
    Applicant: Nuance Communications, Inc.
    Inventors: Tapio I. Koivuniemi, Tuomas A. Tuononen, Jarkko Koivikko, Teijo J. Kinnunen
  • Publication number: 20140052453
    Abstract: Methods and apparatus for searching for content to display on a digitally-tunable electronic device configured to display scheduled programming content. The method comprises receiving a search query from a user, and determining, based on the search query, an action the user wants to perform. The method further comprises determining one or more data sources to search based, at least in part, on the action the user wants to perform, and searching based, at least in part, on the search query, the one or more data sources for the content to display on the electronic device.
    Type: Application
    Filed: August 16, 2012
    Publication date: February 20, 2014
    Inventors: Tapio I. Koivuniemi, Tuomas A. Tuononen, Jarkko Koivikko, Teijo J. Kinnunen
  • Publication number: 20140039898
    Abstract: Methods and apparatus for voice-enabling a web application, wherein the web application includes one or more web pages rendered by a web browser on a computer. At least one information source external to the web application is queried to determine whether information describing a set of one or more supported voice interactions for the web application is available, and in response to determining that the information is available, the information is retrieved from the at least one information source. Voice input for the web application is then enabled based on the retrieved information.
    Type: Application
    Filed: August 2, 2012
    Publication date: February 6, 2014
    Applicant: Nuance Communications, Inc.
    Inventors: David E. Reich, Christopher Hardy
  • Publication number: 20140039882
    Abstract: The instant application includes computationally-implemented systems and methods that include managing adaptation data, wherein the adaptation data is correlated to at least one aspect of speech of a particular party, facilitating transmission of the adaptation data to a target device, wherein the adaptation data is configured to be applied to the target device to assist in execution of a speech-facilitated transaction, facilitating reception of adaptation result data that is based on at least one aspect of the speech-facilitated transaction between the particular party and the target device, determining whether to modify the adaptation data at least partly based on the adaptation result data, and facilitating transmission of at least a portion of modified adaptation data to a receiving device. In addition to the foregoing, other aspects are described in the claims, drawings, and text.
    Type: Application
    Filed: August 1, 2012
    Publication date: February 6, 2014
    Inventors: Royce A. Levien, Richard T. Lord, Robert W. Lord, Mark A. Malamud
  • Publication number: 20140039892
    Abstract: In one embodiment, a human interactive proof portal 140 may use a biometric input to determine whether a user is a standard user or a malicious actor. The human interactive proof portal 140 may receive an access request 302 for an online data service 122 from a user device 110. The human interactive proof portal 140 may send a proof challenge 304 to the user device 110 for presentation to a user. The human interactive proof portal 140 may receive from the user device 110 a proof response 306 having a biometric metadata description 430 based on a biometric input from the user.
    Type: Application
    Filed: August 2, 2012
    Publication date: February 6, 2014
    Applicant: Microsoft Corporation
    Inventors: Chad Mills, Robert Sim, Scott Laufer, Sung Chung
  • Publication number: 20140029701
    Abstract: Systems, methods, and devices synchronize data streams by hashing received data frames to generate a sequence of hash values, comparing the generated hash value sequence to a hash value sequence received in a control stream, and processing data frames when the hash value sequences match. A source device and multiple receiver devices may synchronize audio data encoded in data frames, applying a hash function to each data frame to generate a first sequence of hash values, transmitting the data frames on a first channel and the first sequence of hash values on a control channel, receiving the data frames and the first sequence of hash values in the receiver devices, applying the hash algorithm to received data frames to generate a second sequence of hash values, comparing the first and second sequences of hash values, and processing data frames when the first and second sequences of hash values match.
    Type: Application
    Filed: July 29, 2012
    Publication date: January 30, 2014
    Inventors: Adam E. NEWHAM, Joel Benjamin LINSKY, Rohit SAUHTA, Brian F. MILLER, Kevin Wayne BARTIG
  • Publication number: 20140019139
    Abstract: A blood glucose meter with a simplified programmable voice function, including: a microprocessor; a memory that is both programmable and re-programmable coupled to the microprocessor; and an audio output device coupled to the microprocessor and the memory; wherein a language algorithm and a plurality of language components specific to a language selected by a user are disposed within the memory; and wherein the language algorithm and the plurality of language components are utilized to provide an audio output through the audio output device in the language selected by the user. The language algorithm is operable for determining which language components are utilized to provide the audio output and in what order based on the language selected by the user. Optionally, the audio output is generated by the microprocessor and the memory using a pulse-width modulation scheme and/or the like.
    Type: Application
    Filed: July 12, 2012
    Publication date: January 16, 2014
    Applicant: PRODIGY DIABETES CARE, LLC
    Inventors: Ramzi ABULHAJ, Moo Nam KO, William BAXTER, Amr Yehia Mohamed SHEHAB
  • Publication number: 20140012586
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for determining hotword suitability. In one aspect, a method includes receiving speech data that encodes a candidate hotword spoken by a user, evaluating the speech data or a transcription of the candidate hotword, using one or more predetermined criteria, generating a hotword suitability score for the candidate hotword based on evaluating the speech data or a transcription of the candidate hotword, using one or more predetermined criteria, and providing a representation of the hotword suitability score for display to the user.
    Type: Application
    Filed: August 6, 2012
    Publication date: January 9, 2014
    Applicant: GOOGLE INC.
    Inventors: Andrew E. Rubin, Johan Schalkwyk, Maria Carolina Parada San Martin
  • Publication number: 20130346084
    Abstract: Technologies are described herein for enhancing a user presence status determination. Visual data may be received from a depth camera configured to be arranged within a three-dimensional space. A current user presence status of a user in the three-dimensional space may be determined based on the visual data. A previous user presence status of the user may be transformed to the current user presence status, responsive to determining the current user presence status of the user.
    Type: Application
    Filed: June 22, 2012
    Publication date: December 26, 2013
    Applicant: MICROSOFT CORPORATION
    Inventors: Anne Marie Renée Archambault, Jeffrey Scott Berg, Xiping Zuo, Abhishek Agrawal
  • Publication number: 20130346085
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, and including hardware devices performing a mouth click sound based human-device interaction. In one aspect, receiving at least one mouth click sound signals from a human user, by an acoustic-to-electric sensor of a computing device, and processing the received signals. The received mouth click sound signals may be accompanied by other mouth click sound signals, and other interaction signals.
    Type: Application
    Filed: June 23, 2012
    Publication date: December 26, 2013
    Inventor: Zoltan Stekkelpak
  • Publication number: 20130339007
    Abstract: Embodiments herein include receiving a request to modify an audio characteristic associated with a first user for a voice communication system. One or more suggested modified audio characteristics may be provided for the first user, based on, at least in part, one or more audio preferences established by another user. An input of one or more modified audio characteristics may be received for the first user for the voice communication system. A user-specific audio preference may be associated with the first user for voice communications on the voice communication system, the user-specific audio preference including the one or more modified audio characteristics.
    Type: Application
    Filed: June 18, 2012
    Publication date: December 19, 2013
    Applicant: International Business Machines Corporation
    Inventors: Ruthie D. Lyle, Patrick Joseph O'Sullivan, Lin Sun
  • Publication number: 20130332172
    Abstract: An accessory is configured to receive a request. The accessory transmits information associated with the request to a portable device. An automated assistant application executed by the portable device can interpret the request and provide a report. The portable device can transmit the report to the accessory. The report may include one or more results determined by the automated assistant.
    Type: Application
    Filed: October 1, 2012
    Publication date: December 12, 2013
    Applicant: Apple Inc.
    Inventors: Jude A. Prakash, Shailesh Rathi, Daniel De Rocha Rosario, Sylvain R.Y. Louboutin
  • Publication number: 20130325481
    Abstract: A method of providing navigation on an electronic device when the display screen is locked. The method receives a verbal request to start navigation while the display is locked. The method identifies a route from a current location to a destination based on the received verbal request. While the display screen is locked, the method provides navigational directions on the electronic device from the current location of the electronic device to the destination. Some embodiments provide a method for processing a verbal search request. The method receives a navigation-related verbal search request and prepares a sequential list of the search results based on the received request. The method then provides audible information to present a search result from the sequential list. The method presents the search results in a batch form until the user selects a search result, the user terminates the search, or the search items are exhausted.
    Type: Application
    Filed: September 30, 2012
    Publication date: December 5, 2013
    Applicant: APPLE INC.
    Inventors: Marcel van Os, Sarah G. Barbour, Brady A. Law, Bradford A. Moore
  • Publication number: 20130321390
    Abstract: A system and method are disclosed for augmenting a reading experience in a mixed reality environment. In response to predefined verbal or physical gestures, the mixed reality system is able to answer a user's questions or provide additional information relating to what the user is reading. Responses may be displayed to the user on virtual display slates in a border or around the reading material without obscuring text or interfering with the user's reading experience.
    Type: Application
    Filed: May 31, 2012
    Publication date: December 5, 2013
    Inventors: Stephen G. Latta, Ryan L. Hastings, Cameron G. Brown, Aaron Krauss, Daniel J. McCulloch, Ben J. Sugden
  • Publication number: 20130325462
    Abstract: A system and method for assigning one or more tags to an image file. In one aspect, a server computer receives an image file captured by a client device. In one embodiment, the image file includes an audio component embedded therein by the client device, where the audio component was spoken by a user of the client device as a tag of the image file. The server computer determines metadata associated with the image file and identifies a dictionary of potential textual tags from the metadata. The server computer determines a textual tag from the audio component and from the dictionary of potential textual tags. The server computer then associates the textual tag with the image file as additional metadata.
    Type: Application
    Filed: May 31, 2012
    Publication date: December 5, 2013
    Applicant: Yahoo! inc.
    Inventors: Oren Somekh, Nadav Golbandi, Liran Katzir, Ronny Lempel, Yoelle Maarek
  • Publication number: 20130317828
    Abstract: The effectiveness of targeted content delivery at a multi-user interface can be directly linked to a proper targeting of users. A way of improving targeted content delivery at a multi-user interface can be to determine which users should be targeted based on one or more criteria. The present technology provides various methodologies for selecting one or more users associated with a multi-user interface to receive targeted content. Such users can be selected based on criteria associated with a ranking or priority of the users, criteria associated with an analysis of their interactions with the multi-user interface, criteria based on their most common characteristics, or any combination thereof. The user characteristics associated with such identified used can then be utilized to determine which content should be delivered to the multi-user interface.
    Type: Application
    Filed: May 25, 2012
    Publication date: November 28, 2013
    Applicant: Apple Inc.
    Inventors: Michael Froimowitz Greenzeiger, Mehul K. Sanghavi, Ravindra Phulari
  • Publication number: 20130304479
    Abstract: Methods and systems for determining intent in voice and gesture interfaces are described. An example method includes determining that a gaze direction is in a direction of a gaze target, and determining whether a predetermined time period has elapsed while the gaze direction is in the direction of the gaze target. The method may also include providing an indication that the predetermined time period has elapsed when the predetermined time period has elapsed. According to the method, a voice or gesture command that is received after the predetermined time period has elapsed may be determined to be an input for a computing system. Additional example systems and methods are described herein.
    Type: Application
    Filed: May 8, 2012
    Publication date: November 14, 2013
    Applicant: GOOGLE INC.
    Inventors: Eric Teller, Daniel Aminzade
  • Publication number: 20130297319
    Abstract: A mobile device having at least one microphone sensor and a method for controlling the same are disclosed. The method includes receiving at least two audio signals through the at least one microphone sensor within a predetermined time period, recognizing input directions and voice command from the at least two audio signals sequentially, determining whether the recognized input directions and voice command match to preset input directions and preset voice command mapped to the preset directions, sequentially for the at least two received audio signals, and executing a preset control command, if the recognized input directions and voice command match to the preset input directions and voice command.
    Type: Application
    Filed: July 9, 2012
    Publication date: November 7, 2013
    Inventor: Yongsin KIM
  • Publication number: 20130297296
    Abstract: Methods and apparatus for signal processing are disclosed. Source separation can be performed to extract source signals from mixtures of source signals by way of independent component analysis. Source direction information is utilized in the separation process, and independent component analysis techniques described herein use multivariate probability density functions to preserve the alignment of frequency bins in the source separation process. It is emphasized that this abstract is provided to comply with the rules requiring an abstract that will allow a searcher or other reader to quickly ascertain the subject matter of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims.
    Type: Application
    Filed: May 4, 2012
    Publication date: November 7, 2013
    Applicant: Sony Computer Entertainment Inc.
    Inventors: Jaekwon Yoo, Ruxin Chen
  • Publication number: 20130297320
    Abstract: An additive three-dimensional fabrication system includes voice control for user interaction. This voice-controlled interface can enable a variety of voice-controlled functions and operations, while supporting interactions specific to consumer-oriented fabrication processes.
    Type: Application
    Filed: July 23, 2012
    Publication date: November 7, 2013
    Inventors: Anthony James Buser, Nathaniel B. Pettis
  • Publication number: 20130297298
    Abstract: Methods and apparatus for signal processing are disclosed. Source separation can be performed to extract source signals from mixtures of source signals by way of independent component analysis. Source separation described herein involves mixed multivariate probability density functions that are mixtures of component density functions having different parameters corresponding to frequency components of different sources, different time segments, or some combination thereof.
    Type: Application
    Filed: May 4, 2012
    Publication date: November 7, 2013
    Applicant: Sony Computer Entertainment Inc.
    Inventors: Jaekwon Yoo, Ruxin Chen
  • Publication number: 20130290000
    Abstract: A method is disclosed for controlling a voice-activated device by interpreting a spoken command as a series of voiced and non-voiced intervals. A responsive action is then performed according to the number of voiced intervals in the command. The method is well-suited to applications having a small number of specific voice-activated response functions. Applications using the inventive method offer numerous advantages over traditional speech recognition systems including speaker universality, language independence, no training or calibration needed, implementation with simple microcontrollers, and extremely low cost. For time-critical applications such as pulsers and measurement devices, where fast reaction is crucial to catch a transient event, the method provides near-instantaneous command response, yet versatile voice control.
    Type: Application
    Filed: April 30, 2012
    Publication date: October 31, 2013
    Inventor: David Edward Newman
  • Publication number: 20130282380
    Abstract: Current human-to-machine interfaces enable users to interact with a company's database and enter into a series of transactions (e.g., purchasing products/services and paying bills). Each transaction may require several operations or stages requiring user input or interaction. Some systems enable a user to enter a voice input parameter providing multiple operations of instruction (e.g., single natural language command). However, users of such a system do not know what types of commands the system is capable of accepting. Embodiments of the present invention facilitate communications for user transactions by determining a user's goal transaction and presenting a visual representation of a voice input parameter for the goal transaction. The use of visual representations notifies the user of the system's capability of accepting single natural language commands and the types of commands the system is capable of accepting, thereby enabling a user to complete a transaction in a shorter period of time.
    Type: Application
    Filed: April 20, 2012
    Publication date: October 24, 2013
    Applicant: Nuance Communications, Inc.
    Inventors: David Andrew Mauro, Simona Gandrabur
  • Publication number: 20130282381
    Abstract: Generally, human-to-machine interfaces are configured to accept speech input from a user. However, such interfaces, e.g., web browsers, must be configured to enable acceptance of speech input from the user. Some interfaces, such as mobile browsers, have less configuration adaptability and are not able to be configured to accept speech input from a user. Embodiments of the present invention speech-enable human-to-machine interfaces by loading content of the human-to-machine interface and adding logic configured to enable speech interaction with the content to the interface. The embodiment then activates speech interaction with the content via the logic for the user. Thus, embodiments of the present invention enable speech interaction with interfaces that are not configured to be adapted to allow speech interaction and are able to enable the speech interaction in a seamless manner.
    Type: Application
    Filed: April 20, 2012
    Publication date: October 24, 2013
    Applicant: Nuance Communications, Inc.
    Inventors: David Andrew Mauro, Henri Bouvier
  • Publication number: 20130282371
    Abstract: A method is disclosed herein for recognizing a repeated utterance in a mobile computing device via a processor. A first utterance is detected being spoken into a first mobile computing device. Likewise, a second utterance is detected being spoken into a second mobile computing device within a predetermined time period. The second utterance substantially matches the first spoken utterance and the first and second mobile computing devices are communicatively coupled to each other. The processor enables capturing, at least temporarily, a matching utterance for performing a subsequent processing function. The performed subsequent processing function is based on a type of captured utterance.
    Type: Application
    Filed: April 20, 2012
    Publication date: October 24, 2013
    Applicant: Motorola Mobility, Inc.
    Inventors: Rachid M Alameh, Jiri Slaby, Hisashi D. Watanabe
  • Publication number: 20130262127
    Abstract: A content processing service may analyze an item of original content and identify several objects, attributes of those objects, and relationships between those objects present in the item of original content. The content processing service may also analyze a source graph, such as a social graph or supplemental graph, and identify several objects, attributes of those objects, and relationships between objects present in the source graph. The content processing service may customize the item of original content by selecting an original object and selecting a source graph object. One or more of the attributes or relationships of the selected original object in the item of original content may be replaced by one or more of the attributes or relationships of the selected source graph object. To customize items of audio content, audio content associated with the source graph object may replace audio content associated with the target graph object.
    Type: Application
    Filed: March 29, 2012
    Publication date: October 3, 2013
    Inventors: Douglas S. Goldstein, Ajay Arora, Douglas Hwang, Guy A. Story, JR., Shirley C. Yang
  • Publication number: 20130262124
    Abstract: System and method to search audio data, including: receiving audio data representing speech; receiving a search query related to the audio data; compiling, by use of a processor, the search query into a hierarchy of scored speech recognition sub-searches; searching, by use of a processor, the audio data for speech identified by one or more of the sub-searches to produce hits; and combining, by use of a processor, the hits by use of at least one combination function to provide a composite search score of the audio data. The combination function may include an at-least-M-of-N function that produces a high score when at least M of N function inputs exceed a predetermined threshold value. The composite search score employ a soft time window such as a spline function.
    Type: Application
    Filed: September 27, 2012
    Publication date: October 3, 2013
    Applicant: AURIX LIMITED
    Inventor: Keith Michael Ponting
  • Publication number: 20130238341
    Abstract: An electronic device includes a music play module that plays music and a voice recorder that records ambient voice around the electronic device. The electronic device further includes a music control module that identifies voice characteristics of the ambient voice, and controls the music play module to pause the playing of the music when the voice characteristics of the ambient voice match pre-configured voice reference information.
    Type: Application
    Filed: April 27, 2012
    Publication date: September 12, 2013
    Applicants: HON HAI PRECISION INDUSTRY CO., LTD., FU TAI HUA INDUSTRY (SHENZHEN) CO., LTD.
    Inventor: QIANG YOU
  • Publication number: 20130231930
    Abstract: A computer implemented method and apparatus for automatically filtering an audio input to make a filtered recording comprising: identifying words used in an audio input, determining whether each identified word is contained in a dictionary of banned words, and creating a filtered recording as an audio output, wherein each word identified in the audio input that is found in the dictionary of banned words, is automatically deleted or replaced in the audio output used to make the filtered recording.
    Type: Application
    Filed: March 1, 2012
    Publication date: September 5, 2013
    Applicant: Adobe Systems Inc.
    Inventor: Antonio Sanso
  • Publication number: 20130231927
    Abstract: Implementations of systems, method and devices described herein enable enhancing the intelligibility of a target voice signal included in a noisy audible signal received by a hearing aid device or the like. In particular, in some implementations, systems, methods and devices are operable to generate a machine readable formant based codebook. In some implementations, the method includes determining whether or not a candidate codebook tuple includes a sufficient amount of new information to warrant either adding the candidate codebook tuple to the codebook or using at least a portion of the candidate codebook tuple to update an existing codebook tuple. Additionally and/or alternatively, in some implementations systems, methods and devices are operable to reconstruct a target voice signal by detecting formants in an audible signal, using the detected formants to select codebook tuples, and using the formant information in the selected codebook tuples to reconstruct the target voice signal.
    Type: Application
    Filed: August 20, 2012
    Publication date: September 5, 2013
    Inventors: PIERRE ZAKARAUSKAS, ALEXANDER ESCOTT, CLARENCE S.H. CHU, SHAWN E. STEVENSON
  • Publication number: 20130226589
    Abstract: A sound-activated control system includes an audio receiver and a command discriminator. The receiver is configured to receive an audio waveform and to produce a digital audio waveform therefrom. The command discriminator is configured to detect a temporally and/or spectrally compact nonphonetic audio command within the digital audio waveform and to control a voice-activated system an action in response to the nonphonetic command.
    Type: Application
    Filed: February 29, 2012
    Publication date: August 29, 2013
    Applicant: NVIDIA Corporation
    Inventor: Henry P. Largey
  • Publication number: 20130225999
    Abstract: The embodiments of the ultrasound imaging diagnostic apparatus include at least one non-touch input device for receiving a predetermined gesture as an input command. An optional sequence of predetermined gestures is inputted as an operational command and or data to the embodiments of the ultrasound imaging diagnostic apparatus. A gesture is optionally combined with other conventional input modes through devices such as a microphone, a mouse, a keyboard, a button, a panel switch, a touch command screen, a foot switch, a trackball, and the like.
    Type: Application
    Filed: February 29, 2012
    Publication date: August 29, 2013
    Applicants: TOSHIBA MEDICAL SYSTEMS CORPORATION, KABUSHIKI KAISHA TOSHIBA
    Inventors: Zoran BANJANIN, Raymond F. WOODS
  • Publication number: 20130226588
    Abstract: A method is provided for a simulated conversation by a pre-recorded audio navigator, with particular application to informational and entertainment settings. A monitor may utilize a navigation interface to select pre-recorded responses in the voice of a character represented by a performer. The pre-recorded responses may then be queued and sent to a speaker proximate to the performer. By careful organization of an audio database including audio buckets and script-based navigation with shifts for tailoring to specific guest user profiles and environmental contexts, a convincing and dynamic simulated conversation may be carried out while providing the monitor with a user-friendly navigation interface. Thus, highly specialized training is not necessary and flexible scaling to large-scale deployments is readily supported.
    Type: Application
    Filed: February 28, 2012
    Publication date: August 29, 2013
    Inventors: Holger Irmler, Asa K. Kalama, Raymond J. Scanlon, Brent D. Strong, Cory J. Rouse, Renée M. Johnson, Andrew Stone