Modification Of At Least One Characteristic Of Speech Waves (epo) Patents (Class 704/E21.001)
E Subclasses
-
Patent number: 11657408Abstract: Arrangements for synchronously tracking and controlling events across multiple computer systems are provided. In some examples, a user may register with a system and user data may be received. In some arrangements, historical data associated with the user may also be received. Machine learning may be used to analyze the historical data and/or user data and a first recommendation for an item may be generated and transmitted to the user. Upon receiving acceptance of the recommendation, the system may request data from one or more entities. For instance, entity data associated with current inventory, availability of items, layout of locations, and the like, may be received. Based on the received data, a list of items for capture and/or an item capture route may be generated. In some examples, the item capture route may include step-by-step or map-based instructions to capture the items on the list.Type: GrantFiled: January 7, 2020Date of Patent: May 23, 2023Assignee: Bank of America CorporationInventors: Manu Kurian, Matthew E. Carroll
-
Patent number: 11658928Abstract: A virtual content creation method according to an embodiment of the present invention includes, by a server, receiving a model content including at least one of a text, an SMS, a voice-recorded MP3 file, a picture, and a video of a model; by the server, extracting a model feature including at least one of a text feature, a voice feature, an image feature, and a video feature from the model content; and when a user wants to communicate with the model, by the server, being operated based on deep learning or artificial intelligence to allow the user to input a user content to the server, determine a user state by detecting an emotional state of the user from the user content, and transform the model content into the virtual content using the model feature or the user state.Type: GrantFiled: August 7, 2020Date of Patent: May 23, 2023Inventor: Kab Cheon Choe
-
Patent number: 11580999Abstract: An audio signal encoding method performed by an encoder includes identifying an audio signal of a time domain in units of a block, generating a combined block by combining i) a current original block of the audio signal and ii) a previous original block chronologically adjacent to the current original block, extracting a first residual signal of a frequency domain from the combined block using linear predictive coding of a time domain, overlapping chronologically adjacent first residual signals among first residual signals converted into a time domain, and quantizing a second residual signal of a time domain extracted from the overlapped first residual signal by converting the second residual signal of the time domain into a frequency domain using linear predictive coding of a frequency domain.Type: GrantFiled: May 26, 2021Date of Patent: February 14, 2023Assignee: Electronics and Telecommunications Research InstituteInventors: Seung Kwon Beack, Jongmo Sung, Mi Suk Lee, Tae Jin Lee, Woo-taek Lim, Inseon Jang
-
Patent number: 11540003Abstract: Methods, apparatus, and systems are disclosed for synchronizing streaming media content. An example apparatus includes a storage device, and a processor to execute instructions to identify a first source streaming broadcast media to a first computing device based on an audio fingerprint of audio associated with the broadcast media, identify sources broadcasting the broadcast media streaming to the first computing device, the sources available to a second computing device including the processor, select a second source of the identified sources for streaming the broadcast media to the second computing device, the second source different than the first source, detect termination of the streaming of the broadcast media on the first computing device, the termination corresponding to a termination time of the broadcast media, and automatically start, by using the selected second source, streaming of the broadcast media to the second computing device at the termination time.Type: GrantFiled: March 18, 2021Date of Patent: December 27, 2022Assignee: Gracenote, Inc.Inventors: Suresh Jeyachandran, Roger Tsai, Paul Emmanuel Quinn, Markus K. Cremer
-
Patent number: 11514332Abstract: A method, computer program product, and system for a cognitive dialoguing avatar, the method including identifying a user, a target entity, and a user goal, initiating communication with the target entity, evaluating cognitively a question from a dialog with the target entity, determining cognitively an answer to the question by evaluating stored user information to progress to the user goal, communicating the determined answer to the target entity.Type: GrantFiled: March 26, 2018Date of Patent: November 29, 2022Assignee: International Business Machines CorporationInventors: Adam T. Clark, Nathaniel D. Lee, Daniel J. Strauss
-
Patent number: 11488281Abstract: A multichannel interpolator has an input that receives input data that consists of interleaved channel data from a plurality of data channels. A block random access memory (BRAM) stores data samples from the input data received from the input. Input control logic receives the data samples from the input and places the data samples into the BRAM. Interpolator logic interpolates the data samples to produce output data. The output data is interpolated at an interpolation ratio programmed by a user. The interpolator logic includes a phase generator that calculates a value indicating the interpolation ratio, and a fractional template block that provides a fractional template used to interpolate the data samples to produce the output data, the fraction template block selecting, based on the value calculated by the phase generator. The fractional template is used to interpolate the data samples to produce the output data.Type: GrantFiled: February 8, 2021Date of Patent: November 1, 2022Assignee: Keysight Technologies, Inc.Inventor: Garrett Foltz
-
Patent number: 11487832Abstract: Implementations are described herein for analyzing existing interactive web sites to facilitate automatic engagement with those web sites, e.g., by automated assistants or via other user interfaces, with minimal effort from the hosts of those websites. For example, in various implementations, techniques described herein may be used to abstract, validate, maintain, generalize, extend and/or distribute individual actions and “traces” of actions that are useable to navigate through various interactive websites. Additionally, techniques are described herein for leveraging these actions and/or traces to automate aspects of interaction with a third party website.Type: GrantFiled: May 9, 2019Date of Patent: November 1, 2022Assignee: GOOGLE LLCInventors: Gökhan Bakir, Andre Elisseeff, Torsten Marek, João Paulo Pagaime da Silva, Mathias Carlen, Dana Ritter, Lukasz Suder, Ernest Galbrun, Matthew Stokes, Marcin Nowak-Przygodzki, Mugurel-Ionut Andreica, Marius Dumitran
-
Patent number: 11481185Abstract: In one embodiment, a computer-implemented method for editing navigation of a content item is disclosed. The method may include presenting, via a user interface at a client computing device, time-synchronized text pertaining to the content item; receiving an input of a tag for the time-synchronized text of the content item, wherein the tag corresponds to a performer that performs at least a portion of the content item at a timestamp in the time-synchronized text; storing the tag associated with the portion of the content item at the timestamp in the time-synchronized text of the content item; and responsive to receiving a request to play the content item: playing the content item via a media player presented in the user interface, and concurrently presenting the time-synchronized text and the tag in the user interface, wherein the tag is presented as a graphical user element in the user interface.Type: GrantFiled: June 17, 2021Date of Patent: October 25, 2022Assignee: Musixmatch S.P.A.Inventors: Marco Paglia, Paolo Spazzini, Pierpaolo Di Panfilo, Niche Chathong, Daria Babco
-
Patent number: 11474841Abstract: Methods, apparatus, systems, and computer-readable media are provided for providing context specific schema files that allow an automated assistant to broker human-to-computer dialogs between a user and an application that is separate from the automated assistant. The context specific schema file can provide the automated assistant with sufficient data to be responsive to user queries without necessarily communicating with a remote device, such as a server. Multiple different context specific schema files can be made available to the automated assistant according to a context in which a user is interacting with the automated assistant. In this way, latency otherwise exhibited by the automated assistant can be mitigated by providing the automated assistant with the information needed to respond to a user without continually retrieving the information over a network.Type: GrantFiled: January 23, 2019Date of Patent: October 18, 2022Assignee: GOOGLE LLCInventors: Justin Lewis, Scott Davies
-
Patent number: 11461708Abstract: There is provided a controller that executes: calculating, based on information about behavior of a user, likelihood about future movement of the user; and outputting information about a service for the movement of the user based on the calculated likelihood.Type: GrantFiled: December 23, 2020Date of Patent: October 4, 2022Assignee: TOYOTA JIDOSHA KABUSHIKI KAISHAInventors: Yurika Tanaka, Daiki Kaneichi
-
Patent number: 11444839Abstract: System for optimizing bandwidth during an online meeting comprises a plurality of data processing systems, wherein each of the data processing systems is associated with a user and comprises a processor, and a memory comprising a digital client and a digital client display interface, wherein the processor causes the digital client to publish an audio-visual stream comprising a video component and an audio component from the corresponding data processing system. A first data processing system, among the plurality of data processing systems is configured to receive an instruction to optimize the bandwidth by limiting the number of data processing systems from which an audio-visual stream is to be played in the first digital client display interface. Further, the first data processing system may play, in the first digital client display interface, an audio-visual stream from each of the number of data processing systems as instructed by the first user.Type: GrantFiled: May 27, 2021Date of Patent: September 13, 2022Inventor: Kishore Daggubati
-
Patent number: 11437030Abstract: Selectively performing voice recognition using one device among multiple devices that recognize and execute the voice recognition based on at least one of apparatus information of the multiple devices and a function parsed from a result of the voice recognition. Thereby, only a single preferable device in an environment in which multiple devices exist, which can service the user input via voice recognition, actually responds to the voice input and services the voice input of the user.Type: GrantFiled: October 17, 2018Date of Patent: September 6, 2022Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventor: Chan-hee Choi
-
Patent number: 9043733Abstract: In one example, a method includes receiving an indication of an input gesture detected at a presence-sensitive input device, where the input gesture includes one or more input points and each input point is detected at a respective location of the presence-sensitive input device. The method may also include determining a focal point of the input gesture, and determining a radius length. The method may also include determining a shape centered at the focal point and having a size determined based on the radius length. The method may also include responding to a change in a geometric property of the shape by scaling information included in a graphical user interface, where the scaling of the information being centered at the focal point.Type: GrantFiled: March 15, 2013Date of Patent: May 26, 2015Assignee: Google Inc.Inventors: Winson Chung, Adam William Powell, Svetoslav Ganov, Michael Adam Cohen
-
Patent number: 8994522Abstract: The described method and system provide for HMI steering for a telematics-equipped vehicle based on likelihood to exceed eye glance guidelines. By determining whether a task is likely to cause the user to exceed eye glance guidelines, alternative HMI processes may be presented to a user to reduce ASGT and EORT and increase compliance with eye glance guidelines. By allowing a user to navigate through long lists of items through vocal input, T9 text input, or heuristic processing rather than through conventional presentation of the full list, a user is much more likely to comply with the eye glance guidelines. This invention is particularly useful in contexts where users may be searching for one item out of a plurality of potential items, for example, within the context of hands-free calling contacts, playing back audio files, or finding points of interest during GPS navigation.Type: GrantFiled: May 26, 2011Date of Patent: March 31, 2015Assignees: General Motors LLC, GM Global Technology Operations LLCInventors: Steven C. Tengler, Bijaya Aryal, Scott P. Geisler, Michael A. Wuergler
-
Patent number: 8749405Abstract: An object of the invention is also a navigation system having an input device for the input of an input scale value, having a display device for displaying road map information according to a selected display scale value and having a processor device, wherein the number of enterable input scale values is larger than the number of the selectable display scale values.Type: GrantFiled: March 9, 2012Date of Patent: June 10, 2014Assignee: Bayerische Motoren Werke AktiengesellschaftInventors: Karsten Knebel, Liza Hassel, Frank Wolf
-
Publication number: 20140129231Abstract: A computer program product comprises computer usable program code for receiving data describing a proposed electronic transaction between first and second communications devices. Additional computer usable program code is provided for generating a first audio signal by sound detected by a first microphone of the first communications device, and for generating a second audio signal by sound detected by a second microphone that is part of the second communications device. Still further computer usable program code provides for authenticating that the first communications device and the second communications device are in the same proximity in response to determining that the first and second audio signals were produced by the same sound event, and for completing the proposed electronic transaction between the first and second communications device in response to authenticating that the first and second communications devices are in close proximity.Type: ApplicationFiled: November 2, 2012Publication date: May 8, 2014Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Dean F. Herring, Ethan G. Holder, Brad M. Johnson, III, Adrian X. Rodriguez, Jeffrey J. Smith
-
Publication number: 20140122085Abstract: Embodiments of the present general inventive concept provide a voice controlled vibration data analyzer system, including a vibration sensor to detect vibration data from a machine-under-test, a data acquisition unit to receive the vibration data from the vibration sensor, and a control unit having a user interface to receive manual and audio input from a user, and to communicate information relating to the machine-under-test, the control unit executing commands in response to the manual or audio input to control the data acquisition unit and/or user interface to output an audio or visual message relating to a navigation path of multiple machines to be tested, to collect and process the vibration data, and to receive manual or audio physical observations from the user to characterize collected vibration data.Type: ApplicationFiled: October 26, 2012Publication date: May 1, 2014Applicant: Azima Holdings, Inc.Inventors: Kenneth Ralph Piety, K. C. Dahl
-
Publication number: 20140114664Abstract: Embodiments of methods and systems for dominant speaker identification in video conferencing are described. In one embodiment, the computer-implemented method includes identifying one or more dominant speakers in a video conference. The method may also include generating a list of the one or more dominant speakers. Additionally, the method may include communicating the list of one or more dominant speakers to clients in a video conferencing system. In a further embodiment, the method includes communicating the list of the one or more dominant speakers to a client in response to the client joining the video conference.Type: ApplicationFiled: October 20, 2012Publication date: April 24, 2014Applicant: MICROSOFT CORPORATIONInventors: Humayun M. Khan, Jiannan Zheng, Timothy M. Moore
-
Publication number: 20140100853Abstract: An interactive voice response system, comprising: a processor configured to control the output of voice prompts for transmission to a user; an alphanumeric string generator controllable by the processor to generate a random or pseudo-random alphanumeric string for outputting by the processor to a user in natural language form; an input module for receiving a user response and configured to recognize alphanumeric characters in the user response and to output a recognized string of one or more alphanumeric characters recognized in the user response; and a validation module.Type: ApplicationFiled: October 5, 2012Publication date: April 10, 2014Applicant: TOUCH NETWORKS PTY LTDInventor: Jason Andrew Van
-
Publication number: 20140098233Abstract: An access control reader enhances audio data captured by a beamforming microphone array. The access control reader determines a direction to a user and then utilizes beamforming in the direction of the user to enhance the user's voice. The user's enhanced voice is then transmitted to security personnel or a control system to validate the user's identity, in one example.Type: ApplicationFiled: October 5, 2012Publication date: April 10, 2014Applicant: SENSORMATIC ELECTRONICS, LLCInventors: Walter A. Martin, Martin J. Donaghy
-
Publication number: 20140095166Abstract: In a method for deep tagging a recording, a computer records audio comprising speech from one or more people. The computer detects a non-speech sound within the audio. The computer determines that the non-speech sound corresponds to a type of sound, and in response, associates a descriptive term with a time of occurrence of the non-speech sound within the recorded audio to form a searchable tag. The computer stores the searchable tag as metadata of the recorded audio.Type: ApplicationFiled: September 28, 2012Publication date: April 3, 2014Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Denise A. Bell, Lisa Seacat DeLuca, Jana H. Jenkins, Jeffrey A. Kusnitz
-
Publication number: 20140095153Abstract: Methods and apparatus to provide speech privacy are disclosed. An example method includes forming a sampling block based on a first received audio sample, the sampling block representing speech of a user, creating, with a processor, a mask based on the sampling block, the mask to reduce the intelligibility of the speech of the user, wherein the mask is created by converting the sampling block from a time domain to a frequency domain to form a frequency domain sampling block, identifying a first peak within the frequency domain sampling block, demodulating the frequency domain sampling block at the first peak to form a first envelope of the sampling block, distorting the first envelope to form a first distorted envelope, and emitting an acoustic representation of the mask via a speaker.Type: ApplicationFiled: September 28, 2012Publication date: April 3, 2014Inventor: Rafael de la Guardia Gonzales
-
Publication number: 20140081643Abstract: Systems, methods, and non-transitory computer-readable storage media for determining expertise through speech analytics. The system associates speakers with respective segments of an audio conversation to yield associated speaker segments. The system also identifies a number of times a speaker has spoken about a topic in the audio conversation by searching the associated speaker segments for a term associated with the topic. The system then ranks the speaker as an expert in the topic when the number of times the speaker has spoken about the topic in the audio conversation exceeds a threshold. The audio conversation can include a compilation of a plurality of audio conversations. Moreover, the system can tag the associated speaker segments having the term with keyword tags and match a respective segment from the associated speaker segments with the speaker, the respective segment having a keyword tag.Type: ApplicationFiled: September 14, 2012Publication date: March 20, 2014Applicant: Avaya Inc.Inventors: Ajita JOHN, Michael J. SAMMON, Reinhard KLEMM, Doree Duncan SELIGMANN
-
Publication number: 20140074481Abstract: A method is disclosed for identifying a spoken command by detecting intervals of voiced and unvoiced sound, and then comparing the order of voiced and unvoiced sounds to a set of templates. Each template represents one of the predetermined acceptable commands of the application, and is associated with a predetermined action. When the order of voiced and unvoiced intervals in the spoken command matches the order in one of the templates, the associated action is thus selected. Silent intervals in the command may also be included for enhanced recognition. Efficient protocols are disclosed for discriminating voiced and unvoiced sounds, and for detecting the beginning and ending of each sound interval in the command, and for comparing the command sequence to the templates. In a sparse-command application, this method provides fast and robust recognition, and can be implemented with low-cost hardware and extremely minimal software.Type: ApplicationFiled: September 12, 2012Publication date: March 13, 2014Inventor: David Edward Newman
-
Publication number: 20140067404Abstract: A system and method for selectively applying Intensity Stereo coding to an audio signal is described. The system and method make decisions on whether to apply Intensity Stereo coding to each scale factor band of the audio signal based on (1) the number of bits necessary to encode each scale factor band using Intensity Stereo coding, (2) spatial distortions generated by using Intensity Stereo coding with each scale factor band, and (3) switching distortions for each scale factor band resulting from switching Intensity Stereo coding on or off in relation to a previous scale factor band.Type: ApplicationFiled: September 4, 2012Publication date: March 6, 2014Applicant: Apple Inc.Inventor: Frank M. Baumgarte
-
Publication number: 20140067403Abstract: A method of managing speech interfaces to computer-based services includes beginning a first speech session that is carried out in a vehicle over a short-range wireless connection between a vehicle occupant and a mobile device; detecting an initiation of a second speech session while the first speech session is being carried out; determining an assigned priority level of the first speech session relative to an assigned priority level of the second speech session; and when the assigned priority level of the second speech session has a higher priority than the assigned priority level of the first speech session, carrying out a session-appropriate action on the first speech session.Type: ApplicationFiled: September 6, 2012Publication date: March 6, 2014Applicant: GM GLOBAL TECHNOLOGY OPERATIONS LLCInventors: Denis R. Burke, Danilo Gurovich, Daniel E. Rudman, Keith A. Fry, Shane M. McCutchen, Marco T. Carnevale, Mukesh Gupta
-
Publication number: 20140052450Abstract: Methods and apparatus for providing a search interface for an electronic device including a tuner configured to tune the electronic device to receive scheduled programming content. A search query is received and one or more data sources including information about media content are searched based, at least in part, on the search query. The results of the search are presented on a user interface using a time-based axis and a time-independent axis.Type: ApplicationFiled: August 16, 2012Publication date: February 20, 2014Applicant: Nuance Communications, Inc.Inventors: Yuen-Keen CHEONG, Steven A. HATCH, Hoi L. YOUNG, Tapio I. KOIVUNIEMI
-
Publication number: 20140052453Abstract: Methods and apparatus for searching for content to display on a digitally-tunable electronic device configured to display scheduled programming content. The method comprises receiving a search query from a user, and determining, based on the search query, an action the user wants to perform. The method further comprises determining one or more data sources to search based, at least in part, on the action the user wants to perform, and searching based, at least in part, on the search query, the one or more data sources for the content to display on the electronic device.Type: ApplicationFiled: August 16, 2012Publication date: February 20, 2014Inventors: Tapio I. Koivuniemi, Tuomas A. Tuononen, Jarkko Koivikko, Teijo J. Kinnunen
-
Publication number: 20140052451Abstract: Methods and apparatus for providing a search interface for an electronic device including a tuner configured to tune the electronic device to receive scheduled programming content. A search query is received and one or more data sources including information about media content are searched based, at least in part, on the search query. The results of the search are presented on a user interface using a time-based axis and a time-independent axis.Type: ApplicationFiled: August 16, 2012Publication date: February 20, 2014Applicant: Nuance Communications, Inc.Inventors: Yuen-Keen Cheong, Steven A. Hatch, Hoi L. Young, Tapio I. Koivuniemi
-
Publication number: 20140052452Abstract: Methods and apparatus for providing a search interface for an electronic device including a tuner configured to tune the electronic device to receive scheduled programming content. A search query is received and one or more data sources including information about media content are searched based, at least in part, on the search query. The results of the search are presented on a user interface using a time-based axis and a time-independent axis.Type: ApplicationFiled: August 16, 2012Publication date: February 20, 2014Applicant: Nuance Communications, Inc.Inventors: Tapio I. Koivuniemi, Tuomas A. Tuononen, Jarkko Koivikko, Teijo J. Kinnunen
-
Publication number: 20140052438Abstract: In a computer system that permits multiple audio capture applications to get an audio capture feed concurrently, an audio manager manages audio capture and/or audio playback in reaction to trigger events. For example, a trigger event indicates an application has started, stopped or otherwise changed a communication stream, or indicates an application has gained, lost or otherwise changed focus or visibility in a user interface, or indicates a user change. In response to a trigger event, the audio manager applies a set of rules to determine which audio capture application is allowed to get an audio capture feed. Based on the decisions, the audio manager manages the audio capture feed for the applications. The audio manager also sends a notification to each of the audio capture applications that has registered for notifications, so as to indicate whether the application is allowed to get the audio capture feed.Type: ApplicationFiled: August 20, 2012Publication date: February 20, 2014Applicant: Microsoft CorporationInventors: Frank Yerrace, Kishore Kotteri, Ryan Beberwyck, Gerrit Swaneveld, John Bregar, Rian Chung
-
Publication number: 20140039882Abstract: The instant application includes computationally-implemented systems and methods that include managing adaptation data, wherein the adaptation data is correlated to at least one aspect of speech of a particular party, facilitating transmission of the adaptation data to a target device, wherein the adaptation data is configured to be applied to the target device to assist in execution of a speech-facilitated transaction, facilitating reception of adaptation result data that is based on at least one aspect of the speech-facilitated transaction between the particular party and the target device, determining whether to modify the adaptation data at least partly based on the adaptation result data, and facilitating transmission of at least a portion of modified adaptation data to a receiving device. In addition to the foregoing, other aspects are described in the claims, drawings, and text.Type: ApplicationFiled: August 1, 2012Publication date: February 6, 2014Inventors: Royce A. Levien, Richard T. Lord, Robert W. Lord, Mark A. Malamud
-
Publication number: 20140039892Abstract: In one embodiment, a human interactive proof portal 140 may use a biometric input to determine whether a user is a standard user or a malicious actor. The human interactive proof portal 140 may receive an access request 302 for an online data service 122 from a user device 110. The human interactive proof portal 140 may send a proof challenge 304 to the user device 110 for presentation to a user. The human interactive proof portal 140 may receive from the user device 110 a proof response 306 having a biometric metadata description 430 based on a biometric input from the user.Type: ApplicationFiled: August 2, 2012Publication date: February 6, 2014Applicant: Microsoft CorporationInventors: Chad Mills, Robert Sim, Scott Laufer, Sung Chung
-
Publication number: 20140039898Abstract: Methods and apparatus for voice-enabling a web application, wherein the web application includes one or more web pages rendered by a web browser on a computer. At least one information source external to the web application is queried to determine whether information describing a set of one or more supported voice interactions for the web application is available, and in response to determining that the information is available, the information is retrieved from the at least one information source. Voice input for the web application is then enabled based on the retrieved information.Type: ApplicationFiled: August 2, 2012Publication date: February 6, 2014Applicant: Nuance Communications, Inc.Inventors: David E. Reich, Christopher Hardy
-
Publication number: 20140029701Abstract: Systems, methods, and devices synchronize data streams by hashing received data frames to generate a sequence of hash values, comparing the generated hash value sequence to a hash value sequence received in a control stream, and processing data frames when the hash value sequences match. A source device and multiple receiver devices may synchronize audio data encoded in data frames, applying a hash function to each data frame to generate a first sequence of hash values, transmitting the data frames on a first channel and the first sequence of hash values on a control channel, receiving the data frames and the first sequence of hash values in the receiver devices, applying the hash algorithm to received data frames to generate a second sequence of hash values, comparing the first and second sequences of hash values, and processing data frames when the first and second sequences of hash values match.Type: ApplicationFiled: July 29, 2012Publication date: January 30, 2014Inventors: Adam E. NEWHAM, Joel Benjamin LINSKY, Rohit SAUHTA, Brian F. MILLER, Kevin Wayne BARTIG
-
Publication number: 20140019139Abstract: A blood glucose meter with a simplified programmable voice function, including: a microprocessor; a memory that is both programmable and re-programmable coupled to the microprocessor; and an audio output device coupled to the microprocessor and the memory; wherein a language algorithm and a plurality of language components specific to a language selected by a user are disposed within the memory; and wherein the language algorithm and the plurality of language components are utilized to provide an audio output through the audio output device in the language selected by the user. The language algorithm is operable for determining which language components are utilized to provide the audio output and in what order based on the language selected by the user. Optionally, the audio output is generated by the microprocessor and the memory using a pulse-width modulation scheme and/or the like.Type: ApplicationFiled: July 12, 2012Publication date: January 16, 2014Applicant: PRODIGY DIABETES CARE, LLCInventors: Ramzi ABULHAJ, Moo Nam KO, William BAXTER, Amr Yehia Mohamed SHEHAB
-
Publication number: 20140012586Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for determining hotword suitability. In one aspect, a method includes receiving speech data that encodes a candidate hotword spoken by a user, evaluating the speech data or a transcription of the candidate hotword, using one or more predetermined criteria, generating a hotword suitability score for the candidate hotword based on evaluating the speech data or a transcription of the candidate hotword, using one or more predetermined criteria, and providing a representation of the hotword suitability score for display to the user.Type: ApplicationFiled: August 6, 2012Publication date: January 9, 2014Applicant: GOOGLE INC.Inventors: Andrew E. Rubin, Johan Schalkwyk, Maria Carolina Parada San Martin
-
Publication number: 20130346085Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, and including hardware devices performing a mouth click sound based human-device interaction. In one aspect, receiving at least one mouth click sound signals from a human user, by an acoustic-to-electric sensor of a computing device, and processing the received signals. The received mouth click sound signals may be accompanied by other mouth click sound signals, and other interaction signals.Type: ApplicationFiled: June 23, 2012Publication date: December 26, 2013Inventor: Zoltan Stekkelpak
-
Publication number: 20130346084Abstract: Technologies are described herein for enhancing a user presence status determination. Visual data may be received from a depth camera configured to be arranged within a three-dimensional space. A current user presence status of a user in the three-dimensional space may be determined based on the visual data. A previous user presence status of the user may be transformed to the current user presence status, responsive to determining the current user presence status of the user.Type: ApplicationFiled: June 22, 2012Publication date: December 26, 2013Applicant: MICROSOFT CORPORATIONInventors: Anne Marie Renée Archambault, Jeffrey Scott Berg, Xiping Zuo, Abhishek Agrawal
-
Publication number: 20130339007Abstract: Embodiments herein include receiving a request to modify an audio characteristic associated with a first user for a voice communication system. One or more suggested modified audio characteristics may be provided for the first user, based on, at least in part, one or more audio preferences established by another user. An input of one or more modified audio characteristics may be received for the first user for the voice communication system. A user-specific audio preference may be associated with the first user for voice communications on the voice communication system, the user-specific audio preference including the one or more modified audio characteristics.Type: ApplicationFiled: June 18, 2012Publication date: December 19, 2013Applicant: International Business Machines CorporationInventors: Ruthie D. Lyle, Patrick Joseph O'Sullivan, Lin Sun
-
Publication number: 20130332172Abstract: An accessory is configured to receive a request. The accessory transmits information associated with the request to a portable device. An automated assistant application executed by the portable device can interpret the request and provide a report. The portable device can transmit the report to the accessory. The report may include one or more results determined by the automated assistant.Type: ApplicationFiled: October 1, 2012Publication date: December 12, 2013Applicant: Apple Inc.Inventors: Jude A. Prakash, Shailesh Rathi, Daniel De Rocha Rosario, Sylvain R.Y. Louboutin
-
Publication number: 20130321390Abstract: A system and method are disclosed for augmenting a reading experience in a mixed reality environment. In response to predefined verbal or physical gestures, the mixed reality system is able to answer a user's questions or provide additional information relating to what the user is reading. Responses may be displayed to the user on virtual display slates in a border or around the reading material without obscuring text or interfering with the user's reading experience.Type: ApplicationFiled: May 31, 2012Publication date: December 5, 2013Inventors: Stephen G. Latta, Ryan L. Hastings, Cameron G. Brown, Aaron Krauss, Daniel J. McCulloch, Ben J. Sugden
-
Publication number: 20130325481Abstract: A method of providing navigation on an electronic device when the display screen is locked. The method receives a verbal request to start navigation while the display is locked. The method identifies a route from a current location to a destination based on the received verbal request. While the display screen is locked, the method provides navigational directions on the electronic device from the current location of the electronic device to the destination. Some embodiments provide a method for processing a verbal search request. The method receives a navigation-related verbal search request and prepares a sequential list of the search results based on the received request. The method then provides audible information to present a search result from the sequential list. The method presents the search results in a batch form until the user selects a search result, the user terminates the search, or the search items are exhausted.Type: ApplicationFiled: September 30, 2012Publication date: December 5, 2013Applicant: APPLE INC.Inventors: Marcel van Os, Sarah G. Barbour, Brady A. Law, Bradford A. Moore
-
Publication number: 20130325462Abstract: A system and method for assigning one or more tags to an image file. In one aspect, a server computer receives an image file captured by a client device. In one embodiment, the image file includes an audio component embedded therein by the client device, where the audio component was spoken by a user of the client device as a tag of the image file. The server computer determines metadata associated with the image file and identifies a dictionary of potential textual tags from the metadata. The server computer determines a textual tag from the audio component and from the dictionary of potential textual tags. The server computer then associates the textual tag with the image file as additional metadata.Type: ApplicationFiled: May 31, 2012Publication date: December 5, 2013Applicant: Yahoo! inc.Inventors: Oren Somekh, Nadav Golbandi, Liran Katzir, Ronny Lempel, Yoelle Maarek
-
Publication number: 20130317828Abstract: The effectiveness of targeted content delivery at a multi-user interface can be directly linked to a proper targeting of users. A way of improving targeted content delivery at a multi-user interface can be to determine which users should be targeted based on one or more criteria. The present technology provides various methodologies for selecting one or more users associated with a multi-user interface to receive targeted content. Such users can be selected based on criteria associated with a ranking or priority of the users, criteria associated with an analysis of their interactions with the multi-user interface, criteria based on their most common characteristics, or any combination thereof. The user characteristics associated with such identified used can then be utilized to determine which content should be delivered to the multi-user interface.Type: ApplicationFiled: May 25, 2012Publication date: November 28, 2013Applicant: Apple Inc.Inventors: Michael Froimowitz Greenzeiger, Mehul K. Sanghavi, Ravindra Phulari
-
Publication number: 20130304479Abstract: Methods and systems for determining intent in voice and gesture interfaces are described. An example method includes determining that a gaze direction is in a direction of a gaze target, and determining whether a predetermined time period has elapsed while the gaze direction is in the direction of the gaze target. The method may also include providing an indication that the predetermined time period has elapsed when the predetermined time period has elapsed. According to the method, a voice or gesture command that is received after the predetermined time period has elapsed may be determined to be an input for a computing system. Additional example systems and methods are described herein.Type: ApplicationFiled: May 8, 2012Publication date: November 14, 2013Applicant: GOOGLE INC.Inventors: Eric Teller, Daniel Aminzade
-
Publication number: 20130297320Abstract: An additive three-dimensional fabrication system includes voice control for user interaction. This voice-controlled interface can enable a variety of voice-controlled functions and operations, while supporting interactions specific to consumer-oriented fabrication processes.Type: ApplicationFiled: July 23, 2012Publication date: November 7, 2013Inventors: Anthony James Buser, Nathaniel B. Pettis
-
Publication number: 20130297298Abstract: Methods and apparatus for signal processing are disclosed. Source separation can be performed to extract source signals from mixtures of source signals by way of independent component analysis. Source separation described herein involves mixed multivariate probability density functions that are mixtures of component density functions having different parameters corresponding to frequency components of different sources, different time segments, or some combination thereof.Type: ApplicationFiled: May 4, 2012Publication date: November 7, 2013Applicant: Sony Computer Entertainment Inc.Inventors: Jaekwon Yoo, Ruxin Chen
-
Publication number: 20130297319Abstract: A mobile device having at least one microphone sensor and a method for controlling the same are disclosed. The method includes receiving at least two audio signals through the at least one microphone sensor within a predetermined time period, recognizing input directions and voice command from the at least two audio signals sequentially, determining whether the recognized input directions and voice command match to preset input directions and preset voice command mapped to the preset directions, sequentially for the at least two received audio signals, and executing a preset control command, if the recognized input directions and voice command match to the preset input directions and voice command.Type: ApplicationFiled: July 9, 2012Publication date: November 7, 2013Inventor: Yongsin KIM
-
SOURCE SEPARATION BY INDEPENDENT COMPONENT ANALYSIS IN CONJUNCTION WITH SOURCE DIRECTION INFORMATION
Publication number: 20130297296Abstract: Methods and apparatus for signal processing are disclosed. Source separation can be performed to extract source signals from mixtures of source signals by way of independent component analysis. Source direction information is utilized in the separation process, and independent component analysis techniques described herein use multivariate probability density functions to preserve the alignment of frequency bins in the source separation process. It is emphasized that this abstract is provided to comply with the rules requiring an abstract that will allow a searcher or other reader to quickly ascertain the subject matter of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims.Type: ApplicationFiled: May 4, 2012Publication date: November 7, 2013Applicant: Sony Computer Entertainment Inc.Inventors: Jaekwon Yoo, Ruxin Chen