Patents by Inventor Carlos M. Avendano
Carlos M. Avendano has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240107254Abstract: A method performed by an electronic device in a room. The method performs an enrollment process in which a spatial profile of a location of an artificial sound source is created and performs an identification process that determines whether a sound event within the room is produced by the artificial sound source by 1) capturing the sound event using a microphone array and 2) determining a likelihood that the sound event occurred at the location of the artificial sound source.Type: ApplicationFiled: December 7, 2023Publication date: March 28, 2024Inventors: Hassan Taherian, Jonathan Huang, Carlos M. Avendano
-
Patent number: 11941968Abstract: An electronic device includes a processor, and a memory containing instructions that, when executed by the processor, cause the electronic device to learn a sound emitted by a legacy device and to issue an output when the electronic device subsequently hears the sound. For example, the electronic device can receive a training input and extract a compact representation of a sound in the training input, which the device stores. The device can receive an audio signal corresponding to an observed acoustic scene and extract a representation of the observed acoustic scene from the audio signal. The electronic device can determine whether the sound is present in the observed acoustic scene at least in part from a comparison of the representation of the observed acoustic scene with the representation of the sound. The electronic device emits a selected output responsive to determining that the sound is present in the acoustic scene.Type: GrantFiled: January 30, 2023Date of Patent: March 26, 2024Assignee: Apple Inc.Inventors: Hyung-Suk Kim, Daniel C. Klingler, Miquel Espi Marques, Carlos M. Avendano
-
Publication number: 20240029754Abstract: Implementations of the subject technology provide systems and methods for providing audio source separation for audio input, such as for audio devices having limited power and/or computing resources. The subject technology may allow an audio device to leverage processing and/or power resources of a companion device that is communicatively coupled to the audio device. The companion device may identify a noise condition of the audio device, select a source separation model based on the noise condition, and provide the source separation model to the audio device. In this way, the audio device can provide audio source separation functionality using a relatively small footprint source separation model that is specific to the noise condition in which the audio device is operated.Type: ApplicationFiled: October 3, 2023Publication date: January 25, 2024Inventors: Carlos M. AVENDANO, John WOODRUFF, Jonathan HUANG, Mehrez SOUDEN, Andreas KOUTROUVELIS
-
Publication number: 20240005902Abstract: Microphone signals of a primary headphone are processed and either a first transparency mode of operation is activated or a second transparency mode of operation. In another aspect, a processor enters different configurations in response to estimated ambient acoustic noise being lower or higher than a threshold, wherein in a first configuration a transparency audio signal is adapted via target voice and wearer voice processing (TVWVP) of a microphone signal to boost detected speech frequencies in the transparency audio signal, and in a second configuration the TVWVP is controlled to, as the estimated ambient acoustic noise increases, reduce boosting of, or not boost at all, the detected speech frequencies in the transparency audio signal. Other aspects are also described and claimed.Type: ApplicationFiled: June 30, 2023Publication date: January 4, 2024Inventors: John Woodruff, Andreas Koutrouvelis, Fatemeh Pishdadian, Jonathan D. Sheaffer, Yang Lu, Carlos M. Avendano, Nasim Radmanesh
-
Publication number: 20240005903Abstract: Microphone signals of a primary headphone are processed and either a first transparency mode of operation is activated or a second transparency mode of operation. In another aspect, a processor enters different configurations in response to estimated ambient acoustic noise being lower or higher than a threshold, wherein in a first configuration a transparency audio signal is adapted via target voice and wearer voice processing (TVWVP) of a microphone signal to boost detected speech frequencies in the transparency audio signal, and in a second configuration the TVWVP is controlled to, as the estimated ambient acoustic noise increases, reduce boosting of, or not boost at all, the detected speech frequencies in the transparency audio signal. Other aspects are also described and claimed.Type: ApplicationFiled: June 30, 2023Publication date: January 4, 2024Inventors: Yang Lu, Carlos M. Avendano, Tony S. Verma
-
Patent number: 11863961Abstract: A method performed by an electronic device in a room. The method performs an enrollment process in which a spatial profile of a location of an artificial sound source is created and performs an identification process that determines whether a sound event within the room is produced by the artificial sound source by 1) capturing the sound event using a microphone array and 2) determining a likelihood that the sound event occurred at the location of the artificial sound source.Type: GrantFiled: December 5, 2022Date of Patent: January 2, 2024Assignee: Apple Inc.Inventors: Hassan Taherian, Jonathan Huang, Carlos M. Avendano
-
Publication number: 20230360641Abstract: The subject disclosure provides systems and methods for generating and storing learned embeddings of audio inputs to an electronic device. The electronic device may generate and store encoded versions of audio inputs and learned embeddings of the audio inputs. When a new audio input is obtained, the electronic device can generate an encoded version of the new audio input, compare the encoded version of the new audio input to the stored encoded versions of prior audio inputs, and if the encoded version of the new audio input matches one of the stored encoded versions of the prior audio inputs, the electronic device can provide a stored learned embedding that corresponds to the one of the stored encoded versions of the prior audio inputs to a detection model at the electronic device. The cached embeddings can be provided to locally trained models for detecting individual sounds using electronic devices.Type: ApplicationFiled: May 4, 2022Publication date: November 9, 2023Inventors: Daniel C. KLINGLER, Carlos M. AVENDANO, Jonathan HUANG, Miquel ESPI MARQUES
-
Patent number: 11810588Abstract: Implementations of the subject technology provide systems and methods for providing audio source separation for audio input, such as for audio devices having limited power and/or computing resources. The subject technology may allow an audio device to leverage processing and/or power resources of a companion device that is communicatively coupled to the audio device. The companion device may identify a noise condition of the audio device, select a source separation model based on the noise condition, and provide the source separation model to the audio device. In this way, the audio device can provide audio source separation functionality using a relatively small footprint source separation model that is specific to the noise condition in which the audio device is operated.Type: GrantFiled: January 31, 2022Date of Patent: November 7, 2023Assignee: Apple Inc.Inventors: Carlos M. Avendano, John Woodruff, Jonathan Huang, Mehrez Souden, Andreas Koutrouvelis
-
Publication number: 20230186904Abstract: An electronic device has one or more microphones that pick up a sound. At least one feature extractor processes the audio signals from the microphones, that contain the picked up the sound, to determine several features for the sound. The electronic device also includes a classifier that has a machine learning model which is configured to determine a sound classification, such as artificial versus natural for the sound, based upon at least one of the determined features. Other aspects are also described and claimed.Type: ApplicationFiled: November 22, 2022Publication date: June 15, 2023Inventors: Daniel C. Klingler, Carlos M. Avendano, Hyung-Suk Kim, Miquel Espi Marques
-
Publication number: 20230177942Abstract: An electronic device includes a processor, and a memory containing instructions that, when executed by the processor, cause the electronic device to learn a sound emitted by a legacy device and to issue an output when the electronic device subsequently hears the sound. For example, the electronic device can receive a training input and extract a compact representation of a sound in the training input, which the device stores. The device can receive an audio signal corresponding to an observed acoustic scene and extract a representation of the observed acoustic scene from the audio signal. The electronic device can determine whether the sound is present in the observed acoustic scene at least in part from a comparison of the representation of the observed acoustic scene with the representation of the sound. The electronic device emits a selected output responsive to determining that the sound is present in the acoustic scene.Type: ApplicationFiled: January 30, 2023Publication date: June 8, 2023Inventors: Hyung-Suk KIM, Daniel C. KLINGLER, Miquel ESPI MARQUES, Carlos M. AVENDANO
-
Patent number: 11568731Abstract: An electronic device includes a processor, and a memory containing instructions that, when executed by the processor, cause the electronic device to learn a sound emitted by a legacy device and to issue an output when the electronic device subsequently hears the sound. For example, the electronic device can receive a training input and extract a compact representation of a sound in the training input, which the device stores. The device can receive an audio signal corresponding to an observed acoustic scene and extract a representation of the observed acoustic scene from the audio signal. The electronic device can determine whether the sound is present in the observed acoustic scene at least in part from a comparison of the representation of the observed acoustic scene with the representation of the sound. The electronic device emits a selected output responsive to determining that the sound is present in the acoustic scene.Type: GrantFiled: May 11, 2020Date of Patent: January 31, 2023Assignee: APPLE INC.Inventors: Hyung-Suk Kim, Daniel C. Klingler, Miquel Espi Marques, Carlos M. Avendano
-
Patent number: 11533577Abstract: A method performed by an electronic device in a room. The method performs an enrollment process in which a spatial profile of a location of an artificial sound source is created and performs an identification process that determines whether a sound event within the room is produced by the artificial sound source by 1) capturing the sound event using a microphone array and 2) determining a likelihood that the sound event occurred at the location of the artificial sound source.Type: GrantFiled: May 20, 2021Date of Patent: December 20, 2022Assignee: APPLE INC.Inventors: Hassan Taherian, Jonathan Huang, Carlos M. Avendano
-
Publication number: 20220391758Abstract: The subject disclosure provides systems and methods for providing locally trained models for detecting individual sounds using electronic devices. Local detection of individual sounds with a detection model at an electronic device can be provided by obtaining training samples for the detection model with the electronic device, and generating additional negative and positive training samples based on the obtained training samples. A two-stage detection process may be provided, in which a trigger model at a device compares an audio input to a reference sound to trigger a detection model at the device. The detection of individual sounds with a detection model at an electronic device can also leverage audio capture capabilities of multiple devices in an acoustic scene to capture multiple concurrent training samples.Type: ApplicationFiled: May 4, 2022Publication date: December 8, 2022Inventors: Jonathan HUANG, Miquel ESPI MARQUES, Carlos M. AVENDANO, Kevin M. DURAND, David FINDLAY, Vasudha KOWTHA, Daniel C. KLINGLER, Yichi ZHANG
-
Patent number: 11521598Abstract: An electronic device has one or more microphones that pick up a sound. At least one feature extractor processes the audio signals from the microphones, that contain the picked up the sound, to determine several features for the sound. The electronic device also includes a classifier that has a machine learning model which is configured to determine a sound classification, such as artificial versus natural for the sound, based upon at least one of the determined features. Other aspects are also described and claimed.Type: GrantFiled: September 9, 2019Date of Patent: December 6, 2022Assignee: APPLE INC.Inventors: Daniel C. Klingler, Carlos M. Avendano, Hyung-Suk Kim, Miquel Espi Marques
-
Publication number: 20220377483Abstract: A method performed by an electronic device in a room. The method performs an enrollment process in which a spatial profile of a location of an artificial sound source is created and performs an identification process that determines whether a sound event within the room is produced by the artificial sound source by 1) capturing the sound event using a microphone array and 2) determining a likelihood that the sound event occurred at the location of the artificial sound source.Type: ApplicationFiled: May 20, 2021Publication date: November 24, 2022Inventors: Hassan Taherian, Jonathan Huang, Carlos M. Avendano
-
Publication number: 20220270629Abstract: Implementations of the subject technology provide systems and methods for providing audio source separation for audio input, such as for audio devices having limited power and/or computing resources. The subject technology may allow an audio device to leverage processing and/or power resources of a companion device that is communicatively coupled to the audio device. The companion device may identify a noise condition of the audio device, select a source separation model based on the noise condition, and provide the source separation model to the audio device. In this way, the audio device can provide audio source separation functionality using a relatively small footprint source separation model that is specific to the noise condition in which the audio device is operated.Type: ApplicationFiled: January 31, 2022Publication date: August 25, 2022Inventors: Carlos M. AVENDANO, John WOODRUFF, Jonathan HUANG, Mehrez SOUDEN, Andreas KOUTROUVELIS
-
Publication number: 20210020018Abstract: An electronic device includes a processor, and a memory containing instructions that, when executed by the processor, cause the electronic device to learn a sound emitted by a legacy device and to issue an output when the electronic device subsequently hears the sound. For example, the electronic device can receive a training input and extract a compact representation of a sound in the training input, which the device stores. The device can receive an audio signal corresponding to an observed acoustic scene and extract a representation of the observed acoustic scene from the audio signal. The electronic device can determine whether the sound is present in the observed acoustic scene at least in part from a comparison of the representation of the observed acoustic scene with the representation of the sound. The electronic device emits a selected output responsive to determining that the sound is present in the acoustic scene.Type: ApplicationFiled: May 11, 2020Publication date: January 21, 2021Inventors: Hyung-Suk Kim, Daniel C. Klingler, Miquel Espi Marques, Carlos M. Avendano
-
Patent number: 10861210Abstract: Embodiments of the present disclosure can provide systems, methods, and computer-readable medium for providing audio and/or video effects based at least in part on facial features and/or voice feature characteristics of the user. For example, video and/or an audio signal of the user may be recorded by a device. Voice audio features and facial feature characteristics may be extracted from the voice audio signal and the video, respectively. The facial features of the user may be used to modify features of a virtual avatar to emulate the facial feature characteristics of the user. The extracted voice audio features may modified to generate an adjusted audio signal or an audio signal may be composed from the voice audio features. The adjusted/composed audio signal may simulate the voice of the virtual avatar. A preview of the modified video/audio may be provided at the user's device.Type: GrantFiled: July 11, 2018Date of Patent: December 8, 2020Assignee: Apple Inc.Inventors: Carlos M. Avendano, Sean A. Ramprashad
-
Publication number: 20200090644Abstract: An electronic device has one or more microphones that pick up a sound. At least one feature extractor processes the audio signals from the microphones, that contain the picked up the sound, to determine several features for the sound. The electronic device also includes a classifier that has a machine learning model which is configured to determine a sound classification, such as artificial versus natural for the sound, based upon at least one of the determined features. Other aspects are also described and claimed.Type: ApplicationFiled: September 9, 2019Publication date: March 19, 2020Inventors: Daniel C. Klingler, Carlos M. Avendano, Hyung-Suk Kim, Miquel Espi Marques
-
Publication number: 20180336713Abstract: Embodiments of the present disclosure can provide systems, methods, and computer-readable medium for providing audio and/or video effects based at least in part on facial features and/or voice feature characteristics of the user. For example, video and/or an audio signal of the user may be recorded by a device. Voice audio features and facial feature characteristics may be extracted from the voice audio signal and the video, respectively. The facial features of the user may be used to modify features of a virtual avatar to emulate the facial feature characteristics of the user. The extracted voice audio features may modified to generate an adjusted audio signal or an audio signal may be composed from the voice audio features. The adjusted/composed audio signal may simulate the voice of the virtual avatar. A preview of the modified video/audio may be provided at the user's device.Type: ApplicationFiled: July 11, 2018Publication date: November 22, 2018Applicant: Apple Inc.Inventors: Carlos M. Avendano, Sean A. Ramprashad