Detect Speech In Noise Patents (Class 704/233)
  • Patent number: 11393492
    Abstract: A method for establishing a voice activity detection model includes obtaining a training audio file and a target result of the training audio file, framing the training audio file to obtain an audio frame, extracting an audio feature of the audio frame, the audio feature comprising at least two types of features, inputting the extracted audio feature as an input to a deep neural network model, performing information processing on the audio feature through a hidden layer of the deep neural network model, and outputting the processed audio feature through an output layer of the deep neural network model, to obtain a training result; determining a bias between the training result and the target result, and inputting the bias as an input to an error back propagation mechanism, and updating weights of the hidden layer until the deep neural network model reaches a preset condition.
    Type: Grant
    Filed: November 8, 2019
    Date of Patent: July 19, 2022
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LTD
    Inventor: Haibo Liu
  • Patent number: 11386904
    Abstract: Deterioration of voice extraction performance when positions of a plurality of microphones are changed is prevented. A signal processing device according to an embodiment of the present technology includes a voice extraction unit that performs voice extraction from signals of a plurality of microphones, in which the voice extraction unit uses, when respective positions of the plurality of microphones are changed to positions where other microphones have been present, respective signals of the plurality of microphones as signals of the other microphones. Thus, it is possible to cancel the effect of changing the positions of respective microphones on the voice extraction.
    Type: Grant
    Filed: March 19, 2019
    Date of Patent: July 12, 2022
    Assignee: Sony Corporation
    Inventor: Kazuya Tateishi
  • Patent number: 11386896
    Abstract: Systems and methods are disclosed. A digitized human vocal expression of a user and digital images are received over a network from a remote device. The digitized human vocal expression is processed to determine characteristics of the human vocal expression, including: pitch, volume, rapidity, a magnitude spectrum identify, and/or pauses in speech. Digital images are received and processed to detect characteristics of the user face, including detecting if one or more of the following is present: a sagging lip, a crooked smile, uneven eyebrows, and/or facial droop. Based at least on part on the human vocal expression characteristics and face characteristics, a determination is made as to what action is to be taken. A cepstrum pitch may be determined using an inverse Fourier transform of a logarithm of a spectrum of a human vocal expression signal. The volume may be determined using peak heights in a power spectrum of the human vocal expression.
    Type: Grant
    Filed: January 29, 2020
    Date of Patent: July 12, 2022
    Assignee: The Notebook, LLC
    Inventor: Karen Elaine Khaleghi
  • Patent number: 11373670
    Abstract: Example embodiments disclosed herein relate to filter coefficient updating in time domain filtering. A method of processing an audio signal is disclosed. The method includes obtaining a predetermined number of target gains for a first portion of the audio signal by analyzing the first portion of the audio signal. Each of the target gains is corresponding to a subband of the audio signal. The method also includes determining filter coefficients for time domain filtering the first portion of the audio signal so as to approximate a frequency response given by the target gains. The filter coefficients are determined by iteratively selecting at least one target gain from the target gains and updating the filter coefficient based on the selected at least one target gain. Corresponding system and computer program product for processing an audio signal are also disclosed.
    Type: Grant
    Filed: May 6, 2019
    Date of Patent: June 28, 2022
    Assignee: Dolby Laboratories Licensing Corporation
    Inventors: Dong Shi, Xuejing Sun
  • Patent number: 11375322
    Abstract: The present application relates to a hearing aid adapted to be worn in or at an ear of a hearing aid user and/or to be fully or partially implanted in the head of the hearing aid user.
    Type: Grant
    Filed: February 28, 2020
    Date of Patent: June 28, 2022
    Assignee: Oticon A/S
    Inventors: Thomas Lunner, Lars Bramsløw
  • Patent number: 11350885
    Abstract: A method includes identifying, by an electronic device, one or more segments within a first audio recording that includes one or more non-speech segments and one or more speech segments. The method also includes generating, by the electronic device, one or more synthetic speech segments that include natural speech audio characteristics and that preserve one or more non-private features of the one or more speech segments. The method also includes generating, by the electronic device, an obfuscated audio recording by replacing the one or more speech segments with the one or more synthetic speech segments while maintaining the one or more non-speech segments, wherein the one or more synthetic speech segments prevent recognition of some content of the obfuscated audio recording.
    Type: Grant
    Filed: February 6, 2020
    Date of Patent: June 7, 2022
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Korosh Vatanparvar, Viswam Nathan, Ebrahim Nematihosseinabadi, Md Mahbubur Rahman, Jilong Kuang
  • Patent number: 11348575
    Abstract: A speaker recognition method and apparatus receives a first voice signal of a speaker, generates a second voice signal by enhancing the first voice signal through speech enhancement, generates a multi-channel voice signal by associating the first voice signal with the second voice signal, and recognizes the speaker based on the multi-channel voice signal.
    Type: Grant
    Filed: June 11, 2020
    Date of Patent: May 31, 2022
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Sung-Jae Cho, Kyuhong Kim, Jaejoon Han
  • Patent number: 11335350
    Abstract: An apparatus includes processor(s) to: perform pre-processing operations including derive an audio noise level of speech audio of a speech data set, derive a first relative weighting for first and second segmentation techniques for identifying likely sentence pauses in the speech audio based on the audio noise level, and select likely sentence pauses for a converged set of likely sentence pauses from likely sentence pauses identified by the first and/or second segmentation techniques based on the first relative weighting; and perform speech-to-text processing operations including divide the speech data set into data segments representing speech segments of the speech audio based on the converged set of likely sentence pauses, and derive a second relative weighting based on the audio noise level for selecting words indicated by an acoustic model or by a language model as being most likely spoken in the speech audio for inclusion in a transcript.
    Type: Grant
    Filed: October 12, 2021
    Date of Patent: May 17, 2022
    Assignee: SAS INSTITUTE INC.
    Inventors: Xiaolong Li, Xiaozhuo Cheng, Xu Yang
  • Patent number: 11327050
    Abstract: Disclosed herein are systems and methods for mechanical failure monitoring, detection, and classification in electronic assemblies. In some embodiments, a mechanical monitoring apparatus may include: a fixture to receive an electronic assembly; an acoustic sensor; and a computing device communicatively coupled to the acoustic sensor, wherein the acoustic sensor is to detect an acoustic emission waveform generated by a mechanical failure of the electronic assembly during testing.
    Type: Grant
    Filed: February 20, 2018
    Date of Patent: May 10, 2022
    Assignee: Intel Corporation
    Inventors: Kyle Yazzie, Rajesh Kumar Neerukatti, Naga Sivakumar Yagnamurthy, David C. McCoy, Pramod Malatkar, Frank P. Prieto
  • Patent number: 11328736
    Abstract: Disclosed are a method and an apparatus of denoising, and the method includes: receiving a first voice signal picked up by a microphone; if it is detected, with the first voice signal, that a sensor is in an operation state, subtracting an interference noise signal from the first voice signal to obtain a first voice signal with the interference removed therefrom, where the interference noise signal is an interference noise signal generated with regard to the microphone during an operation of the sensor, and the sensor and the microphone are packaged in one module; and outputting the first voice signal with the interference removed therefrom. By implementing the solution provided in the present disclosure, interference in a signal collected by a microphone when the microphone and a sensor in a module operate together is reduced and the small size of the module, packaged with the microphone and sensor, is guaranteed.
    Type: Grant
    Filed: August 28, 2017
    Date of Patent: May 10, 2022
    Assignee: WEIFANG GOERTEK MICROELECTRONICS CO., LTD.
    Inventors: Dexin Wang, Xiangju Xu, Luyu Duanmu
  • Patent number: 11322138
    Abstract: A voice awakening method and device are provided. According to an embodiment, the method includes: receiving voice information of a user; obtaining an awakening confidence level corresponding to the voice information based on the voice information; determining, on the basis of the awakening confidence level, whether the voice information is suspected wake-up voice information; and performing, in response to determining the voice information being the suspected wake-up voice information, a secondary determination on the voice information to obtain a secondary determination result, and determining whether to perform a wake-up operation on the basis of the secondary determination result. The embodiment implements a secondary verification on the voice information, thereby reducing the probability that the smart device is mistakenly awakened.
    Type: Grant
    Filed: February 6, 2019
    Date of Patent: May 3, 2022
    Assignees: Baidu Online Network Technology (Beijing) Co., Ltd., Shanghai Xiaodu Technology Co., Ltd.
    Inventors: Jun Li, Rui Yang, Lifeng Zhao, Xiaojian Chen, Yushu Cao
  • Patent number: 11322168
    Abstract: A dual microphone signal processing arrangement for reducing reverberation is described. Time domain microphone signals are developed from a pair of sensing microphones. These are converted to the time-frequency domain to produce complex value spectra signals. A binary gain function applies frequency-specific energy ratios between the spectra signals to produce transformed spectra signals. A sigmoid gain function based on an inter-microphone coherence value between the transformed spectra signals is applied to the transformed spectra signals to produce coherence adapted spectra signals. And an inverse time-frequency transformation is applied to the coherence adjusted spectra signals to produce time-domain reverberation-compensated microphone signals with reduced reverberation components.
    Type: Grant
    Filed: August 9, 2019
    Date of Patent: May 3, 2022
    Assignee: MED-EL Elektromedizinische Geraete GmbH
    Inventors: Kostas Kokkinakis, Joshua Stohl
  • Patent number: 11312164
    Abstract: A method is provided for extending the frequency band of an audio signal during a decoding or improvement process. The method includes obtaining the decoded signal in a first frequency band, referred to as a low band. Tonal components and a surround signal are extracted from the signal from the low-band signal, and the tonal components and the surround signal are combined by adaptive mixing using energy-level control factors to obtain an audio signal, referred to as a combined signal. The low-band decoded signal before the extraction step or the combined signal after the combination step are extended over at least one second frequency band which is higher than the first frequency band. Also proved are a frequency-band extension device which implements the described method and a decoder including a device of this type.
    Type: Grant
    Filed: July 13, 2020
    Date of Patent: April 26, 2022
    Assignee: Koninklijke Philips N.V.
    Inventors: Magdalena Kaniewska, Stephane Ragot
  • Patent number: 11308978
    Abstract: Methods, apparatus, systems and articles of manufacture are disclosed for distributed automatic speech recognition. An example apparatus includes a detector to process an input audio signal and identify a portion of the input audio signal including a sound to be evaluated, the sound to be evaluated organized into a plurality of audio features representing the sound. The example apparatus includes a quantizer to process the audio features using a quantization process to reduce the audio features to generate a reduced set of audio features for transmission. The example apparatus includes a transmitter to transmit the reduced set of audio features over a low-energy communication channel for processing.
    Type: Grant
    Filed: August 5, 2019
    Date of Patent: April 19, 2022
    Assignee: INTEL CORPORATION
    Inventors: Binuraj K. Ravindran, Francis M. Tharappel, Prabhakar R. Datta, Tobias Bocklet, Maciej Muchlinski, Tomasz Dorau, Josef G. Bauer, Saurin Shah, Georg Stemmer
  • Patent number: 11308946
    Abstract: Methods and an apparatus for performing feature extraction on speech in a microphone signal with embedded noise processing to reduce the amount of processing are provided. In embodiments, feature extraction and the noise estimate use an output of the same Fourier Transform, such that the noise filtering of the speech is embedded with the feature extraction of the speech.
    Type: Grant
    Filed: July 25, 2019
    Date of Patent: April 19, 2022
    Assignee: Cerence Operating Company
    Inventors: Jianzhong Teng, Xiao-Lin Ren, Xingui Zeng, Yi Gao
  • Patent number: 11302298
    Abstract: A signal processing method for an earphone includes: a motion state of a wearer of the earphone is detected by using an acceleration sensor arranged inside the earphone; a first microphone and a second microphone both arranged outside the earphone detect wind noise conditions corresponding to different frequency bands; and according to the motion state of the wearer of the earphone and the wind noise conditions corresponding to different frequency bands, operating modes of a feedforward filter and a feedback filter inside the earphone are adjusted, herein the feedforward filter and the feedback filter are configured for active noise cancellation of the earphone.
    Type: Grant
    Filed: February 19, 2021
    Date of Patent: April 12, 2022
    Assignee: Beijing Xiaoniao Tingting Technology Co., LTD.
    Inventors: Song Liu, Na Li, Bo Li
  • Patent number: 11295137
    Abstract: A system for exploiting visual information for enhancing audio signals via source separation and beamforming is disclosed. The system may obtain visual content associated with an environment of a user, and may extract, from the visual content, metadata associated with the environment. The system may determine a location of the user based on the extracted metadata. Additionally, the system may load, based on the location, an audio profile corresponding to the location of the user. The system may also load a user profile of the user that includes audio data associated with the user. Furthermore, the system may cancel, based on the audio profile and user profile, noise from the environment of the user. Moreover, the system may include adjusting, based on the audio profile and user profile, an audio signal generated by the user so as to enhance the audio signal during a communications session of the user.
    Type: Grant
    Filed: November 2, 2020
    Date of Patent: April 5, 2022
    Assignee: AT&T INIELLECTUAL PROPERTY I, L.P.
    Inventors: Dimitrios Dimitriadis, Donald J. Bowen, Lusheng Ji, Horst J. Schroeter
  • Patent number: 11289109
    Abstract: Embodiments of the disclosure provide systems and methods for audio signal processing. An exemplary system may include a communication interface configured to receiving a first audio signal acquired from an audio source through a first channel, and a second audio signal acquired from the same audio source through a second channel. The system may also include at least one processor coupled to the communication interface. The at least one processor may be configured to determine channel features based on the first audio signal and the second audio signal individually and determine a cross-channel feature based on the first audio signal and the second audio signal collectively. The at least one processor may further be configured to concatenate the channel features and the cross-channel feature and estimate spectral-spatial masks for the first channel and the second channel using the concatenated channel features and the cross-channel feature.
    Type: Grant
    Filed: April 24, 2020
    Date of Patent: March 29, 2022
    Assignee: BEIJING DIDI INFINITY TECHNOLOGY AND DEVELOPMENT CO., LTD.
    Inventors: Chengyun Deng, Hui Song, Yi Zhang, Yongtao Sha
  • Patent number: 11282528
    Abstract: One embodiment provides a method, including: receiving, at an information handling device, user input comprising a potential wake word; determining, using a processor, whether the potential wake word is associated with a stored wake word; and responsive to determining that the potential wake word is associated with the stored wake word, activating, based on the potential wake word, a digital assistant associated with the information handling device. Other aspects are described and claimed.
    Type: Grant
    Filed: August 14, 2017
    Date of Patent: March 22, 2022
    Assignee: Lenovo (Singapore) Pte. Ltd.
    Inventors: Ryan Charles Knudson, Russell Speight VanBlon, Roderick Echols, Jonathan Gaither Knox
  • Patent number: 11270696
    Abstract: An audio device with at least one microphone adapted to receive sound from a sound field and create an output, and a processing system that is responsive to the output of the microphone. The processing system is configured to use a signal processing algorithm to detect a wakeup word, and modify the signal processing algorithm that is used to detect the wakeup word if the sound field changes.
    Type: Grant
    Filed: July 1, 2019
    Date of Patent: March 8, 2022
    Assignee: Bose Corporation
    Inventors: Ricardo Carreras, Alaganandan Ganeshkumar
  • Patent number: 11257485
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for neural network adaptive beamforming for multichannel speech recognition are disclosed. In one aspect, a method includes the actions of receiving a first channel of audio data corresponding to an utterance and a second channel of audio data corresponding to the utterance. The actions further include generating a first set of filter parameters for a first filter based on the first channel of audio data and the second channel of audio data and a second set of filter parameters for a second filter based on the first channel of audio data and the second channel of audio data. The actions further include generating a single combined channel of audio data. The actions further include inputting the audio data to a neural network. The actions further include providing a transcription for the utterance.
    Type: Grant
    Filed: December 10, 2019
    Date of Patent: February 22, 2022
    Assignee: Google LLC
    Inventors: Bo Li, Ron J. Weiss, Michiel A. U. Bacchiani, Tara N. Sainath, Kevin William Wilson
  • Patent number: 11257487
    Abstract: Techniques are described herein for enabling the use of “dynamic” or “context-specific” hot words to invoke an automated assistant. In various implementations, an automated assistant may be executed in a default listening state at least in part on a user's computing device(s). While in the default listening state, audio data captured by microphone(s) may be monitored for default hot words. Detection of the default hot word(s) transitions of the automated assistant into a speech recognition state. Sensor signal(s) generated by hardware sensor(s) integral with the computing device(s) may be detected and analyzed to determine an attribute of the user. Based on the analysis, the automated assistant may transition into an enhanced listening state in which the audio data may be monitored for enhanced hot word(s). Detection of enhanced hot word(s) triggers the automated assistant to perform a responsive action without requiring detection of default hot word(s).
    Type: Grant
    Filed: August 21, 2018
    Date of Patent: February 22, 2022
    Assignee: GOOGLE LLC
    Inventor: Diego Melendo Casado
  • Patent number: 11257497
    Abstract: The present disclosure provides a voice wake-up processing method, an apparatus and a storage medium. After acquiring voice wake-up signals collected by audio input devices in at least two audio zones, an electronic device may correct, based on to-be-woken-up audio zones obtained from amplitudes of the voice wake-up signals collected by the audio input devices in the at least two audio zones, a to-be-woken-up audio zone identified using a voice engine, avoiding that audio zones in which a plurality of audio input devices collecting voice wake-up signals produced from a same user are located are all woken up, therefore, it is possible to improve accuracy of a voice wake-up result obtained by the electronic device. Therefore, the present disclosure can solve the technical problem that a vehicle-mounted terminal has low voice wake-up accuracy due to an insufficient degree of sound isolation between audio zones of the vehicle-mounted terminal.
    Type: Grant
    Filed: December 23, 2019
    Date of Patent: February 22, 2022
    Assignee: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD.
    Inventors: Hanying Peng, Nengjun Ouyang
  • Patent number: 11257512
    Abstract: Systems and methods include a first voice activity detector operable to detect speech in a frame of a multichannel audio input signal and output a speech determination, a constrained minimum variance adaptive filter operable to receive the multichannel audio input signal and the speech determination and minimize a signal variance at the output of the filter, thereby producing an equalized target speech signal, a mask estimator operable to receive the equalized target speech signal and the speech determination and generate a spectral-temporal mask to discriminate a target speech from noise and interference speech, and a second activity voice detector operable to detect voice in a frame of the speech discriminated signal. An audio input sensor array including a plurality of microphones, each microphone generating a channel of the multichannel audio input signal. A sub-band analysis module operable to decompose each of the channels into a plurality of frequency sub-bands.
    Type: Grant
    Filed: January 6, 2020
    Date of Patent: February 22, 2022
    Assignee: SYNAPTICS INCORPORATED
    Inventors: Francesco Nesta, Alireza Masnadi-Shirazi
  • Patent number: 11250038
    Abstract: An interactive question and answer (Q&A) service provides pairs of questions and corresponding answers related to the content of a web page. The service includes pre-configured Q&A pairs derived from a deep learning framework that includes a series of neural networks trained through joint and transfer learning to generate questions for a given text passage. In addition, pre-configured Q&A pairs are generated from historical web access patterns and sources related to the content of the web page.
    Type: Grant
    Filed: August 13, 2018
    Date of Patent: February 15, 2022
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC.
    Inventors: Payal Bajaj, Gearard Boland, Anshul Gupta, Matthew Glenn Jin, Eduardo Enrique Noriega De Armas, Jason Shaver, Neelakantan Sundaresan, Roshanak Zilouchian Moghaddam
  • Patent number: 11250877
    Abstract: A method for generating a health indicator for at least one person of a group of people, the method comprising: receiving, at a processor, captured sound, where the captured sound is sound captured from the group of people; comparing the captured sound to a plurality of sound models to detect at least one non-speech sound event in the captured sound, each of the plurality of sound models associated with a respective health-related sound type; determining metadata associated with the at least one non-speech sound event; assigning the at least one non-speech sound event and the metadata to at least one person of the group of people; and outputting a message identifying the at least one non-speech event and the metadata to a health indicator generator module to generate a health indicator for the at least one person to whom the at least one non-speech sound event is assigned.
    Type: Grant
    Filed: July 25, 2019
    Date of Patent: February 15, 2022
    Assignee: AUDIO ANALYTIC LTD
    Inventors: Christopher Mitchell, Joe Patrick Lynas, Sacha Krstulovic, Amoldas Jasonas, Julian Harris
  • Patent number: 11240609
    Abstract: An audio device that includes a music classifier that determines when music is present in an audio signal is disclosed. The audio device is configured to receive audio, process the received audio, and to output the processed audio to a user. The processing may be adjusted based on the output of the music classifier. The music classifier utilizes a plurality of decision making units, each operating on the received audio independently. The decision making units are simplified to reduce the processing, and therefore the power, necessary for operation. Accordingly each decision making unit may be insufficient to determine music alone but in combination may accurately detect music while consuming power at a rate that is suitable for a mobile device, such as a hearing aid.
    Type: Grant
    Filed: June 3, 2019
    Date of Patent: February 1, 2022
    Assignee: SEMICONDUCTOR COMPONENTS INDUSTRIES, LLC
    Inventors: Pejman Dehghani, Robert L. Brennan
  • Patent number: 11238883
    Abstract: A method and a system for dialogue enhancement of an audio signal, comprising receiving (step S1) the audio signal and a text content associated with dialogue occurring in the audio signal, generating (step S2) parameterized synthesized speech from the text content, and applying (step S3) dialogue enhancement to the audio signal based on the parameterized synthesized speech. With the invention text captions, subtitles, or other forms of text content included in an audio stream, can be used to significantly improve dialogue enhancement on the playback side.
    Type: Grant
    Filed: May 23, 2019
    Date of Patent: February 1, 2022
    Assignee: Dolby Laboratories Licensing Corporation
    Inventors: Timothy Alan Port, Winston Chi Wai Ng, Mark William Gerrard
  • Patent number: 11222625
    Abstract: Systems and methods for training a control panel to recognize user defined and preprogrammed sound patterns are provided. Such systems and methods can include the control panel operating in a learning mode, receiving initial ambient audio from a region, and saving the initial ambient audio as an audio pattern in a memory device of the control panel. Such systems and methods can also include the control panel operating in an active mode, receiving subsequent ambient audio from the region, using an audio classification model to make an initial determination as to whether the subsequent ambient audio matches or is otherwise consistent with the audio pattern, determining whether the initial determination is correct, and when the control panel determines that the initial determination is incorrect, modifying or updating the audio classification model for improving the accuracy in detecting future consistency with the audio pattern.
    Type: Grant
    Filed: April 15, 2019
    Date of Patent: January 11, 2022
    Assignee: Ademco Inc.
    Inventors: Pradyumna Sampath, Ramprasad Yelchuru, Purnaprajna R. Mangsuli
  • Patent number: 11222654
    Abstract: A method for voice detection, the method may include (a) generating an in-ear signal that represents a signal sensed by an in-ear microphone and fed to a feedback active noise cancellation (ANC) circuit; (b) generating at least one additional signal, based on at least one out of a playback signal and a pickup signal sensed by a voice pickup microphone; and (c) generating a voice indicator based on the in-ear signal and the at least one additional signal.
    Type: Grant
    Filed: January 13, 2020
    Date of Patent: January 11, 2022
    Assignee: DSP GROUP LTD.
    Inventors: Assaf Ganor, Ori Elyada
  • Patent number: 11222624
    Abstract: A server may provide a voice recognition service. The server may include a memory configured for storing a plurality of voice recognition models, a communication device configured for communicating a plurality of voice recognition devices, and an artificial intelligence device configured for providing a voice recognition service to the plurality of voice recognition devices, acquiring use-related information regarding a first voice recognition device (from among the plurality of voice recognition devices), and changing a voice recognition model corresponding to the first voice recognition device from a first voice recognition model to a second voice recognition model based on the use-related information.
    Type: Grant
    Filed: August 20, 2019
    Date of Patent: January 11, 2022
    Assignee: LG ELECTRONICS INC.
    Inventors: Jaehong Kim, Hangil Jeong
  • Patent number: 11204736
    Abstract: The systems and methods described relate to the concept that smart devices can be used to 1) sense various types of phenomena like sound, blue light exposure, RF and microwave radiation, and 2) in real-time analyze, report and/or control outputs (e.g., displays or speakers). The systems are configurable and use standard computing devices, such as wearable electronics, tablet computers, and mobile phones to measure various frequency bands across multiple points, allowing a single user to visualize and/or adjust environmental conditions.
    Type: Grant
    Filed: October 17, 2019
    Date of Patent: December 21, 2021
    Assignee: ZOPHONOS INC.
    Inventor: Levaughn Denton
  • Patent number: 11205411
    Abstract: A method for processing audio signal includes that: audio signals emitted respectively from at least two sound sources are acquired through at least two microphones to obtain respective original noisy signals of the at least two microphones; sound source separation is performed on the respective original noisy signals of the at least two microphones to obtain respective time-frequency estimated signals of the at least two sound sources; a mask value of the time-frequency estimated signal of each sound source in the original noisy signal of each microphone is determined based on the respective time-frequency estimated signals; the respective time-frequency estimated signals of the at least two sound sources are updated based on the respective original noisy signals of the at least two microphones and the mask values; and the audio signals emitted respectively from the at least two sound sources are determined.
    Type: Grant
    Filed: May 29, 2020
    Date of Patent: December 21, 2021
    Assignee: Beijing Xiaomi Intelligent Technology Co., Ltd.
    Inventor: Haining Hou
  • Patent number: 11188718
    Abstract: A collective emotional engagement detection arrangement is provided for determining emotions of users in group conversations. A computer-implemented method includes determining a first conversation velocity of communications through conversation channels over a first time period for a group discussion between user computers; determining that a conversation velocity of the communications has increased to a second conversation velocity of communications which exceeds a predetermined threshold, and has remained above the predetermined threshold for at least a second time period; determining, aggregated emotions of the users during the second time period; and providing an output to a moderator of the group discussion indicating that the second conversation velocity of the communications has exceeded the predetermined threshold for at least the second time period, and indicating the aggregated emotions of the users during the second time period.
    Type: Grant
    Filed: September 27, 2019
    Date of Patent: November 30, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Ilse M. Breedvelt-Schouten, John A. Lyons, Jana H. Jenkins, Jeffrey A. Kusnitz
  • Patent number: 11189303
    Abstract: A multi-microphone algorithm for detecting and differentiating interference sources from desired talker speech in advanced audio processing for smart home applications is described. The approach is based on characterizing a persistent interference source when sounds repeated occur from a fixed spatial location relative to the device, which is also fixed. Some examples of such interference sources include TV, music system, air-conditioner, washing machine, and dishwasher. Real human talkers, in contrast, are not expected to remain stationary and speak continuously from the same position for a long time. The persistency of an acoustic source is established based on identifying historically-recurring inter-microphone frequency-dependent phase profiles in multiple time periods of the audio data. The detection algorithm can be used with a beamforming processor to suppress the interference and for achieving voice quality and automatic speech recognition rate improvements in smart home applications.
    Type: Grant
    Filed: September 25, 2017
    Date of Patent: November 30, 2021
    Assignee: Cirrus Logic, Inc.
    Inventors: Narayan Kovvali, Seth Suppappola
  • Patent number: 11184244
    Abstract: The current document is directed to methods and systems that employ network metrics collected by distributed-computer-system metrics-collection services to determine a service-call-based topology for distributed service-oriented applications. In a described implementation, network metrics are collected over a number of network-metric monitoring periods. Independent component analysis is used to extract, from the collected network metrics, signals corresponding to sequences of service calls initiated by calls to the application-programming interface of a distributed service-oriented application. The signals, in combination with call traces obtained from a distributed-services call-tracing utility or service, are then used to construct representations of distributed-service-oriented-application topologies. The distributed-service-oriented-application topologies provide a basis for any additional types of distributed-computer-system functionalities, utilities, and facilities.
    Type: Grant
    Filed: February 19, 2020
    Date of Patent: November 23, 2021
    Assignee: VMware, Inc.
    Inventors: Susobhit Panigrahi, Reghuram Vasanthakumari, Arihant Jain
  • Patent number: 11164591
    Abstract: A speech enhancement method includes determining a first spectral subtraction parameter based on a power spectrum of a speech signal containing noise and a power spectrum of a noise signal, determining a second spectral subtraction parameter based on the first spectral subtraction parameter and a reference power spectrum, and performing, based on the power spectrum of the noise signal and the second spectral subtraction parameter, spectral subtraction on the speech signal containing noise, where the reference power spectrum includes a predicted user speech power spectrum and/or predicted environmental noise power. Regularity of a power spectrum feature of a user speech of a terminal device and/or regularity of a power spectrum feature of noise in an environment in which a user is located are considered.
    Type: Grant
    Filed: January 18, 2018
    Date of Patent: November 2, 2021
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Weixiang Hu, Lei Miao
  • Patent number: 11150869
    Abstract: Aspects of the present disclosure relate to voice command filtering. One or more directions of background noise for a location of a voice command device are determined. The one or more directions of background noise are stored as one or more blocked directions. A voice input is received at the location of the voice command device. A direction the voice input is being received from is determined and compared to the one or more blocked directions. The voice input is ignored in response to the direction of the voice input being received from corresponding to a direction of the one or more blocked directions, unless the received voice input is in a recognized voice.
    Type: Grant
    Filed: February 14, 2018
    Date of Patent: October 19, 2021
    Assignee: International Business Machines Corporation
    Inventors: Eunjin Lee, Daniel Cunnington, John J. Wood, Giacomo G. Chiarella
  • Patent number: 11146298
    Abstract: A signal generator device includes a digital signal waveform generator to produce a digital signal waveform, a first frequency band signal path having a first frequency band filter to receive the digital signal waveform and to pass first frequency band components of the digital signal waveform, and a first digital-to-analog converter to receive the first frequency band components of the digital signal waveform and to produce a first frequency band analog signal, a second frequency band signal path having a second frequency band filter to receive the digital signal waveform and to pass second frequency band components of the digital signal waveform, a second digital-to-analog converter to receive the second frequency band components of the digital signal waveform and to produce a second frequency band analog signal, and a combining element to combine the first frequency band analog signal and the second frequency band analog signal to produce a wideband analog signal.
    Type: Grant
    Filed: September 30, 2019
    Date of Patent: October 12, 2021
    Assignee: Tektronix, Inc.
    Inventor: Gregory A. Martin
  • Patent number: 11146907
    Abstract: A system for identifying the contribution of a given sound source to a composite audio track, the system comprising an audio input unit operable to receive an input composite audio track comprising two or more sound sources, including the given sound source, an audio generation unit operable to generate, using a model of a sound source, an approximation of the contribution of the given sound source to the composite audio track, an audio comparison unit operable to compare the generated audio to at least a portion of the composite audio track to determine whether the generated audio provides an approximation of the composite audio track that meets a threshold degree of similarity, and an audio identification unit operable to identify, when the threshold is met, the generated audio as a suitable representation of the contribution of the sound source to the composite audio track.
    Type: Grant
    Filed: April 3, 2020
    Date of Patent: October 12, 2021
    Assignee: Sony Interactive Entertainment Inc.
    Inventors: Fabio Cappello, Oliver Hume
  • Patent number: 11138992
    Abstract: This application discloses a voice activity detection method. The method includes receiving speech data, the speech data including a multi-frame speech signal; determining energy and spectral entropy of a frame of speech signal; calculating a square root of the energy of the speech signal and/or calculating a square root of the spectral entropy of the frame of the speech signal; determining a spectral entropy-energy square root of the frame of the speech signal based on at least one of the square root of the energy and the square root of the spectral entropy; and determining that the frame of the speech signal is an unvoiced frame if the spectral entropy-energy square root of the speech signal is less than a first threshold, or that it is a voiced frame if the spectral entropy-energy square root of the speech signal is greater than or equal to the first threshold.
    Type: Grant
    Filed: October 28, 2019
    Date of Patent: October 5, 2021
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventor: Jizhong Liu
  • Patent number: 11132998
    Abstract: A voice recognition device includes: a first feature vector calculating unit (2) for calculating a first feature vector from voice data input; an acoustic likelihood calculating unit (4) for calculating an acoustic likelihood of the first feature vector by using an acoustic model used for calculating an acoustic likelihood of a feature vector; a second feature vector calculating unit (3) for calculating a second feature vector from the voice data; a noise degree calculating unit (6) for calculating a noise degree of the second feature vector by using a discriminant model used for calculating a noise degree indicating whether a feature vector is noise or voice; a noise likelihood recalculating unit (8) for recalculating an acoustic likelihood of noise on the basis of the acoustic likelihood of the first feature vector and the noise degree of the second feature vector; and a collation unit (9) for performing collation with a pattern of a vocabulary word to be recognized, by using the acoustic likelihood calcula
    Type: Grant
    Filed: March 24, 2017
    Date of Patent: September 28, 2021
    Assignee: MITSUBISHI ELECTRIC CORPORATION
    Inventors: Toshiyuki Hanazawa, Tomohiro Narita
  • Patent number: 11122367
    Abstract: In a method for controlling an audio system of a vehicle, an intention to communicate and/or a voice of at least one of a specific, occupant of the vehicle and/or of an occupant on a specific seat of the vehicle, are/is sensed, and at least one audio signal of the vehicle is changed as a function of the sensed intention to communicate and/or the voice of the occupant.
    Type: Grant
    Filed: October 2, 2019
    Date of Patent: September 14, 2021
    Assignee: Bayerische Motoren Werke Aktiengesellschaft
    Inventor: Alexander Augst
  • Patent number: 11114108
    Abstract: A method includes extracting, from multiple microphone input, a hyperset of features of acoustic sources, using the extracted features to identify separable clusters associated with acoustic scenarios, and classifying subsequent input as one of the acoustic scenarios using the hyperset of features. The acoustic scenarios include a desired spatially moving/non-moving talker, and an undesired spatially moving/non-moving acoustic source. The hyperset of features includes both spatial and voice biometric features. The classified acoustic scenario may be used in a robotics application or voice assistant device desired speech enhancement or interference signal cancellation. Specifically, the classification of the acoustic scenarios can be used to adapt a beamformer, e.g., step size adjustment. The hyperset of features may also include visual biometric features extracted from one or more cameras viewing the acoustic sources.
    Type: Grant
    Filed: May 11, 2020
    Date of Patent: September 7, 2021
    Assignee: Cirrus Logic, Inc.
    Inventors: Ghassan Maalouli, Samuel P. Ebenezer
  • Patent number: 11114089
    Abstract: A method, system, and computer program product for applying a profile to an assistive device based on a multitude of cues includes: gathering audio inputs surrounding an assistive device; analyzing, by the assistive device, the audio inputs; determining, based on the analyzing, scenario cues; classifying a current environment surrounding the assistive device from the scenario cues; comparing the current environment to device profiles of the assistive device; determining, based on the comparing, a matching profile; and, in response to determining the matching profile, executing the matching profile on the assistive device.
    Type: Grant
    Filed: November 19, 2018
    Date of Patent: September 7, 2021
    Assignee: International Business Machines Corporation
    Inventors: Matthew Chapman, Chengxuan Xing, Andrew J. Daniel, Ashley Harrison
  • Patent number: 11114093
    Abstract: An intelligent voice recognition method, voice recognition apparatus and intelligent computing device are disclosed. An intelligent voice recognition method according to an embodiment of the present invention obtains a microphone detection signal, recognizes a voice of a user from the microphone detection signal and outputs a response related to the voice on the basis of a result of recognition of the voice, wherein the microphone detection signal includes noise, and a microphone detection signal including only the voice obtained by removing the noise from the microphone detection signal is recognized. Accordingly, only a voice of a user can be effectively separated from a microphone detection signal detected through a microphone of the voice recognition apparatus.
    Type: Grant
    Filed: August 29, 2019
    Date of Patent: September 7, 2021
    Assignee: LG ELECTRONICS INC.
    Inventor: Wonchul Kim
  • Patent number: 11094336
    Abstract: A sound analysis apparatus includes a sound acquirer configured to acquire a sound signal, a measurer configured to output time-series data of numerical values representing volumes based on the sound signal, and a calculator configured to perform calculation for analyzing the time-series data output from the measurer, wherein the calculator performs the calculation in a case of a first state in which a measured value that is the numerical value output from the measurer is included within an analysis target range that is a range in which the measured value is determined to be an analysis target, and wherein the calculator does not perform the calculation in a case of a second state in which the measured value is not included within the analysis target range.
    Type: Grant
    Filed: August 15, 2019
    Date of Patent: August 17, 2021
    Assignee: Yokogawa Electric Corporation
    Inventors: Yuko Ito, Hiroki Yoshino
  • Patent number: 11089404
    Abstract: A sound processing apparatus includes n number of microphones that are disposed correspondingly to n number of persons and that mainly collect sound signals uttered by respective relevant persons, a filter that suppresses crosstalk components included in a talker sound signal collected by a microphone corresponding to at least one talker using the sound signals collected by the n number of microphones, a parameter updater that updates a parameter of the filter for suppressing the crosstalk components and stores an update result in the memory in a case where a predetermined condition including time at which at least one talker talks is satisfied, and a sound output controller that outputs the sound signals, acquired by subtracting the crosstalk components by the filter from the talker sound signals based on the update result, from a speaker.
    Type: Grant
    Filed: January 24, 2020
    Date of Patent: August 10, 2021
    Assignee: PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD.
    Inventors: Masanari Miyamoto, Hiromasa Ohashi, Naoya Tanaka
  • Patent number: 11074927
    Abstract: A computer implemented method, computer system and computer program product are provided for acoustic event detection in polyphonic acoustic data, according to the method, polyphonic acoustic data is inputted by one or more processing units into a trained neural network trained by labeled monophonic acoustic data, a first output from a hidden layer of the trained neural network is obtained by one or more processing units, and at least one acoustic classification of the polyphonic acoustic data is determined by one or more processing units based on the first output and a feature dictionary learnt from the trained neural network.
    Type: Grant
    Filed: October 31, 2017
    Date of Patent: July 27, 2021
    Assignee: International Business Machines Corporation
    Inventors: Xiao Xing Liang, Ning Zhang, Yu Ling Zheng, Yu Chen Zhou
  • Patent number: 11069353
    Abstract: A system and method performs multilingual wakeword detection by determining a language corresponding to the wakeword. A first wakeword-detection component, which may execute using a digital-signal processor, determines that audio data includes a representation of the wakeword and determines a language corresponding to the wakeword. A second, more accurate wakeword-detection component may then process the audio data using the language to confirm that it includes the representation of the wakeword. The audio data may then be sent to a remote system for further processing.
    Type: Grant
    Filed: May 6, 2019
    Date of Patent: July 20, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Yixin Gao, Ming Sun, Jason Krone, Shiv Naga Prasad Vitaladevuni, Yuzong Liu