Patents Examined by Michelle M Koeth
  • Patent number: 11367029
    Abstract: A system and method are presented for adaptive skill level assignments of agents in contact center environments. A client and a service collaborate to automatically determine the effectiveness of an agent handling an interaction that has been routed using skills-based routing. Evaluation operations may be performed including emotion detection, transcription of audio to text, keyword analysis, and sentiment analysis. The results of the evaluation are aggregated with other information such as the interaction's duration, agent skills and agent skill levels, and call requirement skills and skill levels, to update the agent's profile which is then used for subsequent routing operations.
    Type: Grant
    Filed: February 26, 2020
    Date of Patent: June 21, 2022
    Inventors: James Murison, Johnson Tse, Gaurav Mehrotra, Anthony Lam
  • Patent number: 11361677
    Abstract: A computing device, method, and a non-transitory computer readable medium for articulation training for hearing impaired persons is disclosed. The computing device comprises a database including stored mel-frequency cepstral representations of audio recordings associated with text and/or images related to the audio recordings, a microphone configured to receive audible inputs and a display. The computing device is operatively connected to the database, the microphone and the display. The computing device includes circuitry and program instructions stored therein which when executed by one or more processors, cause the system to receive an audible input from the microphone, convert the audible input to a mel-frequency cepstral representation, search the database for a match of the mel-frequency cepstral representation to a stored mel-frequency cepstral representation and display the text and/or images related to the stored mel-frequency cepstral representation when the match is found.
    Type: Grant
    Filed: November 10, 2021
    Date of Patent: June 14, 2022
    Assignee: King Abdulaziz University
    Inventor: Wadee Saleh Ahmed Alhalabi
  • Patent number: 11355033
    Abstract: A method comprises inputting an audio signal into a machine learning circuit to compress the audio signal into a sequence of actuator signals. The machine learning circuit being trained by: receiving a training set of acoustic signals and pre-processing the training set of acoustic signals into pre-processed audio data. The pre-processed audio data including at least a spectrogram. The training further includes training the machine learning circuit using the pre-processed audio data. The neural network has a cost function based on a reconstruction error and a plurality of constraints. The machine learning circuit generates a sequence of haptic cues corresponding to the audio input. The sequence of haptic cues is transmitted to a plurality of cutaneous actuators to generate a sequence of haptic outputs.
    Type: Grant
    Filed: April 10, 2018
    Date of Patent: June 7, 2022
    Assignee: Meta Platforms, Inc.
    Inventors: Brian Alexander Knott, Venkatasiva Prasad Chakkabala
  • Patent number: 11348576
    Abstract: A system configured to process an incoming spoken utterance and to coordinate among multiple speechlet components to execute an action of the utterance, where a trained model considers user history and preference information to select the primary speechlet to execute the action as well as any intermediate speechlets that may be provide input data to the speechlet that will ultimately perform the action. The trained model may also consider current dialog information, feedback data, or other data when determining how to process a dialog.
    Type: Grant
    Filed: December 6, 2017
    Date of Patent: May 31, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Bradford Lynch, Adam D. Baran, Kevindra Pal Singh, Udai Sen Mody
  • Patent number: 11328709
    Abstract: A system for improving dysarthria speech intelligibility and method thereof, are provided. In the system, user only needs to provides a set of paired corpus including a reference corpus and a patient corpus, and a speech disordering module can automatically generate a new corpus completely synchronous with the reference corpus, and the new corpus can be used as a training corpus for training a dysarthria voice conversion model. The present invention does not need to use a conventional corpus alignment technology or a manual manner to perform pre-processing on the training corpus, so that manpower cost and time cost can be reduced, and synchronization of the training corpus can be ensured, thereby improving both training and conversion qualities of the voice conversion model.
    Type: Grant
    Filed: August 20, 2019
    Date of Patent: May 10, 2022
    Assignee: NATIONAL CHUNG CHENG UNIVERSITY
    Inventors: Tay-Jyi Lin, Ching-Hau Sung, Che-Chia Pai, Ching-Wei Yeh
  • Patent number: 11328714
    Abstract: Processing data for speech recognition by generating hypotheses from input data, assigning each hypothesis, a score according to a confidence level value and hypothesis ranking, executing a pass/fail grammar test against each hypothesis, generating replacement hypotheses according to grammar test failures, assigning each replacement hypothesis a score according to a number of hypothesis changes, and providing a set of hypotheses, wherein the set comprises at least one replacement hypotheses.
    Type: Grant
    Filed: January 2, 2020
    Date of Patent: May 10, 2022
    Assignee: International Business Machines Corporation
    Inventors: Andrew R. Freed, Marco Noel, Victor Povar
  • Patent number: 11322157
    Abstract: A method of speaker authentication comprises: receiving a speech signal; dividing the speech signal into segments; and, following each segment, obtaining an authentication score based on said segment and previously received segments, wherein the authentication score represents a probability that the speech signal comes from a specific registered speaker. In response to an authentication request, an authentication result is output based on the authentication score.
    Type: Grant
    Filed: June 6, 2017
    Date of Patent: May 3, 2022
    Assignee: Cirrus Logic, Inc.
    Inventors: Carlos Vaquero Avilés-Casco, David Martínez González, Ryan Roberts
  • Patent number: 11315552
    Abstract: This disclosure describes systems and techniques receiving a request for information from a user and, in response, outputting the requested information along with unsolicited, interesting content that is related to, yet nonresponsive to, the requested information. In some instances, if the requested information is unknown, the techniques may output an indication that the information is unknown, followed by the additional, unsolicited, interesting content.
    Type: Grant
    Filed: March 23, 2018
    Date of Patent: April 26, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Michael Martin George, David Garfield Uffelman, Deepak Maini, Kyle Beyer, Amarpaul Singh Sandhu
  • Patent number: 11302311
    Abstract: An artificial intelligence apparatus for recognizing speech of a user includes a microphone, and a processor configured to receive, via the microphone, a sound signal corresponding to the speech of the user, acquire personalize identification information corresponding to the speech, recognize the speech from the sound signal using a global language model, calculate a reliability for the recognition, and if the calculated reliability exceeds a predetermined first reference value, update a personalized language model corresponding to the personalize identification information using the recognition result.
    Type: Grant
    Filed: August 21, 2019
    Date of Patent: April 12, 2022
    Assignee: LG ELECTRONICS INC.
    Inventors: Boseop Kim, Jaehong Kim
  • Patent number: 11302312
    Abstract: A new model is introduced into a particular domain that receives a routing of a dialog from a speech processing component. A method associated with the model includes running a set of test utterances through the speech processing component that enables a spoken language dialog with a user to establish a base line score associated with processing for the set of test utterances. The speech processing component determines an intent of the user and routes the spoken language dialog to a network-based domain based on the intent. The method includes establishing an automatic test run of the set of test utterances to obtain a current score and, when a threshold associated with a difference between the current score and the base line score is breached, switching, at the network-based domain, from the false accept detection model to a second model.
    Type: Grant
    Filed: September 27, 2019
    Date of Patent: April 12, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Ajay Soni, Xi Chen, Jingqian Zhao, Liu Yang, Prathap Ramachandra, Ruiqi Luo
  • Patent number: 11295730
    Abstract: A method is described that includes processing text and speech from an input utterance using local overrides of default dictionary pronunciations. Applying this method, a word-level grammar used to process the tokens specifies at least one local word phonetic variant that applies within a specific production rule and, within a local context of the specific production rule, the local word phonetic variant overrides one or more default dictionary phonetic versions of the word. This method can be applied to parsing utterances where the pronunciation of some words depends on their syntactic or semantic context.
    Type: Grant
    Filed: August 1, 2019
    Date of Patent: April 5, 2022
    Assignee: SoundHound, Inc.
    Inventors: Keyvan Mohajer, Christopher Wilson, Bernard Mont-Reynaud
  • Patent number: 11257483
    Abstract: Spoken language understanding techniques include training a dynamic neural network mask relative to a static neural network using only post-deployment training data such that the mask zeroes out some of the weights of the static neural network and allows some other weights to pass through and applying a dynamic neural network corresponding to the masked static neural network to input queries to identify outputs for the queries.
    Type: Grant
    Filed: March 29, 2019
    Date of Patent: February 22, 2022
    Assignee: Intel Corporation
    Inventors: Krzysztof Czarnowski, Munir Georges
  • Patent number: 11257491
    Abstract: This application relates generally to modifying visual data based on audio commands and more specifically, to performing complex operations that modify visual data based on one or more audio commands. In some embodiments, a computer system may receive an audio input and identify an audio command based on the audio input. The audio command may be mapped to one or more operations capable of being performed by a multimedia editing application. The computer system may perform the one or more operations to edit to received multimedia data.
    Type: Grant
    Filed: November 29, 2018
    Date of Patent: February 22, 2022
    Assignee: ADOBE INC.
    Inventors: Sarah Kong, Yinglan Ma, Hyunghwan Byun, Chih-Yao Hsieh
  • Patent number: 11250219
    Abstract: Methods, computer program products, and systems are presented. The methods include, for instance: obtaining a style feed including a plurality of original works by an author. An author-style model for the author is built based on the style feed by use of a selected neural network, and a publication is generated in the style of the author based on the author-style model.
    Type: Grant
    Filed: May 10, 2019
    Date of Patent: February 15, 2022
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Laurence Plant, Stefan Harrer, Sean Rory Costello, James David Cleaver
  • Patent number: 11244123
    Abstract: A computer-implemented method may include obtaining, by processor, a first message composed in a first language and obtaining, by the processor, a translated first message. The translated first message may include a translation of the first message from the first language to a second language. The method may further include determining, by the processor, that the translated first message includes a translation-generated additional meaning, and notifying, by the processor, the first user of the determination.
    Type: Grant
    Filed: June 5, 2019
    Date of Patent: February 8, 2022
    Assignee: International Business Machines Corporation
    Inventors: Schayne Bellrose, Pasquale A. Catalano, Prach Jerry Chuaypradit, Andrew Gerald Crimmins, Preston Lane, Michael Lapointe, Francesca Wisniewski
  • Patent number: 11239871
    Abstract: The gain of an amplifier in a receiver operating in a cellular communication system is controlled by determining one or more gain variability metrics, which are then used to produce first and second threshold values. A frequency difference between a current carrier frequency and a target carrier frequency is ascertained and then compared to the threshold values. Target gain setting production is based on comparison results: If the frequency difference is larger than the first threshold, a first automatic gain control algorithm is performed; if the frequency difference is smaller than the first threshold and larger than the second threshold, a second automatic gain control algorithm is performed, wherein the second automatic gain control algorithm uses a current gain setting as a starting point; and if the frequency difference is smaller than both the first and second thresholds, the current gain setting is used as the target gain setting.
    Type: Grant
    Filed: April 11, 2016
    Date of Patent: February 1, 2022
    Assignee: TELEFONAKTIEBOLAGET LM ERICSSON (publ)
    Inventors: Peter Alriksson, Joakim Axmon
  • Patent number: 11227601
    Abstract: A computer-implement voice command authentication method is provided. The method includes obtaining a sound signal stream; calculating a Signal-to-Noise Ratio (SNR) value of the sound signal stream; converting the sound signal stream into a Mel-Frequency Cepstral Coefficients (MFCC) stream; calculating a Dynamic Time Warping (DTW) distance corresponding to the MFCC stream according to the MFCC stream and one of a plurality of sample streams generated by the Gaussian Mixture Model with Universal Background Model (GMM-UBM); calculating, according to the MFCC stream and the sample streams, a Log-likelihood ratio value corresponding to the MFCC stream as a GMM-UBM score; determining whether the sound signal stream passes a voice command authentication according to the GMM-UBM score, the DTW distance and the SNR value; in response to determining that the sound signal stream passes the voice command authentication, determining that the sound signal stream is a voice stream spoken from a legal user.
    Type: Grant
    Filed: September 21, 2019
    Date of Patent: January 18, 2022
    Assignee: Merry Electronics(Shenzhen) Co., Ltd.
    Inventors: Evelyn Kurniawati, Sasiraj Somarajan
  • Patent number: 11221823
    Abstract: A method includes receiving a voice input at an electronic device. An ambiguity of the voice input is determined. The ambiguity is resolved based on contextual data. The contextual data includes at least one of: an image, a non-voice input comprising a gesture, a pointer of a pointing device, a touch, or a combination thereof.
    Type: Grant
    Filed: December 28, 2017
    Date of Patent: January 11, 2022
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Arun Rakesh Yoganandan, Kumi Akiyoshi, Tais C. Mauk, Chang Long Zhu Jin, Jung Won Hong
  • Patent number: 11221822
    Abstract: Embodiments of the present disclosure disclose a method and apparatus for controlling a page. A specific embodiment of the method comprises: receiving voice information from a terminal and element information of at least one element in a displayed page; performing voice recognition on the voice information to acquire a voice recognition result, in response to determining the voice information being used for controlling the displayed page; matching the voice recognition result with the element content information of the at least one element; and generating page control information in response to determining successfully matching the voice recognition result with the element content information of the at least one element, and sending the page control information to the terminal to allow the terminal to control the displayed page based on the page control information.
    Type: Grant
    Filed: December 28, 2017
    Date of Patent: January 11, 2022
    Assignee: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD.
    Inventors: Yan Zhang, Binyuan Du, Fei Wang, Jing Li, Gaofei Cheng
  • Patent number: 11211061
    Abstract: Voice control in a multi-talker and multimedia environment is disclosed. In one aspect, there is provided a method comprising: receiving a microphone signal for each zone in a plurality of zones of an acoustic environment; generating a processed microphone signal for each zone in the plurality of zones of the acoustic environment, the generating including removing echo caused by audio transducers in the acoustic environment from each of the microphone signals, and removing interference from each of the microphone signals; and performing speech recognition on the processed microphone signals.
    Type: Grant
    Filed: January 7, 2019
    Date of Patent: December 28, 2021
    Assignee: 2236008 Ontario Inc.
    Inventors: Xueman Li, Mark Robert Every, Darrin Kenneth John Fry