Specialized Models Patents (Class 704/250)
  • Patent number: 11443750
    Abstract: A method and apparatus for authenticating a user based on an utterance input includes obtaining an input audio signal based on the utterance input of the user; obtaining, from the input audio signal, at least one audio signal of an utterance section and at least one audio signal of a non-utterance section; generating environment information indicating an environment in which the utterance input is received, based on the at least one audio signal of the non-utterance section; obtaining a result of a comparison between the generated environment information and registration environment information indicating an environment in which a registration utterance input corresponding to a previously registered registration audio signal corresponding to the user is received; adjusting an authentication criterion for authenticating the user based on the result of the comparison; and authenticating the user based on the adjusted authentication criterion and the input audio signal.
    Type: Grant
    Filed: December 2, 2019
    Date of Patent: September 13, 2022
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Keunseok Cho, Jaeyoung Roh, Donghan Jang, Jiwon Hyung, Jaewon Lee
  • Patent number: 11417327
    Abstract: An electronic apparatus is provided. The electronic device includes: a storage configured to store recognition related information and misrecognition related information of a trigger word for entering a speech recognition mode; and a processor configured to identify whether or not the speech recognition mode is activated on the basis of characteristic information of a received uttered speech and the recognition related information, identify a similarity between text information of the received uttered speech and text information of the trigger word, and update at least one of the recognition related information or the misrecognition related information on the basis of whether or not the speech recognition mode is activated and the similarity.
    Type: Grant
    Filed: November 27, 2019
    Date of Patent: August 16, 2022
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventor: Chanhee Choi
  • Patent number: 11410645
    Abstract: A user device configured to perform wake-up word detection in a target language.
    Type: Grant
    Filed: November 10, 2016
    Date of Patent: August 9, 2022
    Assignee: Cerence Operating Company
    Inventors: Xiao-Lin Ren, Jianzhong Teng
  • Patent number: 11404067
    Abstract: A method of operating an electronic device and an electronic device thereof are provided. The method includes receiving a first voice signal of a first user, authenticating whether the first user has authority to control the electronic device, based on the first voice signal, and determining an instruction corresponding to the first voice signal based on an authentication result and controlling the electronic device according to the instruction. The electronic device includes a receiver configured to receive a first voice signal of a first user and at least one processor configured to authenticate whether the first user has authority to control the electronic device based on the first voice signal, determine an instruction corresponding to the first voice signal, and control the electronic device according to the instruction.
    Type: Grant
    Filed: June 15, 2020
    Date of Patent: August 2, 2022
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Anas Toma, Ahmad Abu Shariah, Hadi Jadallah
  • Patent number: 11335353
    Abstract: A score indicating a likelihood that a first subject is the same as a second subject may be calibrated to compensate for aging of the first subject between samples of age-sensitive biometric characteristics. Age of the first subject obtained at a first sample time and age of the second subject obtained at a second sample time may be averaged, and an age approximation may be generated based on at least the age average and an interval between the first and second samples. The age approximation, the interval between the first and second sample times, and an obtained gender of the subject are used to calibrate the likelihood score.
    Type: Grant
    Filed: June 1, 2020
    Date of Patent: May 17, 2022
    Assignee: PINDROP SECURITY, INC.
    Inventors: Elie Khoury, Matthew Garland
  • Patent number: 11328711
    Abstract: A user adaptive conversation apparatus generating a talk for a conversation based on emotional and ethical states of a user. A voice recognition unit converts a talk of the user in a conversational situation into a natural language script form to generate talk information. An artificial visualization unit generates situation information by recognizing talking situation from a video and generates intention information indicating an intention of the talk. A natural language analysis unit converts the situation information and the intention information into the natural language script form. A natural language analysis unit analyzes the talk information, the intention information, and the situation information.
    Type: Grant
    Filed: July 5, 2019
    Date of Patent: May 10, 2022
    Assignee: KOREA ELECTRONICS TECHNOLOGY INSTITUTE
    Inventors: Saim Shin, Hyedong Jung, Jinyea Jang
  • Patent number: 11322155
    Abstract: A method and apparatus for establishing a voiceprint model, a computer device, and a storage medium are described herein. The method includes: collecting speech acoustic features in a speech signal to form a plurality of cluster structures; calculating an average value and a standard deviation of the plurality of cluster structures and then performing coordinate transformation and activation function calculation to obtain a feature vector; and obtaining a voiceprint model based on the feature vector.
    Type: Grant
    Filed: July 6, 2018
    Date of Patent: May 3, 2022
    Assignee: PING AN TECHNOLOGY (SHENZHEN) CO., LTD.
    Inventors: Yuanzhe Cai, Jianzong Wang, Ning Cheng, Jing Xiao
  • Patent number: 11233490
    Abstract: A method includes detecting an input that triggers a virtual assistance (VA) on an electronic downdevice (ED) to perform a task that includes outputting audio content through a speaker associated with the ED. The method includes identifying a type of the audio content to be outputted through the speaker. The method includes determining whether a registered user of the ED is present in proximity to the ED. Each registered user is associated with a unique user identifier. The method includes, in response to determining that no registered user is present in proximity to the ED, outputting the audio content via the speaker at a current volume level of the ED. The method includes in response to determining that a registered user is in proximity to the ED, outputting the audio content at a selected, preferred volume level based on pre-determined or pre-established volume preference settings of the registered user.
    Type: Grant
    Filed: November 21, 2019
    Date of Patent: January 25, 2022
    Assignee: Motorola Mobility LLC
    Inventors: Wendel Malta De Assis, Amit Kumar Agrawal, Lucie Corinne Evelyne Papelier
  • Patent number: 11200903
    Abstract: A method of speaker verification comprises receiving an audio signal representing speech. While the audio signal is being received, features of the received audio signal are extracted. The extracted features, of at least a part of the received audio signal corresponding to the speech of at least one speaker, are summarised, and the summarised extracted features are stored. In response to a request for a speaker verification process relating to at least one enrolled user, the speaker verification process is performed using the previously summarised features.
    Type: Grant
    Filed: February 4, 2020
    Date of Patent: December 14, 2021
    Assignee: Cirrus Logic, Inc.
    Inventors: Carlos Vaquero Avilés-Casco, Michael Page
  • Patent number: 11194968
    Abstract: The present invention concerns a text analysis system, the text analysis system being adapted for utilizing a topic model to provide a document representation. The topic model is based on learning performed on a text corpus utilizing hidden layer representations associated to words of the text corpus, wherein each hidden layer representation pertains to a specific word of the text corpus and is based on a word environment including words occurring before and after the specific word in a text of the text corpus.
    Type: Grant
    Filed: May 31, 2018
    Date of Patent: December 7, 2021
    Assignee: SIEMENS AKTIENGESELLSCHAFT
    Inventors: Florian Büttner, Pankaj Gupta
  • Patent number: 11176950
    Abstract: Disclosed herein are an apparatus and method for recognizing a voice speaker. The apparatus for recognizing a voice speaker includes a voice feature extraction unit configured to extract a feature vector from a voice signal inputted through a microphone; and a speaker recognition unit configured to calculate a speaker recognition score by selecting a reverberant environment from multiple reverberant environment learning data sets based on the feature vector extracted by the voice feature extraction unit and to recognize a speaker by assigning a weight depending on the selected reverberant environment to the speaker recognition score.
    Type: Grant
    Filed: March 20, 2019
    Date of Patent: November 16, 2021
    Assignee: Hyundai Mobis Co., Ltd.
    Inventors: Yu Jin Jung, Ki Hee Park, Chang Won Lee, Doh Hyun Kim, Tae Kyung Kim, Tae Yoon Son, Joon Hyuk Chang, Joon Young Yang
  • Patent number: 11170131
    Abstract: Techniques for ensuring differential privacy in top-K selection are provided. In one technique, multiple items and multiple counts are identified in response to a query. For each count, which corresponds to a different item, a noise value is generated and added to the count to generate a noisy value, and the noisy value is added to a set of noisy values that is initially empty. A particular noise value is generated for a particular count and added to the particular count to generate a noisy threshold. The particular noise value is generated using a different technique than the technique used to generate each noise value in the set. Based on the noisy threshold, a subset of the noisy values is identified, where each noisy value in the subset is less than the noisy threshold. A response to the query is generated that excludes items that correspond to the subset.
    Type: Grant
    Filed: July 31, 2019
    Date of Patent: November 9, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Ryan M. Rogers, David Anthony Durfee, Sean S. Peng, Ya Xu
  • Patent number: 11164587
    Abstract: A personal assistant operation is provided for teaching a personal assistant device names preferred by the user for sensor activated devices. For this purpose, a method includes the personal assistant device receiving a request from a user to activate a requested device which the user has identified with a requested name which is unrecognized by the personal assistant device, determining a most likely candidate device from a list of candidate devices to activate in response to the request, activating the determined most likely candidate device, and identifying and saving the requested name as the name of the most likely candidate device in response to receiving confirmation from the user that the determined most likely candidate device is the requested device.
    Type: Grant
    Filed: January 15, 2019
    Date of Patent: November 2, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Bret A. Greenstein, Lisa Seacat DeLuca, Amit Fisher, Steven R. Abrams
  • Patent number: 11120792
    Abstract: An electronic device and method are disclosed. The electronic device includes a communication interface, a microphone, a speaker, a processor and a memory. The processor executes the method, including receiving a first message through a communication interface from an external device, in response to receiving a first user input associated with the first message through an input interface including a microphone, converting first text included in the first message into speech, outputting the speech through an output interface of the electronic device including a speaker, determining whether to execute an additional operation associated with the first message, based on at least one of a state of the first electronic device, the first message, and a second user input received via the microphone, and initiate a conversational mode for executing the additional operation based at least partly on the determination.
    Type: Grant
    Filed: March 8, 2019
    Date of Patent: September 14, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Jihyun Kim, Gajin Song, Dongho Jang, Sangwook Kang, Hyunwoo Kang, Soojung Choi
  • Patent number: 11080591
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing sequences using convolutional neural networks. One of the methods includes, for each of the time steps: providing a current sequence of audio data as input to a convolutional subnetwork, wherein the current sequence comprises the respective audio sample at each time step that precedes the time step in the output sequence, and wherein the convolutional subnetwork is configured to process the current sequence of audio data to generate an alternative representation for the time step; and providing the alternative representation for the time step as input to an output layer, wherein the output layer is configured to: process the alternative representation to generate an output that defines a score distribution over a plurality of possible audio samples for the time step.
    Type: Grant
    Filed: September 6, 2017
    Date of Patent: August 3, 2021
    Assignee: DeepMind Technologies Limited
    Inventors: Aaron Gerard Antonius van den Oord, Sander Etienne Lea Dieleman, Nal Emmerich Kalchbrenner, Karen Simonyan, Oriol Vinyals, Lasse Espeholt
  • Patent number: 11068735
    Abstract: A reliability calculation apparatus calculates a reliability of a recognized object that is recognized as being present on a route on which a vehicle travels. The reliability of the recognized object is used for driving control of the vehicle. In the reliability calculation apparatus, a provisional reliability calculating unit determines a feature quantity of each of a plurality of recognized objects, selects a plurality of pairs of recognized objects from the plurality of recognized objects, determines a difference in feature quantity of each pair of recognized objects, and calculates a reliability that decreases as the difference in feature quantity increases as a provisional reliability of each pair of recognized objects. An integrated reliability calculating unit calculates an integrated reliability of each recognized object from the provisional reliabilities of the plurality of pairs of recognized objects.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: July 20, 2021
    Assignee: DENSO CORPORATION
    Inventors: Masaya Okada, Shunya Kumano, Hiroaki Niino, Mitsuhiro Tokimasa, Takumi Uematsu
  • Patent number: 11038824
    Abstract: The method includes receiving sender media that was recorded by a sender device associated with a sender. The method further comprises playing, by a recipient device, the sender media for a recipient. The method further comprises detecting that the recipient is speaking. The method further comprises recording recipient media based on detecting that the recipient is speaking. The method further comprises determining a location in the sender media at which the recipient media is to be included. The method further comprises generating combined media that includes at least a portion of the sender media and the recipient media at the location.
    Type: Grant
    Filed: September 13, 2018
    Date of Patent: June 15, 2021
    Assignee: Google LLC
    Inventor: Matthew Sharifi
  • Patent number: 11010562
    Abstract: One embodiment provides a method, including: receiving a story, wherein the story is received in a text-based format; identifying conversational blocks contained within the story, wherein the identifying conversational blocks comprises (i) identifying conversational turns within the story, (ii) identifying topic switches between conversational turns, and (iii) grouping conversational turns occurring between topic switches into a conversational block; for each of the conversational blocks, obtaining at least one image illustrating the corresponding conversational block, wherein the obtaining comprises (i) extracting concepts from the conversational block and (ii) identifying images illustrating the concepts; and generating a visual storyline representing the story by ordering the images obtained for each of the conversational blocks, based upon an order of the conversational blocks within the story, wherein the generating comprises refining the images to maintain consistency between the images.
    Type: Grant
    Filed: February 8, 2019
    Date of Patent: May 18, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Smitkumar Narotambhai Marvaniya, Padmanabha Venkatagiri Seshadri, Vijay Ekambaram
  • Patent number: 10958466
    Abstract: Methods and apparatuses for environmental control systems are described. In one example, a method includes monitoring a voice communications call between a local call participant located in a building space and a remote call participant. The method includes detecting a change in a user state of the local call participant from the voice communications call. The method further includes adjusting an environmental parameter in the building space responsive to detecting the change in the user state of the local call participant.
    Type: Grant
    Filed: May 3, 2018
    Date of Patent: March 23, 2021
    Assignee: Plantronics, Inc.
    Inventors: Marcus Dennis Rose, Evan Harris Benway
  • Patent number: 10923130
    Abstract: An electronic device and method are disclosed, including a communication circuit, a speaker, a microphone and a processor. The processor implements the method, including receiving a first voice input through the microphone, executing first voiceprint authentication on the first voice input including determining whether the first voice input matches voice information corresponding to a user stored in the electronic device, when the first voice input does not match the voice information, transmit a request message to at least one external electronic device for execution of a second voiceprint authentication, receive a response message indicating whether the first voice input is authenticated under the second voiceprint authentication, receive a second voice input through the microphone, and transmit second data related to the second voice input to an external server through the communication circuit for Automatic Speech Recognition (ASR) or Natural Language Understanding (NLU) on the second data.
    Type: Grant
    Filed: February 8, 2019
    Date of Patent: February 16, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Gyucheol Choi, Youngkyoo Kim
  • Patent number: 10884503
    Abstract: Methods, computing devices, and computer-program products are provided for implementing a virtual personal assistant. In various implementations, a virtual personal assistant can be configured to receive sensory input, including at least two different types of information. The virtual personal assistant can further be configured to determine semantic information from the sensory input, and to identify a context-specific framework. The virtual personal assistant can further be configured to determine a current intent. Determining the current intent can include using the semantic information and the context-specific framework. The virtual personal assistant can further be configured to determine a current input state. Determining the current input state can include using the semantic information and one or more behavioral models. The behavioral models can include one or more interpretations of previously-provided semantic information.
    Type: Grant
    Filed: October 24, 2016
    Date of Patent: January 5, 2021
    Assignee: SRI International
    Inventors: Ajay Divakaran, Amir Tamrakar, Girish Acharya, William Mark, Greg Ho, Jihua Huang, David Salter, Edgar Kalns, Michael Wessel, Min Yin, James Carpenter, Brent Mombourquette, Kenneth Nitz, Elizabeth Shriberg, Eric Law, Michael Frandsen, Hyong-Gyun Kim, Cory Albright, Andreas Tsiartas
  • Patent number: 10733230
    Abstract: Approaches presented herein enable automatic creation of metadata for contents of a video. More specifically, a video and a script corresponding to the video are obtained. A location corresponding to an object in at least one shot of the video is extracted. This at least one shot includes a series of adjacent frames. The extracted location is saved as an annotation area in an annotation knowledge base. An element of a plot of the video is extracted from the script. This element of the plot is derived from content of the video in combination with content of the script. The extracted element of the plot is saved in a narrative knowledge base.
    Type: Grant
    Filed: December 31, 2018
    Date of Patent: August 4, 2020
    Assignee: INHA UNIVERSITY RESEARCH AND BUSINESS FOUNDATION
    Inventor: Geun Sik Jo
  • Patent number: 10726848
    Abstract: Disclosed herein are methods of diarizing audio data using first-pass blind diarization and second-pass blind diarization that generate speaker statistical models, wherein the first pass-blind diarization is on a per-frame basis and the second pass-blind diarization is on a per-word basis, and methods of creating acoustic signatures for a common speaker based only on the statistical models of the speakers in each audio session.
    Type: Grant
    Filed: January 22, 2018
    Date of Patent: July 28, 2020
    Assignee: VERINT SYSTEMS LTD.
    Inventors: Alex Gorodetski, Oana Sidi, Ron Wein, Ido Shapira
  • Patent number: 10692503
    Abstract: A voice data processing method and apparatus are provided. The method includes obtaining an I-Vector vector of each of voice samples, and determining a target seed sample in the voice samples. A first cosine distance is calculated between an I-Vector vector of the target seed sample and an I-Vector vector of a target remaining voice sample, where the target remaining voice sample is a voice sample other than the target seed sample in the voice samples. A target voice sample is filtered from the voice samples or the target remaining voice sample according to the first cosine distance, to obtain a target voice sample whose first cosine distance is greater than a first threshold.
    Type: Grant
    Filed: March 3, 2017
    Date of Patent: June 23, 2020
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Xingming Jin, Wei Li, Fangmai Zheng, Fuzhang Wu, Bilei Zhu, Binghua Qian, Ke Li, Yongjian Wu, Feiyue Huang
  • Patent number: 10685657
    Abstract: Systems and methods for analyzing digital recordings of the human voice in order to find characteristics unique to an individual. A biometrics engine may use an analytics service in a contact center to supply audio streams based on configured rules and providers for biometric detection. The analytics service may provide call audio data and attributes to connected engines based on a provider-set of selection rules. The connected providers send call audio data and attributes through the analytics service. The engines are notified when a new call is available for processing and can then retrieve chunks of audio data and call attributes by polling an analytics service interface. A mathematical model of the human vocal tract in the call audio data is created and/or matched against existing models. The result is analogous to a fingerprint, i.e., a pattern unique to an individual to within some level of probability.
    Type: Grant
    Filed: February 15, 2019
    Date of Patent: June 16, 2020
    Assignee: VERINT AMERICAS INC
    Inventors: Jamie Richard Williams, Robert John Barnes, Ian Linsdell, Scott M. Bluman
  • Patent number: 10685666
    Abstract: A mechanism is described for facilitating automatic gain adjustment in audio systems according to one embodiment. A method of embodiments, as described herein, includes determining status of one or more of gain settings, mute settings, and boost settings associated with one or more microphones based on a configuration of a computing device including a voice-enabled device. The method may further comprise recommending adjustment of microphone gain based on the configuration and the status of one or more of the gain, mute, and boost settings, and applying the recommended adjustment of the microphone gain.
    Type: Grant
    Filed: April 6, 2018
    Date of Patent: June 16, 2020
    Assignee: INTEL CORPORATION
    Inventors: Przemyslaw Maziewski, Adam Kupryjanow, Lukasz Kurylo, Pawel Trella
  • Patent number: 10666594
    Abstract: Non-limiting examples of the present disclosure describe proactive action by an intelligent personal assistant application/service to improve functionality of one or more applications. In one example, an intelligent personal assistant service may interface with a messaging application to analyze a message thread within the messaging application. The intelligent personal assistant service may analyze the message thread by evaluating context of message content within the message thread. Analysis of the message thread may occur proactively without requiring an explicit request for assistance from a user of a processing device. In response to the analyzing of the message thread, the intelligent personal assistant service may proactively provide a cue that includes content retrieved by the intelligent personal assistant service. An input may be received to include the cue within the message thread. In response to receiving the input, the cue may be displayed within the message thread.
    Type: Grant
    Filed: November 13, 2018
    Date of Patent: May 26, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Zachary Adam Pfriem, Mark Jozef Vitazko, Jared Frew, Jason Michael Nelson
  • Patent number: 10650055
    Abstract: A wearable sound capturing and retrieval system that includes a wearable sound capturing device that comprises a data collection device including at least one microphone configured for capturing sound data adjacent a user in at least a substantially continuous manner. The system may, for example: (1) store the captured sound data; (2) convert the captured sound data to captured textual data; (3) index data selected from: one or more segments of captured sound data and one or more segments of captured textual data; and (3) facilitate retrieval of at least a portion of the indexed data, wherein facilitating the retrieval includes (I) scanning the indexed data to identify one or more key phrases, (ii) retrieving one or more segments of indexed data that was communicated by the user at least about contemporaneously with the one or more key phrases, and (iii) saving the one or more segments of indexed data.
    Type: Grant
    Filed: February 8, 2019
    Date of Patent: May 12, 2020
    Assignee: Viesoft, Inc.
    Inventor: Anthony Vierra
  • Patent number: 10621985
    Abstract: A voice recognition device for a vehicle includes: an input device receiving a command and a negative interjection uttered by a user, converting the command into a first voice signal, and converting the negative interjection into a second voice signal; a storage device storing a negative context, an interjection context, and an acoustic model; and a control device receiving the first voice signal, detecting a first start point and a first end point of the first voice signal, receiving the second voice signal after the detection of the first start point and the first end point of the first voice signal, detecting a second start point and a second end point of the second voice signal, and recognizing the second voice signal based on at least one of the negative context, the interjection context, and the acoustic model when the reception of the first voice signal and the second voice signal is completed.
    Type: Grant
    Filed: May 29, 2018
    Date of Patent: April 14, 2020
    Assignees: Hyundai Motor Company, Kia Motors Corporation
    Inventor: Jae Min Joh
  • Patent number: 10565080
    Abstract: A method for monitoring a condition of a system or process includes acquiring sensor data from a plurality of sensors disposed within the system (S41 and S44). The acquired sensor data is streamed in real-time to a computer system (S42 and S44). A discriminative framework is applied to the streaming sensor data using the computer system (S43 and S45). The discriminative framework provides a probability value representing a probability that the sensor data is indicative of an anomaly within the system. The discriminative framework is an integration of a Kalman filter with a logistical function (S41).
    Type: Grant
    Filed: June 11, 2013
    Date of Patent: February 18, 2020
    Assignee: SIEMENS AKTIENGESELLSCHAFT
    Inventors: Chao Yuan, Amit Chakraborty, Holger Hackstein, Leif Wiebking
  • Patent number: 10515655
    Abstract: Techniques for selecting an emotion type code associated with semantic content in an interactive dialog system. In an aspect, fact or profile inputs are provided to an emotion classification algorithm, which selects an emotion type based on the specific combination of fact or profile inputs. The emotion classification algorithm may be rules-based or derived from machine learning. A previous user input may be further specified as input to the emotion classification algorithm. The techniques are especially applicable in mobile communications devices such as smartphones, wherein the fact or profile inputs may be derived from usage of the diverse function set of the device, including online access, text or voice communications, scheduling functions, etc.
    Type: Grant
    Filed: September 4, 2017
    Date of Patent: December 24, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Edward Un, Max Leung
  • Patent number: 10511585
    Abstract: Techniques are provided for smoothing discretized values used, for example, for authentication or identity assurance. An illustrative method comprises obtaining at least one probability of transitioning between at least two discretized values of a given feature; computing a smoothed feature score for the given feature for a transition from a first one of the discretized values to a second one of the discretized values based on the probability of the transition from the first discrete value to the second discrete value; and performing one or more of authenticating a user and verifying an identity of a user based at least in part on the smoothed feature score. The probabilities of transitioning between the discretized values are optionally stored in a transition matrix. Feature scores for first and second discretized values are optionally weighted based on the probability of the transition from the first discrete value to the second discrete value.
    Type: Grant
    Filed: April 27, 2017
    Date of Patent: December 17, 2019
    Assignee: EMC IP Holding Company LLC
    Inventors: Shay Amram, Carmit Sahar, Anatoly Gendelev, Idan Achituve
  • Patent number: 10475441
    Abstract: A voice end-point detection device, a system and a method are provided. The voice end-point detection system includes a processor that is configured to determine an end-point detection time to detect an end-point of speaking of a user that varies vary for each user and for each domain. The voice end-point detection system is configured to perform voice recognition and a database (DB) is configured to store data for the voice recognition by the processor.
    Type: Grant
    Filed: November 13, 2017
    Date of Patent: November 12, 2019
    Assignees: Hyundai Motor Company, Kia Motors Corporation
    Inventors: Kyung Chul Lee, Jae Min Joh
  • Patent number: 10474946
    Abstract: Methods, systems, apparatuses, and computer program products are provided for altering the behavior of an electronic personal assistant based on a situation associated with a mobile device. A situation is sensed with a plurality of sensors to generate sensor data. A situation score is calculated based on the sensor data. Behavior of an electronic personal assistant is altered based on the calculated situation score. In one aspect, the situation is a driving situation in which a driver drives a vehicle on a roadway. In such case, a driving situation score is calculated based on the sensor data, and behavior of the electronic personal assistant is altered based on the calculated driving situation score, such as suspending interactions by the electronic personal assistant with the driver to avoid the driver being distracted.
    Type: Grant
    Filed: June 24, 2016
    Date of Patent: November 12, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Tae Hong Joo, Tarek Z. Elabbady, Mona Soliman Habib
  • Patent number: 10366693
    Abstract: Disclosed herein are methods of diarizing audio data using first-pass blind diarization and second-pass blind diarization that generate speaker statistical models, wherein the first pass-blind diarization is on a per-frame basis and the second pass-blind diarization is on a per-word basis, and methods of creating acoustic signatures for a common speaker based only on the statistical models of the speakers in each audio session.
    Type: Grant
    Filed: January 22, 2018
    Date of Patent: July 30, 2019
    Assignee: Verint Systems LTD.
    Inventors: Alex Gorodetski, Ido Shapira, Ron Wein, Oana Sidi
  • Patent number: 10304445
    Abstract: A wearable utterance training system includes a wearable utterance training device. The system may, for example: (1) receive one or more target utterances from the user; (2) detect a use of one of the one or more target utterances by the user; and (3) in response, provide one or more responsive effects. The one or more responsive effects may include, for example: (1) providing one or more shocks to the user using the wearable utterance training device; (2) initiating a transfer of money between an account associated with the user and a third party account; (3) creating a public disclosure of the utterance (e.g., by posting the disclosure on one or more social media websites) and/or (4) playing a recording of the user's use of the target utterance or other sound.
    Type: Grant
    Filed: October 13, 2016
    Date of Patent: May 28, 2019
    Assignee: Viesoft, Inc.
    Inventor: Anthony Vierra
  • Patent number: 10304460
    Abstract: According to an embodiment, a conference support system includes a recognizer, a classifier, a first caption controller, a second caption controller, and a display controller. The recognizer is configured to recognize text data corresponding speech from a speech section and configured to distinguish between the speech section and a non-speech section in speech data. The classifier is configured to classify the text data into first utterance data representing a principal utterance and second utterance data representing another utterance. The first caption controller is configured to generate first caption data for displaying the first utterance data without waiting for identification of the first utterance data to finish. The second caption controller is configured to generate second caption data for displaying the second utterance data after identification of the second utterance data finishes. The display controller is configured to control a display of the first caption data and the second caption data.
    Type: Grant
    Filed: February 23, 2017
    Date of Patent: May 28, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Taira Ashikawa, Kosei Fume, Masayuki Ashikawa, Hiroshi Fujimura
  • Patent number: 10276149
    Abstract: Systems, methods, and devices for dynamically outputting TTS content are disclosed. A speech-controlled device captures a spoken command, and sends audio data corresponding thereto to a server(s). The server(s) determines output content responsive to the spoken command. The server(s) may also determine a user that spoke the command and determine an average speech characteristic (e.g., tone, pitch, speed, number of words, etc.) used by the user when speaking commands. The server(s) may also determine a speech characteristic of the presently spoken command, as well as determine a difference between the speech characteristic of the presently spoken command and the average speech characteristic of the user. The server(s) may then cause the speech-controlled device to output audio based on the difference.
    Type: Grant
    Filed: December 21, 2016
    Date of Patent: April 30, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Nancy Yi Liang, Aaron Takayanagi Barnet
  • Patent number: 10269345
    Abstract: This relates to systems and processes for operating an automated assistant to process messages. In one example process, an electronic device receives a communication including a text string and determines whether a portion of the text string is associated with a data type of a plurality of data types. The data type is associated with at least one task. In accordance with a determination that the portion of the text string is associated with the data type, the electronic device receives a user input indicative of a task of the at least one task, and in response, causes the task to be performed based on the portion of the text string. In accordance with a determination that the portion of the text string is not associated with the data type, the electronic device foregoes causing the task to be performed based on the portion of the text string.
    Type: Grant
    Filed: September 19, 2016
    Date of Patent: April 23, 2019
    Assignee: Apple Inc.
    Inventors: Jose A. Castillo Sanchez, Garett R. Nell, Kimberly D. Beverett
  • Patent number: 10249314
    Abstract: A voice conversion system for generating realistic, natural-sounding target speech is disclosed. The voice conversion system preferably comprises a neural network for converting the source speech data to estimated target speech data; a global variance correction module; a modulation spectrum correction module; and a waveform generator. The global variance correction module is configured to scale and shift (or normalize and de-normalize) the estimated target speech based on (i) a mean and standard deviation of the source speech data, and further based on (ii) a mean and standard deviation of the estimated target speech data. The modulation spectrum correction module is configured to apply a plurality of filters to the estimated target speech data after it has been scaled and shifted by the global variance correction module. Each filter is designed to correct the trajectory representing the curve of one MCEP coefficient over time.
    Type: Grant
    Filed: July 21, 2017
    Date of Patent: April 2, 2019
    Assignee: OBEN, INC.
    Inventor: Sandesh Aryal
  • Patent number: 10176163
    Abstract: Embodiments herein include a natural language computing system that provides a diagnosis for a participant in the conversation which indicates the likelihood that the participant exhibited a symptom of autism. To provide the diagnosis, the computing system includes a diagnosis system that performs a training process to generate a machine learning model which is then used to evaluate a textual representation of the conversation. For example, the diagnosis system may receive one or more examples of baseline conversations that exhibit symptoms of autisms and those that do not. The diagnosis system may annotate and the baseline conversations and identify features that are used to identify the symptoms of autism. The system generates a machine learning model that weights the features according to whether the identified features are, or are not, an indicator of autism.
    Type: Grant
    Filed: December 19, 2014
    Date of Patent: January 8, 2019
    Assignee: International Business Machines Corporation
    Inventors: Adam T. Clark, Brian J. Cragun, Anthony W. Eichenlaub, John E. Petri, John C. Unterholzner
  • Patent number: 10158593
    Abstract: Non-limiting examples of the present disclosure describe proactive action by an intelligent personal assistant application/service to improve functionality of one or more applications. In one example, an intelligent personal assistant service may interface with a messaging application to analyze a message thread within the messaging application. The intelligent personal assistant service may analyze the message thread by evaluating context of message content within the message thread. Analysis of the message thread may occur proactively without requiring an explicit request for assistance from a user of a processing device. In response to the analyzing of the message thread, the intelligent personal assistant service may proactively provide a cue that includes content retrieved by the intelligent personal assistant service. An input may be received to include the cue within the message thread. In response to receiving the input, the cue may be displayed within the message thread.
    Type: Grant
    Filed: April 8, 2016
    Date of Patent: December 18, 2018
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Zachary Adam Pfriem, Mark Jozef Vitazko, Jared Frew, Jason Michael Nelson
  • Patent number: 10102760
    Abstract: This disclosure is directed to a system includes a receiver configured to receive audio data from a vehicle. In some examples, the system includes processing circuitry configured to determine an expected maneuver for the vehicle based on the audio data. In some examples, the processing circuitry is further configured to determine whether to output an alert based on the expected maneuver determined from the audio data.
    Type: Grant
    Filed: August 23, 2017
    Date of Patent: October 16, 2018
    Assignee: Honeywell International Inc.
    Inventors: Stanislav Foltan, Robert Sosovicka, Eva Josth Adamova
  • Patent number: 10048079
    Abstract: A destination determination device for a vehicle includes: a communication unit that performs a wireless communication with a mobile terminal used by an occupant in the vehicle and having a destination search function; a search condition acquisition unit that acquires destination search conditions from the mobile terminal through the communication unit; a search unit that performs a destination search based on an AND search of a combined search condition in which a plurality of acquired destination search conditions are combined together when the search condition acquisition unit acquires the plurality of destination search condition; and a search result output unit that outputs a search result of the destination search performed by the search unit under the combined search condition.
    Type: Grant
    Filed: June 2, 2015
    Date of Patent: August 14, 2018
    Assignee: DENSO CORPORATION
    Inventors: Takamitsu Suzuki, Takahira Katoh, Takeshi Yamamoto, Yuuko Nakamura
  • Patent number: 9967724
    Abstract: A method and apparatus for changing a persona of a digital assistant is provided herein. During operation a digital assistant will determine a public-safety incident type and then change its persona based on the public-safety incident type.
    Type: Grant
    Filed: May 8, 2017
    Date of Patent: May 8, 2018
    Assignee: MOTOROLA SOLUTIONS, INC.
    Inventors: Guo Dong Gan, Kong Yong Foo, Mun Yew Tham, Bing Qin Lim
  • Patent number: 9916830
    Abstract: Techniques for altering audio being output by a voice-controlled device, or another device, to enable more accurate automatic speech recognition (ASR) by the voice-controlled device. For instance, a voice-controlled device may output audio within an environment using a speaker of the device. While outputting the audio, a microphone of the device may capture sound within the environment and may generate an audio signal based on the captured sound. The device may then analyze the audio signal to identify speech of a user within the signal, with the speech indicating that the user is going to provide a subsequent command to the device. Thereafter, the device may alter the output of the audio (e.g., attenuate the audio, pause the audio, switch from stereo to mono, etc.) to facilitate speech recognition of the user's subsequent command.
    Type: Grant
    Filed: January 13, 2016
    Date of Patent: March 13, 2018
    Assignee: Amazon Technologies, Inc.
    Inventors: Gregory Michael Hart, William Spencer Worley, III
  • Patent number: 9711145
    Abstract: The subject matter of this specification can be implemented in, among other things, a computer-implemented method for correcting words in transcribed text including receiving speech audio data from a microphone. The method further includes sending the speech audio data to a transcription system. The method further includes receiving a word lattice transcribed from the speech audio data by the transcription system. The method further includes presenting one or more transcribed words from the word lattice. The method further includes receiving a user selection of at least one of the presented transcribed words. The method further includes presenting one or more alternate words from the word lattice for the selected transcribed word. The method further includes receiving a user selection of at least one of the alternate words. The method further includes replacing the selected transcribed word in the presented transcribed words with the selected alternate word.
    Type: Grant
    Filed: November 14, 2016
    Date of Patent: July 18, 2017
    Assignee: Google Inc.
    Inventors: Michael J. LeBeau, William J. Byrne, John Nicholas Jitkoff, Brandon M. Ballinger, Trausti T. Kristjansson
  • Patent number: 9679569
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for a dynamic threshold for speaker verification are disclosed. In one aspect, a method includes the actions of receiving, for each of multiple utterances of a hotword, a data set including at least a speaker verification confidence score, and environmental context data. The actions further include selecting from among the data sets, a subset of the data sets that are associated with a particular environmental context. The actions further include selecting a particular data set from among the subset of data sets based on one or more selection criteria. The actions further include selecting, as a speaker verification threshold for the particular environmental context, the speaker verification confidence score. The actions further include providing the speaker verification threshold for use in performing speaker verification of utterances that are associated with the particular environmental context.
    Type: Grant
    Filed: November 3, 2016
    Date of Patent: June 13, 2017
    Assignee: Google Inc.
    Inventors: Jakob Nicolaus Foerster, Diego Melendo Casado
  • Patent number: 9495127
    Abstract: Methods, computer program products and systems are described for converting speech to text. Sound information is received at a computer server system from an electronic device, where the sound information is from a user of the electronic device. A context identifier indicates a context within which the user provided the sound information. The context identifier is used to select, from among multiple language models, a language model appropriate for the context. Speech in the sound information is converted to text using the selected language model. The text is provided for use by the electronic device.
    Type: Grant
    Filed: December 22, 2010
    Date of Patent: November 15, 2016
    Assignee: Google Inc.
    Inventors: Brandon M. Ballinger, Johan Schalkwyk, Michael H. Cohen, Cyril Georges Luc Allauzen
  • Patent number: 9465794
    Abstract: Disclose is a mobile terminal and control method thereof for inputting a voice to automatically generate a message to be sent during conversation using a mobile messenger, and it may include a microphone for inputting a user's voice, a display unit for displaying a mobile messenger; and a controller for inputting and recognizing a user's voice when a mobile messenger is implemented and then converting into a message to display the message on a message input window of the mobile messenger, and sending the displayed message to the other party which has been preset, and displaying the message sent to the other party and a message received from the other party in the sending and receiving order on a send/receive display window of the mobile messenger.
    Type: Grant
    Filed: May 17, 2010
    Date of Patent: October 11, 2016
    Assignee: LG ELECTRONICS INC.
    Inventors: Sun-Hwa Cha, Jong-Keun Youn