Application Patents (Class 704/270)
  • Patent number: 11328716
    Abstract: When a confirmation response to a user speech is generated, a confirmation response capable of reducing a possibility of the occurrence of misunderstanding by adding additional information acquired from dictionary data or an application execution results is generated. Provided is an information processing device including an audio input unit that receives a user speech, a confirmation response generating unit that generates a confirmation response for requesting confirmation of a speaking user to the user speech, and an audio output unit that outputs the confirmation response generated by the confirmation response generating unit, in which the confirmation response generating unit executes generation of a confirmation response using additional information acquired from at least any of dictionary data and an application execution result.
    Type: Grant
    Filed: October 5, 2018
    Date of Patent: May 10, 2022
    Assignee: SONY CORPORATION
    Inventor: Keisuke Touyama
  • Patent number: 11328732
    Abstract: A method for generating a summary text composition can include obtaining historical reading data of a user. The method can include generating, based on the historical reading data, a reading proficiency level of the user. The method can include selecting, based on the reading proficiency level, a summarization model from a set of summarization models. The method can include obtaining a target composition. The target composition can be selected from the group consisting of a literary work, a video recording, and an audio recording. The method can include generating, by the summarization model, the summary text composition. The summary text composition can correspond to the target composition and have a first reading level classification that matches the reading proficiency level. The method can include transmitting the summary text composition to a computing device.
    Type: Grant
    Filed: September 16, 2020
    Date of Patent: May 10, 2022
    Assignee: International Business Machines Corporation
    Inventors: Yufang Hou, Beat Buesser, Bei Chen, Akihiro Kishimoto
  • Patent number: 11328709
    Abstract: A system for improving dysarthria speech intelligibility and method thereof, are provided. In the system, user only needs to provides a set of paired corpus including a reference corpus and a patient corpus, and a speech disordering module can automatically generate a new corpus completely synchronous with the reference corpus, and the new corpus can be used as a training corpus for training a dysarthria voice conversion model. The present invention does not need to use a conventional corpus alignment technology or a manual manner to perform pre-processing on the training corpus, so that manpower cost and time cost can be reduced, and synchronization of the training corpus can be ensured, thereby improving both training and conversion qualities of the voice conversion model.
    Type: Grant
    Filed: August 20, 2019
    Date of Patent: May 10, 2022
    Assignee: NATIONAL CHUNG CHENG UNIVERSITY
    Inventors: Tay-Jyi Lin, Ching-Hau Sung, Che-Chia Pai, Ching-Wei Yeh
  • Patent number: 11315567
    Abstract: An electronic device and an method of the electronic device are provided, where the electronic device maintains a context that does not reflect a request for a secret conversation, in response to the request for the secret conversation being received from a first user, and generates a response signal to a voice signal of a second user based on the maintained context, in response to an end of the secret conversation with the first user.
    Type: Grant
    Filed: May 12, 2020
    Date of Patent: April 26, 2022
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Sang Hyun Yoo, Young-Seok Kim, Jeong-Hoon Park, Jehun Jeon, Junhwi Choi
  • Patent number: 11301266
    Abstract: The present invention can receive a user input comprising a request to perform a task using at least one of the application programs via at least one of a display or a microphone, transmit data related to the user input to an external server through a communication circuit, transmit version information on at least one of the application programs to the external server via the communication circuit together with the data transmission or after the data transmission, receive a pass rule as a response to the request from the external server via the communication circuit, and store instructions that cause an electronic device to perform the task by allowing the electronic device to have the sequence. The pass rule comprises information on a sequence of states of the electronic device for performing the task, and at least one of the states may be one determined by the external server on the basis of at least a portion of the version information.
    Type: Grant
    Filed: March 27, 2018
    Date of Patent: April 12, 2022
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Kyounggu Woo, Hojun Jaygarl, Jehun Jeon, Hyunjin Park, Sangwook Kang, Kyungtae Kim, Jaeyung Yeo
  • Patent number: 11302324
    Abstract: Disclosed are a speech processing method and apparatus therefor which execute an installed artificial intelligence algorithm and/or machine learning algorithm to perform speech processing in a 5G communication environment.
    Type: Grant
    Filed: November 27, 2019
    Date of Patent: April 12, 2022
    Assignee: LG ELECTRONICS INC.
    Inventors: Ye Jin Kim, Hyun Yu, Byeong Ha Kim
  • Patent number: 11303649
    Abstract: A method, computer program product, and a system where a processor(s) determine that a user of a given computing device has been authenticated to initiate an application session, that the application session is open, and that the application session has a timeout mechanism triggered by inactivity (i.e., the session not receiving a selection within a first predefined period of time). The processor(s) determines that the application session will automatically timeout within a second predefined period of time based on the inactivity. The processor(s) monitors activities of the user during pendency of the application session with the application including physical and computing activities of the user. The processor(s) determines that at least one activity of the activities indicates engagement of the user with the application session. The processor(s) prevent the timeout mechanism from being triggered during a duration of the at least one activity.
    Type: Grant
    Filed: May 30, 2019
    Date of Patent: April 12, 2022
    Assignee: International Business Machines Corporation
    Inventors: Jennifer M. Hatfield, Michael Bender, Todd P. Seager, Jeremy R. Fox
  • Patent number: 11301620
    Abstract: A non-transitory computer-readable recording medium has stored therein a program that causes a computer to execute a process including determining, in response to detection of switching of displayed image data, whether a first annotation is added to first image data which is displayed before the switching, storing the first annotation in a storage in association with the first image data in a case where it is determined that the first annotation is added to the first image data, determining whether first specific image data exists in the storage, the first specific image data satisfying a condition with respect to second image data which is displayed after the switching, and superimposing, in a case where it is determined that the first specific image data exists, a second annotation stored in the storage in association with the first specific image data on the second image data to display the second annotation.
    Type: Grant
    Filed: May 20, 2021
    Date of Patent: April 12, 2022
    Assignee: FUJITSU LIMITED
    Inventors: Koki Hatada, Keiju Okabayashi
  • Patent number: 11302343
    Abstract: A signal analysis device includes an estimation unit that models a sound source position occurrence probability matrix Q using a product of a sound source position probability matrix B and a sound source existence probability matrix A, and estimates at least one of the sound source position probability matrix B and the sound source existence probability matrix A based on the modeling, the sound source position occurrence probability matrix Q being composed of probabilities of arrival of a signal from each sound source position candidate per frame, which is a time section, with respect to a plurality of sound source position candidates. The sound source position probability matrix B being composed of probabilities of arrival of a signal from each sound source position candidate per sound source with respect to a plurality of sound sources.
    Type: Grant
    Filed: April 4, 2019
    Date of Patent: April 12, 2022
    Assignee: NIPPON TELEGRAPH AND TELEPHONE CORPORATION
    Inventors: Nobutaka Ito, Tomohiro Nakatani, Shoko Araki
  • Patent number: 11295748
    Abstract: A speaker recognition device includes a memory, and a processor. The memory stores enrolled key phrase data corresponding to utterances of a key phrase by enrolled users,and text-dependent and text-independent acoustic speaker models of the enrolled users. The processor is operatively connected to the memory, and executes instructions to authenticate a speaker as an enrolled user, which includes detecting input key phrase data corresponding to a key phrase uttered by the speaker, computing text-dependent and text-independent scores for the speaker using speech models of the enrolled user, computing a confidence score, and authenticating or rejecting the speaker as the enrolled user based on whether the confidence score indicates that the input key phrase data corresponds to the speech from the enrolled user.
    Type: Grant
    Filed: December 14, 2018
    Date of Patent: April 5, 2022
    Assignee: Robert Bosch GmbH
    Inventors: Zhongnan Shen, Fuliang Weng, Gengyan Bei, Pongtep Angkititrakul
  • Patent number: 11295721
    Abstract: A system for use in video game development to generate expressive speech audio comprises a user interface configured to receive user-input text data and a user selection of a speech style. The system includes a machine-learned synthesizer comprising a text encoder, a speech style encoder and a decoder. The machine-learned synthesizer is configured to generate one or more text encodings derived from the user-input text data, using the text encoder of the machine-learned synthesizer; generate a speech style encoding by processing a set of speech style features associated with the selected speech style using the speech style encoder of the machine-learned synthesizer; combine the one or more text encodings and the speech style encoding to generate one or more combined encodings; and decode the one or more combined encodings with the decoder of the machine-learned synthesizer to generate predicted acoustic features.
    Type: Grant
    Filed: April 3, 2020
    Date of Patent: April 5, 2022
    Assignee: ELECTRONIC ARTS INC.
    Inventors: Siddharth Gururani, Kilol Gupta, Dhaval Shah, Zahra Shakeri, Jervis Pinto, Mohsen Sardari, Navid Aghdaie, Kazi Zaman
  • Patent number: 11289085
    Abstract: A method of automatically delineating turns in a multi-turn dialogue between a user and a conversational computing interface. Audio data encoding speech of the user in the multi-turn dialogue is received. The audio data is analyzed to recognize, in the speech of the user, an utterance followed by a silence. The utterance is recognized as a last utterance in a turn of the multi-turn dialogue responsive to the silence exceeding a context-dependent duration dynamically updated based on a conversation history of the multi-turn dialogue and features of the received audio, wherein the conversation history includes one or more previous turns of the multi-turn dialogue taken by the user and one or more previous turns of the multi-turn dialogue taken by the conversational computing interface.
    Type: Grant
    Filed: November 26, 2019
    Date of Patent: March 29, 2022
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Charles Clayton Wooters, Steven Andrew Wegmann, Michael Jack Newman, David Leo Wright Hall, Jean Crawford, Laurence Steven Gillick
  • Patent number: 11289082
    Abstract: Described herein is a system for adapting an output to a user input over a period of time based on how often the user interacts with the system. The system may determine a user's level of familiarity of the system, and may determine to personalize the output to a user request based on his level of familiarity. The user's level of familiarity may be determined by analyzing historical interactions between the user and the system. The level of personalization applied to the output may be determined based on the user's level of familiarity. As user becomes more familiar with the system, the output may be more personalized.
    Type: Grant
    Filed: November 7, 2019
    Date of Patent: March 29, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Andrea Klein Lacy, Timothy Whalin
  • Patent number: 11282504
    Abstract: The present disclosure is generally directed a system to detect activation phrases within input audio signals transmitted over a low-bandwidth network. The system can use a two-stage activation phrase detection process. First a sensing device, which can include a plurality of microphones for detecting an input audio signal, can detect an input audio signal that includes a candidate activation phrase. Second, the sensing device can transmit the recordings of the input audio signal to a client device for confirmation that the input audio signal includes the activation phrase.
    Type: Grant
    Filed: May 4, 2020
    Date of Patent: March 22, 2022
    Assignee: GOOGLE LLC
    Inventors: Jeremy Payne, Tomer Amarilio
  • Patent number: 11282507
    Abstract: Described techniques select portions of an audio stream for transmission to a trained machine learning application, which generates response recommendations in real-time. This real-time response is facilitated by the system identifying, selecting and transmitting those portions of the audio stream likely to be most relevant to the conversation. Portions of an audio stream less likely to be relevant to the conversation are identified accordingly and not transmitted. The system may identify the relevant portions of an audio stream by detecting events in a contemporaneous event stream, use a trained machine learning model to identify events in an audio stream, or both.
    Type: Grant
    Filed: June 21, 2021
    Date of Patent: March 22, 2022
    Assignee: CRESTA INTELLIGENCE INC.
    Inventors: Tianlin Shi, Kenneth George Oetzel
  • Patent number: 11275431
    Abstract: A control method of a personal authentication apparatus and a personal authentication apparatus stores a piece of authentication information of a predetermined user. The method and apparatus estimates an emotion felt by a target user, based on a physical quantity acquired by a sensor, generates identification information that identifies an object that causes the target user to feel the estimated emotion, based on the physical quantity acquired by the sensor or presentation information presented by a presenter, and determines whether the target user is the predetermined user based on the estimated emotion and the generated identification information and the piece of authentication information of the predetermined user.
    Type: Grant
    Filed: May 11, 2020
    Date of Patent: March 15, 2022
    Assignee: PANASONIC INTELLECTUAL PROPERTY CORPORATION OF AMERICA
    Inventors: Aki Yoneda, Koichi Kusukame, Shinichi Shikii, Nawatt Silawan, Shinji Uchida, Tatsuo Itoh
  • Patent number: 11276419
    Abstract: A computing device receives a video feed. The video feed is divided into a sequence of video segments. For each video segment, visual features of the video segment are extracted. A predicted spectrogram is generated based on the extracted visual features. A synthetic audio waveform is generated from the predicted spectrogram. All synthetic audio waveforms of the video feed are concatenated to generate a synthetic soundtrack that is synchronized with the video feed.
    Type: Grant
    Filed: July 30, 2019
    Date of Patent: March 15, 2022
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Yang Zhang, Chuang Gan, Sijia Liu, Dakuo Wang
  • Patent number: 11266330
    Abstract: The present disclosure generally relates to techniques and user interfaces for interacting with research studies. In some embodiments, an electronic device displays a user interface that includes a task view with active tasks from multiple research studies. In some embodiments, an electronic device, while displaying a research study user interface, displays an indication of a problem that prevents enrollment in the research study when enrollment problem criteria are met. In some embodiments, an electronic device, while performing a hearing test, suspends the test and displays a restart affordance when the ambient noise level exceeds a threshold.
    Type: Grant
    Filed: August 11, 2020
    Date of Patent: March 8, 2022
    Assignee: Apple Inc.
    Inventors: Matthew W. Crowley, Pablo F. Caro, Charmian Bondoc Naguit
  • Patent number: 11272362
    Abstract: A system for implicit authentication for a mobile device associated with a user, wherein the implicit authentication is behavioral, biometric and task-based and includes at least one authentication task selected so as to leverage the user's muscle memory. The mobile device comprises a touchscreen; a transaction authentication information unit; one or more sensors coupled to the transaction authentication information unit; and an anomaly detector coupled to the transaction authentication information unit. The sensors comprise one or more touchscreen sensors coupled to the touchscreen, an accelerometer, and a gyroscope, and are used to obtain and transmit one or more sets of data to the transaction authentication information unit. The sets of data are associated with one or more performances of the authentication task by the user. The anomaly detector generates an authentication model using the one or more data sets transmitted to the transaction authentication information unit.
    Type: Grant
    Filed: February 3, 2020
    Date of Patent: March 8, 2022
    Assignee: Zighra Inc.
    Inventors: Deepak Chandra Dutt, Anil Buntwal Somayaji, Michael John Kendal Bingham
  • Patent number: 11269413
    Abstract: The present invention relates to a brain-computer interface system and a method for recognizing a conversation intention of a user using the same in addition to inferring the waveform of word sound intended by a user from an imagined speech brainwave associated with a word intended by a user, since the user can intuitively recognize the sentence he/she wants to speak through the imagined speech by classifying words that are most relevant to the imagined speech brainwave of the user in a database in which a word often used by the user or frequently used in a specific situation is stored and by generating a sentence intended by the user by recognizing the words classified in this way, it is possible to perform communication by only thoughts of the user.
    Type: Grant
    Filed: November 7, 2019
    Date of Patent: March 8, 2022
    Assignee: Korea University Research and Business Foundation
    Inventors: Seong-Whan Lee, Ji-Hoon Jeong, No-Sang Kwak, Seo-Hyun Lee
  • Patent number: 11270706
    Abstract: A voice controlled assistant has a housing to hold one or more microphones, one or more speakers, and various computing components. The housing has an elongated cylindrical body extending along a center axis between a base end and a top end. The microphone(s) are mounted in the top end and the speaker(s) are mounted proximal to the base end. The microphone(s) and speaker(s) are coaxially aligned along the center axis. The speaker(s) are oriented to output sound directionally toward the base end and opposite to the microphone(s) in the top end. The sound may then be redirected in a radial outward direction from the center axis at the base end so that the sound is output symmetric to, and equidistance from, the microphone(s).
    Type: Grant
    Filed: May 13, 2020
    Date of Patent: March 8, 2022
    Assignee: Amazon Technologies, Inc.
    Inventor: Timothy Theodore List
  • Patent number: 11265414
    Abstract: The present disclosure provides a method and device for waking up a voice recognition function in a mobile terminal. The method includes: receiving a trigger instruction of an input interface, and activating the input interface; acquiring a rotation angle of the mobile terminal in a first rotation direction with an x-axis as a rotation axis, in which the x-axis is a widthwise direction of a screen of the mobile terminal; and activating the voice recognition function when the rotation angle of the mobile terminal in the first rotation direction with the x-axis as the rotation axis is greater than or equal to a first trigger threshold.
    Type: Grant
    Filed: January 2, 2020
    Date of Patent: March 1, 2022
    Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.
    Inventors: Kang Wang, Minqiang Chang
  • Patent number: 11264029
    Abstract: Embodiments herein relate to a local assistant system responding to voice input using an ear-wearable device. The system detects a wake-up signal and receives a first voice input communicating a first query content. The system includes speech recognition circuitry to determine the first query content, speech generation circuitry, and an input database of locally-handled user inputs. If the first audio input matches one of the locally-handled user inputs, then the system takes a local responsive action. If the first audio input does not match one of the locally-handled user inputs, then the system transmits at least a portion of the first query content over a wireless network to a network resource.
    Type: Grant
    Filed: January 2, 2020
    Date of Patent: March 1, 2022
    Assignee: Starkey Laboratories, Inc.
    Inventors: Achintya Kumar Bhowmik, David Alan Fabry, Amit Shahar, Justin R. Burwinkel, Jeffrey Paul Solum, Thomas Howard Burns
  • Patent number: 11256874
    Abstract: Various embodiments of the present disclosure performing conversation sentiment monitoring for a conversation data object. In various embodiments, a text block that can be resized is identified within a conversation data object and successive regularized sentiment profile generation iterations are performed until a regularized sentiment score of the block exceeds a regularized sentiment score threshold. A current regularized sentiment profile generation iteration involves determining a regularized sentiment score for the block based on an initial sentiment score, a subjectivity probability value, and, optionally, a stage-wise penalty factor. A determination is then made as to whether the score exceeds the threshold. If so, then a regularized sentiment profile of the conversation data object is updated based on the regularized sentiment score. If not, then the text block is resized and a subsequent regularized sentiment profile generation iteration is performed based on the resized block.
    Type: Grant
    Filed: June 16, 2020
    Date of Patent: February 22, 2022
    Assignee: Optum Technology, Inc.
    Inventors: Ninad D. Sathaye, Raghav Bali, Piyush Gupta, Krishnamohan Nandiraju
  • Patent number: 11250848
    Abstract: Control of navigation of a content item is carried out by detection of non-verbal sound events. On the basis of receipt of one or more non-verbal sound event reports, a navigation tool is provided with a corresponding sequence of navigation commands. The correspondence between navigation command sequences and non-verbal sound events is established through analysis or markup of the content item.
    Type: Grant
    Filed: September 24, 2019
    Date of Patent: February 15, 2022
    Assignee: AUDIO ANALYTIC LTD
    Inventors: Christopher James Mitchell, Sacha Krstulovic, Cagdas Bilen, Neil Cooper, Julian Harris, Arnoldas Jasonas, Joe Patrick Lynas
  • Patent number: 11250493
    Abstract: A method includes receiving, at a social media software module operating on a user device, input from a first user associated with a payment from the first user to a second user, wherein the social media software module provides a social media interaction which enables communication between the first user and the second user, receiving, in connection with use of the social media software module for the social media interaction, a payment authorization from the first user for the payment to the second user and initiating, from the social media software module and based on the payment authorization, a blockchain-based payment from the first user to the second user. This approach incorporates a cryptocurrency wallet into a social media application to reduce friction and enable easy payments between users of the social media application using a cryptocurrency for at least part of the transfer of value from one user to another.
    Type: Grant
    Filed: July 31, 2019
    Date of Patent: February 15, 2022
    Assignee: Monticello Enterprises LLC
    Inventors: Thomas M. Isaacson, Ryan Connell Durham
  • Patent number: 11238854
    Abstract: Methods, apparatus, and computer readable media are described related to recording, organizing, and making audio files available for consumption by voice-activated products. In various implementations, in response to receiving an input from a first user indicating that the first user intends to record audio content, audio content may be captured and stored. Input may be received from the first user indicating at least one identifier for the audio content. The stored audio content may be associated with the at least one identifier. A voice input may be received from a subsequent user. In response to determining that the voice input has particular characteristics, speech recognition may be biased in respect of the voice input towards recognition of the at least one identifier. In response to recognizing, based on the biased speech recognition, presence of the at least one identifier in the voice input, the stored audio content may be played.
    Type: Grant
    Filed: December 14, 2016
    Date of Patent: February 1, 2022
    Assignee: Google LLC
    Inventors: Vikram Aggarwal, Barnaby James
  • Patent number: 11240104
    Abstract: Systems, devices, media, and methods are presented for assigning configuration parameters to unknown devices by comparing characteristics of unknown devices to characteristics of known devices. Characteristics of an unknown device are compared to characteristics of known devices to identify a similar device having known configuration parameters. The unknown devise is then assigned the configuration parameters of the identified similar device.
    Type: Grant
    Filed: May 21, 2019
    Date of Patent: February 1, 2022
    Assignee: Snap Inc.
    Inventors: Michael Cieslak, Jiayao Yu, Kai Chen, Farnaz Azmoodeh, Michael David Marr, Jun Huang, Zahra Ferdowsi
  • Patent number: 11238866
    Abstract: A method and apparatus for intelligently alerting individuals within a public-safety communication system is provided herein. During operation a query result is determined to have certain keywords. When this happens, prior conversations of individuals over various talkgroups are analyzed. The query terms and/or the query result are compared to the prior conversations, and a determination is made if any prior conversations are relevant to the query terms and/or the query results. If so, identities are determined for those individuals who were involved in the relevant prior conversations, and a notification of the results of the query are sent to electronic devices associated with the individuals who were involved in the relevant prior conversations.
    Type: Grant
    Filed: June 17, 2019
    Date of Patent: February 1, 2022
    Assignee: MOTOROLA SOLUTIONS, INC.
    Inventors: Craig F Siddoway, Srividya G Subramanyam
  • Patent number: 11232155
    Abstract: Generating and/or recommending command bundles for a user of an automated assistant. A command bundle comprises a plurality of discrete actions that can be performed by an automated assistant. One or more of the actions of a command bundle can cause transmission of a corresponding command and/or other data to one or more devices and/or agents that are distinct from devices and/or agents to which data is transmitted based on other action(s) of the bundle. Implementations determine command bundles that are likely relevant to a user, and present those command bundles as suggestions to the user. In some of those implementations, a machine learning model is utilized to generate a user action embedding for the user, and a command bundle embedding for each of a plurality of command bundles. Command bundle(s) can be selected for suggestion based on comparison of the user action embedding and the command bundle embeddings.
    Type: Grant
    Filed: November 22, 2019
    Date of Patent: January 25, 2022
    Assignee: Google LLC
    Inventor: Yuzhao Ni
  • Patent number: 11232162
    Abstract: A computer-implemented communications method is discussed. The method includes receiving a typed user input in a texting conversation on an electronic communication channel between two or more users of electronic communications devices; analyzing the user input to identify whether the user input contains a reserved word for triggering a search; submitting a search query received in association with the reserved word to a search engine, if the user input contains a reserved word; receiving one or more search results from the search engine; and providing, in the electronic communication channel for incorporation into the texting conversation, at least one of the one or more search results.
    Type: Grant
    Filed: March 3, 2020
    Date of Patent: January 25, 2022
    Assignee: Google LLC
    Inventors: Maryam Garrett, Richard A. Miner
  • Patent number: 11232798
    Abstract: A language proficiency analyzer automatically evaluates a person's language proficiency by analyzing that person's oral communications with another person. The analyzer first enhances the quality of an audio recording of a conversation between the two people using a neural network that automatically detects loss features in the audio and adds those loss features back into the audio. The analyzer then performs a textual and audio analysis on the improved audio. Through textual analysis, the analyzer uses a multi-attention network to determine how focused one person is on the other and how pleased one person is with the other. Through audio analysis, the analyzer uses a neural network to determine how well one person pronounced words during the conversation.
    Type: Grant
    Filed: May 21, 2020
    Date of Patent: January 25, 2022
    Assignee: Bank of America Corporation
    Inventors: Madhusudhanan Krishnamoorthy, Harikrishnan Rajeev
  • Patent number: 11232101
    Abstract: A method, an apparatus and a system for information retrieval are provided. The method for information retrieval may comprise: receiving a query in a natural language form from a user (210); extracting a plurality of feature vectors from the query based on a plurality of knowledge entries in a knowledge base associated with the user and a built-in featurization source, one feature vector per knowledge entry (220); obtaining, with a pre-trained language understanding model, a plurality of language understanding results based on the plurality of feature vectors (230); and selecting a knowledge entry corresponding to the query among the plurality of entries based on the plurality of language understanding results (240). The method may extract more features and combine information retrieval and language understanding in one shot to improve efficiency.
    Type: Grant
    Filed: October 10, 2016
    Date of Patent: January 25, 2022
    Assignee: Microsoft Technology Licensing, LLC
    Inventor: Yong Ni
  • Patent number: 11227602
    Abstract: This disclosure describes transcribing speech using audio, image, and other data. A system is described that includes an audio capture system configured to capture audio data associated with a plurality of speakers, an image capture system configured to capture images of one or more of the plurality of speakers, and a speech processing engine. The speech processing engine may be configured to recognize a plurality of speech segments in the audio data, identify, for each speech segment of the plurality of speech segments and based on the images, a speaker associated with the speech segment, transcribe each of the plurality of speech segments to produce a transcription of the plurality of speech segments including, for each speech segment in the plurality of speech segments, an indication of the speaker associated with the speech segment, and analyze the transcription to produce additional data derived from the transcription.
    Type: Grant
    Filed: November 20, 2019
    Date of Patent: January 18, 2022
    Assignee: Facebook Technologies, LLC
    Inventors: Vincent Charles Cheung, Chengxuan Bai, Yating Sasha Sheng
  • Patent number: 11226787
    Abstract: The present disclosure generally relates to a computer-implemented system for intelligently retaining and recalling memory data. An exemplary method comprises receiving, via a microphone of an electronic device, a speech input of the user; receiving a text input of the user; constructing a first instance of a memory data structure based on the speech input; constructing a second instance of the memory data structure based on the text input; adding the first instance and the second instance of the memory data structure to a memory stack of the user; displaying a user interface for retrieving memory data of the user; receiving, via the user interface, a beginning of a statement from the user; retrieving a particular instance of the memory data structure from the memory stack based on the beginning of the statement; and automatically displaying a completion of the statement.
    Type: Grant
    Filed: July 20, 2021
    Date of Patent: January 18, 2022
    Assignee: Human AI Labs, Inc.
    Inventors: Suman Kanuganti, Xiaoran Zhang, Kristie Kaiser
  • Patent number: 11221690
    Abstract: Methods and apparatuses for a mobile device (e.g., a mobile phone) supporting virtual peripherals are described. The mobile device may include a projecting component configured to project images to external surfaces outside of the mobile device. In some cases, the mobile device may project an image of a document on a first external surface and an image of a keyboard on a second external surface. Moreover, the mobile device may include an imaging component to receive inputs based on sensing an input device overlaid on the projected image (e.g., the keyboard). Accordingly, the mobile device may function as a computer with an external monitor and a keyboard, virtually added to the mobile device. In some cases, the mobile device may serve two or more users or purposes at a time. For example, the mobile device may operate as an infotainment device of a vehicle.
    Type: Grant
    Filed: May 20, 2020
    Date of Patent: January 11, 2022
    Assignee: Micron Technology, Inc.
    Inventors: Elsie de la Garza Villarreal, Claudia A. Delaney, Madison E. Wale, Bhumika Chhabra
  • Patent number: 11218802
    Abstract: A mobile device capable of capturing voice commands includes a beamformer for determining audio data corresponding to one or more directions and a beam selector for selecting in which direction a source of target audio lies. The device determines, based on data from one or more sensors, an angle through which the device has rotated. Based on the angle and one or more rotation-compensation functions, the device interpolates audio data corresponding to the one or more directions to compensate for the rotation such that the direction corresponding to the source of target audio remains selected.
    Type: Grant
    Filed: September 25, 2018
    Date of Patent: January 4, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Srivatsan Kandadai, Amit Singh Chhetri, Trausti Thor Kristjansson
  • Patent number: 11216242
    Abstract: An audio output system according to an embodiment has functions as: a receiving unit that receives voice data including a voice or sound data including a sound as output-data to be output; a content identification information generating unit that generates content identification information for identifying content that is determined from the output-data; a status identification information generating unit that generates status identification information for identifying a status from at least one of sensor data and input data; a parameter information generating unit that generates parameter information used for controlling output of the output-data from the content identification information and the status identification information; and an output control unit that controls audio output of the output-data by using the parameter information.
    Type: Grant
    Filed: February 19, 2019
    Date of Patent: January 4, 2022
    Assignees: KABUSHIKI KAISHA TOSHIBA, TOSHIBA DIGITAL SOLUTIONS CORPORATION
    Inventor: Yuki Yamada
  • Patent number: 11212340
    Abstract: Methods, configured mobile devices, and computer-readable media for displaying content on mobile devices are disclosed. An exemplary method for displaying content on a mobile device includes steps for receiving a first content request provided by a native app executing on one or more processors of a mobile device; computing a priority of the native app to display app content on a lock screen of the mobile device; and in the event that the native app has higher priority than other requesting apps to display app content on the lock screen of the mobile device, transmitting a second content request to a content server; receiving app content from the content server in response to the second content request; and displaying the app content on the lock screen of the mobile device.
    Type: Grant
    Filed: April 8, 2019
    Date of Patent: December 28, 2021
    Assignee: Adenda Media Inc.
    Inventor: Francis George Doumet
  • Patent number: 11210613
    Abstract: Embodiments of the present invention are directed to a computer-implemented machine-learning method and system for automatically creating and updating tasks by reading signals from external data sources and understanding what users are doing. Embodiments of the present invention are directed to a computer-implemented machine-learning method and system for automatically completing tasks by reading signals from external sources and understanding when an existing task has been executed. Tasks created are representable and explainable in a human readable format that can be shown to users and used to automatically fill productivity applications including but not limited to task managers, to-do lists, project management, time trackers, and daily planners. Tasks created are representable in a way that can be interpreted by a machine such as a computer system or an artificial intelligence so that external systems can be delegated or connected to the system.
    Type: Grant
    Filed: August 25, 2017
    Date of Patent: December 28, 2021
    Assignee: DIALPAD UK LIMITED
    Inventors: Michele Sama, Arseni Anisimovich, Tim Porter, Theodosia Togia, James Hammerton
  • Patent number: 11206372
    Abstract: The embodiments of the disclosure provide a projection-type video conference system, which includes a video conference device and a cloud service system. The cloud service system is configured to provide services to the video conference device. The video conference device includes: a camera assembly configured to acquire image information of a conference scene and generate a conference video; an audio input assembly configured to collect voice signals of the conference scene, the voice signals including a recognizable voice instruction; a projection assembly configured to display the conference video. The video conference device can recognize the voice instruction, and perform, according to the voice instruction, an operation corresponding thereto. The video conference device may be controlled through voice, and it has high integration and is convenient to carry.
    Type: Grant
    Filed: January 27, 2021
    Date of Patent: December 21, 2021
    Assignee: AMPULA INC.
    Inventor: Yajun Zhang
  • Patent number: 11204955
    Abstract: Delivering content based on predicted effect can include analyzing, using computer hardware, digital content before providing the digital content to a device of a user, predicting, using the computer hardware, an effect that the digital content will have on a circadian rhythm of the user, based on the predicted effect, modifying, using the computer hardware, the digital content to generate modified digital content, and providing, using the computer hardware, the modified digital content to the device of the user.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: December 21, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Jeremy R. Fox, Trudy L. Hewitt, John Rice, Liam S. Harpur
  • Patent number: 11200905
    Abstract: A speech recognition unit converts speech uttered by a speaker to text data. An intention estimation unit extracts a nickname from the text data, and searches a database for registration information that satisfies a condition corresponding to the nickname. A search result output presents the searched for registration information to the speaker that corresponding to the nickname.
    Type: Grant
    Filed: February 6, 2018
    Date of Patent: December 14, 2021
    Assignee: Nissan Motor Co., Ltd.
    Inventor: Masaru Tasaki
  • Patent number: 11202131
    Abstract: Methods, systems, and computer-readable media for artificially generating a revoiced media stream and maintaining original volume changes of a character in the revoiced media stream are provided. For example, a media stream including an individual speaking may be obtained. A transcript of the media stream may be obtained. The transcript of the media stream may be translated to a target language. A revoiced media stream in which the translated transcript in the target language is spoken by a virtual entity may be generated, wherein a ratio of the volume levels between first and second sets of words in the revoiced media stream is substantially identical to the ratio of volume levels between corresponding first and second utterances in the received media stream.
    Type: Grant
    Filed: March 9, 2020
    Date of Patent: December 14, 2021
    Assignee: VIDUBLY LTD
    Inventors: Ron Zass, Ben Avi Ingel
  • Patent number: 11201849
    Abstract: A device can include a processor; memory accessible to the processor; a microphone operatively coupled to the processor; a network interface operatively coupled to the processor; circuitry that parses audio signals received via the microphone for a name and a type of device; circuitry that analyzes network information associated with remote devices accessible via the network interface based at least in part on at least one of the name and the type of device to associate the name and the type of device with one of the remote devices; and circuitry that stores the name and the type of device to the memory in association with the one of the remote devices.
    Type: Grant
    Filed: January 15, 2018
    Date of Patent: December 14, 2021
    Assignee: Lenovo (Singapore) Pte. Ltd.
    Inventors: Jeffrey Earle Skinner, Aaron Michael Stewart, Jonathan Jen-Wei Yu, Lincoln Penn Hancock, Song Wang, Ming Qian, John Weldon Nicholson, David Alexander Schwarz
  • Patent number: 11195529
    Abstract: A communication system (100) formed of a radio management system (104) interoperates with a radio management codeplug database (106) and a confusability analyzer (108) for the creation and storage of voice recognition target strings for uploading to one or more land mobile portable radios (110) prior to field-deployment. Once the radios are deployed to the field, unrecognized voice command entries are determined, stored locally at the portable radio, and uploaded to cloud based storage (122). Analysis of the cloud based data is performed through a voice control analytics engine (124) to detect patterns associated with the unrecognized voice commands. The unrecognized voice command pattern is processed though a resolution action engine (126) to generate one or more resolution actions for the pattern.
    Type: Grant
    Filed: March 17, 2020
    Date of Patent: December 7, 2021
    Assignee: MOTOROLA SOLUTIONS, INC.
    Inventors: Darrell J. Stogner, Ellis A. Pinder, Matthew E. Simms, Melanie King
  • Patent number: 11194466
    Abstract: The invention relates to a method for entering commands into an electronic device with a touch-sensitive display unit on which an image can be displayed and at least an area of the displayed image can be selected by means of a pointer and in which segmentation can thereby be carried out in a targeted manner and with a speech recognition unit by means of the commands can be generated for the electronic device. According to the invention, the user of the electronic device selects at least a partial area of a displayed image, and the selected area is subjected to an image analysis following a simultaneous acoustic input.
    Type: Grant
    Filed: December 17, 2019
    Date of Patent: December 7, 2021
    Assignee: Patty's GmbH
    Inventor: Andreas Perschk
  • Patent number: 11196742
    Abstract: The present application discloses a method, device, and system for data transmission. The system includes a first terminal having an associated terminal ID and a server. At least one of the first terminal and the server establishes a channel between the first terminal and the server based at least in part on the terminal ID, the channel corresponding to a communication channel that is to be used by the first terminal and the server to communicate data. The terminal ID is used to authenticate the first terminal in connection with the channel being established between the first terminal and the server.
    Type: Grant
    Filed: December 21, 2018
    Date of Patent: December 7, 2021
    Assignee: BANMA ZHIXING NETWORK (HONGKONG) CO., LIMITED
    Inventors: Yanming Cai, Chunhui Zhang
  • Patent number: 11196753
    Abstract: Implementations of the present disclosure discloses user identity verification method selection method and device. In an implementation, times and corresponding verification methods of successful user identity verifications are recorded, one or more verification methods of successful user identity verifications performed in a predetermined time period based on the recorded times and corresponding verification methods are identified. Whether at least one verification method is included in both the identified one or more verification methods of successful user identity verifications and a list of available verification methods are determined. One or more available verification methods different from the identified one or more verification methods are used for user identity verification if no verification method is included in both the identified one or more verification methods of successful user identity verifications and the list of available verification methods.
    Type: Grant
    Filed: December 19, 2019
    Date of Patent: December 7, 2021
    Assignee: Advanced New Technologies Co., Ltd.
    Inventors: Jian Jiang, Xin Lei, Hongfa Sun
  • Patent number: 11189286
    Abstract: As noted above, example techniques relate to toggling a cloud-based VAS between enabled and disabled modes. An example implementation involves a NMD detecting that the housing is in a first orientation and enabling a first mode. Enabling the first mode includes disabling voice input processing via a cloud-based VAS and enabling local voice input processing. In the first mode, the NMD captures sound data associated with a first voice input and detects, via a local natural language unit, that the first voice input comprises sound data matching one or more keywords. The NMD determines an intent of the first voice input and performs a first command according to the determined intent. The NMD may detect that the housing is in a second orientation and enables the second mode. Enabling the second mode includes enabling voice input processing via the cloud-based VAS.
    Type: Grant
    Filed: October 22, 2019
    Date of Patent: November 30, 2021
    Assignee: Sonos, Inc.
    Inventors: Fiede Schillmoeller, Connor Smith