Speech To Image Patents (Class 704/235)
  • Patent number: 10964318
    Abstract: A system and method to receive a spoken utterance and convert the spoken utterance into a recognized speech results through an automatic speech recognition service. A spoken utterance into a recognized speech result through an automatic speech recognition service. The recognized speech results are interpreted through a natural language processing module. A normalizer processes the recognized speech results that transforms the recognized speech interpretations into predefined form for a given automatic speech recognition domain and further determines which automatic speech recognition domains or the recognized speech results are processed by a dedicated dialogue management proxy module or a conversation module.
    Type: Grant
    Filed: January 4, 2018
    Date of Patent: March 30, 2021
    Assignee: BlackBerry Limited
    Inventor: Darrin Kenneth John Fry
  • Patent number: 10957310
    Abstract: The technology disclosed relates to authoring of vertical applications of natural language understanding (NLU), which analyze text or utterances and construct their meaning. In particular, it relates to new programming constructs and tools and data structures implementing those new applications.
    Type: Grant
    Filed: March 15, 2013
    Date of Patent: March 23, 2021
    Assignee: SoundHound, Inc.
    Inventors: Keyvan Mohajer, Seyed Majid Emami, Chris Wilson, Bernard Mont-Reynaud
  • Patent number: 10956481
    Abstract: Described herein are technologies that facilitate effective use (e.g., indexing and searching) of non-text machine data (e.g., audio/visual data) in an event-based machine-data intake and query system.
    Type: Grant
    Filed: April 28, 2017
    Date of Patent: March 23, 2021
    Assignee: SPLUNK Inc.
    Inventor: Adam Oliner
  • Patent number: 10957305
    Abstract: An information processing method and an electronic device are provided. The method includes: obtaining audio data collected by a slave device; obtaining contextual data corresponding to the slave device; and obtaining a recognition result of recognizing the audio data based on the contextual data. The contextual data characterizes a voice environment of the audio data collected by the slave device.
    Type: Grant
    Filed: June 16, 2017
    Date of Patent: March 23, 2021
    Assignee: LENOVO (BEIJING) CO., LTD.
    Inventor: Weixing Shi
  • Patent number: 10949626
    Abstract: The present disclosure provides a global simultaneous interpretation method and production thereof, the method includes the following steps: receiving a calling request sent by a terminal by a smart phone, connecting the calling request, and establishing a calling connection; receiving a first voice information transmitted through the calling connection by the smart phone, and when the first voice information is identified and is determined as a non-specified language, translating the first voice information into a second voice information of a specified language; and playing the second voice information by using a speaker device by the smart phone.
    Type: Grant
    Filed: March 12, 2019
    Date of Patent: March 16, 2021
    Assignee: WING TAK LEE SILICONE RUBBER TECHNOLOGY (SHENZHEN) CO., LTD
    Inventor: Tak Nam Liu
  • Patent number: 10950228
    Abstract: Methods and systems for receiving shouted-out user responses to broadcast entertainment content, and for determining the responsiveness of those responses in relation to the broadcast content. In particular, entertainment broadcasts can be accompanied by mark-up data that represents various events within a given broadcast, which can be compared to the shouted-out responses to determine their accuracy. For example, if a game show was broadcast and an individual started shouting out answers during the broadcast, embodiments disclosed herein could utilize a voice-controlled electronic device that captures the shouted-out answers and passes them on to a language processing system that determines whether they are correct by comparing the answers to the mark-up data.
    Type: Grant
    Filed: June 28, 2017
    Date of Patent: March 16, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Alfred Yong-Hock Tan, Matthew Luker, David Allen Markley
  • Patent number: 10943590
    Abstract: The present invention relates to a washing machine and a server system that recommend a laundry course and washing tip information in consideration of information on a kind of laundry and a degree of contamination inputted by using artificial intelligence through speech recognition, and a method for controlling such washing machine and server system. The present invention extracts a cloth-word indicating a kind of clothes of laundry and a stain-word indicating a kind of contaminant. Then, the present invention determines a laundry course in consideration of the cloth-word. Then, washing tip information on the stain-word is searched in a pre-stored database and the searched washing tip information is provided to a user. As a result, the user can be notified of a washing method capable of effectively removing the contaminant.
    Type: Grant
    Filed: July 12, 2019
    Date of Patent: March 9, 2021
    Assignee: LG Electronics Inc.
    Inventor: Heungkyu Lee
  • Patent number: 10936936
    Abstract: A system and method of configuring a graphical control structure for controlling a machine learning-based automated dialogue system includes configuring a root dialogue classification node that performs a dialogue intent classification task for utterance data input; configuring a plurality of distinct dialogue state classification nodes that are arranged downstream of the root dialogue classification node; configuring a graphical edge connection between the root dialogue classification node and the plurality of distinct state dialogue classification nodes that graphically connects each of the plurality of distinct state dialogue classification nodes to the root dialogue classification node, wherein (i) the root dialogue classification node, (ii) the plurality of distinct classification nodes, (iii) and the transition edge connections define a graphical dialogue system control structure that governs an active dialogue between a user and the machine learning-based automated dialogue system.
    Type: Grant
    Filed: November 13, 2019
    Date of Patent: March 2, 2021
    Assignee: Clinc, Inc.
    Inventors: Parker Hill, Jason Mars, Lingjia Tang, Michael A. Laurenzano, Johann Hauswald, Yiping Kang, Yunqi Zhang
  • Patent number: 10936812
    Abstract: An approach is provided that receives words that are input by a user of an application with the words being displayed on a display device. Each of the words are compared to words from a dictionary. Based on the comparisons, words that are not found in the dictionary and only appear a single time are highlighted as being misspelled words. However, words that are not in the dictionary but appear multiple times in the document are highlighted differently to indicate that these words are possible misspelled words with the difference in highlighting allowing the user to easily discern between misspelled and possibly misspelled words.
    Type: Grant
    Filed: January 10, 2019
    Date of Patent: March 2, 2021
    Assignee: International Business Machines Corporation
    Inventors: Kyle M. Brake, Stephen A. Boxwell, Stanley J. Vernier, Keith G. Frost
  • Patent number: 10937414
    Abstract: Systems and methods for text input based on neuromuscular information. The system includes a plurality of neuromuscular sensors, arranged on one or more wearable devices, wherein the plurality of neuromuscular sensors is configured to continuously record a plurality of neuromuscular signals from a user, at least one storage device configured to store one or more trained statistical models, and at least one computer processor programmed to obtain the plurality of neuromuscular signals from the plurality of neuromuscular sensors, provide as input to the one or more trained statistical models, the plurality of neuromuscular signals or signals derived from the plurality of neuromuscular signals, and determine based, at least in part, on an output of the one or more trained statistical models, one or more linguistic tokens.
    Type: Grant
    Filed: May 8, 2018
    Date of Patent: March 2, 2021
    Assignee: Facebook Technologies, LLC
    Inventors: Adam Berenzweig, Alan Huan Du, Jeffrey Scott Seely
  • Patent number: 10929009
    Abstract: An electronic device is provided. The electronic device includes a housing, a touch screen display that includes a first edge and a second edge, a microphone, at least one speaker, a wireless communication circuit, a memory, and a processor operably connected with the touch screen display, the microphone, the at least one speaker, the wireless communication circuit, and the memory. The processor is configured to output a home screen including a plurality of application icons in a matrix pattern. The processor is configured receive an input from the first edge to the second edge. The processor is configured output a user interface on the touch screen display that includes a button that allows user to call a first operation and a plurality of cards. To call the first operation the processor is configured to receive a user input, transmit data and receive a response, and perform a task.
    Type: Grant
    Filed: April 30, 2018
    Date of Patent: February 23, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Young Seok Lim, Hong Seok Kwon, Ho Min Moon, Mi Jung Park, Woo Young Park, Ki Hyoung Son, Won Ick Ahn, Pil Seung Yang, Jae Seok Yoon, Gi Soo Lee, Sun Jung Lee, Jae Hyeok Lee, Hyun Yeul Lee, Hyeon Cheon Jo, Doo Soon Choi, Kyung Wha Hong, Da Som Lee, Yong Joon Jeon
  • Patent number: 10930288
    Abstract: Aspects of the disclosure provide systems and methods for facilitating dictation. Speech input may be provided to an audio input device of a computing device. A speech recognition engine at the computing device may obtain text corresponding to the speech input. The computing device may transmit the text to a remotely-located storage device. A login webpage that includes a session identifier may be accessed from a target computing device also located remotely relative to the storage device. The session identifier may be transmitted to the storage device and, in response, a text display webpage may be received at the target computing device. The text display webpage may include the speech-derived text and may be configured to automatically copy the text to a copy buffer of the target computing device. The speech-derived text may also be provided to native applications at target computing devices or NLU engines for natural language processing.
    Type: Grant
    Filed: April 7, 2020
    Date of Patent: February 23, 2021
    Assignee: Nuance Communications, Inc.
    Inventors: Markus Vogel, Andreas Neubacher
  • Patent number: 10930263
    Abstract: This disclosure describes techniques for replicating characteristics of an actor or actresses voice across different languages. The disclosed techniques have the practical application of enabling automatic generation of dubbed video content for multiple languages, with particular speakers in each dubbing having the same voice characteristics as the corresponding speakers in the original version of the video content.
    Type: Grant
    Filed: March 28, 2019
    Date of Patent: February 23, 2021
    Assignee: Amazon Technologies, Inc.
    Inventor: Hooman Mahyar
  • Patent number: 10932098
    Abstract: A wireless access point supports media conferencing for wireless User Equipment (UE). The wireless access point wirelessly exchanges timing signaling with the wireless UE to synchronize the wireless UE. After the wireless UE is synchronized, the wireless access point wirelessly exchanges connect signaling with the wireless UE to receive an Establishment Cause and a Wireless Network Identifier from the wireless UE. The wireless access point selects a media conferencing Mobility Management Entity (MME) when the Establishment Cause is associated with the media conferencing MME. The wireless access point selects a data MME based on the Wireless Network Identifier when the Establishment Cause is not associated with the media conferencing MME. The wireless access point exchanges network signaling with the data MME or the media conferencing MME. The wireless access point wirelessly exchanges user data with the wireless UE under control of the data MME or the media conferencing MME.
    Type: Grant
    Filed: February 15, 2018
    Date of Patent: February 23, 2021
    Assignee: Sprint Communications Company L.P.
    Inventor: Rajil Malhotra
  • Patent number: 10930272
    Abstract: A technique for semantic search and retrieval that is event-based, wherein is event is composed of a sequence of observations that are user speech or physical actions. Using a first set of conversations, a machine learning model is trained against groupings of utterances therein to generate a speech act classifier. Observation sequences therein are organized into groupings of events and configured for subsequent event recognition. A set of second (unannotated) conversations are then received. The set of second conversations is evaluated using the speech act classifier and information retrieved from the event recognition to generate event-level metadata that comprises, for each utterance or physical action within an event, one or more associated tags. In response to a query, a search is performed against the metadata. Because the metadata is derived from event recognition, the search is performed against events learned from the set of first conversations.
    Type: Grant
    Filed: October 15, 2020
    Date of Patent: February 23, 2021
    Assignee: Drift.com, Inc.
    Inventors: Jeffrey D. Orkin, Christopher M. Ward, Elias Torres
  • Patent number: 10930270
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for processing audio waveforms. In some implementations, a time-frequency feature representation is generated based on audio data. The time-frequency feature representation is input to an acoustic model comprising a trained artificial neural network. The trained artificial neural network comprising a frequency convolution layer, a memory layer, and one or more hidden layers. An output that is based on output of the trained artificial neural network is received. A transcription is provided, where the transcription is determined based on the output of the acoustic model.
    Type: Grant
    Filed: August 15, 2019
    Date of Patent: February 23, 2021
    Assignee: Google LLC
    Inventors: Tara N. Sainath, Ron J. Weiss, Andrew W. Senior, Kevin William Wilson
  • Patent number: 10923106
    Abstract: An audio synthesis method adapted to video characteristics is provided. The audio synthesis method according to an embodiment includes: extracting characteristics x from a video in a time-series way; extracting characteristics p of phonemes from a text; and generating an audio spectrum characteristic St used to generate an audio to be synthesized with a video at a time t, based on correlations between an audio spectrum characteristic St-1, which is used to generate an audio to be synthesized with a video at a time t?1, and the characteristics x. Accordingly, an audio can be synthesized according to video characteristics, and speech according to a video can be easily added.
    Type: Grant
    Filed: January 24, 2019
    Date of Patent: February 16, 2021
    Assignee: Korea Electronics Technology Institute
    Inventors: Jong Yeol Yang, Young Han Lee, Choong Sang Cho, Hye Dong Jung
  • Patent number: 10923118
    Abstract: An audio input method includes: in an audio-input mode, receiving a first audio input by a user, recognizing the first audio to generate a first recognition result, and displaying corresponding verbal content to the user based on the first recognition result; and in an editing mode, receiving a second audio input by the user and recognizing and generating a second recognition result, converting the second recognition result to an editing instruction, and executing a corresponding operation based on the editing operation. The audio-input mode and the editing mode are switchable.
    Type: Grant
    Filed: November 17, 2016
    Date of Patent: February 16, 2021
    Assignee: BEIJING SOGOU TECHNOLOGY DEVELOPMENT CO., LTD.
    Inventors: Liping Li, Suhang Wang, Congxian Yan, Lei Yang, Min Liu, Hong Zhao, Jia Yao
  • Patent number: 10924605
    Abstract: Systems and methods for providing and facilitating multi-mode communication are disclosed. Users may initiate, receive and/or respond to messages and message notifications on a computing device using multi-mode interactions executed through either a device display or a wearable device such as a headset with enhanced functionality. Contextual prompts guide the user interaction with the computing device using on-board or remote voice recognition text-to-speech and speech-to-text processing and playback. Voice and text data are packaged and transmitted to the network.
    Type: Grant
    Filed: January 11, 2018
    Date of Patent: February 16, 2021
    Assignee: ONVOCAL, INC.
    Inventors: William Wang Graylin, Bogdan Sima, Pichrachana Sun, Andrew Molloy
  • Patent number: 10916249
    Abstract: A method of processing a speech signal for speaker recognition in an electronic apparatus includes: obtaining a speech signal of a first user; extracting a speech feature comprising a feature value from the speech signal; comparing the speech feature extracted from the speech signal of the first user with a predetermined reference value; selecting a first user feature that corresponds to the speech feature of the first user compared with the reference value; generating a recommended phrase used for speaker recognition based on the first user feature; and outputting the recommended phrase.
    Type: Grant
    Filed: February 1, 2019
    Date of Patent: February 9, 2021
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Youngho Han, Keunseok Cho, Jaeyoung Roh, Namhoon Kim, Chiyoun Park, Jongyoub Ryu
  • Patent number: 10917607
    Abstract: This disclosure describes techniques that include modifying text associated with a sequence of images or a video sequence to thereby generate new text and overlaying the new text as captions in the video sequence. In one example, this disclosure describes a method that includes receiving a sequence of images associated with a scene occurring over a time period; receiving audio data of speech uttered during the time period; transcribing into text the audio data of the speech, wherein the text includes a sequence of original words; associating a timestamp with each of the original words during the time period; generating, responsive to input, a sequence of new words; and generating a new sequence of images by overlaying each of the new words on one or more of the images.
    Type: Grant
    Filed: October 14, 2019
    Date of Patent: February 9, 2021
    Assignee: Facebook Technologies, LLC
    Inventors: Vincent Charles Cheung, Marc Layne Hemeon, Nipun Mathur
  • Patent number: 10917519
    Abstract: A method and system to transcribe communications the method comprising the steps of obtaining an audio message originating at a first device during a voice communication session between the first device and a second device, providing the audio message to a first speech recognition system to generate a first transcript of the audio message, directing the first transcript to the second device, in response to obtaining an indication that indicates a quality of the first transcript is below a quality threshold, using a second speech recognition system to generate a second transcript based on the audio message while continuing to provide the audio data to the first speech recognition system to generate the first transcript and, in response to occurrence of an event that indicates the second transcript is to be directed to the second device, directing the second transcript to the second device instead of directing the first transcript.
    Type: Grant
    Filed: August 9, 2019
    Date of Patent: February 9, 2021
    Assignee: Ultratec, Inc.
    Inventors: Robert M. Engelke, Kevin R. Colwell, Christopher Engelke, Robert P Leistiko
  • Patent number: 10909162
    Abstract: Described herein are technologies that facilitate effective use (e.g., indexing and searching) of non-text machine data (e.g., audio/visual data) in an event-based machine-data intake and query system.
    Type: Grant
    Filed: April 28, 2017
    Date of Patent: February 2, 2021
    Assignee: SPLUNK Inc.
    Inventor: Adam Oliner
  • Patent number: 10910001
    Abstract: A voice recognition device including: a recognizer which recognizes a movement of a mouth of an utterer; a detector which detects a noise among a sound around the device; and a controller which controls a voice recognition timing based on the movement of the mouth of the utterer recognized by the recognizer and the noise among the sound around the device detected by the detector.
    Type: Grant
    Filed: December 23, 2018
    Date of Patent: February 2, 2021
    Assignee: CASIO COMPUTER CO., LTD.
    Inventor: Keisuke Shimada
  • Patent number: 10902831
    Abstract: Methods and apparatus to classify media based on a pitch-independent timbre attribute from a media signal are disclosed. An example apparatus includes means for accessing a media signal; and means for: determining a spectrum of audio corresponding to the media signal; and determining a timbre-independent pitch attribute of audio of the media signal based on an inverse transform of a complex argument of a transform of the spectrum.
    Type: Grant
    Filed: March 17, 2020
    Date of Patent: January 26, 2021
    Assignee: The Nielsen Company (US), LLC
    Inventor: Zafar Rafii
  • Patent number: 10891954
    Abstract: Embodiments for managing a voice response system by one or more processors are described. At least one sound is detected. A signal that is representative of at least a portion of the at least one detected sound is received. A voice communication is determined based on the at least one detected sound and the signal. A response to the determined voice communication is determined.
    Type: Grant
    Filed: January 3, 2019
    Date of Patent: January 12, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Shikhar Kwatra, Jeremy Fox, Paul Krystek, Sarbajit Rakshit
  • Patent number: 10891106
    Abstract: Aspects of the subject technology relate to systems and methods for processing voice input data. Voice input data is received from a computing. An intended task is determined based on the received voice input data. Contextual information related to the intended task is obtained. A plurality of services to be accessed at the computing device is determined based on the intended task and the obtained contextual information. Instructions associated with the plurality of services are provided for transmission to the computing device for execution at the computing device.
    Type: Grant
    Filed: October 13, 2015
    Date of Patent: January 12, 2021
    Assignee: Google LLC
    Inventors: Alexander Friedrich Kuscher, Santhosh Balasubramanian, Tiantian Zha
  • Patent number: 10885898
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for receiving audio data including an utterance, obtaining context data that indicates one or more expected speech recognition results, determining an expected speech recognition result based on the context data, receiving an intermediate speech recognition result generated by a speech recognition engine, comparing the intermediate speech recognition result to the expected speech recognition result for the audio data based on the context data, determining whether the intermediate speech recognition result corresponds to the expected speech recognition result for the audio data based on the context data, and setting an end of speech condition and providing a final speech recognition result in response to determining the intermediate speech recognition result matches the expected speech recognition result, the final speech recognition result including the one or more expected speech recognition results indicated b
    Type: Grant
    Filed: September 21, 2017
    Date of Patent: January 5, 2021
    Assignee: Google LLC
    Inventors: Petar Aleksic, Glen Shires, Michael Buchanan
  • Patent number: 10867610
    Abstract: A method for facilitating a remote conference includes receiving a digital video and a computer-readable audio signal. A face recognition machine is operated to recognize a face of a first conference participant in the digital video, and a speech recognition machine is operated to translate the computer-readable audio signal into a first text. An attribution machine attributes the text to the first conference participant. A second computer-readable audio signal is processed similarly, to obtain a second text attributed to a second conference participant. A transcription machine automatically creates a transcript including the first text attributed to the first conference participant and the second text attributed to the second conference participant.
    Type: Grant
    Filed: June 29, 2018
    Date of Patent: December 15, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Adi Diamant, Karen Master Ben-Dor, Eyal Krupka, Raz Halaly, Yoni Smolin, Ilya Gurvich, Aviv Hurvitz, Lijuan Qin, Wei Xiong, Shixiong Zhang, Lingfeng Wu, Xiong Xiao, Ido Leichter, Moshe David, Xuedong Huang, Amit Kumar Agarwal
  • Patent number: 10867596
    Abstract: A voice assistant system includes a server apparatus performing voice assistant and a plurality of devices, in which the server apparatus and the devices are communicatively connected to each other. The plurality of devices each records the same user's speech through a microphone, and then transmits recorded data of the same user's speech to the server apparatus. The server apparatus receives the recorded data transmitted from each of the plurality of devices, and then voice-recognizes two or more of the received recorded data in accordance with a predetermined standard to thereby interpret the contents of the user's speech to perform the voice assistant.
    Type: Grant
    Filed: August 9, 2018
    Date of Patent: December 15, 2020
    Assignee: Lenovo (Singapore) PTE. LTD.
    Inventors: Masaharu Yoneda, Kazuhiro Kosugi, Koji Kawakita
  • Patent number: 10861438
    Abstract: Methods and systems for correcting transcribed text. One method includes receiving audio data from one or more audio data sources and transcribing the audio data based on a voice model to generate text data. The method also includes making the text data available to a plurality of users over at least one computer network and receiving corrected text data over the at least one computer network from the plurality of users. In addition, the method can include modifying the voice model based on the corrected text data.
    Type: Grant
    Filed: November 29, 2017
    Date of Patent: December 8, 2020
    Assignee: III Holdings 1, LLC
    Inventor: Paul M. Hager
  • Patent number: 10860786
    Abstract: The growing amount of communication data generated by inmates in controlled environments makes a timely and effective investigation and analysis more and more difficult. The present disclosure provides details of a system and method to investigate and analyze the communication data in a correctional facility timely and effectively. Such a system receives both real time communication data and recorded communication data, processes and investigates the data automatically, and stores the received communication data and processed communication data in a unified data server. Such a system enables a reviewer to review, modify and insert markers and comments for the communication data. Such a system further enables the reviewer to search the communication data and create scheduled search reports.
    Type: Grant
    Filed: June 1, 2017
    Date of Patent: December 8, 2020
    Assignee: Global Tel*Link Corporation
    Inventor: Stephen Lee Hodge
  • Patent number: 10860526
    Abstract: Disclosed are systems and software that provide a high-performance, extensible file format and web API for remote data access and a visual interface for data viewing, query, and analysis. The described system can support storage of raw spectroscopic data such as neural recording data, MSI data, metadata, and derived analyses in a single, self-describing format that may be compatible by a large range of analysis software.
    Type: Grant
    Filed: November 27, 2013
    Date of Patent: December 8, 2020
    Assignee: The Regents of the University of California
    Inventors: Benjamin P. Bowen, Oliver Ruebel
  • Patent number: 10861451
    Abstract: One embodiment provides a method, including: receiving, at an information handling device, an audible command to perform a function; determining, using a processor, at least one aspect associated with the audible command that prevents performance of the function; and providing, based on the determining, a suggested modification to the audible command. Other aspects are described and claimed.
    Type: Grant
    Filed: March 22, 2018
    Date of Patent: December 8, 2020
    Assignee: Lenovo (Singapore) Pte. Ltd.
    Inventors: John Carl Mese, Nathan J. Peterson, Russell Speight VanBlon
  • Patent number: 10861444
    Abstract: Systems and methods are described for determining whether to activate a voice activated device based on a speaking cadence of the user. When the user speaks with a first cadence the system may determine that the user does not intend to activate the device and may accordingly not to trigger a voice activated device. When the user speaks with a second cadence the system may determine that the user does wish to trigger the device and may accordingly trigger the voice activated device.
    Type: Grant
    Filed: September 24, 2018
    Date of Patent: December 8, 2020
    Assignee: Rovi Guides, Inc.
    Inventors: Edison Lin, Rowena Young, Kanchan Sripathy, Reda Harb
  • Patent number: 10853024
    Abstract: An electronic device is provided. The electronic device includes a microphone, a touch screen display, a processor, and memory.
    Type: Grant
    Filed: March 21, 2019
    Date of Patent: December 1, 2020
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Yoonjung Choi, Iseul Song
  • Patent number: 10847152
    Abstract: An electronic device is provided. The electronic device includes a communication module, a microphone receiving a voice input according to user speech, a memory storing information about an operation of the speech recognition service, a display, and a processor electrically connected with the communication module, the microphone, the memory, and the display. The processor is configured to calculate a specified numerical value associated with the operation of the speech recognition service, to transmit information about the numerical value to a first external device processing the voice input, and to transmit a request for a function, which corresponds to the calculated numerical value, of at least one function associated with the speech recognition service stepwisely provided from the first external device depending on a numerical value, to the first external device to refine a function of the speech recognition service supported by the electronic device.
    Type: Grant
    Filed: March 28, 2018
    Date of Patent: November 24, 2020
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Byong Mok Oh, Han Jun Ku, Kyoung Gu Woo, Sang Hoon Lee, In Jong Rhee, Seok Yeong Jung, Kyung Tae Kim, Ji Hyun Kim, Dong Ho Jang
  • Patent number: 10838779
    Abstract: Disclosed is a system for automatically executing multiple functional units of operation from multiple applications to complete a task that requires data from a variety of sources. The system stores information about operations that may be executed on a client device. When executing a sequence of operations, for example in response to a user request, the system may automatically store output from each operation. When executing subsequent operations in the sequence of operations, the system can determine whether output from a previously executed operation is appropriate for use as input to a new operation. The system can thus save data from one application and automatically use that data as input to a different application. Additionally, the system may determine subsequent operations to execute in response to previously observed output values.
    Type: Grant
    Filed: December 15, 2017
    Date of Patent: November 17, 2020
    Assignee: Brain Technologies, Inc.
    Inventors: Sheng Yue, Yuan Lin
  • Patent number: 10839809
    Abstract: Bandwidth-efficient (i.e., compressed) representations of audio data can be utilized for near real-time presentation of the audio on one or more receiving devices. Persons identified as having speech represented in the audio data can have trained speech models provided to the devices. These trained models can be used to classify the compressed audio in order to improve the quality to correspond more closely to the uncompressed version, without experiencing lag that might otherwise be associated with transmission of the uncompressed audio. The uncompressed audio is also received, with potential lag, and is used to further train the speech models in near real time. The ability to utilize the uncompressed audio as it is received prevents a need to store or further transmit the audio data for offline processing, and enables the further trained model to be used during the communication session.
    Type: Grant
    Filed: December 12, 2017
    Date of Patent: November 17, 2020
    Assignee: AMAZON TECHNOLOGIES, INC.
    Inventors: Madhav Jha, Edo Liberty
  • Patent number: 10839023
    Abstract: Provided is an avatar service system and method that are provided through a network. The avatar service system may include a request receiving unit to receive a request for an avatar to perform an action, a data extracting unit to extract metadata and image data corresponding to the request from the database storing the metadata with respect to the action of the avatar and the image data for a plurality of layers forming the avatar, and an avatar action processing unit to generate and provide action data for applying, to the avatar, the action of the avatar corresponding to the request using the extracted metadata and the extracted image data.
    Type: Grant
    Filed: November 8, 2017
    Date of Patent: November 17, 2020
    Assignee: LINE Corporation
    Inventors: Seung Young Lee, Changhoon Shin, Suk Kyoung Eom
  • Patent number: 10839788
    Abstract: Systems and methods for accent and dialect modification are disclosed. Discussed are a method for selecting a target dialect and accent to use to modify voice communications based on a context and a method for selectively modifying one or more words in voice communications in one dialect and accent with one or more vocal features of a different accent.
    Type: Grant
    Filed: December 13, 2018
    Date of Patent: November 17, 2020
    Assignee: I2X GMBH
    Inventors: Christoph Johann Feinauer, Samuel Frank Jet Brown
  • Patent number: 10841250
    Abstract: A user profile and messaging bots can be joined to participate in a group chat session. A first natural language requirement set can be received from the profile and sent to the bots. A first solution set from a first bot can be processed, with the set including a first natural language script describing a first solution to the requirement set. A second solution set from a second bot can be processed, with the set including a second natural language script that describes a second solution to the requirement set. The first and second scripts can be sent to the profile, and at least part of each solution set can be sent to the other bot. A selection of the first bot for providing a solution to the requirement set can be received from the profile, and an indication of the selection can be transmitted to the first bot.
    Type: Grant
    Filed: June 11, 2019
    Date of Patent: November 17, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Yuval Pinchas Borsutsky, Keren Damari, Benny Schlesinger, Avichai Cohen
  • Patent number: 10834145
    Abstract: A method, computer program product, and computer system for monitoring, at a computing device, at least a portion of a collaboration session provided by one or more participants of the collaboration session. At least the portion of the collaboration session is analyzed to determine a recommendation associated with at least the portion of the collaboration session. A source for information associated with the recommendation is searched based upon, at least in part, analyzing at least the portion of the collaboration session. At least one participant of the one or more participants is presented the recommendation and the information associated with the recommendation.
    Type: Grant
    Filed: April 3, 2019
    Date of Patent: November 10, 2020
    Assignee: International Business Machines Corporation
    Inventors: Donald E. Buddenbaum, Peter F. Haggar, Heather M. Kreger, Arnaud J. Le Hors, John V. Meegan, Keith A. Wells
  • Patent number: 10832679
    Abstract: One embodiment provides a computer program product for improving accuracy of a transcript of a spoken interaction. The computer program product comprises a computer readable storage medium having program instructions embodied therewith. The program instructions are executable by a processor to cause the processor to identify a plurality of patterns in the transcript. The plurality of patterns are indicative of a group of acoustically similar words in the transcript and a corresponding local, sequential context of the group of acoustically similar words. The program instructions are further executable by the processor to cause the processor to predict conditional probabilities for the group of acoustically similar words based on a predictive model and the plurality of patterns, detect one or more transcription errors in the transcript based on the conditional probabilities, and correct the one or more transcription errors by applying a multi-pass correction on the one or more transcription errors.
    Type: Grant
    Filed: November 20, 2018
    Date of Patent: November 10, 2020
    Assignee: International Business Machines Corporation
    Inventors: Margaret H. Szymanski, Robert J. Moore, Sunhwan Lee, Pawan Chowdhary, Shun Jiang, Guangjie Ren, Raphael Arar
  • Patent number: 10825456
    Abstract: A method and apparatus are provided for assisting a text writing operation by using voice recognition. The method includes displaying an input text according to a key input or a touch input in a text writing mode on a text display window; recognizing a voice input while displaying the input text according to the key input or the touch input on the text display window; and assisting a preset text writing operation according to the recognized voice input while displaying the input text according to the key input or the touch input on the text display window. Assisting the preset text writing operation comprises, in response to a first part of the recognized voice input matching a pre-stored command, displaying a result obtained based on a second part of the recognized voice input, together with the input text according to the key input or the touch input, on the text display window.
    Type: Grant
    Filed: January 15, 2019
    Date of Patent: November 3, 2020
    Assignee: Samsung Electronics Co., Ltd
    Inventor: Sung-Joon Won
  • Patent number: 10812422
    Abstract: A head-mounted device (HMD) of a first user has a transparent display. The HMD determines location information of a second user relative to the HMD of the first user. The second user is located within a predefined distance of the HMD. The location information identifies a distance and a direction of the second user relative to the HMD. The HMD receives audio content from the second user, generates augmented reality (AR) content based on the audio content, and displays the AR content in the transparent display based on the location information of the second user. The AR content appears coupled to the second user.
    Type: Grant
    Filed: August 31, 2017
    Date of Patent: October 20, 2020
    Assignee: RPX Corporation
    Inventor: Brian Mullins
  • Patent number: 10811009
    Abstract: An utterance is analyzed to identify an absence of a known invocation phrase. A skill set is constructed in response to the absence, the skill set including a first skill corresponding to the utterance and a first skill score corresponding to a likelihood that the first skill corresponds to the utterance. The first skill score is adjusted, based on the presence of the first skill in a skill history, where the skill history stores a set of history skills in an order of recency of use of each history skill in the set of history skills. The first skill score is adjusted, based on an association of the first skill with a default installed skill. An installed skill is selected, based on the adjusted first skill score, the installed skill performing an action in response to the utterance.
    Type: Grant
    Filed: June 27, 2018
    Date of Patent: October 20, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Bryce A. Curtis, Stewart Nickolas
  • Patent number: 10796687
    Abstract: Methods, systems, and computer-readable media for voice-activated selective memory for voice-capturing devices are disclosed. A first voice input from a voice-capturing device is received, via a network, at a service provider environment comprising one or more services. The first voice input comprises one or more utterances from a user of the voice-capturing device. A representation of the first voice input is stored. A second voice input from the voice-capturing device is received, via the network, at the service provider environment. The second voice input represents a command to disregard the first voice input. Based on the second voice input, the representation of the first voice input is deleted.
    Type: Grant
    Filed: September 6, 2017
    Date of Patent: October 6, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Jonathan Alan Leblang, Kevin Crews, Qi Zhou, Gary Zhong
  • Patent number: 10789045
    Abstract: A non-transitory computer-readable recording medium storing computer-readable instructions may be read by a controller of an information processing device in which an image processing program and a voice-recognition program have been installed. The computer-readable instructions may cause the information processing device to obtain the text data from the voice-recognition program in a case where the voice-recognition program creates the text data. The computer-readable instructions may cause the information processing device to determine whether the obtained text data includes a first string and a second string. The computer-readable instructions may cause the information processing device to cause a display of the information processing device to display a specific image in a case where it is determined that the text data includes the first string and the second string.
    Type: Grant
    Filed: December 2, 2019
    Date of Patent: September 29, 2020
    Assignee: Brother Kogyo Kabushiki Kaisha
    Inventor: Yu Matsuo
  • Patent number: 10789954
    Abstract: According to one or more aspects of the present disclosure, operations related to providing transcriptions may include obtaining a first transcription of first audio obtained by a first device during a communication session conducted between the first device and a second device. The operations may further include providing the first transcription for presentation of the first transcription by a display device during the communication session. In addition, the operations may include providing, in response to a transcription quality indication, a second transcription of second audio obtained by the second device during the communication session for presentation of the second transcription by the display device during the communication session.
    Type: Grant
    Filed: August 29, 2018
    Date of Patent: September 29, 2020
    Assignee: Sorenson IP Holdings, LLC
    Inventor: Michael Holm