Word Recognition Patents (Class 704/251)
  • Patent number: 11222175
    Abstract: A method, system and computer program product for recognizing terms in a specified corpus. In one embodiment, the method comprises providing a set of known terms t?T, each of the known terms t belonging to a set of types ? (t)={?1, . . . }, wherein each of the terms is comprised of a list of words, t=w1, w2, . . . , wn, and the union of all the words for all the terms is a word set W. The method further comprises using the set of terms T and the set of types to determine a set of pattern-to-type mappings p??; and using the set of pattern-to-type mappings to recognize terms in the specified corpus and, for each of the recognized terms in the specified corpus, to recognize one or more of the types ? for said each recognized term.
    Type: Grant
    Filed: May 24, 2019
    Date of Patent: January 11, 2022
    Assignee: International Business Machines Corporation
    Inventors: Michael Glass, Alfio M Gliozzo
  • Patent number: 11222623
    Abstract: A speech keyword recognition method includes: obtaining first speech segments based on a to-be-recognized speech signal; obtaining first probabilities respectively corresponding to the first speech segments by using a preset first classification model. A first probability of a first speech segment is obtained from probabilities of the first speech segment respectively corresponding to pre-determined word segmentation units of a pre-determined keyword.
    Type: Grant
    Filed: May 27, 2020
    Date of Patent: January 11, 2022
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Jun Wang, Dan Su, Dong Yu
  • Patent number: 11222635
    Abstract: An electronic device of the present invention comprises: a housing; a touchscreen display; a microphone; at least one speaker; a button disposed on a portion of the housing or set to be displayed on the touchscreen display; a wireless communication circuit; a processor; and a memory. The electronic device is configured to store an application program including a user interface for receiving a text input. When the user interface is not displayed on the touchscreen display, the electronic device enables a user to receive a user input through the button, receives user speech through the microphone, and then provides data on the user speech to an external server including an automatic speech recognition system and an intelligence system. An instruction for performing a task generated by the intelligence system in response to the user speech is received from the server.
    Type: Grant
    Filed: February 1, 2018
    Date of Patent: January 11, 2022
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Sang-Ki Kang, Jang-Seok Seo, Kook-Tae Choi, Hyun-Woo Kang, Jin-Yeol Kim, Chae-Hwan Li, Kyung-Tae Kim, Dong-Ho Jang, Min-Kyung Hwang
  • Patent number: 11211046
    Abstract: A mistranscription generated by a speech recognition system is identified. A received utterance is matched to a first utterance member within a set of known utterance members. The matching operation matches fewer than the first plural number of words in the received utterance and the received utterance varies in a first particular manner as compared to a first word in a first slot in the first utterance member. The received utterance is sent to a mistranscription analyzer component which increments evidence that the received utterance is evidence of a mistranscription. Once the incremented evidence for the mistranscription exceeds a threshold, future received utterances containing the mistranscription are treated as though the first word was recognized.
    Type: Grant
    Filed: January 13, 2020
    Date of Patent: December 28, 2021
    Assignee: International Business Machines Corporation
    Inventors: Andrew Aaron, Shang Guo, Jonathan Lenchner, Maharaj Mukherjee
  • Patent number: 11205415
    Abstract: An electronic apparatus which includes a memory configured to store first voice recognition information related to a first language and second voice recognition information related to a second language, and a processor to obtain a first text corresponding to a user voice that is received on the basis of first voice recognition information. The processor, based on an entity name being included in the user voice according to the obtained first text, identifies a segment in the user voice in which the entity name is included, and obtains a second text corresponding to the identified segment of the user voice on the basis of the second voice recognition information, and obtains control information corresponding to the user voice on the basis of the first text and the second text.
    Type: Grant
    Filed: October 25, 2019
    Date of Patent: December 21, 2021
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Chansik Bok, Jihun Park
  • Patent number: 11200909
    Abstract: A method is disclosed. The proposed method includes: providing an initial speech corpus including plural utterances; based on a condition of maximum a posteriori (MAP), according to respective sequences of syllable duration, syllable duration prosodic state, syllable tone, base-syllable type, and break type of the kth utterance, using a probability of an ISR of the kth utterance xk to estimate an estimated value {circumflex over (x)}k of the xk; and through the MAP condition, according to respective sequences of syllable duration, syllable duration prosodic state, syllable tone, base-syllable type, and break type of the given lth breath group/prosodic phrase group (BG/PG) of the kth utterance, using a probability of an ISR of the lth BG/PG of the kth utterance xk,l to estimate an estimated value {circumflex over (x)}k,l of the xk,l wherein the {circumflex over (x)}k,l is the estimated value of local ISR, and a mean of a prior probability model of the {circumflex over (x)}k,l is the {circumflex over (x)}k.
    Type: Grant
    Filed: August 30, 2019
    Date of Patent: December 14, 2021
    Assignee: NATIONAL YANG MING CHIAO TUNG UNIVERSITY
    Inventors: Chen-Yu Chiang, Guan-Ting Liou, Yih-Ru Wang, Sin-Horng Chen
  • Patent number: 11200382
    Abstract: This application discloses a prosodic pause prediction method, a prosodic pause prediction device and an electronic device. The specific implementation scheme includes: obtaining a first matrix by mapping a to-be-tested text sequence through a trained embedding layer, where the to-be-tested text sequence includes a to-be-tested input text and an identity of a to-be-tested speaker; inputting the first matrix into a trained attention model, and determining a semantic representation matrix by the trained attention model; and, performing prosodic pause prediction based on the semantic representation matrix and outputting a prosodic pause prediction result of each word in the to-be-tested input text.
    Type: Grant
    Filed: May 8, 2020
    Date of Patent: December 14, 2021
    Assignee: Baidu Online Network Technology (Beijing) Co., Ltd.
    Inventors: Zhipeng Nie, Yanyao Bian, Zhanjie Gao, Changbin Chen
  • Patent number: 11194825
    Abstract: A distributed sequential pattern data mining framework mines user data to determine statistically-relevant sequential patterns which are used to correlate the sequential patterns to a particular outcome. The correlation is provided by a statistical model, a binary predictive model and/or a logistic regression model which uses the sequential patterns to learn the behavior of end users during their usage of a software application.
    Type: Grant
    Filed: September 23, 2018
    Date of Patent: December 7, 2021
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC.
    Inventors: Shengyu Fu, Sai Tulasi Neppali, Neelakantan Sundaresan, Siyu Yang
  • Patent number: 11188923
    Abstract: Aspects of the disclosure relate to real-time knowledge-based widget prioritization and display. A computing platform may detect, via a computing device, a voice-based interaction between an enterprise agent and a customer. Then, the computing platform may cause, via the computing device, the voice-based interaction to be captured as audio data. The computing platform may then transform the audio data to textual data. Subsequently, the computing platform may identify, in the textual data, a customer query. Then, the computing platform may retrieve, in real-time and based on the voice-based interaction and from a repository of widgets, a first widget, where the first widget includes information at least partially responsive to the customer query. Then, the computing platform may display, to the enterprise agent and via a graphical user interface in use by the enterprise agent, the first widget.
    Type: Grant
    Filed: August 29, 2019
    Date of Patent: November 30, 2021
    Assignee: Bank of America Corporation
    Inventors: Gaurav Bansal, Shekhar Singh Mehra, Vinod Maghnani, Sandeep Kumar Chauhan
  • Patent number: 11189288
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for processing multimodal input. A system configured to practice the method continuously monitors an audio stream associated with a gesture input stream, and detects a speech event in the audio stream. Then the system identifies a temporal window associated with a time of the speech event, and analyzes data from the gesture input stream within the temporal window to identify a gesture event. The system processes the speech event and the gesture event to produce a multimodal command. The gesture in the gesture input stream can be directed to a display, but is remote from the display. The system can analyze the data from the gesture input stream by calculating an average of gesture coordinates within the temporal window.
    Type: Grant
    Filed: January 15, 2020
    Date of Patent: November 30, 2021
    Assignee: Nuance Communications, Inc.
    Inventors: Michael Johnston, Derya Ozkan
  • Patent number: 11189287
    Abstract: Provided are an optimization method, apparatus, device for a wake-up model and a storage medium, which allow for: acquiring a training set and a verification set; performing an iterative training on the wake-up model according to the training set and the verification set; during the iterative training, periodically updating the training set and the verification set according to the wake-up model and a preset corpus database, and continuing performing the iterative training on the wake-up model according to the updated training set and verification set; and outputting the wake-up model when a preset termination condition is reached. The embodiments of the present disclosure, by periodically updating the training set and the verification set according to the wake-up model and the preset corpus database during an iteration, may improve optimization efficiency and effects of the wake-up model, thereby improving stability and adaptability of the wake-up model and avoiding overfitting.
    Type: Grant
    Filed: December 4, 2019
    Date of Patent: November 30, 2021
    Assignees: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD., SHANGHAI XIAODU TECHNOLOGY CO. LTD.
    Inventor: Yongchao Zhang
  • Patent number: 11183187
    Abstract: The present invention provides a dialog system comprising a speech receiving step in which the dialog system receives input of a speech of a human, a first speech determination step in which the dialog system determines a first speech which is a speech in response to the speech of the human, a first speech presentation step in which the first speech is presented by a first agent, a reaction acquisition step in which the dialog system acquires a reaction of the human to the first speech, a second speech determination step in which the dialog system determines, when the reaction of the human is a reaction indicating that the first speech is not a speech in response to the speech of the human, a second speech which is different from the first speech, and a second speech presentation step in which the second speech is presented by the second agent.
    Type: Grant
    Filed: May 19, 2017
    Date of Patent: November 23, 2021
    Assignees: NIPPON TELEGRAPH AND TELEPHONE CORPORATION, OSAKA UNIVERSITY
    Inventors: Hiroaki Sugiyama, Toyomi Meguro, Junji Yamato, Yuichiro Yoshikawa, Hiroshi Ishiguro
  • Patent number: 11182555
    Abstract: A sequence processing method and apparatus are provided. The sequence processing method includes determining a word of a first R-node corresponding to a root node based on an input sequence, generating first I-nodes that are connected to the first R-node and include relative position information with respect to the word of the first R-node, determining a word of a second R-node to correspond to each of the first I-nodes, and determining an output sequence corresponding to the input sequence based on the determined words.
    Type: Grant
    Filed: April 9, 2020
    Date of Patent: November 23, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Hwidong Na, Min-Joong Lee
  • Patent number: 11176141
    Abstract: An aspect provides a method, including: receiving, at an input component of an information handling device, user input comprising one or more words; identifying, using a processor of the information handling device, an emotion associated with the one or more words; creating, using the processor, an emotion tag including the emotion associated with the one or more words; storing the emotion tag in a memory; analyzing one or more emotion tags; and modifying an operation of an application based on the analyzing. Other embodiments are described and claimed.
    Type: Grant
    Filed: May 16, 2016
    Date of Patent: November 16, 2021
    Assignee: Lenovo (Singapore) Pte. Ltd.
    Inventors: Suzanne Marion Beaumont, Russell Speight VanBlon, Rod D. Waltermann
  • Patent number: 11176214
    Abstract: Methods, apparatuses, and computer program products are described herein that are configured to express a linguistic description of set of points within a spatial area in an output text. In some example embodiments, a method is provided that comprises generating one or more descriptors and/or one or more combinations of descriptors that are configured to linguistically describe at least a portion of a set of points within a spatial area. The method of this embodiment may also include scoring each of the one or more descriptors and/or one or more combinations of the one or more descriptors. The method of this embodiment may also include selecting a descriptor or combination of descriptors that has the highest score when compared to other descriptors or combination of descriptors, providing the descriptor or combination of descriptors satisfies a threshold.
    Type: Grant
    Filed: May 1, 2015
    Date of Patent: November 16, 2021
    Assignee: ARRIA DATA2TEXT LIMITED
    Inventors: Gowri Somayajulu Sripada, Neil Burnett
  • Patent number: 11176520
    Abstract: A method may include configuring a processor to monitor, in an application, composition of an electronic communication addressed to a second user from a first user, the electronic communication associated with a set of parameters; determine an intent of the electronic communication based on the set of parameters; search an associative data structure to retrieve content associated with the intent, the content previously transmitted to a third user from the first user or content(s) received from a fourth user(s); and present a suggestion in the application to include the retrieved content in the electronic communication
    Type: Grant
    Filed: April 18, 2019
    Date of Patent: November 16, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventor: Manoj Ramakrishnan
  • Patent number: 11164562
    Abstract: A system for entity-level clarification in conversation services includes a memory having instructions therein. The system also includes at least one processor in communication with the memory. The at least one processor is configured to execute the instructions to receive a conversation services training example set, build an entity usage map using the conversation services training example set, receive a user utterance, and, responsive to a reception of the user utterance, generate a clarification response using the entity usage map. The at least one processor is also configured to execute the instructions to provide the clarification response to a user.
    Type: Grant
    Filed: January 10, 2019
    Date of Patent: November 2, 2021
    Assignee: International Business Machines Corporation
    Inventors: Carmine M. DiMascio, Donna K. Byron, Benjamin L. Johnson, Florian Pinel
  • Patent number: 11164561
    Abstract: A method and system for building a speech recognizer, and a speech recognition method and system are proposed. The method for building a speech recognizer includes: reading and parsing each grammar file, and building a network of each grammar; reading an acoustic syllable mapping relationship table, and deploying the network of each grammar as a syllable network; performing a merge minimization operation for each syllable network to form a sound element decoding network; forming the speech recognizer by using the sound element decoding network and a language model. The technical solutions of the present disclosure may be applied to exhibit strong extensibility, support an N-Gram language model, support a class model, present flexible use, and adapt for an embedded recognizer in a vehicle-mounted environment.
    Type: Grant
    Filed: August 19, 2019
    Date of Patent: November 2, 2021
    Assignee: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD.
    Inventors: Zhijian Wang, Sheng Qian
  • Patent number: 11159685
    Abstract: A display control device includes a display section, a first receiving section, a second receiving section, and a performing section. The display section displays an object. The first receiving section receives non-voice input specifying a first operation on the object. The second receiving section receives voice input specifying a second operation on the object. The performing section performs, on the object, a complex operation specified by the non-voice input and the voice input.
    Type: Grant
    Filed: March 27, 2020
    Date of Patent: October 26, 2021
    Assignee: KYOCERA Document Solutions Inc.
    Inventors: Nobuto Fujita, Kenji Kiyose, Sumio Yamada, Takayuki Mashimo, Ryota Seike, Koji Kuroda
  • Patent number: 11151988
    Abstract: Techniques for implementing multiple wakeword detectors on a single device are described. A digital signal processor (DSP) of the device may implement a wakeword detection component to detect when captured speech includes a wakeword. A companion application installed on the device may implement a wakeword detection component trained using speech of a user of the device. If the DSP's wakeword detection component detects a wakeword in speech, the companion application's wakeword detection component may be used to determine whether the wakeword was spoken by the user of the device. If the companion application's wakeword detection component determines the user spoke the wakeword, audio data representing the speech may be sent to at least one server(s) for processing.
    Type: Grant
    Filed: January 31, 2020
    Date of Patent: October 19, 2021
    Assignee: Amazon Technolgies, Inc.
    Inventors: Deepak Yavagal, Ajith Prabhakara, John Gray
  • Patent number: 11137978
    Abstract: An electronic device includes a processor, and a memory. The memory may store instructions that, cause the processor to display a user interface including items, receive a first user utterance while the user interface is displayed, wherein the first user utterance includes a first request for executing a first task by using at least one item, transmit first data related to the first user utterance to an external server, receive a first response from the external server, wherein the first response includes information on a first sequence of states of the electronic device for executing the first task and further includes numbers and locations of the items in the user interface, and execute the first task including an operation of allowing the application program to select the one or the plurality of items based on the numbers or the locations.
    Type: Grant
    Filed: April 27, 2018
    Date of Patent: October 5, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Kwang Yong Lee, Jung Hoe Kim, Soo Bin Park, Kyoung Gu Woo, Seong Min Je
  • Patent number: 11120802
    Abstract: An approach is provided that receives an audio stream and utilizes a voice activation detection (VAD) process to create a digital audio stream of voices from at least two different speakers. An automatic speech recognition (ASR) process is applied to the digital stream with the ASR process resulting in the spoken words to which a speaker turn detection (STD) process is applied to identify a number of speaker segments with each speaker segment ending at a word boundary. A speaker clustering algorithm is then applied to the speaker segments to associate one of the speakers with each of the speaker segments.
    Type: Grant
    Filed: November 21, 2017
    Date of Patent: September 14, 2021
    Assignee: International Business Machines Corporation
    Inventors: Kenneth W. Church, Dimitrios B. Dimitriadis, Petr Fousek, Miroslav Novak, George A. Saon
  • Patent number: 11113098
    Abstract: The present disclosure relates to the field of a multi-chip system, and provides an interrupt processing method, a master chip, a slave chip, and a multi-chip system. An interrupt processing method is applied to a master chip and includes: when an interrupt transport request sent by a slave chip through an interrupt line is detected, obtaining all current interrupt requests (irq_s_1-irq_s_N) of the slave chip, the interrupt request (irq_s_1_-irq_s_N) is generated by a first peripheral (4) of the slave chip; obtaining an interrupt subroutine corresponding to each of the interrupt requests (irq_s_1-irq_s_N), and processing the corresponding interrupt request (irq_s_1-irq_s_N) by using the interrupt subroutine. In the embodiments of the present disclosure, all the interrupt requests (irq_s_1-irq_s_N) of the slave chip are mapped to the master chip, so that the interrupt processing flow of the peripheral on the slave chip is simplified.
    Type: Grant
    Filed: November 26, 2019
    Date of Patent: September 7, 2021
    Assignee: SHENZHEN GOODIX TECHNOLOGY CO., LTD.
    Inventors: Zhibing Liang, Yifan Li, Zekai Chen
  • Patent number: 11113672
    Abstract: A system and method to provide computer support for a meeting of invitees comprises accessing one or more sensory data streams providing digitized sensory data responsive to an activity of one or more of the invitees during the meeting, the one or more sensory data streams including at least one audio stream. The method also comprises subjecting the at least one audio stream to phonetic and situational computer modeling to recognize a sequence of words in the audio stream and to assign each word to an invitee, subjecting the sequence of words to semantic computer modeling to recognize a sequence of directives in the sequence of words, and releasing one or more output data streams based on the sequence of directives, the one or more output data streams including one or more notifications.
    Type: Grant
    Filed: March 22, 2018
    Date of Patent: September 7, 2021
    Inventors: Robert Alexander Sim, Marcello Mendes Hasegawa, Ryen William White, Mudit Jain, Tomer Hermelin, Adi Gerzi Rosenthal, Sagi Hilleli
  • Patent number: 11114100
    Abstract: Methods, apparatus, and computer readable media are described related to automated assistants that proactively incorporate, into human-to-computer dialog sessions, unsolicited content of potential interest to a user. In various implementations, based on content of an existing human-to-computer dialog session between a user and an automated assistant, an entity mentioned by the user or automated assistant may be identified. Fact(s)s related to the entity or to another entity that is related to the entity may be identified based on entity data contained in database(s). For each of the fact(s), a corresponding measure of potential interest to the user may be determined. Unsolicited natural language content may then be generated that includes one or more of the facts selected based on the corresponding measure(s) of potential interest. The automated assistant may then incorporate the unsolicited content into the existing human-to-computer dialog session or a subsequent human-to-computer dialog session.
    Type: Grant
    Filed: August 23, 2019
    Date of Patent: September 7, 2021
    Assignee: GOOGLE LLC
    Inventors: Vladimir Vuskovic, Stephan Wenger, Zineb Ait Bahajji, Martin Baeuml, Alexandru Dovlecel, Gleb Skobeltsyn
  • Patent number: 11109104
    Abstract: Novel techniques are described for viewer compositing using media playback systems for enhanced media recommendation and consumption. For example, a display device can be in communication with a media recommendation and consumption compositor (MRCC) system. When a group of viewers desires a shared media consumption experience, the MRCC system can detect the group of viewers and can obtain respective viewer profiles, which can be used to generate a composite profile representing a composite of the group of viewers. The MRCC system can determine an available content space indicating the content available for consumption and can compute a content recommendation space as a function of the composite viewer profile and the available content space that defines recommended content options for the composited group of viewers. A recommendation interface can be output to indicate recommended content options for selecting and viewing.
    Type: Grant
    Filed: July 10, 2019
    Date of Patent: August 31, 2021
    Assignee: DISH Network L.L.C.
    Inventor: John Rishea
  • Patent number: 11100296
    Abstract: Provided is a processor-implemented method of generating a natural language, the method including generating a latent variable from an embedding vector that corresponds to an input utterance, determining attention information related to the input utterance by applying the generated latent variable to a neural network model, and outputting a natural language response that corresponds to the input utterance based on the calculated attention information.
    Type: Grant
    Filed: July 16, 2018
    Date of Patent: August 24, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Jehun Jeon, Young-Seok Kim, Sang Hyun Yoo, Junhwi Choi
  • Patent number: 11086596
    Abstract: Provided are a display apparatus, a control method thereof, a server, and a control method thereof. The display apparatus includes: a processor which processes a signal; a display which displays an image based on the processed signal; a first command receiver which receives a voice command; a storage which stores a plurality of voice commands said by a user; a second command receiver which receives a user's manipulation command; and a controller which, upon receiving the voice command, displays a list of the stored plurality of voice commands, selects one of the plurality of voice commands of the list according to the received user's manipulation command and controls the processor to process based on the selected voice command.
    Type: Grant
    Filed: September 11, 2018
    Date of Patent: August 10, 2021
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Do-wan Kim, Oh-yun Kwon, Tae-hwan Cha
  • Patent number: 11087760
    Abstract: A system of multi-modal transmission of packetized data in a voice activated data packet based computer network environment is provided. A natural language processor component can parse an input audio signal to identify a request and a trigger keyword. Based on the input audio signal, a direct action application programming interface can generate a first action data structure, and a content selector component can select a content item. An interface management component can identify first and second candidate interfaces, and respective resource utilization values. The interface management component can select, based on the resource utilization values, the first candidate interface to present the content item. The interface management component can provide the first action data structure to the client computing device for rendering as audio output, and can transmit the content item converted for a first modality to deliver the content item for rendering from the selected interface.
    Type: Grant
    Filed: November 26, 2019
    Date of Patent: August 10, 2021
    Assignee: Google, LLC
    Inventors: Gaurav Bhaya, Robert Stets
  • Patent number: 11075876
    Abstract: Embodiments provide a social networking platform offering various services, such as, facilitating aggregation and management of a user's interaction on one or more social networking platforms, offering enhanced control over the level of privacy associated with the flow of user data, offering tools to customize the user's exposure to advertisement-related content on the social networking platform(s), integrating features to control aspects of how data/content is presented to and visualized by the user, empowering the user to multicast direct messages to other users without the other users having to meet certain constraints, empowering the user to create and/or join a group based on messaging threads, and the like. One or more of these enhanced services/features are associated with a powerful framework of authentication/permission model for access control.
    Type: Grant
    Filed: April 15, 2021
    Date of Patent: July 27, 2021
    Assignee: SGROUPLES, INC.
    Inventors: Jonathan Wolfe, Mark Weinstein
  • Patent number: 11068655
    Abstract: A text recognition method and apparatus, and a storage medium are provided. The method includes: obtaining sample text data, the sample text data comprising a plurality of sample phrases; and generating a recognition model based on the sample phrases by performing training on a plurality of training nodes. Generating the recognition model includes respectively obtaining, by each of the plurality of training nodes, recognition coefficients of the sample phrases distributed to the corresponding training node; and determining, by the plurality of training nodes, model parameters of the recognition model according to the recognition coefficients of the sample phrases. The method also includes obtaining to-be-recognized text data; inputting the text data to the recognition model; and obtaining recognized target text data output by the recognition model and corresponding to the text data.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: July 20, 2021
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventor: Zhao Yang
  • Patent number: 11064006
    Abstract: A listening device that identifies, based on receiving the digitized voice stream, a first keyword of a plurality of keywords in the digitized voice stream. For example, the keyword may be a name associated with a service provider (e.g., “Google”). In response to identifying the first keyword of the plurality of keywords in the digitized voice stream the listening device identifies a first communication address of a first communication server of a first service provider associated with the first keyword of the plurality of keywords in the digitized voice stream. The listening server then routes the digitized voice stream and/or information associated with the digitized voice stream to the first communication server of the first service provider using the first communication address.
    Type: Grant
    Filed: November 13, 2019
    Date of Patent: July 13, 2021
    Assignee: Flex Ltd.
    Inventors: Mesut Gorkem Eraslan, Bruno Dias Leite
  • Patent number: 11064339
    Abstract: An emergency event detection and response system detects an occurrence of an event associated with a user and initiates an emergency response flow. A user may be associated with a wearable device and have in his home a base station and portable or stationary wireless devices containing sensors capable of detecting an emergency event. The emergency event may be detected based on voice or non-voice audio input from the user, data monitoring by the wearable device, base station, and/or portable or stationary wireless device, or by physical button press. Responsive to determining that an emergency event has occurred, the system triggers an emergency response flow by notifying a call center and contacting one or more caregivers associated with the user. Caregivers may access a response system application to receive updates regarding the detected emergency and to contact the user and/or a provider associated with the call center.
    Type: Grant
    Filed: April 12, 2020
    Date of Patent: July 13, 2021
    Assignee: Aloe Care Health, Inc.
    Inventors: Lasse Hamre, Raymond Eugene Spoljaric, Evan Samuel Schwartz, Ryan Christopher Haigh, Alexander Neville Sassoon, Sveinung Kval Bakken
  • Patent number: 11056113
    Abstract: A conversation guidance method of a speech recognition system may include managing a user domain based on speech recognition function information and situation information corrected from a system mounted on a vehicle, generating a conversation used for speech recognition based on the user domain, and guiding a user with the generated conversation.
    Type: Grant
    Filed: May 16, 2019
    Date of Patent: July 6, 2021
    Assignees: Hyundai Motor Company, Kia Corporation
    Inventors: Kyung Chul Lee, Jae Min Joh
  • Patent number: 11049495
    Abstract: There is provided a system and method for processing and/or recognizing acoustic signals. The method comprises obtaining at least one pre-existing speech recognition model; adapting and/or training the at least one pre-existing speech recognition model incrementally when new, previously unseen, user-specific data is received, the data comprising input acoustic signals and/or user action demonstrations and/or semantic information about a meaning of the acoustic signals, wherein the at least one model is incrementally updated by associating new input acoustic signals with input semantic frames to enable recognition of changed input acoustic signals. The method further comprises adapting to a user's vocabulary over time by learning new words and/or removing words no longer being used by the user, generating a semantic frame from an input acoustic signal according to the at least one model, and mapping the semantic frame to a predetermined action.
    Type: Grant
    Filed: March 17, 2017
    Date of Patent: June 29, 2021
    Assignee: Fluent.ai Inc.
    Inventors: Vikrant Tomar, Vincent P. G. Renkens, Hugo R. J. G. Van Hamme
  • Patent number: 11044364
    Abstract: A system for providing help includes a preprogrammed kit that includes at least one digital assistant and a virtual private network repeater for connecting to a data provider for connecting the digital assistant to a server. A plurality of agent computers is connected to the server by a data network. The digital assistant is preprogrammed with a skill for recognizing a preprogrammed specific utterance and the digital assistant is pre-configured to connect with the virtual private network repeater. After the preprogrammed specific utterance is detected by the digital assistant, the digital assistant initiates a request for help to the server and upon receiving the request for the help, the server assigns one of the agent computers and forwards the request for help to the one of the agent computers.
    Type: Grant
    Filed: April 24, 2020
    Date of Patent: June 22, 2021
    Assignee: Ways Investments, LLC
    Inventor: Mark Edward Gray
  • Patent number: 11031013
    Abstract: Method starts with processing, by a processor, audio signal to generate audio caller utterance and transcribed caller utterance. Processor generates identified task based on transcribed caller utterance. Processor samples audio caller utterance to generate samples of audio caller utterance. Processor generates loudness result based on loudness values of samples using loudness neural network associated with identified task. Processor generates pitch result based on pitch values of samples using pitch neural network associated with identified task. Processor generates tone result for each word in transcribed caller utterance using tone neural network associated with identified task. Using task completion probability neural network associated with identified task, processor generates task completion probability result that is based on at least one of: loudness result, pitch result, or tone result. Other embodiments are disclosed herein.
    Type: Grant
    Filed: June 17, 2019
    Date of Patent: June 8, 2021
    Assignee: Express Scripts Strategic Development, Inc.
    Inventors: Christopher M. Myers, Danielle L. Smith
  • Patent number: 11031000
    Abstract: An artificial intelligence (AI) system configured to simulate functions of a human brain, such as recognition, determination, etc., by using a machine learning algorithm, such as deep learning, etc., and an application thereof. The AI system includes a method performed by a device to transmit and receive audio data to and from another device includes obtaining a voice input that is input by a first user of the device, obtaining recognition information indicating a meaning of the obtained voice input, transmitting the obtained voice input to the other device, determining whether an abnormal situation occurs, in which a second user of the other device does not understand the transmitted voice input, and transmitting the obtained recognition information to the other device, based on a result of the determination.
    Type: Grant
    Filed: December 18, 2019
    Date of Patent: June 8, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Jae-deok Kim, Mee-jeong Park
  • Patent number: 11011170
    Abstract: The present invention provides a speech processing method, which includes: receiving speech information; recognizing the speech information to convert the speech information into a plurality of pieces of text information; displaying at least one pieces of the text information; and receiving a selection signal to respond according to selected text information. The present invention further provides a speech processing device. According to the speech processing method and device provided by embodiments of the present invention, a user can select and confirm by clicking on displayed text information after the speech information is recognized, so that the selected text information is guaranteed to be coincident with a received speech information, thus further increasing accuracy and convenience of the speech processing and providing the user with a good experience effect.
    Type: Grant
    Filed: September 20, 2017
    Date of Patent: May 18, 2021
    Assignees: ZHEJIANG GEELY HOLDING GROUP CO., LTD., ZHEJIANG GEELY AUTOMOBILE RESEARCH INSTITUTE CO., LTD.
    Inventor: Shufu Li
  • Patent number: 11003354
    Abstract: Method and system are provided for compensating for user hand tremors when using a hand-held electronic device having a user interface display. The method includes monitoring position data of a user's finger in relation to the user interface display over time as the finger approaches an element in the user interface display. A target element is determined by predicting an intended movement of the user's finger and the method dynamically enlarges the determined target element in the user interface display as the user's finger approaches the user interface display.
    Type: Grant
    Filed: July 24, 2019
    Date of Patent: May 11, 2021
    Assignee: International Business Machines Corporation
    Inventors: Samuel D. Goulden, William E. Sawyer, Andrew J. Seymour, Peter Weller
  • Patent number: 10991366
    Abstract: A method, performed by a dialogue processing device, of processing dialogue associated with a user based on dialog act information, the method comprises receiving speech information, corresponding to speech of the user, including a plurality of sentence units; identifying a first sentence unit and a second sentence unit, of the plurality of sentence units, based on receiving the speech information; extracting a first dialog act indicative of an intention of the first sentence unit and extracting a second dialog act indicative of an intention of the second sentence unit; extracting a first dialog act indicative of an intention of the first sentence unit and extracting a second dialog act indicative of an intention of the second sentence unit; processing the first sentence unit and the second unit in a sequence according to respective priority orders assigned based on a number of empty slots of dialogue frames of the sentence units.
    Type: Grant
    Filed: July 2, 2018
    Date of Patent: April 27, 2021
    Assignee: SAMSUNG SDS CO., LTD.
    Inventors: Han Hoon Kang, Eun Hye Ji, Na Rae Kim, Jae Young Yang
  • Patent number: 10976997
    Abstract: An electronic device includes a housing, a touch screen display exposed through a first area of the housing, a microphone exposed through a second area of the housing, a wireless communication circuit, a processor electrically connected to the touch screen display, the microphone, and the wireless communication circuit, and a memory electrically connected to the processor and storing instructions.
    Type: Grant
    Filed: July 24, 2018
    Date of Patent: April 13, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Kyu Young Kim, Tae Kwang Um, Jae Yung Yeo
  • Patent number: 10977687
    Abstract: Embodiments disclosed herein are related to computing systems, and methods for determining patterns in received data that are indicative of common characteristics of the one or more users of a computing system. Data from first users of the computing system is received. The received data defines information about the first users. The type of the data that is received is determined by the first users. The received data is analyzed to determine one or more patterns in the received data. The one or more patterns are indicative of one or more common characteristics shared by the first users. Information related to the determined one or more patterns is provided to second users. The information includes a DID for each of the f first users that may be used by the second users to communicate with the first users.
    Type: Grant
    Filed: October 8, 2018
    Date of Patent: April 13, 2021
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Ankur Patel, Daniel James Buchner
  • Patent number: 10978060
    Abstract: A method is provided in accordance with an aspect of the present disclosure. The method includes detecting at least one voice input from a user of an electronic device, transforming the at least one voice input into a text structure including at least one word, and determining a current context scope of the electronic device. The method also includes comparing the text structure to a plurality of existing text structures, where each of the existing text structure associated with a command for an action on the electronic device. The method further includes identifying, when the text structure matches with at least one of the existing text structures, a command to correspond to the at least one voice input from the user, and performing an action on the electronic device based on the identified command.
    Type: Grant
    Filed: January 31, 2014
    Date of Patent: April 13, 2021
    Assignee: Hewlett-Packard Development Company, L.P.
    Inventors: Syed S Azam, Yetian Huang
  • Patent number: 10978094
    Abstract: The present disclosure provides systems and methods for selecting and presenting content items based on user input. The method includes receiving first input intended to identify a desired content item among content items associated with metadata, determining that an input portion has an importance measure exceeding a threshold, and providing feedback identifying the input portion. The method further includes receiving second input, and inferring user intent to alter or supplement the first input with the second input. The method further includes, upon inferring intent to alter the first input, determining an alternative query by modifying the first input based on the second input, and, upon inferring intent to supplement the first input, determining an alternative query by combining the first input and the second input. The method further includes selecting and presenting a subset of content items based on comparing the alternative query and metadata associated with the subset.
    Type: Grant
    Filed: October 18, 2018
    Date of Patent: April 13, 2021
    Assignee: Veveo, Inc.
    Inventors: Murali Aravamudan, Girish Welling, Daren Gill, Sankar Ardhanari, Rakesh Barve, Sashikumar Venkataraman
  • Patent number: 10970490
    Abstract: A conversation for simulation by an artificial intelligence-based tutor is automatically generated. The conversation relates to content on which the artificial intelligence-based tutor is being trained. The automatically generating the conversation includes automatically generating inputs from a simulated student regarding the content, and automatically generating on behalf of the artificial intelligence-based tutor expected responses to the inputs generated from the simulated student. The conversation is provided as an output to be input to the artificial intelligence-based tutor for simulation by the artificial intelligence-based tutor.
    Type: Grant
    Filed: May 16, 2019
    Date of Patent: April 6, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Saubhik Kundu, Rajesh Barri, Vivek R. Pratap, Rishi Arora, Saugata Das, Tarannum M. Landge
  • Patent number: 10964308
    Abstract: A speech processing apparatus is provided in which, while face feature points are extracted from moving image data obtained by imaging a speaker's face, for each frame, a first generation network for generating face feature points of the corresponding frame based on speech feature data extracted from uttered speech of the speaker for each frame is generated, and whether the first generation network is appropriate is evaluated using an identification network, then, a second generation network for generating the uttered speech from a plurality of uncertain settings including at least text representing utterance content of the uttered speech and information indicating emotions included in the uttered speech, a plurality of types of fixed settings which define speech quality, and the face feature points generated by the first generation network evaluated as appropriate, is generated, and whether the second generation network is appropriate is evaluated using the identification network.
    Type: Grant
    Filed: October 29, 2018
    Date of Patent: March 30, 2021
    Inventor: Ken-ichi Kainuma
  • Patent number: 10957317
    Abstract: A computing system can determine a vehicle command based on a received spoken language command and determined confidence levels. The computing system can operate a vehicle based on the vehicle command. The computing system can further determine the spoken language command by processing audio spectrum data corresponding to spoken natural language with an automatic speech recognition (ASR) system.
    Type: Grant
    Filed: October 18, 2018
    Date of Patent: March 23, 2021
    Assignee: Ford Global Technologies, LLC
    Inventors: Lisa Scaria, Ryan Burke, Praveen Narayanan, Francois Charette
  • Patent number: 10943592
    Abstract: A computer speech output control method, system, and non-transitory computer readable medium, include a computer speech output unit configured to output a computer speech, a human speech monitoring circuit configured to determine whether ambient human conversation including human-to-human speech is occurring, and an interruption determining circuit configured to determine whether to cause the computer speech output unit to output the computer speech based on a status of the human conversation.
    Type: Grant
    Filed: October 31, 2017
    Date of Patent: March 9, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Christopher J. Hardee, Steven Robert Joroff, Pamela Ann Nesbitt, Scott Edward Schneider
  • Patent number: 10942702
    Abstract: This relates to systems and processes for using a virtual assistant to arbitrate among and/or control electronic devices. In one example process, a first electronic device samples an audio input using a microphone. The first electronic device broadcasts a first set of one or more values based on the sampled audio input. Furthermore, the first electronic device receives a second set of one or more values, which are based on the audio input, from a second electronic device. Based on the first set of one or more values and the second set of one or more values, the first electronic device determines whether to respond to the audio input or forego responding to the audio input.
    Type: Grant
    Filed: August 23, 2018
    Date of Patent: March 9, 2021
    Assignee: Apple Inc.
    Inventors: Kurt Piersol, Ryan M. Orr, Daniel J. Mandel