Patents Examined by Jonathan C Kim
  • Patent number: 11600276
    Abstract: One embodiment provides a method for predicting a next action in a conversation system that includes obtaining, by a processor, information from conversation logs and a conversation design. The processor further creates a dialog graph based on the conversation design. Weights and attributes for edges in the dialog graph are determined based on the information from the conversation logs and adding user input and external context information to an edge attributes set. An unrecognized user input is analyzed and a next action is predicted based on dialog nodes in the dialog graph and historical paths. A guiding conversation response is generated based on the predicted next action.
    Type: Grant
    Filed: January 11, 2021
    Date of Patent: March 7, 2023
    Assignee: International Business Machines Corporation
    Inventors: Lei Huang, Robert J. Moore, Guangjie Ren, Shun Jiang
  • Patent number: 11574135
    Abstract: The present disclosure provides a method, apparatus, electronic device and readable storage medium for translation and relates to translation technologies. In the embodiments of the present disclosure, the at least one knowledge element is obtained according to associated information of content to be translated, and respective knowledge element in the at least one knowledge element comprise an element of the first language type and an element of the second language type so that the at least one knowledge element can be used to obtain a translation result of the content to be translated. Since the at least one knowledge element obtained in advance is taken as global information of the translation task of this time, it can be ensured that the translation result of the same content to be translated is consistent, thereby improving the quality of the translation result.
    Type: Grant
    Filed: April 29, 2020
    Date of Patent: February 7, 2023
    Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.
    Inventors: Haifeng Wang, Hua Wu, Zhongjun He, Hao Xiong
  • Patent number: 11568853
    Abstract: Disclosed is a voice recognition method and apparatus using artificial intelligence. A voice recognition method using artificial intelligence may include: generating a utterance by receiving a voice command of a user; obtaining a user's intention by analyzing the generated utterance; deriving an urgency level of the user on the basis of the generated utterance and prestored user information; generating a first response in association with the user's intention; obtaining main vocabularies included in the first response; generating a second response by using the main vocabularies and the urgency level of the user; determining a speech rate of the second response on the basis of the urgency level of the user; and outputting the second response according to the speech rate by synthesizing the second response to a voice signal.
    Type: Grant
    Filed: July 29, 2020
    Date of Patent: January 31, 2023
    Assignee: LG ELECTRONICS INC.
    Inventor: Jonghoon Chae
  • Patent number: 11557286
    Abstract: A speech recognition method includes receiving speech data, obtaining, from the received speech data, a candidate text including at least one word and a phonetic symbol sequence associated with a pronunciation of a target word included in the received speech data, using a speech recognition model, replacing the phonetic symbol sequence included in the candidate text with a replacement word corresponding to the phonetic symbol sequence, and determining a target text corresponding to the received speech data based on a result of the replacing.
    Type: Grant
    Filed: December 30, 2019
    Date of Patent: January 17, 2023
    Assignee: Samsung Electronics Co., Ltd.
    Inventor: Jihyun Lee
  • Patent number: 11531813
    Abstract: A method, an electronic device and a readable storage medium for creating a label marking model are disclosed. The method for creating the label marking model includes: obtaining text data and determining a word or phrase to be marked in the text data; according to the word or phrase to be marked, constructing a first training sample of the text data corresponding to a word or phrase replacing task and a second training sample corresponding to a label marking task; training a neural network model with a plurality of the first training samples and a plurality of the second training samples, respectively, until a loss function of the word or phrase replacing task and a loss function of the label marking task satisfy a preset condition, to obtain the label marking model.
    Type: Grant
    Filed: September 9, 2020
    Date of Patent: December 20, 2022
    Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.
    Inventors: Xinchao Xu, Haifeng Wang, Hua Wu, Zhanyi Liu
  • Patent number: 11532315
    Abstract: An autocorrelation calculation unit 21 calculates an autocorrelation RO(i) from an input signal. A prediction coefficient calculation unit 23 performs linear prediction analysis by using a modified autocorrelation R?O(i) obtained by multiplying a coefficient wO(i) by the autocorrelation RO(i). It is assumed here, for each order i of some orders i at least, that the coefficient wO(i) corresponding to the order i is in a monotonically increasing relationship with an increase in a value that is negatively correlated with a fundamental frequency of the input signal of the current frame or a past frame.
    Type: Grant
    Filed: December 14, 2020
    Date of Patent: December 20, 2022
    Assignee: NIPPON TELEGRAPH AND TELEPHONE CORPORATION
    Inventors: Yutaka Kamamoto, Takehiro Moriya, Noboru Harada
  • Patent number: 11514908
    Abstract: A voice command recognition device and a method thereof are provided. The voice command recognition device includes a processor that registers one or more voice commands selected by analysis of one or more voice commands repeatedly used by a user or a voice command utterance pattern of the user to generate one package command and a storage storing data or an algorithm for speech recognition by the processor.
    Type: Grant
    Filed: May 6, 2020
    Date of Patent: November 29, 2022
    Assignees: HYUNDAI MOTOR COMPANY, KIA MOTORS CORPORATION
    Inventors: Seo Hwan Choi, Jae Min Joh
  • Patent number: 11516039
    Abstract: An embodiment of the present invention comprises: a communication module for communicating with at least one external device; a microphone for receiving a user utterance; a memory for storing performance mode information having been configured in the electronic device; and a processor electrically connected to the communication module, the microphone, and the memory, wherein the processor is configured to: receive, through the microphone, a second user utterance associated with task execution; transmit first data associated with the second user utterance to an external device; receive, from the external device, second data associated with at least a part of processing of the first data; identify a first work load allocated to the electronic device at the time of receiving the second data; and compare a second work load required for processing the second data and the first work load, so as to control the performance mode. In addition, various embodiments recognized through the specification are possible.
    Type: Grant
    Filed: March 6, 2019
    Date of Patent: November 29, 2022
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Kyounggu Woo, Inchoon Yeo
  • Patent number: 11514905
    Abstract: Provided is an information processing apparatus that has an utterance function or controls the utterance function. The information processing apparatus includes a sending unit that sends interactive information regarding a voice agent, a receiving unit that receives interactive information regarding another voice, and a control unit that controls an utterance timing of the voice agent on the basis of the interactive information regarding another voice agent received by the receiving unit. The control unit causes utterance by the voice agent to stand by on the basis of the interactive information received from another voice agent. Moreover, the control unit causes the interactive information to be continuously sent during the utterance by the voice agent and during interaction between the voice agent and a user.
    Type: Grant
    Filed: August 28, 2018
    Date of Patent: November 29, 2022
    Assignee: SONY CORPORATION
    Inventors: Noriko Totsuka, Kazuya Tateishi, Yuichiro Koyama
  • Patent number: 11501754
    Abstract: Desired vehicle destinations may be determined from spoken dialogs. A speech input may be received from a user through a voice user interface. Current utterance variables may be obtained by tokenizing the user speech input. One or more of a plurality of utterance templates for a reply to the user speech input may be determined by a trained automatic agent based on the plurality of current utterance variables. One of a plurality of Application Programming Interfaces (API) to call and one or more parameters for the API to call with may be determine by the trained automatic agent based on the plurality of current utterance variables. A response may be obtained from the API call. A context string for the reply to the user speech input by the trained automatic agent may be constructed based on the utterance templates and the response of the API call.
    Type: Grant
    Filed: July 7, 2020
    Date of Patent: November 15, 2022
    Assignee: Beijing DiDi Infinity Technology and Development Co., Ltd.
    Inventors: Kevin Knight, Arkady Arkhangorodsky, Amittai Axelrod, Christopher Chu, Scot Fang, Yiqi Huang, Ajay Nagesh, Xing Shi, Boliang Zhang
  • Patent number: 11501773
    Abstract: In one aspect, a playback device includes a voice assistant service (VAS) wake-word engine and a command keyword engine. The playback device detects, via the command keyword engine, a first command keyword, and determines whether one or more playback conditions corresponding to the first command keyword are satisfied. Based on (a) detecting the first command keyword and (b) determining that the one or more playback conditions corresponding to the first command keyword are satisfied, the playback device playback device performs a first playback command corresponding to the first command keyword. When the playback device detects, via the wake-word engine, a wake-word in voice input, the playback device streams sound data corresponding to at least a portion of the voice input to one or more remote servers associated with the VAS.
    Type: Grant
    Filed: March 9, 2020
    Date of Patent: November 15, 2022
    Assignee: Sonos, Inc.
    Inventors: Connor Smith, John Tolomei, Kurt Soto
  • Patent number: 11495223
    Abstract: An electronic device according to various embodiments may comprise a memory in which one or more applications are installed, a communication circuit, and a processor, wherein the processor is configured to acquire audio data during execution of a designated application among the one or more applications, wherein the acquiring of audio data comprises an operation of storing, in the memory, at least a portion including multiple pieces of phoneme information among the audio data, when a designated condition is satisfied, transmit the at least portion to an external electronic device so that the external electronic device generates designated information for execution of at least one application among the one or more applications by using at least a part of the multiple pieces of phoneme information stored before the designated condition is satisfied, and on the basis of the designated information, execute the at least one application in relation to the designated application.
    Type: Grant
    Filed: November 1, 2018
    Date of Patent: November 8, 2022
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Hyun Gi Ahn, Joo Yoo Kim, Ji Eun Kim, Dong Hyun Roh, Kyung Sub Min, Seung Eun Lee
  • Patent number: 11488600
    Abstract: The invention relates to a method for an efficient dialogue configuration and interpretation in a computer-supported automated dialogue system. The invention offers, inter alia, the advantage that the work complexity when generating dialogues, and thus also the error propensity, is reduced. Furthermore, it is possible to take the context of sub-dialogs into consideration and dynamically generate a dialogue depending on said context. Furthermore, the method can automatically initiate a dialogue with a human user. The invention further relates to a correspondingly designed system assembly and to a computer program product with control commands which carry out the method or operate the system assembly.
    Type: Grant
    Filed: May 26, 2019
    Date of Patent: November 1, 2022
    Assignee: GK Easydialog
    Inventor: Christoph Neumann
  • Patent number: 11468894
    Abstract: The present teaching relates to method, system, medium, and implementations for enabling communication with a user. Information representing surrounding of a user engaged in an on-going dialogue is received via the communication platform, wherein the information includes a current response from the user in the on-going dialogue and is acquired from a current scene in which the user is present and captures characteristics of the user and the current scene. Relevant features are extracted from the information. A state of the user is estimated based on the relevant features and a dialogue context surrounding the current scene is determined based on the relevant features. A feedback directed to the current response of the user is generated based on the state of the user and the dialogue context.
    Type: Grant
    Filed: December 27, 2018
    Date of Patent: October 11, 2022
    Assignee: DMAI, INC.
    Inventor: Nishant Shukla
  • Patent number: 11450315
    Abstract: Various embodiments of the disclosure disclose a method and apparatus for processing a voice recognition service in an electronic device. According to various embodiments of the disclosure, an electronic device may include a microphone, a memory, and a processor operatively coupled to the microphone and the memory. The processor may be configured to wake-up on the basis of detection of a wake-up word, process a first task corresponding to a first voice command of a user on the basis of the wake-up, set a wait time during which a follow-up command can be received on the basis of the processing of the first task, detect a second voice command of the user during the wait time, analyze a conversational context on the basis of the first voice command and second voice command, and process a second task on the basis of a result of the analysis. Various embodiments are possible.
    Type: Grant
    Filed: December 8, 2017
    Date of Patent: September 20, 2022
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Yong Ho Kim, Sourabh Pateriya, Sunah Kim, Gahyun Joo, Sang-Woong Hwang, Say Jang
  • Patent number: 11443115
    Abstract: One embodiment provides a method that includes receiving adjusted labeled data based on emotional tone factors. Words are analyzed using a tone latent Dirichlet allocation (T-LDA) model that models tone intensity using the emotional tone factors and integrating the adjusted labeled data. Representative words are provided for each emotional tone factor based on using the T-LDA model. The representative words are obtained using the T-LDA model based on determining posterior probabilities and adjusting the posterior probabilities based on an auxiliary topic.
    Type: Grant
    Filed: February 19, 2020
    Date of Patent: September 13, 2022
    Assignee: International Business Machines Corporation
    Inventors: Peifeng Yin, Zhe Liu, Anbang Xu, Taiga Nakamura
  • Patent number: 11437030
    Abstract: Selectively performing voice recognition using one device among multiple devices that recognize and execute the voice recognition based on at least one of apparatus information of the multiple devices and a function parsed from a result of the voice recognition. Thereby, only a single preferable device in an environment in which multiple devices exist, which can service the user input via voice recognition, actually responds to the voice input and services the voice input of the user.
    Type: Grant
    Filed: October 17, 2018
    Date of Patent: September 6, 2022
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventor: Chan-hee Choi
  • Patent number: 11429345
    Abstract: Techniques for remotely executing a secondary-device driver for generating commands for a secondary device are described herein. For instance, a secondary device (or “appliance”) may reside within an environment, along with a device to which the secondary device communicatively couples. The device may be configured to send control signals to the secondary device for causing the secondary device to perform certain operations. For instance, a user in the environment may provide, to the device, a request that the secondary device perform a certain operation. The device, which may lack some or all of a device driver associated with the secondary device, may then work with a remote service that executes the device driver for the purpose of receiving a command from the device driver and sending the command along to the secondary device. Upon receiving the command, the secondary device may perform the operation.
    Type: Grant
    Filed: October 18, 2019
    Date of Patent: August 30, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Rohan Mutagi, Michael Dale Whiteley, He Lu, Brian James Butler, Fred Torok, Willy Lew Yuk Vong, David Ross Bronaugh, Christopher Ryan Nies, Shikher Sitoke
  • Patent number: 11422772
    Abstract: Techniques for causing different devices to perform different operations using a single voice command are described herein. In some instances, a user may define a “scene”, in which a user sets different devices to different states and then associates an utterance with those states or with the operations performed by the devices to reach those states. For instance, a user may dim a light, turn on his television, and turn on his set-top box before sending a request to a local device or to a remote service to associate those settings with a predefined utterance, such as “my movie scene”. Thereafter, the user may cause the light to dim, the television to turn on, and the set-top box to turn on simply by issuing the voice command “execute my movie scene”.
    Type: Grant
    Filed: May 28, 2019
    Date of Patent: August 23, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Rohan Mutagi, He Lu, Fred Torok, Willy Lew Yuk Vong, David Ross Bronaugh, Bo Li
  • Patent number: 11417317
    Abstract: Aspects described herein may relate to the determination of data that is indicative of a greater range of speech properties than input text data. The determined data may be used as input to one or more speech processing tasks, such as model training, model validation, model testing, or classification. For example, after a model is trained based on the determined data, the model's performance may exhibit more resilience to a wider range of speech properties. The determined data may include one or more modified versions of the input text data. The one or more modified versions may be associated with the one or more speakers or accents and/or may be associated with one or more levels of semantic similarity in relation to the input text data. The one or more modified versions may be determined based on one or more machine learning algorithms.
    Type: Grant
    Filed: February 20, 2020
    Date of Patent: August 16, 2022
    Assignee: Capital One Services, LLC
    Inventors: Christopher Larson, Tarek Aziz Lahlou, Diana Mingels, Zachary Kulis, Erik T. Mueller