Abstract: A method of determining controversy voice share is provided. The method comprises identifying keywords related to a controversy topic and identifying information sources relevant to the controversy topic based on occurrences of the keywords within each source. An external voice score is calculated for an organization regarding the controversy topic, wherein the external voice score is calculated based on information sources that are external to the organization. An internal voice score is calculated for the organization regarding the controversy topic, wherein the internal voice score is calculated based on information sources internal to the organization. An overall voice share of the organization is then calculated from the external voice score and internal voice score. The overall voice share of the organization represents a proportion of influence the organization contributes to public commentary on the controversy topic relative to other organizations within a same industry sector.
Abstract: Systems and processes for providing personalized suggestions indicating that a task may be performed using a digital assistant of an electronic device are provided. An example method includes, at an electronic device with a display, receiving context data associated with the electronic device; determining, based on the context data, a task that may be performed by a digital assistant of the electronic device in response to a natural-language expression; determining, based on the context data, whether suggestion criteria associated with the determined task are satisfied; and in accordance with a determination that the suggestion criteria are satisfied, providing a suggestion indicating that the determined task may be performed using the digital assistant of the electronic device.
Type:
Grant
Filed:
October 2, 2019
Date of Patent:
July 12, 2022
Assignee:
Apple Inc.
Inventors:
Kirill Novitchenko, Kyle W. Lu, Nikrouz Ghotbi, Trungtin Tran, Lun Cui, Matias Hernan Sulzberger, Marek Rastislav Telgarsky, Briana C. Chapman, Kevin Siu Wai Li, Jared Crawford, Robert A. Walker
Abstract: The present disclosure relates to a field of intelligent communications, and discloses a speech communication system and method with human-machine coordination, which resolve a problem of bad client experience because great differences occur after a switchover in a call through a prior human-machine coordination and time of a client is wasted.
Abstract: In one example, a method includes method comprising: receiving audio data generated by a microphone of a current computing device; identifying, based on the audio data, one or more computing devices that each emitted a respective audio signal in response to speech reception being activated at the current computing device; and selecting either the current computing device or a particular computing device from the identified one or more computing devices to satisfy a spoken utterance determined based on the audio data.
Abstract: Information loss in speech to text conversion and Inability to preserve vocal emotion information without changing the artificial intelligence model infrastructure in a conventional speech to speech translation system are essential drawback of the conventional techniques. Embodiments of the invention provide direct speech to speech translation system is disclosed. Direct speech to speech translation system uses a one-tier approach, creating a unified-model for whole application. The single-model ecosystem takes in audio (mel spectrogram) as an input and gives out audio (mel spectrogram) as an output. This solves the bottleneck problem by not converting speech directly to text but having text as a byproduct of speech to speech translation, preserving phonetic information along the way. This model also uses pre-processing and post-processing scripts but only for the whole model. This model needs parallel audio samples in two languages.
Type:
Grant
Filed:
December 24, 2021
Date of Patent:
June 14, 2022
Inventors:
Sandeep Dhawan, Kapil Dhawan, Dennis Reutter, Chris Beckman, Ahsan Memon
Abstract: A signal activity detector (SAD) combines at least three decision signals to generate a combined decision signal as input to a hangover addition circuit of the SAD. Each of the decision signals indicating whether or not activity is detected in the input signal according to respective decision criteria. The SAD sends the combined decision signal to the hangover addition circuit to generate a final decision signal of the SAD as to whether or not activity is detected in the input signal.
Abstract: Provided is a home appliance that performs a washing or drying operation based on a start command input through artificial intelligence and voice recognition a method for controlling such a home appliance. According to the present disclosure, the home appliance and the method for controlling such a home appliance include a sensor that senses a state of a door installed in a washing unit and a controller that executes an operation command based on a change of a state of the door sensed by the sensor. At this time, when the door is changed from an open state to a close state after the power of the home appliance is turned on, the controller controls the operation of the washing unit based on the operation command extracted from the voice signal of the user. Thus, the present disclosure may enable performing an command to start an operation input based on voice recognition.
Type:
Grant
Filed:
July 24, 2019
Date of Patent:
May 17, 2022
Assignee:
LG ELECTRONICS INC.
Inventors:
Hyosang Shin, Sungjun Yun, Yongwoon Jang, Byung Woo Jun
Abstract: Provided is a system for recording a voice about a working state, etc., linking the recorded voice to position information, and converting the recorded voice into a text. In a voice detail recording system (1), a control unit (10) executes an acquisition module (11) to acquire a voice and a position where the voice is uttered. In addition, the control unit (10) executes a voice recognition module (12) to perform voice recognition on the acquired voice. In addition, the control unit (10) executes a specification module (13) to specify Web content linked to the position acquired by executing the acquisition module (11). In addition, the control unit (10) executes a recording module (14) to record details obtained by executing the voice recognition module (12) to perform the voice recognition in the Web content specified by executing the specification module (13).
Abstract: A voice information processing method includes: receiving, by a terminal, voice information, and converting the voice information into text information; obtaining a field probability that the text information belongs to each of preset M event fields; obtaining a prior probability that the text information belongs to each of N event fields, where N?M; obtaining confidence that the text information belongs to each of the N event fields; calculating, based on the field probability, the prior probability, and the confidence that the text information belongs to each of the N event fields, values of N probabilities that the text information separately belongs to the N event fields; and outputting a semantic understanding result of performing semantic understanding based on an event field corresponding to a largest probability value in the N probability values.
Abstract: Embodiments of the present disclosure relate to a method and apparatus for operating a smart terminal. The method for operating a smart terminal includes: parsing a currently inputted user voice command to a current text command; querying a historical text command record matching the current text command from a historical text command record set corresponding to a historical user voice command set; parsing the queried historical text command record to obtain a user's intention; sending a user request generated based on the user's intention to a resource database; and commanding an action component to take an action based on a response returned from the resource database and the user's intention.
Abstract: An electronic communication method and system are disclosed. Exemplary methods can be used by hearing or speech-impaired participants in a conference, e.g., collaboration, environment to facilitate communication between conference participants.
Abstract: A processor may receive a social media post related to a monitored system. The processor may automatically process at least one element of the social media post by applying at least one natural language processing algorithm. The processor may determine a meaning of the social media post based on an output of the at least one natural language processing algorithm. The processor may automatically route the social media post to at least one recipient based on the meaning.
Type:
Grant
Filed:
November 1, 2018
Date of Patent:
April 5, 2022
Assignee:
INTUIT INC.
Inventors:
Gregory Dean Yenney, Radha Damera, Chandana Davuluri, Michael Han, Vaishnavi Rajagopal, Vinayak Thakkar, Suman Malla
Abstract: Example embodiments are disclosed where a narrative generation platform is integrated with a conversational interface such as a Chatbot to support interactive and dynamic narrative generation in response to speech inputs from users. Such a system can be further integrated with a visualization platform to allow the speech input through the conversational interface to interactively and dynamically focus and control visualizations and/or narratives that accompany the visualizations.
Type:
Grant
Filed:
July 2, 2020
Date of Patent:
March 29, 2022
Assignee:
NARRATIVE SCIENCE INC.
Inventors:
Lawrence A. Birnbaum, Kristian J. Hammond
Abstract: The present disclosure provides various embodiments of methods for intelligent active speaker identification and information handling systems (IHSs) utilizing such methods. In general, the methods disclosed herein may be used to accurately identify an active speaker in a communication session with an application or an IHS, regardless of whether the active speaker is alone, in a group environment, or using someone else's system or login to participate in the communication session. The methods disclosed herein may use voice processing technology and one or more voice identification databases (VIDs) to identify the active speaker in a communication session. In some embodiments, the disclosed methods may display the identity of the active speaker to other users or participants in the same communication session. In other embodiments, the disclosed methods may dynamically switch between user profiles or accounts during the communication session based on the identity of the active speaker.
Abstract: A method, system and computer program product for editing a text using speech recognition includes receiving, by a computer, a first voice input from a user comprising a first target word. The computer identifies instances of the first target word within the text and assigns a first numerical indicator to each instance of the first target word within the text. A selection is received from the user including the first numerical indicator corresponding to a starting point of a selection area. The computer receives a second voice input from the user including a second target word, identifies instances of the second target word within the text, assigns a second numerical indicator to each instance of the second target word, and receives a selection from the user including the second numerical indicator corresponding to an ending point of the selection area.
Type:
Grant
Filed:
September 25, 2019
Date of Patent:
March 29, 2022
Assignee:
International Business Machines Corporation
Inventors:
JunXing Yang, XueJun Zhong, Wei Sun, ZhiXia Wang
Abstract: In one example, a method includes method comprising: receiving audio data generated by a microphone of a current computing device; identifying, based on the audio data, one or more computing devices that each emitted a respective audio signal in response to speech reception being activated at the current computing device; and selecting either the current computing device or a particular computing device from the identified one or more computing devices to satisfy a spoken utterance determined based on the audio data.
Abstract: Exemplary embodiments relate to using a voice user interface for creating a poll to send to users. A poll requestor can create a poll using voice instructions and specify response options and setup information. After creating the poll, the system broadcasts it to users who can respond to the poll. The poll may be broadcasted while a user is receiving content. The users can respond using a voice user interface. The system compiles the responses and provides them to the poll requestor.
Type:
Grant
Filed:
June 26, 2019
Date of Patent:
February 15, 2022
Assignee:
Amazon Technologies, Inc.
Inventors:
Eric Sungmin Kim, Greg Roberts, Daniel Zi Wang, Dylan Robert House, Luis Valencia, Paul Damian Tidwell, John Benjamin Dalin, Steven T. Rabuchin, James Robert Blair, Vicky Ge
Abstract: A method, computer system, and computer program product for form-based conversation system design are provided. The embodiment may include ingesting, by a processor, a plurality of forms from a given domain. The embodiment may also include extracting indicators of required input fields from the ingested plurality of forms. The embodiment may further include generating a required input list based on the extracted indicators of the required input fields to update a size of the required input list. The embodiment may also include determining transactional intents based on the required input list. The embodiments may further include generating a dialog flow that satisfies the determined transactional intents.
Type:
Grant
Filed:
April 16, 2019
Date of Patent:
February 15, 2022
Assignee:
INTERNATIONAL BUSINESS MACHINES CORPORATION
Inventors:
Andrew R. Freed, Corville O. Allen, Joseph Kozhaya, Shikhar Kwatra
Abstract: A method for form-based conversation system design is provided. The embodiment may include ingesting, by a processor, a plurality of forms from a given domain. The embodiment may also include extracting indicators of required input fields from the ingested plurality of forms. The embodiment may further include generating a required input list based on the extracted indicators of the required input fields to update a size of the required input list. The embodiment may also include determining transactional intents based on the required input list. The embodiments may further include generating a dialog flow that satisfies the determined transactional intents.
Type:
Grant
Filed:
July 12, 2019
Date of Patent:
February 15, 2022
Assignee:
INTERNATIONAL BUSINESS MACHINES CORPORATION
Inventors:
Andrew R. Freed, Corville O. Allen, Joseph Kozhaya, Shikhar Kwatra