Speech Recognition Depending On Application Context, E.g., In A Computer, Etc. (epo) Patents (Class 704/E15.044)
-
Patent number: 12154567Abstract: The present disclosure is generally related to a data processing system to validate vehicular functions in a voice activated computer network environment. The data processing system can improve the efficiency of the network by discarding action data structures and requests that invalid prior to their transmission across the network. The system can invalidate requests by comparing attributes of a vehicular state to attributes of a request state.Type: GrantFiled: March 28, 2023Date of Patent: November 26, 2024Assignee: GOOGLE LLCInventors: Haris Ramic, Vikram Aggarwal, Moises Morgenstern Gali, David Roy Schairer, Yao Chen
-
Patent number: 12126813Abstract: A method for channel aware semantic coding (CASC) by a user equipment (UE), comprising: determining a quality level for a channel for a time period in which a video frame is being transmitted over the channel; determine, based on the quality level, one or more semantic elements to include in a semantic transcript stream (STS); encode the video frame with the one or more elements of the STS; and transmit the STS to a remote device.Type: GrantFiled: September 21, 2022Date of Patent: October 22, 2024Assignee: Apple Inc.Inventors: Sabine Roessel, Christian Drewes, Matthias Sauer
-
Patent number: 12118321Abstract: Methods and systems for multimodal collaborative conversational dialogue are disclosed. The multimodal collaborative conversational dialogue system include a multimodal avatar interface and one or more sensors, which obtains one or more multimodal inputs. A multimodal semantic parser generates one or more logical form representations based on the one or more multimodal inputs. A collaborative dialogue manager infers a goal of the user from the one or more logical form representations, and develops a plan including communicative actions and non-communicative actions with regard to the goal. The multimodal avatar interface outputs one or more multimodal collaborative plan-based dialogue system-generated communications with respect to execution of at least one communicative action. The collaborative dialogue manager maintains a collaborative dialogue with the user until obtainment of the goal.Type: GrantFiled: December 27, 2023Date of Patent: October 15, 2024Assignee: Openstream Inc.Inventors: Philip R. Cohen, Lucian Galescu, Rajasekhar Tumuluri
-
Patent number: 12093607Abstract: An audio control system can control interactions with an application or device using keywords spoken by a user of the device. The audio control system can use machine learning models (e.g., a neural network model) trained to recognize one or more keywords. Which machine learning model is activated can depend on the active location in the application or device. Responsive to detecting keywords, different actions are performed by the device, such as navigation to a pre-specified area of the application.Type: GrantFiled: July 29, 2022Date of Patent: September 17, 2024Assignee: Snap Inc.Inventors: Xin Chen, Yurii Monastyrshyn, Fedir Poliakov, Shubham Vij
-
Patent number: 12087304Abstract: An electronic device for providing content including an image and a voice is disclosed. The electronic device comprises: a display configured to display an image; a memory in which a voice recognition module including various executable instructions is stored; and a processor configured to acquire expected words that will possibly be included in a voice, based on information about content, using the expected words to perform voice recognition for the voice through the voice recognition module, and displaying, on the display, text converted from the voice based on the voice recognition.Type: GrantFiled: October 8, 2021Date of Patent: September 10, 2024Assignee: Samsung Electronics Co., Ltd.Inventor: Youngdae Kim
-
Patent number: 12080298Abstract: Systems and methods for processing speech transcription in a speech processing system are disclosed. Transcriptions of utterances is received and identifications to the transcriptions are assigned. In response to receiving an indication of an erroneous transcribed utterance in at least one of the transcriptions, an audio receiver is automatically activated for receiving a second utterance. In response to receiving the second utterance, an audio file of the second utterance and a corresponding identification of the erroneous transcribed utterance are transmitted to a speech recognition system for a second transcription, and the erroneous transcribed utterance is replaced with the second transcription.Type: GrantFiled: September 6, 2023Date of Patent: September 3, 2024Assignee: Rovi Guides, Inc.Inventors: Sukanya Agarwal, Vikram Makam Gupta
-
Patent number: 12067980Abstract: Disclosed is an electronic device. An electronic device according to an embodiment comprises: a display; a microphone; a wireless communication circuit; a processor operatively connected to the display, the microphone, and the communication circuit; and a memory operatively connected to the processor, wherein the memory may store a plurality of user-customized tasks mapped to one of a selected word, phrase, or sentence, and may store instructions that, when executed, cause the processor to receive a first voice input according to user's utterance through the microphone, obtain context information related to the electronic device when one of the word, phrase, or sentence is included in the first voice input, select one task from among the plurality of tasks on the basis of at least in part on the context information, and perform the selected one task by using at least a part of the electronic device. Other embodiments understood through the present specification are also possible.Type: GrantFiled: May 23, 2019Date of Patent: August 20, 2024Assignee: Samsung Electronics Co., Ltd.Inventor: Jeongpyo Lee
-
Patent number: 12046242Abstract: Audio message processing methods and apparatuses are provided, where a method may include a server recognizing types of communication messages transmitted between communicating counterparties; when a type of any communication message is an audio type, the server acquiring the any communication message, and converting the any communication message to corresponding text content; and upon determining that any communicating party has a conversion need for the any communication message, the server sending the text content to the any communicating party. Through technical solutions of the present disclosure, text conversion may be performed upon audio messages in advance, thereby increasing response speed for audio conversion requests of users.Type: GrantFiled: May 11, 2021Date of Patent: July 23, 2024Assignee: Alibaba Group Holding LimitedInventors: Daping Zhang, Lili Zhang, Yixin Huang, Yun Chen, Jiandong Lai, Haohua Zhong
-
Patent number: 12026482Abstract: How to correctly understand natural language query and catch the useful information from it is a long-standing research question in the chatbot field. However, in many cases only a small set of words are sufficient for closed-domain tasks. In this paper, we propose Minimal Dictionary Language (MDL): a simple language architecture which allows users to express their needs through certain basic syntax. For the pilot study, we selected Enron email open dataset[2] for our experimental use and apply MDL as the language for email retrieval task. We also built a front-end search interface by Python Flask and Vue.js for test and evaluation purposes.Type: GrantFiled: December 18, 2020Date of Patent: July 2, 2024Assignee: THE UNIVERSITY OF NORTH CAROLINA AT CHAPEL HILLInventors: Javed Mostafa, Xiaopeng Lu, Sandeep Avula
-
Patent number: 12003804Abstract: To provide an information processing device, an information processing method, and a computer program that assist a device operation of a user. An information processing device includes an input unit that receives input from a user, and a presentation unit that presents information regarding a command for operating a predetermined device on the basis of the input. The input unit receives input by at least one of voice or gesture from a user. The presentation unit presents a candidate of a command including at least one of voice or gesture and partially matching the input.Type: GrantFiled: August 16, 2019Date of Patent: June 4, 2024Assignee: SONY CORPORATIONInventors: Andrew Shin, Naoki Ide
-
Patent number: 11996103Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for voice recognition. In one aspect, a method includes the actions of receiving a voice input; determining a transcription for the voice input, wherein determining the transcription for the voice input includes, for a plurality of segments of the voice input: obtaining a first candidate transcription for a first segment of the voice input; determining one or more contexts associated with the first candidate transcription; adjusting a respective weight for each of the one or more contexts; and determining a second candidate transcription for a second segment of the voice input based in part on the adjusted weights; and providing the transcription of the plurality of segments of the voice input for output.Type: GrantFiled: July 11, 2022Date of Patent: May 28, 2024Assignee: Google LLCInventors: Petar Aleksic, Pedro J. Moreno Mengibar
-
Patent number: 11996097Abstract: A system and method performs multilingual wakeword detection by determining a language corresponding to the wakeword. A first wakeword-detection component, which may execute using a digital-signal processor, determines that audio data includes a representation of the wakeword and determines a language corresponding to the wakeword. A second, more accurate wakeword-detection component may then process the audio data using the language to confirm that it includes the representation of the wakeword. The audio data may then be sent to a remote system for further processing.Type: GrantFiled: June 28, 2021Date of Patent: May 28, 2024Assignee: Amazon Technologies, Inc.Inventors: Yixin Gao, Ming Sun, Jason Krone, Shiv Naga Prasad Vitaladevuni, Yuzong Liu
-
Patent number: 11978449Abstract: An intelligent server is provided.Type: GrantFiled: November 4, 2021Date of Patent: May 7, 2024Assignee: Samsung Electronics Co., Ltd.Inventors: Jinho Choi, Gajin Song
-
Patent number: 11972759Abstract: Mitigating mistranscriptions resolves errors in a transcription of the audio portion of a video based on a semantic matching with contextualized data electronically garnered from one or more sources other than the audio portion of the video. A mistranscription is identified using a pretrained word embedding model that maps words to an embedding space derived from the contextualizing data. A similarity value for each vocabulary word of a multi-word vocabulary of the pretrained word embedding model is determined in relation to the mistranscription. Candidate words are selected based on the similarity values, each indicating a closeness of a corresponding vocabulary word to the mistranscription. The textual rendering is modified by replacing the mistranscription with a candidate word that, based on average semantic similarity values, is more similar to the mistranscription than is each other candidate word.Type: GrantFiled: December 2, 2020Date of Patent: April 30, 2024Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Shikhar Kwatra, Vijay Ekambaram, Hemant Kumar Sivaswamy, Rodrigo Goulart Silva
-
Patent number: 11973806Abstract: Techniques are disclosed relating to automatically altering a displayed user interface for an event. A server computer system may cause, via a conferencing service, display of a user interface for an event having a group of users accessing the conferencing service via a plurality of user devices, the displayed interface including an indication of a video feed of a user in the group of users that is currently active. The system may store, in a database, data for the event, including content of audio and video feeds of users in the event. The system may analyze a set of characteristics included in the content of the audio and video feeds. The system may alter, while the indication of the video feed of the user is being displayed, aspects of the displayed user interface other than the indication, where the altering is performed based on the analyzing.Type: GrantFiled: January 19, 2023Date of Patent: April 30, 2024Assignee: Toucan Events Inc.Inventors: Ivo Walter Rothschild, Paul Robert Murphy, Asahi Sato, Antonia Theodora Hellman, Ethan Duncan He-Li Hellman, Steven Emmanuel Hellman
-
Patent number: 11955124Abstract: An example electronic device includes a housing; a touchscreen display; a microphone; at least one speaker; a button disposed on a portion of the housing or set to be displayed on the touchscreen display; a wireless communication circuit; a processor; and a memory. When a user interface is not displayed on the touchscreen display, the electronic device enables a user to receive a user input through the button, receives user speech through the microphone, and then provides data on the user speech to an external server. An instruction for performing a task is received from the server. When the user interface is displayed on the touchscreen display, the electronic device enables the user to receive the user input through the button, receives user speech through the microphone, and then provides data on the user speech to the external server.Type: GrantFiled: January 10, 2022Date of Patent: April 9, 2024Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventors: Sang-Ki Kang, Jang-Seok Seo, Kook-Tae Choi, Hyun-Woo Kang, Jin-Yeol Kim, Chae-Hwan Li, Kyung-Tae Kim, Dong-Ho Jang, Min-Kyung Hwang
-
Patent number: 11932256Abstract: The disclosure generally pertains to systems and methods for identifying a location of an occupant in a vehicle. In an example method, a processor deconvolves a vocal utterance by an occupant of a vehicle and also determines an angle of arrival of the vocal utterance. The location of the occupant in the vehicle is then identified by the processor based on evaluating the deconvolved vocal utterance and the angle of arrival of the vocal utterance. Deconvolving the vocal utterance can involve applying a cabin impulse response to the vocal utterance for eliminating undesirable effects that may be imposed upon the vocal utterance by acoustic characteristics of the cabin of the vehicle (echo, sound reflections, sound damping, reverberation etc.). In some applications, the processor may refer to a lookup table to estimate a location of the occupant in the vehicle.Type: GrantFiled: November 18, 2021Date of Patent: March 19, 2024Assignee: Ford Global Technologies, LLCInventors: Ranjani Rangarajan, Leah Busch, Karthik Krishnamurthy, Nikhitha Bekkanti
-
Patent number: 11935529Abstract: Techniques for virtual assistant execution of ambiguous commands is provided. A voice instruction from a user may be received at a virtual assistant. The voice instruction may request the virtual assistant to perform a command. The command that is most likely being requested by the voice instruction from the user is identified. An ordered set of actions to execute when performing the command may be retrieved. Each action of the ordered set of actions may indicate if the action is reversible. Each action of the ordered set of actions may be executed in order until a not reversible action is reached or no further actions are in the ordered set of actions.Type: GrantFiled: June 15, 2021Date of Patent: March 19, 2024Assignee: MOTOROLA SOLUTIONS, INC.Inventors: Ying Bin Tan, Chew How Lim, Yih Farn Ghee, Joe Yie Chong
-
Patent number: 11922127Abstract: According to an embodiment, an electronic device comprises: a memory, a communication module comprising communication circuitry, and a processor operatively connected with the memory and the communication module. The processor is configured to control the electronic device to: obtain a utterance text corresponding to utterance speech, obtain an intent of the utterance text and emotion information based on the utterance speech and the utterance text, obtain a response text for the utterance text based on the intent of the utterance text and the emotion information, obtain a markup language including information about an output unit of text of the response text based on at least one of the intent of the utterance text, the emotion information, or the response text, and add the markup language to the response text and provide the response text. The text output unit is at least one selected from among a phoneme unit, a consonant and vowel unit, a syllable unit, or a word unit.Type: GrantFiled: May 21, 2021Date of Patent: March 5, 2024Assignee: Samsung Electronics Co., Ltd.Inventors: Kichul Kim, Yoonjae Park, Jooyong Byeon, Youngkyu Kim, Byungkeon Park, Soowon Jang, Changyong Jeong, Sungbin Jin, Jaeyung Yeo
-
Patent number: 11875792Abstract: A computer implemented method, computer system, and computer program product for executing a voice command. A number of processor units displays a view of a location with voice command devices in response to detecting the voice command from a user. The number of processor units displays a voice command direction for the voice command in the view of the location. The number of processor units changes the voice command direction in response to a user input. The number of processor units identifies a voice command device from the voice command devices in the location based on the voice command direction to form a selected voice command device. The number of processor units executes the voice command using the selected voice command device.Type: GrantFiled: August 17, 2021Date of Patent: January 16, 2024Assignee: International Business Machines CorporationInventors: Clement Decrop, Jeremy R. Fox, Tushar Agrawal, Sarbajit K. Rakshit
-
Patent number: 11869499Abstract: An information processing apparatus includes an extracting unit (133) that extracts a changing message related to a change in macro data (M), the changing message including at least one piece of first information indicating a function to be executed, and second information linked to the first information, from a user speech; a presuming unit (134) that presumes an element to be changed in the macro data (M) based on the changing message extracted by the extracting unit (133); and a changing unit (135) that changes the element to be changed in the macro data (M) presumed by the presuming unit (134), based on the changing message.Type: GrantFiled: July 1, 2019Date of Patent: January 9, 2024Assignee: Sony CorporationInventors: Yuhei Taki, Hiro Iwase, Kunihito Sawai, Masaki Takase, Akira Miyashita
-
Patent number: 11854422Abstract: A method and a device for information interaction. The method comprises: in response to receiving an oral practice request initiated by a user, outputting task information for indicating a target oral practice task (201), wherein the task information corresponds to task intention information and task keyword information; acquiring voice information inputted by the user with regard to the task information (202); recognizing the voice information, so as to determine user intention information and user keyword information corresponding to the user (203); generating a matching result for indicating whether the user has completed the target oral practice task (204), wherein the matching result is obtained by the following step: respectively matching the user intention information with the task intention information, and the user keyword information with the task keyword information, so as to obtain the matching result; and presenting the matching result to the user (205).Type: GrantFiled: August 15, 2022Date of Patent: December 26, 2023Assignee: DOUYIN VISION CO., LTD.Inventors: Haoran Huang, Xi Luo, Fuxiang Li, Hang Li
-
Patent number: 11854573Abstract: Techniques for performing conversation recovery of a system/user exchange are described. In response to determining that an action responsive to a user input cannot be performed, a system may determine a topic to recommend to a user. The topic may be unrelated to the original substance of the user input. The system may have access to various data representing a context in which a user provides an input to the system. The system may use these inputs and various data at runtime to make a determination regarding whether a user should be recommended a topic, as well as what that topic should be. The system may cause a question be output to the user, with the question asking the user about the topic, for example whether the user would like a song played, whether the user would like to hear information about a particular individual (e.g., artist), whether the user would like to know about a particular skill (e.g.Type: GrantFiled: September 10, 2020Date of Patent: December 26, 2023Assignee: Amazon Technologies, Inc.Inventors: Gregory Newell, Eliav Kahan, Ravi Chandra Reddy Yasa, David Suarez, Joel Toledano
-
Patent number: 11847124Abstract: Techniques for contextual search on multimedia content are provided. An example method includes extracting entities associated with multimedia content, wherein the entities include values characterizing one or more objects represented in the multimedia content, generating one or more query rewrite candidates based on the extracted entities and one or more terms in a query related to the multimedia content, providing the one or more query rewrite candidates to a search engine, scoring the one or more query rewrite candidates, ranking the scored one or more query rewrite candidates based on their respective scores, rewriting the query related to the multimedia content based on a particular ranked query rewrite candidate and providing for display, responsive to the query related to the multimedia content, a result set from the search engine based on the rewritten query.Type: GrantFiled: November 19, 2021Date of Patent: December 19, 2023Assignee: GOOGLE LLCInventors: Gökhan Hasan Bakir, Károly Csalogány, Behshad Behzadi
-
Patent number: 11823678Abstract: Techniques for determining a command or intent likely to be subsequently invoked by a user of a system are described. A user inputs a command (either via a spoken utterance or textual input) to a system. The system determines content responsive to the command. The system also determines a second command or corresponding intent likely to be invoked by the user subsequent to the previous command. Such determination may involve analyzing pairs of intents, with each pair being associated with a probability that one intent of the pair will be invoked by a user subsequent to a second intent of the pair. The system then outputs first content responsive to the first command and second content soliciting the user as to whether the system to execute the second command.Type: GrantFiled: February 28, 2022Date of Patent: November 21, 2023Assignee: Amazon Technologies, Inc.Inventors: Anjishnu Kumar, Xing Fan, Arpit Gupta, Ruhi Sarikaya
-
Patent number: 11823704Abstract: Coordinating signal processing among computing devices in a voice-driven computing environment is provided. A first and second digital assistant can detect an input audio signal, perform a signal quality check, and provide indications that the first and second digital assistants are operational to process the input audio signal. A system can select the first digital assistant for further processing. The system can receive, from the first digital assistant, data packets including a command. The system can generate, for a network connected device selected from a plurality of network connected devices, an action data structure based on the data packets, and transmit the action data structure to the selected network connected device.Type: GrantFiled: March 18, 2021Date of Patent: November 21, 2023Assignee: GOOGLE LLCInventors: Anshul Kothari, Gaurav Bhaya, Tarun Jain
-
Patent number: 11804215Abstract: An example process includes: receiving a first natural language input; initiating, by a digital assistant operating on the electronic device, a first task based on the first natural language input; determining whether the first task is of a predetermined type; and in accordance with a determination that the first task is of a predetermined type: determining whether one or more criteria are satisfied; and providing a response to the first natural language input, where providing the response includes: in accordance with a determination that the one or more criteria are not satisfied, outputting a first sound indicative of the initiated first task and a first verbal response indicative of the initiated first task; and in accordance with a determination that the one or more criteria are satisfied, outputting the first sound without outputting the first verbal response.Type: GrantFiled: September 21, 2022Date of Patent: October 31, 2023Assignee: Apple Inc.Inventors: Daniel A. Castellani, James N. Jones, Pedro Mari, Jessica J. Peck, Hugo D. Verweij, Garrett L. Weinberg, Mitchell R. Lerner
-
Patent number: 11798549Abstract: Embodiments include systems and methods for receiving an action item trigger by a user of a conferencing application; and in response to receiving the action item trigger, generating spoken words from audio data of a session of the conferencing application; normalizing the spoken words; generating higher-level representations of the normalized spoken words; determining semantic similarities of the higher-level representations of the normalized spoken words and higher level representations of normalized action words of an action word list; ranking options for top spoken words and action words based at least in part on the semantic similarities; identifying candidates for action words and/or phrases from the top spoken words and action words; and parsing the candidates to generate one or more action items.Type: GrantFiled: March 19, 2021Date of Patent: October 24, 2023Assignee: Mitel Networks CorporationInventors: Jonathan Braganza, Kevin Lee, Logendra Naidoo
-
Patent number: 11790914Abstract: The present disclosure generally relates to voice-control for electronic devices. In some embodiments, the method includes, in response to detecting a plurality of utterances, associating the plurality of operations with a first stored operation set and detecting a second set of one or more inputs corresponding to a request to perform the operations associated with the first stored operation set; and performing the plurality of operations associated with the first stored operation set, in the respective order.Type: GrantFiled: September 22, 2022Date of Patent: October 17, 2023Assignee: Apple Inc.Inventors: Kevin Bartlett Aitken, Clare T. Kasemset
-
Patent number: 11763804Abstract: A method of leveraging a dialogue history of a conversational computing interface to execute an updated dialogue plan. The method comprises maintaining an annotated dialogue history of the conversational computing interface. The annotated dialogue history includes a plurality of traced steps defining a data-flow including input data used to execute a context-dependent operation and output data recorded from a previous execution of the context-dependent operation. The method further comprises recognizing an updated dialogue plan including a prefix of executable steps and an updated executable step following the prefix. The method further comprises automatically computer-recognizing that the prefix of executable steps of the updated dialogue plan matches a corresponding prefix of traced steps in the annotated dialogue history. The method further comprises re-using the data-flow from the prefix of traced steps in the annotated dialogue history to automatically determine input data of the updated executable step.Type: GrantFiled: June 29, 2020Date of Patent: September 19, 2023Assignee: Microsoft Technology Licensing, LLCInventors: David Leo Wright Hall, Pengyu Chen, Jason Andrew Wolfe, Jayant Sivarama Krishnamurthy
-
Patent number: 11755756Abstract: Systems and methods for sensitive data management are disclosed. A voice-enabled device may generate audio data representing a request from a user utterance. A remote system may perform speech-processing operations, including obtaining responsive text data from a third-party application. In examples, a sensitivity designation may be received from the third-party application, which may cause the remote system to encrypt the responsive text data, redact the text data, and/or remove the text data from the remote system after the response is provided to the voice-enabled device.Type: GrantFiled: August 14, 2020Date of Patent: September 12, 2023Assignee: Amazon Technologies, Inc.Inventors: Jason Cline, Yolando Pereira, Arvind Kumar Babel, Bharanidharan Arul Janakiammal, Rohan Manish Chandra, Gary Scot Henderson
-
Patent number: 11749286Abstract: AM and LM parameters to be used for adapting an ASR model are derived for each audio segment of an audio stream comprising multiple audio programs. A set of identifiers, including a speaker identifier, a speaker domain identifier and a program domain identifier, is obtained for each audio segment. The set of identifiers are used to select most suitable AM and LM parameters for the particular audio segment. The embodiments enable provision of maximum constraints on the AMs and LMs and enable adaptation of the ASR model on the fly for audio streams of multiple audio programs, such as broadcast audio. This means that the embodiments enable selecting AM and LM parameters that are most suitable in terms of ASR performance for each audio segment.Type: GrantFiled: December 6, 2021Date of Patent: September 5, 2023Assignee: Telefonaktiebolaget LM Ericsson (publ)Inventors: Volodya Grancharov, Erlendur Karlsson, Sigurdur Sverrisson, Maxim Teslenko, Konstantinos Vandikas, Aneta Vulgarakis Feljan
-
Patent number: 11734502Abstract: Systems and methods to maintain amends to an annotation as discrete chronological events are disclosed. Exemplary implementations may: obtain a selection of a first annotation template for a first annotation via a client computing platform; generate a root node based on the selection of the first annotation template; obtain a first command to update to the first annotation; append a first update node to the root node responsive to the first command; obtain a second command to update to the first annotation; append a second update node to the first node responsive to the second command; receive an indication to present the first annotation; generate, responsive to the indication, the first annotation by populating the first annotation template included in the root node based on the first node set and in sequential order indicated by the edges; and effectuate presentation of the first annotation.Type: GrantFiled: December 1, 2022Date of Patent: August 22, 2023Assignee: Suki AI, Inc.Inventors: Badarinarayan Parthasarathi Burli, Harish Chandra Thuwal, Sai Chaitanya Ramineni
-
Patent number: 11726806Abstract: A display apparatus is provided. The display apparatus according to an embodiment includes a display, and a processor configured to control the display to display a UI screen including a plurality of text objects, control the display to display a text object in a different language from a preset language among the plurality of text objects, along with a preset number, and in response to a recognition result of a voice uttered by a user including the displayed number, perform an operation relating to a text object corresponding to the displayed number.Type: GrantFiled: September 2, 2020Date of Patent: August 15, 2023Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventors: Yang-soo Kim, Suraj Singh Tanwar
-
Patent number: 11720324Abstract: Various embodiments of the present invention relate to an apparatus and a method for displaying an electronic document for processing a user's voice command in an electronic device. The electronic device includes an input device; a display; and a processor, wherein the processor may be configured to detect a voice command of a user using the input device, if outputting an electronic document corresponding to the voice command, identify at least one input field in the electronic document, determine guide information based on information of the at least one input field, and display the electronic document comprising the guide information using the display. Other embodiments may be possible.Type: GrantFiled: January 2, 2019Date of Patent: August 8, 2023Assignee: Samsung Electronics Co., Ltd.Inventors: Kyungtae Kim, Seonho Lee, Yoonjeong Choi, Hosung You, Bunam Jeon, Taeho Ha, Changho Lee
-
Patent number: 11715455Abstract: A machine has a processor and a memory connected to the processor. The memory stores instructions executed by the processor to supply a name page in response to a request from an administrator machine. Name page updates are received from the administrator machine. The name page updates include participants and associated network contact information for the participants. A code is utilized to form a link to the name page. Prompts for textual name information and audio name information are supplied to a client machine that activates the link to the name page. Textual name information and audio name information are received from the client machine. The textual name information and audio name information are stored in association with the name page. Navigation tools are supplied to facilitate access to the textual name information and audio name information.Type: GrantFiled: October 12, 2020Date of Patent: August 1, 2023Assignee: NAMECOACH, INC.Inventor: Praveen Shanbhag
-
Patent number: 11711469Abstract: Methods, computer program products, and systems are presented.Type: GrantFiled: May 10, 2021Date of Patent: July 25, 2023Assignee: International Business Machines CorporationInventors: Ryan Brink, Andrew R. Freed, Marco Noel
-
Patent number: 11690578Abstract: System and methods for controlling healthcare devices and systems using voice commands are presented. In some aspects a listening device may receive voice command from a person. The voice command may be translated into human readable or machine readable text via a speech-to-text service. A control component may receive the text and send device-specific instructions to a medical device associated with a patient based on the translated voice command. In response to the instructions, the medical device may take an action on a patient. Some examples of actions taken may include setting an alarm limit on a monitor actively monitoring a patient and adjusting the amount of medication delivered by an infusion pump. Because these devices may be controlled using a voice command, in some cases, no physical or manual interaction is needed with the device. As such, multiple devices may be hands-free controlled from any location.Type: GrantFiled: February 2, 2021Date of Patent: July 4, 2023Assignee: CERNER INNOVATION, INC.Inventors: Chad Hays, Randy Lantz
-
Patent number: 11676592Abstract: A natural-language voice chatbot engages a consumer in a voice dialogue. The chatbot is customized for engaging the specific consumer based on features and characteristics of that specific consumer's speech and a lexicon associated with terms, words, and commands for item ordering. The consumer can perform voice queries for specific items and/or specific establishments for placing a pre-staged order with the chatbot. Once the consumer selects options with a specific establishment, a pre-staged order is provided to the corresponding establishment on behalf of the user. Location data for a consumer-operated device is monitored and when it is determined that the consumer will arrive at the establishment within a time period required by the establishment to prepare the pre-staged order, a message is sent to the establishment to begin preparing the pre-staged order.Type: GrantFiled: November 25, 2020Date of Patent: June 13, 2023Assignee: NCR CorporationInventors: Jodessiah Sumpter, Christian McDaniel, Kendall Marie Rose, Shaundell D. Thompson
-
Patent number: 11664024Abstract: An artificial intelligence device may receive first voice data corresponding to first voice uttered by a user from a first peripheral device, acquire a first intention corresponding to the first voice data, transmit a first search result corresponding to the first intention to the first peripheral device, receive second voice data corresponding to second voice uttered by the user from a second peripheral device, acquire a second intention corresponding to the received second voice data, and transmit a search result corresponding to the second intention to the second peripheral device depending on whether the second intention is an interactive intention associated with the first intention.Type: GrantFiled: August 20, 2020Date of Patent: May 30, 2023Assignee: LG ELECTRONICS INC.Inventors: Hyangjin Lee, Myeongok Son, Jaekyung Lee
-
Patent number: 11657803Abstract: Disclosed is a speech recognition method performed by one or more processors of a computing device, the speech recognition method including: performing first speech recognition on voice information to obtain first text information on the voice information; receiving feedback regarding the first text information; and generating final text information for the voice information based on the received feedback, in which the first speech recognition includes real-time speech recognition, and is performed through a neural network model of a first structure.Type: GrantFiled: November 2, 2022Date of Patent: May 23, 2023Assignee: ActionPower Corp.Inventors: Hyungwoo Kim, Dongchan Shin
-
Patent number: 11657107Abstract: Methods and systems for generating search results are disclosed. In some examples, one or more keywords are extracted from one or more stored reviews associated with a merchant offering. A first extracted keyword is associated with a stored listing of the merchant offering. The first extracted keyword may be absent from the stored listing. In response to a search query that includes the first keyword, a set of search results is provided, where the set of search results includes the listing associated with the first keyword.Type: GrantFiled: April 5, 2022Date of Patent: May 23, 2023Assignee: SHOPIFY INC.Inventors: Siavash Ghorbani, Carl Johan Gustavsson
-
Patent number: 11651765Abstract: Techniques and apparatuses for recognizing accented speech are described. In some embodiments, an accent module recognizes accented speech using an accent library based on device data, uses different speech recognition correction levels based on an application field into which recognized words are set to be provided, or updates an accent library based on corrections made to incorrectly recognized speech.Type: GrantFiled: October 14, 2020Date of Patent: May 16, 2023Assignee: Google Technology Holdings LLCInventor: Kristin A. Gray
-
Patent number: 11651770Abstract: The present disclosure is generally related to a data processing system to validate vehicular functions in a voice activated computer network environment. The data processing system can improve the efficiency of the network by discarding action data structures and requests that invalid prior to their transmission across the network. The system can invalidate requests by comparing attributes of a vehicular state to attributes of a request state.Type: GrantFiled: September 14, 2020Date of Patent: May 16, 2023Assignee: GOOGLE LLCInventors: Haris Ramic, Vikram Aggarwal, Moises Morgenstern Gali, David Roy Schairer, Yao Chen
-
Patent number: 11647249Abstract: The present disclosure relates to methods and devices for testing video data being rendered at or using a media device. A plurality of video frames to be rendered is received, each frame comprising one or more primary screen objects and at least one further screen object. The received frames are rendered at or using the media device wherein the at least one further screen object is superimposed on the one or more primary screen objects of a given frame during rendering. The rendered frames are provided to a data model. Extracted metadata indicating the presence or absence of further screen objects in the rendered video frames is the output of the data model. The data model is also provided with original metadata associated with the video frames prior to rendering. The rendering of each further screen object is then tested based on the original metadata and extracted metadata relating to a given video frame.Type: GrantFiled: August 9, 2019Date of Patent: May 9, 2023Assignee: NAGRAVISION S.A.Inventors: Douglas Gore, Ping Zou
-
Patent number: 11631407Abstract: Smart speaker system mechanisms, associated with a smart speaker device comprising an audio capture device, are provided for processing audio sample data captured by the audio capture device. The mechanisms receive, from the audio capture device of the smart speaker device, an audio sample captured from a monitored environment. The mechanisms classify a sound in the audio sample data as a type of sound based on performing a joint analysis of a plurality of different characteristics of the sound and matching results of the joint analysis to criteria specified in a plurality of sound models. The mechanisms determine, based on the classification of the sound, whether a responsive action is to be performed based on the classification of the sound. In response to determining that a responsive action is to be performed, the mechanisms initiate performance of the responsive action by the smart speaker system.Type: GrantFiled: September 9, 2020Date of Patent: April 18, 2023Assignee: International Business Machines CorporationInventors: Michael S. Gordon, James Kozloski, Ashish Kundu, Clifford A. Pickover, Komminist Weldemariam
-
Patent number: 11620340Abstract: Systems and methods for a media guidance application that generates results in multiple languages for search queries. In particular, the media guidance application resolves multiple language barriers by taking automatic and manual user language settings and applying those settings to a variety of potential search results.Type: GrantFiled: July 13, 2020Date of Patent: April 4, 2023Assignee: Rovi Product CorporationInventor: Arun Sreedhara
-
Patent number: 11616756Abstract: Systems, methods, and computer-readable storage media for enabling secure transfer of Internet domains between registrars. An example method can include receiving, at a registry, a request from a first registrar for information associated with an object recorded in the registry and registered by the first registrar, then generating, at the registry, an authorization code, the authorization code having an expiration. The registry can then transmit, to the first registrar, the authorization code, which in turn can be given to the registrant. The registrant can forward the authorization code to the second registrar, and the registry can receive, from a second registrar before the expiration has been reached: the authorization code and a transfer request for the object, the transfer request identifying a transfer of the object from the first registrar to the second registrar.Type: GrantFiled: April 14, 2022Date of Patent: March 28, 2023Assignee: VeriSign, Inc.Inventors: James Gould, Srikanth Veeramachaneni, Matthew Pozun
-
Patent number: 11610590Abstract: AM and LM parameters to be used for adapting an ASR model are derived for each audio segment of an audio stream comprising multiple audio programs. A set of identifiers, including a speaker identifier, a speaker domain identifier and a program domain identifier, is obtained for each audio segment. The set of identifiers are used to select most suitable AM and LM parameters for the particular audio segment. The embodiments enable provision of maximum constraints on the AMs and LMs and enable adaptation of the ASR model on the fly for audio streams of multiple audio programs, such as broadcast audio. This means that the embodiments enable selecting AM and LM parameters that are most suitable in terms of ASR performance for each audio segment.Type: GrantFiled: March 30, 2021Date of Patent: March 21, 2023Assignee: Telefonaktiebolaget LM Ericsson (publ)Inventors: Volodya Grancharov, Erlendur Karlsson, Sigurdur Sverrisson, Maxim Teslenko, Konstantinos Vandikas, Aneta Vulgarakis Feljan
-
Patent number: 11595459Abstract: A conferencing system terminal device includes a communication device receiving one or more videoconference feeds depicting one or more subjects engaged in a videoconference from one or more remote electronic devices. The conferencing system terminal device includes a contextual information extraction engine extracting contextual information of the videoconference from the one or more videoconference feeds. One or more processors automatically apply overlay indicia generated from the contextual information to at least one videoconference feed during the videoconference. A display presents the at least one videoconference feed after the overlay indicia is applied.Type: GrantFiled: October 14, 2021Date of Patent: February 28, 2023Assignee: Motorola Mobility LLCInventors: Alexandre Novaes Olivieri, Amit Kumar Agrawal