Procedures Used During A Speech Recognition Process, E.g., Man-machine Dialogue, Etc. (epo) Patents (Class 704/E15.04)
  • Patent number: 12254888
    Abstract: Techniques are described herein for multi-factor audio watermarking. A method includes: receiving audio data; processing the audio data to generate predicted output that indicates a probability of one or more hotwords being present in the audio data; determining that the predicted output satisfies a threshold that is indicative of the one or more hotwords being present in the audio data; in response to determining that the predicted output satisfies the threshold, processing the audio data using automatic speech recognition to generate a speech transcription feature; detecting a watermark that is embedded in the audio data; and in response to detecting the watermark: determining that the speech transcription feature corresponds to one of a plurality of stored speech transcription features; and in response to determining that the speech transcription feature corresponds to one of the plurality of stored speech transcription features, suppressing processing of a query included in the audio data.
    Type: Grant
    Filed: September 26, 2023
    Date of Patent: March 18, 2025
    Assignee: GOOGLE LLC
    Inventors: Aleks Kracun, Matthew Sharifi
  • Patent number: 12248796
    Abstract: This disclosure describes one or more implementations of systems, non-transitory computer-readable media, and methods that perform language guided digital image editing utilizing a cycle-augmentation generative-adversarial neural network (CAGAN) that is augmented using a cross-modal cyclic mechanism. For example, the disclosed systems generate an editing description network that generates language embeddings which represent image transformations applied between a digital image and a modified digital image. The disclosed systems can further train a GAN to generate modified images by providing an input image and natural language embeddings generated by the editing description network (representing various modifications to the digital image from a ground truth modified image). In some instances, the disclosed systems also utilize an image request attention approach with the GAN to generate images that include adaptive edits in different spatial locations of the image.
    Type: Grant
    Filed: July 23, 2021
    Date of Patent: March 11, 2025
    Assignee: Adobe Inc.
    Inventors: Ning Xu, Zhe Lin
  • Patent number: 12243512
    Abstract: An incorrect conversion dictionary generating system includes a term input unit, a voice conversion unit 5 that converts an input term to voice to obtain input-term voice, a voice analysis unit that receives the input-term voice output from the voice conversion unit, performs a voice analysis to convert the input-term voice to a term, and obtains a voice analyzed term, and an incorrectly converted term determining unit that receives the input term from the term input unit or the voice conversion unit, receives the voice analyzed term from the voice analysis unit, and determines the voice analyzed term as an incorrectly converted term of the input term when the input term does not match the voice analyzed term. The incorrect conversion dictionary generating system receives the input term and the incorrectly converted term, associates the input term with the incorrectly converted term, and stores in an incorrect conversion dictionary.
    Type: Grant
    Filed: February 10, 2020
    Date of Patent: March 4, 2025
    Assignee: Interactive Solutions Corp.
    Inventor: Kiyoshi Sekine
  • Patent number: 12233661
    Abstract: Disclosed is a digital pen that tracks the user's writing and provides useful feedback based on the user's writing. In one embodiment, the pen may provide feedback when the user has written a misspelled word, invalid mathematical expression, or any noncompliant expression. The pen may also provide feedback relating to the user's handwriting. The feedback may be visual, auditory, or tactile, and may be realtime or delayed. Statistics relating to the user's performance may be tracked, uploaded to external devices, and shared with others. This allows the user and interested parties to track the user's progress over time. The disclosed pen will be useful in educational settings.
    Type: Grant
    Filed: October 2, 2023
    Date of Patent: February 25, 2025
    Inventor: Lauren Michelle Neubauer
  • Patent number: 12230279
    Abstract: Techniques for authenticating users at devices that interact with the users via voice input. For instance, the described techniques may allow a voice-input device to safely verify the identity of a user by engaging in a back-and-forth conversation. The device or another device coupled thereto may then verify the accuracy of the responses from the user during the conversation, as well as compare an audio signature associated with the user's responses to a pre-stored audio signature associated with the user. By utilizing multiple checks, the described techniques are able to accurately and safely authenticate the user based solely on an audible conversation between the user and the voice-input device.
    Type: Grant
    Filed: August 6, 2021
    Date of Patent: February 18, 2025
    Assignee: Amazon Technologies, Inc.
    Inventor: Preethi Parasseri Narayanan
  • Patent number: 12216988
    Abstract: In some embodiments, a method is provided for updating an editing parameter for a model for automatically suggesting revisions to text data. The method may include displaying, on a graphical user interface (GUI) of a user device, one or more interactive input elements, wherein each of the one or more input elements is associated with an editing parameter for a model for automatically suggesting revisions to text data. The method may include receiving, via the GUI, an input from a selected input element of the one or more input elements, wherein the input comprises an indication of a value for a selected editing parameter associated with the selected input element. The method may include updating the selected editing parameter for the model based on the value. The method may include using the model with the updated selected editing parameter to apply an edit operation to an obtained text-under-analysis.
    Type: Grant
    Filed: May 8, 2023
    Date of Patent: February 4, 2025
    Assignee: BLACKBOILER, INC.
    Inventors: Liam Roshan Dunan Emmart, Jonathan Herr, Daniel P. Broderick, Daniel Edward Simonson
  • Patent number: 12211499
    Abstract: A method and electronic device for a vehicle are disclosed herein. The electronic device is mounted in the vehicle and includes a display, a memory storing voice commands, and a processor. The processor implements the method, including: obtaining at least one of vehicle driving information, occupant information, or display output information, generating one or more short commands by shortening one or more of the voice commands, based on the obtained at least one of the vehicle driving information, the occupant information and the display output information, and controlling the display to display one or more voice command guidance user interface (UI) displaying the one or more short commands.
    Type: Grant
    Filed: May 19, 2022
    Date of Patent: January 28, 2025
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Minsu Jung, Hyunwoo Kang, Taejun Kim, Jonghee Choi, Woong Lee
  • Patent number: 12211483
    Abstract: The present application provides a robot that a user feels more strongly to be a living being. A robot includes a speech generating unit that generates speech and a speech output unit that outputs the generated speech. The robot outputs speech generated by the robot rather than outputting speech prepared in advance. Because of this, speech that is in accordance with sensor information can be generated and output, or speech unique to the robot can be generated and output.
    Type: Grant
    Filed: February 26, 2021
    Date of Patent: January 28, 2025
    Assignee: GROOVE X, INC.
    Inventors: Kaname Hayashi, John Belmonte, Atsuya Kose, Masaya Matsuura, Hideya Minamiji, Taiji Fukaya, Takuma Miyamoto
  • Patent number: 12211502
    Abstract: Systems and processes for operating a virtual assistant to provide natural assistant interaction are provided. In accordance with one or more examples, a method includes, at an electronic device with one or more processors and memory: receiving a first audio stream including one or more utterances; determining whether the first audio stream includes a lexical trigger; generating one or more candidate text representations of the one or more utterances; determining whether at least one candidate text representation of the one or more candidate text representations is to be disregarded by the virtual assistant. If at least one candidate text representation is to be disregarded, one or more candidate intents are generated based on candidate text representations of the one or more candidate text representations other than the to be disregarded at least one candidate text representation.
    Type: Grant
    Filed: June 21, 2023
    Date of Patent: January 28, 2025
    Assignee: Apple Inc.
    Inventors: Juan Carlos Garcia, Paul S. McCarthy, Kurt Piersol
  • Patent number: 12198697
    Abstract: The method S200 can include: at an aircraft, receiving an audio utterance from air traffic control S210, converting the audio utterance to text, determining commands from the text using a question-and-answer model S240, and optionally controlling the aircraft based on the commands S250. The method functions to automatically interpret flight commands from the air traffic control (ATC) stream.
    Type: Grant
    Filed: March 19, 2024
    Date of Patent: January 14, 2025
    Assignee: Merlin Labs, Inc.
    Inventors: Michael Pust, Joseph Bondaryk, Matthew George
  • Patent number: 12198674
    Abstract: Disclosed are a speech synthesis method and apparatus, and a storage medium.
    Type: Grant
    Filed: March 18, 2020
    Date of Patent: January 14, 2025
    Assignees: BEIJING JINGDONG SHANGKE INFORMATION TECHNOLOGY CO., LTD., BEIJING JINGDONG CENTURY TRADING CO., LTD.
    Inventors: Zhizheng Wu, Wei Song
  • Patent number: 12198689
    Abstract: A system includes a processor; a memory; an electronic network; an application including computing instructions configured to be executed by the processor; and a multilingual dialogue server accessible by the application, comprising a language detection model, and an automatic speech recognition model; wherein the computing instructions of the app, when executed by the one or more processors, are configured to cause the one or more processors to: receive input; analyze the input using the trained model; select a speech model; analyze the input; generate metadata; and transmit the metadata. A method includes receiving input; analyzing the input; selecting a speech model; analyzing the input; generating metadata; and transmitting the metadata. A non-transitory computer readable medium containing program instructions that when executed, cause a computer to: receive input; analyze the input; select a trained speech model; analyze the input; generate metadata; and transmit the metadata.
    Type: Grant
    Filed: August 10, 2021
    Date of Patent: January 14, 2025
    Assignee: Summer Institute of Linguistics, Inc.
    Inventors: Daniel Lee Whitenack, Stephen Anthony Moitozo, II, Joshua John Nemecek
  • Patent number: 12197876
    Abstract: Systems and methods for managing conversations between a robot computing device and a user are disclosed. Exemplary implementations may: initiate a first-time user experience sequence with the user; teach the user the robot computing capabilities and/or characteristics; initiate, utilizing a dialog manager, a conversation with the user; receive, one or more command files from the user via one or more microphones; and generate conversation response files and communicating the generated conversation files to the dialog manager in response to the one or more received user global command files to initiate an initial conversation exchange.
    Type: Grant
    Filed: January 16, 2024
    Date of Patent: January 14, 2025
    Assignee: Embodied, Inc.
    Inventors: Stefan A Scherer, Mario E Munich, Paolo Pirjanian, Kevin D Saunders, Wilson Harron, Marissa Kohan
  • Patent number: 12190885
    Abstract: Configurable core domains of a speech processing system are described. A core domain output data format for a given command is originally configured with default content portions. When a user indicates additional content should be output for the command, the speech processing system creates a new output data format for the core domain. The new output data format is user specific and includes both default content portions as well as user preferred content portions.
    Type: Grant
    Filed: September 18, 2023
    Date of Patent: January 7, 2025
    Assignee: Amazon Technology, Inc.
    Inventors: Rohan Mutagi, Felix Wu, Rongzhou Shen, Neelam Satish Agrawal, Vibhunandan Gavini, Pablo Carballude Gonzalez
  • Patent number: 12189843
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for assisting a visually impaired user in interacting with a visual content entity are provided. One of the methods includes capturing, by an augmented reality device worn or held by the visually impaired user, a plurality of images of the visual content entity; notifying the visually impaired user of availability of the visual content entity and descriptive content associated with the visual content entity; receiving an indication from the visually impaired user to provide the descriptive content associated with the visual content entity; providing the descriptive content associated with the visual content entity and interactional guidance to the visually impaired user for interacting with the descriptive content; tracking one or more interactions by the visually impaired user with the descriptive content; and updating the interactional guidance provided to the visually impaired user based on the tracking.
    Type: Grant
    Filed: October 7, 2022
    Date of Patent: January 7, 2025
    Assignee: Chegg, Inc.
    Inventor: Vincent Le Chevalier
  • Patent number: 12184474
    Abstract: An information technology (IT) component associated with a first alert having an alert type is identified. A first list of recommended actions associated with the IT component is output. The first list includes a recommended action. A first user input of a user-selected action is received. An alert-to-component likelihood between the IT component and the alert type is decreased based on a determination that the first list does not include the user-selected action. The IT component is identified as being associated with a second alert based on the alert-to-component likelihood exceeding an alert-to-component likelihood threshold. A second list of recommended actions associated with the IT component is output. The second list does not include the recommended action and includes the user-selected action. A second user input of the user-selected action is received. A request to execute the user-selected action is transmitted.
    Type: Grant
    Filed: December 11, 2023
    Date of Patent: December 31, 2024
    Assignee: PagerDuty, Inc.
    Inventors: Vijay Shankar Venkataraman, Sanghamitra Goswami, Irena Grabovitch-Zuyev
  • Patent number: 12175968
    Abstract: Techniques for selecting a skill to execute in response to a natural language input are described. A system may receive a natural language input, determine profile data associated with the natural language input, and determine the profile data indicates a locale and at least first language and second languages. The system determines first and second sets of skills corresponding to the locale/first language and locale/second language, respectively. The system determines a first group of skill candidates corresponding to a portion of the first set of skills, and determines a second group of skill candidates corresponding to a portion of the second set of skills. The system performs ranking processing across the first and second groups of skills to determine a best skill for responding to the natural language input. Thus, in some situations, the skill invoked may not correspond to the same language represented in the natural language input.
    Type: Grant
    Filed: March 26, 2021
    Date of Patent: December 24, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Mohamed Farouk AbdelHady, Qian Hu, Mohamed Thahir Peer Mohamed, Wei Xiao, Zheng Gao, Radhika Arava, Xibin Gao
  • Patent number: 12165651
    Abstract: Systems and methods for optimizing voice detection via a network microphone device (NMD) based on a selected voice-assistant service (VAS) are disclosed herein. In one example, the NMD detects sound via individual microphones and selects a first VAS to communicate with the NMD. The NMD produces a first sound-data stream based on the detected sound using a spatial processor in a first configuration. Once the NMD determines that a second VAS is to be selected over the first VAS, the spatial processor assumes a second configuration for producing a second sound-data stream based on the detected sound. The second sound-data stream is then transmitted to one or more remote computing devices associated with the second VAS.
    Type: Grant
    Filed: June 30, 2023
    Date of Patent: December 10, 2024
    Assignee: Sonos, Inc.
    Inventors: Connor Kristopher Smith, Kurt Thomas Soto, Charles Conor Sleith
  • Patent number: 12165637
    Abstract: Systems and methods are described herein for providing media guidance. Control circuitry may receive a first voice input and access a database of topics to identify 5 a first topic associated with the first voice input. A user interface may generate a first response to the first voice input, and subsequent to generating the first response, the control circuitry may receive a second voice input. The control circuitry may 10 determine a match between the second voice input and an interruption input such as a period of silence or a keyword or a phrase, such as “Ahh,”, “Umm,”, or “Hmm.” The user interface may generate a second response that is associated with a second topic related to the first 15 topic. By interrupting the conversation and changing the subject from time to time, media guidance systems can appear to be more intelligent and human.
    Type: Grant
    Filed: July 6, 2023
    Date of Patent: December 10, 2024
    Assignee: Adeia Guides Inc.
    Inventors: Charles Dawes, Walter R. Klappert
  • Patent number: 12165634
    Abstract: A computer device acquires speech content. The device performs feature extraction on the speech content to obtain an intermediate feature. The intermediate feature is used for indicating an audio expression characteristic of the speech content. The device decodes the intermediate feature based on an attention mechanism to obtain a first word graph network. The device performs feature mapping on the intermediate feature based on pronunciation of the speech content to obtain a second word graph network. The device determines a recognition result of the speech content according to candidate word connection relationships indicated by the first word graph network and the second word graph network.
    Type: Grant
    Filed: November 2, 2022
    Date of Patent: December 10, 2024
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Xilin Zhang, Bo Liu, Shuo Liu
  • Patent number: 12159629
    Abstract: According to one embodiment, an information processing apparatus includes a processor. The processor generates a template, regarding a recording data sheet including a plurality of items, for one or more of the items that can be specified, with reference to an input order of input target items selected from the items. The processor performs a speech recognition on an utterance of a user and generate a speech recognition result. The processor determines an input target range relating to one more items specified by the utterance of the user among the items based on the template and the speech recognition result.
    Type: Grant
    Filed: February 25, 2022
    Date of Patent: December 3, 2024
    Assignee: KABUSHIKI KAISHA TOSHIBA
    Inventors: Nayuko Watanabe, Toshiaki Nakasu
  • Patent number: 12154133
    Abstract: The present invention provides improved techniques for identifying online advertisement laundering and online advertisement injection that may occur during the serving of advertisements to browsers. For example, in one embodiment, a method of identifying fraudulent advertising activities may comprise obtaining information identifying a computer system of every party involved in supplying an advertisement to an end user device, determining whether a computer system of a first party and a computer system of a second party provide websites and are not intermediary computer systems, determining whether the computer system of the first party provides a website of a type undesirable to advertisers, determining whether the computer system of the second party has usage and traffic characteristics indicative of not being an intermediary computer system, and determining a frequency of recurrence of the determined conditions.
    Type: Grant
    Filed: November 7, 2014
    Date of Patent: November 26, 2024
    Assignee: DoubleVerify, Inc.
    Inventors: Matthew McLaughlin, Roy Kalman Rosenfeld
  • Patent number: 12141207
    Abstract: The present invention allows appropriate acquisition of focus points in a dialogue.
    Type: Grant
    Filed: August 14, 2019
    Date of Patent: November 12, 2024
    Assignee: NIPPON TELEGRAPH AND TELEPHONE CORPORATION
    Inventors: Setsuo Yamada, Yoshiaki Noda, Takaaki Hasegawa
  • Patent number: 12143538
    Abstract: A communication logging system facilitates communications between two or more users. A secured database stores communications from each user, and the system notifies the intended recipients of the communications. The system records times of creation of the communications, times of access to the communications, and who created and/or accessed the communications. The communications may not be altered or deleted once recorded, so there is a permanent source of reliable evidence as to the content of the communications and information associated therewith.
    Type: Grant
    Filed: June 13, 2022
    Date of Patent: November 12, 2024
    Inventor: Karen Paulson
  • Patent number: 12142275
    Abstract: One or more associated identifiers are determined based on one or more associated tag types of an interactable element of web content. The determined one or more associated identifiers are associated with the interactable element. Based on the association of the determined one or more associated identifiers with the interactable element, one of the interactable element is matched to a received speech input. An action is performed with respect to the interactable element based on the matching.
    Type: Grant
    Filed: January 16, 2023
    Date of Patent: November 12, 2024
    Assignee: ServiceNow, Inc.
    Inventors: Jebakumar Mathuram Santhosm Swvigaradoss, Satya Sarika Sunkara, Ankit Goel, Jason Aloia, Rishabh Verma
  • Patent number: 12125405
    Abstract: In some embodiments, a method includes receiving, at a processor, audio data related to a verbal conversation between a first user and a second user and converting, using a natural language processing algorithm, the audio data into a set of text data portions. The method further includes analyzing the set of text data portions to determine a domain of the verbal conversation and a set of concepts associated with the domain and retrieving a set of educational content files associated with the set of concepts to provide educational content related to and during the verbal conversation to the first user or the second user. The method includes sending at least one notification to cause at least one pause of the verbal conversation and automatically sending, during the at least one pause of the verbal conversation, a signal to playback at least one educational content file.
    Type: Grant
    Filed: April 1, 2022
    Date of Patent: October 22, 2024
    Assignee: Chegg, Inc.
    Inventor: Vincent Le Chevalier
  • Patent number: 12118994
    Abstract: Implementations set forth herein relate to an automated assistant that can provide suggestions for a user to interact with the automated assistant to control applications while in a vehicle. The suggestions can be provided to encourage hands-free interactions with the applications, by suggesting an assistant input that invokes the automated assistant to operate as an interface between the user and the applications. Assistant suggestions can be based on a context of a user and/or a context of the vehicle, such as content of a display interface of a device that the user is accessing while in the vehicle. For instance, the automated assistant can determine that an action that the user has employed an application to perform can be initialized more safely and/or in less time by utilizing a particular assistant input. This particular assistant input can then be rendered at an interface of a vehicle computing device.
    Type: Grant
    Filed: February 21, 2022
    Date of Patent: October 15, 2024
    Assignee: GOOGLE LLC
    Inventors: Sriram Natarajan, Yuxin Yu, Josh Brown, David Notario
  • Patent number: 12118991
    Abstract: Implemented is a configuration of being able to reliably notify a device characteristic to a user by naturally generating and outputting a system speech including the device characteristic during an interaction with the user. The configuration includes a data processing section that generates and outputs a system speech including device characteristic information. The data processing section selects device characteristic information that is relevant to what is talked about during an interaction between the user and the information processing device, and generates and outputs a system speech including the selected device characteristic information. The data processing section receives an input of information of a user selection made to an option presented to the user, includes device characteristic information that is relevant to the inputted selection information into a system speech to be executed in an interaction with the user, and outputs the system speech.
    Type: Grant
    Filed: May 29, 2019
    Date of Patent: October 15, 2024
    Assignee: SONY CORPORATION
    Inventor: Mari Saito
  • Patent number: 12112769
    Abstract: Provided is a method of performing automatic interpretation based on speaker separation by a user terminal, the method including: receiving a first speech signal including at least one of a user speech of a user and a user surrounding speech around the user from an automatic interpretation service providing terminal, separating the first speech signal into speaker-specific speech signals, performing interpretation on the speaker-specific speech signals in a language selected by the user on the basis of an interpretation mode, and providing a second speech signal generated as a result of the interpretation to at least one of a counterpart terminal and the automatic interpretation service providing terminal according to the interpretation mode.
    Type: Grant
    Filed: November 19, 2021
    Date of Patent: October 8, 2024
    Assignee: Electronics and Telecommunications Research Institute
    Inventors: Jeong Uk Bang, Seung Yun, Sang Hun Kim, Min Kyu Lee, Joon Gyu Maeng
  • Patent number: 12112745
    Abstract: An electronic device is disclosed. The present electronic device comprises: a voice receiving unit; and a processor, wherein the processor: when a user's voice is received through the voice receiving unit, determines an accumulation level of utterance history information corresponding to the characteristics of the user's voice; when the accumulation level of utterance history information is below a predetermined threshold level, provides response information corresponding to the user's voice on the basis of user information related to the characteristics of the user's voice; and when the accumulation level of utterance history information is equal to or higher than the predetermined threshold level, provides response information corresponding to the user's voice on the basis of the user information and the utterance history information.
    Type: Grant
    Filed: September 9, 2019
    Date of Patent: October 8, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Jisun Park, Minjin Rho
  • Patent number: 12105746
    Abstract: Embodiments are directed to surveying security environments. A subject index that includes entries may be generated based on a survey of a content system. A question a client may be matched to entries in the subject index. A prompt associated with the content system may be generated based on the entries, the data sources, or the question. Query models may be employed to obtain data associated with the question from the data sources. Other prompts may be generated based on the data from the data sources to generate candidate answers based on the question and the data from the data sources. An evaluation prompt that includes the candidate answers and the question may be generated to rank the candidate answers for correctness. Answers may be determined based on the ranking of the candidate questions such that top ranked candidate answers are provided to the client.
    Type: Grant
    Filed: May 1, 2024
    Date of Patent: October 1, 2024
    Assignee: Dropzone.ai, Inc.
    Inventors: Xue Jun Wu, Changhwan Oh, Eric Joseph Hammerle, Robert Rowland Foley
  • Patent number: 12106013
    Abstract: Certain aspects of the present disclosure provide techniques for executing a function in a software application through a conversational user interface based on a knowledge graph associated with the function. An example method generally includes receiving a request to execute a function in a software application through a conversational user interface. A graph definition of the function is retrieved from a knowledge engine. Input is iteratively requested through the conversational user interface for each parameter of the parameters identified in the graph definition of the function based on a traversal of the graph definition of the function. Based on a completeness graph associated with the function, it is determined that the requested inputs corresponding to the parameters identified in the graph definition of the function have been provided through the conversational user interface. The function is executed using the requested inputs as parameters for executing the function.
    Type: Grant
    Filed: September 30, 2021
    Date of Patent: October 1, 2024
    Assignee: INTUIT INC.
    Inventors: Cynthia Joann Osmon, Roger C. Meike, Sricharan Kallur Palli Kumar, Gregory Kenneth Coulombe
  • Patent number: 12101280
    Abstract: The present disclosure provides a method and an apparatus for providing responses in an event-related session. The event is associated with a predefined domain, and the session comprises an electronic conversational agent and at least one participant. At least one message from the at least one participant may be detected. A set of candidate responses may be retrieved, from an index set being based on the domain, according to the at least one message. The set of candidate responses may be optimized through filtering the set of candidate responses according to predetermined criteria. A response to the at least one message may be selected from the filtered set of candidate responses. The selected response may be provided in the session.
    Type: Grant
    Filed: April 2, 2020
    Date of Patent: September 24, 2024
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Jie Zhang, Jianyong Wang, Peng Chen, Zeyu Shang, Ye Yu
  • Patent number: 12100395
    Abstract: Some implementations relate to an automated assistant that can provide assistant suggestions based on content that the user has requested for the automated assistant to render at a computing device. In some instances, an assistant suggestion can correspond to a portion of the content that is not being rendered at the computing device but that is part of the content that the user is accessing. In other instances, an assistant suggestion can be based on data that can indicate how one or more users may have interacted with their automated assistant when viewing the content. Therefore, an assistant suggestion can correspond to an assistant operation that may have been initialized by another user who had previously viewed the content. In some implementations, an arrangement of the assistant suggestions can be modified according to which part of the content a user is viewing and/or user distance.
    Type: Grant
    Filed: November 30, 2021
    Date of Patent: September 24, 2024
    Assignee: GOOGLE LLC
    Inventors: Stacy Cronin, Cliff Kuang, Abraham Wallin, Katrina Egert-Pecot, Sean Joseph Klassen, Caleb Misclevitz, Karan Patel, Ian Baker, Simon Sok, Ryan Allen
  • Patent number: 12094192
    Abstract: One or more multi-layer systems are used to perform inference. A multi-layer system may correspond to a node that receives a set of sensory input data for hierarchical processing, and may be grouped to perform processing for sensory input data. Inference systems at lower layers of a multi-layer system pass representation of objects to inference systems at higher layers. Each inference system can perform inference and form their own versions of representations of objects, regardless of the level and layer of the inference systems. The set of candidate objects for each inference system is updated to those consistent with feature-location representations for the sensors as well as object representations at lower layers. The set of candidate objects is also updated to those consistent with candidate objects from other inference systems, such as inference systems at other layers of the hierarchy or inference systems included in other multi-layer systems.
    Type: Grant
    Filed: July 20, 2021
    Date of Patent: September 17, 2024
    Assignee: Numenta, Inc.
    Inventors: Jeffrey C. Hawkins, Subutai Ahmad
  • Patent number: 12079225
    Abstract: Provided is a conversation information generation device capable of easily generating supplemental information that supplements conversation information. A conversation device (100) includes a graph database (106) that structurally stores a plurality of node information by using edge information indicating a mutual relationship. The conversation device (100) then extracts a word from text information to be supplemented, and determines whether the word is a specific word needed to be supplemented. When the word is determined as the specific word, the conversation device (100) references the graph database (106) and generates supplemental information for supplementing the word on the basis of the edge information deriving from the node information coinciding with the word and of other node information.
    Type: Grant
    Filed: August 1, 2019
    Date of Patent: September 3, 2024
    Assignee: NTT DOCOMO, INC.
    Inventors: Miyu Sato, Kanako Oonishi
  • Patent number: 12080268
    Abstract: A system is provided for determining customized notification output based on user preferences and event information. A user may request a system to provide a notification when a future event occurs. The system may determine a custom notification/output template for the notification, where a portion of the template is determined based on a user preference. When the event occurs, the system may generate a notification using the template, where portions of the notification are determined using event information and user preferences.
    Type: Grant
    Filed: March 30, 2020
    Date of Patent: September 3, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Vinaya Nadig, Rafael Colorado Vargas, Yugandhar Maram, Samarth Bhargava
  • Patent number: 12081712
    Abstract: An information processing apparatus instructs, if an audio input device such as a microphone is in use by the information processing apparatus, a printing apparatus to which a print job is transmitted to operate in a silent mode to reduce operation noise occurring in printing.
    Type: Grant
    Filed: June 6, 2023
    Date of Patent: September 3, 2024
    Assignee: Canon Kabushiki Kaisha
    Inventor: Hiroaki Morimoto
  • Patent number: 12080187
    Abstract: A reading comprehension system may include an authoring tool to help generate adaptable dialogs and a reading tool to conduct adaptable dialog sessions with students. The authoring tool may receive and process stories to generate labeled stories and information models. The information models may provide the conceptual structures that an effective reader should build while reading and understanding a story. The system may use dialog models for general dialogs and information models for story specific dialogs to guide adaptable dialog sessions with students. During the adaptable dialog sessions, the system may constantly assess and guide the student's progress in the understanding the current story and in general reading comprehension development. Using the labeled stories and dialog sessions as training data, the system may learn how to dialog effectively with the students, to gather an evolving understanding of the student's abilities, and to acquire knowledge about the world or the story.
    Type: Grant
    Filed: July 27, 2018
    Date of Patent: September 3, 2024
    Assignee: Elemental Cognition Inc.
    Inventors: David Ferrucci, David Melville, Gregory Burnham
  • Patent number: 12070688
    Abstract: A data processing apparatus includes input circuitry to receive audio data for a plurality of respective dialogue recordings for a video game, classification circuitry comprising one or more machine learning models to receive at least a portion of the audio data for each dialogue recording and trained to output classification data indicative of a quality classification of a dialogue recording in dependence upon one or more properties of the audio data for the dialogue recording, and storage circuitry to store identification data for one or more of the plurality of dialogue recordings in dependence upon the classification data.
    Type: Grant
    Filed: October 12, 2021
    Date of Patent: August 27, 2024
    Assignee: Sony Interactive Entertainment Inc.
    Inventors: Marina Villanueva Barreiro, Michael Lee Jones, Oliver Hume, Fabio Cappello, Danjeli Schembri
  • Patent number: 12073334
    Abstract: A method includes: obtaining a text entered by a user; determining at least one topic related to the text; determining a target dialogue robot from the plurality of dialogue robots based on the at least one topic related to the text and a predefined mapping relationship between a dialogue robot and a topic, where a target topic corresponding to the target dialogue robot is some or all of the at least one topic related to the text; allocating the text to the target dialogue robot; and obtaining a reply for the text from the target dialogue robot, where the reply is generated by the target dialogue robot based on at least one semantic understanding of the text.
    Type: Grant
    Filed: January 18, 2022
    Date of Patent: August 27, 2024
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Lifeng Shang, Zhengdong Lu, Hang Li
  • Patent number: 12068001
    Abstract: Techniques for detecting certain acoustic events from audio data are described. A system may perform event aggregation for certain types of events before sending an output to a device representing the event is detected. The system may bypass the event aggregation process for certain types of events that the system may detect with a high level of confidence. In such cases, the system may send an output to the device when the event is detected. The system may be used to detect acoustic events representing presence of a person or other harmful circumstances (such as, fire, smoke, etc.) in a home, an office, a store, or other types of indoor settings.
    Type: Grant
    Filed: September 8, 2023
    Date of Patent: August 20, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Harshavardhan Sundar, Sheetal Laad, Jialiang Bao, Ming Sun, Chao Wang, Chungnam Chan, Cengiz Erbas, Mathias Jourdain, Nipul Bharani, Aaron David Wirshba
  • Patent number: 12061636
    Abstract: A dialogue system, comprising: a first input configured to obtain first input data relating to speech or text provided by a user through a first interface; a first output configured to provide first output data relating to speech or text information specified by a determined dialogue act through the first interface; one or more processors, configured to: receive second input data through a second interface; store information specifying one or more configuration settings based on the second input data; and perform a dialogue method using a dialogue platform, the dialogue method comprising: determining dialogue information from the first input data; determining a dialogue act based on the determined dialogue information using a dialogue management module, wherein determining the dialogue act comprises selecting a next state from a plurality of states stored in the dialogue management module, wherein at least some of the plurality of states comprise information specifying a dialogue act and at least some of the
    Type: Grant
    Filed: May 15, 2023
    Date of Patent: August 13, 2024
    Assignee: POLYAI LIMITED
    Inventors: Tsung-Hsien Wen, Ivan Vulić, Nikola Mrk{hacek over (s)}ić, Pei-Hao Su, Pawel Franciszek Budzianowski, R{hacek over (a)}zvan-Emanuel Kusztos, Paul Julian Annetts, Ho Man Yau, Catherine Rachel Oxley, Emmanuel Sevrin, Vincent Yohann Dollet, I{circumflex over (n)}igo Casanueva Perez, Benjamin Peter Levin, Duong Hà Anh Nguyên, Swaroop Jagadeesh, Qian Zheng, Joshua Luke Jeffries Hook, Samuel John Coope
  • Patent number: 12062373
    Abstract: Introduced here are computer programs and associated computer-implemented techniques for facilitating the creation of a master transcription (or simply “transcript”) that more accurately reflects underlying audio by comparing multiple independently generated transcripts. The master transcript may be used to record and/or produce various forms of media content, as further discussed below. Thus, the technology described herein may be used to facilitate editing of text content, audio content, or video content. These computer programs may be supported by a media production platform that is able to generate the interfaces through which individuals (also referred to as “users”) can create, edit, or view media content. For example, a computer program may be embodied as a word processor that allows individuals to edit voice-based audio content by editing a master transcript, and vice versa.
    Type: Grant
    Filed: December 18, 2020
    Date of Patent: August 13, 2024
    Assignee: Descript, Inc.
    Inventors: Kundan Kumar, Vicki Anand
  • Patent number: 12045572
    Abstract: Disclosed is a system and method for processing out of scope or out of domain user inquiries with a first virtual assistant, which may include the steps of receiving a user request at a user device and converting the user request into a user inquiry, interpreting the user inquiry with an anomaly detection system to generate an interpretation of the user inquiry, forming a question from the interpretation using a textual composition module, accessing a dataset of text-based descriptions of a scope of the first virtual assistant using a textual composition module and a scope of an external source using a textual composition module, querying the dataset for an answer to the question, and when the answer is found in the description of the scope of an external source, transmitting the user inquiry to the external source for processing to generate a response to the user inquiry.
    Type: Grant
    Filed: June 21, 2021
    Date of Patent: July 23, 2024
    Assignee: MeetKai, Inc.
    Inventor: James Kaplan
  • Patent number: 12039228
    Abstract: The present application relates to an electronic device and a non-transitory storage medium. The electronic device of the embodiments includes a display processing unit and a voice operation processing unit. The display processing unit is configured to display on the display unit a first operation image including presentation of an operation item for operation input; if a predetermined condition is satisfied in the presentation of the first operation image, a second operation image is displayed on the display, the second operation image being an operation image in which text information corresponding to the operation item in the first operation image is highlighted. In the presentation of the second operation image, the voice operation processing unit is configured to receive the voice operation input corresponding to the text information highlighted in the second operation image as an operation on the video apparatus.
    Type: Grant
    Filed: December 15, 2021
    Date of Patent: July 16, 2024
    Assignees: Hisense Visual Technology Co., Ltd., TVS REGZA Corporation
    Inventors: Shuichi Hisatomi, Yu Onodera, Gen Omura, Hayato Nishimura, Toshihiro Sawaoka, Tetsuya Kitamura
  • Patent number: 12039977
    Abstract: An embodiment call termination apparatus includes a call termination word existence determination device configured to determine whether a call termination word exists in a call voice of a user, and a controller configured to compare the call termination word of the user with a call termination example word previously registered in a vehicle to control whether to terminate a call.
    Type: Grant
    Filed: July 29, 2021
    Date of Patent: July 16, 2024
    Assignees: HYUNDAI MOTOR COMPANY, KIA CORPORATION
    Inventors: Seong Soo Yae, Seo Hwan Choi, Hyun Woo Lee
  • Patent number: 12033637
    Abstract: Implementations described herein relate to an application and/or automated assistant that can identify arrangement operations to perform for arranging text during speech-to-text operations—without a user having to expressly identify the arrangement operations. In some instances, a user that is dictating a document (e.g., an email, a text message, etc.) can provide a spoken utterance to an application in order to incorporate textual content. However, in some of these instances, certain corresponding arrangements are needed for the textual content in the document. The textual content that is derived from the spoken utterance can be arranged by the application based on an intent, vocalization features, and/or contextual features associated with the spoken utterance and/or a type of the application associated with the document, without the user expressly identifying the corresponding arrangements.
    Type: Grant
    Filed: June 3, 2021
    Date of Patent: July 9, 2024
    Assignee: GOOGLE LLC
    Inventors: Victor Carbune, Krishna Sapkota, Behshad Behzadi, Julia Proskurnia, Jacopo Sannazzaro Natta, Justin Lu, Magali Boizot-Roche, Márius {hacek over (S)}ajgalík, Nicolo D'Ercole, Zaheed Sabur, Luv Kothari
  • Patent number: 12032909
    Abstract: Techniques are disclosed for generating a syntax for a neuro-linguistic model of input data obtained from one or more sources. A stream of words of a dictionary built from a sequence of symbols are received. The symbols are generated from an ordered stream of normalized vectors generated from input data. Statistics for combinations of words co-occurring in the stream are evaluated. The statistics includes a frequency upon which the combinations of words co-occur. A model of combinations of words based on the evaluated statistics is updated. The model identifies statistically relevant words. A connected graph is generated. Each node in the connected graph represents one of the words in the stream. Edges connecting the nodes represent a probabilistic relationship between words in the stream. Phrases are identified based on the connected graph.
    Type: Grant
    Filed: September 20, 2021
    Date of Patent: July 9, 2024
    Assignee: Intellective Ai, Inc.
    Inventors: Ming-Jung Seow, Gang Xu, Tao Yang, Wesley Kenneth Cobb
  • Patent number: 12026666
    Abstract: Provided are an electronic device capable of providing text information corresponding to a user voice through a user interface and a method of controlling the electronic device. Specifically, an electronic device according to the present disclosure, when an image including at least one object is obtained, analyzes the image to identify the at least one object included in the image, and when a user voice is received, performs voice recognition on the user voice to obtain text information corresponding to the user voice, then identifies an object corresponding to the user voice among the at least one object included in the image, and displays a memo user interface (UI) including text information on an area corresponding to the object identified as corresponding to the user voice among areas on a display.
    Type: Grant
    Filed: December 19, 2022
    Date of Patent: July 2, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Minkyu Shin, Sangyoon Kim, Dokyun Lee, Changwoo Han, Jonguk Yoo, Jaewon Lee