Patents Examined by Ibrahim Siddo
  • Patent number: 11862169
    Abstract: Providing speech-to-text (STT) transcription by a user endpoint device includes initiating an audio communication between an enterprise server and the user endpoint device, the audio communication comprising a voice interaction between a user associated with the user endpoint device and an agent associated with an agent device to which the enterprise server routes the audio communication; performing a first STT of at least a portion of the voice interaction to produce a first transcribed speech in a first language; concurrent with performing the first STT, performing, by the user endpoint device, a second STT of the at least the portion of the voice interaction to produce a second transcribed speech in a second language different than the first language, and transmitting the at least the portion of the voice interaction and at least the first transcribed speech from the user endpoint device to the enterprise server.
    Type: Grant
    Filed: September 11, 2020
    Date of Patent: January 2, 2024
    Assignee: Avaya Management L.P.
    Inventors: Valentine C. Matula, Pushkar Yashavant Deole, Sandesh Chopdekar, Navin Daga
  • Patent number: 11853652
    Abstract: Systems and methods to implement commands based on selection sequences to a user interface are disclosed. Exemplary implementations may: store, electronic storage, a library of terms utterable by users that facilitate implementation of intended results; obtain audio information representing sounds captured by a client computing platform; detect the spoken terms uttered by the user present within the audio information; determine whether the spoken terms detected are included in the library of terms; responsive to determination that the spoken terms are not included in the library of terms, effectuate presentation of an error message via the user interface; record a selection sequence that the user performs subsequent to the presentation of the error message that causes a result; correlate the selection sequence with the spoken terms based on the selection sequence recorded subsequent to error message to generate correlation; and store the correlation to the electronic storage.
    Type: Grant
    Filed: March 15, 2023
    Date of Patent: December 26, 2023
    Assignee: Suki AI, Inc.
    Inventors: Jatin Chhugani, Ganesh Satish Mallya, Alan Diec, Vamsi Reddy Chagari, Sudheer Tumu, Nithyanand Kota, Maneesh Dewan
  • Patent number: 11853405
    Abstract: Devices and methods to track contact between persons using acoustic voiceprint identification from stored audio signals are presented herein.
    Type: Grant
    Filed: September 1, 2021
    Date of Patent: December 26, 2023
    Assignee: Staton Techiya LLC
    Inventors: Steven Goldstein, John Patrick Keady
  • Patent number: 11849196
    Abstract: A method and system to automatically convert a presentation with slide materials to a digitized notetaking resource, by inputting a media stream from a presentation to a compute server, converting the media stream by segmenting the video into smaller segments, transcribing audio of the presenter's speech into text. Time stamp metadata is associated to elements of the segmented video (and, if available, slide data), audio, and transcribed text, and the elements are time ordered. A user interface is provided displaying elements of the segmented video/slide data and transcribed text. The user interface enables playback of the elements of the segmented video/slide data, audio of the presenter's speech, and transcribed text, wherein playback items are time-matched. Different times can be selected by a user, wherein the selected elements are made prominent in the display, with the audio of the presenter's speech also being time-matched to the selection.
    Type: Grant
    Filed: September 11, 2020
    Date of Patent: December 19, 2023
    Assignee: EDUCATIONAL VISION TECHNOLOGIES, INC.
    Inventors: Monal Mahesh Parmar, Jason John Bunk
  • Patent number: 11830518
    Abstract: A sound data processing method includes acquiring sound data of a target by input. The sound data processing method further includes: generating similar sound data that becomes a similar sound similar to the sound data of the target, based on the sound data of the target; and performing machine learning by using the acquired sound data of the target and the generated similar sound data as learning sound data, and generating a learning model for performing classification determination related to the sound data of the target.
    Type: Grant
    Filed: July 18, 2019
    Date of Patent: November 28, 2023
    Assignee: Panasonic Intellectual Property Management Co., Ltd.
    Inventor: Ryota Fujii
  • Patent number: 11823680
    Abstract: Implementations relate to an automated assistant that can respond to communications received via a third party application and/or other third party communication modality. The automated assistant can determine that the user is participating in multiple different conversations via multiple different third party communication services. In some implementations, conversations can be processed to identify particular features of the conversations. When the automated assistant is invoked to provide input to a conversation, the automated assistant can compare the input to the identified conversation features in order to select the particular conversation that is most relevant to the input. In this way, the automated assistant can assist with any of multiple disparate conversations that are each occurring via a different third party application.
    Type: Grant
    Filed: January 30, 2023
    Date of Patent: November 21, 2023
    Assignee: GOOGLE LLC
    Inventors: Victor Carbune, Matthew Sharifi
  • Patent number: 11817100
    Abstract: A computer-implemented method of voice-to-text tagging for transcription of a human voice signal by one of an automatic speech recognition system, or a natural conversation, includes generating a speech-to-text verbal transcript of the verbal vocalizations of an audio signal at a verbal timestamp of the audio signal. A voice-to-text non-verbal transcript of the one or more non-verbal vocalizations of the audio signal is generated at a non-verbal timestamp of the audio signal. The non-verbal transcript and the verbal transcript are combined to generate an enhanced transcript.
    Type: Grant
    Filed: June 26, 2020
    Date of Patent: November 14, 2023
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Robert John Moore, Shun Jiang, Chung-hao Tan, Eric Young Liu, Saurabh Mishra, Lei Huang, Guangjie Ren
  • Patent number: 11810562
    Abstract: Systems and processes for selectively processing and responding to a spoken user input are provided. In one example, audio input containing a spoken user input can be received at a user device. The spoken user input can be identified from the audio input by identifying start and end-points of the spoken user input. It can be determined whether or not the spoken user input was intended for a virtual assistant based on contextual information. The determination can be made using a rule-based system or a probabilistic system. If it is determined that the spoken user input was intended for the virtual assistant, the spoken user input can be processed and an appropriate response can be generated. If it is instead determined that the spoken user input was not intended for the virtual assistant, the spoken user input can be ignored and/or no response can be generated.
    Type: Grant
    Filed: August 30, 2021
    Date of Patent: November 7, 2023
    Assignee: Apple Inc.
    Inventors: Philippe P. Piernot, Justin G. Binder
  • Patent number: 11810549
    Abstract: A computing device trains a position optimization model for determining, from among a plurality of positions, one or more optimal positions on a face based on a training data set including facial skin strain data at the plurality of positions. The computing device trains a speech classification model for classifying a voice from the facial skin strain data based on the training data at the one or more optimal positions determined by the position optimization model among the training data set.
    Type: Grant
    Filed: June 3, 2021
    Date of Patent: November 7, 2023
    Assignees: Samsung Electronics Co., Ltd., SEOUL NATIONAL UNIVERSITY R&DB FOUNDATION
    Inventors: Sungroh Yoon, Eunji Kim, Heeseung Kim
  • Patent number: 11803709
    Abstract: A method, computer program product and computer system to provide topic guide during document drafting is provided. A processor retrieves at least one section of text from a document. A processor receives a target topic for the document. A processor extracts at least one local topic from the at least one section of text. A processor generates a semantic network comprising the at least one local topic and the target topic. A processor determines a deviation value for the at least one local topic based on a distance between the at least one local topic and the target topic in the semantic network. A processor, in response to the deviation value exceeding a threshold value, alerts a user that the at least one section of text from the document is off-topic from the target topic.
    Type: Grant
    Filed: September 23, 2021
    Date of Patent: October 31, 2023
    Assignee: International Business Machines Corporation
    Inventors: Xiang Yu Yang, Wen Jie Hao, Zhong Fang Yuan, Wang Hu Dang, Deng Xin Luo, Jia Yong Xie, Wen Wang
  • Patent number: 11800025
    Abstract: Image processing apparatuses have a unit for registering settings in applications or a unit for registering application settings on a portal screen, but do not allow these units to function in a coordinated manner. For example, if a user who frequently uses a certain application registers settings of the application, the settings may also be useful to other users. However, users who mainly use their own portal screens cannot invoke the settings. Settings of an application registered as a shared custom button are automatically registered as frequently used settings in the application. Similarly, settings registered as frequently used settings in an application are automatically registered as a shared custom button.
    Type: Grant
    Filed: September 13, 2022
    Date of Patent: October 24, 2023
    Assignee: CANON KABUSHIKI KAISHA
    Inventor: Kazuki Kado
  • Patent number: 11798571
    Abstract: Provided is an acoustic signal processing technique for performing a signal transformation suitable for desired signal processing (e.g., sound source enhancement processing) on a signal, and then performing the desired signal processing on the transformed signal. An acoustic signal processing device performs signal processing M which is a desired target on an input acoustic signal x.
    Type: Grant
    Filed: December 2, 2019
    Date of Patent: October 24, 2023
    Assignee: NIPPON TELEGRAPH AND TELEPHONE CORPORATION
    Inventors: Yuma Koizumi, Noboru Harada
  • Patent number: 11798536
    Abstract: A computer-implemented method, a computer system and a computer program product annotate media files with convenient pause points. The method includes acquiring a text file version of an audio narration file. The text file version includes a pause point history of a plurality of prior users. The method also includes generating a list of pause points based on the pause point history. In addition, the method includes determining a tone of voice being used by a speaker at each pause point using natural language processing algorithms. The method further includes determining a set of convenient pause points based on the list of pause points and the determined tone of voice. Lastly, the method includes inserting the determined set of convenient pause points into the audio narration file.
    Type: Grant
    Filed: June 14, 2021
    Date of Patent: October 24, 2023
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Jonathan Asghar, Belinda Marie Vennam
  • Patent number: 11790916
    Abstract: Systems and methods for processing speech transcription in a speech processing system are disclosed. Transcriptions of utterances is received and identifications to the transcriptions are assigned. In response to receiving an indication of an erroneous transcribed utterance in at least one of the transcriptions, an audio receiver is automatically activated for receiving a second utterance. In response to receiving the second utterance, an audio file of the second utterance and a corresponding identification of the erroneous transcribed utterance are transmitted to a speech recognition system for a second transcription, and the erroneous transcribed utterance is replaced with the second transcription.
    Type: Grant
    Filed: May 4, 2020
    Date of Patent: October 17, 2023
    Assignee: ROVI GUIDES, INC.
    Inventors: Sukanya Agarwal, Vikram Makam Gupta
  • Patent number: 11790914
    Abstract: The present disclosure generally relates to voice-control for electronic devices. In some embodiments, the method includes, in response to detecting a plurality of utterances, associating the plurality of operations with a first stored operation set and detecting a second set of one or more inputs corresponding to a request to perform the operations associated with the first stored operation set; and performing the plurality of operations associated with the first stored operation set, in the respective order.
    Type: Grant
    Filed: September 22, 2022
    Date of Patent: October 17, 2023
    Assignee: Apple Inc.
    Inventors: Kevin Bartlett Aitken, Clare T. Kasemset
  • Patent number: 11790901
    Abstract: Described herein are dialog systems, and techniques for providing such dialog systems, that are suitable for use on standalone computing devices. In some embodiments, a dialog system includes a dialog manager, which takes as input an input logical form, which may be a representation of user input. The dialog manager may include a dialog state tracker, an execution subsystem, a dialog policy subsystem, and a context stack. The dialog state tracker may generate an intermediate logical form from the input logical form combined with a context from the context stack. The context stack may maintain a history of a current dialog, and thus, the intermediate logical form may include contextual information potentially missing from the input logical form. The execution subsystem may execute the intermediate logical form to produce an execution result, and the dialog policy subsystem may generate an output logical form based on the execution result.
    Type: Grant
    Filed: December 30, 2022
    Date of Patent: October 17, 2023
    Assignee: Oracle International Corporation
    Inventors: Thanh Long Duong, Mark Edward Johnson, Vu Cong Duy Hoang, Tuyen Quang Pham, Yu-Heng Hong, Vladislavs Dovgalecs, Guy Bashkansky, Jason Eric Black, Andrew David Bleeker, Serge Le Huitouze
  • Patent number: 11783819
    Abstract: Disclosed are various approaches for generating a text transcript of a soundtrack. The soundtrack can correspond to an event in a conferencing service. Language models can be trained on data that is specific to organizations, users within the organization, and metadata associated with an agenda for the event. The metadata can include texts, attachments, and other data associated with the event. The language models can be arranged into a convolutional neural network and output a text transcript. The text transcript can be used to retrain the language models for subsequent use.
    Type: Grant
    Filed: July 2, 2021
    Date of Patent: October 10, 2023
    Assignee: VMware, INC.
    Inventors: Rohit Pradeep Shetty, Ravish Chawla, Adam Chow
  • Patent number: 11783828
    Abstract: Systems and methods for determining whether to combine responses from multiple automated assistants. An automated assistant may be invoked by a user utterance, followed by a query, which is provided to a plurality of automated assistants. A first response is received from a first automated assistant and a second response is received from a second automated assistant. Based on similarity between the responses, a primary automated assistant determines whether to combine the responses into a combined response. Once the combined response has been generated, one or more actions are performed in response to the combined response.
    Type: Grant
    Filed: April 15, 2021
    Date of Patent: October 10, 2023
    Assignee: GOOGLE LLC
    Inventors: Matthew Sharifi, Victor Carbune
  • Patent number: 11769005
    Abstract: Techniques are provided for assessing uniqueness of information using string-based collection frequency techniques. One method comprises obtaining multiple collections of documents from at least one data source; determining a collection frequency for a given character string based on a number of the collections comprising the given character string relative to a total number of the collections; assigning a uniqueness rating to the given character string based at least in part on a comparison of the collection frequency of the given character string to a collection frequency of one or more additional character strings in one or more of the plurality of collections; and performing an automated action using the given character string based on the assigned uniqueness rating. The automated action may comprise protecting the given character string and/or identifying the given character string as important information satisfying one or more importance criteria.
    Type: Grant
    Filed: May 29, 2020
    Date of Patent: September 26, 2023
    Assignee: EMC IP Holding Company LLC
    Inventors: Shou-Huey Jiang, Wenjin Liu, Chao Su
  • Patent number: 11769500
    Abstract: Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for providing augmented reality content in association with travel. The program and method provide for receiving, by a messaging application, a request to perform a scan operation in association with an image captured by a device camera; determining a travel parameter associated with the request and an attribute of an object depicted in the image; selecting an augmented reality content item based on at least one of the travel parameter or the attribute, the augmented reality content item being configured to present augmented reality content based on speech input; receiving the speech input; obtaining at least one of a transcription or translation of the speech input; and presenting the augmented reality content item, including the transcription or translation, in association with the image.
    Type: Grant
    Filed: April 8, 2021
    Date of Patent: September 26, 2023
    Assignee: Snap Inc.
    Inventors: Virginia Drummond, Ilteris Kaan Canberk, Jean Luo, Alek Matthiessen, Celia Nicole Mourkogiannis