Patents Examined by Jesse S Pullias
  • Patent number: 12223015
    Abstract: A computer-implemented method includes receiving a document insight request that requests document insights for a corpus of documents. The document insight request includes the corpus of documents, a set of entities contained within each document of the corpus of documents, and document insight request parameters that includes a confidence value threshold. The method also includes generating the document insights for the corpus of documents based on the confidence value threshold. Here, the document insights include an accuracy target and a user review rate target. The method also includes transmitting the document insights to the user device causing a graphical user interface to display the document insights on the user device.
    Type: Grant
    Filed: February 16, 2022
    Date of Patent: February 11, 2025
    Assignee: GOOGLE LLC
    Inventors: Emmanouil Koukoumidis, Nikolaos Kofinas, Evan Huang, Kiran Bellare, Xiao Liu, Michael Lanning, Lukas Rutishauser
  • Patent number: 12223963
    Abstract: A method of a local recognition system controlling a host device to perform one or more operations is provided. The method includes receiving, by the local recognition system, a query, performing speech recognition on the received query by implementing, by the local recognition system, a local language context comprising a set of words comprising descriptions in terms of components smaller than the words, and performing speech recognition, using the local language context, to create a transcribed query. Further, the method includes controlling the host device in dependence upon the speech recognition performed on the transcribed query.
    Type: Grant
    Filed: June 12, 2020
    Date of Patent: February 11, 2025
    Assignee: ScoutHound AI IP, LLC
    Inventors: Keyvan Mohajer, Timothy Stonehocker, Bernard Mont-Reynaud
  • Patent number: 12223945
    Abstract: A speech recognition method is provided. The method may include: obtaining speech data and a speech recognition result of the speech data, the speech data including speech of a plurality of speakers, and the speech recognition result including a plurality of words; determining speaking time of each of the plurality of speakers by processing the speech data; determining, based on the speaking times of the plurality of speakers and the speech recognition result, a corresponding relationship between the plurality of words and the plurality of speakers; determining, based on the corresponding relationship, at least one conversion word from the plurality of words, each of the at least one conversion word corresponding to at least two of the plurality of speakers; and re-determining the corresponding relationship between the plurality of words and the plurality of speakers based on the at least one conversion word.
    Type: Grant
    Filed: April 23, 2022
    Date of Patent: February 11, 2025
    Assignee: HITHINK ROYALFLUSH INFORMATION NETWORK CO., LTD.
    Inventors: Jinlong Wang, Xinkang Xu, Xinhui Hu
  • Patent number: 12217747
    Abstract: Disclosed is an electronic device including a communication interface, a memory, a microphone, a speaker, a display, a main processor, and a sub-processor activating the main processor by recognizing a wake-up word included in a voice input. The at least one memory stores instructions that, when executed, cause the main processor to receive a first voice input to register the wake-up word, when the first voice input does not include a specified word, to receive a second voice input including a word identical to the first voice input, through the microphone, to generate a wake-up word recognition model for recognizing the wake-up word, and to store the generated wake-up word recognition model in the at least one memory, and when the first voice input includes the specified word, to output information for requesting a third voice input, through the speaker or the display.
    Type: Grant
    Filed: August 23, 2019
    Date of Patent: February 4, 2025
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Euisuk Chung, Sangki Kang, Sunghwan Baek, Seokyeong Jung, Kyungtae Kim
  • Patent number: 12217760
    Abstract: A method for audio processing includes receiving a recording of a teleconference among multiple participants over a network, including an audio stream containing speech uttered by the participants and information outside the audio stream. The method further includes processing the audio stream to identify speech segments interspersed with intervals of silence, extracting speaker identifications from the information outside the audio stream in the received recording, labeling a first set of the identified speech segments from the audio stream with the speaker identifications, extracting acoustic features from the speech segments in the first set, learning a correlation between the speaker identifications labelled to the segments in the first set and the extracted acoustic features, and labeling a second set of the identified speech segments using the learned correlation, to indicate the participants who spoke during the speech segments in the second set.
    Type: Grant
    Filed: January 30, 2022
    Date of Patent: February 4, 2025
    Assignee: GONGIO Ltd.
    Inventors: Eilon Reshef, Hanan Shteingart, Zohar Shay, Shlomi Medalion
  • Patent number: 12210838
    Abstract: A computer-implemented method is provided for estimating output confidence of a black box Application Programming Interface (API). The method includes generating paraphrases for an input text. The method further includes calculating a distance between the input text and each respective one of the paraphrases. The method also includes sorting the paraphrases in ascending order of the distance. The method additionally includes selecting a top predetermined number of the paraphrases. The method further includes inputting the input text and the selected paraphrases into the API to obtain an output confidence score for each of the input text and the selected paraphrases. The method also includes estimating, by a hardware processor, the output confidence of the input text from a robustness of output scores of the input text and the selected paraphrases.
    Type: Grant
    Filed: August 15, 2023
    Date of Patent: January 28, 2025
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Yohei Ikawa, Issei Yoshida, Sachiko Yoshihama, Miki Ishikawa, Kohichi Kamijoh
  • Patent number: 12204866
    Abstract: Techniques for conversational-based searching are described. A system may receive a first spoken user input, and may determine that the first spoken user input corresponds to a request for information associated with an entity. The system may retrieve item results corresponding to the entity. The system may determine a suggested user input based on the retrieved item results, and may determine output data corresponding to the suggested user input. The system may send output data to a user device, where the output data includes the item results and the suggested user input. The system may receive a second spoken user input, and may determine that the second spoken user input corresponds to the suggested user input. In response, the system may send the previously determined output data to the device.
    Type: Grant
    Filed: September 10, 2021
    Date of Patent: January 21, 2025
    Assignee: Amazon Technologies, Inc.
    Inventors: Srinivasa Sandeep Atluri, Constantin Daniel Marcu, Kevin Small, Kemal Oral Cansizlar, Vijit Singh, Li Zhou, Aritra Biswas, Bhanu Pratap Jain
  • Patent number: 12204856
    Abstract: Data such as unstructured text is received that includes a sequence of sentences. This received data is then tokenized into a plurality of tokens. The received data is segmented using a hierarchical transformer network model including a token transformer, a sentence transformer, and a segmentation classifier. The token transformer contextualizes tokens within sentences and yields sentence embeddings. The sentences transformer contextualizes sentence representations based on the sentence embedddings. The segmentation classifier predicts segments of the received data based on the contextualized sentence representations. Data can be provided which characterizes the segmentation of the received data. Related apparatus, systems, techniques and articles are also described.
    Type: Grant
    Filed: September 23, 2021
    Date of Patent: January 21, 2025
    Assignee: Educational Testing Service
    Inventors: Swapna Somasundaran, Goran Glavaš
  • Patent number: 12204862
    Abstract: Systems and methods are provided for generating and training a relation extraction model configured to extract document-level relations. Systems obtain a knowledge database that comprises a plurality of entity tuples and a plurality of relation types, use the knowledge database to generate annotated relation instances based on relation instances that are identified in a set of unlabeled text, generate a training dataset comprising the annotated relation instances and the set of unlabeled text, and generate the machine learning model via modular self-supervision. Systems and methods are also provided for using a relation extraction model to extract document-level relations in specific use scenarios, such as for extracting drug response relations from full-text medical research articles.
    Type: Grant
    Filed: July 16, 2021
    Date of Patent: January 21, 2025
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Sheng Zhang, Cliff Richard Wong, Naoto Usuyama, Sarthak Jain, Tristan Josef Naumann, Hoifung Poon
  • Patent number: 12198671
    Abstract: In some implementations, a language proficiency of a user of a client device is determined by one or more computers. The one or more computers then determines a text segment for output by a text-to-speech module based on the determined language proficiency of the user. After determining the text segment for output, the one or more computers generates audio data including a synthesized utterance of the text segment. The audio data including the synthesized utterance of the text segment is then provided to the client device for output.
    Type: Grant
    Filed: April 28, 2023
    Date of Patent: January 14, 2025
    Assignee: Google LLC
    Inventors: Matthew Sharifi, Jakob Nicolaus Foerster
  • Patent number: 12190863
    Abstract: Automated systems and methods are provided for processing speech, comprising obtaining a trained machine learning model that has been trained using a cumulative historical data structure corresponding to at least one digitally-encoded speech representation for a plurality of telecommunications interactions conducted by a plurality of agent-side participants, which includes a first data corresponding to a score variable and a second data corresponding to a plurality of driver variables; applying the trained machine learning model: to a subset of data in the cumulative historical data structure that corresponds to a first agent-side participant of the plurality of agent-side participants, to generate a performance classification score and/or a performance direction classification score, to identify an intervention-target agent-side participant from among the plurality of agent-side participants, and to the cumulative historical data structure to identify an intervention training plan; and conducting at least on
    Type: Grant
    Filed: May 23, 2022
    Date of Patent: January 7, 2025
    Assignee: Conduent Business Services, LLC
    Inventors: Dennis F. Quebe, Jian Feng, Ambrish Gupta, Ashwin Subramanyam
  • Patent number: 12182526
    Abstract: Implementations relate to effectively localizing system responses, that include dynamic information, to target language(s), such that the system responses are grammatical and/or natural in the target language(s). Some of those implementations relate to various techniques for resource efficient generation of templates for a target language. Some versions of those implementations relate to resource efficient generation of target language natural language generation (NLG) templates and, more particularly, to techniques that enable a human user to generate a target language NLG template more efficiently and/or with greater accuracy. The more efficient target language NLG template generation enables less utilization of various client device resources and/or can mitigate the risk of flawed NLG templates being provided for live use in one or more systems.
    Type: Grant
    Filed: May 12, 2021
    Date of Patent: December 31, 2024
    Assignee: GOOGLE LLC
    Inventors: Katherine Vadella, Joshua Andrews, Max Copperman, Gabrielle Gayles, Shanjian Li, Jieyu Lu, Luchuan Xu
  • Patent number: 12182183
    Abstract: The present application provides a robot response method, apparatus, device and storage medium. The method includes: obtaining, by a robot, current query voice; extracting semantic information of the current query voice; matching the semantic information of the current query voice with multiple semantic information clusters stored in advance to get a matched target semantic information cluster, where each semantic information cluster includes: at least one Q&A instance, and each Q&A instance includes: semantic information corresponding to a historical query voice and a query question selected in a query list corresponding to the historical query voice; and obtaining, by the robot, the number of times each query question was selected in the target semantic information cluster, determining, according to the number of times each query question was selected, a target query question corresponding to the current query voice, and outputting a query response corresponding to the target query question.
    Type: Grant
    Filed: April 20, 2020
    Date of Patent: December 31, 2024
    Assignee: JINGDONG TECHNOLOGY HOLDING CO., LTD.
    Inventor: Yuyu Zheng
  • Patent number: 12183320
    Abstract: A method for generating synthetic speech for text through a user interface is provided. The method may include receiving one or more sentences, determining a speech style characteristic for the received one or more sentences, and outputting a synthetic speech for the one or more sentences that reflects the determined speech style characteristic. The one or more sentences and the determined speech style characteristic may be inputted to an artificial neural network text-to-speech synthesis model and the synthetic speech may be generated based on the speech data outputted from the artificial neural network text-to-speech synthesis model.
    Type: Grant
    Filed: January 20, 2021
    Date of Patent: December 31, 2024
    Assignee: NEOSAPIENCE, INC.
    Inventors: Taesu Kim, Younggun Lee
  • Patent number: 12177379
    Abstract: A centralized and robust threat assessment tool is disclosed to perform comprehensive analysis of previously-stored and subsequent communication data, activity data, and other relevant information relating to inmates within a controlled environment facility. As part of the analysis, the system detects certain keywords and key interactions with the dataset in order to identify particular criminal proclivities of the inmate. Based on the identified proclivities, the system assigns threat scores to inmate that represents a relative likelihood that the inmate will carry out or be drawn to certain threats and/or criminal activities. This analysis provides a predictive tool for assessing an inmate's ability to rehabilitate. Based on the analysis, remedial measures can be taken in order to correct an inmate's trajectory within the controlled environment and increase the likelihood of successful rehabilitation, as well as to prevent potential criminal acts.
    Type: Grant
    Filed: January 5, 2023
    Date of Patent: December 24, 2024
    Assignee: Global Tel*Link Corporation
    Inventor: Mitch Volkart
  • Patent number: 12164828
    Abstract: A method in an interactive computing-system includes pre-processing an input natural-language (NL) from a user command based on natural language processing (NLP) for classifying speech information and non-speech information, obtaining an NLP result from the user command, fetching a device specific information from one or more IoT devices operating in an environment based on the NLP result, generating one or more contextual parameters based on the NLP result and the device specific information, selecting at least one speaker embedding stored in a database for the one or more IoT devices based on the one or more contextual parameters, and outputting the selected at least one speaker embedding for playback to the user.
    Type: Grant
    Filed: June 10, 2021
    Date of Patent: December 10, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Sourabh Tiwari, Akshit Jindal, Saksham Goyal, Vinay Vasanth Patage, Ravibhushan B. Tayshete
  • Patent number: 12165661
    Abstract: There is inter alia disclosed an apparatus for spatial audio encoding which can receive or determine for one or more audio signals (102), spatial audio parameters (106) on a sub band basis for providing spatial audio reproduction, the spatial audio parameters can comprise a coherence value (112) for each sub band of a plurality of subbands (202) of a frame. The apparatus then determines a significance measure for the coherence values (401) of the plurality of sub bands of the frame and uses the significance measure to determine whether to encode (403) the coherence values of the plurality of sub bands of the frame.
    Type: Grant
    Filed: March 26, 2020
    Date of Patent: December 10, 2024
    Assignee: NOKIA TECHNOLOGIES OY
    Inventors: Mikko-Ville Laitinen, Adriana Vasilache
  • Patent number: 12159251
    Abstract: The present disclosure includes systems, apparatuses, and methods for event identification. In some aspects, a method includes receiving data including text and performing natural language processing on the received data to generate processed data that indicates one or more sentences. The method also includes generating, based on a first keyword set, a second keyword set having more keywords than the first keyword set. The method further includes, for each of the first and second keyword sets: detecting one or more keywords and one or more entities included in the processed data, determining one or more matched pairs based on the detected keywords and entities, and extracting a sentence, such as a single sentence or multiple sentences, from a document based on the one or more sentences indicated by the processed data. The method may also include outputting at least one extracted sentence.
    Type: Grant
    Filed: September 6, 2022
    Date of Patent: December 3, 2024
    Assignee: Thomson Reuters Enterprise Centre GmbH
    Inventors: Berk Ekmekci, Eleanor Hagerman, Blake Stephen Howald
  • Patent number: 12158902
    Abstract: Methods, systems, and computer programs are presented for searching the content of voice conversations. The conversations are translated into text and analysis of the conversation is performed to identify information in the conversation. The information identified includes turns taking data in the conversation and states identified within each state. A powerful user interface (UI) is provided to review the conversations and add annotations that tag the different turns. Additionally, parameter values are extracted from the text. A powerful search engine is provided with multiple search options, such as searching for text, searching by state within the conversation, searching by parameters extracted from the conversation, or a combination thereof.
    Type: Grant
    Filed: July 19, 2021
    Date of Patent: December 3, 2024
    Assignee: Twilio Inc.
    Inventors: Luke Percival de Oliveira, Umair Akeel, Alfredo Láinez Rodrigo, Nicolas Acosta Amador, Sahil Kumar, Liat Barda Dremer, Byeongung Ahn, Tyler Cole
  • Patent number: 12154582
    Abstract: A system and method code an object-based audio signal comprising audio objects in response to audio streams with associated metadata. In the system and method, a metadata processor codes the metadata and generates information about bit-budgets for the coding of the metadata of the audio objects. An encoder codes the audio streams while a bit-budget allocator is responsive to the information about the bit-budgets for the coding of the metadata of the audio objects from the metadata processor to allocate bitrates for the coding of the audio streams by the encoder.
    Type: Grant
    Filed: July 7, 2020
    Date of Patent: November 26, 2024
    Inventor: Vaclav Eksler