Patents Examined by Jesse S Pullias
-
Patent number: 12223015Abstract: A computer-implemented method includes receiving a document insight request that requests document insights for a corpus of documents. The document insight request includes the corpus of documents, a set of entities contained within each document of the corpus of documents, and document insight request parameters that includes a confidence value threshold. The method also includes generating the document insights for the corpus of documents based on the confidence value threshold. Here, the document insights include an accuracy target and a user review rate target. The method also includes transmitting the document insights to the user device causing a graphical user interface to display the document insights on the user device.Type: GrantFiled: February 16, 2022Date of Patent: February 11, 2025Assignee: GOOGLE LLCInventors: Emmanouil Koukoumidis, Nikolaos Kofinas, Evan Huang, Kiran Bellare, Xiao Liu, Michael Lanning, Lukas Rutishauser
-
Patent number: 12223963Abstract: A method of a local recognition system controlling a host device to perform one or more operations is provided. The method includes receiving, by the local recognition system, a query, performing speech recognition on the received query by implementing, by the local recognition system, a local language context comprising a set of words comprising descriptions in terms of components smaller than the words, and performing speech recognition, using the local language context, to create a transcribed query. Further, the method includes controlling the host device in dependence upon the speech recognition performed on the transcribed query.Type: GrantFiled: June 12, 2020Date of Patent: February 11, 2025Assignee: ScoutHound AI IP, LLCInventors: Keyvan Mohajer, Timothy Stonehocker, Bernard Mont-Reynaud
-
Patent number: 12223945Abstract: A speech recognition method is provided. The method may include: obtaining speech data and a speech recognition result of the speech data, the speech data including speech of a plurality of speakers, and the speech recognition result including a plurality of words; determining speaking time of each of the plurality of speakers by processing the speech data; determining, based on the speaking times of the plurality of speakers and the speech recognition result, a corresponding relationship between the plurality of words and the plurality of speakers; determining, based on the corresponding relationship, at least one conversion word from the plurality of words, each of the at least one conversion word corresponding to at least two of the plurality of speakers; and re-determining the corresponding relationship between the plurality of words and the plurality of speakers based on the at least one conversion word.Type: GrantFiled: April 23, 2022Date of Patent: February 11, 2025Assignee: HITHINK ROYALFLUSH INFORMATION NETWORK CO., LTD.Inventors: Jinlong Wang, Xinkang Xu, Xinhui Hu
-
Patent number: 12217747Abstract: Disclosed is an electronic device including a communication interface, a memory, a microphone, a speaker, a display, a main processor, and a sub-processor activating the main processor by recognizing a wake-up word included in a voice input. The at least one memory stores instructions that, when executed, cause the main processor to receive a first voice input to register the wake-up word, when the first voice input does not include a specified word, to receive a second voice input including a word identical to the first voice input, through the microphone, to generate a wake-up word recognition model for recognizing the wake-up word, and to store the generated wake-up word recognition model in the at least one memory, and when the first voice input includes the specified word, to output information for requesting a third voice input, through the speaker or the display.Type: GrantFiled: August 23, 2019Date of Patent: February 4, 2025Assignee: Samsung Electronics Co., Ltd.Inventors: Euisuk Chung, Sangki Kang, Sunghwan Baek, Seokyeong Jung, Kyungtae Kim
-
Patent number: 12217760Abstract: A method for audio processing includes receiving a recording of a teleconference among multiple participants over a network, including an audio stream containing speech uttered by the participants and information outside the audio stream. The method further includes processing the audio stream to identify speech segments interspersed with intervals of silence, extracting speaker identifications from the information outside the audio stream in the received recording, labeling a first set of the identified speech segments from the audio stream with the speaker identifications, extracting acoustic features from the speech segments in the first set, learning a correlation between the speaker identifications labelled to the segments in the first set and the extracted acoustic features, and labeling a second set of the identified speech segments using the learned correlation, to indicate the participants who spoke during the speech segments in the second set.Type: GrantFiled: January 30, 2022Date of Patent: February 4, 2025Assignee: GONGIO Ltd.Inventors: Eilon Reshef, Hanan Shteingart, Zohar Shay, Shlomi Medalion
-
Patent number: 12210838Abstract: A computer-implemented method is provided for estimating output confidence of a black box Application Programming Interface (API). The method includes generating paraphrases for an input text. The method further includes calculating a distance between the input text and each respective one of the paraphrases. The method also includes sorting the paraphrases in ascending order of the distance. The method additionally includes selecting a top predetermined number of the paraphrases. The method further includes inputting the input text and the selected paraphrases into the API to obtain an output confidence score for each of the input text and the selected paraphrases. The method also includes estimating, by a hardware processor, the output confidence of the input text from a robustness of output scores of the input text and the selected paraphrases.Type: GrantFiled: August 15, 2023Date of Patent: January 28, 2025Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Yohei Ikawa, Issei Yoshida, Sachiko Yoshihama, Miki Ishikawa, Kohichi Kamijoh
-
Patent number: 12204866Abstract: Techniques for conversational-based searching are described. A system may receive a first spoken user input, and may determine that the first spoken user input corresponds to a request for information associated with an entity. The system may retrieve item results corresponding to the entity. The system may determine a suggested user input based on the retrieved item results, and may determine output data corresponding to the suggested user input. The system may send output data to a user device, where the output data includes the item results and the suggested user input. The system may receive a second spoken user input, and may determine that the second spoken user input corresponds to the suggested user input. In response, the system may send the previously determined output data to the device.Type: GrantFiled: September 10, 2021Date of Patent: January 21, 2025Assignee: Amazon Technologies, Inc.Inventors: Srinivasa Sandeep Atluri, Constantin Daniel Marcu, Kevin Small, Kemal Oral Cansizlar, Vijit Singh, Li Zhou, Aritra Biswas, Bhanu Pratap Jain
-
Patent number: 12204856Abstract: Data such as unstructured text is received that includes a sequence of sentences. This received data is then tokenized into a plurality of tokens. The received data is segmented using a hierarchical transformer network model including a token transformer, a sentence transformer, and a segmentation classifier. The token transformer contextualizes tokens within sentences and yields sentence embeddings. The sentences transformer contextualizes sentence representations based on the sentence embedddings. The segmentation classifier predicts segments of the received data based on the contextualized sentence representations. Data can be provided which characterizes the segmentation of the received data. Related apparatus, systems, techniques and articles are also described.Type: GrantFiled: September 23, 2021Date of Patent: January 21, 2025Assignee: Educational Testing ServiceInventors: Swapna Somasundaran, Goran Glavaš
-
Patent number: 12204862Abstract: Systems and methods are provided for generating and training a relation extraction model configured to extract document-level relations. Systems obtain a knowledge database that comprises a plurality of entity tuples and a plurality of relation types, use the knowledge database to generate annotated relation instances based on relation instances that are identified in a set of unlabeled text, generate a training dataset comprising the annotated relation instances and the set of unlabeled text, and generate the machine learning model via modular self-supervision. Systems and methods are also provided for using a relation extraction model to extract document-level relations in specific use scenarios, such as for extracting drug response relations from full-text medical research articles.Type: GrantFiled: July 16, 2021Date of Patent: January 21, 2025Assignee: Microsoft Technology Licensing, LLCInventors: Sheng Zhang, Cliff Richard Wong, Naoto Usuyama, Sarthak Jain, Tristan Josef Naumann, Hoifung Poon
-
Patent number: 12198671Abstract: In some implementations, a language proficiency of a user of a client device is determined by one or more computers. The one or more computers then determines a text segment for output by a text-to-speech module based on the determined language proficiency of the user. After determining the text segment for output, the one or more computers generates audio data including a synthesized utterance of the text segment. The audio data including the synthesized utterance of the text segment is then provided to the client device for output.Type: GrantFiled: April 28, 2023Date of Patent: January 14, 2025Assignee: Google LLCInventors: Matthew Sharifi, Jakob Nicolaus Foerster
-
Patent number: 12190863Abstract: Automated systems and methods are provided for processing speech, comprising obtaining a trained machine learning model that has been trained using a cumulative historical data structure corresponding to at least one digitally-encoded speech representation for a plurality of telecommunications interactions conducted by a plurality of agent-side participants, which includes a first data corresponding to a score variable and a second data corresponding to a plurality of driver variables; applying the trained machine learning model: to a subset of data in the cumulative historical data structure that corresponds to a first agent-side participant of the plurality of agent-side participants, to generate a performance classification score and/or a performance direction classification score, to identify an intervention-target agent-side participant from among the plurality of agent-side participants, and to the cumulative historical data structure to identify an intervention training plan; and conducting at least onType: GrantFiled: May 23, 2022Date of Patent: January 7, 2025Assignee: Conduent Business Services, LLCInventors: Dennis F. Quebe, Jian Feng, Ambrish Gupta, Ashwin Subramanyam
-
Patent number: 12182526Abstract: Implementations relate to effectively localizing system responses, that include dynamic information, to target language(s), such that the system responses are grammatical and/or natural in the target language(s). Some of those implementations relate to various techniques for resource efficient generation of templates for a target language. Some versions of those implementations relate to resource efficient generation of target language natural language generation (NLG) templates and, more particularly, to techniques that enable a human user to generate a target language NLG template more efficiently and/or with greater accuracy. The more efficient target language NLG template generation enables less utilization of various client device resources and/or can mitigate the risk of flawed NLG templates being provided for live use in one or more systems.Type: GrantFiled: May 12, 2021Date of Patent: December 31, 2024Assignee: GOOGLE LLCInventors: Katherine Vadella, Joshua Andrews, Max Copperman, Gabrielle Gayles, Shanjian Li, Jieyu Lu, Luchuan Xu
-
Patent number: 12182183Abstract: The present application provides a robot response method, apparatus, device and storage medium. The method includes: obtaining, by a robot, current query voice; extracting semantic information of the current query voice; matching the semantic information of the current query voice with multiple semantic information clusters stored in advance to get a matched target semantic information cluster, where each semantic information cluster includes: at least one Q&A instance, and each Q&A instance includes: semantic information corresponding to a historical query voice and a query question selected in a query list corresponding to the historical query voice; and obtaining, by the robot, the number of times each query question was selected in the target semantic information cluster, determining, according to the number of times each query question was selected, a target query question corresponding to the current query voice, and outputting a query response corresponding to the target query question.Type: GrantFiled: April 20, 2020Date of Patent: December 31, 2024Assignee: JINGDONG TECHNOLOGY HOLDING CO., LTD.Inventor: Yuyu Zheng
-
Patent number: 12183320Abstract: A method for generating synthetic speech for text through a user interface is provided. The method may include receiving one or more sentences, determining a speech style characteristic for the received one or more sentences, and outputting a synthetic speech for the one or more sentences that reflects the determined speech style characteristic. The one or more sentences and the determined speech style characteristic may be inputted to an artificial neural network text-to-speech synthesis model and the synthetic speech may be generated based on the speech data outputted from the artificial neural network text-to-speech synthesis model.Type: GrantFiled: January 20, 2021Date of Patent: December 31, 2024Assignee: NEOSAPIENCE, INC.Inventors: Taesu Kim, Younggun Lee
-
Patent number: 12177379Abstract: A centralized and robust threat assessment tool is disclosed to perform comprehensive analysis of previously-stored and subsequent communication data, activity data, and other relevant information relating to inmates within a controlled environment facility. As part of the analysis, the system detects certain keywords and key interactions with the dataset in order to identify particular criminal proclivities of the inmate. Based on the identified proclivities, the system assigns threat scores to inmate that represents a relative likelihood that the inmate will carry out or be drawn to certain threats and/or criminal activities. This analysis provides a predictive tool for assessing an inmate's ability to rehabilitate. Based on the analysis, remedial measures can be taken in order to correct an inmate's trajectory within the controlled environment and increase the likelihood of successful rehabilitation, as well as to prevent potential criminal acts.Type: GrantFiled: January 5, 2023Date of Patent: December 24, 2024Assignee: Global Tel*Link CorporationInventor: Mitch Volkart
-
Patent number: 12164828Abstract: A method in an interactive computing-system includes pre-processing an input natural-language (NL) from a user command based on natural language processing (NLP) for classifying speech information and non-speech information, obtaining an NLP result from the user command, fetching a device specific information from one or more IoT devices operating in an environment based on the NLP result, generating one or more contextual parameters based on the NLP result and the device specific information, selecting at least one speaker embedding stored in a database for the one or more IoT devices based on the one or more contextual parameters, and outputting the selected at least one speaker embedding for playback to the user.Type: GrantFiled: June 10, 2021Date of Patent: December 10, 2024Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventors: Sourabh Tiwari, Akshit Jindal, Saksham Goyal, Vinay Vasanth Patage, Ravibhushan B. Tayshete
-
Patent number: 12165661Abstract: There is inter alia disclosed an apparatus for spatial audio encoding which can receive or determine for one or more audio signals (102), spatial audio parameters (106) on a sub band basis for providing spatial audio reproduction, the spatial audio parameters can comprise a coherence value (112) for each sub band of a plurality of subbands (202) of a frame. The apparatus then determines a significance measure for the coherence values (401) of the plurality of sub bands of the frame and uses the significance measure to determine whether to encode (403) the coherence values of the plurality of sub bands of the frame.Type: GrantFiled: March 26, 2020Date of Patent: December 10, 2024Assignee: NOKIA TECHNOLOGIES OYInventors: Mikko-Ville Laitinen, Adriana Vasilache
-
Patent number: 12159251Abstract: The present disclosure includes systems, apparatuses, and methods for event identification. In some aspects, a method includes receiving data including text and performing natural language processing on the received data to generate processed data that indicates one or more sentences. The method also includes generating, based on a first keyword set, a second keyword set having more keywords than the first keyword set. The method further includes, for each of the first and second keyword sets: detecting one or more keywords and one or more entities included in the processed data, determining one or more matched pairs based on the detected keywords and entities, and extracting a sentence, such as a single sentence or multiple sentences, from a document based on the one or more sentences indicated by the processed data. The method may also include outputting at least one extracted sentence.Type: GrantFiled: September 6, 2022Date of Patent: December 3, 2024Assignee: Thomson Reuters Enterprise Centre GmbHInventors: Berk Ekmekci, Eleanor Hagerman, Blake Stephen Howald
-
Patent number: 12158902Abstract: Methods, systems, and computer programs are presented for searching the content of voice conversations. The conversations are translated into text and analysis of the conversation is performed to identify information in the conversation. The information identified includes turns taking data in the conversation and states identified within each state. A powerful user interface (UI) is provided to review the conversations and add annotations that tag the different turns. Additionally, parameter values are extracted from the text. A powerful search engine is provided with multiple search options, such as searching for text, searching by state within the conversation, searching by parameters extracted from the conversation, or a combination thereof.Type: GrantFiled: July 19, 2021Date of Patent: December 3, 2024Assignee: Twilio Inc.Inventors: Luke Percival de Oliveira, Umair Akeel, Alfredo Láinez Rodrigo, Nicolas Acosta Amador, Sahil Kumar, Liat Barda Dremer, Byeongung Ahn, Tyler Cole
-
Patent number: 12154582Abstract: A system and method code an object-based audio signal comprising audio objects in response to audio streams with associated metadata. In the system and method, a metadata processor codes the metadata and generates information about bit-budgets for the coding of the metadata of the audio objects. An encoder codes the audio streams while a bit-budget allocator is responsive to the information about the bit-budgets for the coding of the metadata of the audio objects from the metadata processor to allocate bitrates for the coding of the audio streams by the encoder.Type: GrantFiled: July 7, 2020Date of Patent: November 26, 2024Inventor: Vaclav Eksler