Creating Patterns For Matching Patents (Class 704/243)
-
Patent number: 11373043Abstract: According to an embodiment of a present disclosure, a method for generating and utilizing a text fingerprint performed by a computing device is disclosed. The method comprises the steps of: dividing text data into one or more segments based on a predetermined text segmentation algorithm; determining a mapping value assigned to one or more subsegments that form the divided segment based on a predetermined mapping algorithm; generating a coordinate value for each of the one or more segments based on the determined mapping value; and generating the virtual fingerprint having a phonetic feature for the text data based on the generated coordinate value. That is, whether a plurality of the text data are similar to each other can be easily determined, when the text data also has a unique virtual fingerprint based on a pronunciation, in the same way every person has a unique fingerprint.Type: GrantFiled: December 28, 2017Date of Patent: June 28, 2022Assignee: SEJIN MIND INC.Inventor: Kyung Pyo Ko
-
Patent number: 11367441Abstract: An electronic apparatus and a control method thereof are provided. A method of controlling an electronic apparatus according to an embodiment of the disclosure includes: receiving input of a first utterance, identifying a first task for the first utterance based on the first utterance, providing a response to the first task based on a predetermined response pattern, receiving input of a second utterance, identifying a second task for the second utterance based on the second utterance, determining the degree of association between the first task and the second task, and setting a response pattern for the first task based on the second task based on the determined degree of association satisfying a predetermined condition. The control method of an electronic apparatus may use an artificial intelligence model trained according to at least one of machine learning, a neural network, or a deep learning algorithm.Type: GrantFiled: November 1, 2019Date of Patent: June 21, 2022Assignee: Samsung Electronics Co., Ltd.Inventors: Yeonho Lee, Kyenghun Lee, Saebom Jang, Silas Jeon
-
Patent number: 11347784Abstract: A method for generating a set of concepts related to a target concept includes accessing a set of candidate concepts, embedding the target concept and the set of candidate concepts in a semantic vector space, selecting one or more intermediate concepts from the set of candidate concepts in response to determining whether each embedded candidate concept in the set of embedded candidate concepts satisfies a predetermined relationship with the embedded target concept, and filtering the one or more intermediate concepts to yield the set of concepts related to the target concept. The method may further include generating a multiple-choice question in which the target concept corresponds to a correct answer choice and the set of concepts related to the target concept correspond to distractors.Type: GrantFiled: December 15, 2021Date of Patent: May 31, 2022Assignee: CEREGO JAPAN KABUSHIKI KAISHAInventors: Michael A. Yen, Iain M. Harlow, Andrew Smith Lewis, Paul T. Mumma
-
Patent number: 11335352Abstract: A voice identity feature extractor training method includes extracting a voice feature vector of training voice, The method may include determining a corresponding I-vector according to the voice feature vector of the training voice. The method may include adjusting a weight of a neural network model by using the I-vector as a first target output of the neural network model, to obtain a first neural network model, The method may include obtaining a voice feature vector of target detecting voice and determining an output result of the first neural network model for the voice feature vector of the target detecting voice. The method may include determining an I-vector latent variable. The method may include estimating a posterior mean of the I-vector latent variable, and adjusting a weight of the first neural network model using the posterior mean as a second target output, to obtain a voice identity feature extractor.Type: GrantFiled: October 16, 2019Date of Patent: May 17, 2022Assignee: Tencent Technology (Shenzhen) Company LimitedInventors: Na Li, Jun Wang
-
Patent number: 11328712Abstract: Provided are techniques for domain specific correction of output from automatic speech recognition. An output of an automatic speech recognition engine is received. An alphanumeric sequence is extracted from the output, where the alphanumeric sequence represents an erroneous translation by the automatic speech recognition engine. Candidates for the alphanumeric sequence are generated. The candidates are ranked based on scores associated with the candidates. A candidate of the candidates having a highest score of the scores is selected. The output is corrected by replacing the alphanumeric sequence with the selected candidate. The corrected output is returned.Type: GrantFiled: August 2, 2019Date of Patent: May 10, 2022Assignee: International Business Machines CorporationInventors: Anbumunee Ponniah, Abhishek Singh, Nithin Mathew, Balasubramaniam Gurumurthy, Sunil Mayanna
-
Patent number: 11328007Abstract: Embodiments generally relate to the generation of a domain-specific phrasal dictionary. In some embodiments, a method includes receiving text from a user, wherein the text includes unstructured text of a natural language. The method further includes parsing the text into text chunks. The method further includes sending the text chunks to the user. The method further includes receiving one or more phrase categories and one or more predetermined phrases from the user, wherein each predetermined phrase of the one or more predetermined phrases corresponds to at least one phrase category of the one or more phrase categories. The method further includes comparing the predetermined phrases with the text chunks. The method further includes assigning at least one phrase category of the one or more phrase categories to at least one text chunk. The method further includes sending at least one text chunk and the at least one phrase category that is assigned to the at least one text chunk to the user.Type: GrantFiled: February 4, 2019Date of Patent: May 10, 2022Assignee: International Business Machines CorporationInventors: Dwi Sianto Mansjur, Scott Robert Carrier, Brendan Bull, Andrew Hicks
-
Patent number: 11315547Abstract: Provided is a system for generating speech recognition training data, the system including: a speech data processing module receiving speech data from a user terminal and performing data preprocessing on the received speech data; an auto speech recognition (ASR) interfacing module transmitting the preprocessed speech data to a plurality of ASR engines and acquiring a confidence score and transcription data of the speech data from the plurality of ASR engines; an ASR result evaluating module determining whether the speech data and the transcription data match each other; and a training data managing unit generating training data as a pair of the speech data and the transcription data determined to match each other.Type: GrantFiled: September 19, 2019Date of Patent: April 26, 2022Assignee: 42 MARU INC.Inventors: Dong Hwan Kim, Hyun Ok Kim, You Kyung Kwon
-
Patent number: 11308934Abstract: A method includes receiving text input data for conversion into synthesized speech and determining, using a hotword-aware model trained to detect a presence of a hotword assigned to a user device, whether a pronunciation of the text input data includes the hotword. The hotword is configured to initiate a wake-up process on the user device for processing the hotword and/or one or more other terms following the hotword in the audio input data. When the pronunciation of the text input data includes the hotword, the method also includes generating an audio output signal from the text input data and providing the audio output signal to an audio output device to output the audio output signal. The audio output signal when captured by an audio capture device of the user device, configured to prevent initiation of the wake-up process on the user device.Type: GrantFiled: June 25, 2018Date of Patent: April 19, 2022Assignee: Google LLCInventors: Matthew Sharifi, Aleksandar Kracun
-
Patent number: 11308939Abstract: A system and method performs wakeword detection and automatic speech recognition using the same acoustic model. A mapping engine maps phones/senones output by the acoustic model to phones/senones corresponding to the wakeword. A hidden Markov model (HMM) may determine that the wakeword is present in audio data; the HMM may have multiple paths for multiple wakewords or may have multiple models. Once the wakeword is detected, ASR is performed using the acoustic model.Type: GrantFiled: September 25, 2018Date of Patent: April 19, 2022Assignee: Amazon Technologies, Inc.Inventors: Yixin Gao, Ming Sun, Varun Nagaraja, Gengshen Fu, Chao Wang, Shiv Naga Prasad Vitaladevuni
-
Patent number: 11302327Abstract: Apparatus and methods for interactive voice recognition. The apparatus and methods may include a canonical phrase derivation engine configured to derive canonical phrases from voice data. The apparatus may include an input engine configured to parse utterances. The apparatus may include a knowledge extraction engine to disambiguate the utterances into words, form a sequence from the words, extract context from the sequence, pair the sequence with a phrase of the canonical phrases, merge the sequence and the phrase to form a hybrid phrase, vectorize the hybrid phrase into a vector, and feed the vector into a non-linear classification engine to determine an intent corresponding to the utterances.Type: GrantFiled: June 22, 2020Date of Patent: April 12, 2022Assignee: Bank of America CorporationInventors: Emad Noorizadeh, Ramakrishna R. Yannam
-
Patent number: 11289097Abstract: The present disclosure provides various embodiments of methods for intelligent active speaker identification and information handling systems (IHSs) utilizing such methods. In general, the methods disclosed herein may be used to accurately identify an active speaker in a communication session with an application or an IHS, regardless of whether the active speaker is alone, in a group environment, or using someone else's system or login to participate in the communication session. The methods disclosed herein may use voice processing technology and one or more voice identification databases (VIDs) to identify the active speaker in a communication session. In some embodiments, the disclosed methods may display the identity of the active speaker to other users or participants in the same communication session. In other embodiments, the disclosed methods may dynamically switch between user profiles or accounts during the communication session based on the identity of the active speaker.Type: GrantFiled: August 28, 2018Date of Patent: March 29, 2022Assignee: Dell Products L.P.Inventors: Douglas J. Peeler, Srinivas Kamepalli
-
Patent number: 11289098Abstract: A method with registration includes: receiving a speech signal of a speaker; synthesizing the received speech signal and a noise signal to generate a synthesized signal; generating a feature vector based on the synthesized signal; and constructing a registration database (DB) corresponding to the speaker based on the generated feature vector.Type: GrantFiled: February 4, 2020Date of Patent: March 29, 2022Assignee: Samsung Electronics Co., Ltd.Inventors: Sung-Un Park, Kyuhong Kim
-
Patent number: 11270074Abstract: Implemented are an apparatus and a method that enable highly accurate intent estimation of a user utterance. An utterance learning adaptive processing unit analyzes a plurality of user utterances input from a user, generates learning data in which entity information included in a user utterance with an unclear intent is associated with a correct intent, and stores the generated learning data is a storage unit. The utterance learning adaptive processing unit generates learning data in which an intent, acquired from a response utterance from the user to an apparatus utterance after input of a first user utterance with an unclear intent, is recorded in association with entity information included in the first user utterance. The learning data is recorded to include superordinate semantic concept information of the entity information. At the time of estimating an intent for a new user utterance, learning data with similar superordinate semantic concept information is used.Type: GrantFiled: October 26, 2018Date of Patent: March 8, 2022Assignee: SONY CORPORATIONInventors: Hiro Iwase, Shinichi Kawano, Yuhei Taki, Kunihito Sawai
-
Patent number: 11263403Abstract: A method includes identifying a set of identigens for each word of a first phrase of a phrase group to produce a first plurality of sets of identigens and determining whether first and second identigen rules are applicable to the first plurality of sets of identigens. When the first and second identigen rules are applicable to the first plurality of sets of identigens, the method further includes identifying a set of identigens for each word of a second phrase of the phrase group to produce a second plurality of sets of identigens and determining that the first identigen rules are applicable to the second plurality of sets of identigens. The method further includes identifying one valid identigen of each set of identigens of the first plurality of sets of identigens by applying the first identigen rules to the first plurality of sets of identigens to produce a first entigen group.Type: GrantFiled: September 17, 2019Date of Patent: March 1, 2022Assignee: entigenlogic LLCInventors: Frank John Williams, David Ralph Lazzara, Dennis Arlen Roberson, Stephen Emerson Sundberg, Stephen Chen, Karl Olaf Knutson, Jessy Thomas, David Michael Corns, II, Andrew Chu, Gary W. Grube
-
Patent number: 11255690Abstract: A plurality of abbreviated names are generated for evaluation based on a full name of a point-of-interest (POI) on a map. A plurality of address names comprising the full name of the POI or any of the abbreviated names to be evaluated are obtained from a predetermined area of the POI. A phrase status vector used to indicate a location status of the target phrase in each particular address name is calculated for a target phrase based on each address name, the target phrase including the full name of the POI or any of the abbreviated names to be evaluated. A similarity is calculated between a phrase status vector for the full name of the POI and a phrase status vector. A particular abbreviated name corresponding with a calculated similarity greater than a predetermined threshold is associated with the full name of the POI corresponding to the calculated similarity.Type: GrantFiled: October 26, 2020Date of Patent: February 22, 2022Assignee: Advanced New Technologies Co., Ltd.Inventors: Guanhai Zhong, Hui Li
-
Patent number: 11258867Abstract: A system and method for detecting providing a multi-region software as a service model in a cloud computing system, including for example, obtaining a list of allowed regions that can host data of a user from a source region; and automatically assigning a device associated with a new user from the source region to a computing system located in one of the allowed regions. The source region may be determined based on location information obtained from the device during registration.Type: GrantFiled: January 14, 2021Date of Patent: February 22, 2022Assignee: Softimize Ltd.Inventors: Guy Vinograd, Maor Last, Yoav Ofan, Ben Rosencveig, Olga Shpigel, David Ronen
-
Patent number: 11243613Abstract: A method includes monitoring a plurality of system inputs, and detecting a behavioral pattern performed by a user and associated with the plurality of system inputs, When the behavioral pattern is detected, the method includes associating, in a memory, a gesture with at least one action, the at least one action being determined by the plurality of system inputs, and, upon detecting the gesture, executing the action associated with the gesture.Type: GrantFiled: July 29, 2019Date of Patent: February 8, 2022Assignees: HONDA MOTOR CO., LTD., EDGE 3 TECHNOLOGIES LLCInventors: Stuart Masakazu Yamamoto, Tarek A. El Dokor
-
Patent number: 11238877Abstract: Proposed are a generative adversarial network-based speech bandwidth extender and extension method. A generative adversarial network-based speech bandwidth extension method, according to an embodiment, comprises the steps of: extracting feature vectors from a narrowband (NB) signal and a wideband (WB) signal of a speech; estimating the feature vector of the wideband signal from the feature vector of the narrowband signal; and learning a deep neural network classification model for discriminating the estimated feature vector of the wideband signal from the actually extracted feature vector of the wideband signal and the actually extracted feature vector of the narrowband signal.Type: GrantFiled: May 17, 2018Date of Patent: February 1, 2022Assignee: IUCF-HYU (INDUSTRY-UNIVERSITY COOPERATION FOUNDATION HANYANG UNIVERSITY)Inventors: Joon-Hyuk Chang, Kyoungjin Noh
-
Patent number: 11232786Abstract: Systems and methods to improve the performance of an automatic speech recognition (ASR) system using a confusion index indicative of the amount of confusion between words are described, where a confusion index (CI) or score is calculated by receiving a first word (Word1) and a second word (Word2), calculating an acoustic score (A12) indicative of the phonetic difference between Word1 and Word2, calculating a weighted language score (W(U1+U2), indicative of a weighted likelihood (or word frequency) of Word1 and Word2 occurring in the corpus, the confusion index CI incorporating both the acoustic score and the weighted language score, such that the CI for words that sound alike and have a high likelihood of occurring in the corpus will be higher than the CI for words that sound alike and do not have a high likelihood of occurring in the corpus.Type: GrantFiled: November 27, 2019Date of Patent: January 25, 2022Assignee: DISNEY ENTERPRISES, INC.Inventors: Sanchita Tiwari, Chang Shu
-
Patent number: 11232780Abstract: A method for predicting parametric vocoder parameter includes receiving a text utterance having one or more words, each word having one or more syllables, and each syllable having one or more phonemes. The method also includes receiving, as input to a vocoder model, prosodic features that represent an intended prosody for the text utterance and a linguistic specification. The prosodic features include a duration, pitch contour, and energy contour for the text utterance, while the linguistic specification includes sentence-level linguistic features, word-level linguistic features for each word, syllable-level linguistic features for each syllable, and phoneme-level linguistic features for each phoneme. The method also includes predicting vocoder parameters based on the prosodic features and the linguistic specification.Type: GrantFiled: September 26, 2020Date of Patent: January 25, 2022Assignee: Google LLCInventors: Rakesh Iyer, Vincent Wan
-
Patent number: 11227580Abstract: The present invention provides a device for estimating the deterioration factor of speech recognition accuracy by estimating an acoustic factor that leads to a speech recognition error. The device extracts an acoustic feature amount for each frame from an input speech, calculates a posterior probability for each acoustic event for the acoustic feature amount for each frame, corrects the posterior probability by filtering the posterior probability for each acoustic event using a time-series filter with weighting coefficients developed in the time axis, outputs a set of speech recognition results with a recognition score, outputs a feature amount for the speech recognition results for each frame, calculates and outputs a principal deterioration factor class for the speech recognition accuracy for each frame on the basis of the corrected posterior probability, the feature amount for speech recognition results for each frame, and the acoustic feature amount for each frame.Type: GrantFiled: February 6, 2019Date of Patent: January 18, 2022Assignee: NIPPON TELEGRAPH AND TELEPHONE CORPORATIONInventors: Takashi Nakamura, Takaaki Fukutomi
-
Patent number: 11217227Abstract: System and method for generating disambiguated terms in automatically generated transcriptions including instructions within a knowledge domain and employing the system are disclosed.Type: GrantFiled: November 8, 2019Date of Patent: January 4, 2022Assignee: Suki AI, Inc.Inventor: Ahmad Badary
-
Patent number: 11211058Abstract: Described herein is a system for prompting a user for clarification when an automatic speech recognition (ASR) system encounters ambiguity with respect to the user's input. The feedback provided by the user is used to retrain machine-learning models and/or to generate new machine-learning models. Based on the type of ambiguity, the system may determine to retrain one or more ASR models that are widely used by the system or to generate/update one or more user-specific models that are used to process inputs from one or more particular users.Type: GrantFiled: September 20, 2019Date of Patent: December 28, 2021Assignee: Amazon Technologies, Inc.Inventors: Aaron Eakin, Angela Sun, Ankur Gandhe, Ariya Rastrow, Chenlei Guo, Xing Fan
-
Patent number: 11211045Abstract: Provided is an artificial intelligence apparatus for predicting a performance of a voice recognition model in a user environment including: a memory configured to store a performance prediction model; and a processor configured to: obtain first controlled environment data including first controlled environment factors corresponding to a first controlled voice recognition environment and a first controlled voice recognition performance of a target voice recognition model in the first controlled voice recognition environment; obtain first user environment factors corresponding to a first user environment, in which the performance is to be predicted; predict, using the performance prediction model, a first user voice recognition performance of the target voice recognition model in the first user voice recognition environment from the obtained first controlled environment data and the first user environment factors; and output the predicted first user voice recognition performance.Type: GrantFiled: May 29, 2019Date of Patent: December 28, 2021Assignee: LG ELECTRONICS INC.Inventor: Jonghoon Chae
-
Patent number: 11189302Abstract: A speech emotion detection system may obtain to-be-detected speech data. The system may generate speech frames based on framing processing and the to-be-detected speech data. The system may extract speech features corresponding to the speech frames to form a speech feature matrix corresponding to the to-be-detected speech data. The system may input the speech feature matrix to an emotion state probability detection model. The system may generate, based on the speech feature matrix and the emotion state probability detection model, an emotion state probability matrix corresponding to the to-be-detected speech data. The system may input the emotion state probability matrix and the speech feature matrix to an emotion state transition model. The system may generate an emotion state sequence based on the emotional state probability matrix, the speech feature matrix, and the emotional state transition model. The system may determine an emotion state based on the emotion state sequence.Type: GrantFiled: October 11, 2019Date of Patent: November 30, 2021Assignee: Tencent Technology (Shenzhen) Company LimitedInventor: Haibo Liu
-
Patent number: 11176321Abstract: Language models may be run with an input set of words in a given sentence. Each of the language models can predict a set of next candidate words to follow the input set or words. Based on the sets of next candidate words predicted by the language models and an actual next word, language guidance can be provided.Type: GrantFiled: May 2, 2019Date of Patent: November 16, 2021Assignee: International Business Machines CorporationInventors: Toby Kurien, Richard Allen Young, Maletsabisa Molapo, Ismail Yunus Akhalwaya
-
Patent number: 11176960Abstract: A system for distinguishing between a human voice generated command and an electronic speaker generated command is provided. An exemplary system comprises a microphone array for receiving an audio signal collection, preprocessing circuitry configured for converting the audio signal collection into processed recorded audio signals, energy balance metric determination circuitry configured for calculating a final energy balance metric based on the processed recorded audio signals, and energy balance metric evaluation circuitry for outputting a command originator signal based at least in part on the final energy balance metric.Type: GrantFiled: June 18, 2019Date of Patent: November 16, 2021Assignee: University of Florida Research Foundation, IncorporatedInventors: Patrick G. Traynor, Logan E. Blue, Luis Vargas
-
Patent number: 11165779Abstract: Aspects of the present invention disclose a method, computer program product, and system for modifying a prohibited list. The method includes one or more processors determining a trend in usage of a command based on an analysis of aggregated usage information of the command from a plurality of listening devices, wherein the determined trend indicates an amount of usage of the command over a period of time. The method further includes determining whether the determined trend in usage of the command deviates from a usage threshold for the command. The method further includes in response to determining that the determined trend in usage of the command does deviate from the usage threshold for the command, modifying a prohibited list based on an amount of deviation from the usage threshold usage of the command.Type: GrantFiled: November 29, 2018Date of Patent: November 2, 2021Assignee: International Business Machines CorporationInventors: Michael Bender, Jeremy R. Fox, Kulvir Singh Bhogal
-
Patent number: 11158303Abstract: In an approach to soft-forgetting training, one or more computer processors train a first model utilizing one or more training batches wherein each training batch of the one or more training batches comprises one or more blocks of information. The one or more computer processors, responsive to a completion of the training of the first model, initiate a training of a second model utilizing the one or more training batches. The one or more computer processors jitter a random block size for each block of information for each of the one or more training batches for the second model. The one or more computer processors unroll the second model over one or more non-overlapping contiguous jittered blocks of information. The one or more computer processors, responsive to the unrolling of the second model, reduce overfitting for the second model by applying twin regularization.Type: GrantFiled: August 27, 2019Date of Patent: October 26, 2021Assignee: International Business Machines CorporationInventors: Kartik Audhkhasi, George Andrei Saon, Zoltan Tueske, Brian E. D. Kingsbury, Michael Alan Picheny
-
Patent number: 11158305Abstract: Generally discussed herein are devices, systems, and methods for wake word verification. A method can include receiving, at a server, a message from a device indicating that an utterance of a user-defined wake word was detected at the device, the message including (a) audio samples or features extracted from the audio samples and (b) data indicating the user-defined wake word, retrieving or generating, at the server, a custom decoding graph for the user-defined wake word, wherein the decoding graph and the static portion of the wake word verification model form a custom wake word verification model for the user-defined wake word, executing the wake word verification model to determine a likelihood that the wake word was uttered, and providing a message to the device indicating whether wake was uttered based on the determined likelihood.Type: GrantFiled: July 25, 2019Date of Patent: October 26, 2021Assignee: Microsoft Technology Licensing, LLCInventors: Khuram Shahid, Kshitiz Kumar, Teng Yi, Veljko Miljanic, Huaming Wang, Yifan Gong, Hosam Adel Khalil
-
Patent number: 11151984Abstract: The invention discloses a multi-language mixed speech recognition method, which belongs to the technical field of speech recognition; the method comprises: step S1, configuring a multi-language mixed dictionary including a plurality of different languages; step S2, performing training according to the multi-language mixed dictionary and multi-language speech data including a plurality of different languages to form an acoustic recognition model; step S3, performing training according to multi-language text corpus including a plurality of different languages to form a language recognition model; step S4, forming the speech recognition system by using the multi-language mixed dictionary, the acoustic recognition model and the language recognition model; and subsequently, recognizing mixed speech by using the speech recognition system, and outputting a corresponding recognition result.Type: GrantFiled: January 26, 2018Date of Patent: October 19, 2021Assignee: YUTOU TECHNOLOGY (HANGZHOU) CO., LTD.Inventors: Fan Lichun, Meng Meng, Peng Gao
-
Patent number: 11133022Abstract: A method may include dividing input audio into frames and calculating a characteristic value for each of the frames. The method may include establishing a voting matrix having a first dimension representing a quantity of segments of sample audio and a second dimension representing a quantity of frames of each segment. The method may include marking voting labels in the voting matrix corresponding to frames of the sample audio when the characteristic values of corresponding frames of the input audio and sample audio match. The method may include determining a frame to be a recognition result when a sum of the voting labels at a corresponding position is higher than a threshold.Type: GrantFiled: January 6, 2021Date of Patent: September 28, 2021Assignee: ADVANCED NEW TECHNOLOGIES CO., LTD.Inventors: Zhijun Du, Nan Wang
-
Patent number: 11107475Abstract: An exemplary automatic speech recognition (ASR) system may receive an audio input including a segment of speech. The segment of speech may be independently processed by general ASR and domain-specific ASR to generate multiple ASR results. A selection between the multiple ASR results may be performed based on respective confidence levels for the general ASR and domain-specific ASR. As incremental ASR is performed, a composite result may be generated based on general ASR and domain-specific ASR.Type: GrantFiled: May 9, 2019Date of Patent: August 31, 2021Assignee: Rovi Guides, Inc.Inventor: Jeffry Copps Robert Jose
-
Patent number: 11100291Abstract: A query-processing server that interprets natural language expressions supports the extension of a first semantic grammar (for a particular type of expression), which is declared extensible, by a second semantic grammar (for another type of expression). When an extension is requested, the query-processing server checks that the two semantic grammars have compatible semantic types. The developers need not have any knowledge of each other, or about their respective grammars. Performing an extension may be done by yet another party, such as the query-processing server, or another server, independently of all previous parties. The use of semantic grammar extensions provides a way to expand the coverage and functionality of natural language interpretation in a simple and flexible manner, so that new forms of expression may be supported, and seamlessly combined with pre-existing interpretations. Finally, in some implementations, this is done without loss of efficiency.Type: GrantFiled: July 8, 2019Date of Patent: August 24, 2021Assignee: SOUNDHOUND, INC.Inventors: Keyvan Mohajer, Christopher S. Wilson, Bernard Mont-Reynaud
-
Patent number: 11069342Abstract: A method for training a voice data set is provided. A first test set of data selected from a first voice data set, and a first voice model parameter obtained by performing first voice model training based on a first voice data set, are obtained. Data from a second voice data set is randomly selected to generate a second test set. Further, second voice model training is performed based on the second voice data set and the first voice model parameter when the second test set and the first test set satisfy a similarity condition.Type: GrantFiled: June 10, 2019Date of Patent: July 20, 2021Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITEDInventors: Tao Sun, Yueteng Kang, Xiaoming Zhang, Li Zhang
-
Patent number: 11068668Abstract: The disclosed computer-implemented method for performing natural language translation in AR may include accessing an audio input stream that includes words spoken by a speaking user in a first language. The method may next include performing active noise cancellation on the words in the audio input stream so that the spoken words are suppressed before reaching a listening user. Still further, the method may include processing the audio input stream to identify the words spoken by the speaking user, and translating the identified words spoken by the speaking user into a second, different language. The method may also include generating spoken words in the second, different language using the translated words, and replaying the generated spoken words in the second language to the listening user. Various other methods, systems, and computer-readable media are also disclosed.Type: GrantFiled: October 25, 2018Date of Patent: July 20, 2021Assignee: Facebook Technologies, LLCInventors: Andrew Lovitt, Antonio John Miller, Philip Robinson, Scott Selfon
-
Patent number: 11062705Abstract: According to one embodiment, an information processing apparatus includes one or more processors configured to detect a trigger from a voice signal, the trigger indicating start of voice recognition; and to perform voice recognition of a recognition sound section subsequent to a trigger sound section including the detected trigger, referring to a trigger and voice recognition dictionary corresponding to the trigger.Type: GrantFiled: February 27, 2019Date of Patent: July 13, 2021Assignee: Kabushiki Kaisha ToshibaInventors: Nayuko Watanabe, Takehiko Kagoshima, Hiroshi Fujimura
-
Patent number: 11055745Abstract: Techniques for linguistic personalization of messages for targeted campaigns are described. In one or more implementations, dependencies between keywords and modifiers are extracted, from one or more segment-specific texts and a product-specific text, to build language models for the one or more segment specific texts and the product specific text. Modifiers with a desired sentiment are extracted from the product specific text and transformation points are identified in a message skeleton. Then one or more of the extracted modifiers are inserted to modify one or more identified keywords in the message skeleton to create a personalized message for a target segment of the targeted marketing campaign.Type: GrantFiled: December 10, 2014Date of Patent: July 6, 2021Assignee: Adobe Inc.Inventors: Rishiraj Saha Roy, J. Guna Prasaad, Aishwarya Padmakumar, Ponnurangam Kumaraguru
-
Patent number: 11048869Abstract: Methods and systems for a transportation vehicle are provided. One method includes receiving a user input for a valid communication session by a processor executable, digital assistant at a device on a transportation vehicle; tagging by the digital assistant, the user input words with a grammatical connotation; generating an action context, a filter context and a response context by a neural network, based on the tagged user input; storing by the digital assistant, a key-value pair for a parameter of the filter context at a short term memory, based on an output from the neural network; updating by the digital assistant, the key-value pair at the short term memory after receiving a reply to a follow-up request and another output from the trained neural network; and providing a response to the reply by the digital assistant.Type: GrantFiled: December 27, 2019Date of Patent: June 29, 2021Assignee: Panasonic Avionics CorporationInventors: Rawad Hilal, Gurmukh Khabrani, Chin Perng
-
Patent number: 11043211Abstract: A speech recognition method includes obtaining captured voice information, and determining semantic information of the captured voice information; segmenting the captured voice information to obtain voice segments when the semantic information does not satisfy a preset rule, and extracting voiceprint information of the voice segments; obtaining an unmatched voiceprint information from a local voiceprint database; matching the voiceprint information of the voice segments with the unmatched voiceprint information to determine a set of filtered voice segments having voiceprint information that successfully matches the unmatched voiceprint information; combining the set of filtered voice segments to obtain combined voice information, and determining combined semantic information of the combined voice information; and using the combined semantic information as a speech recognition result when the combined semantic information satisfies the preset rule.Type: GrantFiled: June 14, 2019Date of Patent: June 22, 2021Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITEDInventor: Qiusheng Wan
-
Patent number: 11031108Abstract: A medicine management method includes acquiring user information of medicine usage corresponding to at least one assigned medicine; acquiring user information of medicine using reactions; and displaying the user information of medicine usage with first time information corresponding to different time points of using the at least one assigned medicine and the user information of medicine using reactions with second time information corresponding to different time points of occurring medicine using reactions concurrently.Type: GrantFiled: February 7, 2017Date of Patent: June 8, 2021Assignee: HTC CorporationInventors: Tsung-Hsiang Liu, Ya-Han Yang, Hao-Ting Chang, Chih-Wei Cheng, Ting-Jung Chang
-
Patent number: 11031029Abstract: A pitch detection method. Such a pitch detection method may have M-PWVT-TEO algorithm to detect a pitch value from a speech signal, apply a partial auto-correlation to a current signal with the pitch value to compensate the delay of the pitch value. Also, the pitch detection method may apply a full auto-correlation to the speech signal where the pitch value is not detected to recover on-sets of the speech signal.Type: GrantFiled: August 30, 2019Date of Patent: June 8, 2021Assignee: OmniSpeech LLCInventor: Vahid Khanagha
-
Patent number: 11017022Abstract: Methods and systems are disclosed in which audio broadcasts are converted into audio segments, for example, based on segment content. These audio segments are indexed, so as to be searchable, as computer searchable segments, for example, by network search engines and other computerized search tools.Type: GrantFiled: January 29, 2017Date of Patent: May 25, 2021Assignee: SubPLY Solutions Ltd.Inventors: Gal Klein, Rachel Ludmer
-
Patent number: 11017781Abstract: Techniques are provided for reverberation compensation for far-field speaker recognition. A methodology implementing the techniques according to an embodiment includes receiving an authentication audio signal associated with speech of a user and extracting features from the authentication audio signal. The method also includes scoring results of application of one or more speaker models to the extracted features. Each of the speaker models is trained based on a training audio signal processed by a reverberation simulator to simulate selected far-field environmental effects to be associated with that speaker model. The method further includes selecting one of the speaker models, based on the score, and mapping the selected speaker model to a known speaker identification or label that is associated with the user.Type: GrantFiled: October 6, 2018Date of Patent: May 25, 2021Assignee: INTEL CORPORATIONInventors: Gokcen Cilingir, Narayan Biswal
-
Patent number: 11010820Abstract: Systems, methods, and apparatus are disclosed for generating and processing natural language requests. A request processing system processes a received natural language request to identify an intent of the natural language request and a confidence level of the identified intent. In response to the confidence level of the identified intent not satisfying a threshold level, the request processing system sends the natural language request to the fulfillment computing device for further processing by a person associated with the fulfillment computing device. In response to the confidence level satisfying the threshold level, the request processing system proceeds with fulfilling the request per the identified intent.Type: GrantFiled: May 5, 2016Date of Patent: May 18, 2021Assignee: TRANSFORM SR BRANDS LLCInventors: Edward Lampert, Eui Chung, Bharath Sridharan
-
Patent number: 10978053Abstract: A system determines user intent from a received conversation element. A plurality of distinct intent labels are generated for the received conversation element. The generated plurality of distinct intent labels are divided into a plurality of interpretation partitions with overlapping semantic content. for each interpretation partition of the plurality of interpretation partitions, a set of maximal coherent subgroups are defined that do not disagree on labels for terms in each subgroup, a score is computed for each maximal coherent subgroup of the defined set of maximal coherent subgroups, and a maximal coherent subgroup is selected from the set of maximal coherent subgroups based on the computed score. Intent labels are aggregated from the selected maximal coherent subgroup of each interpretation partition of the plurality of interpretation partitions to define a multiple intent interpretation of the received conversation element.Type: GrantFiled: October 13, 2020Date of Patent: April 13, 2021Assignee: SAS Institute Inc.Inventors: Jared Michael Dean Smythe, Richard Welland Crowell
-
Patent number: 10964310Abstract: A method of updating speech recognition data including a language model used for speech recognition, the method including obtaining language data including at least one word; detecting a word that does not exist in the language model from among the at least one word; obtaining at least one phoneme sequence regarding the detected word; obtaining components constituting the at least one phoneme sequence by dividing the at least one phoneme sequence into predetermined unit components; determining information regarding probabilities that the respective components constituting each of the at least one phoneme sequence appear during speech recognition; and updating the language model based on the determined probability information.Type: GrantFiled: March 16, 2020Date of Patent: March 30, 2021Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventors: Chi-youn Park, Il-hwan Kim, Kyung-min Lee, Nam-hoon Kim, Jae-won Lee
-
Patent number: 10950229Abstract: A voice command application allows a user to configure an infotainment system to respond to customized voice commands. The voice command application exposes a library of functions to the user which the infotainment system can execute via interaction with the vehicle. The voice command application receives a selection of one or more functions and then receives a speech sample of the voice command. The voice command application generates sample metadata that includes linguistic elements of the voice command, and then generates a command specification. The command specification indicates the selected functions and the sample metadata for storage in a database. Subsequently, the voice command application receives the voice command from the user and locates the associated command specification in the database. The voice command application then extracts the associated set of functions and causes the vehicle to execute those functions to perform vehicle operations.Type: GrantFiled: August 25, 2017Date of Patent: March 16, 2021Assignee: Harman International Industries, IncorporatedInventors: Rajesh Biswal, Arindam Dasgupta
-
Patent number: 10938389Abstract: A method for controlling operation of a power switch includes obtaining, by one or more processors of a power switch, data indicative of one or more non-contact gestures. The method includes determining, by the one or more processors, a control action based at least in part on the data indicative of the one or more non-contact gestures. The method includes implementing, by the one or more processors, the control action.Type: GrantFiled: December 19, 2018Date of Patent: March 2, 2021Assignee: Hubbell IncorporatedInventors: Shawn Monteith, Michael Tetreault, Daniel Gould, Nicholas Kraus
-
Patent number: 10922488Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for computing numeric representations of words. One of the methods includes obtaining a set of training data, wherein the set of training data comprises sequences of words; training a classifier and an embedding function on the set of training data, wherein training the embedding function comprises obtained trained values of the embedding function parameters; processing each word in the vocabulary using the embedding function in accordance with the trained values of the embedding function parameters to generate a respective numerical representation of each word in the vocabulary in the high-dimensional space; and associating each word in the vocabulary with the respective numeric representation of the word in the high-dimensional space.Type: GrantFiled: March 25, 2019Date of Patent: February 16, 2021Assignee: Google LLCInventors: Tomas Mikolov, Kai Chen, Gregory S. Corrado, Jeffrey A. Dean