Patents Examined by L. Thomas
  • Patent number: 10950229
    Abstract: A voice command application allows a user to configure an infotainment system to respond to customized voice commands. The voice command application exposes a library of functions to the user which the infotainment system can execute via interaction with the vehicle. The voice command application receives a selection of one or more functions and then receives a speech sample of the voice command. The voice command application generates sample metadata that includes linguistic elements of the voice command, and then generates a command specification. The command specification indicates the selected functions and the sample metadata for storage in a database. Subsequently, the voice command application receives the voice command from the user and locates the associated command specification in the database. The voice command application then extracts the associated set of functions and causes the vehicle to execute those functions to perform vehicle operations.
    Type: Grant
    Filed: August 25, 2017
    Date of Patent: March 16, 2021
    Assignee: Harman International Industries, Incorporated
    Inventors: Rajesh Biswal, Arindam Dasgupta
  • Patent number: 10943600
    Abstract: A system or method for manipulating audiovisual data using transcript information. The system or method performs the following actions. Creating a computer-generated transcript of audio data from the audiovisual data, the computer-generated transcript includes a plurality of words, at least some words of the plurality of words are associated with a respective timestamp and a confidence score. Receiving a traditional transcript of the audio data, the traditional transcript includes a plurality of words that are not associated with timestamps. Identifying one or more words from the plurality of words of the computer-generated transcript that match words from the plurality of words of the traditional transcript. Associating the timestamp of the one or more words of the computer-generated transcript with the matching word of the traditional transcript. Processing the audiovisual data using the traditional transcript and the associated timestamps.
    Type: Grant
    Filed: February 28, 2019
    Date of Patent: March 9, 2021
    Assignee: Axon Enterprise, Inc.
    Inventors: Joseph Charles Dimino, Jr., Sayce William Falk, Leo Thomas Rossignac-Milon
  • Patent number: 10943068
    Abstract: A computing system is provided. The computing system includes a processor configured to execute one or more programs and associated memory. The processor is configured to execute neural network system that includes a first neural network and a second neural network. The processor is configured to receive input text, and for each of a plurality of text spans within the input text: identify a vector of semantic entities and a vector of entity mentions; define an n-ary relation between entity mentions including subrelations; and determine mention-level representation vectors in the text spans that satisfy the n-ary relation or subrelations. The processor is configured to: aggregate the mention-level representation vectors over all of the text spans to produce entity-level representation vectors; input to the second neural network the entity-level representation vectors; and output a prediction of a presence of the n-ary relation for the semantic entities in the input text.
    Type: Grant
    Filed: March 29, 2019
    Date of Patent: March 9, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Hoifung Poon, Cliff Wong, Robin Jia
  • Patent number: 10909986
    Abstract: Methods, computer program products, and systems are presented. The methods include, for instance: obtaining an input text for an output speech. The number of words and syllables are counted in each sentence, and a mean sentence length of the input text is calculated. Each sentence length is checked against the mean sentence length and a variation for each sentence is calculated. For the input text, the consumability-readability score is produced as an average of variations for all sentences in the input text. The consumability-readability score indicates the level of satisfaction for the listener of the output speech based on the input text.
    Type: Grant
    Filed: November 29, 2018
    Date of Patent: February 2, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Craig M. Trim, John M. Ganci, Jr., Anna Chaney, Stefan Van Der Stockt
  • Patent number: 10885286
    Abstract: A computer-implemented method that may be carried out by a reader device or other computing device includes identifying boundaries of words and sentences of displayed content and receiving a language selection of a language for translating a selected unit of text in the displayed content; translating the unit of text to the selected language; retrieving from a picture dictionary, a pictogram associated with a tag applied to the word or at least one word from the sentence of the selected unit of text; generating pronunciation files for both an original language of the unit of text and the selected language; and displaying, in context with the displayed content, the unit of text in the original language and the selected language, the pictogram, and icons for requesting to play the pronunciation files. The selected unit of text can be a word or a sentence.
    Type: Grant
    Filed: October 12, 2018
    Date of Patent: January 5, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Paul Ronald Ray, Michael Tholfsen, Guillaume Simonnet
  • Patent number: 10878199
    Abstract: A word vector processing method is provided. Word segmentation is performed on a corpus to obtain words, and n-gram strokes corresponding to the words are determined. Each n-gram stroke represents n successive strokes of a corresponding word. Word vectors of the words and stroke vectors of the n-gram strokes are initialized corresponding to the words. After performing the word segmentation, the n-gram strokes are determined, and the word vectors and stroke vectors are determined, training the word vectors and the stroke vectors.
    Type: Grant
    Filed: September 30, 2019
    Date of Patent: December 29, 2020
    Assignee: Advanced New Technologies Co., Ltd.
    Inventors: Shaosheng Cao, Xiaolong Li
  • Patent number: 10867134
    Abstract: A multicore CPU of a text string data analyzing device: loads a plurality of blocks obtained by dividing a text string dictionary into a memory; executes, in parallel on block groups executable independently of each other, an entry registration process of registering, character by character, unregistered text strings of text string data as new entries in the blocks in order from last characters; and outputs, as BW transformed data of the text string dictionary in which the text string data is already registered, a text string obtained by coupling text strings registered in entries of the blocks in a state in which no unregistered text strings of the blocks exists.
    Type: Grant
    Filed: September 2, 2016
    Date of Patent: December 15, 2020
    Assignee: HITACHI HIGH-TECH CORPORATION
    Inventor: Kouichi Kimura
  • Patent number: 10824816
    Abstract: A semantic parsing method and a semantic parsing apparatus. The method includes: receiving semantic information; parsing the semantic information to obtain a primary text feature; calculating a similarity degree between the primary text feature and a linguistic data feature; using a linguistic data to which a linguistic data feature most similar to the primary text feature belongs, as a primary parsing text; extracting a feature of the primary parsing text to obtain a secondary text feature; obtaining a service feature matching the secondary text feature; obtaining a weight value associated with the obtained service feature; using a service field to which a service feature having a greatest weight value belongs, as the service field to which the primary parsing text belongs; and parsing the primary parsing text in the service field to which the primary parsing text belongs, to obtain a secondary parsing text.
    Type: Grant
    Filed: April 20, 2017
    Date of Patent: November 3, 2020
    Assignees: Beijing Jingdong Shangke Information Technology Co., Ltd., Beijing Jingdong Century Trading Co., Ltd.
    Inventors: Xi Zang, Xin Zhou
  • Patent number: 10825450
    Abstract: An electronic computing device for providing a response to an audio query where the response is determined to have a public safety impact. The electronic computing device includes a microphone and an electronic processor. The electronic processor is configured to receive audio data including a speech segment and determine a plurality of possible meanings of the speech segment. Each possible meaning is associated with a probability that the possible meaning is a correct meaning for the speech segment and a first possible meaning is associated with a first probability that is higher than a second probability associated with a second possible meaning. The electronic processor is also configured to determine public safety impact context information of the second possible meaning. The electronic processor is further configured to output a second response associated with the second possible meaning without first outputting a first response associated with the first possible meaning.
    Type: Grant
    Filed: October 25, 2018
    Date of Patent: November 3, 2020
    Assignee: MOTOROLA SOLUTIONS, INC.
    Inventors: Haim Kahlon, Alexander Aperstein, David Lev, Tamar Mordel
  • Patent number: 10817665
    Abstract: Systems and methods are provided for detecting inaccuracy in a product title, comprising identifying, by running a string algorithm on a title associated with a product, at least one product type associated with the product, predicting, using a machine learning algorithm, at least one product type associated with the product based on the title, detecting an inaccuracy in the title, based on at least one of the identification or the prediction, and outputting, to a remote device, a message indicating that the title comprises the inaccuracy. Running the string algorithm may comprise receiving a set of strings, generating a tree based on the received set of strings, receiving the title, and traversing the generated tree using the title to find a match. Using the machine learning algorithm may comprise identifying words in the title, learning a vector representation for each character n-gram of each word, and summing each character n-gram.
    Type: Grant
    Filed: May 8, 2020
    Date of Patent: October 27, 2020
    Assignee: Coupang Corp.
    Inventors: Shusi Yu, Jing Li
  • Patent number: 10818292
    Abstract: A method for term-dependent output of information based on a voice input to a specific group includes the steps of: capturing the voice input; analyzing the captured voice input for the presence of a group-specific key term, associated with the specific group; and on detection of the group-specific key term in the analyzed voice input, outputting the information based on the voice input to the specific group.
    Type: Grant
    Filed: June 28, 2019
    Date of Patent: October 27, 2020
    Assignee: Andreas Stihl AG & Co. KG
    Inventors: Harald Mang, Michael Reinert, Martin Tippelt
  • Patent number: 10811020
    Abstract: A voice signal decoding device includes a first decoder, a second decoder, a signal switch, and a noise adder. The first decoder decodes first encoded data encoded by a first encoding method. The second decoder decodes second encoded data encoded by a second encoding method. The second encoded data has a narrower band than a band of the first encoded data. The signal switch switches an output signal of the first decoder and an output signal of the second decoder. The noise adder adds a noise signal to a high-frequency band in the output signal of the second decoder when the signal switch switches an output signal from the output signal of the first decoder to the output signal of the second decoder. The high-frequency band is a band where a signal component is lacking as compared with the output signal of the first decoder.
    Type: Grant
    Filed: May 24, 2016
    Date of Patent: October 20, 2020
    Assignee: PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD.
    Inventors: Hiroyuki Ehara, Takuya Kawashima
  • Patent number: 10783877
    Abstract: A system for categorizing words into clusters includes a receiver to receive a set of sentences formed by a plurality of words. The set of sentences is indicative of interaction of a user with a virtual assistant. A categorizer categorizes the plurality of words into a first set of clusters by using a first clustering technique, and categorizes the plurality of words into a second set of clusters by using a second clustering technique. A detector detects words that appear in similar clusters after categorization by the first clustering technique and the second clustering technique. Similarity of clusters is based on a nature of words forming the clusters. A generator generates a confidence score for each of the plurality of words based on the detection. The confidence score of a word is indicative of accuracy of the categorization of the word.
    Type: Grant
    Filed: July 24, 2018
    Date of Patent: September 22, 2020
    Assignee: ACCENTURE GLOBAL SOLUTIONS LIMITED
    Inventors: Anshul Solanki, Akanksha Juneja, Bibudh Lahiri, Anurag Tripathi, Sonam Gupta, Rinki Arya
  • Patent number: 10783329
    Abstract: The present disclosure relates to a method, device and computer readable storage medium for presenting an emotion. The method for presenting the emotion includes obtaining a first emotion presentation instruction, wherein the first emotion presentation instruction includes at least one first emotion presentation modality and at least one emotional style, and the at least one first emotion presentation modality includes a text emotion presentation modality; and presenting an emotion corresponding to one or more of the at least one emotional style according to each of the at least one first emotion presentation modality. The present disclosure can realize text-based multi-modal emotion presentation modes, and thus user experience can be improved.
    Type: Grant
    Filed: August 1, 2018
    Date of Patent: September 22, 2020
    Assignee: SHANGHAI XIAOI ROBOT TECHNOLOGY CO., LTD.
    Inventors: Hui Wang, Yuning Wang, Pinpin Zhu
  • Patent number: 10776588
    Abstract: A smartphone-based telephone translation system includes a translation machine that starts a call translation process when the call translation is set to ON. The system detects an incoming call or dials, translates a voice signal to obtain a translation text for the ON call state. The voice signal of the caller is received and recognized. A voice signal is translated to obtain a translation text for the ON call state. The localized voice signal is recognized, using a TTS machine to obtain a reply voice signal and sent to the caller. The system performs voice recognition and text translation on the call voice, translates the language into localized language, and performs the function of calling between different languages in real time, so as to realize a telephone communication between two persons speaking different languages. The system supports communication with a hearing-impaired person or deaf-mute person.
    Type: Grant
    Filed: July 23, 2018
    Date of Patent: September 15, 2020
    Assignee: SHENZHEN ZHIYUAN TECHNOLOGY CO., LTD.
    Inventors: Liang Yu, Chuan Wang
  • Patent number: 10777207
    Abstract: The present disclosure discloses a method and apparatus for verifying information. The method includes: acquiring a target combination of question and answer, the target combination of question and answer comprising a question and an answer to the question; sending the question in the target combination of question and answer to a terminal device; receiving to-be-confirmed voice information sent by the terminal device, and recognizing the to-be-confirmed voice information to generate a to-be-confirmed answer; and determining that the to-be-confirmed voice information is verified, if the to-be-confirmed answer matches the answer in the target combination of question and answer.
    Type: Grant
    Filed: August 20, 2018
    Date of Patent: September 15, 2020
    Assignee: Baidu Online Network Technology (Beijing) Co., Ltd.
    Inventors: Yang Mu, Wenyu Wang, Lan Li
  • Patent number: 10755052
    Abstract: The present disclosure provides a semantic disambiguation method, a semantic disambiguation device, a server and a storage medium. The method includes the following. Semantic analysis is performed on a current dialog message input by a user. It is determined, based on an analysis result, whether a current service field corresponding to the current dialog message is an ambiguous field. When the current service field is the ambiguous field, disambiguation is performed on the current service field based on one or more predetermined historical service fields corresponding to one or more historical dialog messages input by the user.
    Type: Grant
    Filed: July 24, 2018
    Date of Patent: August 25, 2020
    Assignee: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD.
    Inventors: Liping Deng, Peixuan Shi
  • Patent number: 10699715
    Abstract: An artificial neural network architecture is provided for processing raw audio waveforms to create speaker representations that are used for text-independent speaker verification and recognition. The artificial neural network architecture includes a strided convolution layer, first and second sequentially connected residual blocks, a transformer layer, and a final fully connected (FC) layer. The strided convolution layer is configured to receive raw audio waveforms from a speaker. The first and the second residual blocks both include multiple convolutional and max pooling layers. The transformer layer is configured to aggregate frame level embeddings to an utterance level embedding. The output of the FC layer creates a speaker representation for the speaker whose raw audio waveforms were inputted into the strided convolution layer.
    Type: Grant
    Filed: January 27, 2020
    Date of Patent: June 30, 2020
    Assignee: Alphonso Inc.
    Inventors: Aashiq Muhamed, Susmita Ghose
  • Patent number: 10573046
    Abstract: The present invention makes it possible to see an original text string which (i) is contained in a captured image and (ii) has been translated, even after a translation of the original text string is displayed. The text string decoration display control section (12) decorates a part indicating a text string contained in a captured image and causes the decorated part to be displayed. The translation image generating section (13) generates a translation image showing a result of translating the text string into another language. The translation display control section (16) switches between display and non-display of the translation image in accordance with an input carried out by a user.
    Type: Grant
    Filed: June 4, 2018
    Date of Patent: February 25, 2020
    Assignee: SHARP KABUSHIKI KAISHA
    Inventors: Kiyofumi Ohtsuka, Tadao Nagasawa
  • Patent number: 10482883
    Abstract: A voice to text model used by a voice-enabled electronic device is dynamically and in a context-sensitive manner updated to facilitate recognition of entities that potentially may be spoken by a user in a voice input directed to the voice-enabled electronic device. The dynamic update to the voice to text model may be performed, for example, based upon processing of a first portion of a voice input, e.g., based upon detection of a particular type of voice action, and may be targeted to facilitate the recognition of entities that may occur in a later portion of the same voice input, e.g., entities that are particularly relevant to one or more parameters associated with a detected type of voice action.
    Type: Grant
    Filed: May 2, 2018
    Date of Patent: November 19, 2019
    Assignee: GOOGLE LLC
    Inventors: Yuli Gao, Sangsoo Sung, Prathab Murugesan