Word Recognition Patents (Class 704/251)
  • Patent number: 10722185
    Abstract: A user, such as an elderly person, may be assisted by an assistance device in his or her caregiving environment that operates in conjunction with one or more server computers. The assistance device may execute a schedule of assistance actions where each assistance action is associated with a time and is executed at that time to assist the user. An assistance action may present an input request to a user, process a voice input of the user, and analyze the voice input to determine that the voice input corresponds to a negative response event, a positive response event, or a non-response event. Based on the categorization of one or more voice inputs as negative response events, positive response events, or non-response events, it may be determined to notify a caregiver of the user, for example where the user has not responded to a number of assistance actions.
    Type: Grant
    Filed: November 3, 2017
    Date of Patent: July 28, 2020
    Assignee: LifePod Solutions, Inc.
    Inventor: Dennis Fountaine
  • Patent number: 10715713
    Abstract: A processing device causes a selection of a multimedia module to be detected. The processing device renders, during a first mode of operation, a dynamic navigation flow control in association with first multimedia content of the multimedia module, wherein the rendered dynamic navigation flow control indicates a first navigation position. The processing device stores a first result to memory obtained by execution of an interactive event during the first mode of operation. Based at least on the first result, a second mode of operation is entered. During the second mode of operation, the dynamic navigation flow control is re-rendered in association with second multimedia content of the multimedia module, wherein the re-rendered dynamic navigation flow control indicates a second navigation position.
    Type: Grant
    Filed: April 24, 2019
    Date of Patent: July 14, 2020
    Assignee: Breakthrough PerformanceTech, LLC
    Inventors: John DiGiantomasso, Martin L. Cohen
  • Patent number: 10714086
    Abstract: Some implementations are directed to selective invocation of a particular third-party (3P) agent by an automated assistant to achieve an intended action determined by the automated assistant during a dynamic dialog between the automated assistant and a user. In some of those implementations, the particular 3P agent is invoked with value(s) for parameter(s) that are determined during the dynamic dialog; and/or the particular 3P agent is selected, from a plurality of candidate 3P agents, for invocation based on the determined value(s) for the parameter(s) and/or based on other criteria. In some of those implementations, the automated assistant invokes the particular 3P agent by transmitting, to the particular 3P agent, a 3P invocation request that includes the determined value(s) for the parameter(s).
    Type: Grant
    Filed: January 10, 2019
    Date of Patent: July 14, 2020
    Assignee: GOOGLE LLC
    Inventors: Ulas Kirazci, Bo Wang, Steve Chen, Sunil Vemuri, Barnaby James, Valerie Nygaard
  • Patent number: 10713329
    Abstract: In some examples, a system performs language processing of text of an information page to determine whether the text refers to an online resource, and in response to determining that the text refers to the online resource, identify the text as an implicit reference to the online resource. The system derives a link to the online resource based on the implicit reference, the derived link useable in accessing the online resource.
    Type: Grant
    Filed: October 30, 2018
    Date of Patent: July 14, 2020
    Assignee: LONGSAND LIMITED
    Inventors: Brian Cowe, Sean Blanchflower
  • Patent number: 10706086
    Abstract: Techniques for simulating a user in a conversation are described. A user simulation service and a conversation agent service conduct a dialog. The user simulation service compares a current sequence of stored labels corresponding to statements in the dialog with a plurality of candidate sequences of labels corresponding to statements in a plurality of candidate dialogs to identify a matching sequence of labels. The user simulation sequence identifies a base sequence of labels that includes the matching sequence of labels to identify a label corresponding to an act in the base sequence of labels following the matching sequence of labels. The user simulation service issues the act to the conversation agent service to be used as a simulated user act.
    Type: Grant
    Filed: March 12, 2018
    Date of Patent: July 7, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Devin Didericksen, Jared Kane Kramer, Oleg Rokhlenko, Kevin Michael Small, Li Zhou
  • Patent number: 10699076
    Abstract: Embodiments of the specification disclose a risk address identification method and apparatus, and an electronic device. The risk address identification method includes: acquiring an address word sequence corresponding to an input address; determining an address word in the address word sequence, the determined address word matching a risk word corresponding to a risk address; generating an observation sequence corresponding to the address word sequence according to the determined address word; processing the observation sequence using a hidden Markov model obtained based on semantics learning before and after address words, to obtain a decision vector, wherein the decision vector represents probabilities of the risk address being matched by address words contained in the address word sequence; and identifying whether the input address is a risk address by making a classification decision on the decision vector.
    Type: Grant
    Filed: January 29, 2020
    Date of Patent: June 30, 2020
    Assignee: Alibaba Group Holding Limited
    Inventor: Qingqing Sun
  • Patent number: 10692496
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for suppressing hotwords are disclosed. In one aspect, a method includes the actions of receiving audio data corresponding to playback of an utterance. The actions further include providing the audio data as an input to a model (i) that is configured to determine whether a given audio data sample includes an audio watermark and (ii) that was trained using watermarked audio data samples that each include an audio watermark sample and non-watermarked audio data samples that do not each include an audio watermark sample. The actions further include receiving, from the model, data indicating whether the audio data includes the audio watermark. The actions further include, based on the data indicating whether the audio data includes the audio watermark, determining to continue or cease processing of the audio data.
    Type: Grant
    Filed: May 21, 2019
    Date of Patent: June 23, 2020
    Assignee: Google LLC
    Inventors: Alexander H. Gruenstein, Taral Pradeep Joglekar, Vijayaditya Peddinti, Michiel A. U. Bacchiani
  • Patent number: 10685648
    Abstract: Mechanisms are provided, in a smart speaker system having at least one smart speaker device comprising an audio capture device, and smart speaker system logic, for processing audio sample data captured by the audio capture device. The audio capture device captures an audio sample from a monitored environment and one or more sensor devices capture sensor data representing non-verbal attention indicators associated with a speaker of a speech portion of the audio sample. The smart speaker system logic evaluates the non-verbal attention indicators of the sensor data to determine whether or not the speech portion of the audio sample is directed to the smart speaker device. In response to determining that the speech portion of the audio sample is directed to the smart speaker device, a cognitive system associated with the smart speaker system generates a response to the speech portion.
    Type: Grant
    Filed: November 8, 2017
    Date of Patent: June 16, 2020
    Assignee: International Business Machines Corporation
    Inventors: John J. Andersen, Dogukan Erenel, Richard O. Lyle, Connie Yee
  • Patent number: 10672393
    Abstract: A system, apparatus, method, and computer program product for a speaking aid. The system including network interface circuitry to receive speech input from a user. The speech input includes a partial sentence with a missing word or the partial sentence with a stuttered word. The system also includes a processor coupled to the network interface circuitry and one or more memory devices coupled to the processor. The one or more memory devices include instructions, that when executed by the processor, cause the system to detect a stutter or pause in the speech input, predict the stuttered word or the missing word, present a predicted word from an n-best list to the user; and if a prompt is received from the user, present a next word from the n-best list until the user speaks a correct word to replace the stutter or the pause.
    Type: Grant
    Filed: January 12, 2018
    Date of Patent: June 2, 2020
    Assignee: Intel Corporation
    Inventors: Ze'ev Rivlin, Vered Bar Bracha, Douglas Gabel, Jonathan Huang, Sylvia Downing, Binuraj Ravindran, Willem Beltman
  • Patent number: 10672385
    Abstract: A method for implementing a speaker-independent speech recognition system with reduced latency is provided. The method includes capturing voice data at a carry-on-device from a user during a pre-flight check-in performed by the user for an upcoming flight; extracting features associated with the user from the captured voice data at the carry-on-device; uplinking the extracted features to the speaker-independent speech recognition system onboard the aircraft; and adapting the extracted features with an acoustic feature model of the speaker-independent speech recognition system.
    Type: Grant
    Filed: September 4, 2015
    Date of Patent: June 2, 2020
    Assignee: Honeywell International Inc.
    Inventors: Bharath Sundararajan, Vasantha Selvi Paulraj, Rengarajan Mohan
  • Patent number: 10674014
    Abstract: A system for providing help includes at least one digital assistant, a wireless data interface for connecting the digital assistants to a server, and battery backup. A plurality of agent computers is connected to the server by a data network. Each of the at least one digital assistants is preprogrammed and centrally managed with at least one skill for recognizing a preprogrammed specific utterance and each of the at least one digital assistants is pre-configured to connect with the wireless data interface. After the preprogrammed specific utterance is detected by one of the digital assistants, that digital assistant initiates a request for help to the server and upon receiving the request for the help, the server assigns one of the agent computers and forwards the request for help to the one of the agent computers.
    Type: Grant
    Filed: October 18, 2019
    Date of Patent: June 2, 2020
    Assignee: Ways Investments, LLC
    Inventor: Mark Edward Gray
  • Patent number: 10665243
    Abstract: A system for subvocalized speech recognition includes a plurality of sensors, a controller and a processor. The sensors are coupled to a near-eye display (NED) and configured to capture non-audible and subvocalized commands provided by a user wearing the NED. The controller interfaced with the plurality of sensors is configured to combine data acquired by each of the plurality of sensors. The processor coupled to the controller is configured to extract one or more features from the combined data, compare the one or more extracted features with a pre-determined set of commands, and determine a command of the user based on the comparison.
    Type: Grant
    Filed: October 23, 2017
    Date of Patent: May 26, 2020
    Assignee: Facebook Technologies, LLC
    Inventors: Eric Michael Whitmire, Laura Cristina Trutoiu
  • Patent number: 10650192
    Abstract: The invention discloses a method and device for recognizing a domain named entity. The main technical solution is: conducting word segmentation on a text to be recognized; tagging each segmented word in the text to be recognized according to a tag set of a domain corresponding to the text to be recognized, where the tag set includes a basic tag set based on domain named entity recognition and a tag set belongs to the corresponding domain, and the basic tag set includes a location tag of related words which constitute the domain named entity; and extracting the tagged segmented words according to a domain named entity extraction rule and constituting the domain named entity.
    Type: Grant
    Filed: December 2, 2016
    Date of Patent: May 12, 2020
    Assignee: Beijing Gridsum Technology Co., Ltd.
    Inventors: Wenbin Xu, Xin He
  • Patent number: 10643617
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for voice recognition. In one aspect, a method includes the actions of receiving a voice input; determining a transcription for the voice input, wherein determining the transcription for the voice input includes, for a plurality of segments of the voice input: obtaining a first candidate transcription for a first segment of the voice input; determining one or more contexts associated with the first candidate transcription; adjusting a respective weight for each of the one or more contexts; and determining a second candidate transcription for a second segment of the voice input based in part on the adjusted weights; and providing the transcription of the plurality of segments of the voice input for output.
    Type: Grant
    Filed: March 14, 2019
    Date of Patent: May 5, 2020
    Assignee: Google LLC
    Inventors: Petar Aleksic, Pedro J. Moreno Mengibar
  • Patent number: 10643620
    Abstract: A speech recognition method includes: storing at least one acoustic model (AM); obtaining, from a device located outside the ASR server, a device ID for identifying the device; obtaining speech data from the device; selecting an AM based on the device ID; performing speech recognition on the speech data by using the selected AM; and outputting a result of the speech recognition.
    Type: Grant
    Filed: May 6, 2015
    Date of Patent: May 5, 2020
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Tae-yoon Kim, Chang-woo Han, Jae-won Lee
  • Patent number: 10629198
    Abstract: It is inter alia disclosed to perform at least one of operating an interaction process with a user of the medical apparatus and determining, based on a representation of at least one instruction given by the user, at least one instruction operable by the medical apparatus. Therein, the at least one of the operating and the determining at least partially depends on operational data. It is further disclosed to receive modification information for modifying at least a part of the operational data, wherein the modification information is at least partially determined based on an analysis of a representation of at least one instruction given by the user.
    Type: Grant
    Filed: October 24, 2017
    Date of Patent: April 21, 2020
    Assignee: Sanofi-Aventis Deutschland GmbH
    Inventor: Michael George
  • Patent number: 10629201
    Abstract: An apparatus configured for correcting an utterance error of a user, which is configured for extracting a sematic frame with high accuracy even though the utterance of the user is incomplete, and a method thereof may include a voice activity detector configured to extract a voice after distinguishing a voice part and a mute part in a user utterance; a voice recognizer configured to output a text corresponding to the voice extracted by the voice activity detector after recognizing the voice; a named-entity recognizer configured to recognize a named entity of each text output from the voice recognizer; a text normalizer configured to normalize each text based on the named entity of each text recognized by the named entity recognizer; and a text determiner configured to determine a final text among normalized texts.
    Type: Grant
    Filed: March 23, 2018
    Date of Patent: April 21, 2020
    Assignees: Hyundai Motor Company, Kia Motors Corporation
    Inventors: Sung Soo Park, Jae Min Joh
  • Patent number: 10622008
    Abstract: An audio processing apparatus includes a first-section detection unit configured to detect a first section that is a section in which the power of a spatial spectrum in a sound source direction is higher than a predetermined amount of power on the basis of an audio signal of a plurality of channels, a speech state determination unit configured to determine a speech state on the basis of an audio signal within the first section, a likelihood calculation unit configured to calculate a first likelihood that a type of sound source according to an audio signal within the first section is voice and a second likelihood that the type of sound source is non-voice, and a second-section detection unit configured to determine whether or not a second section in which power is higher than average the power of a speech section is a voice section on the basis of the first likelihood and the second likelihood within the second section.
    Type: Grant
    Filed: June 27, 2016
    Date of Patent: April 14, 2020
    Assignee: HONDA MOTOR CO., LTD.
    Inventors: Keisuke Nakamura, Kazuhiro Nakadai
  • Patent number: 10614803
    Abstract: The present disclosure provides a wake-on-voice method, a terminal and a storage medium. The method includes: acquiring a wake-up voice configured to wake up a smart terminal; performing an analysis on an acoustic feature of the wake-up voice by using a preset acoustic model and a preset wake-up word recognition network of the smart terminal, so as to acquire a confidence coefficient of the acoustic feature of the wake-up voice with respect to an acoustic feature of a preset wake-up word; determining whether the confidence coefficient falls in a preset range of moderate confidence coefficients, if yes, uploading the wake-up voice to a remote server; and determining whether a linguistic feature obtained by analyzing the wake-up voice using a linguistic model matches to a linguistic feature of the preset wake-up word, if yes, receiving an instruction to wake up the smart terminal generated by the remote server.
    Type: Grant
    Filed: December 27, 2017
    Date of Patent: April 7, 2020
    Assignee: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD.
    Inventors: Bo Xie, Sheng Qian, Yang Sun, Yan Xie
  • Patent number: 10614810
    Abstract: Early selection of operating parameters for improving accuracy of transcriptions generated by automatic speech recognition (ASR) systems. In one embodiment, a server receives an audio recording that includes speech, taken over a period spanning at least two hours. During the first hour, the server receives a ground truth transcription of a certain segment of the audio recording, created by a transcriber after listening to the certain segment. The server operates an ASR system a plurality of times, using a plurality of sets of operating parameters, to generate a plurality of respective transcriptions of the certain segment. The server evaluates accuracies of the plurality of transcriptions with respect to the ground truth transcription, and selects an optimal set of operating parameters. The server may then apply the optimal set of operating parameters to transcribe additional segments of the audio recording utilizing the ASR system.
    Type: Grant
    Filed: October 7, 2019
    Date of Patent: April 7, 2020
    Assignee: Verbit Software Ltd.
    Inventors: Eric Ariel Shellef, Yaakov Kobi Ben Tsvi, Iris Getz, Tom Livne, Eli Asor, Elad Shtilerman
  • Patent number: 10609221
    Abstract: A system for processing voice responses is disclosed. The system is configured to store a correlation table identifying relationships between self-service routines, tags, and corresponding actions. The system receives a call from a user and issues a query in response to the call. The system receives an utterance from the user in response to the user and determines whether the utterance matches a pre-defined response. If there is no match, the system analyzes the utterance with a pre-defined statistical language model and identifies a service tag for the utterance. The system then associates the utterance with the service tag and a self-service routine that is associated with the call. The system identifies an action from the correlation table that correlates to the service tag and the self-service routine.
    Type: Grant
    Filed: November 11, 2019
    Date of Patent: March 31, 2020
    Assignee: Bank of America Corporation
    Inventors: Kyle A. Tobin, Robert E. Lutzkow, Jr., Robert S. Morse, Jitendra Kumar Padam, Scott Steven Randrup
  • Patent number: 10607596
    Abstract: A mistranscription generated by a speech recognition system is identified. A first class of utterance members is provided for use by the speech recognition system, each utterance class member consisting of a respective number of words. The first class is defined by a first common meaning and a first common system response if a class member of the first class is recognized. If the speech recognition system matches a received utterance to a first class member of the first class, the received utterance is sent to a mistranscription analyzer. The received utterance contains a mistranscription as compared to the first class member. Evidence is incremented by the mistranscription analyzer that the received utterance is evidence of the mistranscription of the first class member. If the incremented evidence for the mistranscription of the first class member exceeds a first threshold, a second class member is added to the first class of utterance members based on the mistranscription of the first class member.
    Type: Grant
    Filed: January 7, 2018
    Date of Patent: March 31, 2020
    Assignee: International Business Machines Corporation
    Inventors: Andrew Aaron, Shang Guo, Jonathan Lenchner, Maharaj Mukherjee
  • Patent number: 10602287
    Abstract: An audio transmission system of the present invention includes: an unmanned aircraft; a steering terminal that is used to steer the unmanned aircraft; and an information terminal configured to access a server providing a service for translating text data created in an official language of a country into a designated language other than the official language, and converting the translated text data into audio data, the information terminal being able to download the audio data from the server. The information terminal transmits the audio data in the designated language downloaded from the server to the unmanned aircraft, and the unmanned aircraft outputs the audio data in the designated language as an audio message from a speaker attached to an airframe.
    Type: Grant
    Filed: March 15, 2019
    Date of Patent: March 24, 2020
    Assignee: QUALITYSOFT CORPORATION
    Inventor: Makoto Miyawaki
  • Patent number: 10593320
    Abstract: A mistranscription generated by a speech recognition system is identified. A set of known utterance members is provided for use by a speech recognition system. Each utterance member consists of a respective plural number of words. A received utterance is matched to a first utterance member within the set of known utterance members. The first utterance member is a closest matching utterance member and has a first plural number of words. The matching operation matches fewer than the first plural number of words in the received utterance and the received utterance varies in a first particular manner as compared to a first word in a first slot in the first utterance member. The received utterance is sent to a mistranscription analyzer component which increments evidence that the received utterance is evidence of a mistranscription.
    Type: Grant
    Filed: January 7, 2018
    Date of Patent: March 17, 2020
    Assignee: International Business Machines Corporation
    Inventors: Andrew Aaron, Shang Guo, Jonathan Lenchner, Maharaj Mukherjee
  • Patent number: 10572592
    Abstract: Disclosed is a method for providing at least one word linguistically associated with at least one searched word belonging to a set of words. After having queried (325) a first database of expressions to obtain a set of expressions including the at least one searched word and obtaining the set of expressions, a second database is queried (340), for each expression of at least an expression subset of the obtained set of expressions, to obtain at least one word linguistically associated with the at least one searched word and obtaining the at least one word linguistically associated with the at least one searched word. Next, at least one obtained word linguistically associated with the at least one searched word is selected (350).
    Type: Grant
    Filed: February 2, 2017
    Date of Patent: February 25, 2020
    Inventor: Theo Hoffenberg
  • Patent number: 10553219
    Abstract: A voice recognition apparatus, a voice recognition method, and a non-transitory computer readable recording medium are provided. The voice recognition apparatus includes a storage configured to store a preset threshold value for voice recognition; a voice receiver configured to receive a voice signal of an uttered voice; and a voice recognition processor configured to recognize a voice recognition starting word from the received voice signal, perform the voice recognition on the voice signal in response to a similarity score, which represents a recognition result of the recognized voice recognition starting word, being greater than or equal to the stored preset threshold value, and change the preset threshold value based on the recognition result of the voice recognition starting word.
    Type: Grant
    Filed: July 19, 2016
    Date of Patent: February 4, 2020
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventor: Chi-sang Jung
  • Patent number: 10553210
    Abstract: A system, apparatus, and method for processing a natural language, and a computer readable recording medium are provided. The system includes a user device configured to receive a compound or complex sentence and a natural language processing apparatus configured to generate a plurality of control commands for the user device based on whether operations intended by a user are sequentially performable in the compound or complex sentence received from the user device.
    Type: Grant
    Filed: July 18, 2016
    Date of Patent: February 4, 2020
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Ji-hye Chung, Kyung-duk Kim
  • Patent number: 10546576
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for speech endpointing based on word comparisons are described. In one aspect, a method includes the actions of obtaining a transcription of an utterance. The actions further include determining, as a first value, a quantity of text samples in a collection of text samples that (i) include terms that match the transcription, and (ii) do not include any additional terms. The actions further include determining, as a second value, a quantity of text samples in the collection of text samples that (i) include terms that match the transcription, and (ii) include one or more additional terms. The actions further include classifying the utterance as a likely incomplete utterance or not a likely incomplete utterance based at least on comparing the first value and the second value.
    Type: Grant
    Filed: October 9, 2018
    Date of Patent: January 28, 2020
    Assignee: Google LLC
    Inventors: Michael Buchanan, Pravir Kumar Gupta, Christopher Bo Tandiono
  • Patent number: 10546578
    Abstract: An artificial intelligence (AI) system configured to simulate functions of a human brain, such as recognition, determination, etc., by using a machine learning algorithm, such as deep learning, etc., and an application thereof. The AI system includes a method performed by a device to transmit and receive audio data to and from another device includes obtaining a voice input that is input by a first user of the device, obtaining recognition information indicating a meaning of the obtained voice input, transmitting the obtained voice input to the other device, determining whether an abnormal situation occurs, in which a second user of the other device does not understand the transmitted voice input, and transmitting the obtained recognition information to the other device, based on a result of the determination.
    Type: Grant
    Filed: December 20, 2017
    Date of Patent: January 28, 2020
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Jae-deok Kim, Mee-jeong Park
  • Patent number: 10540140
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for processing multimodal input. A system configured to practice the method continuously monitors an audio stream associated with a gesture input stream, and detects a speech event in the audio stream. Then the system identifies a temporal window associated with a time of the speech event, and analyzes data from the gesture input stream within the temporal window to identify a gesture event. The system processes the speech event and the gesture event to produce a multimodal command. The gesture in the gesture input stream can be directed to a display, but is remote from the display. The system can analyze the data from the gesture input stream by calculating an average of gesture coordinates within the temporal window.
    Type: Grant
    Filed: July 17, 2017
    Date of Patent: January 21, 2020
    Assignee: NUANCE COMMUNICATIONS, INC.
    Inventors: Michael Johnston, Derya Ozkan
  • Patent number: 10540599
    Abstract: A method may include extracting feature information from a user history including multiple goal messages. Each of the multiple goal messages may include visual media and text, and may be associated with a goal indicated as completed. Extracting the feature information from the user history may include performing feature detection on the visual media of the multiple goal messages. The feature information from the user history may be analyzed for correlations between features and the completion indications of the associated goals. Feature information may be extracted from a generic goal message. A probability of a user indicating completion of a goal associated with the generic goal message may be determined based at least in part on the feature information from the generic goal message and the correlations between the features included in the multiple goal messages and the indications of the associated goals having been completed.
    Type: Grant
    Filed: April 7, 2016
    Date of Patent: January 21, 2020
    Assignee: FUJITSU LIMITED
    Inventors: Ajay Chander, Ramya Malur Srinivasan
  • Patent number: 10539693
    Abstract: Described herein is a method of processing a gamma ray spectrum acquired from a target. The method comprises determining whether the gamma ray spectrum of the target belongs to a first class of a plurality of classes, the first class containing reference gamma ray spectra of one or more radionuclide sources of interest, using optimal loading coefficients associated with the one or more radionuclide sources of interest, wherein the optimal loading coefficients have been obtained using Fisher linear discriminant analysis, and generating an output signal dependent on the determining.
    Type: Grant
    Filed: July 6, 2012
    Date of Patent: January 21, 2020
    Assignee: Australian Nuclear Science and Technology Organization
    Inventor: David Boardman
  • Patent number: 10535349
    Abstract: Network connected devices are controlled. A command is received. A graph model is applied to identify a subset of nodes related to the command. The graph model includes a plurality of nodes that each correspond to a device and the graph model includes a plurality of edges that specify relationships between the plurality of nodes. The subset of nodes is activated in response to the command.
    Type: Grant
    Filed: November 7, 2018
    Date of Patent: January 14, 2020
    Assignee: BrainofT Inc.
    Inventors: Ashutosh Saxena, Brendan Berman, Deng Deng, Lukas Kroc
  • Patent number: 10535345
    Abstract: An interactive method/system generates a fictional story. A user interface receives human speech and transmits machine-generated speech. A processor(s) is programmed to execute functions that include parsing the human speech into fragments thereof and identifying a primary fragment from the fragments wherein the primary fragment includes a verb. A generalized intent is associated with the verb and at least one object is associated with the generalized intent. The generalized intent and each object associated with the generalized intent are stored. An open-ended question is generated based on the generalized intent wherein the open-ended question is provided to the user interface for transmission as machine-generated speech. After the above steps are repeated for a number of cycles, a sequence of sentences is generated using each generalized intent and object(s) associated with the generalized intent. The sequence of sentences is transmitted as machine-generated speech from the user interface.
    Type: Grant
    Filed: October 20, 2017
    Date of Patent: January 14, 2020
    Inventor: Yingjia Liu
  • Patent number: 10529329
    Abstract: A method and apparatus for improving the performance of voice recognition in a mobile device are provided. The method of recognizing a voice includes: monitoring the usage pattern of a user of a device for inputting a voice; selecting predetermined words from among words stored in the device based on the result of monitoring, and storing the selected words; and recognizing a voice based on an acoustic model and predetermined words. In this way, a voice can be recognized by using prediction of whom the user mainly makes a call to. Also, by automatically modeling the device usage pattern of the user and applying the pattern to vocabulary for voice recognition based on probabilities, the performance of voice recognition, as actually felt by the user, can be enhanced.
    Type: Grant
    Filed: November 3, 2017
    Date of Patent: January 7, 2020
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Kyu-hong Kim, Jeong-su Kim, Ick-sang Han
  • Patent number: 10529335
    Abstract: An auto-complete method for a spoken complete value entry is provided. A processor receives a possible complete value entry having a unique subset, prompts a user to speak the spoken complete value entry, receives a spoken subset of the spoken complete value entry, compares the spoken subset with the unique subset of the possible complete value entry, and automatically completes the spoken complete value entry to match the possible complete value entry if the unique subset matches the spoken subset. The spoken subset has a predetermined minimum number of characters.
    Type: Grant
    Filed: April 9, 2019
    Date of Patent: January 7, 2020
    Assignee: HAND HELD PRODUCTS, INC.
    Inventors: Matthew Nichols, Alexander Nikolaus Mracna, Kurt Charles Miller, Russell Evans, Bernard Kriley, Luke Sadecky, Brian Manuel, Navaneetha Myaka
  • Patent number: 10529315
    Abstract: The present disclosure relates to methods, non-transitory computer readable media, and devices for text-to-speech conversion of electronic documents. An electronic document comprising one or more pages comprising a plurality of characters and a plurality of first segments is received. The plurality of characters is segmented into a plurality of second segments based on first metadata associated with the plurality of characters. A first relationship between each of the plurality of second segments is identified based on the first metadata associated with the plurality of characters, second metadata associated with the plurality of first segments, and spatial information associated with the plurality of segments. A reading sequence of the electronic document is determined based on the first relationship. An audio is then generated based on the reading sequence of the electronic document.
    Type: Grant
    Filed: October 3, 2017
    Date of Patent: January 7, 2020
    Assignee: Wipro Limited
    Inventor: Dhruv Premi
  • Patent number: 10523773
    Abstract: Systems, apparatuses, and methods are provided for enhancing users' overall experiences with physical items by supplementing their physical experiences with digital experiences. According to an embodiment, a user uses an electronic device to scan a smart tag associated with an item to obtain an item identifier of the item. The electronic device sends the item identifier to a server, which selects digital content related to the item and sends the selected digital content to the user's electronic device for display.
    Type: Grant
    Filed: October 17, 2018
    Date of Patent: December 31, 2019
    Assignee: adidas AG
    Inventors: Jon H. Werner, Christian Dibenedetto, Amy Jones Vaterlaus, Stephen M. Vincent
  • Patent number: 10522148
    Abstract: A communications device and method are provided for converting speech to text and applying corrections to the text. The communications device may include at least one audio interface, such a microphone and/or speaker, and at least one communications subsystem, as well as a controller or processor operative to receive speech input using the at least one audio interface, convert the speech input to input text, correct the input text to corrected text, and send the corrected text over a network using the communications subsystem. The corrected text may involve the application of proposed modification, such as a grammatical correction or ambiguity resolution, to the input text. The application of the proposed modification may be based upon the receipt of an instruction to accept or reject the proposed correction or resolution. The instruction may be a spoken instruction.
    Type: Grant
    Filed: February 10, 2016
    Date of Patent: December 31, 2019
    Assignee: BlackBerry Limited
    Inventor: Jennawae McLean
  • Patent number: 10514268
    Abstract: An in-vehicle terminal sends a speech input from a voice input unit as a voice signal to a relay server using a short-range wireless communication unit. The relay server converts the voice data received from the in-vehicle terminal into a string with a voice recognition unit, and makes an inquiry, via a communication unit, to an address database of a database server as to whether the string is a geographical condition. If the string converted by the voice recognition unit is a geographical condition based on the inquiry result of the communication unit, a POI data base is searched based on the geographical condition and a stored non-geographical condition, and the geographical condition is stored in a search history. If the string is a non-geographical condition, the POI database is searched based on the non-geographical condition and the geographical condition stored in the search history, and the non-geographical condition is stored in the search history.
    Type: Grant
    Filed: September 15, 2015
    Date of Patent: December 24, 2019
    Assignee: Clarion Co., Ltd.
    Inventors: Takashi Yamaguchi, Susumu Kojima
  • Patent number: 10504503
    Abstract: A method performed by a speech recognizing apparatus to recognize speech includes: obtaining a distance from the speech recognizing apparatus to a user generating a speech signal; determining a normalization value for the speech signal based on the distance; normalizing a feature vector extracted from the speech signal based on the normalization value; and performing speech recognition based on the normalized feature vector.
    Type: Grant
    Filed: May 3, 2017
    Date of Patent: December 10, 2019
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Hodong Lee, Hoshik Lee
  • Patent number: 10496648
    Abstract: Methods, systems, and computer readable media are described herein for processing a query on a parent table and associated child tables to provide efficient search engine functionality. Search grams are determined based on the query and parent records and child records are scanned for the search grams. Gram scores are computed based on a degree of matching between the parent and child records and the search grams.
    Type: Grant
    Filed: February 8, 2017
    Date of Patent: December 3, 2019
    Assignee: TIBCO Software Inc.
    Inventors: Fletcher Foti, Bruce Zuidema
  • Patent number: 10496759
    Abstract: A language translation application on a user device includes a user interface that provides relevant textual and graphical feedback mechanisms associated with various states of voice input and translated speech.
    Type: Grant
    Filed: May 8, 2018
    Date of Patent: December 3, 2019
    Assignee: Google LLC
    Inventors: Alexander Jay Cuthbert, Sunny Goyal, Matthew Morton Gaba, Joshua J. Estelle, Masakazu Seno
  • Patent number: 10492054
    Abstract: A system for providing help includes at least one digital assistant and a wireless data interface for connecting the digital assistants to a server. A plurality of agent computers is connected to the server by a data network. Each of the at least one digital assistants is preprogrammed and centrally managed with at least one skill for recognizing a preprogrammed specific utterance and each of the at least one digital assistants is pre-configured to connect with the wireless data interface. After the preprogrammed specific utterance is detected by one of the digital assistants, that digital assistant initiates a request for help to the server and upon receiving the request for the help, the server assigns one of the agent computers and forwards the request for help to the one of the agent computers.
    Type: Grant
    Filed: March 5, 2019
    Date of Patent: November 26, 2019
    Assignee: Ways Investments, LLC
    Inventor: Mark Edward Gray
  • Patent number: 10489438
    Abstract: The disclosed embodiments illustrate methods of data processing for text classification of a target domain. The method includes generating a plurality of clusters from a plurality of first text segments corresponding to a plurality of source domains, based on an association of the plurality of first text segments with a plurality of categories. The method further includes computing a similarity score of each of a plurality of second text segments corresponding to the target domain for each of the plurality of clusters. The method further includes identifying a pre-specified count of clusters from the plurality of clusters, based on the computed similarity score. Further, the method includes training a first classifier by utilizing first text segments in the identified pre-specified count of clusters, wherein the trained first classifier is utilized to automatically classify the plurality of second text segments into categories associated with the identified pre-specified count of clusters.
    Type: Grant
    Filed: May 19, 2016
    Date of Patent: November 26, 2019
    Assignee: CONDUENT BUSINESS SERVICES, LLC
    Inventors: Himanshu Sharad Bhatt, Manjira Sinha, Shourya Roy
  • Patent number: 10492013
    Abstract: A system and method of testing one or more operations of a vehicle system module (VSM) using computer-based audio recognition techniques, wherein the method is carried out by a computer that includes or is connected to a microphone, wherein the VSM includes or is connected to an audio speaker, and wherein the method includes the steps of: carrying out one or more initial VSM test actions; listening for one or more audio signals from the VSM audio speaker using the computer microphone; processing the one or more audio signals using computer-based audio-recognition technique(s); and determining whether the one or more performed VSM operations correspond to one or more VSM test actions.
    Type: Grant
    Filed: September 14, 2017
    Date of Patent: November 26, 2019
    Assignee: GM GLOBAL TECHNOLOGY OPERATIONS LLC
    Inventors: Robert G. Qiu, Meijun Wang, Gagandeep S. Chabbewal
  • Patent number: 10482882
    Abstract: Methods, apparatus, and computer readable media are described related to automated assistants that proactively incorporate, into human-to-computer dialog sessions, unsolicited content of potential interest to a user. In various implementations, based on content of an existing human-to-computer dialog session between a user and an automated assistant, an entity mentioned by the user or automated assistant may be identified. Fact(s)s related to the entity or to another entity that is related to the entity may be identified based on entity data contained in database(s). For each of the fact(s), a corresponding measure of potential interest to the user may be determined. Unsolicited natural language content may then be generated that includes one or more of the facts selected based on the corresponding measure(s) of potential interest. The automated assistant may then incorporate the unsolicited content into the existing human-to-computer dialog session or a subsequent human-to-computer dialog session.
    Type: Grant
    Filed: November 29, 2017
    Date of Patent: November 19, 2019
    Assignee: Google LLC
    Inventors: Vladimir Vuskovic, Stephan Wenger, Zineb Ait Bahajji, Martin Baeuml, Alexandru Dovlecel, Gleb Skobeltsyn
  • Patent number: 10477028
    Abstract: A system for processing voice responses is disclosed. The system is configured to store a correlation table identifying relationships between self-service routines, tags, and corresponding actions. The system receives a call from a user and issues a query in response to the call. The system receives an utterance from the user in response to the user and determines whether the utterance matches a pre-defined response. If there is no match, the system analyzes the utterance with a pre-defined statistical language model and identifies a service tag for the utterance. The system then associates the utterance with the service tag and a self-service routine that is associated with the call. The system identifies an action from the correlation table that correlates to the service tag and the self-service routine.
    Type: Grant
    Filed: July 13, 2018
    Date of Patent: November 12, 2019
    Assignee: Bank of America Corporation
    Inventors: Kyle A. Tobin, Robert E. Lutzkow, Jr., Robert S. Morse, Jitendra Padam, Scott Steven Randrup
  • Patent number: 10468031
    Abstract: An approach is provided that receives an audio stream and utilizes a voice activation detection (VAD) process to create a digital audio stream of voices from at least two different speakers. An automatic speech recognition (ASR) process is applied to the digital stream with the ASR process resulting in the spoken words to which a speaker turn detection (STD) process is applied to identify a number of speaker segments with each speaker segment ending at a word boundary. The STD process analyzes a number of speaker segments using a language model that determines when speaker changes occur. A speaker clustering algorithm is then applied to the speaker segments to associate one of the speakers with each of the speaker segments.
    Type: Grant
    Filed: November 21, 2017
    Date of Patent: November 5, 2019
    Assignee: International Business Machines Corporation
    Inventors: Kenneth W. Church, Dimitrios B. Dimitriadis, Petr Fousek, Miroslav Novak, George A. Saon
  • Patent number: 10460722
    Abstract: A method for selective transmission of audio data to a speech processing server uses detection of an acoustic trigger in the audio data in determining the data to transmit. Detection of the acoustic trigger makes use of an efficient computation approach that reduces the amount of run-time computation required, or equivalently improves accuracy for a given amount of computation, by combining a “time delay” structure in which intermediate results of computations are reused at various time delays, thereby avoiding computation of computing new results, and decomposition of certain transformations to require fewer arithmetic operations without sacrificing significant performance. For a given amount of computation capacity the combination of these two techniques provides improved accuracy as compared to current approaches.
    Type: Grant
    Filed: June 30, 2017
    Date of Patent: October 29, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Ming Sun, David Snyder, Yixin Gao, Nikko Strom, Spyros Matsoukas, Shiv Naga Prasad Vitaladevuni