Speech To Image Patents (Class 704/235)
  • Patent number: 10684136
    Abstract: A computer-implemented method includes receiving an input, from a user, in the form of a destination to be navigated to; calculating a route to the destination based on the input; recognizing at least one object on a route taken by the user; processing data from the received input, the calculated route, and the recognized at least one object; and providing an output to the user based on the recognized at least one object, the output being based on natural language processing.
    Type: Grant
    Filed: February 28, 2017
    Date of Patent: June 16, 2020
    Assignee: International Business Machines Corporation
    Inventors: Julia S. Baldauf, Fatemeh Jalali, Benjamin Scott Mashford, Mahsa Salehi
  • Patent number: 10679610
    Abstract: A method for eyes-off training of a dictation system includes translating an audio signal featuring speech audio of a speaker into an initial recognized text using a previously-trained general language model. The initial recognized text is provided to the speaker for error correction. The audio signal is re-translated into an updated recognized text using a specialized language model biased to recognize words included in the corrected text. The general language model is retrained in an “eyes-off” manner, based on the audio signal and the updated recognized text.
    Type: Grant
    Filed: July 16, 2018
    Date of Patent: June 9, 2020
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Hemant Malhotra, Shuangyu Chang, Pradip Kumar Fatehpuria
  • Patent number: 10679609
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for natural language processing. One of the method includes receiving a voice input from a user device; generating a recognition output; receiving a user selection of one or more terms in the recognition output; receiving a user input of one or more letters replacing the user selected one or more terms; determining suggested correction candidates based in part on the user input and the voice input; and providing one or more suggested correction candidates to the user device as suggested corrected recognition outputs.
    Type: Grant
    Filed: December 3, 2019
    Date of Patent: June 9, 2020
    Assignee: Google LLC
    Inventors: Evgeny A. Cherepanov, Jakob Nicolaus Foerster, Vikram Sridar, Ishai Rabinovitz, Omer Tabach
  • Patent number: 10672394
    Abstract: The subject matter of this specification can be implemented in, among other things, a computer-implemented method for correcting words in transcribed text including receiving speech audio data from a microphone. The method further includes sending the speech audio data to a transcription system. The method further includes receiving a word lattice transcribed from the speech audio data by the transcription system. The method further includes presenting one or more transcribed words from the word lattice. The method further includes receiving a user selection of at least one of the presented transcribed words. The method further includes presenting one or more alternate words from the word lattice for the selected transcribed word. The method further includes receiving a user selection of at least one of the alternate words. The method further includes replacing the selected transcribed word in the presented transcribed words with the selected alternate word.
    Type: Grant
    Filed: December 21, 2017
    Date of Patent: June 2, 2020
    Assignee: Google LLC
    Inventors: Michael J. LeBeau, William J. Byrne, John Nicholas Jitkoff, Brandon M. Ballinger, Trausti T. Kristjansson
  • Patent number: 10665241
    Abstract: Being able to rapidly and accurately transcribe long audio recordings, such as same-day transcription of multi-hour legal depositions, is a challenging task. Hybrid transcription, which involves automatic speech recognition (ASR) systems generating initial transcriptions that are then reviewed by human transcribers, can be used to tackle this challenge. However, hybrid transcription may be stymied when the transcribers cannot resolve certain issues in the ASR-generated transcriptions. This disclosure describes rapid resolution of transcription-related inquiries of transcribers. In one embodiment, a computer receives an audio recording that includes speech of multiple people in a room and generates transcriptions of segments of the audio recording utilizing an ASR system. These transcriptions are provided for review of transcribers.
    Type: Grant
    Filed: October 7, 2019
    Date of Patent: May 26, 2020
    Assignee: Verbit Software Ltd.
    Inventors: Eric Ariel Shellef, Yaakov Kobi Ben Tsvi, Iris Getz, Tom Livne, Roman Himmelreich
  • Patent number: 10659585
    Abstract: A system and method for making a call to a service provider on behalf of a user is disclosed. The system and method include using an intelligent voice assistant to call the service provider and having the intelligent voice assistant navigate an interactive voice response system to reach a representative. The system and method also include using the intelligent voice assistant to reconnect a user with a representative after an interrupted call.
    Type: Grant
    Filed: May 7, 2019
    Date of Patent: May 19, 2020
    Assignee: United States Automobile Association (USAA)
    Inventors: Richard Daniel Graham, Ruthie D. Lyle
  • Patent number: 10657952
    Abstract: Techniques are provided for reducing the latency of automatic speech recognition using hypothesis score trend analysis. A methodology implementing the techniques according to an embodiment includes generating complete-phrase hypotheses and partial-phrase hypotheses, along with associated likelihood scores, based on a segment of speech. The method also includes selecting the complete-phrase hypothesis associated with the highest of the complete-phrase hypotheses likelihood scores, and selecting the partial-phrase hypothesis associated with the highest of the partial-phrase hypotheses likelihood scores. The method further includes calculating a relative likelihood score based on a ratio of the likelihood score associated with the selected complete-phrase hypothesis to the likelihood score associated with the selected partial-phrase hypothesis.
    Type: Grant
    Filed: February 9, 2018
    Date of Patent: May 19, 2020
    Assignee: Intel IP Corporation
    Inventors: Joachim Hofer, Georg Stemmer, Josef G. Bauer, Munir Nikolai Alexander Georges
  • Patent number: 10657204
    Abstract: A cognitive assistant system, computer program product, and computer-implemented method configured to improve the accuracy of a statement. In one example, the system receives from a user an original statement that includes statement entities and generates an original confidence estimate (CE) for a correctness of the original statement. The system generates a replacement statement by selecting a suspect statement entity, selecting a replacement entity for the suspect statement entity, and replacing the suspect statement entity with the replacement entity in the original statement. The system generates a replacement CE for a correctness of the replacement statement. If the replacement CE exceeds the original CE by a tuning factor, the system sends the replacement statement to the user. If not, the system repeats generating the replacement statement and generating the replacement CE.
    Type: Grant
    Filed: July 17, 2018
    Date of Patent: May 19, 2020
    Assignee: International Business Machines Corporation
    Inventors: Kyle M. Brake, Stephen A. Boxwell, Keith G. Frost, Stanley J. Vernier
  • Patent number: 10650100
    Abstract: A computer-implemented method, computer system, and computer program product for improving a natural language generation knowledge base is provided. The method may include detecting user input comprising a natural language expression, generating an erroneous natural language response based on the user input and the knowledge base, determining user feedback corresponding to the erroneous natural language response, wherein the user feedback comprises an indication of an end-user reaction to the erroneous natural language response, determining an improvable performance metric with respect to the knowledge base, and updating the knowledge base based on the improvable performance metric, wherein the knowledge base comprises an explicit model of language corresponding to the erroneous natural language response, and wherein updating the knowledge base comprises updating the explicit model of language based on the user input, the erroneous natural language response, and the user feedback.
    Type: Grant
    Filed: June 8, 2018
    Date of Patent: May 12, 2020
    Assignee: International Business Machines Corporation
    Inventors: Alaa Abou Mahmoud, Paul R. Bastide, Matthew E. Broomhall, Robert E. Loredo, Fang Lu
  • Patent number: 10650831
    Abstract: Various embodiments of the technology described herein alleviate the need to specifically request enrollment information from a user to enroll the user in a voice biometric authentication program. For example, the system can receive a voice interaction from a user where the voice interaction includes a request or a command having one or more portions. After the user is authenticated using non-voice biometric authentication information, the system enrolls the user into a voice biometric authentication program for at least one portion of the request or the command. The system also enrolls the user into the voice biometric authentication program for other requests or commands that include one of the at least one portion of the request or the command and portions of one or more second requests or commands for which the user has been enrolled into the voice biometric authentication program.
    Type: Grant
    Filed: May 30, 2018
    Date of Patent: May 12, 2020
    Assignee: United Services Automobile Association (USAA)
    Inventors: Zakery Layne Johnson, Maland Keith Mortensen, Gabriel Carlos Fernandez, Debra Randall Casillas, Sudarshan Rangarajan, Thomas Bret Buckingham
  • Patent number: 10642889
    Abstract: A method for information processing includes receiving in a computer a corpus of recorded conversations, with two or more speakers participating in each conversation. Respective frequencies of occurrence of multiple words in each of a plurality of chunks in each of the recorded conversations are computed. Based on the frequencies of occurrence of the words over the conversations in the corpus, an optimal set of topics to which the chunks can be assigned is derived, such that the optimal set maximizes a likelihood that the chunks will be generated by the topics in the set. A recorded conversation from the corpus is segmented using the derived topics into a plurality of segments, such that each segment is classified as belonging to a particular topic in the optimal set.
    Type: Grant
    Filed: December 7, 2017
    Date of Patent: May 5, 2020
    Assignee: GONG I.O LTD.
    Inventors: Eilon Reshef, Zvi Marx
  • Patent number: 10643616
    Abstract: The technology of the present application provides a method and apparatus to manage speech resources. The method includes using a text recognizer to detect a change in a speech application that requires the use of different resources. On detection of the change, the method loads the different resources without the user needing to exit the currently executing speech application.
    Type: Grant
    Filed: April 11, 2018
    Date of Patent: May 5, 2020
    Assignee: nVoq Incorporated
    Inventor: Charles Corfield
  • Patent number: 10636427
    Abstract: Examples described herein improve the way in which a transcript is generated and displayed so that the context of a conversation taking place during a meeting or another type of collaboration event can be understood by a person that reviews the transcript (e.g., reads or browses through the transcript). The techniques described herein use voice recognition to identify a user that is speaking during the meeting. Accordingly, when the speech of the user is converted to text for the transcript, the text can be attributed to the identified user. The techniques described herein further configure a graphical user interface layout, in which the transcript can be displayed. The graphical user interface layout enables users to better understand the context of a conversation that takes place during a meeting.
    Type: Grant
    Filed: June 22, 2018
    Date of Patent: April 28, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Christopher Andrews Jung, Benjamin David Smith, Erez Kikin Gil
  • Patent number: 10629178
    Abstract: Methods and apparatus to classify media based on a pitch-independent timbre attribute from a media signal are disclosed. An example apparatus includes an interface to access a media signal; and an audio characteristic extractor to determine a spectrum of audio corresponding to the media signal; and determine a timbre-independent pitch attribute of the audio based on an inverse transform of a complex argument of a transform of the spectrum.
    Type: Grant
    Filed: October 21, 2019
    Date of Patent: April 21, 2020
    Assignee: The Nielsen Company (US), LLC
    Inventor: Zafar Rafii
  • Patent number: 10629203
    Abstract: Voice information is received, via a network from a first voice assistant. The voice information comprises a digitized voice stream and/or information associated with the digitized voice stream. A keyword is identified in the voice information. The keyword is used to create a first bookmark. In response to identifying the first keyword in the voice information, a name for the first bookmark is received in the voice information, a command for the first bookmark is received in the voice information; and the bookmark is stored in a memory.
    Type: Grant
    Filed: June 29, 2018
    Date of Patent: April 21, 2020
    Assignee: Flex, Ltd.
    Inventor: Anoop Krishnan Nair
  • Patent number: 10614808
    Abstract: Aspects of the disclosure provide systems and methods for facilitating dictation. Speech input may be provided to an audio input device of a computing device. A speech recognition engine at the computing device may obtain text corresponding to the speech input. The computing device may transmit the text to a remotely-located storage device. A login webpage that includes a session identifier may be accessed from a target computing device also located remotely relative to the storage device. The session identifier may be transmitted to the storage device and, in response, a text display webpage may be received at the target computing device. The text display webpage may include the speech-derived text and may be configured to automatically copy the text to a copy buffer of the target computing device. The speech-derived text may also be provided to native applications at target computing devices or NLU engines for natural language processing.
    Type: Grant
    Filed: June 30, 2014
    Date of Patent: April 7, 2020
    Assignee: Nuance Communications, Inc.
    Inventors: Markus Vogel, Andreas Neubacher
  • Patent number: 10607611
    Abstract: When transcribing large audio files, such as in the case of legal depositions, there are often many transcribers to choose from. Embodiments described herein enable calculation of expected accuracy of transcriptions by transcribers, which can be used to guide the selection of transcribers for specific tasks. In one embodiment, a computer receives a segment of an audio recording that includes speech of a person, and identifies an accent of the person and a topic of the segment. The computer generates feature values based on data that includes the accent and the topic, and utilizes a model to calculate, based on the feature values, an expected accuracy of a transcription of the segment by a certain transcriber. The model is generated based on training data that includes segments of previous audio recordings and values of accuracies of transcriptions, by the certain transcriber, of the segments.
    Type: Grant
    Filed: October 7, 2019
    Date of Patent: March 31, 2020
    Assignee: Verbit Software Ltd.
    Inventors: Eric Ariel Shellef, Yaakov Kobi Ben Tsvi, Iris Getz, Tom Livne, Elisha Yehuda Rosensweig
  • Patent number: 10606554
    Abstract: A method and apparatus for providing voice command functionality to an interactive whiteboard appliance is provided. An interactive whiteboard appliance comprises: one or more processors; a non-transitory computer-readable medium having instructions embodied thereon, the instructions when executed by the one or more processors cause performance of: detecting, during execution of an annotation window on the interactive whiteboard appliance, a voice input received from a user; storing, in an audio packet, a recording of the voice input; transmitting the audio packet to a speech-to-text service; receiving, from the speech-to-text service, a command string comprising a transcription of the recording of the voice input; using voice mode command processing in a command processor, identifying, from the command string, an executable command that is executable by the interactive whiteboard appliance; causing the application of the interactive whiteboard appliance to execute the executable command.
    Type: Grant
    Filed: August 21, 2019
    Date of Patent: March 31, 2020
    Assignee: RICOH COMPANY, LTD.
    Inventors: Rathnakara Malatesha, Lana Wong, Hiroshi Kitada
  • Patent number: 10602237
    Abstract: Systems and methods for an ephemeral digital story channel may include (1) maintaining a primary story channel for a user of a social networking platform in which the user may post digital story compositions, (2) determining that a particular time period coincides with a life event of the user, (3) maintaining, during the time period, an additional story channel for the life event in which the user and the user's contacts may post digital story compositions, such that the social networking platform is maintaining both the primary story channel and the additional story channel during the time period, and (4) discontinuing, after the time period, the additional story channel while continuing to maintain the primary story channel. Various other methods, systems, and computer-readable media are also disclosed.
    Type: Grant
    Filed: December 10, 2018
    Date of Patent: March 24, 2020
    Assignee: Facebook, Inc.
    Inventor: Debashish Paul
  • Patent number: 10600419
    Abstract: Techniques for performing command processing are described. A system receives, from a device, input data corresponding to a command. The system determines NLU processing results associated with multiple applications. The system also determines NLU confidences for the NLU processing results for each application. The system sends NLU processing results to a portion of the multiple applications, and receives output data or instructions from the portion of the applications. The system ranks the portion of the applications based at least in part on the NLU processing results associated with the portion of the applications as well as the output data or instructions provided by the portion of the applications. The system may also rank the portion of the applications using other data. The system causes content corresponding to output data or instructions provided by the highest ranked application to be output to a user.
    Type: Grant
    Filed: September 22, 2017
    Date of Patent: March 24, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Ruhi Sarikaya, Rohit Prasad, Kerry Hammil, Spyridon Matsoukas, Nikko Strom, Frédéric Johan Georges Deramat, Stephen Frederick Potter, Young-Bum Kim
  • Patent number: 10599823
    Abstract: Systems and methods are disclosed for coordinating applications by detecting a touch input and a related touch input a device locations associated with the applications, one of which is a user authentication application. A parameter may be established with the user authentication application so that the other application runs with the at least one parameter.
    Type: Grant
    Filed: November 18, 2016
    Date of Patent: March 24, 2020
    Assignee: InvenSense, Inc.
    Inventor: Karthik Katingari
  • Patent number: 10592555
    Abstract: A query is received from a user at a data service engine. The query includes a string of characters. A number of candidate topics are identified by the data service engine based on the query. A similarity score is determined between the query and each of the plurality of candidate topics based on a Vector Propagation On a Click Graph (VPCG) model trained based on user click data. A number of candidate topics are ranked based on the similarity scores. One or more topics are selected from the ranked candidate topics. The selected topics are outputted via a user interface (UI).
    Type: Grant
    Filed: April 22, 2019
    Date of Patent: March 17, 2020
    Assignee: Alibaba Group Holding Limited
    Inventors: Wangshu Zhang, Zhiwei Shi, Junhong Liu, Jie Cai, Xiang Hu, Defeng Mao
  • Patent number: 10587417
    Abstract: Disclosed are a document encryption prompt method and system. The document encryption prompt method comprises: detecting text content of a target document, and if the text content of the target document comprises a preset sensitive word, determining that the document is a sensitive document; and prompting a user to encrypt the sensitive document. A document encryption prompt system comprises: a detecting unit and a prompting unit. The detecting unit is configured to detect text content of a target document, and if the text content of the target document comprises a preset sensitive word, determine that the document is a sensitive document; and the prompting unit is configured to prompt the user to encrypt the sensitive document. According to this application, a user does not need to manually determine whether to encrypt a document, and the user can be prompted in time to perform encryption processing on the document, thereby improving the security of the document.
    Type: Grant
    Filed: November 25, 2014
    Date of Patent: March 10, 2020
    Inventors: Yuntong Chao, Dizuo Ou, Zhishou Huang, Xupeng Huang, Xuewen Hong, Huan Liu, Yafei Li, Ting Fang
  • Patent number: 10586535
    Abstract: Systems and processes for operating a digital assistant are provided. In one example, a method includes receiving a first speech input from a user. The method further includes identifying context information and determining a user intent based on the first speech input and the context information. The method further includes determining whether the user intent is to perform a task using a searching process or an object managing process. The searching process is configured to search data, and the object managing process is configured to manage objects. The method further includes, in accordance with a determination the user intent is to perform the task using the searching process, performing the task using the searching process; and in accordance with the determination that the user intent is to perform the task using the object managing process, performing the task using the object managing process.
    Type: Grant
    Filed: September 21, 2016
    Date of Patent: March 10, 2020
    Assignee: Apple Inc.
    Inventors: Aram D. Kudurshian, Bronwyn Jones, Elizabeth Caroline Furches Cranfill, Harry J. Saddler
  • Patent number: 10579730
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for natural language processing. One of the methods includes receiving a first voice input from a user device; generating a first recognition output; receiving a user selection of one or more terms in the first recognition output; receiving a second voice input spelling a correction of the user selection; determining a corrected recognition output for the selected portion; and providing a second recognition output that merges the first recognition output and the corrected recognition output.
    Type: Grant
    Filed: January 25, 2019
    Date of Patent: March 3, 2020
    Assignee: Google LLC
    Inventors: Evgeny A. Cherepanov, Gleb Skobeltsyn, Jakob Foerster, Petar Aleksic, Assaf Hurwitz Michaely
  • Patent number: 10579834
    Abstract: A computer-implemented method and an apparatus facilitate customer intent prediction. The method includes receiving natural language communication provided by a customer on at least one enterprise related interaction channel. Textual data corresponding to the natural language communication is generated by converting one or more non-textual portions in the natural language communication to a text form. One or more processing operations are performed on the textual data to generate normalized text. The normalized text is configured to facilitate interpretation of the natural language communication. At least one intention of the customer is predicted, at least in part, based on the normalized text and a reply is provisioned to the customer based on the predicted intention. The reply is provisioned to the customer on the at least one enterprise related interaction channel in response to the natural language communication.
    Type: Grant
    Filed: October 26, 2016
    Date of Patent: March 3, 2020
    Assignee: [24]7.ai, Inc.
    Inventor: Anmol Walia
  • Patent number: 10572527
    Abstract: Systems and methods for improving search results from proprietary search engine technologies and proprietary machine classifiers, without ingesting, copying, or storing, the data to be searched, are described herein. A user sends a query to a proprietary search engines and gets a result document set back. The user may apply a user model for classifying a result document set to generate a result document for review of a user. The reviewed document may be added to a user training corpus, which is then used to retrain the user model. The retrained user model may be applied by the user to generate the next result document for user review and so on until the user model converges to generate relevant documents reliably. Once the user model converges, the user may apply the now reliable user model to generate multiple relevant documents for the user.
    Type: Grant
    Filed: May 31, 2019
    Date of Patent: February 25, 2020
    Assignee: RINA SYSTEMS, LLC.
    Inventors: Terence M Carr, Leo Zamansky
  • Patent number: 10573335
    Abstract: Methods, systems and apparatuses are provided to perform a continuous-to-continuous mapping of neural signal data received from one or more body sensors connected to an user wherein the one or more body sensors monitors at least neural activities of the user of a sub-vocalized voice at a sensory level and sends the neural signal data to a processor. The processor receives the neural signal data in an iterative closed loop to train the processor and to generate a sufficiently large data set in the neural signal domain to link to a produced voice domain. The processor constructs a common feature space which associates the neural signal domain with the produced voice domain wherein the common feature space implicitly extracts features related to audio communications for linking neural signal domain data to the produced voice data without requiring any prior feature classification of the received neural signal data.
    Type: Grant
    Filed: March 20, 2018
    Date of Patent: February 25, 2020
    Assignee: HONEYWELL INTERNATIONAL INC.
    Inventors: Mohammad Moghadamfalahi, Umut Orhan, Michael Dillard
  • Patent number: 10565317
    Abstract: Embodiments provide methods and apparatus for improving responses of automated conversational agents. The method includes generating a vector representation of a conversational input provided by a user. The vector representation is used to determine an intent of the conversational input. Further, annotators generate bait sentences that cover multiple aspects of the intent. Then, sentences in a data pool are accessed. The bait sentences and the data pool sentences are converted into a first and a second set of vector representations, respectively. The first and the second set of vector representations are compared to retrieve a list of similar sentences. The list of similar sentences includes one or more sentences of the data pool that are semantically similar to the bait sentences. The list of similar sentences is analyzed for updating the intent data and thereby improve the responses.
    Type: Grant
    Filed: May 7, 2019
    Date of Patent: February 18, 2020
    Assignee: Moveworks, Inc.
    Inventors: Zhan Liu, Jiang Chen
  • Patent number: 10567600
    Abstract: An image processing device includes: a displayer; and a hardware processor that obtains voice recognition data that is a voice recognition result related to a voice vocalized in a state in which at least one operation screen is displayed in the displayer, determines a search target character string on the basis of the voice recognition data, executes search processing of searching for one voice operation command that agrees with the search target character string among a plurality of voice operation commands including a voice operation command group related to a first screen related to the image processing device, and a voice operation command group related to a second screen displayed according to user's operation for the first screen, and executes processing corresponding to the one voice operation command that has been searched for by the hardware processor.
    Type: Grant
    Filed: February 11, 2019
    Date of Patent: February 18, 2020
    Assignee: Konica Minolta, Inc.
    Inventor: Hozuma Nakajima
  • Patent number: 10564991
    Abstract: A method of creating content adapted multimedia guidance for a web document, comprising: analyzing location and type of dynamic components in a web document for the creation of a flow dataset, that maps at least one trigger node representing a user initiated trigger event and at least one action node linked to the trigger node and representing instruction to present guidance related to one of the dynamic components in response to the user initiated trigger event; identifying a loading event during which the web document is loaded on a client terminal of a user; extracting at least one generated value from the dynamic components; building an audio sequence associated with the action node by selecting audio segments according to the generated value and arranging the audio segments according to the flow dataset; and when the trigger is activated, playing the audio sequence on the client terminal.
    Type: Grant
    Filed: April 27, 2016
    Date of Patent: February 18, 2020
    Assignee: Toonimo Inc.
    Inventors: Dan Kotlicki, Ohad Rozen, Edi Buslovich
  • Patent number: 10567564
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for connecting an interactive wearable device with a network. In one aspect, a method includes loading content from a playlist; recognizing contextual information relating to the content; determining the location of the user; requesting supplemental content via a network based on the contextual information and the location; displaying supplemental information to a user; interacting with the supplemental information at least in part via an interactive apparatus.
    Type: Grant
    Filed: May 30, 2018
    Date of Patent: February 18, 2020
    Assignee: Muzik, Inc.
    Inventor: Jason Hardi
  • Patent number: 10560656
    Abstract: In some implementations, a user device can be configured to create media messages with automatic titling. For example, a user can create a media messaging project that includes multiple video clips. The video clips can be generated based on video data and/or audio data captured by the user device and/or based on pre-recorded video data and/or audio data obtained from various storage locations. When the user device captures the audio data for a clip, the user device can obtain a speech-to-text transcription of the audio data in near real time and present the transcription data (e.g., text) overlaid on the video data while the video data is being captured or presented by the user device.
    Type: Grant
    Filed: March 15, 2018
    Date of Patent: February 11, 2020
    Assignee: Apple Inc.
    Inventors: Joseph-Alexander P. Weil, Andrew L. Harding, David Black, James Brasure, Joash S. Berkeley, Katherine K. Ernst, Richard Salvador, Stephen Sheeler, William D. Cummings, Xiaohuan Corina Wang, Robert L. Clark, Kevin M. O'Neil
  • Patent number: 10555106
    Abstract: Embodiments relate to a headset that filters sounds according to a direction of a gaze of a user wearing the headset. The user wears the headset including an eye tracking unit and one or more microphones. The eye tracking unit tracks an orientation of an eye of the user to determine the direction of the gaze of the user. The direction of the gaze may be different from a facing direction of the headset. According to the determined direction of the gaze of the user, input sound signals generated by the microphones can be beamformed to amplify or emphasize sound originating from the direction of the gaze.
    Type: Grant
    Filed: January 8, 2018
    Date of Patent: February 4, 2020
    Assignee: Facebook Technologies, LLC
    Inventor: Ravish Mehra
  • Patent number: 10546579
    Abstract: According to an embodiment, a verification system includes a storage controller, first and second receivers, a comparator, a response constructor, a response generator, and an output controller. The storage controller stores, in a storage, first response data and first situation data associated with the first response data. The first receiver receives second response data. The comparator determines a similarity between second situation data indicating a second context for using the second response data and the first situation data. The response constructor constructs response content information comprising the second response data and the first response data associated with the first situation data having the similarity equal to or greater than a threshold. The second receiver receives speech data. The response generator generates a response sentence corresponding to the speech data using the response content information. The output controller outputs for display one or more response sentences.
    Type: Grant
    Filed: August 28, 2017
    Date of Patent: January 28, 2020
    Assignee: KABUSHIKI KAISHA TOSHIBA
    Inventors: Takami Yoshida, Kenji Iwata, Hiroshi Fujimura
  • Patent number: 10546137
    Abstract: A method and an apparatus for protecting an application program password of a mobile terminal are described. The method includes: positioning an application interface view in an application program of a mobile terminal, so as to obtain a currently activated application interface view; determining whether the currently activated application interface view is an input box; and if yes, acquiring a key input event triggered in the input box and obtaining an input character by using the key input event; and replacing the character with a fake password character. A related apparatus is also described.
    Type: Grant
    Filed: November 15, 2016
    Date of Patent: January 28, 2020
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Zhao Hu, Xin Jiang, Hao Wu, Siwei Zhou
  • Patent number: 10540979
    Abstract: A device includes a memory, a receiver, a processor, and a display. The memory is configured to store a speaker model. The receiver is configured to receive an input audio signal. The processor is configured to determine a first confidence level associated with a first portion of the input audio signal based on the speaker model. The processor is also configured to determine a second confidence level associated with a second portion of the input audio signal based on the speaker model. The display is configured to present a graphical user interface associated with the first confidence level or associated with the second confidence level.
    Type: Grant
    Filed: April 16, 2015
    Date of Patent: January 21, 2020
    Assignee: Qualcomm Incorporated
    Inventors: Erik Visser, Lae-Hoon Kim, Minho Jin, Yinyi Guo
  • Patent number: 10535337
    Abstract: A method includes determining a content of a speech of a user on the basis of text data including a word string acquired as a recognition result of the speech, selecting a correction method suitable for the determined content of the speech from among multiple correction methods for correcting false recognition contained in the text data, and correcting the false recognition contained in the text data using the selected correction method.
    Type: Grant
    Filed: March 3, 2017
    Date of Patent: January 14, 2020
    Assignee: PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD.
    Inventors: Mayu Yokoya, Katsuyoshi Yamagami
  • Patent number: 10534623
    Abstract: A method, performed by at least one computer, the method comprising using the at least one computer to perform acts of accessing information specifying at least one user-specified condition specified by a user and at least one corresponding user-specified action, the user-specified action to be performed when the user-specified condition is met; determining whether the at least one user-specified condition is met; and when it is determined that the at least one user-specified condition is met, causing a virtual assistant executing on a mobile device different from the at least one computer to perform the at least one user-specified action.
    Type: Grant
    Filed: December 16, 2013
    Date of Patent: January 14, 2020
    Assignee: Nuance Communications, Inc.
    Inventors: Kenneth S. Harper, Fares Jaradeh, Holger Quast, Carey Radebaugh, Sean P. Brown
  • Patent number: 10535352
    Abstract: A computer-implemented method includes associating, using a processor, one or more words in an electronic agenda template to at least one agenda item indicative of a point for discussion. The processor captures a real-time interaction comprising speech from one or more participants of a plurality of discussion participants into a digital representation. The processor isolates a portion of the real-time interaction from the digital representation. The portion is associated with a single speaker of the plurality of discussion participants. The processor makes at least one match between an isolated portion of the real-time interaction and the at least one agenda item. The processor determines an intent of the single speaker from the isolated portion and matching the determined intent of the single speaker to the at least one agenda item on the electronic agenda template, and generates discussion minutes output based on the matched intent and agenda item.
    Type: Grant
    Filed: November 16, 2017
    Date of Patent: January 14, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Sharathchandra Pankanti, Stefan Ravizza, Erik Rueger
  • Patent number: 10528446
    Abstract: A mechanism is provided for delivering troubleshooting support that captures a user's interactions with a local computing device. Responsive to receiving a set of events associated with an issue being experienced by the user of the local computing device, each event in the set of events is presented to a technician. Each event in the set of events is flagged as either correct or incorrect. Responsive to the incorrect flagged event being an operation performed by the local computing device, a communicative connecting is established to the local computing device in order that the technician corrects a future operation corresponding to the operation performed. Responsive to the incorrect flagged event being an action performed by the user of the local computing device, an analysis is sent in order to inform the user of incorrect actions so as to not perform the incorrect actions in the future.
    Type: Grant
    Filed: August 7, 2017
    Date of Patent: January 7, 2020
    Assignee: International Business Machines Corporation
    Inventors: Miron Aloni, Rotem Klein, Daniel Lereya, Vladimir Shalikashvili, Moshe Weiss
  • Patent number: 10528810
    Abstract: A method to determine whether a user is experiencing difficulty visually resolving content is disclosed. The method includes capturing one or more images of the user while the user is viewing the content. The method also includes obtaining facial parameters related to a visual acuity of the user from the captured one or more images. The method further includes determining whether the user is experiencing difficulty visually resolving the content based on the obtained one or more facial parameters. The method is implemented in a device such as a smartphone, tablet computer, or television. The facial parameters include information about the extent to which the user has their eyes open or closed, whether the user is wearing glasses, and the distance at which the user is viewing the content.
    Type: Grant
    Filed: January 25, 2018
    Date of Patent: January 7, 2020
    Assignee: Samsung Electronics Co., Ltd.
    Inventor: Young Eun Cho
  • Patent number: 10529337
    Abstract: Symbol sequences are estimated using a computer-implemented method including detecting one or more candidates of a target symbol sequence from a speech-to-text data, extracting a related portion of each candidate from the speech-to-text data, detecting repetition of at least a partial sequence of each candidate within the related portion of the corresponding candidate, labeling the detected repetition with a repetition indication, and estimating whether each candidate is the target symbol sequence, using the corresponding related portion including the repetition indication of each of the candidates.
    Type: Grant
    Filed: January 7, 2019
    Date of Patent: January 7, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Kenneth W. Church, Gakuto Kurata, Bhuvana Ramabhadran, Abhinav Sethy, Masayuki Suzuki, Ryuki Tachibana
  • Patent number: 10521481
    Abstract: In one aspect, a method includes (a) accessing, by a computing system, a social media (SM) platform, (b) selecting a SM content item published on the SM platform, (c) responsive to the selection, accessing a webpage of the SM platform, (d) detecting that the webpage contains a nested link object, (e) selecting the nested link object to display the hidden SM content item(s) on the webpage, (f) responsive to selecting the nested link object, evaluating the displayed hidden SM content item(s) on the webpage to determine whether the selected SM content item is displayed on the webpage, (g) repeating acts (d)-(f) until the computing system determines that the selected SM content item is displayed on the webpage, and (h) responsive to determining that the selected SM content item is displayed on the webpage, retrieving the selected SM content item via the webpage.
    Type: Grant
    Filed: July 15, 2016
    Date of Patent: December 31, 2019
    Assignee: Tribune Broadcasting Company, LLC
    Inventors: Hank J. Hundemer, Dana Lasher
  • Patent number: 10516709
    Abstract: The present technology automatically shares materials at the start of a videoconference without requiring a participant to find the materials or instruct the videoconferencing application to share the materials. The conference materials can be automatically shared without any conference participant involvement. The present technology automatically associates materials included in a calendar invitation to the conference or in a shared space referenced in the calendar invitation. These materials can be automatically shared when the conference launches.
    Type: Grant
    Filed: June 29, 2017
    Date of Patent: December 24, 2019
    Assignee: CISCO TECHNOLOGY, INC.
    Inventors: Brian Powell, Jonathan Rosenberg, Amit Barave
  • Patent number: 10510340
    Abstract: Techniques for using a dynamic wakeword detection threshold are described. A server(s) may receive audio data corresponding to an utterance from a device in response to the device detecting a wakeword using a wakeword detection threshold. The server(s) may then determine the device should use a lower wakeword detection threshold for a duration of time. In addition to sending the device output data responsive to the utterance, the server(s) may send the device an instruction to use the lower wakeword detection threshold for the duration of time. Alternatively, the server(s) may train a machine learning model to determine when the device should use a lower wakeword detection threshold. The server(s) may send the trained machine learned model to the device for use at runtime.
    Type: Grant
    Filed: December 5, 2017
    Date of Patent: December 17, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Gengshen Fu, Shiv Naga Prasad Vitaladevuni, Paul McIntyre, Shuang Wu
  • Patent number: 10510344
    Abstract: Method and systems are provided for interpreting speech data. A method and system for recognizing speech involving a filter module to generate a set of processed audio data based on raw audio data; a translation module to provide a set of translation results for the raw audio data; and a decision module to select the text data that represents the raw audio data. A method for minimizing noise in audio signals received by a microphone array is also described. A method and system of automatic entry of data into one or more data fields involving receiving a processed audio data; and operating a processing module to: search in a trigger dictionary for a field identifier that corresponds to the trigger identifier; identify a data field associated with a data field identifier corresponding to the field identifier; and providing content data associated with the trigger identifier to the identified data field.
    Type: Grant
    Filed: December 7, 2018
    Date of Patent: December 17, 2019
    Assignee: Interdev Technologies Inc.
    Inventors: Janet M. Rice, Peng Liang, Terence W. Kuehn
  • Patent number: 10509627
    Abstract: Methods, computer program products, and systems are presented. The method computer program products, and systems can include, for instance: obtaining an audio input, the audio input representing sound emanating from a key press based user interface; generating a context pattern based on the audio input; performing classification of the context pattern to classify the context pattern as belonging to a signature pattern classification, wherein the signature pattern classification specifies a user activity; and providing an output based on the performing classification.
    Type: Grant
    Filed: November 22, 2017
    Date of Patent: December 17, 2019
    Assignee: International Business Machines Corporation
    Inventors: Tamer E. Abuelsaad, Gregory J. Boss, John E. Moore, Jr., Randy A. Rendahl
  • Patent number: 10511643
    Abstract: Disclosed techniques provide for the management of immersion levels and notifications of conference activities that are of interest to a user. In some embodiments, a system can monitor a conference session, that a user is not attending, on the user's behalf to determine when activity that is of interest to the user occurs during the conference session. The system can then send a notification to the user to inform her of the identified activity. In some embodiments, the notification may enable the user to control her immersion level with respect to the conference session. For example, the notification may include user interface element(s) that the user may select to observe or even join the conference session. In some embodiments, the notification may enable the user to directly share content with attendees of the conference session without having to observe or join the conference session.
    Type: Grant
    Filed: May 18, 2017
    Date of Patent: December 17, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Jason Thomas Faulkner, Mark Robert Swift
  • Patent number: 10509862
    Abstract: The present disclosure generally relates to dynamic phrase expansion for language input. In one example process, a user input comprising text of a first symbolic system is received. The process determines, based on the text, a plurality of sets of one or more candidate words of a second symbolic system. The process determines, based on at least a portion of the plurality of sets of one or more candidate words, a plurality of expanded candidate phrases. Each expanded candidate phrase comprises at least one word of a respective set of one or more candidate words of the plurality of sets of one or more candidate words and one or more predicted words based on the at least one word of the respective set of one or more candidate words. One or more expanded candidate phrases of the plurality of expanded candidate phrases are displayed for user selection.
    Type: Grant
    Filed: September 21, 2016
    Date of Patent: December 17, 2019
    Assignee: Apple Inc.
    Inventors: Xin Wang, Jun Hatori