Application Patents (Class 704/270)
  • Patent number: 10659898
    Abstract: A method provides binaural sound to a person through electronic earphones. The binaural sound localizes to a sound localization point (SLP) in empty space that is away from but proximate to the person. When an event occurs, the binaural sound switches or changes to stereo sound, to mono sound, or to altered binaural sound.
    Type: Grant
    Filed: November 25, 2019
    Date of Patent: May 19, 2020
    Inventors: Philip Scott Lyren, Glen A. Norris
  • Patent number: 10656602
    Abstract: A power tool accident prevention system receiving images from a static camera of a setup of a power tool, the system comprising: a processor; and a memory, the memory storing instructions to cause the processor to: analyze the images to identify inherent dangers in the setup of the power tool; identify at least one potential cause of an accident based on the identified inherent dangers; and activate an emergency safety measure of the power tool to avoid the at least one potential cause of the accident.
    Type: Grant
    Filed: February 26, 2019
    Date of Patent: May 19, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Vittorio Castelli, Radu Florian, Gourab Kundu, Taesun Moon, Avirup Sil
  • Patent number: 10650822
    Abstract: A server controlling an external device is provided. The server includes a communicator; a processor; a memory which stores at least one natural language understanding (NLU) engine for generating a command corresponding to a user's utterance. The server receives, from a pairing device paired to the external device, the user's utterance controlling the external device and information about at least one external device registered with the pairing device, via the communicator, determines an NLU engine corresponding to the external device, from among the at least one NLU engine, based on the user's utterance controlling the external device and the information about the at least one external device, and generates the command controlling the external device based on the user's utterance, by using the determined NLU engine.
    Type: Grant
    Filed: August 29, 2017
    Date of Patent: May 12, 2020
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventor: Dong-hyun Choi
  • Patent number: 10650603
    Abstract: Controlling a mixed reality (MR), virtual reality (VR), or augmented reality (AR) (collectively, MR) environment visualization may involve obtaining a plurality of sensor data from a plurality of data sources; processing the obtained plurality of sensor data using a plurality of data analyzers to identify at least one feature; generating a plurality of annotated data sets, wherein the annotated data sets contain an annotation of the at least one feature; aggregating the plurality of annotated data sets to correlate the at least one feature as a common feature across different annotated data sets of the plurality of annotated data sets; and based at least on the common feature, providing a visualization output representing the MR (VR, AR, or specifically mixed-reality) environment. Disclosed examples enable utilizing of disparate data types from different data sources, localizing the different data a common space, and aggregating results for visualization and/or further analysis.
    Type: Grant
    Filed: August 9, 2018
    Date of Patent: May 12, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Anthony Ernst, George O. Johnston, Valeriy Trofimov, Gregory Smith
  • Patent number: 10645224
    Abstract: A system and method of categorizing communications involves evaluating one or more portions of a communication. Systems include a language module structured to analyze the one or more portions of the communication for at least one of a language characteristic or a non-word symbol within a text of the communication, an acoustic analysis module structured to analyze the one or more portions of the communication for an acoustic characteristic, and a determination module structured to determine at least one category label to associate with the one or more portions of the communication using at least one of the language characteristic or the non-word symbol, and the acoustic characteristic.
    Type: Grant
    Filed: April 17, 2019
    Date of Patent: May 5, 2020
    Assignee: CallMiner, Inc.
    Inventors: Michael C. Dwyer, Erik A. Strand, Scott R. Wolf, Frank Salinas, Jeffrey A. Gallino, Scott A. Kendrick, Shaoyu Xue
  • Patent number: 10645494
    Abstract: An active control system of a vehicular dual microphone and a control method thereof control a beamforming area of a microphone installed adjacent to each of a driver seat and a passenger seat to be directed toward a person (utterer) who utters a voice command. The control system and method use sitting height information and seat position information of the driver or the passenger and use passenger seat riding information. The control system determines output from a mouth shape change detection unit and the user and seat position detection unit, controls microphone beamforming of the driver seat according to the determined microphone beamforming control mode, and selectively adjusts a specific microphone gain according to the determined microphone beamforming control mode to separately control microphone beamforming of the driver seat and the passenger seat.
    Type: Grant
    Filed: June 13, 2019
    Date of Patent: May 5, 2020
    Assignees: HYUNDAI MOTOR COMPANY, KIA MOTORS CORPORATION
    Inventors: Hyung Suk Oh, Hong Ju Kim
  • Patent number: 10643164
    Abstract: Computer program products for performing workflows using a mobile device, without requiring tactile input from the user. The workflow is thus “touchless” and may utilize input preferably including optical data and/or audio data. Tactile input may be optionally provided in some embodiments, but the workflow is configured to be performed without any tactile input. Accordingly, in one embodiment, a computer program product includes a computer readable medium having computer readable and/or executable program instructions embodied therewith, the program instructions being configured to cause a processor to: invoke a mobile application using a mobile device; receive auditory input via the mobile device; and perform a business workflow via the mobile application based on the auditory input. Corresponding systems and computer program product embodiments configured for touchless mobile workflows are also described.
    Type: Grant
    Filed: March 2, 2018
    Date of Patent: May 5, 2020
    Assignee: KOFAX, INC.
    Inventors: Anthony Macciola, Jan W. Amtrup, Jiyong Ma
  • Patent number: 10638206
    Abstract: A customized video stream is provided. A second content is searched for that includes the plurality of entities, and a plurality of user data associated with the social media account. A hierarchy of additional content is generated based, at least in part, on the second content that is associated with the plurality of entities, and the user data associated with the social media account, wherein the hierarchy of additional content contains a plurality of second content. A customized video stream is generated based, at least in part, on the video stream and the hierarchy of additional content. The customized video stream is populated on a display screen of the user device, wherein the video stream is displayed on the display screen of the user device, and the hierarchy of additional content is displayed in a portion of the display screen and displayed concurrently with the video stream.
    Type: Grant
    Filed: January 28, 2019
    Date of Patent: April 28, 2020
    Assignee: International Business Machines Corporation
    Inventors: Gray Cannon, Marco Aurelio Stelmar Netto, Vagner Figueredo De Santana
  • Patent number: 10629200
    Abstract: A method for performing voice analysis includes storing a simulation file having a script for conducting a training session with a user. The method also stores desired attributes associated with the simulation file. The method also includes retrieving the simulation file and providing a user interface to conduct the voice analysis using the simulation file. The method further includes receiving one or more voice impressions from a user and analyzing at least one of the voice impressions of the user. The method additionally includes determining attributes of the at least one voice impression in response to analyzing the at least one voice impression and comparing the determined attributes to the desired attributes associated with the simulation file. The method provides feedback to the user based on the comparison.
    Type: Grant
    Filed: March 7, 2018
    Date of Patent: April 21, 2020
    Assignee: SalesBoost, LLC
    Inventor: Margaret L Brooks
  • Patent number: 10628191
    Abstract: A system permitting virtual server images to be reused and repurposed to provide for different application environments for performance-based testing by providing a plurality of virtual hard drive images.
    Type: Grant
    Filed: June 15, 2018
    Date of Patent: April 21, 2020
    Assignee: Prometric LLC
    Inventor: Christopher Crowhurst
  • Patent number: 10630838
    Abstract: A system and method for engaging in an automated dialog with a user. A processor retrieves a preset dialog flow that includes various blocks directing the dialog with the user. The processor provides a prompt to the user based on a current block of the dialog flow, receives an action from the user in response to the prompt, and retrieves a classification/decision tree corresponding to the dialog flow. The classification tree has a plurality of nodes mapped to the blocks of the dialog flow. Each of the nodes represents a user intent. The processor computes a probability for each of the nodes based on the action from the user. A particular one of the nodes is then selected based on the computed probabilities. A target block of the dialog flow is further identified based on the selected node, and a response is output in response to the identified target block.
    Type: Grant
    Filed: May 22, 2018
    Date of Patent: April 21, 2020
    Inventors: Conor McGann, Ioana Grigoropol, Mariya Orshansky, Ankit Pat
  • Patent number: 10628635
    Abstract: A conversational holographic assistant displays a computer-generated human-like hologram capable of carrying on conversations in real-time with a user, remembering things about past interactions with the user to make the conversation realistic and more human-like. The system is intended to be utilized in retail, office, business and personal settings to provide service to users. The holographic person can interact with retail point of sale systems in order to process orders on behalf of a user, provide users with detailed directions, and answer questions about the holographic person's surroundings, among other things. In some settings, the holographic person can integrate with existing internet of things devices.
    Type: Grant
    Filed: October 3, 2017
    Date of Patent: April 21, 2020
    Assignee: Valyant AI, Inc.
    Inventors: Robley Theron Carpenter, II, Benjamin William Thielker
  • Patent number: 10621421
    Abstract: A computer system that performs in dialogue with a user and provides a prescribed service, comprising: an imaging device; a computer; and a generation device generating dialogue content on a basis of an algorithm for generating dialogue content. The computer couples to a database that stores an authentication image used for an authentication process that uses an image. The computer calculates a distance between the user and the imaging device; executes an attribute estimation process in a case where the distance is larger than a threshold, selects the algorithm on the basis of results of the attribute estimation process, and issues a notification of the selected algorithm to the generation device.
    Type: Grant
    Filed: June 7, 2018
    Date of Patent: April 14, 2020
    Assignee: Hitachi, Ltd.
    Inventors: Yuto Komatsu, Yoshitaka Hiramatsu
  • Patent number: 10623935
    Abstract: The present storage management system, described in detail in the following specification and accompanying drawings, includes an active (self-powered) Bluetooth® BLE transceiver beacon (hereinafter referred to as “beacon”), which, according to a preferred embodiment has a thin flat form factor with a self-adhesive-backing. A beacon is secured to each storage box and continuously transmits an “advertisement packet” to the surrounding area. A user's smart device (e.g., a smart phone) picks up the different signals sent by each beacon and is able to communicate with select ones. A software application allows a user to catalog items to be stored so that each is effectively “linked” to a particular storage box. At a later time, any of the catalogued items may be selected and the corresponding storage container in which the item resides is instructed to notify its location within the storage area using light and sound.
    Type: Grant
    Filed: April 17, 2018
    Date of Patent: April 14, 2020
    Inventors: Phillip Lucas Williams, Scott Sullivan
  • Patent number: 10621984
    Abstract: Implementations relate to generating and/or executing a customized interactive dialog application. The customized interactive dialog application may be generated from a state mapping tool that allows a user to generate custom states and custom transitions between the custom states. A customized configuration description is then generated based on the generated custom states and custom transitions. Further, a default configuration description is identified that includes additional or alternative states and transitions. In executing the customized interactive dialog application, dialog turns are generated based on the states and transition information, with the customized configuration description taking precedence and the default configuration description being utilized for any undefined states and/or transitions. Implementations additionally or alternatively relate to generating and/or executing a custom agent based on generated custom states and custom transitions, and a default configuration description.
    Type: Grant
    Filed: April 23, 2018
    Date of Patent: April 14, 2020
    Assignee: GOOGLE LLC
    Inventors: Uri First, Yang Sun
  • Patent number: 10621978
    Abstract: Systems, methods and platforms for generating and improving conversational dialogs between a computer system and a user to create more intelligent, realistic human-like conversations. The systems, methods and platforms achieve a realistic human conversation by identify parameters such as the user's culture, age and known interests, identify a corpus that contains the knowledge appropriate for achieving conversational goals to a perceived topic of interest and generate expressions into dialog that may be extracted from human generated sources. The system appropriately matches the targeted user's parameters and topic based on the input of the user initiating the dialog interaction. The system may analyze and/or record the emotion of the user, cognitive context, gestures and the user's expressions, which may be performed using one or more recording systems to increase the accuracy of topic identification and the appropriate dialog responses.
    Type: Grant
    Filed: November 22, 2017
    Date of Patent: April 14, 2020
    Assignee: International Business Machines Corporation
    Inventors: Edgar Adolfo Zamora Duran, Franz Friedrich Liebinger Portela, Yanil Zeledon, Roxana Monge Nunez
  • Patent number: 10614492
    Abstract: Systems and methods are provided for notifying a wireless communication device user regarding an account payment. One method comprises dialing a phone number of a wireless communication device, preferably a cell phone. A communication relating to an account payment that includes a balance due is then transmitted or delivered to a wireless communication device user. Optionally, at the same time or a later time, input from the wireless communication device user may be received regarding payment of the balance due. Such notification methods may be carried out by a host computer system which has a record of when the account payment is due and dials and transmits within a predetermined time frame. Generally, such methods will further comprise pre-registering or enrolling the wireless communication device user.
    Type: Grant
    Filed: August 18, 2010
    Date of Patent: April 7, 2020
    Assignees: First Data Corporation, The Western Union Company
    Inventors: Patricia A. Frazier, Kenneth Algiene
  • Patent number: 10614099
    Abstract: Systems and methods are provided for populating user information onto an electronic form using human interactions via touch, voice, gestures or an input device. The electronic form is selected by the user for completion using a user profile of stored data. When a form field requires a manual input—such as a form field with multiple potential values—a user is prompted to complete the field using one or more of the human interactions to allow the user to easily complete the field. These human interactions may include touching the form field with a finger on a touchscreen user interface, speaking the form field name, gesturing or selecting via the input device to generate a window of different potential values, and then touching, speaking, gesturing of selecting via the input device the value that the user prefers.
    Type: Grant
    Filed: October 23, 2013
    Date of Patent: April 7, 2020
    Assignee: UBIQ Security, Inc.
    Inventors: Linda Eigner, William Eigner, Eric Tobias, Charles Kahle, Anthony F. Iasi
  • Patent number: 10614804
    Abstract: A voice controlled room automation system that includes a speaker device situated in a guest room, a hotel automation controller operatively coupled to one or more components in the guest room, and a web service operatively coupled to the speaker device and the hotel automation controller. The web service is configured to receiving voice commands from the speaker device, process the voice command using speech recognition, interpret the voice command to determine a corresponding command for the hotel automation controller, and transmit the corresponding command to the hotel automation controller. The hotel automation controller is configured to receive the corresponding command from the web service, and to carry out the corresponding command by interacting with one or more of the components in the room. In some cases, the hotel automation controller may be configured to initiate an announcement or query on the speaker device via the web service.
    Type: Grant
    Filed: January 23, 2018
    Date of Patent: April 7, 2020
    Assignee: Honeywell International Inc.
    Inventors: Soumitri Kolavennu, Ryan Gardner, Michael R. Sobanko, Scott Knecht, Arun Mahasenan
  • Patent number: 10614729
    Abstract: A method and system for accessing information related to a physical document. An electronic copy of an existing physical document is identified and located. The electronic copy of the physical document is an exact replica of the physical document. A selected part of the physical document is identified and in response, data related to the selected part of the physical document is retrieved from the electronic copy of the physical document. The retrieved data is presented visually to a visually impaired person or orally to a blind person on the user system, which enables the visually impaired person to see or hear, respectively, the retrieved data.
    Type: Grant
    Filed: February 21, 2019
    Date of Patent: April 7, 2020
    Assignee: International Business Machines Corporation
    Inventors: Fernando Incertis Carro, Sharon M. Trewin
  • Patent number: 10616701
    Abstract: A method of determining whether a sound has been generated by a loudspeaker comprises receiving an audio signal representing at least a part of the sound. The audio signal is separated into different frequency bands. The signal content of different frequency bands are compared. Based on said comparison, frequency-based variations in signal content indicative of use of a loudspeaker are identified.
    Type: Grant
    Filed: November 13, 2018
    Date of Patent: April 7, 2020
    Assignee: Cirrus Logic, Inc.
    Inventor: John Paul Lesso
  • Patent number: 10609073
    Abstract: Classifying electronic communications is disclosed. An electronic communication is received. A first likelihood that a potential recipient of the electronic communication would conclude that the communication was transmitted on behalf of an authoritative entity is determined. An assessment of a second likelihood that the received communication was transmitted with authorization from the purported authoritative entity is performed. The electronic communication is classified based at least in part on the first and second likelihoods.
    Type: Grant
    Filed: February 11, 2019
    Date of Patent: March 31, 2020
    Assignee: ZAPFRAUD, INC.
    Inventor: Bjorn Markus Jakobsson
  • Patent number: 10607633
    Abstract: In accordance with an example embodiment of the present invention, disclosed is a method and an apparatus for voice activity detection (VAD). The VAD comprises creating a signal indicative of a primary VAD decision and determining hangover addition. The determination on hangover addition is made in dependence of a short term activity measure and/or a long term activity measure. A signal indicative of a final VAD decision is then created.
    Type: Grant
    Filed: June 7, 2018
    Date of Patent: March 31, 2020
    Assignee: Telefonaktiebolaget LM Ericsson (publ)
    Inventor: Martin Sehlstedt
  • Patent number: 10599393
    Abstract: The subject disclosure relates to user input into a computer system, and a technology by which one or more users interact with a computer system via a combination of input modalities. When the input data of two or more input modalities are related, they are combined to interpret an intended meaning of the input. For example, speech when combined with one input gesture has one intended meaning, e.g., convert the speech to verbatim text for consumption by a program, while the exact speech when combined with a different input gesture has a different meaning, e.g., convert the speech to a command that controls the operation of that same program.
    Type: Grant
    Filed: August 1, 2018
    Date of Patent: March 24, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Oscar E. Murillo, Janet E. Galore, Jonathan C. Cluts, Colleen G. Estrada, Michael Koenig, Jack Creasey, Subha Bhattacharyay
  • Patent number: 10598504
    Abstract: The present disclosure relates to a vehicle control device for a vehicle having a voice input unit, an audio output unit, and a display. The vehicle control device may include a communication unit configured to communicate with at least one of the voice input unit, the audio output unit, and the display, and a processor configured to control the communication unit to output a message for inducing an utterance of a passenger aboard the vehicle in at least one of an auditory manner and a visual manner when a preset event occurs, and identify the passenger based on receiving a response of the passenger to the message through the voice input unit, and perform a different function based on the identified passenger.
    Type: Grant
    Filed: November 29, 2017
    Date of Patent: March 24, 2020
    Assignee: LG Electronics Inc.
    Inventors: Jiwon Oh, Jinhua Piao, Eunjoung Cho
  • Patent number: 10593423
    Abstract: Mechanisms are provided to implement a natural language request processing engine (NLRPE). The NRLPE performs natural language processing on a portion of unstructured text in an electronic data structure to generate textual characteristics of the portion of unstructured text. The NRLPE annotates at least one phrase in the portion of unstructured text at least by linking the at least one phrase to one or more concepts specified in at least one ontological data structure based on the textual characteristics of the portion of unstructured text. The NRLPE generates a model of the portion of unstructured text based on the one or more concepts linked to the at least one phrase. The NRLPE processes a request for information specifying a concept of interest based on the model of the portion of unstructured text by retrieving the at least one phrase or the at least one merged phrase as a response.
    Type: Grant
    Filed: December 28, 2017
    Date of Patent: March 17, 2020
    Assignee: International Business Machines Corporation
    Inventors: Tyler Baldwin, Yufan Guo
  • Patent number: 10591444
    Abstract: A system for detecting rodent activity in an area. A series of detectors that calibrated to detect rodent vocalizations are placed in a geographic area. One or more of the detectors detect rodent vocalizations and transmits data notifying a central computer of detected vocalizations. The computer is provided with a program that generates a report of incidences and locations of detected vocalizations.
    Type: Grant
    Filed: October 8, 2019
    Date of Patent: March 17, 2020
    Assignee: AP&G Co., Inc.
    Inventors: Jeffrey Frisch, Jonathan Frisch
  • Patent number: 10587594
    Abstract: A service provider or service of the service provider may generate authentication information based on information provided by a user. The information provided by the user may include a selection of a media object, information indicating a segment of the media object, and an indication of an action to be performed by the user. This information may be saved by the service provider and used at some point in time later to determine whether to provide access to restricted computing resources based at least in part on an authentication claim provided by the user. The authentication claim may be generated by a computing device operated by the user and may include information corresponding to the media object, the indicated segment of the media object, and the action performed by the user.
    Type: Grant
    Filed: September 23, 2014
    Date of Patent: March 10, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Jon Arron McClintock, Darren Ernest Canavor, Jesper Mikael Johansson, George Nikolaos Stathakopoulos
  • Patent number: 10588017
    Abstract: A system for implicit authentication for a mobile device associated with a user, wherein the implicit authentication is behavioural, biometric and task-based and includes at least one authentication task selected so as to leverage the user's muscle memory. The mobile device comprises a touchscreen; a transaction authentication information unit; one or more sensors coupled to the transaction authentication information unit; and an anomaly detector coupled to the transaction authentication information unit. The sensors comprise one or more touchscreen sensors coupled to the touchscreen, an accelerometer, and a gyroscope, and are used to obtain and transmit one or more sets of data to the transaction authentication information unit. The sets of data are associated with one or more performances of the authentication task by the user. The anomaly detector generates an authentication model using the one or more data sets transmitted to the transaction authentication information unit.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: March 10, 2020
    Assignee: Zighra Inc.
    Inventors: Deepak Chandra Dutt, Anil Buntwal Somayaji, Michael John Kendal Bingham
  • Patent number: 10582056
    Abstract: A communication channel customer journey may involve receiving a first communication on a communication channel of a plurality of communication channels, wherein the first communication is associated with a contact, receiving at least one additional communication from the contact, wherein the at least one additional communication is on a same communication channel or is on a different communication channel of the plurality of communication channels, linking the first communication and the at least one additional communication by the contact, wherein linking involves storing information from at least one of a recording, a transcript, a copy of one or more of the first communication and the at least one additional communication, a language characteristic, an acoustic characteristic, a category, and a score, and enabling a user to navigate between all of the stored information for the first communication and the at least one additional communication.
    Type: Grant
    Filed: April 17, 2019
    Date of Patent: March 3, 2020
    Assignee: CallMiner, Inc.
    Inventors: Michael C. Dwyer, Erik A. Strand, Scott R. Wolf, Frank Salinas, Jeffrey A. Gallino, Scott A. Kendrick, Shaoyu Xue
  • Patent number: 10579921
    Abstract: Aspects of the disclosure generally relate to computing devices and may be generally directed to devices, systems, methods, and/or applications for learning conversations among two or more conversation participants, storing this knowledge in a knowledgebase (i.e. neural network, graph, sequences, etc.), and enabling a user to simulate a conversation with an artificially intelligent conversation participant.
    Type: Grant
    Filed: January 8, 2017
    Date of Patent: March 3, 2020
    Inventor: Jasmin Cosic
  • Patent number: 10580435
    Abstract: Monitoring and analysis of a user's speech to detect symptoms of a mental health disorder by continuously monitoring a user's speech in real-time to generate audio data based, transcribing the audio data to text and analyzing the text of the audio data to determine a sentiment of the audio data is disclosed. A trained machine learning model may be applied to correlate the text and the determined sentiment to clinical information associated with symptoms of a mental health disorder to determine whether the symptoms are a symptom event. The initial determination may be transmitted to a second device to determine (and/or verify) whether or not the symptom event was falsely recognized. The trained machine learning model may be updated based on a response from the second device.
    Type: Grant
    Filed: November 27, 2017
    Date of Patent: March 3, 2020
    Assignee: International Business Machines Corporation
    Inventors: Maryam Ashoori, Benjamin D. Briggs, Lawrence A. Clevenger, Leigh Anne H. Clevenger
  • Patent number: 10572220
    Abstract: Systems and methods to control a controller, such as a baseboard management controller (BMC), and its host computer with voice. The system may include a mobile device, which is communicatively connected to the controller via a network. The mobile device may receive a voice input, and generate a corresponding instruction by performing voice recognition to the voice input. The instruction may be sent to the controller for controlling the controller or its host computer. For example, the controller may perform a corresponding operation based on the instruction to determine whether the instruction is for the controller or for the host computer. When the instruction is for the host computer, in response to determining the target as the host computing device, the controller may perform keyboard-video-mouse (KVM) redirection to redirect the instruction to the host computer.
    Type: Grant
    Filed: April 12, 2017
    Date of Patent: February 25, 2020
    Assignee: AMERICAN MEGATRENDS INTERNATIONAL, LLC
    Inventors: Chandrasekar Rathineswaran, Viswanathan Swaminathan
  • Patent number: 10558701
    Abstract: Disclosed embodiments provide techniques for suggesting images that can be conveniently included in electronic communication such as social media posts, emails, text messages, and other forms of electronic communication, based on voice input and/or emotion analysis from facial expressions of a user. A speech-to-text process converts a user's spoken utterance to text. Natural language processing performs an entity extraction on the utterance to generate a scenario summary. An image database is queried using the scenario summary and/or emotion analysis based on user facial images. A mapping rule library may be used to perform the image database query. Based on the scenario summary and/or the emotion analysis, images are presented to the user as recommended for including in an electronic communication message such as a social media post, email, or text message.
    Type: Grant
    Filed: February 8, 2017
    Date of Patent: February 11, 2020
    Assignee: International Business Machines Corporation
    Inventors: Jin Long L L He, Yu Hong Y L Li, Xiao Rui Shao, Fei F. Shen, Jian Jun Wang
  • Patent number: 10559309
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for collaboration between multiple voice controlled devices are disclosed. In one aspect, a method includes the actions of identifying, by a first computing device, a second computing device that is configured to respond to a particular, predefined hotword; receiving audio data that corresponds to an utterance; receiving a transcription of additional audio data outputted by the second computing device in response to the utterance; based on the transcription of the additional audio data and based on the utterance, generating a transcription that corresponds to a response to the additional audio data; and providing, for output, the transcription that corresponds to the response.
    Type: Grant
    Filed: December 22, 2016
    Date of Patent: February 11, 2020
    Assignee: Google LLC
    Inventors: Victor Carbune, Pedro Gonnet Anders, Thomas Deselaers, Sandro Feuz
  • Patent number: 10552543
    Abstract: A computer natural language conversational agent authors an event-processing rule by carrying out a dialog in natural language with a user. A data model that customizes a dialog and building of the event-processing rule is received. A partial tree data structure is constructed based on a rule's grammar, and specialized based on tokens extracted from the data model. An utterance is received from a user and interpreted according to the grammar as specialized to the data model. Based on the interpreting of the utterance, the grammar, the data model, and context of interactions with the user, a natural language prompt is determined for the computer natural language conversational agent to output to the user. The partial tree data structure is filled based on the natural language prompt and the utterance from the user. The event-processing rule is generated based on the partial tree data structure filled during the dialog.
    Type: Grant
    Filed: May 10, 2017
    Date of Patent: February 4, 2020
    Assignee: International Business Machines Corporation
    Inventors: Martin J. Hirzel, Avraham E. Shinnar, Jerome Simeon
  • Patent number: 10546236
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for training a deep neural network. One of the methods includes generating a plurality of feature vectors that each model a different portion of an audio waveform, generating a first posterior probability vector for a first feature vector using a first neural network, determining whether one of the scores in the first posterior probability vector satisfies a first threshold value, generating a second posterior probability vector for each subsequent feature vector using a second neural network, wherein the second neural network is trained to identify the same key words and key phrases and includes more inner layer nodes than the first neural network, and determining whether one of the scores in the second posterior probability vector satisfies a second threshold value.
    Type: Grant
    Filed: September 9, 2016
    Date of Patent: January 28, 2020
    Assignee: Google LLC
    Inventor: Alexander H. Gruenstein
  • Patent number: 10542144
    Abstract: An example electronic apparatus includes a display; and a processor, wherein the processor may be configured to obtain a content including at least one object; distinguish the at least one object within the content; display an instruction text in correspondence with a non-text object among the at least one object; and select the non-text object corresponding to the instruction text if a voice command corresponding to the instruction text is inputted.
    Type: Grant
    Filed: February 23, 2017
    Date of Patent: January 21, 2020
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Kwang-Youn Kim, Won-Nam Jang
  • Patent number: 10542145
    Abstract: A system and method of navigating an automated telephone answering system (TAS) may include recording incoming and outgoing portions of a call with the TAS, separating the recording into a series of prompt-response segments, customizing at least one response, and storing the segments as a script for use in subsequent calls to the TAS.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: January 21, 2020
    Assignee: CREATIVE EMPOWERING SOLUTIONS, LLC
    Inventors: Susan Cheek, Shawn Storie
  • Patent number: 10542480
    Abstract: A computer-implemented method includes: monitoring, by a user device, calling activity on the user device; detecting, by the user device and based on the monitoring, that a call has started on the user device; providing, by the user device, a pause instruction to an assistant device based on detecting that the call has started on the user device, causing the assistant device to disable speech response functions; detecting, by the user device and based on the monitoring, that the call has ended on the user device; and providing, by the user device, a resume instruction to the assistant device based on detecting that the call has ended on the user device, causing the assistant device to resume speech response functions.
    Type: Grant
    Filed: May 17, 2019
    Date of Patent: January 21, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Lisa Seacat Deluca, Jeremy A. Greenberger
  • Patent number: 10535343
    Abstract: A method at an electronic device with an audio input system includes: receiving a verbal input at the device; processing the verbal input; transmitting a request to a remote system, the request including information determined based on the verbal input; receiving a response to the request, wherein the response is generated by the remote system in accordance with the information based on the verbal input; and performing an operation in accordance with the response, where one or more of the receiving, processing, transmitting, receiving, and performing are performed by one or more voice processing modules of a voice assistant library executing on the electronic device, the voice processing modules providing a plurality of voice processing operations that are accessible to one or more application programs and/or operating software executing or executable on the electronic device.
    Type: Grant
    Filed: May 10, 2017
    Date of Patent: January 14, 2020
    Assignee: GOOGLE LLC
    Inventor: Kenneth Mixter
  • Patent number: 10530719
    Abstract: A computing device includes an interface configured to interface and communicate with a communication system, a memory that stores operational instructions, and processing circuitry operably coupled to the interface and to the memory that is configured to execute the operational instructions to perform various operations. The computing device processes a message that is provided from a sender and is intended for a recipient associated with another computing device in accordance with topic, emotive content, and/or social content to generate a classification model for the message that includes classification parameter value(s). When appropriate to perform message transformation, the computing device selects a tonal transformation based on the classification parameter value(s) and processes the message in accordance with the tonal transformation to generate a normalized message.
    Type: Grant
    Filed: November 16, 2017
    Date of Patent: January 7, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Kelley Anders, Jeremy R. Fox, Liam S. Harpur, Jonathan Dunne
  • Patent number: 10522150
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for collaboration between multiple voice controlled devices are disclosed. In one aspect, a method includes the actions of identifying, by a first computing device, a second computing device that is configured to respond to a particular, predefined hotword; receiving audio data that corresponds to an utterance; receiving a transcription of additional audio data outputted by the second computing device in response to the utterance; based on the transcription of the additional audio data and based on the utterance, generating a transcription that corresponds to a response to the additional audio data; and providing, for output, the transcription that corresponds to the response.
    Type: Grant
    Filed: December 22, 2016
    Date of Patent: December 31, 2019
    Assignee: Google LLC
    Inventors: Victor Carbune, Pedro Gonnet Anders, Thomas Deselaers, Sandro Feuz
  • Patent number: 10516939
    Abstract: Described is a way to steer speaker array or microphone array based on direction outputs of tiny light sensors attached to TV viewers, noisy-environment workers, or AR/VR system users. More specifically, a light projector is disposed on a ceiling, which sends out different sequential on/off signals for each pixel. Two light sensors are attached to each speaker array or microphone array, and one or more light sensors are attached to each user. Because each projector pixel corresponds to a specific direction, when a light sensor receives sequential signal from the projector, the light sensor can determine its direction corresponding to the projector and report that to the central station. With the speaker/microphone array direction and user direction known, the system can generate proper phase shifts for different speaker signals and generate directional sound for each individual. Similarly, the central station can determine phase shifts for combining audios from different microphones.
    Type: Grant
    Filed: July 3, 2018
    Date of Patent: December 24, 2019
    Assignee: FUJI XEROX CO., LTD.
    Inventors: Qiong Liu, Donald Kimber, Shang Ma
  • Patent number: 10515637
    Abstract: Techniques for dynamically maintaining speech processing data on a local device for frequently input commands are described. A system determines a usage history associated with a user profile. The usage history represents at least a first command. The system determines the first command is associated with an input frequency that satisfies an input frequently threshold. The system also determines the first command is missing from first speech processing data stored by a device associated with the user profile. The system then generates second speech processing data specific to the first command and sends the second speech processing data to the device.
    Type: Grant
    Filed: September 19, 2017
    Date of Patent: December 24, 2019
    Assignee: AMAZON TECHNOLOGIES, INC.
    Inventors: David William Devries, Rajesh Mittal
  • Patent number: 10516938
    Abstract: System and method for assessing speaker spatial orientation are provided. For example, audio data, as well as input from other sensors, may be analyzed to assess speaker spatial orientation. For example, the audio data may be analyzed to determine that two speakers are engaged in conversation. relative direction of one speaker with respect to the other may be obtained. Spatial orientation of at least one of the speakers may be obtained. The spatial orientation may be assessed according to the relative direction and the determination that the two speakers are engaged in conversation. Feedbacks and reports may be provided based on the assessed speaker spatial orientation.
    Type: Grant
    Filed: July 16, 2017
    Date of Patent: December 24, 2019
    Inventor: Ron Zass
  • Patent number: 10515301
    Abstract: Conversion of a large-footprint DNN to a small-print DNN is performed using a variety of techniques, including split-vector quantization. The small-foot print DNN may be distributed to a variety of devices, including mobile devices. Further, the small-footprint DNN may aid a digital assistant on a device in interpreting speech input.
    Type: Grant
    Filed: January 19, 2016
    Date of Patent: December 24, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Jinyu Li, Yifan Gong, Yongqiang Wang
  • Patent number: 10509626
    Abstract: Systems and methods disclosed herein include (i) receiving a voice command via a networked microphone device configured to receive voice commands for a media playback system, wherein the media playback system comprises a primary media playback device configured to perform a function and/or provide a first feedback in response to voice commands received via the networked microphone device, (ii) determining that the primary media playback device is not available to perform a function and/or provide the first feedback in response to the voice command, and (iii) in response to determining that the primary media playback device is not available to perform a function and/or provide the first feedback in response to the voice command, instructing a fallback device to perform a function and/or provide a second feedback in response to the voice command.
    Type: Grant
    Filed: April 18, 2016
    Date of Patent: December 17, 2019
    Inventors: Mark Plagge, Simon Jarvis, Christopher Butts
  • Patent number: 10510337
    Abstract: A method on a mobile device for voice recognition training is described. A voice training mode is entered. A voice training sample for a user of the mobile device is recorded. The voice training mode is interrupted to enter a noise indicator mode based on a sample background noise level for the voice training sample and a sample background noise type for the voice training sample. The voice training mode is returned to from the noise indicator mode when the user provides a continuation input that indicates a current background noise level meets an indicator threshold value.
    Type: Grant
    Filed: April 10, 2018
    Date of Patent: December 17, 2019
    Assignee: Google LLC
    Inventors: Michael E. Gunn, Boris Bekkerman, Mark A. Jasiuk, Pratik M. Kamdar, Jeffrey A. Sierawski
  • Patent number: 10503468
    Abstract: Systems and methods for voice control of computing devices are disclosed. Applications may be downloaded and/or accessed by a device having a display, and content associated with the applications may be displayed. Many applications do not allow for voice commands to be utilized to interact with the displayed content. Improvements described herein allow for non-voice-enabled applications to utilize voice commands to interact with displayed content by determining screen data displayed by the device and utilizing the screen data to determine an intent associated with the application. Directive data to perform an action corresponding to the intent may be sent to the device and may be utilized to perform the action on an object associated with the displayed content.
    Type: Grant
    Filed: December 8, 2017
    Date of Patent: December 10, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Manikandan Thangarathnam, Suraj Gopalakrishnan