Abstract: A method for analyzing a voice sample of a subject to determine a cardiac condition, for example, an arrhythmic condition, comprising extracting at least one voice feature from the voice sample, detecting an effect of the cardiac condition on the at least one voice feature and determining the cardiac condition based on the effect. Disclosed is also a system for determining the cardiac condition in the voice sample provided by the subject. Related apparatus and methods are also described.
Type:
Grant
Filed:
February 12, 2018
Date of Patent:
July 26, 2022
Assignee:
CardioKol Ltd.
Inventors:
Yirmiyahu Hauptman, Alon Goren, Eli Attar, Pinhas Sabach
Abstract: The present specification relates to a smart controlling device capable of utilizing machine learning for voice recognition and a method of controlling therefor. The smart controlling device according to the present invention includes a receiver configured to receive an input including a command trigger, and a controller configured to detect one or more external display devices, select a display device of the detected one or more external display devices, cause a power status of the selected display device to be changed to a first state, and cause a response data corresponding to a first command data received after the command trigger to be output on a display of the selected display device.
Abstract: Example techniques involve suppressing a wake word response to a local wake word. An example implementation involves a playback device receiving audio content for playback by the playback device and providing a sound data stream representing the received audio content to a voice assistant service (VAS) wake-word engine and a local keyword engine. The playback device plays back a first portion of the audio content and detects, via the local keyword engine, that a second portion of the received audio content includes sound data matching one or more particular local keywords. Before the second portion of the received audio content is played back, the playback device disables a local keyword response of the local keyword engine to the one or more particular local keywords and then plays back the second portion of the audio content via one or more speakers.
Abstract: The present specification relates to a smart controlling device capable of utilizing machine learning for voice recognition and a method of controlling therefor. The smart controlling device according to the present invention includes a receiver configured to receive an input including a command trigger, and a controller configured to detect one or more external display devices, select a display device of the detected one or more external display devices, cause a power status of the selected display device to be changed to a first state, and cause a response data corresponding to a first command data received after the command trigger to be output on a display of the selected display device.
Abstract: The present disclosure provides a voice recognition method and a voice recognition apparatus. The voice recognition method includes the following operations of acquiring an identification result of an operator; acquiring an acoustic feature set corresponding to the operator based on the identification result of the operator; and recognizing voice of the operator from received sound based on the acquired acoustic feature set.
Abstract: A method for processing a narrative generated by an artificial Intelligence based natural language generator, to assess relationships between words and phrases in the generated narrative, where necessary, to replace particular words and phrases and more clearly convey a desired intended semantic content of the generated narrative, and/or generate learning data for use by the natural language generator to improve its text generating operation. Based on an input received by the AI based natural language generator, the method generates a narrative, processing the narrative by implementing one or more mechanisms to provide at least one cue in the narrative, in accordance with a plurality of rules to identify semantic content and based on a cue, and the identified semantic content of the narrative, determining how relationships between the words and phrases comprising the narrative could be altered to more clearly convey the semantic content, to realize a directive and communicating the directive.
Abstract: An electronic device for, in response to a user utterance, recommending content suitable to a user, and an operating method thereof are provided. The electronic device includes a user interface, a microphone, a speaker, at least one processor operatively connected with the user interface, the microphone, and the speaker, and at least one memory operatively connected with the processor. The memory stores instructions which, when executed, configure the processor to receive a first user utterance requesting content or suggesting a request for content, through the microphone, provide a list of contents received from at least one content provider, based at least partly on user profile information, receive a user input selecting content from the list of contents, enable a service provided by a content provider of the selected content, after enabling of the service, receive a second user utterance, and provide at least one service.
Abstract: A system, method and computer readable medium for providing translated web content is disclosed. The method on an information processing system includes retrieving a first content in a first language and parsing the first content into a plurality of translatable components. The method further includes generating a unique identifier for each of the plurality of translatable components of the first content and queuing the plurality of translatable components and corresponding unique identifiers for translation into a second language. The method further includes, for each of the plurality of translatable components, storing a translated component and an associated unique identifier corresponding to the translatable component, thereby storing a plurality of translated components and corresponding unique identifiers.
Type:
Grant
Filed:
July 17, 2019
Date of Patent:
April 19, 2022
Inventors:
Enrique Travieso, Adam Rubenstein, Arcadio Andrade, Collin Birdsey
Abstract: A translated media generating and editing system generates playback of audio data in alignment with translated text that has been automatically translated from either a prior translation or a transcription of the audio data.
Type:
Grant
Filed:
December 3, 2019
Date of Patent:
April 12, 2022
Assignee:
Trint Limited
Inventors:
Jeff Kofman, Simon Turvey, John St Leger, Alastair Jardine
Abstract: In response to a user interacting with a tangible peripheral assistant control device (e.g., depressing a button of the device), causing an automated assistant to perform one or more actions. The action(s) performed can be based on input previously provided by the user in configuring the peripheral assistant control device. The action(s) performed in response to interaction with the peripheral assistant control device can vary based on one or more conditions, such as which user is currently active, where the peripheral assistant control device is currently located (which can optionally be inferred based on which of multiple assistant computing devices the button is paired with), and/or the current state of one or more smart devices and/or other devices (e.g., as determined based on a device topology). A utility of the peripheral assistant control device can be automatically extended beyond what was specifically requested by a user during configuration.
Abstract: A system that includes a stand-alone device or a server connected client device are in communication with a server and provide recommendations. The device includes an input component, a storage component, a processor and an output component. The server-connected client device includes an input component that receives the user's request, a communication component that communicates the request to the server and receives the recommendation from the server, and an output component that provides the recommendation to user.
Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for generating author vectors. One of the methods includes obtaining a set of sequences of words, the set of sequences of words comprising a plurality of first sequences of words and, for each first sequence of words, a respective second sequence of words that follows the first sequence of words, wherein each first sequence of words and each second sequence of words has been classified as being authored by a first author; and training a neural network system on the first sequences and the second sequences to determine an author vector for the first author, wherein the author vector characterizes the first author.
Abstract: An apparatus for encoding audio information is provided. The apparatus for encoding audio information includes a selector for selecting a comfort noise generation mode from two or more comfort noise generation modes depending on a background noise characteristic of an audio input signal, and an encoding unit for encoding the audio information, wherein the audio information includes mode information indicating the selected comfort noise generation mode.
Type:
Grant
Filed:
September 25, 2018
Date of Patent:
February 15, 2022
Assignee:
Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.
Inventors:
Emmanuel Ravelli, Martin Dietz, Wolfgang Jaegers, Christian Neukam, Stefan Reuschl
Abstract: Audio encoder for encoding a multichannel signal is shown. The audio encoder includes a downmixer for downmixing the multichannel signal to obtain a downmix signal, a linear prediction domain core encoder for encoding the downmix signal, wherein the downmix signal has a low band and a high band, wherein the linear prediction domain core encoder is configured to apply a bandwidth extension processing for parametrically encoding the high band, a filterbank for generating a spectral representation of the multichannel signal, and a joint multichannel encoder configured to process the spectral representation including the low band and the high band of the multichannel signal to generate multichannel information.
Type:
Grant
Filed:
July 9, 2019
Date of Patent:
February 1, 2022
Assignee:
Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.
Inventors:
Sascha Disch, Guillaume Fuchs, Emmanuel Ravelli, Christian Neukam, Konstantin Schmidt, Conrad Benndorf, Andreas Niedermeier, Benjamin Schubert, Ralf Geiger
Abstract: In response to a user interacting with a tangible peripheral assistant control device (e.g., depressing a button of the device), causing an automated assistant to perform one or more actions. The action(s) performed can be based on input previously provided by the user in configuring the peripheral assistant control device. The action(s) performed in response to interaction with the peripheral assistant control device can vary based on one or more conditions, such as which user is currently active, where the peripheral assistant control device is currently located (which can optionally be inferred based on which of multiple assistant computing devices the button is paired with), and/or the current state of one or more smart devices and/or other devices (e.g., as determined based on a device topology). A utility of the peripheral assistant control device can be automatically extended beyond what was specifically requested by a user during configuration.
Abstract: The present invention relates to a system and a method for determining a stroke based on a voice analysis. According to the present invention, voice data of subjects are collected to extract and analyze voice onset times to determine stroke patients based on voices. The system for determining a stroke generates and collects voice data from test subjects reading a predetermined word that includes a plosive sound. The system for determining a stroke extracts and calculates voice onset times from the voice data to calculate probability parameters for the voice onset times of each of a normal group and a stroke patient group. The system for determining a stroke uses a set of probability parameters to determine an integration section, and calculates probabilities of being in the normal group and the stroke patient group. The system for determining a stroke applies the calculated probabilities to the Bayes theorem to determine whether the subjects are stroke patients.
Type:
Grant
Filed:
August 29, 2019
Date of Patent:
January 11, 2022
Inventors:
Se Jin Park, Seung Nam Min, Kyung Sun Lee, Jung Nam Im, Dong Joon Kim, Sung Kyun Im, Hea Sol Kim, Murali Subramaniyam, Seoung Eun Kim
Abstract: Utterances spoken or sung by a first person can be received, in real time. The detected utterances can be compared to at least a stored sample of utterances spoken or sung by the first person. Based on the comparing, audio of the utterances spoken or sung by the first person can be isolated from a background noise. A volume of the utterances spoken or sung by a first person relative to the background noise can be determined. A key indicator that indicates the volume of the detected utterances spoken or sung by the first person relative to the background noise can be generated. Based on the key indicator, information indicating the volume of the detected utterances spoken or sung by the first person relative to the background noise can be communicated.
Type:
Grant
Filed:
July 16, 2019
Date of Patent:
November 30, 2021
Assignee:
INTERNATIONAL BUSINESS MACHINES CORPORATION
Inventors:
Alan D. Emery, Aditya Sood, Mathews Thomas, Janki Y. Vora
Abstract: Systems and methods for providing intelligent assistance for handling usage modes. In some embodiments, an Information Handling System (IHS) may include: a processor and a memory having program instructions stored thereon that, upon execution, cause the IHS to: detect a headphone state of a user; in response to the detection, capture audio; and in response to the capture, execute a command, at least in part, based upon the headphone state.
Type:
Grant
Filed:
May 1, 2018
Date of Patent:
October 26, 2021
Assignee:
Dell Products, L.P.
Inventors:
Marc Randall Hammons, Todd Erick Swierk, Tyler Ryan Cox
Abstract: An information processing apparatus includes: a speech obtainer which obtains speech of a user; a first controller which, when the first controller recognizes that the speech obtained by the speech obtainer is a first activation word, outputs a speech signal corresponding to a second activation word different from the first activation word; and a second controller which performs an activation process for starting transmission of the speech signal of the speech obtained by the speech obtainer to a VPA cloud server when the second controller recognizes that the speech signal output by the first controller indicates the second activation word.