Patents Examined by Daniel Abebe
-
Patent number: 12210832Abstract: In a method of responding based on sentence paraphrase recognition for dialog system, main keywords of a domain and patterns thereof are recognized and extracted based on a morpheme analysis result in a pre-processing process. Question domains/sub-categories/dialogue-acts are classified using the morpheme analysis result and the extracted main keywords and patterns. Learning a model is performed using classification features of the classified question domains, sub-categories, and dialogue-acts as semantic features of query sentences, and sentence semantic vectors are extracted by measuring similarity between the vectors. A language model of letters is trained and similarity in expression and structure is analyzed by extracting a sentence expression vector based on the letter.Type: GrantFiled: November 20, 2019Date of Patent: January 28, 2025Assignee: Korea Advanced Institute of Science and TechnologyInventors: Hojin Choi, Kyojoong Oh, Youngjun Lee, Soohwan Park
-
Patent number: 12207954Abstract: System and methods for controlling healthcare devices and systems using voice commands are presented. In some aspects a listening device may receive voice command from a person. The voice command may be translated into human readable or machine readable text via a speech-to-text service. A control component may receive the text and send device-specific instructions to a medical device associated with a patient based on the translated voice command. In response to the instructions, the medical device may take an action on a patient. Some examples of actions taken may include setting an alarm limit on a monitor actively monitoring a patient and adjusting the amount of medication delivered by an infusion pump. Because these devices may be controlled using a voice command, in some cases, no physical or manual interaction is needed with the device. As such, multiple devices may be hands-free controlled from any location.Type: GrantFiled: May 9, 2023Date of Patent: January 28, 2025Assignee: CERNER INNOVATION, INC.Inventors: Chad Hays, Randy Lantz
-
Patent number: 12211511Abstract: An artificial intelligence system that conducts an oral question-and-answer session with a single respondent to achieve an educational outcome, such as reviewing material to reinforce ideas or to prepare the respondent for future tests or events, or evaluating the respondent's knowledge or communication skills. The system may use an artificial intelligence engine that can understand and generate natural language, such as a large-language model like ChatGPT™. Text-to-speech and speech-to-text converters may be coupled to the AI engine to enable the system to communicate directly with the respondent using spoken language. The system may provide an interface for entry of session information such as background information on topics and the type of questioning that should be used during the session. The respondent may be given multiple attempts to answer a question and to select the best response, and the system may provide an evaluation and feedback after each response.Type: GrantFiled: October 7, 2024Date of Patent: January 28, 2025Assignee: SANDREW & COMPANY, LLCInventor: Barry Sandrew
-
Patent number: 12197870Abstract: A computer system and method identifies topics in conversations, such as a conversation between a doctor and patient during a medical examination. The system and method generates, based on first text (such as a document corpus including previous clinical documentation), a plurality of sentence embeddings representing a plurality of semantic representations in a plurality of sentences in the training text. The system and method generate a classifier based on the second text, which includes a plurality of sections associated with a plurality of topics, and the plurality of sentence embeddings. The system and method generate, based on a sentence (such as a sentence in a doctor-patient conversation) and the classifier, an identifier of a topic to associate with the first sentence. The system and method may also insert the sentence into a section, associated with the identified topic, in a document (such as a clinical note).Type: GrantFiled: March 18, 2021Date of Patent: January 14, 2025Assignee: Solventum Intellectual Properties CompanyInventors: Thomas S. Polzin, Hua Cheng, Detlef Koll
-
Patent number: 12190075Abstract: An apparatus for processing voice commands includes: a memory configured to store computer-executable instructions, and a processor configured to execute the computer-executable instructions. When executed, the instructions cause the processor to perform: receiving an utterance of a user in an input language set by the user, determining an utterance intent of the utterance by analyzing the utterance with the input language, determining a standard utterance in the input language corresponding to the utterance of the user based on the determined utterance intent, determining whether the input language and an output language are different languages, extracting a standard utterance in the output language corresponding to the determined standard utterance in the input language when the input language and the output language are different, generating an output response in the output language based on the extracted standard utterance in the output language, and outputting the output response.Type: GrantFiled: May 24, 2022Date of Patent: January 7, 2025Assignee: Samsung Electronics Co., Ltd.Inventors: Jooyong Byeon, Seolhee Kim
-
Patent number: 12186254Abstract: A system and method for voice-based control of sexual stimulation devices. In some configurations, the system and method involve receiving voice data, analyzing the voice data to detect spoken commands, and generating control signals based on the commands. In some configurations, the system and method involve receiving voice data, analyzing the voice data for non-speech vocalizations, detecting voice stress patterns, and generating control signals based on the detected patterns. In some configurations, the analyses of the voice data are performed by machine learning algorithms which may be trained on associations between speech and non-speech vocalizations of a user while the user engages in one or more voice-based training tasks, associating speech and non-speech vocalizations with controls of the sexual stimulation device. In some configurations, machine learning algorithms are used to make the associations. In some configurations, data from other biometric sensors is included in the associations.Type: GrantFiled: August 21, 2023Date of Patent: January 7, 2025Inventor: Brian Sloan
-
Patent number: 12183347Abstract: Processing stacked data structures is provided. A system receives an input audio signal detected by a sensor of a local computing device, identifies an acoustic signature, and identifies an account corresponding to the signature. The system establishes a session and a profile stack data structure including a first profile layer having policies configured by a third-party device. The system pushes, to the profile stack data structure, a second profile layer retrieved from the account. The system parses the input audio signal to identify a request and a trigger keyword. The system generates, based on the trigger keyword and the second profile layer, a first action data structure compatible with the first profile layer. The system provides the first action data structure for execution. The system disassembles the profile stack data structure to remove the first profile layer or the second profile layer from the profile stack data structure.Type: GrantFiled: January 13, 2023Date of Patent: December 31, 2024Assignee: GOOGLE LLCInventors: Anshul Kothari, Gaurav Bhaya, Tarun Jain
-
Patent number: 12183322Abstract: A method includes receiving a sequence of acoustic frames characterizing one or more utterances as input to a multilingual automated speech recognition (ASR) model. The method also includes generating a higher order feature representation for a corresponding acoustic frame. The method also includes generating a hidden representation based on a sequence of non-blank symbols output by a final softmax layer. The method also includes generating a probability distribution over possible speech recognition hypotheses based on the hidden representation generated by the prediction network at each of the plurality of output steps and the higher order feature representation generated by the encoder at each of the plurality of output steps. The method also includes predicting an end of utterance (EOU) token at an end of each utterance. The method also includes classifying each acoustic frame as either speech, initial silence, intermediate silence, or final silence.Type: GrantFiled: September 22, 2022Date of Patent: December 31, 2024Assignee: Google LLCInventors: Bo Li, Tara N. Sainath, Ruoming Pang, Shuo-yiin Chang, Qiumin Xu, Trevor Strohman, Vince Chen, Qiao Liang, Heguang Liu, Yanzhang He, Parisa Haghani, Sameer Bidichandani
-
Patent number: 12175996Abstract: Systems and methods for visualizations of music may include one or more processors which receive an audio input, and compute a simulation of a human auditory periphery using the audio input. The processor(s) may generate one or more visual patterns on a visual display, according to the simulation, the one or visual patterns synchronized to the audio input.Type: GrantFiled: October 19, 2022Date of Patent: December 24, 2024Assignee: Oscilloscape, LLCInventors: Edward W. Large, Ji Chul Kim
-
Patent number: 12175993Abstract: A method for decoding an encoded audio bitstream is disclosed. The method includes receiving the encoded audio bitstream and decoding the audio data to generate a decoded lowband audio signal. The method further includes extracting high frequency reconstruction metadata and filtering the decoded lowband audio signal with an analysis filterbank to generate a filtered lowband audio signal. The method also includes extracting a flag indicating whether either spectral translation or harmonic transposition is to be performed on the audio data and regenerating a highband portion of the audio signal using the filtered lowband audio signal and the high frequency reconstruction metadata in accordance with the flag.Type: GrantFiled: May 29, 2024Date of Patent: December 24, 2024Assignee: DOLBY INTERNATIONAL ABInventors: Lars Villemoes, Heiko Purnhagen, Per Ekstrand
-
Patent number: 12170088Abstract: An electronic device is provided. The electronic device according to an embodiment includes a microphone, a communicator comprising communication circuitry, and a processor configured to control the communicator to transmit a control command to an external audio device for reducing an audio output level of the external audio device in response to a trigger signal for starting a voice control mode being received through the microphone and to control the electronic device to operate in the voice control mode.Type: GrantFiled: July 7, 2023Date of Patent: December 17, 2024Assignee: Samsung Electronics Co., Ltd.Inventors: Min-seok Kim, Min-ho Lee
-
Patent number: 12165655Abstract: An artificial intelligence system that conducts dialogs with one or more participants to achieve an educational outcome, such as teaching material with a Socratic question-and-answer method, reviewing material to reinforce ideas or to prepare a participant for future tests or events, or evaluating a participant's knowledge or communication skills. The system may use an artificial intelligence engine that can understand and generate natural language, such as a large-language model like ChatGPT. Text-to-speech and speech-to-text converters may be coupled to the AI engine to enable the system to communicate directly with participants using spoken language. The system may provide an interface for entry of contextual information for a dialog, such as background information on topics for discussion. It may provide an interface for moderation of a dialog, for example by a teacher, lawyer, thesis advisor, etc. System may evaluate the performance of the participants at the end of a dialog.Type: GrantFiled: May 21, 2024Date of Patent: December 10, 2024Assignee: SANDREW & COMPANY, LLCInventor: Barry Sandrew
-
Patent number: 12165662Abstract: A downscaled version of an audio decoding procedure may more effectively and/or at improved compliance maintenance be achieved if the synthesis window used for downscaled audio decoding is a downsampled version of a reference synthesis window involved in the non-downscaled audio decoding procedure by downsampling by the downsampling factor by which the downsampled sampling rate and the original sampling rate deviate, and downsampled using a segmental interpolation in segments of ¼ of the frame length.Type: GrantFiled: May 9, 2023Date of Patent: December 10, 2024Assignee: Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.Inventors: Markus Schnell, Manfred Lutzky, Eleni Fotopoulou, Konstantin Schmidt, Conrad Benndorf, Adrian Tomasek, Tobias Albert, Timon Seidl
-
Patent number: 12159638Abstract: A downscaled version of an audio decoding procedure may more effectively and/or at improved compliance maintenance be achieved if the synthesis window used for downscaled audio decoding is a downsampled version of a reference synthesis window involved in the non-downscaled audio decoding procedure by downsampling by the downsampling factor by which the downsampled sampling rate and the original sampling rate deviate, and downsampled using a segmental interpolation in segments of ¼ of the frame length.Type: GrantFiled: April 25, 2023Date of Patent: December 3, 2024Assignee: Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.Inventors: Markus Schnell, Manfred Lutzky, Eleni Fotopoulou, Konstantin Schmidt, Conrad Benndorf, Adrian Tomasek, Tobias Albert, Timon Seidl
-
Patent number: 12159086Abstract: A device can perform device arbitration, even when the device is unable to communicate with a remote system over a wide area network (e.g., the Internet). Upon detecting a wakeword in an utterance, the device can wait a period of time for data to arrive at the device, which, if received, indicates to the device that another speech interface device in the environment detected an utterance. If the device receives data prior to the period of time lapsing, the device can determine the earliest-occurring wakeword based on multiple wakeword occurrence times, and may designate whichever device that detected the wakeword first as the designated device to perform an action with respect to the user speech. To account for differences in sound capture latency between speech interface devices, a pre-calculated time offset value can be applied to wakeword occurrence time(s) during device arbitration.Type: GrantFiled: November 23, 2022Date of Patent: December 3, 2024Assignee: Amazon Technologies, Inc.Inventor: Stanislaw Ignacy Pasko
-
Patent number: 12154580Abstract: A downscaled version of an audio decoding procedure may more effectively and/or at improved compliance maintenance be achieved if the synthesis window used for downscaled audio decoding is a downsampled version of a reference synthesis window involved in the non-downscaled audio decoding procedure by downsampling by the downsampling factor by which the downsampled sampling rate and the original sampling rate deviate, and downsampled using a segmental interpolation in segments of ¼ of the frame length.Type: GrantFiled: May 9, 2023Date of Patent: November 26, 2024Assignee: Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.Inventors: Markus Schnell, Manfred Lutzky, Eleni Fotopoulou, Konstantin Schmidt, Conrad Benndorf, Adrian Tomasek, Tobias Albert, Timon Seidl
-
Patent number: 12154569Abstract: Example techniques involve a control hierarchy for a “smart” home having smart appliances and related devices, such as wireless illumination devices, home-automation devices (e.g., thermostats, door locks, etc.), and audio playback devices, among others. An example home includes various rooms in which smart devices might be located. Under the example control hierarchy described herein and referred to as “home graph,” a name of a room (e.g., “Kitchen”) may represent a smart device (or smart devices) within that room. In other words, from the perspective of a user, the smart devices within a room are that room. This hierarchy permits a user to refer to a smart device within a given room by way of the name of the room when controlling smart devices within the home using a voice user interface (VUI) or graphical user interface (GUI).Type: GrantFiled: May 1, 2023Date of Patent: November 26, 2024Assignee: Sonos, Inc.Inventors: Robert Lambourne, Dayn Wilberding, Jeffrey Torgerson
-
Patent number: 12154579Abstract: A downscaled version of an audio decoding procedure may more effectively and/or at improved compliance maintenance be achieved if the synthesis window used for downscaled audio decoding is a downsampled version of a reference synthesis window involved in the non-downscaled audio decoding procedure by downsampling by the downsampling factor by which the downsampled sampling rate and the original sampling rate deviate, and downsampled using a segmental interpolation in segments of 1/4 of the frame length.Type: GrantFiled: May 9, 2023Date of Patent: November 26, 2024Assignee: Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.Inventors: Markus Schnell, Manfred Lutzky, Eleni Fotopoulou, Konstantin Schmidt, Conrad Benndorf, Adrian Tomasek, Tobias Albert, Timon Seidl
-
Patent number: 12147521Abstract: The presently disclosed embodiments may include a non-transitory computer readable medium containing instruction that when executed by at least one processor cause the at least one processor to perform thresholding operations for interpretation of facial skin micromovements. The operations may include detecting facial micromovements in an absence of perceptible vocalization associated with the facial micromovements, determining an intensity level of the facial micromovements and comparing the determined intensity level with a threshold. When the intensity level is above the threshold, the operations may include interpreting the facial micromovements and when the intensity level falls beneath the threshold, the operations may include disregarding the facial micromovements.Type: GrantFiled: November 15, 2023Date of Patent: November 19, 2024Assignee: Q (Cue) Ltd.Inventor: Yonatan Wexler
-
Patent number: 12137183Abstract: A call captioning system for captioning a hearing user's (HU's) voice signal during an ongoing call with an assisted user (AU) includes: an AU communication device with a display screen and a caption service activation feature, and a first processor programmed to, during an ongoing call, receive the HU's voice signal. Prior to activating the caption service via the activation feature, the processor uses an automated speech recognition (ASR) engine to generate HU voice signal captions, detect errors in the HU voice signal captions, use the errors to train the ASR software to the HU's voice signal to increase accuracy of the HU captions generated by the ASR engine; and store the trained ASR engine for subsequent use. Upon activating the caption service during the ongoing call, the processor uses the trained ASR engine to generate HU voice signal captions and present them to the AU via the display screen.Type: GrantFiled: March 20, 2023Date of Patent: November 5, 2024Assignee: Ultratec, Inc.Inventors: Robert M. Engelke, Kevin R. Colwell, Christopher R. Engelke