Patents Examined by Marcus T. Riley
  • Patent number: 11972751
    Abstract: Disclosed are a method and an apparatus for detecting a voice end point by using acoustic and language modeling information to accomplish strong voice recognition. A voice end point detection method according to an embodiment may comprise the steps of: inputting an acoustic feature vector sequence extracted from a microphone input signal into an acoustic embedding extraction unit, a phonemic embedding extraction unit, and a decoder embedding extraction unit, which are based on a recurrent neural network (RNN); combining acoustic embedding, phonemic embedding, and decoder embedding to configure a feature vector by the acoustic embedding extraction unit, the phonemic embedding extraction unit, and the decoder embedding extraction unit; and inputting the combined feature vector into a deep neural network (DNN)-based classifier to detect a voice end point.
    Type: Grant
    Filed: June 29, 2020
    Date of Patent: April 30, 2024
    Assignee: IUCF-HYU (INDUSTRY-UNIVERSITY COOPERATION FOUNDATION HANYANG UNIVERSITY)
    Inventors: Joon-Hyuk Chang, Inyoung Hwang
  • Patent number: 11966713
    Abstract: In various embodiments, a computer-implemented language identification and communication system is provided. The system includes an application engine configured for processing data associated with multiple access devices of a population of users who are users seeking human language translation services and users providing human language translation services. A geolocation module is provided for locating a position of various users, such as the different locations of human language translators. The application engine is further programmed for receiving translator selections from user access devices and establishing communication connections between or among different user access devices.
    Type: Grant
    Filed: September 16, 2021
    Date of Patent: April 23, 2024
    Assignee: Zoose Language Technologies LLC
    Inventors: Patrick S. Allocco, Shalini Kadavill
  • Patent number: 11961520
    Abstract: Disclosed are a voice recognition system and a display device using the same. The disclosed voice recognition system includes a plate structure, a vibration sensor, and a voice recognition device. The plate structure vibrates based on propagation of a voice wave generated from a user, and the vibration sensor is provided in contact with the plate structure to detect the vibration of the plate structure. The voice recognition device recognizes voice of the user by receiving a signal output from the vibration sensor.
    Type: Grant
    Filed: December 16, 2022
    Date of Patent: April 16, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Cheheung Kim, Jaehyung Jang, Hyeokki Hong
  • Patent number: 11948573
    Abstract: A method for automatically rectifying in real-time anomalies in natural language processing systems. The method can include determining an output corresponding to a user request from a user device for a user based on a new request template or machine learning. The method further can include retrieving one or more entity rules corresponding to entity data of the user request. The method also can include overwriting entity information of the entity data corresponding to the one or more entity rules. Additionally, the method can include outputting the output. Furthermore, the method can include transmitting, to the user device, a response to the user. Other embodiments are disclosed.
    Type: Grant
    Filed: October 31, 2022
    Date of Patent: April 2, 2024
    Assignee: WALMART APOLLO, LLC
    Inventors: Snehasish Mukherjee, Haoxuan Chen, Phani Ram Sayapaneni, Shankara Bhargava Subramanya
  • Patent number: 11942087
    Abstract: A device performs a method for using image data to aid voice recognition. The method includes the device capturing image data of a vicinity of the device and adjusting, based on the image data, a set of parameters for voice recognition performed by the device. The set of parameters for the device performing voice recognition include, but are not limited to: a trigger threshold of a trigger for voice recognition; a set of beamforming parameters; a database for voice recognition; and/or an algorithm for voice recognition. The algorithm may include using noise suppression or using acoustic beamforming.
    Type: Grant
    Filed: January 13, 2021
    Date of Patent: March 26, 2024
    Assignee: Google Technology Holdings LLC
    Inventors: Robert A. Zurek, Adrian M. Schuster, Fu-Lin Shau, Jincheng Wu
  • Patent number: 11922939
    Abstract: A system and method are disclosed for ignoring a wakeword received at a speech-enabled listening device when it is determined the wakeword is reproduced audio from an audio-playing device. Determination can be by detecting audio distortions, by an ignore flag sent locally between an audio-playing device and speech-enabled device, by and ignore flag sent from a server, by comparison of received audio played audio to a wakeword within an audio-playing device or a speech-enabled device, and other means.
    Type: Grant
    Filed: May 4, 2022
    Date of Patent: March 5, 2024
    Assignee: SoundHound AI IP, LLC
    Inventors: Hsuan Yang, Qindí Zhãng, Warren S. Heit
  • Patent number: 11914923
    Abstract: Techniques for computer system-based conversations are described. In an example, a system receives, from a first device, first data corresponding to a first interaction in a conversation that requests a function. The system causes the first device to output a first response to the first interaction. Prior to an execution of the function, the system determines that the conversation is to be paused and causes the first device to output a first indication that the conversation is paused. Upon determining that the conversation is to be resumed, the system causes a second device to output a second indication that the conversation is resumed. The second device can be the same or different from the first device. The system receives, from the second device, second data corresponding to a second interaction in the conversation and causes the execution of the function based at least in part on the second data.
    Type: Grant
    Filed: December 10, 2021
    Date of Patent: February 27, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Shiveesh Fotedar, Saurabh Rathi, Steven Bishop
  • Patent number: 11914658
    Abstract: The invention relates to a multi-node word segmentation system and method for keyword search. A word distribution proxy server in that system is communication connection with each word segmentation node server respectively, receive a large number of words, and distributes each word to a corresponding word segmentation node server according to a preset word segmentation rule; the client proxy server is communication connection with each word segmentation node server respectively, receives terms to be searched uploaded by a user terminal, and distributes the terms to be searched to the corresponding word segmentation node servers according to a preset word segmentation rule; the segmentation node server inquires the words matched with the terms to be searched, and returns the words obtained by inquiry to the client proxy server; and that client proxy server sends the query words to a corresponding user terminal.
    Type: Grant
    Filed: May 15, 2020
    Date of Patent: February 27, 2024
    Assignee: SHENZHEN SEKORM COMPONENT NETWORK CO., LTD
    Inventor: Lianglei Shang
  • Patent number: 11906320
    Abstract: The present disclosure provides a method and an apparatus for managing navigation broadcast, and a device, related to an Intelligent transportation technology field. A specific implementation solution includes: obtaining a geographical identifier of a user; obtaining a statement-conversion template set corresponding to the geographical identifier based on the geographical identifier of the user; converting a standard navigation broadcast statement based on the statement-conversion template set to generate a geographical navigation broadcast statement; and performing navigation broadcast based on the geographical navigation broadcast statement. Thereby, the navigation broadcasts are matched with respective regions, users in different regions are provided with diversified and personalized navigation broadcasts.
    Type: Grant
    Filed: April 7, 2020
    Date of Patent: February 20, 2024
    Assignee: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD.
    Inventors: Ran Ji, Jizhou Huang, Ying Li, Yongzhi Ji, Lei Jia
  • Patent number: 11907667
    Abstract: A system for assisting sharing of information includes circuitry to: input a plurality of sentences each representing a statement made by one of a plurality of users, the sentence being generated by speaking or writing during a meeting or by extracting from at least one of meeting data, email data, electronic file data, and chat data at any time; determine a statement type of the statement represented by each one of the plurality of sentences, the statement type being one of a plurality of statement types previously determined; select, from among the plurality of sentences being input, one or more sentences each representing a statement of a specific statement type of the plurality of types; and output a list of the selected one or more sentences as key statements of the plurality of sentences.
    Type: Grant
    Filed: August 11, 2022
    Date of Patent: February 20, 2024
    Assignee: RICOH COMPANY, LTD.
    Inventor: Tomohiro Shima
  • Patent number: 11908481
    Abstract: Provided is a method for encoding live-streaming data, including: acquiring first state information associated with a current data frame; generating backup state information by backing up the first state information; generating a first encoded data frame by encoding the current data frame based on a first bit rate and the first state information; generating reset state information by resetting the updated first state information based on the backup state information; generating a second encoded data frame by encoding the current data frame based on a second bit rate and the reset state information; and generating a first target data frame corresponding to the current data frame based on the first encoded data frame and the second encoded data frame.
    Type: Grant
    Filed: January 24, 2022
    Date of Patent: February 20, 2024
    Assignee: BEIJING DAJIA INTERNET INFORMATION TECHNOLOGY CO., LTD.
    Inventors: Wenhao Xing, Chen Zhang
  • Patent number: 11908453
    Abstract: A method and a system for training a machine-learning algorithm (MLA) to determine a user class of a user of an electronic device are provided. The method comprises: receiving a training audio signal representative of a training user utterance; soliciting, by the processor, a plurality of assessor-generated labels for the training audio signal, the given one of the plurality of assessor-generated labels being indicative of whether the training user is perceived to be one of a first class and a second class; generating an amalgamated assessor-generated label for the training audio signal, the amalgamated assessor-generated label being indicative of a label distribution of the plurality of assessor-generated labels between the first class and the second class; generating a training set of data including the training audio signal and the amalgamated assessor-generated to train the MLA to determine the user class of the user producing an in-use user utterance.
    Type: Grant
    Filed: August 23, 2021
    Date of Patent: February 20, 2024
    Assignee: Direct Cursus Technology L.L.C
    Inventors: Vladimir Andreevich Aliev, Stepan Aleksandrovich Kargaltsev, Artem Valerevich Babenko
  • Patent number: 11900062
    Abstract: Described are methods and systems are for generating dynamic conversational queries. For example, as opposed to being a simply reactive system, the methods and systems herein provide means for actively determining a user's intent and generating a dynamic query based on the determined user intent. Moreover, these methods and systems generate these queries in a conversational environment.
    Type: Grant
    Filed: October 1, 2021
    Date of Patent: February 13, 2024
    Assignee: Capital One Services, LLC
    Inventors: Minh Le, Arturo Hernandez Zeledon, Md Arafat Hossain Khan
  • Patent number: 11900518
    Abstract: A method of producing an avatar video, the method comprising the steps of: providing a reference image of a person's face; providing a plurality of characteristic features representative of a facial model X0 of the person's face, the characteristic features defining a facial pose dependent on the person speaking; providing a target phrase to be rendered over a predetermined time period during the avatar video and providing a plurality of time intervals t within the predetermined time period; generating, for each of said times intervals t, speech features from the target phrase, to provide a sequence of speech features; and generating, using the plurality of characteristic features and sequence of speech features, a sequence of facial models Xt for each of said time intervals t.
    Type: Grant
    Filed: September 17, 2019
    Date of Patent: February 13, 2024
    Assignee: VirtTari Limited
    Inventors: Peter Alistair Brady, Hayden Allen-Vercoe, Sathish Sankarpandi, Ethan Dickson
  • Patent number: 11893996
    Abstract: Techniques for generating a personalization identifier that is usable by a skill to customize output of supplemental content to a user, without the skill being able to determine an identity of the user based on the personalization identifier, are described. A personalization identifier may be generated to be specific to a skill, such that different skills receive different personalization identifiers with respect to the same user. The personalization identifier may be generated by performing a one-way hash of a skill identifier, and a user profile identifier and/or a device identifier. User-perceived latency may be reduced by generating the personalization identifier at least partially in parallel to performing ASR processing and/or NLU processing.
    Type: Grant
    Filed: June 29, 2021
    Date of Patent: February 6, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Mark Conrad Kockerbeck, Song Chen, Aditi Srinivasan, Ryan Idrogo-Lam, Jilani Zeribi, John Botros
  • Patent number: 11893989
    Abstract: A system and method for controlling an electronic eyewear device using voice commands receives audio data from a microphone, processes the audio data to identify a wake word, and upon identification of a wake word, processes the audio data to identify at least one action keyword in the audio data. The audio data is provided to one of a plurality of controllers associated with different action keywords or sets of action keywords to implement an action. For example, the audio data may be provided to a settings controller to adjust settings of the electronic eyewear device when the action keyword is indicative of a request to adjust a setting of the electronic eyewear device or to a navigation controller to navigate to the system information of the electronic eyewear device when the action keyword is indicative of a request to navigate to system information of the electronic eyewear device.
    Type: Grant
    Filed: July 13, 2021
    Date of Patent: February 6, 2024
    Assignee: Snap Inc.
    Inventor: Piotr Gurgul
  • Patent number: 11887602
    Abstract: Techniques for performing audio-based device location determinations are described. A system may send, to a first device, a command to output audio requesting a location of the first device be determined. A second device may receive the audio and send, to the system, data representing the second device received the audio, where the received data includes spectral energy data representing a spectral energy of the audio as received by the second device. The system may, using the spectral energy data, determine attenuation data representing an attenuation experienced by the audio as it traveled from the first device to the second device. The system may generate, based on the attenuation data, spatial relationship data representing a spatial relationship between the first device and the second device, where the spatial relationship data is usable to determine a device for outputting a response to a subsequently received user input.
    Type: Grant
    Filed: December 10, 2021
    Date of Patent: January 30, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Brendon Jude Wilson, Henry Michael D Souza, Cindy Angie Hou, Christopher Evans, Sumit Garg, Ravina Chopra
  • Patent number: 11881220
    Abstract: A display device for providing a speech recognition service according to an embodiment of the present disclosure can include a display unit, a network interface unit configured to communicate with a server, and a control unit configured to receive a voice command uttered by a user, acquire usage information of the display device, transmit the voice command and the usage information of the display device to the server through the network interface unit, receive, from the server, an utterance intention based on the voice command and the usage information of the display device, and perform an operation corresponding to the received utterance intention.
    Type: Grant
    Filed: May 15, 2019
    Date of Patent: January 23, 2024
    Assignee: LG ELECTRONICS INC.
    Inventors: Hyangjin Lee, Jaekyung Lee
  • Patent number: 11880651
    Abstract: Taste and smell classification from multilanguage descriptions can be performed by extracting, by one or more processors using natural language processing, a text including one or more words associated with taste and smell perceptions from an input received from a plurality of users. The input includes multilanguage information regarding at least one of changes in smell and changes in taste perceived by each of the plurality of users. Feature vectors are generated for the text extracted from the input using global vectors, and a distance between the feature vectors and a plurality of reference descriptors associated with taste and smell is calculated for determining a similarity between the text and the reference descriptors and creating a training dataset based on which a classification model is generated for categorizing the plurality of users according to the at least one of changes in smell and changes in taste.
    Type: Grant
    Filed: June 23, 2021
    Date of Patent: January 23, 2024
    Assignee: International Business Machines Corporation
    Inventors: Pablo Meyer Rojas, Guillermo Cecchi, Elif Eyigoz, Raquel Norel
  • Patent number: 11881215
    Abstract: Various embodiments of the present invention relate to a method for providing an intelligent assistance service, and an electronic device performing same. According to an embodiment, the electronic device includes a display, a communication interface, at least one processor, and at least one memory, wherein the memory is configured to store a task customized by a user and mapped to any one among a selected word, phrase, or sentence. The memory may store instructions which, when executed, cause the processor to: display a user interface, configured to set or change the task, on the display; display at least one utterance related to the task as text on the user interface; identify and display at least one replaceable parameter in the utterance; receive a user input, which may be used as the parameter, for selecting or inputting at least one item; and store the task including the item.
    Type: Grant
    Filed: October 10, 2019
    Date of Patent: January 23, 2024
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Suneung Park, Taekwang Um, Jaeyung Yeo