Patents Examined by Farzad Kazeminezhad
  • Patent number: 11017799
    Abstract: The present disclosure discloses a method for processing a voice in interior environment of a vehicle, an electronic device and a storage medium. The method includes the following. A reference audio is acquired, and the reference audio is recorded to obtain a recorded audio. A pure voice is acquired. Noise data for each part or period of the recorded audio satisfying a target signal-to-noise ratio condition pertaining to that part is selected from the recorded audio, and the noise data is superimposed to the pure data to obtain a noisy voice. The noisy voice and the reference audio are inputted to an acoustic echo canceller (AEC) module as inputted data. The AEC module is configured to perform an echo cancellation operation on the inputted data to obtain training data having AEC residual noise.
    Type: Grant
    Filed: October 15, 2018
    Date of Patent: May 25, 2021
    Inventor: Hui Song
  • Patent number: 11017172
    Abstract: A method for determining whether a text complies with a query, an apparatus and a computer program product, the method comprising: obtaining a macro comprising a plurality of representations of a query; and executing the macro against a database comprising a simplified representation of at least one phrase within the text, to determine 10 whether the text complies with any of the representations of the query. The simplified representation is presented in the form of: conferrer (carrier; bearer; substrate; descriptor) and provides information pertaining to who, what, when and how with respect to the text associated with the simplified representation.
    Type: Grant
    Filed: August 1, 2018
    Date of Patent: May 25, 2021
    Inventor: Nahum Korda
  • Patent number: 11011157
    Abstract: Techniques are disclosed for generating ASR training data. According to an embodiment, impactful ASR training corpora is generated efficiently, and the quality or relevance of ASR training corpora being generated is increased by leveraging knowledge of the ASR system being trained. An example methodology includes: selecting one of a word or phrase, based on knowledge and/or content of said ASR training corpora; presenting a textual representation of said word or phrase; receiving a speech utterance that includes said word or phrase; receiving a transcript for said speech utterance; presenting said transcript for review (to allow for editing, if needed); and storing said transcript and said audio file in an ASR system training database. The selecting may include, for instance, selecting a word or phrase that is under-represented in said database, and/or based upon an n-gram distribution on a language, and/or based upon known areas that tend to incur transcription mistakes.
    Type: Grant
    Filed: November 13, 2018
    Date of Patent: May 18, 2021
    Assignee: ADOBE INC.
    Inventor: Franck Dernoncourt
  • Patent number: 10999636
    Abstract: Systems, methods, and computer-readable media are disclosed for voice-based content searching. The systems, methods, and computer-readable media described herein may improve user experiences by providing voice-based content searching using mobile devices. In an example method described herein, a computer system may receive a voice search initiation signal from a mobile device over a WiFi protocol connection or a Bluetooth protocol connection. The computer system may monitor a WiFi receiver and a Bluetooth receiver of the computer system for first voice data. The computer system may determine that the first voice data is being received at the WiFi receiver. The computer system may determine that second voice data is received at the Bluetooth receiver while the first voice data is being received at the WiFi receiver, and discard the second voice data.
    Type: Grant
    Filed: November 10, 2014
    Date of Patent: May 4, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Richard W. Kim, Radhika Suresh Karandikar, Aaron Nicholas Olsen, James Shu, Shinji Kimura, Xin Wang
  • Patent number: 10997987
    Abstract: A signal processor comprising: an input terminal, configured to receive an input-signal; a voicing-terminal, configured to receive a voicing-signal representative of a voiced speech component of the input-signal; an output terminal; a delay block, configured to receive the input-signal and provide a filter-input-signal as a delayed representation of the input-signal; a filter block, configured to: receive the filter-input-signal; and provide a noise-estimate-signal by filtering the filter-input-signal; a combiner block, configured to: receive a combiner-input-signal representative of the input-signal; receive the noise-estimate-signal; and combine the combiner-input-signal with the noise-estimate-signal to provide an output-signal to the output terminal; and a filter-control-block, configured to: receive the voicing-signal; receive signalling representative of the input-signal; and set filter coefficients of the filter block in accordance with the voicing-signal and the input-signal.
    Type: Grant
    Filed: May 15, 2018
    Date of Patent: May 4, 2021
    Assignee: NXP B.V.
    Inventors: Ann Elvire F. Spriet, Wouter Joos Tirry
  • Patent number: 10991366
    Abstract: A method, performed by a dialogue processing device, of processing dialogue associated with a user based on dialog act information, the method comprises receiving speech information, corresponding to speech of the user, including a plurality of sentence units; identifying a first sentence unit and a second sentence unit, of the plurality of sentence units, based on receiving the speech information; extracting a first dialog act indicative of an intention of the first sentence unit and extracting a second dialog act indicative of an intention of the second sentence unit; extracting a first dialog act indicative of an intention of the first sentence unit and extracting a second dialog act indicative of an intention of the second sentence unit; processing the first sentence unit and the second unit in a sequence according to respective priority orders assigned based on a number of empty slots of dialogue frames of the sentence units.
    Type: Grant
    Filed: July 2, 2018
    Date of Patent: April 27, 2021
    Assignee: SAMSUNG SDS CO., LTD.
    Inventors: Han Hoon Kang, Eun Hye Ji, Na Rae Kim, Jae Young Yang
  • Patent number: 10978072
    Abstract: According to an embodiment, a display control system includes a recognizer, a calculator, and a display controller. The recognizer is configured to recognize sound generated from a sound source. The calculator is configured to calculate a first direction of the sound source from a position of a user based at least in part on the recognized sound. The display controller is configured to cause a display to output display information that visually expresses the sound in a first display format determined based at least in part on an angle between a second direction in which the user faces and the first direction of the sound source.
    Type: Grant
    Filed: February 15, 2018
    Date of Patent: April 13, 2021
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Masahiro Yamamoto, Kosei Fume
  • Patent number: 10979242
    Abstract: Embodiments of the present disclosure pertain to a personal assistant controller. In one embodiment, the present disclosure includes a computer implemented method comprising receiving a voice audio signal in the personal assistant controller, converting the voice audio signal into a target command corresponding to one of a plurality of personal assistants, wherein different personal assistants comprise different target command protocols for executing different operations on different network enabled appliances, and sending the target command for execution by a backend system corresponding to the one of the plurality of personal assistants, and in accordance therewith, performing an operation on the backend system.
    Type: Grant
    Filed: June 5, 2018
    Date of Patent: April 13, 2021
    Assignee: SAP SE
    Inventors: Alexander Ocher, Andrey Belyy, Viktor Lapitski
  • Patent number: 10978091
    Abstract: A system is provided to realize suppression by selecting wavelets for feature compression in distributed speech recognition. The system comprises a first device and a second device. The first device comprising: a first network module for connecting to a network; an acoustic transducer module for recording speech and outputting frames of recorded signal; and a first processor configured for the following: extracting multiple-dimensional speech features from the frames of the recorded signal to generate multiple feature sequences; applying discrete wavelet transform (DWT) to the feature sequences to obtain a plurality of component data; and transmitting at least one of the plurality of component data via the network, wherein another one of the plurality of component data is not transmitted.
    Type: Grant
    Filed: November 30, 2018
    Date of Patent: April 13, 2021
    Inventors: Yu Tsao, Syu-Siang Wang
  • Patent number: 10949230
    Abstract: A device may be configured to enable a user to select a language, and may fulfill resource requests from applications by selecting, from among resources respectively associated with a language, a resource associated with the selected language of the user. However, this resource selection process may be inadequate if the user selects multiple languages; if a resource associated with the selected language of the user is unavailable, but resources associated with related languages are available; or if the user and/or the application specifies an ordering for the selection among the languages. Presented herein are techniques for performing the resource selection by, for respective languages selected by the user, calculating a weight representing a suitability of the language for the resource request; generating a selection order of the selected languages according to the weights; and selecting a resource based on the position of the associated language in the selection order.
    Type: Grant
    Filed: May 31, 2012
    Date of Patent: March 16, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Rylan Michael Hawkins, Eric Scott Albright, Nick Beal, Peter Geoffrey Constable, Wade H. Curtiss, Erik Fortune, Andrew Stuart Glass, Samuel Aaron Beard, Microsoft Technology Licensing
  • Patent number: 10937432
    Abstract: There is provided an error concealment unit, method, and computer program, for providing an error concealment audio information for concealing a loss of an audio frame in an encoded audio information. In one embodiment, the error concealment unit provides an error concealment audio information for a lost audio frame on the basis of a properly decoded audio frame preceding the lost audio frame. The error concealment unit derives a damping factor on the basis of characteristics of a decoded representation of the properly decoded audio frame preceding the lost audio frame. The error concealment unit performs a fade out using the damping factor.
    Type: Grant
    Filed: September 6, 2018
    Date of Patent: March 2, 2021
    Assignee: Fraunhofer-Gesellschaft zur Foerderung der Angewandten Forschung e.V.
    Inventors: Jérémie Lecomte, Adrian Tomasek
  • Patent number: 10937411
    Abstract: Methods, apparatus, and articles of manufacture to generate voices for artificial speech are disclosed. An example apparatus includes a component storing an identifier, the identifier uniquely identifying the apparatus from a plurality of apparatus, an artificial speech generator to generate a first artificial speech signal representing text, the first artificial speech signal generated based on the identifier, the first artificial speech signal audibly different from artificial speech signals generated by respective ones of the plurality of apparatus for the text, an output device to output an audible signal representing the first artificial speech signal.
    Type: Grant
    Filed: October 2, 2019
    Date of Patent: March 2, 2021
    Inventor: Oleg Pogorelik
  • Patent number: 10923113
    Abstract: Techniques for recommending speechlets to users via voice user interfaces (VUIs) are described. A system may include predetermined rankings of speechlets. The predetermined rankings may be determined offline and may represent reliabilities of the speechlets. During runtime, when the system receives a user input that can be handled by multiple speechlets, the system may determine a type of speechlet configured to handle the user input. The system may also identify information specific to the user's previous interactions with the system. The system may rerank the predetermined rankings using the user specific information, as well as other information. The system may recommend speechlets to the user based on the speechlets' rerankings. The system may first recommend the top reranked speechlet. If the user accepts the recommendation, the system may cause the top reranked speechlet to handle the original user input.
    Type: Grant
    Filed: April 13, 2018
    Date of Patent: February 16, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Srikanth Nori, Kevin Boehm
  • Patent number: 10896428
    Abstract: A dynamic contact management system is provided for analyzing aspects of speech from a customer contact and generating dynamic output based on the analysis. The system can process audio data representing utterances that occur during a customer contact, wherein the processing comprises generating scores based on agent and customer sentiments determined using their respective speech to text data. Based on the processing, the system can generate scores for use in dynamically determining which actions to take, updating displays, analyzing contact outcomes over time, etc.
    Type: Grant
    Filed: December 14, 2017
    Date of Patent: January 19, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Balarajan Balasubramaniam, Brian Benjamin Hartley, Mabel Karen Bo Ming Lee, Srdjan Majkic, Simon Gilles Meyffret, Mohammadrasool Raeesi Nafchi, Derek Tong
  • Patent number: 10896210
    Abstract: An information processing apparatus includes a memory, a reception unit, an acquisition unit, and a presentation unit. The memory stores, as history information, first information including attribute information of a user and service contents desired by the user, second information including service contents suggested to the service contents and a message prompting use of the service contents, and success/failure information indicating whether or not a service has been used and a successful result has been obtained. The reception unit receives the first information. The acquisition unit searches the history information for the first information with a similarity with respect to the received first information that reaches a threshold and acquires the second information corresponding to the found first information. The presentation unit extracts a word, based on the number of appearance times and the success/failure information, from the message included in the acquired second information and presents the word.
    Type: Grant
    Filed: April 2, 2018
    Date of Patent: January 19, 2021
    Assignee: FUJI XEROX CO., LTD.
    Inventor: Yuhei Shimada
  • Patent number: 10885910
    Abstract: Systems, methods, and computer-readable media are disclosed for systems and methods for voice-forward graphical user interface mode management. Example methods include determining that a device is coupled to an accessory device, determining that being coupled to the accessory device causes the device to deactivate a first operating mode and activate a second operating mode, where the second operating mode has a lower content density than the first operating mode, and determining that an application setting of an application executing on the device is causing the device to remain in the first operating mode. Example methods may include determining that a new value is associated with the application setting, and causing the device to activate the second operating mode.
    Type: Grant
    Filed: March 14, 2018
    Date of Patent: January 5, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Kynan Dylan Antos, Jenny Toi Wah Lam, Mark Yoshitake, Ankur Narendra Bhai Vachhani, Blade Imsande McRae, Robert Williams, James Martin Conway, Nedim Fresko, Michael Wendling, Mustafa Husain
  • Patent number: 10854197
    Abstract: Systems and methods for handling away messages with intelligent assistance using voice services. In some embodiments, an Information Handling System (IHS) may include: a processor; and a memory coupled to the processor, the memory having program instructions stored thereon that, upon execution, cause the IHS to: detect the presence of a person; output an audio greeting in response to the detection; receive an audio instruction in response to the audio greeting; transmit the audio instruction to a voice service provider, the voice service provider configured to: (i) convert the audio instruction into a text instruction, and (ii) transmit the text instruction to an intelligent assistance provider; receive a command from the intelligent assistance provider, the intelligent assistance provider configured to generate the command based upon the text instruction; and execute the command.
    Type: Grant
    Filed: May 1, 2018
    Date of Patent: December 1, 2020
    Assignee: Dell Products, L.P.
    Inventors: Marc Randall Hammons, Todd Erick Swierk, Tyler Ryan Cox
  • Patent number: 10825449
    Abstract: Systems and methods for providing characteristic analytics are provided. A data construct including a text object is received. A first subset of classification models in a plurality of classification models is applied to the data construct. The subset of classification models parse the text object into a plurality of text strings. Each respective classification model in the plurality of classification models parses a portion of the text object in accordance with a plurality of heuristic instructions associated with the respective classification model. The text object is evaluated using a reference database that includes a predetermined plurality of text strings. A characteristic of the data construct in the form of a result of the evaluating is provided. The evaluation step further comprises a second disjoint subset of classification models used for the evaluation of the parsed text object.
    Type: Grant
    Filed: October 21, 2019
    Date of Patent: November 3, 2020
    Assignee: CrowdAround Inc.
    Inventor: Anandan Chinnalagu
  • Patent number: 10818307
    Abstract: Method, apparatus, and storage medium for voice imitation are provided. The voice imitation method, includes: separately obtaining a training voice of a source user and training voices of a plurality of imitation users including a target user; determining, according to the training voice of the source user and a training voice of the target user, a conversion rule for converting the training voice of the source user into the training voice of the target user; collecting voice information of the source user; and converting the voice information of the source user into an imitation voice of the target user according to the conversion rule.
    Type: Grant
    Filed: January 11, 2018
    Date of Patent: October 27, 2020
    Inventors: Yuanyuan Liu, Guangjun Liu, Guoli Lu, Fen Fu
  • Patent number: 10796700
    Abstract: An artificial intelligence-based cross-language speech transcription method and apparatus, a device and a readable medium. The method includes pre-processing to-be-transcribed speech data to obtain multiple acoustic features, the to-be-transcribed speech data being represented in a first language; predicting a corresponding translation text after transcription of the speech data according to the multiple acoustic features and a pre-trained cross-language transcription model; wherein the translation text is represented in a second language which is different from the first language. According to the technical solution, it is unnecessary, upon cross-language speech transcription, to perform speech recognition first and then perform machine translation, but to directly perform cross-language transcription according to the pre-trained cross-language transcription model.
    Type: Grant
    Filed: May 14, 2018
    Date of Patent: October 6, 2020
    Inventors: Wei Zou, Xiangang Li, Bin Huang