Patents by Inventor Dongyan Huang

Dongyan Huang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 12385015
    Abstract: Provided is a method for constructing a gE protein-deleted pseudorabies virus (PRV) strain using an adenine base editor (ABE) and use thereof. The method includes: designing an sgRNA sequence using the ABE with a start codon of the gE gene in a PRV as a target site, ligating an enzyme-digested fragment to a double-stranded DNA fragment with sticky ends to obtain a ligation product; and transforming the ligation product into a competent cell to allow plate screening and culture, selecting a resulting positive bacterial strain to allow expanded culture, and extracting a plasmid from a resulting positive bacterial solution; and transferring the plasmid into a target cell to allow the transfection for 24 h, collecting a resulting virus liquid, and centrifuging the virus liquid to collect a resulting supernatant.
    Type: Grant
    Filed: October 1, 2024
    Date of Patent: August 12, 2025
    Assignee: JIANGXI AGRICULTURAL UNIVERSITY
    Inventors: Yu Ye, Peixia Wang, Yuxin Tang, Chuan Zeng, Jun Gu, Jinyan Shen, Yiwen Duan, Yuwei Bai, Dongyan Huang, Deping Song
  • Patent number: 12374319
    Abstract: A speech synthesis method includes: obtaining an acoustic feature sequence of a text to be processed; processing the acoustic feature sequence by using a non-autoregressive computing model in parallel to obtain first audio information of the text to be processed, wherein the first audio information comprises audio corresponding to each segment; processing the acoustic feature sequence and the first audio information by using an autoregressive computing model to obtain a residual value corresponding to each segment; and obtaining second audio information corresponding to an i-th segment based on the first audio information corresponding to the i-th segment and the residual values corresponding to a first to an (i?1)-th segment, wherein a synthesized audio of the text to be processed comprises each of the second audio information, i=1, 2 . . . n, n is a total number of the segments.
    Type: Grant
    Filed: December 28, 2022
    Date of Patent: July 29, 2025
    Assignee: UBTECH ROBOTICS CORP LTD
    Inventors: Wan Ding, Dongyan Huang, Zhiyuan Zhao, Zhiyong Yang
  • Patent number: 12315059
    Abstract: A method for generating a talking head video includes: obtaining a text and an image containing a face of a user; determining a phoneme sequence that corresponds to the text and includes one or more phonemes; determining acoustic features corresponding to the text according to the phoneme sequence, and obtaining synthesized speech corresponding to the text according to the acoustic features; determining a first mouth movement sequence corresponding to the text according to the phoneme sequence, and determining a second mouth movement sequence corresponding to the text according to the acoustic features; creating a facial action video corresponding to the user according to the first mouth movement sequence, the second mouth movement sequence and the image; and processing the synthesized speech and the facial action video synchronously to obtain a talking head video corresponding to the user.
    Type: Grant
    Filed: May 26, 2023
    Date of Patent: May 27, 2025
    Assignee: UBTECH ROBOTICS CORP LTD
    Inventors: Wan Ding, Dongyan Huang, Linhuang Yan, Zhiyong Yang
  • Publication number: 20250136951
    Abstract: Provided is a method for constructing a gE gene-deleted pseudorabies virus (PRV) strain using an adenine base editor (ABE) and use thereof. The method includes: designing an sgRNA sequence using the ABE with a start codon of the gE gene in a PRV as a target site, ligating an enzyme-digested fragment to a double-stranded DNA fragment with sticky ends to obtain a ligation product; and transforming the ligation product into a competent cell to allow plate screening and culture, selecting a resulting positive bacterial strain to allow expanded culture, and extracting a plasmid from a resulting positive bacterial solution; and transferring the plasmid into a target cell to allow the transfection for 24 h, collecting a resulting virus liquid, and centrifuging the virus liquid to collect a resulting supernatant.
    Type: Application
    Filed: October 1, 2024
    Publication date: May 1, 2025
    Applicant: JIANGXI AGRICULTURAL UNIVERSITY
    Inventors: Yu YE, Peixia WANG, Yuxin TANG, Chuan ZENG, Jun GU, Jinyan SHEN, Yiwen DUAN, Yuwei BAI, Dongyan HUANG, Deping SONG
  • Publication number: 20250133337
    Abstract: A sound source localization method includes: obtaining a first audio frame and at least two second audio frames, wherein the first audio frame and the at least two second audio frames are synchronously sampled, the first audio frame is obtained by processing sound signals collected by the first microphone, the at least two second audio frames are obtained by processing sound signals collected by the second microphones; calculating a time delay estimation between the first audio frame and each of the at least two second audio frames; and determining a sound source orientation corresponding to the first audio frame and the at least two second audio frames through a preset time delay-orientation lookup table according to the time delay estimation between the first audio frame and each of the at least two second audio frames.
    Type: Application
    Filed: October 9, 2024
    Publication date: April 24, 2025
    Inventors: ZEHONG ZHENG, Dongyan Huang, Xianjie Yang, Wan Ding
  • Publication number: 20240428493
    Abstract: A method for synthesizing a talking head video includes: obtaining speech data to be synthesized and observation data, wherein the observation data is data obtained through observation other than the speech data; performing feature extraction on the speech data to obtain speech features corresponding to the speech data, and performing feature extraction on the observation data to obtain non-speech features corresponding to the observation data; performing temporal modeling on the speech features and first non-speech features to obtain low-dimensional representations, wherein the first non-speech features are non-speech features that are sensitive to temporal changes; and performing video synthesis based on the low-dimensional representations and second non-speech features, wherein the second non-speech features are non-speech features insensitive to temporal changes.
    Type: Application
    Filed: June 7, 2024
    Publication date: December 26, 2024
    Inventors: WAN DING, Dongyan Huang, Xianjie Yang, Zehong Zheng, Penghul Li
  • Patent number: 12129517
    Abstract: The present disclosure provides a kit for detecting an African swine fever virus (ASFV), and provides a corresponding single guide RNA (sgRNA) with a nucleic acid sequence shown in any one of SEQ ID NO: 1 to SEQ ID NO: 5. The kit is based on loop-mediated isothermal amplification (LAMP)-clustered regularly interspaced short palindromic repeat (CRISPR)/Cas12b and can detect the ASFV in one tube at a constant temperature. The kit only needs to set one reaction temperature, does not open its lid midway, and has high sensitivity and specificity while showing no specificity to other swine viruses. The kit exhibits high efficiency and convenience, and does not rely on a large-scale experimental equipment. Compared with the traditional fluorescence quantitative PCR methods, the kit has a greatly improved sensitivity in detecting 1 copy/?L. The kit realizes visual detection by combining with colloidal gold test strip detection.
    Type: Grant
    Filed: April 8, 2024
    Date of Patent: October 29, 2024
    Assignee: Jiangxi Agricultural University
    Inventors: Yu Ye, Ning Jiang, Yuxin Tang, Dongyan Huang
  • Patent number: 11996112
    Abstract: The present disclosure discloses a voice conversion method. The method includes: obtaining a to-be-converted voice, and extracting acoustic features of the to-be-converted voice; obtaining a source vector corresponding to the to-be-converted voice from a source vector pool, and selecting a target vector corresponding to the target voice from the target vector pool; obtaining acoustic features of the target voice output by the voice conversion model by using the acoustic features of the to-be-converted voice, the source vector corresponding to the to-be-converted voice, and the target vector corresponding to the target voice as an input of the voice conversion model; and obtaining the target voice by converting the acoustic features of the target voice using a vocoder. In addition, a voice conversion apparatus and a storage medium are also provided.
    Type: Grant
    Filed: October 30, 2020
    Date of Patent: May 28, 2024
    Assignee: UBTECH ROBOTICS CORP LTD
    Inventors: Ruotong Wang, Zhichao Tang, Dongyan Huang, Jiebin Xie, Zhiyuan Zhao, Yang Liu, Youjun Xiong
  • Patent number: 11941366
    Abstract: The present disclosure discloses a context-based multi-turn dialogue method.
    Type: Grant
    Filed: November 23, 2020
    Date of Patent: March 26, 2024
    Assignee: UBTECH ROBOTICS CORP LTD
    Inventors: Chi Shao, Dongyan Huang, Wan Ding, Youjun Xiong
  • Publication number: 20230386116
    Abstract: A method for generating a talking head video includes: obtaining a text and an image containing a face of a user; determining a phoneme sequence that corresponds to the text and includes one or more phonemes; determining acoustic features corresponding to the text according to the phoneme sequence, and obtaining synthesized speech corresponding to the text according to the acoustic features; determining a first mouth movement sequence corresponding to the text according to the phoneme sequence, and determining a second mouth movement sequence corresponding to the text according to the acoustic features; creating a facial action video corresponding to the user according to the first mouth movement sequence, the second mouth movement sequence and the image; and processing the synthesized speech and the facial action video synchronously to obtain a talking head video corresponding to the user.
    Type: Application
    Filed: May 26, 2023
    Publication date: November 30, 2023
    Inventors: WAN DING, Dongyan Huang, Linhuang Yan, Zhiyong Yang
  • Patent number: 11763796
    Abstract: A computer-implemented method for speech synthesis, a computer device, and a non-transitory computer readable storage medium are provided. The method includes: obtaining a speech text to be synthesized; obtaining a Mel spectrum corresponding to the speech text to be synthesized according to the speech text to be synthesized; inputting the Mel spectrum into a complex neural network, and obtaining a complex spectrum corresponding to the speech text to be synthesized, wherein the complex spectrum comprises real component information and imaginary component information; and obtaining a synthetic speech corresponding to the speech text to be synthesized, according to the complex spectrum. The method can efficiently and simply complete speech synthesis.
    Type: Grant
    Filed: December 10, 2020
    Date of Patent: September 19, 2023
    Assignee: UBTECH ROBOTICS CORP LTD
    Inventors: Dongyan Huang, Leyuan Sheng, Youjun Xiong
  • Publication number: 20230206895
    Abstract: A speech synthesis method includes: obtaining an acoustic feature sequence of a text to be processed; processing the acoustic feature sequence by using a non-autoregressive computing model in parallel to obtain first audio information of the text, to be processed, wherein the first audio information comprises audio corresponding to each segment; processing the acoustic feature sequence and the first audio information by using an autoregressive computing model to obtain a residual value corresponding to each segment; and obtaining second audio information corresponding to an i-th segment based on the first audio information corresponding to the i-th segment and the residual values corresponding to a first to an (i-1)-th segment, wherein a synthesized audio of the text to be processed comprises each of the second audio information, i=1, 2 . . . n, n is a total number of the segments.
    Type: Application
    Filed: December 28, 2022
    Publication date: June 29, 2023
    Inventors: Wan Ding, Dongyan Huang, Zhiyuan Zhao, Zhiyong Yang
  • Patent number: 11645474
    Abstract: A computer-implemented method for text conversion, a computer device, and a non-transitory computer readable storage medium are provided. The method includes: obtaining a text to be converted; performing a non-standard word recognition on the text to be converted, to determine whether the text to be converted includes a non-standard word; recognizing the non-standard word in the text to be converted by using an eXtreme Gradient Boosting model in response to the text to be converted including the non-standard word; and obtaining a target converted text corresponding to the text to be converted, according to a recognition result outputted by the eXtreme Gradient Boosting model. The method has a faster recognition speed and a higher recognition accuracy compared with the deep learning model.
    Type: Grant
    Filed: December 24, 2020
    Date of Patent: May 9, 2023
    Assignee: UBTECH ROBOTICS CORP LTD
    Inventors: Zhongfa Feng, Dongyan Huang, Youjun Xiong
  • Patent number: 11417316
    Abstract: The present disclosure provides a speech synthesis method as well as an apparatus and a computer readable storage medium using the same. The method includes: obtaining a to-be-synthesized text, and extracting to-be-processed Mel spectrum features of the to-be-synthesized text through a preset speech feature extraction algorithm; inputting the to-be-processed Mel spectrum features into a preset ResUnet network model to obtain first intermediate features; performing an average pooling and a first down sampling on the to-be-processed Mel spectrum features to obtain second intermediate features; taking the second intermediate features and the first intermediate features output by the ResUnet network model as an input to perform a deconvolution and a first up sampling so as to obtain target Mel spectrum features corresponding to the to-be-processed Mel spectrum features; and converting the target Mel spectrum features into a target speech corresponding to the to-be-synthesized text.
    Type: Grant
    Filed: December 8, 2020
    Date of Patent: August 16, 2022
    Assignee: UBTECH ROBOTICS CORP LTD
    Inventors: Dongyan Huang, Leyuan Sheng, Youjun Xiong
  • Patent number: 11367456
    Abstract: The present disclosure provides a streaming voice conversion method as well as an apparatus and a computer readable storage medium using the same. The method includes: obtaining to-be-converted voice data; partitioning the to-be-converted voice data in an order of data obtaining time as a plurality of to-be-converted partition voices, where the to-be-converted partition voice data carries a partition mark; performing a voice conversion on each of the to-be-converted partition voices to obtain a converted partition voice, where the converted partition voice carries a partition mark; performing a partition restoration on each of the converted partition voices to obtain a restored partition voice, where the restored partition voice carries a partition mark; and outputting each of the restored partition voices according to the partition mark carried by the restored partition voice. In this manner, the response time is shortened, and the conversion speed is improved.
    Type: Grant
    Filed: December 3, 2020
    Date of Patent: June 21, 2022
    Assignee: UBTECH ROBOTICS CORP LTD
    Inventors: Jiebin Xie, Ruotong Wang, Dongyan Huang, Zhichao Tang, Yang Liu, Youjun Xiong
  • Publication number: 20220189454
    Abstract: A computer-implemented method for speech synthesis, a computer device, and a non-transitory computer readable storage medium are provided. The method includes: obtaining a speech text to be synthesized; obtaining a Mel spectrum corresponding to the speech text to be synthesized according to the speech text to be synthesized; inputting the Mel spectrum into a complex neural network, and obtaining a complex spectrum corresponding to the speech text to be synthesized, wherein the complex spectrum comprises real component information and imaginary component information; and obtaining a synthetic speech corresponding to the speech text to be synthesized, according to the complex spectrum. The method can efficiently and simply complete speech synthesis.
    Type: Application
    Filed: December 10, 2020
    Publication date: June 16, 2022
    Inventors: Dongyan Huang, Leyuan Sheng, Youjun Xiong
  • Patent number: 11282503
    Abstract: The present disclosure discloses a voice conversion training method. The method includes: forming a first training data set including a plurality of training voice data groups; selecting two of the training voice data groups from the first training data set to input into a voice conversion neural network for training; forming a second training data set including the first training data set and a first source speaker voice data group; inputting one of the training voice data groups selected from the first training data set and the first source speaker voice data group into the network for training; forming the third training data set including the second source speaker voice data group and the personalized voice data group that are parallel corpus with respect to each other; and inputting the second source speaker voice data group and the personalized voice data group into the network for training.
    Type: Grant
    Filed: November 12, 2020
    Date of Patent: March 22, 2022
    Assignee: UBTECH ROBOTICS CORP LTD
    Inventors: Ruotong Wang, Dongyan Huang, Xian Li, Jiebin Xie, Zhichao Tang, Wan Ding, Yang Liu, Bai Li, Youjun Xiong
  • Publication number: 20210200961
    Abstract: The present disclosure discloses a context-based multi-turn dialogue method.
    Type: Application
    Filed: November 23, 2020
    Publication date: July 1, 2021
    Inventors: Chi Shao, Dongyan Huang, Wan Ding, Youjun Xiong
  • Publication number: 20210201925
    Abstract: The present disclosure provides a streaming voice conversion method as well as an apparatus and a computer readable storage medium using the same. The method includes: obtaining to-be-converted voice data; partitioning the to-be-converted voice data in an order of data obtaining time as a plurality of to-be-converted partition voices, where the to-be-converted partition voice data carries a partition mark; performing a voice conversion on each of the to-be-converted partition voices to obtain a converted partition voice, where the converted partition voice carries a partition mark; performing a partition restoration on each of the converted partition voices to obtain a restored partition voice, where the restored partition voice carries a partition mark; and outputting each of the restored partition voices according to the partition mark carried by the restored partition voice. In this manner, the response time is shortened, and the conversion speed is improved.
    Type: Application
    Filed: December 3, 2020
    Publication date: July 1, 2021
    Inventors: Jiebin Xie, Ruotong Wang, Dongyan Huang, Zhichao Tang, Yang Liu, Youjun Xiong
  • Publication number: 20210200962
    Abstract: A computer-implemented method for text conversion, a computer device, and a non-transitory computer readable storage medium are provided. The method includes: obtaining a text to be converted; performing a non-standard word recognition on the text to be converted, to determine whether the text to be converted includes a non-standard word; recognizing the non-standard word in the text to be converted by using an eXtreme Gradient Boosting model in response to the text to be converted including the non-standard word; and obtaining a target converted text corresponding to the text to be converted, according to a recognition result outputted by the eXtreme Gradient Boosting model. The method has a faster recognition speed and a higher recognition accuracy compared with the deep learning model.
    Type: Application
    Filed: December 24, 2020
    Publication date: July 1, 2021
    Inventors: Zhongfa Feng, Dongyan Huang, Youjun Xiong