Patents by Inventor Siyu DING

Siyu DING has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240412002
    Abstract: A method is provided. The method includes: obtaining a first sample dataset; inputting at least one first question text corresponding to at least one piece of first sample data into a dialog model separately to obtain at least one first answer prediction result; inputting each second question text into the dialog model to obtain a second answer prediction result output by the dialog model; inputting the second answer prediction result into a reward model to obtain a score of the second answer prediction result output by the reward model; determining a comprehensive loss based on the at least one first answer prediction result, a first answer text of each of the at least one piece of first sample data, and a score corresponding to each of at least one piece of second sample data; and adjusting at least one parameter of the dialog model based on the comprehensive loss.
    Type: Application
    Filed: June 19, 2024
    Publication date: December 12, 2024
    Applicant: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.
    Inventors: Yanbin ZHAO, Siyu DING, Shuohuan WANG, Yu SUN, Hao TIAN, Hua WU, Haifeng WANG
  • Publication number: 20240358654
    Abstract: The present disclosure discloses a preparation method of porous starch for encapsulating probiotic, and belongs to the field of food processing. The present disclosure provides porous starch with controllable pore size and morphology by regulating the ratio of amylose to amylopectin (quantitative compounding), keeping starch hydroxyl sites exposed (concentration cultivation), and building interfacial tension to stretch starch chains (convection drying), and uses the porous starch to encapsulate probiotics, which can improve the retention rate of probiotics, reduce the loss of probiotics during food processing and transportation, and thus retain the biological functions of probiotics to a maximum extent.
    Type: Application
    Filed: April 21, 2024
    Publication date: October 31, 2024
    Inventors: Enbo Xu, Qingqing Zhu, Jingsong Feng, Haibo Pan, Siyu Yao, Tian Ding, Donghong Liu, Xingqian Ye
  • Patent number: 12131728
    Abstract: The present application provides a method of training a natural language processing model, which relates to a field of artificial intelligence, and in particular to a field of natural language processing. A specific implementation scheme includes: performing a semantic learning for multi-tasks on an input text, so as to obtain a semantic feature for the multi-tasks, wherein the multi-tasks include a plurality of branch tasks; performing a feature learning for each branch task based on the semantic feature, so as to obtain a first output result for each branch task; calculating a loss for each branch task according to the first output result for the branch task; and adjusting a parameter of the natural language processing model according to the loss for each branch task. The present application further provides a method of processing a natural language, an electronic device, and a storage medium.
    Type: Grant
    Filed: May 31, 2022
    Date of Patent: October 29, 2024
    Assignee: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.
    Inventors: Siyu Ding, Chao Pang, Shuohuan Wang, Yanbin Zhao, Junyuan Shang, Yu Sun, Shikun Feng, Hao Tian, Hua Wu, Haifeng Wang
  • Patent number: 12106052
    Abstract: The disclosure discloses a method and an apparatus for generating a semantic representation model, and a storage medium. The detailed implementation includes: performing recognition and segmentation on the original text included in an original text set to obtain knowledge units and non-knowledge units in the original text; performing knowledge unit-level disorder processing on the knowledge units and the non-knowledge units in the original text to obtain a disorder text; generating a training text set based on the character attribute of each character in the disorder text; and training an initial semantic representation model by employing the training text set to generate the semantic representation model.
    Type: Grant
    Filed: March 18, 2021
    Date of Patent: October 1, 2024
    Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.
    Inventors: Shuohuan Wang, Siyu Ding, Yu Sun
  • Patent number: 12086831
    Abstract: The present teaching relates to placing sponsored search results based on correlation of the sponsored search results. A search query is first received at a search engine from a user. One or more keywords are further extracted from the search query. A plurality of sponsored search results related to the one or more keywords are received in response to the search query. The placement of the plurality of sponsored search results are further determined based on correlation of the plurality of sponsored search results, and a search results page containing the plurality of sponsored search results are presented to the user.
    Type: Grant
    Filed: April 27, 2021
    Date of Patent: September 10, 2024
    Assignee: YAHOO ASSETS LLC
    Inventors: Shanshan Zhang, Qingquan Wang, Siyu Zhu, Zhuoye Ding
  • Publication number: 20230252354
    Abstract: A method for pre-training a language model includes: constructing a pre-training language data set, in which the pre-training language data set comprises unsupervised language data and supervised language data; generating a hierarchical multi-template and multi-task language data set based on the pre-training language data set; and pre-training the language model based on the hierarchical multi-template and multi-task language data set.
    Type: Application
    Filed: March 7, 2023
    Publication date: August 10, 2023
    Inventors: Junyuan SHANG, Shuohuan WANG, Siyu DING, Yanbin ZHAO, Chao PANG, Yu SUN, Hao TIAN, Hua WU, Haifeng WANG
  • Publication number: 20230206080
    Abstract: A model training system includes at least one first cluster and a second cluster communicating with the at least first cluster. The at least one first cluster is configured to acquire a sample data set, generate training data according to the sample data set, and send the training data to the second cluster; and the second cluster is configured to train a pre-trained model according to the training data sent by the at least one first cluster.
    Type: Application
    Filed: March 7, 2023
    Publication date: June 29, 2023
    Inventors: Shuohuan WANG, Weibao GONG, Zhihua WU, Yu SUN, Siyu DING, Yaqian HAN, Yanbin ZHAO, Yuang LIU, Dianhai YU
  • Patent number: 11663404
    Abstract: The disclosure provides a text recognition method, an electronic device, and a storage medium. The method includes: obtaining N segments of a sample text; inputting each of the N segments into a preset initial language model in sequence, to obtain first text vector information corresponding to the N segments; inputting each of the N segments into the initial language model in sequence again, to obtain second text vector information corresponding to a currently input segment; in response to determining that the currently input segment has the mask, predicting the mask according to the second text vector information and the first text vector information to obtain a predicted word at a target position corresponding to the mask; training the initial language model according to an original word and the predicted word to generate a long text language model; and recognizing an input text through the long text language model.
    Type: Grant
    Filed: November 23, 2020
    Date of Patent: May 30, 2023
    Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.
    Inventors: Shuohuan Wang, Siyu Ding, Yu Sun, Hua Wu, Haifeng Wang
  • Publication number: 20230040095
    Abstract: A method and apparatus for pre-training a model, a device, a storage medium, and a program product. An embodiment of the method includes: acquiring a sample natural language text; generating N types of prompt words based on the sample natural language text, where N is a positive integer; generating sample input data based on the sample natural language text and the N types of prompt words; and training an initial language model based on the sample input data, to obtain a pre-trained language model.
    Type: Application
    Filed: August 16, 2022
    Publication date: February 9, 2023
    Inventors: Junyuan SHANG, Shuohuan WANG, Siyu DING, Yanbin ZHAO, Chao PANG, Yu Sun
  • Publication number: 20220327290
    Abstract: There is provided a method of training a feature determination model, which relates to a field of deep learning and natural language processing. The method is implemented to include: determining, by a plurality of feature determination layers arranged in stages, a feature vector for each segment in a pre-training text; and pre-training the feature determination model according to the feature vector. A current stage feature vector is determined by a feature determination layer of a current stage according to a preceding segment feature vector determined for a preceding segment, and a preceding stage feature vector determined by a feature determination layer of a preceding stage. A method of training a feature determination model for a target task, a method of performing semantic analysis for a target task, an electronic device, and a computer storage medium are also provided.
    Type: Application
    Filed: June 29, 2022
    Publication date: October 13, 2022
    Inventors: Junyuan SHANG, Shuohuan WANG, Siyu DING
  • Publication number: 20220293092
    Abstract: The present application provides a method of training a natural language processing model, which relates to a field of artificial intelligence, and in particular to a field of natural language processing. A specific implementation scheme includes: performing a semantic learning for multi-tasks on an input text, so as to obtain a semantic feature for the multi-tasks, wherein the multi-tasks include a plurality of branch tasks; performing a feature learning for each branch task based on the semantic feature, so as to obtain a first output result for each branch task; calculating a loss for each branch task according to the first output result for the branch task; and adjusting a parameter of the natural language processing model according to the loss for each branch task. The present application further provides a method of processing a natural language, an electronic device, and a storage medium.
    Type: Application
    Filed: May 31, 2022
    Publication date: September 15, 2022
    Inventors: Siyu DING, Chao PANG, Shuohuan WANG, Yanbin ZHAO, Junyuan SHANG, Yu SUN, Shikun FENG, Hao TIAN, Hua WU, Haifeng WANG
  • Publication number: 20210383064
    Abstract: The disclosure provides a text recognition method, an electronic device, and a storage medium. The method includes: obtaining N segments of a sample text; inputting each of the N segments into a preset initial language model in sequence, to obtain first text vector information corresponding to the N segments; inputting each of the N segments into the initial language model in sequence again, to obtain second text vector information corresponding to a currently input segment; in response to determining that the currently input segment has the mask, predicting the mask according to the second text vector information and the first text vector information to obtain a predicted word at a target position corresponding to the mask; training the initial language model according to an original word and the predicted word to generate a long text language model; and recognizing an input text through the long text language model.
    Type: Application
    Filed: November 23, 2020
    Publication date: December 9, 2021
    Inventors: Shuohuan WANG, Siyu DING, Yu SUN, Hua WU, Haifeng WANG
  • Publication number: 20210334659
    Abstract: The present application discloses a method and an apparatus for adversarial training of a machine learning (ML) model and a medium. The method includes: obtaining input information in a training sample; extracting features of a plurality of input characters in the input information; inputting the features of the plurality of input characters to the ML model, to capture an attention weight on an input character of the plurality of input characters by an attention layer of the ML model; disturbing the attention weight captured by the attention layer, so that the ML model outputs a predicted character according to the attention weight disturbed; and training the ML model according to a difference between the predicted character and a labeled character in the training sample.
    Type: Application
    Filed: July 7, 2021
    Publication date: October 28, 2021
    Inventors: Siyu DING, Shuohuan WANG, Yu SUN
  • Publication number: 20210312139
    Abstract: A method and apparatus of generating a semantic feature, a method and apparatus of training a model, an electronic device, and a storage medium are provided. The method of generating the semantic feature includes: segmenting a target document to obtain a segment sequence of the target document; generating a semantic feature of each document segment in the segment sequence of the target document by using a pre-trained bidirectional semantic encoding model; and acquiring the semantic feature of the target document based on the semantic feature of the each document segment in the segment sequence of the target document. The present disclosure further provides a method of training a bidirectional semantic encoding model.
    Type: Application
    Filed: June 22, 2021
    Publication date: October 7, 2021
    Inventors: Shuohuan WANG, Siyu DING, Junyuan SHANG, Yu SUN
  • Publication number: 20210248484
    Abstract: The disclosure discloses a method and an apparatus for generating a semantic representation model, and a storage medium. The detailed implementation includes: performing recognition and segmentation on the original text included in an original text set to obtain knowledge units and non-knowledge units in the original text; performing knowledge unit-level disorder processing on the knowledge units and the non-knowledge units in the original text to obtain a disorder text; generating a training text set based on the character attribute of each character in the disorder text; and training an initial semantic representation model by employing the training text set to generate the semantic representation model.
    Type: Application
    Filed: March 18, 2021
    Publication date: August 12, 2021
    Inventors: Shuohuan WANG, Siyu DING, Yu SUN