Patents by Inventor Siyu DING
Siyu DING has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240412002Abstract: A method is provided. The method includes: obtaining a first sample dataset; inputting at least one first question text corresponding to at least one piece of first sample data into a dialog model separately to obtain at least one first answer prediction result; inputting each second question text into the dialog model to obtain a second answer prediction result output by the dialog model; inputting the second answer prediction result into a reward model to obtain a score of the second answer prediction result output by the reward model; determining a comprehensive loss based on the at least one first answer prediction result, a first answer text of each of the at least one piece of first sample data, and a score corresponding to each of at least one piece of second sample data; and adjusting at least one parameter of the dialog model based on the comprehensive loss.Type: ApplicationFiled: June 19, 2024Publication date: December 12, 2024Applicant: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.Inventors: Yanbin ZHAO, Siyu DING, Shuohuan WANG, Yu SUN, Hao TIAN, Hua WU, Haifeng WANG
-
Publication number: 20240358654Abstract: The present disclosure discloses a preparation method of porous starch for encapsulating probiotic, and belongs to the field of food processing. The present disclosure provides porous starch with controllable pore size and morphology by regulating the ratio of amylose to amylopectin (quantitative compounding), keeping starch hydroxyl sites exposed (concentration cultivation), and building interfacial tension to stretch starch chains (convection drying), and uses the porous starch to encapsulate probiotics, which can improve the retention rate of probiotics, reduce the loss of probiotics during food processing and transportation, and thus retain the biological functions of probiotics to a maximum extent.Type: ApplicationFiled: April 21, 2024Publication date: October 31, 2024Inventors: Enbo Xu, Qingqing Zhu, Jingsong Feng, Haibo Pan, Siyu Yao, Tian Ding, Donghong Liu, Xingqian Ye
-
Patent number: 12131728Abstract: The present application provides a method of training a natural language processing model, which relates to a field of artificial intelligence, and in particular to a field of natural language processing. A specific implementation scheme includes: performing a semantic learning for multi-tasks on an input text, so as to obtain a semantic feature for the multi-tasks, wherein the multi-tasks include a plurality of branch tasks; performing a feature learning for each branch task based on the semantic feature, so as to obtain a first output result for each branch task; calculating a loss for each branch task according to the first output result for the branch task; and adjusting a parameter of the natural language processing model according to the loss for each branch task. The present application further provides a method of processing a natural language, an electronic device, and a storage medium.Type: GrantFiled: May 31, 2022Date of Patent: October 29, 2024Assignee: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.Inventors: Siyu Ding, Chao Pang, Shuohuan Wang, Yanbin Zhao, Junyuan Shang, Yu Sun, Shikun Feng, Hao Tian, Hua Wu, Haifeng Wang
-
Patent number: 12106052Abstract: The disclosure discloses a method and an apparatus for generating a semantic representation model, and a storage medium. The detailed implementation includes: performing recognition and segmentation on the original text included in an original text set to obtain knowledge units and non-knowledge units in the original text; performing knowledge unit-level disorder processing on the knowledge units and the non-knowledge units in the original text to obtain a disorder text; generating a training text set based on the character attribute of each character in the disorder text; and training an initial semantic representation model by employing the training text set to generate the semantic representation model.Type: GrantFiled: March 18, 2021Date of Patent: October 1, 2024Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.Inventors: Shuohuan Wang, Siyu Ding, Yu Sun
-
Patent number: 12086831Abstract: The present teaching relates to placing sponsored search results based on correlation of the sponsored search results. A search query is first received at a search engine from a user. One or more keywords are further extracted from the search query. A plurality of sponsored search results related to the one or more keywords are received in response to the search query. The placement of the plurality of sponsored search results are further determined based on correlation of the plurality of sponsored search results, and a search results page containing the plurality of sponsored search results are presented to the user.Type: GrantFiled: April 27, 2021Date of Patent: September 10, 2024Assignee: YAHOO ASSETS LLCInventors: Shanshan Zhang, Qingquan Wang, Siyu Zhu, Zhuoye Ding
-
Publication number: 20230252354Abstract: A method for pre-training a language model includes: constructing a pre-training language data set, in which the pre-training language data set comprises unsupervised language data and supervised language data; generating a hierarchical multi-template and multi-task language data set based on the pre-training language data set; and pre-training the language model based on the hierarchical multi-template and multi-task language data set.Type: ApplicationFiled: March 7, 2023Publication date: August 10, 2023Inventors: Junyuan SHANG, Shuohuan WANG, Siyu DING, Yanbin ZHAO, Chao PANG, Yu SUN, Hao TIAN, Hua WU, Haifeng WANG
-
Publication number: 20230206080Abstract: A model training system includes at least one first cluster and a second cluster communicating with the at least first cluster. The at least one first cluster is configured to acquire a sample data set, generate training data according to the sample data set, and send the training data to the second cluster; and the second cluster is configured to train a pre-trained model according to the training data sent by the at least one first cluster.Type: ApplicationFiled: March 7, 2023Publication date: June 29, 2023Inventors: Shuohuan WANG, Weibao GONG, Zhihua WU, Yu SUN, Siyu DING, Yaqian HAN, Yanbin ZHAO, Yuang LIU, Dianhai YU
-
Patent number: 11663404Abstract: The disclosure provides a text recognition method, an electronic device, and a storage medium. The method includes: obtaining N segments of a sample text; inputting each of the N segments into a preset initial language model in sequence, to obtain first text vector information corresponding to the N segments; inputting each of the N segments into the initial language model in sequence again, to obtain second text vector information corresponding to a currently input segment; in response to determining that the currently input segment has the mask, predicting the mask according to the second text vector information and the first text vector information to obtain a predicted word at a target position corresponding to the mask; training the initial language model according to an original word and the predicted word to generate a long text language model; and recognizing an input text through the long text language model.Type: GrantFiled: November 23, 2020Date of Patent: May 30, 2023Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.Inventors: Shuohuan Wang, Siyu Ding, Yu Sun, Hua Wu, Haifeng Wang
-
Publication number: 20230040095Abstract: A method and apparatus for pre-training a model, a device, a storage medium, and a program product. An embodiment of the method includes: acquiring a sample natural language text; generating N types of prompt words based on the sample natural language text, where N is a positive integer; generating sample input data based on the sample natural language text and the N types of prompt words; and training an initial language model based on the sample input data, to obtain a pre-trained language model.Type: ApplicationFiled: August 16, 2022Publication date: February 9, 2023Inventors: Junyuan SHANG, Shuohuan WANG, Siyu DING, Yanbin ZHAO, Chao PANG, Yu Sun
-
Publication number: 20220327290Abstract: There is provided a method of training a feature determination model, which relates to a field of deep learning and natural language processing. The method is implemented to include: determining, by a plurality of feature determination layers arranged in stages, a feature vector for each segment in a pre-training text; and pre-training the feature determination model according to the feature vector. A current stage feature vector is determined by a feature determination layer of a current stage according to a preceding segment feature vector determined for a preceding segment, and a preceding stage feature vector determined by a feature determination layer of a preceding stage. A method of training a feature determination model for a target task, a method of performing semantic analysis for a target task, an electronic device, and a computer storage medium are also provided.Type: ApplicationFiled: June 29, 2022Publication date: October 13, 2022Inventors: Junyuan SHANG, Shuohuan WANG, Siyu DING
-
Publication number: 20220293092Abstract: The present application provides a method of training a natural language processing model, which relates to a field of artificial intelligence, and in particular to a field of natural language processing. A specific implementation scheme includes: performing a semantic learning for multi-tasks on an input text, so as to obtain a semantic feature for the multi-tasks, wherein the multi-tasks include a plurality of branch tasks; performing a feature learning for each branch task based on the semantic feature, so as to obtain a first output result for each branch task; calculating a loss for each branch task according to the first output result for the branch task; and adjusting a parameter of the natural language processing model according to the loss for each branch task. The present application further provides a method of processing a natural language, an electronic device, and a storage medium.Type: ApplicationFiled: May 31, 2022Publication date: September 15, 2022Inventors: Siyu DING, Chao PANG, Shuohuan WANG, Yanbin ZHAO, Junyuan SHANG, Yu SUN, Shikun FENG, Hao TIAN, Hua WU, Haifeng WANG
-
Publication number: 20210383064Abstract: The disclosure provides a text recognition method, an electronic device, and a storage medium. The method includes: obtaining N segments of a sample text; inputting each of the N segments into a preset initial language model in sequence, to obtain first text vector information corresponding to the N segments; inputting each of the N segments into the initial language model in sequence again, to obtain second text vector information corresponding to a currently input segment; in response to determining that the currently input segment has the mask, predicting the mask according to the second text vector information and the first text vector information to obtain a predicted word at a target position corresponding to the mask; training the initial language model according to an original word and the predicted word to generate a long text language model; and recognizing an input text through the long text language model.Type: ApplicationFiled: November 23, 2020Publication date: December 9, 2021Inventors: Shuohuan WANG, Siyu DING, Yu SUN, Hua WU, Haifeng WANG
-
Publication number: 20210334659Abstract: The present application discloses a method and an apparatus for adversarial training of a machine learning (ML) model and a medium. The method includes: obtaining input information in a training sample; extracting features of a plurality of input characters in the input information; inputting the features of the plurality of input characters to the ML model, to capture an attention weight on an input character of the plurality of input characters by an attention layer of the ML model; disturbing the attention weight captured by the attention layer, so that the ML model outputs a predicted character according to the attention weight disturbed; and training the ML model according to a difference between the predicted character and a labeled character in the training sample.Type: ApplicationFiled: July 7, 2021Publication date: October 28, 2021Inventors: Siyu DING, Shuohuan WANG, Yu SUN
-
Publication number: 20210312139Abstract: A method and apparatus of generating a semantic feature, a method and apparatus of training a model, an electronic device, and a storage medium are provided. The method of generating the semantic feature includes: segmenting a target document to obtain a segment sequence of the target document; generating a semantic feature of each document segment in the segment sequence of the target document by using a pre-trained bidirectional semantic encoding model; and acquiring the semantic feature of the target document based on the semantic feature of the each document segment in the segment sequence of the target document. The present disclosure further provides a method of training a bidirectional semantic encoding model.Type: ApplicationFiled: June 22, 2021Publication date: October 7, 2021Inventors: Shuohuan WANG, Siyu DING, Junyuan SHANG, Yu SUN
-
Publication number: 20210248484Abstract: The disclosure discloses a method and an apparatus for generating a semantic representation model, and a storage medium. The detailed implementation includes: performing recognition and segmentation on the original text included in an original text set to obtain knowledge units and non-knowledge units in the original text; performing knowledge unit-level disorder processing on the knowledge units and the non-knowledge units in the original text to obtain a disorder text; generating a training text set based on the character attribute of each character in the disorder text; and training an initial semantic representation model by employing the training text set to generate the semantic representation model.Type: ApplicationFiled: March 18, 2021Publication date: August 12, 2021Inventors: Shuohuan WANG, Siyu DING, Yu SUN