Patents by Inventor Xinying Song
Xinying Song has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250045534Abstract: Implementations relate to a method implemented by one or more processors, the method including: receiving natural language (NL) based input associated with a client device; generating, using a large language model (LLM) and based on processing the NL based input, LLM output; determining, based on the LLM output, a sequence of LLM responses, the sequence of LLM responses including at least one intermediate LLM response and a final LLM response. In some implementations, the method may further include causing the final LLM response to be rendered at the client device. In additional or alternative implementations, the method may further include storing, as an instance of training data for fine-tuning the LLM or an additional LLM, the NL based input along with the final LLM response.Type: ApplicationFiled: October 10, 2023Publication date: February 6, 2025Inventors: Swaroop Mishra, Ragha Kotikalapudi, Sahitya Potluri, Taylor Bos, YaGuang Li, Hanzhao Lin, Steven Zheng, Yu Du, Chen Zhu, Chenkai Kuang, Xinying Song, Heng-Tze Cheng, Ed H. Chi, Quoc Le
-
Publication number: 20240362093Abstract: At least utilizing a custom corpus of documents to condition a large language model (LLM) when generating a response to a user query. In some implementations, a user query associated with a client device is received. An API query for an external application is generated by an LLM based on the user query. The external application has access to a custom corpus of documents comprising a plurality of documents. The external application is queried using the API query. Data representative of one or more documents in the custom corpus of documents is received from the external application in response to the API query. The LLM generates a response to the query that is conditioned on the data representing one or more of the documents in the custom corpus of documents received from the external application. The response to the user query is caused to be rendered on the client device.Type: ApplicationFiled: August 8, 2023Publication date: October 31, 2024Inventors: Hao Zhou, Jamie Hall, Xinying Song, Sahitya Potluri, Yu Du, Heng-Tze Cheng, Quoc Le, Ed H. Chi
-
Publication number: 20240289395Abstract: Implementations relate to helping a large language model generate factual responses to prompts that request factual content is disclosed. The large language model may receive a prompt context, a plurality of encoded context passages as input. The large language model is trained to determine whether or not to utilize the encoded context passages in generating the response. Implementations also relate to different methods of fine-tuning the responses generated by the large language model through query refinements, response re-writes, and evaluation of factual accuracy.Type: ApplicationFiled: December 4, 2023Publication date: August 29, 2024Inventors: Hao Zhou, Shrestha Basu Mallick, Trevor Strohman, Patricia Luisa Romero Domingo, Amirhossein Kiani, Yu Du, Xinying Song, Heng-Tze Cheng, Quoc V. Le, Ed Huai-Hsin Chi, Christopher Jamie Maclean Hall
-
Publication number: 20240054288Abstract: Systems and methods for performing inference for word or wordpiece tokenization are disclosed using a left-to-right longest-match-first greedy process. In some examples, the vocabulary may be organized into a trie structure in which each node includes a precomputed token or token_ID and a fail link, so that the tokenizer can parse the trie in a single pass to generate a list of only those tokens or token_IDs that correspond to the longest matching vocabulary entries in the sample string, without the need for backtracking. In some examples, the vocabulary may be organized into a trie in which each node has a fail link, and any node that would share token(s) or token_ID(s) of a preceding node is instead given a prev_match link that points back to a chain of nodes with those token(s) or token_ID(s).Type: ApplicationFiled: June 5, 2023Publication date: February 15, 2024Inventors: Xinying Song, Yang Song
-
Patent number: 11763083Abstract: Systems and methods for performing inference for word or wordpiece tokenization are disclosed using a left-to-right longest-match-first greedy process. In some examples, the vocabulary may be organized into a trie structure in which each node includes a precomputed token or token ID and a fail link, so that the tokenizer can parse the trie in a single pass to generate a list of only those tokens or token IDs that correspond to the longest matching vocabulary entries in the sample string, without the need for backtracking. In some examples, the vocabulary may be organized into a trie in which each node has a fail link, and any node that would share token(s) or token_ID(s) of a preceding node is instead given a prev_match link that points back to a chain of nodes with those token(s) or token_ID(s).Type: GrantFiled: May 18, 2020Date of Patent: September 19, 2023Assignee: Google LLCInventors: Xinying Song, Yang Song
-
Patent number: 11734066Abstract: Generally discussed herein are devices, systems, and methods for scheduling tasks to be completed by resources. A method can include identifying features of the task, the features including a time-dependent feature and a time-independent feature, the time-dependent feature indicating a time the task is more likely to be successfully completed by the resource, converting the features to feature values based on a predefined mapping of features to feature values in a first memory device, determining, by a gradient boost tree model and based on a first current time and the feature values, a likelihood the resource will successfully complete the task, and scheduling the task to be performed by the resource based on the determined likelihood.Type: GrantFiled: January 8, 2020Date of Patent: August 22, 2023Assignee: Microsoft Technology Licensing, LLCInventors: Jinchao Li, Yu Wang, Karan Srivastava, Jianfeng Gao, Prabhdeep Singh, Haiyuan Cao, Xinying Song, Hui Su, Jaideep Sarkar
-
Publication number: 20230124402Abstract: Systems and methods for performing inference for word or wordpiece tokenization are disclosed using a left-to-right longest-match-first greedy process. In some examples, the vocabulary may be organized into a trie structure in which each node includes a precomputed token or token ID and a fail link, so that the tokenizer can parse the trie in a single pass to generate a list of only those tokens or token IDs that correspond to the longest matching vocabulary entries in the sample string, without the need for backtracking. In some examples, the vocabulary may be organized into a trie in which each node has a fail link, and any node that would share token(s) or token_ID(s) of a preceding node is instead given a prev_match link that points back to a chain of nodes with those token(s) or token_ID(s).Type: ApplicationFiled: May 18, 2020Publication date: April 20, 2023Inventors: Xinying SONG, Yang SONG
-
Patent number: 11629475Abstract: A recyclable pile foundation is provided. The recyclable pile foundation includes several inner cylinders, several outer cylinders and several reciprocating components which are circumferentially distributed between the inner cylinders and the outer cylinders. Each reciprocating component includes several steel collars, a push-pull rod, a hold component and at least one motion component. The motion components are distributed along the push-pull rod. Each motion component includes at least one triangular connection plate, several connection rods, an inner wedge block, an outer wedge block, a motion block and a pointed rod. When the push-pull rod is pushed along its own axis to the pushed position, the pointed rod protrudes from the outer cylinders to increases the friction between the surrounding soil and the recyclable pile foundation. When the push-poll rod is pulled along its own axis to the pulled position, the pointed rods retract back into the outer cylinders.Type: GrantFiled: September 7, 2022Date of Patent: April 18, 2023Assignee: HAINAN UNIVERSITYInventors: Yongwei Wang, Qinxi Dong, Youliang Zhang, Xinying Song
-
Publication number: 20230081869Abstract: A recyclable pile foundation is provided. The recyclable pile foundation includes several inner cylinders, several outer cylinders and several reciprocating components which are circumferentially distributed between the inner cylinders and the outer cylinders. Each reciprocating component includes several steel collars, a push-pull rod, a hold component and at least one motion component. The motion components are distributed along the push-pull rod. Each motion component includes at least one triangular connection plate, several connection rods, an inner wedge block, an outer wedge block, a motion block and a pointed rod. When the push-pull rod is pushed along its own axis to the pushed position, the pointed rod protrudes from the outer cylinders to increases the friction between the surrounding soil and the recyclable pile foundation. When the push-poll rod is pulled along its own axis to the pulled position, the pointed rods retract back into the outer cylinders.Type: ApplicationFiled: September 7, 2022Publication date: March 16, 2023Inventors: Yongwei Wang, Qinxi Dong, Youliang Zhang, Xinying Song
-
Patent number: 11327726Abstract: A workflow engine tool is disclosed that enables scientists and engineers to programmatically author workflows (e.g., a directed acyclic graph, “DAG”) with nearly no overhead, using a simpler script that needs almost no modifications for portability among multiple different workflow engines. This permits users to focus on the business logic of the project, avoiding the distracting tedious overhead related to workflow management (such as uploading modules, drawing edges, setting parameters, and other tasks). The workflow engine tool provides an abstraction layer on top of workflow engines, introducing a binding function that converts a programming language function (e.g., a normal python function) into a workflow module definition. The workflow engine tool infers module instances and induces edge dependencies automatically by inferring from a programming language script to build a DAG.Type: GrantFiled: July 31, 2020Date of Patent: May 10, 2022Assignee: Microsoft Technology Licensing, LLCInventors: Yu Wang, Yu Hu, Haiyuan Cao, Hui Su, Jinchao Li, Xinying Song, Jianfeng Gao
-
Publication number: 20210224047Abstract: A workflow engine tool is disclosed that enables scientists and engineers to programmatically author workflows (e.g., a directed acyclic graph, “DAG”) with nearly no overhead, using a simpler script that needs almost no modifications for portability among multiple different workflow engines. This permits users to focus on the business logic of the project, avoiding the distracting tedious overhead related to workflow management (such as uploading modules, drawing edges, setting parameters, and other tasks). The workflow engine tool provides an abstraction layer on top of workflow engines, introducing a binding function that converts a programming language function (e.g., a normal python function) into a workflow module definition. The workflow engine tool infers module instances and induces edge dependencies automatically by inferring from a programming language script to build a DAG.Type: ApplicationFiled: July 31, 2020Publication date: July 22, 2021Inventors: Yu WANG, Yu HU, Haiyuan CAO, Hui SU, Jinchao LI, Xinying SONG, Jianfeng GAO
-
Patent number: 11068304Abstract: Systems and methods are disclosed for intelligent scheduling of calls to sales leads, leveraging machine learning (ML) to optimize expected results. One exemplary method includes determining, using a connectivity prediction model, call connectivity rate predictions; determining timeslot resources; allocating, based at least on the call connectivity rate predictions and timeslot resources, leads to timeslots in a first time period; determining, within a timeslot and using a lead scoring model, lead prioritization among leads within the timeslot; configuring, based at least on the lead prioritization, the telephone unit with lead information for placing a phone call; and applying a contextual bandit (ML) process to update the connectivity prediction model, the lead scoring model, or both. During subsequent time periods, the updated connectivity prediction and lead scoring models are used, thereby improving expected results over time.Type: GrantFiled: February 25, 2019Date of Patent: July 20, 2021Assignee: Microsoft Technology Licensing, LLCInventors: Jinchao Li, Xinying Song, Ah Young Kim, Haiyuan Cao, Yu Wang, Hui Su, Shahina Ferdous, Jianfeng Gao, Karan Srivastava, Jaideep Sarkar
-
Patent number: 10768908Abstract: A workflow engine tool is disclosed that enables scientists and engineers to programmatically author workflows (e.g., a directed acyclic graph, “DAG”) with nearly no overhead, using a simpler script that needs almost no modifications for portability among multiple different workflow engines. This permits users to focus on the business logic of the project, avoiding the distracting tedious overhead related to workflow management (such as uploading modules, drawing edges, setting parameters, and other tasks). The workflow engine tool provides an abstraction layer on top of workflow engines, introducing a binding function that converts a programming language function (e.g., a normal python function) into a workflow module definition. The workflow engine tool infers module instances and induces edge dependencies automatically by inferring from a programming language script to build a DAG.Type: GrantFiled: February 25, 2019Date of Patent: September 8, 2020Assignee: Microsoft Technology Licensing, LLCInventors: Yu Wang, Yu Hu, Haiyuan Cao, Hui Su, Jinchao Li, Xinying Song, Jianfeng Gao
-
Publication number: 20200273000Abstract: Systems and methods are disclosed for intelligent scheduling of calls to sales leads, leveraging machine learning (ML) to optimize expected results. One exemplary method includes determining, using a connectivity prediction model, call connectivity rate predictions; determining timeslot resources; allocating, based at least on the call connectivity rate predictions and timeslot resources, leads to timeslots in a first time period; determining, within a timeslot and using a lead scoring model, lead prioritization among leads within the timeslot; configuring, based at least on the lead prioritization, the telephone unit with lead information for placing a phone call; and applying a contextual bandit (ML) process to update the connectivity prediction model, the lead scoring model, or both. During subsequent time periods, the updated connectivity prediction and lead scoring models are used, thereby improving expected results over time.Type: ApplicationFiled: February 25, 2019Publication date: August 27, 2020Inventors: Jinchao LI, Xinying SONG, Ah Young KIM, Haiyuan CAO, Yu WANG, Hui SU, Shahina FERDOUS, Jianfeng GAO, Karan SRIVASTAVA, Jaideep SARKAR
-
Publication number: 20200272433Abstract: A workflow engine tool is disclosed that enables scientists and engineers to programmatically author workflows (e.g., a directed acyclic graph, “DAG”) with nearly no overhead, using a simpler script that needs almost no modifications for portability among multiple different workflow engines. This permits users to focus on the business logic of the project, avoiding the distracting tedious overhead related to workflow management (such as uploading modules, drawing edges, setting parameters, and other tasks). The workflow engine tool provides an abstraction layer on top of workflow engines, introducing a binding function that converts a programming language function (e.g., a normal python function) into a workflow module definition. The workflow engine tool infers module instances and induces edge dependencies automatically by inferring from a programming language script to build a DAG.Type: ApplicationFiled: February 25, 2019Publication date: August 27, 2020Inventors: Yu WANG, Yu HU, Haiyuan CAO, Hui SU, Jinchao LI, Xinying SONG, Jianfeng GAO
-
Publication number: 20200142737Abstract: Generally discussed herein are devices, systems, and methods for scheduling tasks to be completed by resources. A method can include identifying features of the task, the features including a time-dependent feature and a time-independent feature, the time-dependent feature indicating a time the task is more likely to be successfully completed by the resource, converting the features to feature values based on a predefined mapping of features to feature values in a first memory device, determining, by a gradient boost tree model and based on a first current time and the feature values, a likelihood the resource will successfully complete the task, and scheduling the task to be performed by the resource based on the determined likelihood.Type: ApplicationFiled: January 8, 2020Publication date: May 7, 2020Inventors: Jinchao Li, Yu Wang, Karan Srivastava, Jinfeng Gao, Prabhdeep Singh, Haiyuan Cao, Xinying Song, Hui Su, Jaideep Sarkar
-
Patent number: 10579423Abstract: Generally discussed herein are devices, systems, and methods for scheduling tasks to be completed by resources. A method can include identifying features of the task, the features including a time-dependent feature and a time-independent feature, the time-dependent feature indicating a time the task is more likely to be successfully completed by the resource, converting the features to feature values based on a predefined mapping of features to feature values in a first memory device, determining, by a gradient boost tree model and based on a first current time and the feature values, a likelihood the resource will successfully complete the task, and scheduling the task to be performed by the resource based on the determined likelihood.Type: GrantFiled: April 2, 2018Date of Patent: March 3, 2020Assignee: Microsoft Technology Licensing, LLCInventors: Jinchao Li, Yu Wang, Karan Srivastava, Jianfeng Gao, Prabhdeep Singh, Haiyuan Cao, Xinying Song, Hui Su, Jaideep Sarkar
-
Patent number: 10579430Abstract: Generally discussed herein are devices, systems, and methods for task routing. A method can include receiving, from a resource, a request for a task, in response to receiving the request, determining whether to retrieve a new task of new tasks stored in a first queue or a backlog task of backlog tasks stored in a second queue based on a combined amount of backlog tasks and new tasks relative to a capacity of the resource or the resources, retrieving the new task or the backlog task from the determined first queue or second queue, respectively, based on the determination, and providing the retrieved task to the resource.Type: GrantFiled: May 7, 2018Date of Patent: March 3, 2020Assignee: Microsoft Technolog Licensing, LLCInventors: Xinying Song, Jaideep Sarkar, Karan Srivastava, Jianfeng Gao, Prabhdeep Singh, Hui Su, Jinchao Li, Andreea Bianca Spataru
-
Publication number: 20190340030Abstract: Generally discussed herein are devices, systems, and methods for task routing. A method can include receiving, from a resource, a request for a task, in response to receiving the request, determining whether to retrieve a new task of new tasks stored in a first queue or a backlog task of backlog tasks stored in a second queue based on a combined amount of backlog tasks and new tasks relative to a capacity of the resource or the resources, retrieving the new task or the backlog task from the determined first queue or second queue, respectively, based on the determination, and providing the retrieved task to the resource.Type: ApplicationFiled: May 7, 2018Publication date: November 7, 2019Inventors: Xinying Song, Jaideep Sarkar, Karan Srivastava, Jianfeng Gao, Prabhdeep Singh, Hui Su, Jinchao Li, Andreea Bianca Spataru
-
Patent number: 10445650Abstract: A processing unit can successively operate layers of a multilayer computational graph (MCG) according to a forward computational order to determine a topic value associated with a document based at least in part on content values associated with the document. The processing unit can successively determine, according to a reverse computational order, layer-specific deviation values associated with the layers based at least in part on the topic value, the content values, and a characteristic value associated with the document. The processing unit can determine a model adjustment value based at least in part on the layer-specific deviation values. The processing unit can modify at least one parameter associated with the MCG based at least in part on the model adjustment value. The MCG can be operated to provide a result characteristic value associated with test content values of a test document.Type: GrantFiled: November 23, 2015Date of Patent: October 15, 2019Assignee: Microsoft Technology Licensing, LLCInventors: Jianfeng Gao, Li Deng, Xiaodong He, Lin Xiao, Xinying Song, Yelong Shen, Ji He, Jianshu Chen