Patents by Inventor En Shi
En Shi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250077780Abstract: A method for invoking a plugin of a large language model includes: acquiring natural language content; performing semantic understanding on the natural language content and detecting whether the natural language content hits a plugin to obtain a first plugin pointed to by the plugin hit result; comparing the first plugin with a second plugin corresponding to the current session understanding task to determine a to-be-executed session understanding task and a third plugin corresponding to the to-be-executed session understanding task; acquiring the language understanding content of the to-be-executed session understanding task and sending the language understanding content to the large language model to obtain the input parameter of the third plugin; and calling the third plugin according to the input parameter of the third plugin to obtain the calling result of the to-be-executed session understanding task.Type: ApplicationFiled: June 20, 2024Publication date: March 6, 2025Inventors: Yongkang Xie, Guming Gao, Penghao Zhao, Xue Xiong, Qian Wang, Dongze Xu, En Shi, Yuxuan Li, Sheng Zhou, Shupeng Li, Yao Wang, Zhou Xin
-
Publication number: 20250045517Abstract: A copywriting generation method, an electronic device and a storage medium are provided and relate to a field of artificial intelligence technology, in particular to fields of deep learning and natural language processing technologies, and may be applied to scenarios of large language models and generative dialogues.Type: ApplicationFiled: June 20, 2024Publication date: February 6, 2025Inventor: En SHI
-
Patent number: 12182546Abstract: A method for model production includes acquiring a related operation for model production from a user interface layer of a model production system, and determining a software platform of the model production system; acquiring a model service corresponding to the related operation by invoking an application programming interface (API) corresponding to the related operation, wherein the API is located between the user interface layer and other layer in the model production system; performing the model service by invoking local resources of the software platform with a tool of the software platform adapted to the model service, to generate a target model; and applying the target model in a target usage scene.Type: GrantFiled: August 16, 2022Date of Patent: December 31, 2024Assignee: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.Inventors: En Shi, Yongkang Xie, Zihao Pan, Shupeng Li, Xiaoyu Chen, Zhengyu Qian, Jingqiu Li
-
Patent number: 12174790Abstract: An apparatus comprises a processing device configured to detect a request for an updated snapshot schedule for an information technology asset, and to determine a current state of the information technology asset comprising a set of snapshot parameters of a current snapshot schedule and one or more performance metric values. The processing device is also configured to generate, utilizing a reinforcement learning framework, an updated parameter value for at least one of the snapshot parameters based at least in part on the current state. The processing device is further configured to monitor performance of the information technology asset utilizing the updated snapshot schedule comprising the updated parameter value for the at least one snapshot parameter, and to update the reinforcement learning framework based at least in part on a subsequent state of the information technology asset determined while monitoring performance of the information technology asset utilizing the updated snapshot schedule.Type: GrantFiled: March 28, 2023Date of Patent: December 24, 2024Assignee: Dell Products L.P.Inventors: Chi Chen, En Shi, Changyue Dai
-
Publication number: 20240419991Abstract: A method is provided that includes: creating a plurality of first model instances of a first service model to be deployed; allocating an inference service for each of a plurality of first model instances from the plurality of inference services; calling, for each first model instance, a loading interface of the inference service allocated for the first model instance to mount a weight file; determining, in response to a user request for a target service model, a target model instance from a plurality of model instances of the target service model to respond to the user request; and calling a target inference service allocated for the target model instance to use computing resources configured for the target inference service to run, in the target model instance, a base model mounted with a target weight file, and obtain a request result of the user request.Type: ApplicationFiled: June 19, 2024Publication date: December 19, 2024Applicant: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.Inventors: Zhenfang CHU, Zhengyu QIAN, En SHI, Mingren HU, Zhengxiong YUAN, Jinqi LI, Yue HUANG, Yang LUO, Guobin WANG, Yang QIAN, Kuan WANG
-
Publication number: 20240411552Abstract: A computer-implemented method for recommending a large model interface configuration includes: obtaining a search space of a model interface configuration and a test data set, wherein the search space comprises at least one candidate model interface and a value range of a hyperparameter; and obtaining a plurality of model interface configuration sets based on the search space, wherein each model interface configuration set comprises a candidate model interface and a value of the hyperparameter; and obtaining a test result corresponding to each model interface configuration set, by using the test data set to test a large model called based on each model interface configuration set; and determining a target interface configuration based on the test results corresponding to the plurality of model interface configuration sets.Type: ApplicationFiled: June 20, 2024Publication date: December 12, 2024Inventors: Jing XIA, Jun LIU, Binbin XU, Kuan SHI, Shupeng LI, Zhengyu QIAN, En SHI
-
Patent number: 12099422Abstract: Techniques for storage testing involve: acquiring a first state of a storage system including first input/output (IO) load information; taking a first action based on the first state, the first action causing the first IO load information to be changed to second IO load information; updating the first action to be a reserved action for the first state if it is obtained based on the second IO load information that the storage system reaches a preset condition; and obtaining an action combination of a plurality of IO load information changes based on a plurality of reserved actions corresponding to a plurality of states, wherein the plurality of states include the first state. Accordingly, the most effective load combination change mode for the storage system can be found automatically and more accurately, so as to find more vulnerabilities of the storage system, thereby improving the efficiency of storage system testing.Type: GrantFiled: February 1, 2023Date of Patent: September 24, 2024Assignee: Dell Products L.P.Inventors: Chi Chen, Changyue Dai, En Shi, Hailan Dong
-
Publication number: 20240311337Abstract: An apparatus comprises a processing device configured to detect a request for an updated snapshot schedule for an information technology asset, and to determine a current state of the information technology asset comprising a set of snapshot parameters of a current snapshot schedule and one or more performance metric values. The processing device is also configured to generate, utilizing a reinforcement learning framework, an updated parameter value for at least one of the snapshot parameters based at least in part on the current state. The processing device is further configured to monitor performance of the information technology asset utilizing the updated snapshot schedule comprising the updated parameter value for the at least one snapshot parameter, and to update the reinforcement learning framework based at least in part on a subsequent state of the information technology asset determined while monitoring performance of the information technology asset utilizing the updated snapshot schedule.Type: ApplicationFiled: March 28, 2023Publication date: September 19, 2024Inventors: Chi Chen, En Shi, Changyue Dai
-
Patent number: 11954011Abstract: An apparatus and a method for executing a customized production line using an artificial intelligence development platform, a computing device and a computer readable storage medium are provided. The apparatus includes: a production line executor configured to generate a native form of the artificial intelligence development platform based on a file set, the native form to be sent to a client accessing the artificial intelligence development platform so as to present a native interactive page of the artificial intelligence development platform; and a standardized platform interface configured to provide an interaction channel between the production line executor and the artificial intelligence development platform. The production line executor is further configured to generate an intermediate result by executing processing logic defined in the file set and to process the intermediate result by interacting with the artificial intelligence development platform via the standardized platform interface.Type: GrantFiled: October 28, 2020Date of Patent: April 9, 2024Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.Inventors: Yongkang Xie, Ruyue Ma, Zhou Xin, Hao Cao, Kuan Shi, Yu Zhou, Yashuai Li, En Shi, Zhiquan Wu, Zihao Pan, Shupeng Li, Mingren Hu, Tian Wu
-
Publication number: 20240028492Abstract: Techniques for storage testing involve: acquiring a first state of a storage system including first input/output (IO) load information; taking a first action based on the first state, the first action causing the first IO load information to be changed to second IO load information; updating the first action to be a reserved action for the first state if it is obtained based on the second IO load information that the storage system reaches a preset condition; and obtaining an action combination of a plurality of IO load information changes based on a plurality of reserved actions corresponding to a plurality of states, wherein the plurality of states include the first state. Accordingly, the most effective load combination change mode for the storage system can be found automatically and more accurately, so as to find more vulnerabilities of the storage system, thereby improving the efficiency of storage system testing.Type: ApplicationFiled: February 1, 2023Publication date: January 25, 2024Inventors: Chi Chen, Changyue Dai, En Shi, Hailan Dong
-
Publication number: 20240005182Abstract: Provided are a streaming media processing method based on inference service, an electronic device, and a storage medium, which relates to the field of artificial intelligence, and in particular, to the field of inference service of artificial intelligence models. The method includes: detecting, in a process of processing a k-th channel of streaming media through an i-th inference service pod, the i-th inference service pod, to obtain a detection result of the i-th inference service pod, i and k being positive integers; determining a replacement object of the i-th inference service pod, in the case where it is determined that the i-th inference service pod is in an abnormal state based on the detection result of the i-th inference service pod; and processing the k-th channel of streaming media through the replacement object of the i-th inference service pod.Type: ApplicationFiled: November 7, 2022Publication date: January 4, 2024Applicant: Beijing Baidu Netcom Science Technology Co., Ltd.Inventors: Jinqi Li, En Shi, Mingren Hu, Zhengyu Qian, Zhengxiong Yuan, Zhenfang Chu, Yue Huang, Yang Luo, Guobin Wang
-
Publication number: 20230401484Abstract: Provided are a data processing method and apparatus, an electronic device, and a storage medium. The data processing method includes acquiring a target directed acyclic graph (DAG) corresponding to the service processing logic of a model self-taught learning service, where the service processing logic includes execution logic for acquiring service data generated by an online released service model, execution logic for training a to-be-trained service model based on the service data, and execution logic for releasing the trained service model online; and performing self-taught learning on the to-be-trained service model according to the target DAG.Type: ApplicationFiled: December 7, 2022Publication date: December 14, 2023Inventors: Chao WANG, Xiangyue Lin, Yang Liang, En Shi, Shuangshuang QIAO
-
Publication number: 20230376726Abstract: Provided are an inference service deployment method, a device and a storage medium, relating to the field of artificial intelligence technology, and in particular to the field of machine learning and inference service technology. The inference service deployment method includes: obtaining performance information of a runtime environment of a deployment end; selecting a target version of an inference service from a plurality of candidate versions of the inference service of a model according to the performance information of the runtime environment of the deployment end; and deploying the target version of the inference service to the deployment end.Type: ApplicationFiled: November 3, 2022Publication date: November 23, 2023Inventors: Zhengxiong YUAN, Zhenfang CHU, Jinqi LI, Mingren HU, Guobin WANG, Yang LUO, Yue HUANG, Zhengyu QIAN, En SHI
-
Publication number: 20230196245Abstract: A method and apparatus for predicting a risk are provided. The method may include: determining an inherent risk probability of a to-be-tested object; building a relationship graph between the to-be-tested object and different associated objects; determining a primary conduction probability between any two directly associated objects in the relationship graph; determining, based on the primary conduction probability between any two directly associated objects in the relationship graph and the relationship graph, a multi-level conduction probability of the to-be-tested object; and determining a target risk probability of the to-be-tested object, based on the inherent risk probability and the multi-level conduction probability of the to-be-tested object.Type: ApplicationFiled: February 15, 2023Publication date: June 22, 2023Inventors: Mengyue LIU, Haibin ZHANG, Penghao ZHAO, Shupeng LI, En SHI
-
Publication number: 20230121576Abstract: Methods, system, and non-transitory processor-readable storage medium for a tuning system are provided herein. An example method includes providing, by a tuning module, a tuning action based on a state associated with a test case, executing, by a tuning agent, the test case using the tuning action, assessing, by a tuning assessment module, the tuning action with respect to a long-term reward, and determining a similarity between a subsequent execution of the test case and the assessment of the tuning action on at least one previous execution of the test case to tune the tuning action.Type: ApplicationFiled: October 25, 2021Publication date: April 20, 2023Applicant: Dell Products L.P.Inventors: Cherry Dai, Amber Jing Li, En Shi
-
Publication number: 20220391182Abstract: A method for model production includes acquiring a related operation for model production from a user interface layer of a model production system, and determining a software platform of the model production system; acquiring a model service corresponding to the related operation by invoking an application programming interface (API) corresponding to the related operation, wherein the API is located between the user interface layer and other layer in the model production system; performing the model service by invoking local resources of the software platform with a tool of the software platform adapted to the model service, to generate a target model; and applying the target model in a target usage scene.Type: ApplicationFiled: August 16, 2022Publication date: December 8, 2022Applicant: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.Inventors: En Shi, Yongkang Xie, Zihao Pan, Shupeng Li, Xiaoyu Chen, Zhengyu Qian, Jingqiu Li
-
Publication number: 20220374742Abstract: A method for running an inference service platform, includes: determining inference tasks to be allocated for the inference service platform, in which the inference service platform includes two or more inference service groups, versions of the inference service groups are different, and the inference service groups are configured to perform a same type of inference services; determining a flow weight of each of the inference service groups, in which the flow weight is configured to indicate a proportion of a number of inference tasks to which the corresponding inference service group need to be allocated in a total number of inference tasks; and allocating the corresponding number of inference tasks in the inference tasks to be allocated to each of the inference service groups based on the flow weight of each of the inference service groups; and performing the inference tasks by the inference service group.Type: ApplicationFiled: August 3, 2022Publication date: November 24, 2022Applicant: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.Inventors: Zhengxiong Yuan, Zhengyu Qian, En Shi, Mingren Hu, Jinqi Li, Zhenfang Chu, Runqing Li, Yue Huang
-
Patent number: 11483530Abstract: A color compensation method includes obtaining a target brightness, a target frame rate and a target pulse number; selecting a plurality of second gamma groups from a plurality of first gamma groups according to the target brightness and the target pulse number, wherein the plurality of first gamma groups respectively correspond to a plurality of frame rates; and calculating the compensation value to compensate the display brightness and color according to the target brightness, the target frame rate, the plurality of second gamma groups and a calculation method.Type: GrantFiled: March 14, 2022Date of Patent: October 25, 2022Assignee: NOVATEK Microelectronics Corp.Inventors: En-Shi Shih, Yen-Tao Liao, Shih-Ting Huang
-
Publication number: 20220309395Abstract: The present disclosure discloses a method and an apparatus for adapting a deep learning model, an electronic device and a medium, which relates to technology fields of artificial intelligence, deep learning, and cloud computing. The specific implementation plan is: obtaining model information of an original deep learning model and hardware information of a target hardware to be adapted; querying a conversion path table according to the model information and the hardware information to obtain a matched target conversion path; and converting, according to the target conversion path, the original deep learning model to an intermediate deep learning model in the conversion path, and converting the intermediate deep learning model to the target deep learning model.Type: ApplicationFiled: September 16, 2020Publication date: September 29, 2022Inventors: Tuobang WU, En SHI, Yongkang XIE, Xiaoyu CHEN, Lianghuo ZHANG, Jie LIU, Binbin XU
-
Patent number: 11455173Abstract: A method for management of an artificial intelligence development platform is provided. The artificial intelligence development platform is deployed with instances of a plurality of model services, and each of the model services is provided with one or more instances. The method includes: acquiring calling information of at least one model service; determining the activity of the at least one model service according to the calling information; and at least deleting all instances of the at least one model service in response to that the determined activity meets a first condition.Type: GrantFiled: March 19, 2021Date of Patent: September 27, 2022Assignee: Beijing Baidu Netcom Science and Technology Co., Ltd.Inventors: Zhengxiong Yuan, En Shi, Yongkang Xie, Mingren Hu, Zhengyu Qian, Zhenfang Chu