Patents by Inventor Kun Deng
Kun Deng has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240140931Abstract: Tricyclic compounds of formula (I), pharmaceutical compositions comprising same, methods for preparing same, and uses thereof, wherein each variable is as defined in the description.Type: ApplicationFiled: January 28, 2022Publication date: May 2, 2024Inventors: Guangxiu DAI, Kun XIAO, Wei DENG
-
Patent number: 11941373Abstract: A deep learning model trained to learn to predict source code is tuned for a target source code generation task through reinforcement learning using a reward score that considers the quality of the source code predicted during the tuning process. The reward score is adjusted to consider code-quality factors and source code metrics. The code-quality factors account for the predicted source code having syntactic correctness, successful compilation, successful execution, successful invocation, readability, functional correctness, and coverage. The source code metrics generate a score based on how close the predicted source code is to a ground truth code.Type: GrantFiled: December 17, 2021Date of Patent: March 26, 2024Assignee: Microsoft Technology Licensing, LLC.Inventors: Shao Kun Deng, Neelakantan Sundaresan, Alexey Svyatkovskiy, Michele Tufano
-
Publication number: 20240028306Abstract: A code completion tool uses a neural transformer model to generate candidate sequences to complete a line of source code. The neural transformer model is trained using a conditional language modeling objective on a large unsupervised dataset that includes source code programs written in several different programming languages. The neural transformer model is used within a beam search that predicts the most likely candidate sequences for a code snippet under development.Type: ApplicationFiled: August 9, 2023Publication date: January 25, 2024Inventors: ALEXEY SVYATKOVSKIY, SHENGYU FU, NEELAKANTAN SUNDARESAN, SHAO KUN DENG
-
Publication number: 20230419683Abstract: Provided are a method and system for automatic driving data collection and closed-loop management. The method includes: obtaining vehicle driving data; preprocessing the vehicle driving data; obtaining incremental data by filtering, through a pre-trained neural network based on a predetermined filtering rule, the preprocessed vehicle driving data; and storing the incremental data or transmitting the incremental data to a cloud. The method can effectively filter high-value incremental data, thereby reducing requirements for data storage volume and/or data transmission bandwidth of the system. The system of the present disclosure can be post-mounted or pre-mounted on a vehicle, and is independent of specific vehicle type of the vehicle. In addition, it is not necessary for the vehicle to be equipped with real-value systems such as high-cost laser radars, which greatly improves use convenience of the system and facilitates rapid and large-scale application.Type: ApplicationFiled: December 1, 2020Publication date: December 28, 2023Inventors: Yue HU, Kun DENG, Jianfeng ZHANG, Xinyu ZHENG
-
Publication number: 20230376685Abstract: Edit automation enhancements may be implemented in source code editors and other text editors. Provisional selections that indicate user intentions are submitted to a suggestion generator with other edit context information, to improve the quality of generated text suggestions and reduce the cognitive load on users. A provisional selection may include a highlighted completion list entry, or document text targeted by a hovering cursor, or metainformation text targeted by the hovering cursor, for example. An inline grey text suggestion driven by provisional selection may be displayed simultaneously with completion list suggestions that were created without regard to provisional selection. Suggestions driven by provisional selection may be interleaved with existing document text. Suggestions may be accepted fully in one gesture, or in parts. Suggestions may be edited by a user before being accepted, driving further suggestion refinement.Type: ApplicationFiled: August 2, 2023Publication date: November 23, 2023Inventors: Mark Alistair WILSON-THOMAS, Jonathan Keith SIMMONS, David Ellis PUGH, Vivian Julia LIM, Anqi LI, Shwetha SRINATH, German David OBANDO CHACON, Jin Woo JANG, Shengyu FU, Shao Kun DENG
-
Patent number: 11809842Abstract: A code completion tool uses a neural transformer model to generate candidate sequences to complete a line of source code. The neural transformer model is trained using a conditional language modeling objective on a large unsupervised dataset that includes source code programs written in several different programming languages. The neural transformer model is used within a beam search that predicts the most likely candidate sequences for a code snippet under development.Type: GrantFiled: January 20, 2022Date of Patent: November 7, 2023Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC.Inventors: Alexey Svyatkovskiy, Shengyu Fu, Neelakantan Sundaresan, Shao Kun Deng
-
Publication number: 20230342287Abstract: A test-driven development system utilizes a neural transformer model with attention to generate method bodies for a focal method given its associated test cases, and optionally a method signature and a docstring of the focal method. The candidate method bodies are validated for syntactic correctness, tested using the given test cases, and tested with a donor class in a target system. Those candidate method bodies passing the validation and testing are then ranked based on a PLUM score that analyzes the candidate method bodies against various quality and performance metrics.Type: ApplicationFiled: June 19, 2023Publication date: October 26, 2023Inventors: COLIN BRUCE CLEMENT, SHAO KUN DENG, NEELAKANTAN SUNDARESAN, ALEXEY SVYATKOVSKIY, MICHELE TUFANO
-
Patent number: 11797426Abstract: A test-driven development system utilizes a neural transformer model with attention to generate method bodies for a focal method given its associated test cases, and optionally a method signature and a docstring of the focal method. The candidate method bodies are validated for syntactic correctness, tested using the given test cases, and tested with a donor class in a target system. Those candidate method bodies passing the validation and testing are then ranked based on a PLUM score that analyzes the candidate method bodies against various quality and performance metrics.Type: GrantFiled: October 22, 2021Date of Patent: October 24, 2023Assignee: MICROSOFT TECHNOLOGY LICENSINGInventors: Colin Bruce Clement, Shao Kun Deng, Neelakantan Sundaresan, Alexey Svyatkovskiy, Michele Tufano
-
Patent number: 11763078Abstract: Edit automation enhancements may be implemented in source code editors and other text editors. Provisional selections that indicate user intentions are submitted to a suggestion generator with other edit context information, to improve the quality of generated text suggestions and reduce the cognitive load on users. A provisional selection may include a highlighted completion list entry, or document text targeted by a hovering cursor, or metainformation text targeted by the hovering cursor, for example. An inline grey text suggestion driven by provisional selection may be displayed simultaneously with completion list suggestions that were created without regard to provisional selection. Suggestions driven by provisional selection may be interleaved with existing document text. Suggestions may be accepted fully in one gesture, or in parts. Suggestions may be edited by a user before being accepted, driving further suggestion refinement.Type: GrantFiled: April 22, 2021Date of Patent: September 19, 2023Assignee: Microsoft Technology Licensing, LLCInventors: Mark Alistair Wilson-Thomas, Jonathan Keith Simmons, David Ellis Pugh, Vivian Julia Lim, Anqi Li, Shwetha Srinath, German David Obando Chacon, Jin Woo Jang, Shengyu Fu, Shao Kun Deng
-
Publication number: 20230281318Abstract: A constrained decoding technique incorporates token constraints into a beam search at each time step of a decoding process in order to generate viable candidate sequences that are syntactically and semantically correct. The token constraints identify source code tokens or sequences of tokens that should appear in a candidate sequence. The token constraints are generated from checking whether a token predicted at each decoding step is feasible for a partial solution based on the production rules of the grammar of the programming language, the syntactic correctness of a partial sequence, and/or static type correctness.Type: ApplicationFiled: March 7, 2022Publication date: September 7, 2023Inventors: COLIN BRUCE CLEMENT, SHAO KUN DENG, XIAOYU LIU, NEELAKANTAN SUNDARESAN, ALEXEY SVYATKOVSKIY
-
Publication number: 20230222334Abstract: A deep learning model is quantized during its training to perform a target software engineering task. During training, a portion of the full-precision floating point weights is quantized into INT4 or INT 8 data types through scalar quantization or product quantization to make the model more resilient to quantization and to reduce the noise between the quantized and full-precision model outputs. In scalar quantization, each sub-block consists of a single weight that is mapped into a codeword of a codebook. In product quantization, an identity matrix and a codebook of centroids is used to map a quantized weight into its original value.Type: ApplicationFiled: January 10, 2022Publication date: July 13, 2023Inventors: COLIN BRUCE CLEMENT, SHAO KUN DENG, NEELAKANTAN SUNDARESAN, ALEXEY SVYATKOVSKIY
-
Publication number: 20230195428Abstract: A deep learning model trained to learn to predict source code is tuned for a target source code generation task through reinforcement learning using a reward score that considers the quality of the source code predicted during the tuning process. The reward score is adjusted to consider code-quality factors and source code metrics. The code-quality factors account for the predicted source code having syntactic correctness, successful compilation, successful execution, successful invocation, readability, functional correctness, and coverage. The source code metrics generate a score based on how close the predicted source code is to a ground truth code.Type: ApplicationFiled: December 17, 2021Publication date: June 22, 2023Inventors: SHAO KUN DENG, NEELAKANTAN SUNDARESAN, ALEXEY SVYATKOVSKIY, MICHELE TUFANO
-
Publication number: 20230128008Abstract: A test-driven development system utilizes a neural transformer model with attention to generate method bodies for a focal method given its associated test cases, and optionally a method signature and a docstring of the focal method. The candidate method bodies are validated for syntactic correctness, tested using the given test cases, and tested with a donor class in a target system. Those candidate method bodies passing the validation and testing are then ranked based on a PLUM score that analyzes the candidate method bodies against various quality and performance metrics.Type: ApplicationFiled: October 22, 2021Publication date: April 27, 2023Inventors: COLIN BRUCE CLEMENT, SHAO KUN DENG, NEELAKANTAN SUNDARESAN, ALEXEY SVYATKOVSKIY, MICHELE TUFANO
-
Publication number: 20230114423Abstract: An automated program repair tool utilizes a neural transformer model with attention to predict the contents of a bug repair in the context of source code having a bug of an identified bug type. The neural transformer model is trained on a large unsupervised corpus of source code using a span-masking denoising optimization objective, and fine-tuned on a large supervised dataset of triplets containing a bug-type annotation, software bug, and repair. The bug-type annotation is derived from an interprocedural static code analyzer. A bug type edit centroid is computed for each bug type and used in the inference decoding phase to generate the bug repair.Type: ApplicationFiled: November 25, 2022Publication date: April 13, 2023Inventors: SHAO KUN DENG, NEELAKANTAN SUNDARESAN, ALEXEY SVYATKOVSKIY, MICHELE TUFANO
-
Patent number: 11599345Abstract: Language interoperability between source code programs not compatible with an interprocedural static code analyzer is achieved through language-independent representations of the programs. The source code programs are transformed into respective intermediate language instructions from which a language-independent control flow graph and a language-independent type environment is created. A program compatible with the interprocedural static code analyzer is generated from the language-independent control flow graph and the language-independent type environment in order to utilize the interprocedural static code analyzer to detect memory safety faults.Type: GrantFiled: November 4, 2021Date of Patent: March 7, 2023Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC.Inventors: Shao Kun Deng, Matthew Glenn Jin, Shuvendu Lahiri, Xiaoyu Liu, Xin Shi, Neelakantan Sundaresan
-
Publication number: 20220398462Abstract: A cloud platform includes several web services that facilitate the automated tuning and deployment of pre-trained deep learning models configured for software engineering tasks. The automated tuning and deployment allow a developer to fine-tune a pre-existing model without having access to the parameters of the pre-existing and the fine-tuned model in a manner that does not require user management input. The cloud platform provides a set of files for each pre-trained models used to automatically build a fine-tuning infrastructure to fine-tune a model and a deployment infrastructure that deploys the fine-tuned model without requiring user input.Type: ApplicationFiled: June 14, 2021Publication date: December 15, 2022Inventors: COLIN BRUCE CLEMENT, SHAO KUN DENG, DAWN DRAIN, NEELAKANTAN SUNDARESAN, ALEXEY SVYATKOVSKIY, YIDING TIAN, MICHELE TUFANO, PAUL AN-CHIEH WANG, CHEN WU, DONGJIANG YOU
-
Patent number: 11526424Abstract: An automated program repair tool utilizes a neural transformer model with attention to predict the contents of a bug repair in the context of source code having a bug of an identified bug type. The neural transformer model is trained on a large unsupervised corpus of source code using a span-masking denoising optimization objective, and fine-tuned on a large supervised dataset of triplets containing a bug-type annotation, software bug, and repair. The bug-type annotation is derived from an interprocedural static code analyzer. A bug type edit centroid is computed for each bug type and used in the inference decoding phase to generate the bug repair.Type: GrantFiled: June 10, 2020Date of Patent: December 13, 2022Assignee: MICROSOFT TECHNOLOGY LICENSING LLC.Inventors: Shao Kun Deng, Neelakantan Sundaresan, Alexey Svyatkovskiy, Michele Tufano
-
Publication number: 20220374208Abstract: A code completion tool uses a neural transformer model with attention to generate syntactically-correct candidates with holes to complete a partially-formed code snippet. The model is trained to predict the expansion of non-terminal symbols of the production rules of the underlying grammar of the code snippet without being constrained to a left-to-right expansion order. A hole is a non-terminal symbol of the grammar of a programming language that marks a position in a candidate where the code completion engine is not certain of the production rule that should be used to expand the non-terminal symbol. The hole allows the code completion engine to expand other non-terminal symbols in a candidate and allow the user to guide the expansion of the holes in a candidate.Type: ApplicationFiled: May 15, 2021Publication date: November 24, 2022Inventors: MILTIADIS ALLAMANIS, DAYA GUO, SHAO KUN DENG, NEELAKANTAN SUNDARESAN, ALEXEY SVYATKOVSKIY
-
Publication number: 20220358286Abstract: Edit automation enhancements may be implemented in source code editors and other text editors. Provisional selections that indicate user intentions are submitted to a suggestion generator with other edit context information, to improve the quality of generated text suggestions and reduce the cognitive load on users. A provisional selection may include a highlighted completion list entry, or document text targeted by a hovering cursor, or metainformation text targeted by the hovering cursor, for example. An inline grey text suggestion driven by provisional selection may be displayed simultaneously with completion list suggestions that were created without regard to provisional selection. Suggestions driven by provisional selection may be interleaved with existing document text. Suggestions may be accepted fully in one gesture, or in parts. Suggestions may be edited by a user before being accepted, driving further suggestion refinement.Type: ApplicationFiled: April 22, 2021Publication date: November 10, 2022Inventors: Mark Alistair WILSON-THOMAS, Jonathan Keith SIMMONS, David Ellis PUGH, Vivian Julia LIM, Anqi LI, Shwetha SRINATH, German David OBANDO CHACON, Jin Woo JANG, Shengyu FU, Shao Kun DENG
-
Patent number: 11400927Abstract: A system includes a computer including a processor and a memory, the memory storing instructions executable by the processor to determine respective threat numbers for each of a plurality of targets based on an angular acceleration of a host vehicle and actuate a component in the host vehicle based on the threat numbers.Type: GrantFiled: January 29, 2018Date of Patent: August 2, 2022Assignee: FORD GLOBAL TECHNOLOGIES, LLCInventors: Kun Deng, Nanjun Liu, Fangjun Jiang, Gary Song, Alex Maurice Miller