Patents by Inventor Baocai YIN
Baocai YIN has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11911902Abstract: A method for obstacle avoidance in degraded environments of robots based on intrinsic plasticity of an SNN is disclosed. A decision network in a synaptic autonomous learning module takes lidar data, distance from a target point and velocity at a previous moment as state input, and outputs the velocity of left and right wheels of the robot through the autonomous adjustment of the dynamic energy-time threshold, so as to carry out autonomous perception and decision making. The method solves the difficulty of the lack of intrinsic plasticity in the SNN, which leads to the difficulty of adapting to degraded environments due to the homeostasis imbalance of the model, is successfully deployed in mobile robots to maintain a stable trigger rate for autonomous navigation and obstacle avoidance in degraded, disturbed and noisy environments, and has validity and applicability on different degraded scenes.Type: GrantFiled: December 20, 2021Date of Patent: February 27, 2024Assignee: DALIAN UNIVERSITY OF TECHNOLOGYInventors: Xin Yang, Jianchuan Ding, Bo Dong, Felix Heide, Baocai Yin
-
Patent number: 11875583Abstract: The present invention belongs to the technical field of 3D reconstruction in the field of computer vision, and provides a dataset generation method for self-supervised learning scene point cloud completion based on panoramas. Pairs of incomplete point cloud and target point cloud with RGB information and normal information can be generated by taking RGB panoramas, depth panoramas and normal panoramas in the same view as input for constructing a self-supervised learning dataset for training of the scene point cloud completion network. The key points of the present invention are occlusion prediction and equirectangular projection based on view conversion, and processing of the stripe problem and point-to-point occlusion problem during conversion. The method of the present invention includes simplification of the collection mode of the point cloud data in a real scene; occlusion prediction idea of view conversion; and design of view selection strategy.Type: GrantFiled: November 23, 2021Date of Patent: January 16, 2024Assignee: DALIAN UNIVERSITY OF TECHNOLOGYInventors: Xin Yang, Tong Li, Baocai Yin, Zhaoxuan Zhang, Boyan Wei, Zhenjun Du
-
Patent number: 11810359Abstract: The present invention belongs to the technical field of computer vision, and provides a video semantic segmentation method based on active learning, comprising an image semantic segmentation module, a data selection module based on the active learning and a label propagation module. The image semantic segmentation module is responsible for segmenting image results and extracting high-level features required by the data selection module; the data selection module selects a data subset with rich information at an image level, and selects pixel blocks to be labeled at a pixel level; and the label propagation module realizes migration from image to video tasks and completes the segmentation result of a video quickly to obtain weakly-supervised data. The present invention can rapidly generate weakly-supervised data sets, reduce the cost of manufacture of the data and optimize the performance of a semantic segmentation network.Type: GrantFiled: December 21, 2021Date of Patent: November 7, 2023Assignee: DALIAN UNIVERSITY OF TECHNOLOGYInventors: Xin Yang, Xiaopeng Wei, Yu Qiao, Qiang Zhang, Baocai Yin, Haiyin Piao, Zhenjun Du
-
Patent number: 11798264Abstract: Dictionary learning method and means for zero-shot recognition can establish the alignment between visual space and semantic space at category layer and image level, so as to realize high-precision zero-shot image recognition. The dictionary learning method includes the following steps: (1) training a cross domain dictionary of a category layer based on a cross domain dictionary learning method; (2) generating semantic attributes of an image based on the cross domain dictionary of the category layer learned in step (1); (3) training a cross domain dictionary of the image layer based on the image semantic attributes generated in step (2); (4) completing a recognition task of invisible category images based on the cross domain dictionary of the image layer learned in step (3).Type: GrantFiled: January 29, 2022Date of Patent: October 24, 2023Assignee: Beijing University of TechnologyInventors: Lichun Wang, Shuang Li, Shaofan Wang, Dehui Kong, Baocai Yin
-
Publication number: 20230166397Abstract: A method for obstacle avoidance in degraded environments of robots based on intrinsic plasticity of an SNN is disclosed. A decision network in a synaptic autonomous learning module takes lidar data, distance from a target point and velocity at a previous moment as state input, and outputs the velocity of left and right wheels of the robot through the autonomous adjustment of the dynamic energy-time threshold, so as to carry out autonomous perception and decision making. The method solves the difficulty of the lack of intrinsic plasticity in the SNN, which leads to the difficulty of adapting to degraded environments due to the homeostasis imbalance of the model, is successfully deployed in mobile robots to maintain a stable trigger rate for autonomous navigation and obstacle avoidance in degraded, disturbed and noisy environments, and has validity and applicability on different degraded scenes.Type: ApplicationFiled: December 20, 2021Publication date: June 1, 2023Inventors: Xin YANG, Jianchuan DING, Bo DONG, Felix HEIDE, Baocai YIN
-
Publication number: 20230131545Abstract: Dictionary learning method and means for zero-shot recognition can establish the alignment between visual space and semantic space at category layer and image level, so as to realize high-precision zero-shot image recognition. The dictionary learning method includes the following steps: (1) training a cross domain dictionary of a category layer based on a cross domain dictionary learning method; (2) generating semantic attributes of an image based on the cross domain dictionary of the category layer learned in step (1); (3) training a cross domain dictionary of the image layer based on the image semantic attributes generated in step (2); (4) completing a recognition task of invisible category images based on the cross domain dictionary of the image layer learned in step (3).Type: ApplicationFiled: January 29, 2022Publication date: April 27, 2023Applicant: Beijing University of TechnologyInventors: Lichun WANG, Shuang LI, Shaofan WANG, Dehui KONG, Baocai YIN
-
Publication number: 20230094308Abstract: The present invention belongs to the technical field of 3D reconstruction in the field of computer vision, and provides a dataset generation method for self-supervised learning scene point cloud completion based on panoramas. Pairs of incomplete point cloud and target point cloud with RGB information and normal information can be generated by taking RGB panoramas, depth panoramas and normal panoramas in the same view as input for constructing a self-supervised learning dataset for training of the scene point cloud completion network. The key points of the present invention are occlusion prediction and equirectangular projection based on view conversion, and processing of the stripe problem and point-to-point occlusion problem during conversion. The method of the present invention includes simplification of the collection mode of the point cloud data in a real scene; occlusion prediction idea of view conversion; and design of view selection strategy.Type: ApplicationFiled: November 23, 2021Publication date: March 30, 2023Inventors: Xin YANG, Tong LI, Baocai YIN, Zhaoxuan ZHANG, Boyan WEI, Zhenjun DU
-
Publication number: 20220212339Abstract: The present invention belongs to the technical field of computer vision and provides a data active selection method for robot grasping. The core content of the present invention is a data selection strategy module, which shares the feature extraction layer of backbone main network and integrates the features of three receptive fields with different sizes. While making full use of the feature extraction module, the present invention greatly reduces the amount of parameters that need to be added. During the training process of the main grasp method detection network model, the data selection strategy module can be synchronously trained to form an end-to-end model. The present invention makes use of naturally existing labeled and unlabeled labels, and makes full use of the labeled data and the unlabeled data. When the amount of the labeled data is small, the network can still be more fully trained.Type: ApplicationFiled: December 29, 2021Publication date: July 7, 2022Inventors: Xin YANG, Boyan WEI, Baocai YIN, Qiang ZHANG, Xiaopeng WEI, Zhenjun DU
-
Publication number: 20220215662Abstract: The present invention belongs to the technical field of computer vision, and provides a video semantic segmentation method based on active learning, comprising an image semantic segmentation module, a data selection module based on the active learning and a label propagation module. The image semantic segmentation module is responsible for segmenting image results and extracting high-level features required by the data selection module; the data selection module selects a data subset with rich information at an image level, and selects pixel blocks to be labeled at a pixel level; and the label propagation module realizes migration from image to video tasks and completes the segmentation result of a video quickly to obtain weakly-supervised data. The present invention can rapidly generate weakly-supervised data sets, reduce the cost of manufacture of the data and optimize the performance of a semantic segmentation network.Type: ApplicationFiled: December 21, 2021Publication date: July 7, 2022Inventors: Xin YANG, Xiaopeng WEI, Yu QIAO, Qiang ZHANG, Baocai YIN, Haiyin PIAO, Zhenjun DU