Patents by Inventor Xiangyang Ji

Xiangyang Ji has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11948079
    Abstract: The present disclosure discloses a multi-agent coordination method. The method includes: performing multiple data collections on N agents to collect E sets of data, where N and E are integers greater than 1; and optimizing neural networks of the N agents using reinforcement learning based on the E sets of data. Each data collection includes: randomly selecting a first coordination pattern from multiple predetermined coordination patterns; obtaining N observations after the N agents act on an environment in the first coordination pattern; determining a first probability and a second probability that a current coordination pattern is the first coordination pattern based on the N observations; and determining a pseudo reward based on the first probability and the second probability. The E sets of data include: a first coordination pattern label indicating the first coordination pattern, the N observations, and the pseudo reward.
    Type: Grant
    Filed: October 19, 2020
    Date of Patent: April 2, 2024
    Inventors: Xiangyang Ji, Shuncheng He
  • Patent number: 11823062
    Abstract: The present disclosure discloses an unsupervised reinforcement learning method and apparatus based on Wasserstein distance. The method includes: obtaining a state distribution in a trajectory obtained with guidance of a current policy of an agent; calculating a Wasserstein distance between the state distribution and a state distribution in a trajectory obtained with another historical policy, and calculating a pseudo reward of the agent based on the Wasserstein distance, replacing a reward fed back from an environment in a target reinforcement learning framework with the pseudo reward, and guiding the current policy of the agent to keep a large distance from the other historical policy. The method uses Wasserstein distance to encourage an algorithm in an unsupervised reinforcement learning framework to obtain diverse policies and skills through training.
    Type: Grant
    Filed: March 21, 2023
    Date of Patent: November 21, 2023
    Assignee: TSINGHUA UNIVERSITY
    Inventors: Xiangyang Ji, Shuncheng He, Yuhang Jiang
  • Publication number: 20230266584
    Abstract: An encoded illumination real-time focusing scanning imaging device includes an LED array, an object stage, an objective lens, a liquid lens, a lens sleeve, and a camera. The object stage is used for placing a sample, the LED array is used for emitting bright field light or encoded light, and the bright field light or the encoded light respectively penetrates through the sample and then successively passes through the objective lens, the liquid lens, and the lens sleeve to reach the camera to shoot a bright field image or an encoded light illumination image by the camera. The present invention also discloses an encoded illumination real-time focusing scanning imaging method, which is performed by using the above-mentioned device. The present invention enables fast and accurate focusing at a low cost.
    Type: Application
    Filed: April 20, 2023
    Publication date: August 24, 2023
    Inventors: Yongbing ZHANG, Kaifa XIN, Xiangyang JI, Haoqian WANG
  • Publication number: 20230154025
    Abstract: The present disclosure relates to the technical field of computer vision and digital image processing, and more particularly to a light field encoded imaging method and apparatus for a scattering scene. The method includes: obtaining scattering light field data, and extracting a refocusing focal stack corresponding to different offset pixels between views from the scattering light field data; transforming an image in the refocusing focal stack into a saturation domain to form a processed focal stack; comparing boundary sharpness degrees of each region in the processed focal stack when refocusing by different offset pixels, and extracting offset pixels corresponding to the sharpest boundary of each region by using a 1-norm operator; and converting the offset pixels into a depth map of a target based on data parameters of the light field.
    Type: Application
    Filed: November 18, 2022
    Publication date: May 18, 2023
    Inventors: Xiangyang JI, Xiaoyu WANG, Yi ZHANG, Xiaocong LIAN
  • Patent number: 11620760
    Abstract: The present invention provides a ranging method based on laser-line scanning imaging to effectively suppress interference of extreme weather on imaging. The method includes the following steps: acquiring priori reference images for a fixed laser-line scanning system, including respectively placing reference whiteboards at different distances, projecting line laser beams to the whiteboards, and acquiring the reference images by using a camera; placing a laser-line scanning device in a real scene, causing the laser-line scanning device to respectively emit line lasers at different angles, and acquiring an image at each scanning angle by using a camera; and performing fusion calculation on the acquired scanning image in the real scene and the priori reference images by using a ranging algorithm based on laser-line scanning, and extracting distance information of a surrounding object, to implement environment perception.
    Type: Grant
    Filed: June 23, 2021
    Date of Patent: April 4, 2023
    Assignee: Tsinghua Shenzhen International Graduate School
    Inventors: Yongbing Zhang, Xizhi Huang, Xiangyang Ji, Haoqian Wang
  • Publication number: 20230005181
    Abstract: Provided are a reinforcement learning-based label-free six-dimensional object pose prediction method and apparatus. The method includes: obtaining a target image to be predicted, the target image being a two-dimensional image including a target object; performing pose prediction based on the target image by using a pre-trained pose prediction model to obtain a prediction result, the pose prediction model being obtained by performing reinforcement learning based on a sample image; and determining a three-dimensional position and a three-dimensional direction of the target object based on the prediction result.
    Type: Application
    Filed: August 5, 2022
    Publication date: January 5, 2023
    Inventors: Xiangyang JI, Jianzhun SHAO
  • Patent number: 11461925
    Abstract: Provided are a pose prediction method and apparatus, and a model training method and apparatus. The pose prediction method includes: performing target identification on a first image to be predicted to determine an area where a target object is located; determining a transformed target image based on the area where the target object is located; inputting the transformed target image into a pose decoupling prediction model for pose prediction; and determining a rotation amount of the target object based on an outputted result of a rotation amount branch network of the pose decoupling prediction model and determining a translation amount of the target object based on an outputted result of rotation amount branch network of the pose decoupling prediction model. By means of decoupling rotation and translation in object pose, the accuracy of pose prediction can be improved.
    Type: Grant
    Filed: February 24, 2022
    Date of Patent: October 4, 2022
    Assignee: TSINGHUA UNIVERSITY
    Inventors: Xiangyang Ji, Zhigang Li
  • Publication number: 20220276093
    Abstract: A high-resolution spectral image fast acquisition apparatus comprises an illumination source, an objective lens, a beam splitter, a single shot spectral image acquisition assembly and a reference image acquisition assembly, wherein the objective lens is used to align a sample to be measured; the illumination source is used to project an illumination light onto the sample to be measured so that the sample to be measured is amplified by the objective lens; wherein one part of amplified light enters the single shot spectral image acquisition assembly so as to acquire a low-resolution spectral cube of the sample to be measured, and another part of the amplified light enters the reference image acquisition assembly to acquire a high-resolution spectral cube. The apparatus enables rapid access to high-resolution spectral images, thereby speeding up the process of using spectral images for medical diagnosis.
    Type: Application
    Filed: May 17, 2022
    Publication date: September 1, 2022
    Inventors: Yongbing ZHANG, Kaifa XIN, Xiangyang JI, Haoqian WANG
  • Publication number: 20220180553
    Abstract: Provided are a pose prediction method and apparatus, and a model training method and apparatus. The pose prediction method includes: performing target identification on a first image to be predicted to determine an area where a target object is located; determining a transformed target image based on the area where the target object is located; inputting the transformed target image into a pose decoupling prediction model for pose prediction; and determining a rotation amount of the target object based on an outputted result of a rotation amount branch network of the pose decoupling prediction model and determining a translation amount of the target object based on an outputted result of rotation amount branch network of the pose decoupling prediction model. By means of decoupling rotation and translation in object pose, the accuracy of pose prediction can be improved.
    Type: Application
    Filed: February 24, 2022
    Publication date: June 9, 2022
    Inventors: Xiangyang JI, Zhigang LI
  • Publication number: 20220121920
    Abstract: The present disclosure discloses a multi-agent coordination method. The method includes: performing multiple data collections on N agents to collect E sets of data, where N and E are integers greater than 1; and optimizing neural networks of the N agents using reinforcement learning based on the E sets of data. Each data collection includes: randomly selecting a first coordination pattern from multiple predetermined coordination patterns; obtaining N observations after the N agents act on an environment in the first coordination pattern; determining a first probability and a second probability that a current coordination pattern is the first coordination pattern based on the N observations; and determining a pseudo reward based on the first probability and the second probability. The E sets of data include: a first coordination pattern label indicating the first coordination pattern, the N observations, and the pseudo reward.
    Type: Application
    Filed: October 19, 2020
    Publication date: April 21, 2022
    Inventors: Xiangyang JI, Shuncheng He
  • Patent number: 11200696
    Abstract: The present disclosure relates to a method and an apparatus for training a 6D pose estimation network based on deep learning iterative matching. The method includes: obtaining a rendered image and a first segmentation mask of a target object by using a 3D model and an initial 6D pose estimation of the target object; inputting the rendered image, the first segmentation mask, an observed image of the target object, and a second segmentation mask of the target object in the observed image into a deep convolutional neural network to obtain a 6D pose estimation, a third segmentation mask and an optical flow; and performing said obtaining and said inputting again by updating the initial 6D pose estimation using the obtained relative 6D pose estimation and replacing the second segmentation mask with the third segmentation mask, to iteratively train the deep convolutional neural network.
    Type: Grant
    Filed: September 17, 2020
    Date of Patent: December 14, 2021
    Assignee: TSINGHUA UNIVERSITY
    Inventors: Xiangyang Ji, Gu Wang, Yi Li
  • Publication number: 20210319573
    Abstract: The present invention provides a ranging method based on laser-line scanning imaging to effectively suppress interference of extreme weather on imaging. The method includes the following steps: acquiring priori reference images for a fixed laser-line scanning system, including respectively placing reference whiteboards at different distances, projecting line laser beams to the whiteboards, and acquiring the reference images by using a camera; placing a laser-line scanning device in a real scene, causing the laser-line scanning device to respectively emit line lasers at different angles, and acquiring an image at each scanning angle by using a camera; and performing fusion calculation on the acquired scanning image in the real scene and the priori reference images by using a ranging algorithm based on laser-line scanning, and extracting distance information of a surrounding object, to implement environment perception.
    Type: Application
    Filed: June 23, 2021
    Publication date: October 14, 2021
    Applicant: Tsinghua Shenzhen International Graduate School
    Inventors: Yongbing ZHANG, Xizhi HUANG, Xiangyang JI, Haoqian WANG
  • Publication number: 20210004984
    Abstract: The present disclosure relates to a method and an apparatus for training a 6D pose estimation network based on deep learning iterative matching. The method includes: obtaining a rendered image and a first segmentation mask of a target object by using a 3D model and an initial 6D pose estimation of the target object; inputting the rendered image, the first segmentation mask, an observed image of the target object, and a second segmentation mask of the target object in the observed image into a deep convolutional neural network to obtain a 6D pose estimation, a third segmentation mask and an optical flow; and performing said obtaining and said inputting again by updating the initial 6D pose estimation using the obtained relative 6D pose estimation and replacing the second segmentation mask with the third segmentation mask, to iteratively train the deep convolutional neural network.
    Type: Application
    Filed: September 17, 2020
    Publication date: January 7, 2021
    Applicant: TSINGHUA UNIVERSITY
    Inventors: Xiangyang JI, Gu WANG, Yi LI
  • Patent number: 10115182
    Abstract: The present invention discloses a depth map super-resolution processing method, including: firstly, respectively acquiring a first original image (S1) and a second original image (S2) and a low resolution depth map (d) of the first original image (S1); secondly, 1) dividing the low resolution depth map (d) into multiple depth image blocks; 2) respectively performing the following processing on the depth image blocks obtained in step 1); 21) performing super-resolution processing on a current block with multiple super-resolution processing methods, to obtain multiple high resolution depth image blocks; 22) obtaining new synthesized image blocks by using an image synthesis technology; 23) upon matching and judgment, determining an ultimate high resolution depth image block; and 3) integrating the high resolution depth image blocks of the depth image blocks into one image according to positions of the depth image blocks in the low resolution depth map (d).
    Type: Grant
    Filed: July 21, 2016
    Date of Patent: October 30, 2018
    Assignee: GRADUATE SCHOOL AT SHENZHEN, TSINGHUA UNIVERSITY
    Inventors: Lei Zhang, Xiangyang Ji, Yangguang Li, Yongbing Zhang, Xin Jin, Haoqian Wang, Guijin Wang
  • Patent number: 8832759
    Abstract: A resource scheduling apparatus, a resource scheduling method, a program requesting method, a program requesting system, and a Set Top Box (STB) are provided. The resource scheduling method includes: distributing bandwidth greater than an inherent code rate of a program respectively requested by a user to a Video On Demand (VOD) program of the user within available bandwidth of a frequency point according to a received VOD user request; and reducing the bandwidth distributed to at least one VOD program to which bandwidth has been distributed according to the received VOD user request for currently requesting the program when the available bandwidth of the frequency point is smaller than an inherent code rate of a currently requested program, so that the available bandwidth of the frequency point is greater than or equal to the inherent code rate of the currently requested program. Thus, a transmission speed of the program and a utilization ratio of the bandwidth are increased.
    Type: Grant
    Filed: September 29, 2010
    Date of Patent: September 9, 2014
    Assignee: Huawei Technologies Co., Ltd.
    Inventors: Xiangyang Ji, Yongxiong Liao, Wei Luo, Zengli Jia
  • Patent number: 8340177
    Abstract: Techniques and tools are described for scalable video coding and decoding. For example, a 3D sub-band video encoder includes an embedded base layer codec as well as temporal sub-band transforms and spatial sub-band transforms. The placement of the base layer codec among the sub-band transforms and the role of the base layer codec in scalable video coding vary depending on implementation. In general, the base layer codec provides efficient compression at low bit rates and produces a base layer compressed video bit stream compatible with existing decoders. At the same time, the 3D sub-band video encoder provides spatial and temporal scalability options at higher bit rates, refining the base layer video. A corresponding 3D sub-band video decoder includes an embedded base layer decoder.
    Type: Grant
    Filed: May 10, 2005
    Date of Patent: December 25, 2012
    Assignee: Microsoft Corporation
    Inventors: Xiangyang Ji, Jizheng Xu, Feng Wu
  • Patent number: 8306124
    Abstract: An encoding method for skipped macroblocks in a video image includes the steps of: adding one indication bit into a picture header for indicating a coding mode for skipped macroblocks in a current image; selecting the coding mode for a macroblock type in the current image according to the number of skipped macroblocks, if it is a run_length coding, then setting the indication bit of the picture header as a status indicating a run_length coding, and encoding the macroblock type in the image by the run_length coding mode; if it is a joint coding, then setting the indication bit of the picture header as status indicating a joint coding and encoding the macroblock type in the image by the joint coding mode of the number of skipped macroblocks and the macroblock type; finally, encoding other data in the current macroblock and writing data into a code stream.
    Type: Grant
    Filed: July 8, 2004
    Date of Patent: November 6, 2012
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Wen Gao, Junhao Zheng, Siwei Ma, Xiangyang Ji, Peng Zhang, Yan Lu
  • Patent number: 8116375
    Abstract: A method for obtaining an image reference block in a code mode of fixed reference frame number includes the steps of: performing motion estimation for each block of a current B frame and obtaining a motion vector MV of a corresponding block of a backward reference frame; discriminating whether the motion vector is beyond a maximum forward reference frame which is possibly pointed by the B frame, if not, then calculating the forward and backward motion vectors in a normal way; if yes, then using the motion vector of the forward reference frame that the B frame can obtain in the same direction to replace the motion vector of the corresponding block in the backward reference, and calculating the forward and the backward motion vectors of the B frame; finally, two image blocks pointed by the final obtained forward and backward motion vectors as the image reference blocks corresponding to the macro block.
    Type: Grant
    Filed: July 19, 2004
    Date of Patent: February 14, 2012
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Wen Gao, Xiangyang Ji, Siwei Ma, Debin Zhao, Yan Lu
  • Patent number: 8005144
    Abstract: The invention discloses a bi-directional prediction method for video coding/decoding. When bi-directional prediction coding at the coding end, firstly the given forward candidate motion vector of the current image block is obtained for every image block of the current B-frame; the backward candidate motion vector is obtained through calculation, and the candidate bi-directional prediction reference block is obtained through bi-directional prediction method; the match is computed within the given searching scope and/or the given matching threshold; finally the optimal matching block is selected to determine the final forward motion vector, and the backward motion vector and the block residual. The present invention achieves the object of bi-directional prediction by coding a single motion vector, furthermore, it will not enhance the complexity of searching for a matching block at the coding end, and may save amount of coding the motion vector and represent the motion of the objects in video more actually.
    Type: Grant
    Filed: July 2, 2004
    Date of Patent: August 23, 2011
    Assignee: Institute of Computing Technology Chinese Academy of Sciences
    Inventors: Xiangyang Ji, Wen Gao, Debin Zhao, Yan Lu, Siwei Ma, Honggang Qi
  • Patent number: 7974344
    Abstract: A “rounding to zero” method can maintain the exact motion vector and can also be achieved by the method without division so as to improve the precision of calculating the motion vector, embody the motion of the object in video more factually, and obtain the more accurate motion vector prediction. Combining with the forward prediction coding and the backward prediction coding, the present invention realizes a new prediction coding mode, which can guarantee the high efficiency of coding in direct mode as well as is convenient for hardware realization, and gains the same effect as the conventional B frame coding.
    Type: Grant
    Filed: July 8, 2004
    Date of Patent: July 5, 2011
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Xiangyang Ji, Wen Gao, Siwei Ma, Debin Zhao, Yan Lu