Patents by Inventor Ronggang Wang

Ronggang Wang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11949868
    Abstract: Embodiments of the present disclosure provide a method and device for selecting a context model of a quantized coefficient end flag. The method comprises: obtaining a scanning position POS of a non-zero coefficient corresponding to current quantized coefficient end flag in a specific scanning order; wherein the scanning position POS is a subscript of the non-zero coefficient in the scanning order; configuring a first context model array, and using a fixed value as the base to calculate the logarithmic value of the value of the scanning position POS plus 1, and according to the logarithmic value, selecting a first context model from the first context model array; and using the first context model to encode or decode a binary symbol of the current quantized coefficient end flag. According to the technical solution of the present application, it is able to improve encoding and decoding efficiency for a quantized coefficient end flag, thereby further improve the efficiency of video encoding and decoding.
    Type: Grant
    Filed: March 7, 2019
    Date of Patent: April 2, 2024
    Assignee: PEKING UNIVERSITY SHENZHEN GRADUATE SCHOOL
    Inventors: Ronggang Wang, Zhenyu Wang, Wen Gao
  • Patent number: 11902524
    Abstract: An embodiment of the present description provides a decoding method and device for a quantization block as well as an electronic device.
    Type: Grant
    Filed: March 7, 2019
    Date of Patent: February 13, 2024
    Assignee: PEKING UNIVERSITY SHENZHEN GRADUATE SCHOOL
    Inventors: Ronggang Wang, Zhenyu Wang, Wen Gao
  • Patent number: 11870994
    Abstract: A method, system, device and computer-readable storage medium for inverse quantization. The method comprises: determining an initial weighted inverse quantization matrix, wherein, the initial weighted inverse quantization matrix is the same as the quantized block in size; setting some matrix elements in the initial weighted inverse quantization matrix to zero to obtain a weighted inverse quantization matrix, wherein, determining the matrix elements that need to be zeroed according to the size of the quantized block; weighted inverse quantizing the quantized coefficients in the quantized block to generate corresponding inverse transform coefficients, wherein, the value of the matrix element corresponding to the position of the quantized coefficient in the weighted inverse quantization matrix is used as a weight coefficient of the weighted inverse quantization.
    Type: Grant
    Filed: March 6, 2019
    Date of Patent: January 9, 2024
    Assignee: PEKING UNIVERSITY SHENZHEN GRADUATE SCHOOL
    Inventors: Ronggang Wang, Zhenyu Wang, Wen Gao
  • Patent number: 11783459
    Abstract: Embodiments of the present application provide a method and a device of inverse tone mapping and an electronic device. The method includes: obtaining one or more low dynamic range images; performing a decomposition operation to the low dynamic range image to acquire a detail layer and a basic layer of the low dynamic range image; restoring the detail layer and the basic layer by using a predetermined first restoration network and a second restoration network to acquire restored detail layer and basic layer; and adjusting the restored detail layer and basic layer by using a predetermined fusion network to acquire an adjusted high dynamic range image. With the technical solution of the present application, the conversion from a low dynamic range image to a high dynamic range image can be more robustly completed without complicated parameter settings.
    Type: Grant
    Filed: June 19, 2019
    Date of Patent: October 10, 2023
    Assignee: PEKING UNIVERSITY SHENZHEN GRADUATE SCHOOL
    Inventors: Ronggang Wang, Chao Wang, Wen Gao
  • Publication number: 20230109825
    Abstract: A method and a device for encoding or decoding based on an inter-frame prediction. The method includes steps of: determining a temporal motion vector prediction value of a to-be-processed coding unit, where the temporal motion vector prediction value is a temporal motion vector prediction value of a sub-block, a temporal motion vector of which is obtainable through prediction, in sub-blocks adjacent to the to-be-processed coding unit and/or sub-blocks in the to-be-processed coding unit; determining a motion vector residual prediction value of the to-be-processed coding unit according to the temporal motion vector prediction value; determining a motion vector of a sub-block in the to-be-processed coding unit according to the temporal motion vector prediction value and the motion vector residual prediction value and performing a motion compensation according to the motion vector of the sub-block in the to-be-processed coding unit to determine a prediction block of the to-be-processed coding unit.
    Type: Application
    Filed: October 25, 2019
    Publication date: April 13, 2023
    Inventors: Ronggang WANG, Xi XIE, Kui FAN, Wen GAO
  • Publication number: 20230025557
    Abstract: Embodiments of the present application provide a method and a device of inverse tone mapping and an electronic device. The method includes: obtaining one or more low dynamic range images; performing a decomposition operation to the low dynamic range image to acquire a detail layer and a basic layer of the low dynamic range image; restoring the detail layer and the basic layer by using a predetermined first restoration network and a second restoration network to acquire restored detail layer and basic layer; and adjusting the restored detail layer and basic layer by using a predetermined fusion network to acquire an adjusted high dynamic range image. With the technical solution of the present application, the conversion from a low dynamic range image to a high dynamic range image can be more robustly completed without complicated parameter settings.
    Type: Application
    Filed: June 19, 2019
    Publication date: January 26, 2023
    Inventors: Ronggang WANG, Chao WANG, Wen GAO
  • Publication number: 20220345706
    Abstract: An embodiment of the present description provides a decoding method and device for a quantization block as well as an electronic device.
    Type: Application
    Filed: March 7, 2019
    Publication date: October 27, 2022
    Inventors: Ronggang WANG, Zhenyu WANG, Wen GAO
  • Patent number: 11451774
    Abstract: An embodiment of the present disclosure discloses an intra-prediction-based encoding and decoding method, apparatus and filter, and belongs to a video coding technical field.
    Type: Grant
    Filed: March 26, 2019
    Date of Patent: September 20, 2022
    Assignee: PEKING UNIVERSITY SHENZHEN GRADUATE SCHOOL
    Inventors: Ronggang Wang, Kui Fan, Guisen Xu, Wen Gao
  • Publication number: 20220245775
    Abstract: A tone mapping method is provided. This method includes: obtaining one or a plurality of high dynamic range images, and determining a storage format of each high dynamic range image; performing an image decomposition on the high dynamic range image to obtain a first component, a second component and a third component, when the storage format of the high dynamic range image is determined as a predetermined storage format; inputting the first component and the second component into a predetermined deep neural network, and using the deep neural network to perform mapping on the first component and the second component respectively to obtain a first mapped component and a second mapped component; and fusing the first mapped component and the second mapped component with the third component to obtain a fused low dynamic range image corresponding to the high dynamic range image.
    Type: Application
    Filed: April 20, 2022
    Publication date: August 4, 2022
    Inventors: Ronggang WANG, Ning ZHANG, Wen GAO
  • Patent number: 11397890
    Abstract: The present application discloses a cross-media retrieval method based on deep semantic space, which includes a feature generation stage and a semantic space learning stage. In the feature generation stage, a CNN visual feature vector and an LSTM language description vector of an image are generated by simulating a perception process of a person for the image; and topic information about a text is explored by using an LDA topic model, thus extracting an LDA text topic vector. In the semantic space learning phase, a training set image is trained to obtain a four-layer Multi-Sensory Fusion Deep Neural Network, and a training set text is trained to obtain a three-layer text semantic network, respectively. Finally, a test image and a text are respectively mapped to an isomorphic semantic space by using two networks, so as to realize cross-media retrieval. The disclosed method can significantly improve the performance of cross-media retrieval.
    Type: Grant
    Filed: August 16, 2017
    Date of Patent: July 26, 2022
    Assignee: Peking University Shenzhen Graduate School
    Inventors: Wenmin Wang, Mengdi Fan, Peilei Dong, Ronggang Wang, Ge Li, Shengfu Dong, Zhenyu Wang, Ying Li, Hui Zhao, Wen Gao
  • Patent number: 11381812
    Abstract: Disclosed is a boundary filtering method for intra prediction, relating to the video encoding technology filed. Whether boundary filtering is performed on an intra prediction block or not is adaptively selected by means of a rate distortion optimization decision; during filtering, a filter coefficient exponentially attenuated relative to distance to boundary is adopted to perform filtering on the first N rows or the first N columns of the intra prediction block by means of an intra prediction block filter, and different filtering strengths are used according to different sizes of the prediction blocks. Therefore, the boundary distortion problem of intra prediction block is solved, the intra prediction precision is improved, and the encoding efficiency of intra prediction block is increased; and the practicability and the robustness of the boundary filtering technology are improved.
    Type: Grant
    Filed: September 25, 2018
    Date of Patent: July 5, 2022
    Assignee: PEKING UNIVERSITY SHENZHEN GRADUATE SCHOOL
    Inventors: Ronggang Wang, Kui Fan, Ge Li, Wen Gao
  • Patent number: 11379711
    Abstract: A video action detection method based on a convolutional neural network (CNN) is disclosed in the field of computer vision recognition technologies. A temporal-spatial pyramid pooling layer is added to a network structure, which eliminates limitations on input by a network, speeds up training and detection, and improves performance of video action classification and time location. The disclosed convolutional neural network includes a convolutional layer, a common pooling layer, a temporal-spatial pyramid pooling layer and a full connection layer. The outputs of the convolutional neural network include a category classification output layer and a time localization calculation result output layer. The disclosed method does not require down-sampling to obtain video clips of different durations, but instead utilizes direct input of the whole video at once, improving efficiency.
    Type: Grant
    Filed: August 16, 2017
    Date of Patent: July 5, 2022
    Assignee: Peking University Shenzhen Graduate School
    Inventors: Wenmin Wang, Zhihao Li, Ronggang Wang, Ge Li, Shengfu Dong, Zhenyu Wang, Ying Li, Hui Zhao, Wen Gao
  • Patent number: 11368661
    Abstract: The embodiment of the present specification discloses an image synthesis method, apparatus and device for free-viewpoint.
    Type: Grant
    Filed: February 14, 2019
    Date of Patent: June 21, 2022
    Assignee: PEKING UNIVERSITY SHENZHEN GRADUATE SCHOOL
    Inventors: Ronggang Wang, Sheng Wang, Zhenyu Wang, Wen Gao
  • Publication number: 20220191503
    Abstract: A method for deriving motion vector is provided, this method includes: obtaining a space domain motion vector prediction and a time domain motion vector prediction of adjacent blocks of a coding unit in a predetermined direction; performing a filtering operation on the space domain motion vector and the time domain motion vector prediction to obtain the space domain motion vector prediction and the time domain motion vector prediction of the filtered adjacent blocks; determining, according to a predetermined inter-frame prediction mode, reference motion vectors of a current block in four side directions by using the space domain motion vector prediction and the time domain motion vector prediction of the filtered adjacent blocks and a coordinate position of the current block in the coding unit; and deriving motion vectors of the current block according to the reference motion vectors and the coordinate position of the current block in the coding unit.
    Type: Application
    Filed: December 22, 2021
    Publication date: June 16, 2022
    Inventors: Ronggang WANG, Xi XIE, Kui FAN, Wen GAO
  • Patent number: 11354541
    Abstract: The present specification discloses a method, apparatus, and device for video frame interpolation.
    Type: Grant
    Filed: March 7, 2019
    Date of Patent: June 7, 2022
    Assignee: PEKING UNIVERSITY SHENZHEN GRADUATE SCHOOL
    Inventors: Ronggang Wang, Haoxian Zhang, Zhenyu Wang, Wen Gao
  • Patent number: 11347979
    Abstract: A method and a device for MCMC framework-based sub-hypergraph matching are provided. Matching of object features is performed by constructing sub-hypergraphs. In a large number of actual images and videos, objects vary constantly, and contain various noise points as well as other interference factors, which makes image object matching and searching very difficult. Perform object feature matching by representing the appearance and positions of objects by sub-hypergraphs allows for faster and more accurate image matching. Furthermore, a sub-hypergraph has several advantages over a graph or a hypergraph: on one hand, a sub-hypergraph has more geometric information (e.g. angle transformation, rotation, scale, etc.) than a graph, and has a lower degree of difficulty and better extensibility than a hypergraph. On the other hand, the disclosed method and device have stronger capabilities to resist interference and good robustness, and are adaptable to more complex settings, especially with outliers.
    Type: Grant
    Filed: March 10, 2016
    Date of Patent: May 31, 2022
    Assignee: Peking University Shenzhen Graduate School
    Inventors: Wenmin Wang, Ruonan Zhang, Ronggang Wang, Ge Li, Shengfu Dong, Zhenyu Wang, Ying Li, Hui Zhao, Wen Gao
  • Patent number: 11350128
    Abstract: The present application provides methods, systems, devices and computer-readable mediums for encoding and decoding transform. A method of the present application comprises: performing transform-encoding on a prediction residual block separately using a plurality of transform matrices, to obtain transformed residual blocks; based on the transformed residual block, determining, by a rate-distortion optimization decision, a transform matrix or a combination of transform matrices matching the residual characteristics of the prediction residual block from the plurality of transform matrices, and determining a transformed residual block to be outputted in a bitstream. Compared with the prior art, the method of the embodiments of the present invention performs a residual transform using transform matrices that are more closely matched with the residual characteristic, thereby improving the expression of the residual signal and improving the coding efficiency of the residual block.
    Type: Grant
    Filed: February 19, 2019
    Date of Patent: May 31, 2022
    Assignee: PEKING UNIVERSITY SHENZHEN GRADUATE SCHOOL
    Inventors: Ronggang Wang, Kui Fan, Guisen Xu, Zhenyu Wang, Wen Gao
  • Patent number: 11330301
    Abstract: The present application provides a method and a device of encoding and decoding based on free viewpoint, and relates to the technical field of video encoding. The method includes: generating a planar splicing image and splice information based on multiple single-viewpoint videos at a server side; generating a planar splicing video based on the planar splicing image; generating camera side information of the planar splicing video based on camera side information existing in the multiple single-viewpoint videos; and encoding the planar splicing video, the splice information and the camera side information of the planar splicing video to generate a planar splicing video bit stream, and decoding a planar splicing video bit stream to acquire a virtual viewpoint according to viewpoint information of a viewer at client side.
    Type: Grant
    Filed: April 8, 2019
    Date of Patent: May 10, 2022
    Assignee: PEKING UNIVERSITY SHENZHEN GRADUATE SCHOOL
    Inventors: Ronggang Wang, Zhenyu Wang, Wen Gao
  • Publication number: 20220116607
    Abstract: The application discloses a method, system, device and computer-readable storage medium for inverse quantization, wherein, in some embodiments, determining an initial weighted inverse quantization matrix, wherein, the initial weighted inverse quantization matrix is the same as the quantized block in size; setting some matrix elements in the initial weighted inverse quantization matrix to zero to obtain a weighted inverse quantization matrix, wherein, determining the matrix elements that need to be zeroed according to the size of the quantized block; weighted inverse quantizing the quantized coefficients in the quantized block to generate corresponding inverse transform coefficients, wherein, the value of the matrix element corresponding to the position of the quantized coefficient in the weighted inverse quantization matrix is used as a weight coefficient of the weighted inverse quantization.
    Type: Application
    Filed: March 6, 2019
    Publication date: April 14, 2022
    Inventors: Ronggang WANG, Zhenyu WANG, Wen GAO
  • Patent number: 11301953
    Abstract: Disclosed are a panoramic video asymmetrical mapping method and a corresponding inverse mapping method that include mapping a spherical surface corresponding to a panoramic image or video A onto a two-dimensional image or video B, projecting the spherical surface onto an isosceles quadrangular pyramid with a square bottom plane, and further projecting the isosceles quadrangular pyramid onto a planar surface, using isometric projection on a main viewpoint region in the projection and using a relatively high sampling density to ensure that the video quality of the region of the main viewpoint is high, while using a relatively low sample density for non-main viewpoint regions so as to reduce bit rate. The panoramic video asymmetrical inverse mapping technique provides a method for mapping from a planar surface to a spherical surface, and a planar surface video may be mapped back to a spherical surface for rendering and viewing.
    Type: Grant
    Filed: May 29, 2018
    Date of Patent: April 12, 2022
    Assignee: PEKING UNIVERSITY SHENZHEN GRADUATE SCHOOL
    Inventors: Ronggang Wang, Yueming Wang, Zhenyu Wang, Wen Gao