Patents by Inventor Yuwen He

Yuwen He has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10972731
    Abstract: Systems and methods are provided for video coding (e.g., encoding and/or decoding). A coding device may include a processor. The processor may be configured to receive a video frame and partition the video frame into a super coding units (SCUs). The processor may be configured to partition each of the SCUs into coding tree units (CTUs) based on a coding mode.
    Type: Grant
    Filed: November 10, 2016
    Date of Patent: April 6, 2021
    Assignee: InterDigital Madison Patent Holdings, SAS
    Inventors: Miaohui Wang, Yuwen He, Yan Ye, Xiaoyu Xiu
  • Patent number: 10939130
    Abstract: Inter-layer motion mapping information may be used to enable temporal motion vector prediction (TMVP) of an enhancement layer of a bitstream. For example, a reference picture and a motion vector (MV) of an inter-layer video block may be determined. The reference picture may be determined based on a collocated base layer video block. For example, the reference picture may be a collocated inter-layer reference picture of the reference picture of the collocated base layer video block. The MV may be determined based on a MV of the collocated base layer video block. For example, the MV may be determined by determining the MV of the collocated base layer video block and scaling the MV of the collocated base layer video block according to a spatial ratio between the base layer and the enhancement layer. TMVP may be performed on the enhancement layer picture using the MV of the inter-layer video block.
    Type: Grant
    Filed: January 9, 2018
    Date of Patent: March 2, 2021
    Assignee: VID SCALE, Inc.
    Inventors: Xiaoyu Xiu, Yan Ye, Yong He, Yuwen He
  • Patent number: 10939096
    Abstract: Systems and methods are described for control-point based intra mode for coding a video bitstream. In an exemplary embodiment, at least two control points in a picture are selected. The control points may be, for example, points at or adjacent to two or more corners of a current block. For each of the control points, an associated intra prediction direction is identified. The intra prediction directions may be encoded in the bitstream, e.g. using differential coding. A derived intra prediction direction is interpolated based on a position of a pixel (or of a block) relative to the control points, and the derived intra prediction direction is used to predict one or more samples in the video. Different interpolation techniques, such as triangular interpolation or bilinear interpolation may be used.
    Type: Grant
    Filed: May 4, 2017
    Date of Patent: March 2, 2021
    Assignee: Interdigital Madison Patent Holdings, SAS
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20210051340
    Abstract: Systems and methods are described for selecting a motion vector (MV) to use in frame-rate up conversion (FRUC) coding of a block of video. In one embodiment, a first set of motion vector candidates is identified for FRUC prediction of the block. A search center is defined based on the first set of motion vector candidates, and a search window is determined, the search window having a selected width and being centered on the search center. A search for a selected MV is performed within the search window. In some embodiments, an initial set of MVs is processed with a clustering algorithm to generate a smaller number of MVs that are used as the first set. The selected MV may be subject to a motion refinement search, which may also be performed over a constrained search range. In additional embodiments, search iterations are constrained to limit complexity.
    Type: Application
    Filed: January 28, 2019
    Publication date: February 18, 2021
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 10917660
    Abstract: Intra planar approach(es) may be used to predict a pixel(s) in a current block. The current block may be associated with a reconstructed left reference line, a reconstructed top reference line, and an non-reconstructed reference line to be predicted. The reconstructed reference lines may have been decoded and may be available. The non-reconstructed reference lines to be predicted may include an non-reconstructed right and/or an non-reconstructed bottom reference lines. A pivot reference pixel may be identified and may be located on an extension of the reconstructed left and/or top reference lines. A reference pixel may be determined and may be located on the reconstructed top and/or left reference lines. Pixels on the non-reconstructed reference line(s) may be predicted based on the pivot reference pixel and the reference pixel. Pixels of the current block may be predicted using the predicted pixels on the right and the bottom reference lines.
    Type: Grant
    Filed: January 9, 2018
    Date of Patent: February 9, 2021
    Assignee: VID SCALE, Inc.
    Inventors: Rahul Vanam, Yuwen He, Yan Ye
  • Publication number: 20210037258
    Abstract: Exemplary embodiments include systems and methods for coding a video comprising a plurality of pictures including a current picture, a first reference picture, and a second reference picture, where each picture includes a plurality of blocks. In one method, for at least a current block in the current picture, a number of available bi-prediction weights is determined based at least in part on a temporal layer and/or a quantization parameter of the current picture. From among available bi-prediction weights a pair of weights are identified. Using the identified weights, the current block is then predicted as a weighted sum of a first reference block in the first reference picture and a second reference block in the second reference picture. Encoding techniques are also described for efficient searching and selection of a pair of bi-prediction weights to use for prediction of a block.
    Type: Application
    Filed: January 23, 2019
    Publication date: February 4, 2021
    Inventors: Yan Zhang, Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20210029378
    Abstract: A video coding device may be configured to perform directional Bi-directional optical flow (BDOF) refinement on a coding unit (CU). The device may determine the direction in which to perform directional BDOF refinement. The device may calculate the vertical direction gradient difference and the horizontal direction gradient difference for the CU. The vertical direction gradient difference may indicate the difference between the vertical gradients for a first reference picture and the vertical gradients for a second reference picture. The horizontal direction gradient difference may indicate the difference between the horizontal gradients for the first reference picture and the horizontal gradients for the second reference picture. The video coding device may determine the direction in which to perform directional BDOF refinement based on the vertical direction gradient difference and the horizontal direction gradient difference.
    Type: Application
    Filed: April 5, 2019
    Publication date: January 28, 2021
    Applicant: VID SCALE, INC.
    Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye
  • Patent number: 10904571
    Abstract: A system, method, and/or instrumentality may be provided for coding a 360-degree video. A picture of the 360-degree video may be received. The picture may include one or more faces associated with one or more projection formats. A first projection format indication may be received that indicates a first projection format may be associated with a first face. A second projection format indication may be received that indicates a second projection format may be associated with a second face. Based on the first projection format, a first transform function associated with the first face may be determined. Based on the second projection format, a second transform function associated with the second face may be determined. At least one decoding process may be performed on the first face using the first transform function and/or at least one decoding process may be performed on the second face using the second transform function.
    Type: Grant
    Filed: May 24, 2018
    Date of Patent: January 26, 2021
    Assignee: VID SCALE, Inc.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 10897629
    Abstract: In an intra-block copy video encoding method, an encoder performs a hash-based search to identify a selected set of candidate blocks for prediction of an input video block. For each of the candidate blocks in the selected set, the encoder determines a correlation between, on the one hand, luma and chroma components of the input video block and, on the other hand, luma and chroma components of the respective candidate blocks. A predictor block is selected based on the correlation and is used to encode the input video block. In different embodiments, the correlation may be the negative of the sum of absolute differences of the components, may include a Jaccard similarity measure between respective pixels, or may be based on a Hamming distance between two high precision hash values of the input video block and the candidate block.
    Type: Grant
    Filed: June 18, 2015
    Date of Patent: January 19, 2021
    Assignee: Vid Scale, Inc.
    Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye, Ralph Neff
  • Patent number: 10887621
    Abstract: Processing video data may include capturing the video data with multiple cameras and stitching the video data together to obtain a 360-degree video. A frame-packed picture may be provided based on the captured and stitched video data. A current sample location may be identified in the frame-packed picture. Whether a neighboring sample location is located outside of a content boundary of the frame-packed picture may be determined. When the neighboring sample location is located outside of the content boundary, a padding sample location may be derived based on at least one circular characteristic of the 360-degree video content and the projection geometry. The 360-degree video content may be processed based on the padding sample location.
    Type: Grant
    Filed: July 7, 2017
    Date of Patent: January 5, 2021
    Assignee: VID SCALE, Inc.
    Inventors: Yuwen He, Yan Ye, Philippe Hanhart, Xiaoyu Xiu
  • Publication number: 20200404341
    Abstract: Cross-plane filtering may be used to restore blurred edges and/or textures in one or both chroma planes using information from a corresponding luma plane. Adaptive cross-plane filters may be implemented. Cross-plane filter coefficients may be quantized and/or signaled such that overhead in a bitstream minimizes performance degradation. Cross-plane filtering may be applied to select regions of a video image (e.g., to edge areas). Cross-plane filters may be implemented in single-layer video coding systems and/or multi-layer video coding systems.
    Type: Application
    Filed: September 2, 2020
    Publication date: December 24, 2020
    Applicant: InterDigital Madison Patent Holdings, SAS
    Inventors: Jie Dong, Yuwen He, Yan Ye
  • Publication number: 20200404321
    Abstract: Systems and methods are described for encoding and decoding video using derived block vectors as predictors in intra block copy mode. In an exemplary encoding method, an encoder identifies at least a first candidate block vector for the prediction of an input video block, where the first candidate block vector points to a first candidate block. The encoder then identifies a first predictive vector (e.g. a block vector or a motion vector) that was used to encode the first candidate block. From the first candidate block vector and the first predictive vector, the encoder generates a derived predictive vector from the first candidate block vector and the first predictive vector. The encoder then encodes the video block in the bit stream using the derived predictive vector for the prediction of the input video block.
    Type: Application
    Filed: September 1, 2020
    Publication date: December 24, 2020
    Inventors: Yuwen He, Yan Ye, Xiaoyu Xiu
  • Publication number: 20200374513
    Abstract: Video coding methods are described for reducing latency in template-based inter coding. In some embodiments, a method is provided for coding a video that includes a current picture and at least one reference picture. For at least a current block in the current picture, a respective predicted value is generated (e.g. using motion compensated prediction) for each sample in a template region adjacent to the current block. Once the predicted values are generated for each sample in the template region, a process is invoked to determine a template-based inter prediction parameter by using predicted values in the template region and sample values the reference picture. This process can be invoked without waiting for reconstructed sample values in the template region. Template-based inter prediction of the current block is then performed using the determined template-based inter prediction parameter.
    Type: Application
    Filed: March 22, 2019
    Publication date: November 26, 2020
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 10841615
    Abstract: Systems, methods, and devices are disclosed for performing adaptive color space conversion and adaptive entropy encoding of LUT parameters. A video bitstream may be received and a first flag may be determined based on the video bitstream. The residual may be converted from a first color space to a second color space in response to the first flag. The residual may be coded in two parts separated by the most significant bits and least significant bits of the residual. The residual may be further coded based on its absolute value.
    Type: Grant
    Filed: October 17, 2018
    Date of Patent: November 17, 2020
    Assignee: VID SCALE, Inc.
    Inventors: Yuwen He, Yan Ye
  • Publication number: 20200344458
    Abstract: A video coding device may be configured to periodically select the frame packing configuration (e.g., face layout and/or face rotations parameters) associated with a RAS, The device may receive a plurality of pictures, which may each comprise a plurality of faces. The pictures may be grouped Into a plurality of RASs. The device may select a frame packing configuration with the lowest cost for a first RAS. For example, the cost of a frame packing configuration may be determined based on the first picture of the first RAS. The device may select a frame packing configuration for a second RAS. The frame packing configuration for the first RAS may be different than the frame packing configuration for the second RAS. The frame packing configuration for the first RAS and the frame packing configuration for the second RAS may be signaled in the video bitstream.
    Type: Application
    Filed: January 14, 2019
    Publication date: October 29, 2020
    Applicant: VID SCALE, INC.
    Inventors: Philippe Hanhart, Yuwen He, Yan Ye
  • Publication number: 20200336738
    Abstract: Systems, methods, and Instrumentalities are described herein for calculating local Illumination compensation (LIC) parameters for bi-predicted coding unit (CU). The LIC parameters may be used to generate adjusted samples for the current CU and to address local illumination changes that may exist among temporal neighboring pictures. LIC parameters may be calculated based on bi-predicted reference template samples and template samples for a current CU. Bi-predicted reference template samples may be generated based on reference template samples neighboring temporal reference CUs. For example, the bi-predicted reference template samples may be generated based on averaging the reference template samples. The reference template samples may correspond to template samples for the current CU. A CU may be or may include a coding block and/or a sub-block that may be derived by dividing the coding block.
    Type: Application
    Filed: January 15, 2019
    Publication date: October 22, 2020
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye, Saurav Bandyopadhyay
  • Publication number: 20200322630
    Abstract: Video data may be palette decoded. Data defining a palette table may be received. The palette table may comprise index values corresponding to respective colors. Palette index prediction data may be received and may comprise data indicating index values for at least a portion of a palette index map mapping pixels of the video data to color indices in the palette table. The palette index prediction data may comprise run value data associating run values with index values for at least a portion of a palette index map. A run value may be associated with an escape color index. The palette index map may be generated from the palette index prediction data at least in part by determining whether to adjust an index value of the palette index prediction data based on a last index value. The video data may be reconstructed in accordance with the palette index map.
    Type: Application
    Filed: June 22, 2020
    Publication date: October 8, 2020
    Applicant: VID SCALE, INC.
    Inventors: Chia-Ming Tsai, Yuwen He, Xiaoyu Xiu, Yan Ye
  • Publication number: 20200322632
    Abstract: Systems, methods, and instrumentalities are disclosed for discontinuous face boundary filtering for 360-degree video coding. A face discontinuity may be filtered (e.g., to reduce seam artifacts) in whole or in part, for example, using coded samples or padded samples on either side of the face discontinuity. Filtering may be applied, for example, as an in-loop filter or a post-processing step. 2D positional information related to two sides of the face discontinuity may be signaled. In a video bitstream so that filtering may be applied independent of projection formats and/or frame packing techniques.
    Type: Application
    Filed: December 18, 2018
    Publication date: October 8, 2020
    Applicant: VID SCALE, INC.
    Inventors: Philippe Hanhart, Yan Ye, Yuwen He
  • Patent number: 10798423
    Abstract: Cross-plane filtering may be used to restore blurred edges and/or textures in one or both chroma planes using information from a corresponding luma plane. Adaptive cross-plane filters may be implemented. Cross-plane filter coefficients may be quantized and/or signaled such that overhead in a bitstream minimizes performance degradation. Cross-plane filtering may be applied to select regions of a video image (e.g., to edge areas). Cross-plane filters may be implemented in single-layer video coding systems and/or multi-layer video coding systems.
    Type: Grant
    Filed: July 1, 2019
    Date of Patent: October 6, 2020
    Assignee: InterDigital Madison Patent Holdings, SAS
    Inventors: Jie Dong, Yuwen He, Yan Ye
  • Publication number: 20200304788
    Abstract: A block may be identified. The block may be partitioned into one or more (e.g., two) sibling nodes (e.g., sibling nodes BO and B1). A partition direction and a partition type for the block may be determined. If the partition type for the block is binary tree (BT), one or more (e.g., two) partition parameters may be determined for sibling node BO. A partition parameter (e.g., a first partition parameter) may be determined for sibling node B1. A decoder may determine whether to receive an indication of a second partition parameter for B1 based on, for example, the partition direction for the block, the partition type for the block, and the first partition parameter for B1. The decoder may derive the second partition parameter based on, for example, the partition direction and type for the block, and the first partition parameter for B1.
    Type: Application
    Filed: November 1, 2018
    Publication date: September 24, 2020
    Applicant: VID SCALE, INC.
    Inventors: Yuwen He, Fanyi Duanmu, Xiaoyu Xiu, Yan Ye