Patents by Inventor Yuwen He

Yuwen He has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11317085
    Abstract: Video coding methods are described for reducing latency in template-based inter coding. In some embodiments, a method is provided for coding a video that includes a current picture and at least one reference picture. For at least a current block in the current picture, a respective predicted value is generated (e.g. using motion compensated prediction) for each sample in a template region adjacent to the current block. Once the predicted values are generated for each sample in the template region, a process is invoked to determine a template-based inter prediction parameter by using predicted values in the template region and sample values the reference picture. This process can be invoked without waiting for reconstructed sample values in the template region. Template-based inter prediction of the current block is then performed using the determined template-based inter prediction parameter.
    Type: Grant
    Filed: March 22, 2019
    Date of Patent: April 26, 2022
    Assignee: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20220116656
    Abstract: Methods, procedures, architectures, apparatuses, systems, devices, interfaces, and computer program products for encoding/decoding data (e.g. a data stream) are provided. A video coding method for predicting a current block includes identifying a first block adjacent to the current block, the first block having motion information, performing motion compensation using the motion information to generate a set of reference samples adjacent to the current block, identifying a first line of reference samples from the set of generated reference samples to be used for intra prediction of the current block, and performing intra prediction of the current block using at least the first line of reference samples.
    Type: Application
    Filed: January 10, 2020
    Publication date: April 14, 2022
    Inventors: Rahul Vanam, Yuwen He
  • Patent number: 11290736
    Abstract: Aspects include a method, apparatus and computer-readable medium of decoding video or blocks of an image, including receiving a bitstream of the image, deriving, for a block of the image in the bitstream, multiple intra-prediction modes (IPMs) to use in decoding the block, determining, based on the multiple IPMs, a final predictor to use in decoding the block, and decoding the block using the final predictor. Other aspects include method, apparatus and computer-readable medium for similarly encoding video or blocks of an image.
    Type: Grant
    Filed: January 13, 2021
    Date of Patent: March 29, 2022
    Assignee: Lemon Inc.
    Inventors: Yang Wang, Kai Zhang, Li Zhang, Yuwen He, Hongbin Liu
  • Publication number: 20220094940
    Abstract: Procedures, methods, architectures, apparatuses, systems, devices, and computer program products directed to improved linear model estimation for template-based video coding are provided. Included therein is a method comprising determining minimum and maximum (“min/max”) values of luma and chroma samples neighboring a coding block, wherein the min/max chroma values correspond to the min/max luma values; determining a first linear model parameter of a template-based video coding technique (i) based on a single look-up table and the min/max chroma values and (ii) at a precision no greater than 16 bits; determining a second linear model parameter of the template-based video coding technique (i) based on the first linear model parameter and the minimum chroma and luma values and (ii) at a precision no greater than 16 bits; and predicting chroma samples of the coding block based on reconstructed luma samples of the coding block and the first and second linear model parameters.
    Type: Application
    Filed: December 20, 2019
    Publication date: March 24, 2022
    Inventors: Saurav Bandyopadhyay, Xiaoyu Xiu, Yuwen He
  • Patent number: 11284089
    Abstract: Coding techniques for 360-degree video. An encoder selects a projection format and maps the 360-degree video to a 2D planar video using the selected projection format. The encoder encodes the 2D planar video in a bitstream and further signals, in the bitstream, parameters identifying the projection format. The parameters identifying the projection format may be signaled in a video parameter set, sequence parameter set, and/or picture parameter set of the bitstream. Different projection formats that may be signaled include formats using geometries such as equirectangular, cubemap, equal-area, octahedron, icosahedron, cylinder, and user-specified polygon. Other parameters that may be signaled include different arrangements of geometric faces or different encoding quality for different faces. Corresponding decoders. Projection parameters may further include relative geometry rotation parameters that define an orientation of the projection geometry.
    Type: Grant
    Filed: August 30, 2017
    Date of Patent: March 22, 2022
    Assignee: VID SCALE, INC.
    Inventors: Philippe Hanhart, Yuwen He, Yan Ye
  • Patent number: 11277635
    Abstract: A video coding system (e.g., an encoder and/or a decoder) may perform face-based sub-block motion compensation for 360-degree video to predict samples (e.g., of a sub-block). The video coding system may receive a 360-degree video content. The 360-degree video content may include a current block. The current block may include a plurality of sub-blocks. The system may determine whether a sub-block mode is used for the current block. The system may predict a sample in the current block based on the sub-block level face association. For a first sub-block in the current block, the system may identify a first location of the first sub-block. The system may associate the first sub-block with a first face based on the identified first location of the first sub-block. The system may predict a first sample in the first sub-block based on the first face that is associated with the first sub-block.
    Type: Grant
    Filed: March 15, 2018
    Date of Patent: March 15, 2022
    Assignee: VID SCALE, Inc.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20220078406
    Abstract: Systems, methods, and instrumentalities are disclosed for processing history-based motion vector prediction (HMVP). A video coding device may generate a history-based motion vector prediction (HMVP) list for a current block. The video coding device derive an HMVP candidate from a previously coded block. The HMVP candidate may include motion information associated with a neighboring block of the current block, one or more reference indices, and a bi-prediction weight index. The video coding device may add the HMVP candidate to the HMVP list for motion compensated prediction of a motion vector associated with the current block. The video coding device use one HMVP selected from the HMVP list to perform motion compensated prediction of the current block. The motion compensated prediction may be performed using the motion information associated with the neighboring block of the current block, the one or more reference indices, and the bi-prediction weight index.
    Type: Application
    Filed: December 19, 2019
    Publication date: March 10, 2022
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Jiancong Luo
  • Publication number: 20220070441
    Abstract: Systems, methods, and instrumentalities are disclosed for a combined inter and intra prediction, A video coding device may receive a motion vector difference (MMVD) mode indication that indicates whether MMVD mode is used to generate inter prediction of a coding unit (CU). The video coding device may receive a combined inter merge/intra prediction (CUP) indication, for example, when the MMVD mode indication indicates that MMVD mode is not used to generate the inter prediction of the CU, The video coding device may determine whether to use triangle merge mode for the CU, for example, based on the MMVD mode indication and/or the CUP indication. On a condition that the CUP indication indicates that CUP is applied for the CU or the MMVD mode indication indicates that MMVD mode is used to generate the inter prediction, the video coding device may disable the triangle merge mode for the CU.
    Type: Application
    Filed: December 20, 2019
    Publication date: March 3, 2022
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Jiancong Luo
  • Publication number: 20220038679
    Abstract: Systems and methods are described for reducing the complexity of using bi-directional optical flow (BIO) in video coding. In some embodiments, bit-width reduction steps are introduced in the BIO motion refinement process to reduce the maximum bit-width used for BIO calculations. In some embodiments, simplified interpolation filters are used to generate predicted samples in an extended region around a current coding unit. In some embodiments, different interpolation filters are used for vertical versus horizontal interpolation. In some embodiments, BIO is disabled for coding units with small heights and/or for coding units that are predicted using a sub-block level inter prediction technique, such as advanced temporal motion vector prediction (ATMVP) or affine prediction.
    Type: Application
    Filed: September 17, 2019
    Publication date: February 3, 2022
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye, Jiancong Luo
  • Publication number: 20220038738
    Abstract: Exemplary embodiments include systems and methods for coding a video comprising a plurality of pictures including a current picture, a first reference picture, and a second reference picture, where each picture includes a plurality of blocks. In one method, for at least a current block in the current picture, a number of available bi-prediction weights is determined based at least in part on a temporal layer and/or a quantization parameter of the current picture. From among available bi-prediction weights a pair of weights are identified. Using the identified weights, the current block is then predicted as a weighted sum of a first reference block in the first reference picture and a second reference block in the second reference picture. Encoding techniques are also described for efficient searching and selection of a pair of bi-prediction weights to use for prediction of a block.
    Type: Application
    Filed: October 18, 2021
    Publication date: February 3, 2022
    Inventors: Yan Zhang, Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20220038771
    Abstract: Power aware adaptation for a power aware video streaming system may be based on the complexity information conveyed in different ways. A complexity level of a data stream, such as a video data stream, may be selected as a function of a remaining battery power of a wireless transmit/receive unit (WTRU) and on a state set of a plurality of state sets that may be stored and/or managed by the WTRU. These state sets may correspond to, for example, different content sources and/or different complexity estimation algorithms and may be used to select the complexity level of the data stream. The data stream may then be received at the selected complexity level. The complexity level and/or a bitrate of the data stream may be adapted to accommodate, for example, the remaining battery power and/or other circumstances. The adaptation may be customized according to the objectives of use cases.
    Type: Application
    Filed: October 18, 2021
    Publication date: February 3, 2022
    Applicant: InterDigital Patent Holdings, Inc.
    Inventors: Yuwen He, Markus Kunstner, Yan Ye, Ralph Neff
  • Patent number: 11223848
    Abstract: 360-degree video content may be coded. A sampling position in a projection format may be determined to code 360-degree video content. For example, a sampling position in a target projection format and a sampling position in a reference projection format may be identified. The sample position in the target projection format may be related to the corresponding sample position in the reference projection format via a transform function. A parameter weight (e.g., a reference parameter weight) for the sampling position in the reference projection format may be identified. An adjustment factor associated with the parameter weight for the sampling position in the reference projection format may be determined. The parameter weight (e.g., adjusted parameter weight) for the sampling position in the target projection format may be calculated. The calculated adjusted parameter weight may be applied to the sampling position in the target projection format when coding the 360-degree video content.
    Type: Grant
    Filed: June 29, 2018
    Date of Patent: January 11, 2022
    Assignee: VID SCALE, Inc.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20220007053
    Abstract: Systems, methods, and instrumentalities are disclosed for performing horizontal geometry padding on a current sample based on receiving a wraparound enabled indication that indicates whether a horizontal wraparound motion compensation is enabled. If the horizontal wraparound motion compensation is enabled based on the wraparound enabled indication, a video coding device may determine a reference sample wraparound offset of a current sample in a picture. The reference sample wraparound offset may indicate a face width of the picture. The video coding device may determine a reference sample location for the current sample based on the reference sample wraparound offset, a picture width of the picture, and a current sample location. The video coding device may predict the current sample based on the reference sample location in a horizontal direction. Repetitive padding or clipping may be used in the vertical direction.
    Type: Application
    Filed: September 25, 2019
    Publication date: January 6, 2022
    Applicant: Vid Scale, Inc.
    Inventors: Philippe Hanhart, Yuwen He, Yan Ye
  • Publication number: 20220007048
    Abstract: Systems, methods, and instrumentalities may be provided for determining whether to bypass bi-directional optical flow (BDOF) if BDOF is used in combination with bi-prediction with coding unit (CU) weights (e.g., generalized bi-prediction (GBi)). A coding system may combine coding modes, coding techniques, and/or coding tools. The coding system may include a wireless transmit/receive unit (WTRU). For example, the coding system may combine BDOF and bi-prediction with GU weights (BCW). BDOF may include refining a motion vector associated with a current CU based at least in part on gradients associated with a location in the current CU. The coding system may determine that BDOF is enabled, and/or that bi-prediction with CU weights is enabled for the current CU. The coding system's determination that bi-prediction with CU weights is enabled and/or that BDOF is enabled may be based on one or more indications.
    Type: Application
    Filed: September 26, 2019
    Publication date: January 6, 2022
    Applicant: VID SCALE, INC.
    Inventors: Yuwen He, Jiancong Luo, Xiaoyu Xiu, Yan Ye
  • Patent number: 11212496
    Abstract: Conversion between different projection formats of a 360-degree video may be performed in a uniform way. The geometric characteristics of the different projection formats may be considered when applying 3D-to-2D and 2D-to-3D mapping. Parameters reflective of the geometric characteristics of the different projection formats may be determined and used in the mapping and/or conversion. The parameters may include a normal vector that is perpendicular to a projection plane, a reference point in the projection plane, and/or unit vectors defined in the projection plane. An architecture with consolidated modules for handling the various projection formats may be provided.
    Type: Grant
    Filed: October 6, 2017
    Date of Patent: December 28, 2021
    Assignee: VID SCALE, Inc.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye, Philippe Hanhart
  • Publication number: 20210337202
    Abstract: Systems, procedures, and instrumentalities may be provided for adaptively adjusting quantization parameters (QPs) for 360-degree video coding. For example, a first luma QP for a first region may be identified. Based on the first luma QP, a first chroma QP for the first region may be determined. A QP offset for a second region may be identified. A second luma QP for the second region may be determined based on the first luma QP and/or the QP offset for the second region. A second chroma QP of the second region may be determined based on the first chroma QP and/or the QP offset for the second region. An inverse quantization may be performed for the second region based on the second luma QP for the second region and/or the second chroma QP for the second region. The QP offset may be adapted based on a spherical sampling density.
    Type: Application
    Filed: June 21, 2018
    Publication date: October 28, 2021
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 11153645
    Abstract: Power aware adaptation for a power aware video streaming system may be based on the complexity information conveyed in different ways. A complexity level of a data stream, such as a video data stream, may be selected as a function of a remaining battery power of a wireless transmit/receive unit (WTRU) and on a state set of a plurality of state sets that may be stored and/or managed by the WTRU. These state sets may correspond to, for example, different content sources and/or different complexity estimation algorithms and may be used to select the complexity level of the data stream. The data stream may then be received at the selected complexity level. The complexity level and/or a bitrate of the data stream may be adapted to accommodate, for example, the remaining battery power and/or other circumstances. The adaptation may be customized according to the objectives of use cases.
    Type: Grant
    Filed: September 24, 2019
    Date of Patent: October 19, 2021
    Assignee: InterDigital Patent Holdings, Inc.
    Inventors: Yuwen He, Markus Kunstner, Yan Ye, Ralph Neff
  • Patent number: 11153602
    Abstract: Exemplary embodiments include systems and methods for coding a video comprising a plurality of pictures including a current picture, a first reference picture, and a second reference picture, where each picture includes a plurality of blocks. In one method, for at least a current block in the current picture, a number of available bi-prediction weights is determined based at least in part on a temporal layer and/or a quantization parameter of the current picture. From among available bi-prediction weights a pair of weights are identified. Using the identified weights, the current block is then predicted as a weighted sum of a first reference block in the first reference picture and a second reference block in the second reference picture. Encoding techniques are also described for efficient searching and selection of a pair of bi-prediction weights to use for prediction of a block.
    Type: Grant
    Filed: January 23, 2019
    Date of Patent: October 19, 2021
    Assignee: Vid Scale, Inc.
    Inventors: Yan Zhang, Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 11134259
    Abstract: Video coding systems and methods are described using an enhanced motion compensated picture. In exemplary embodiments, an enhanced motion compensated picture is generated by applying at least one high-pass filter to the motion compensated picture and adding the output of the filter to the motion compensated picture. Coefficients of the high-pass filter are selected by comparing the enhanced motion compensated picture to an original picture. The selected coefficients may be quantized and entropy coded into a bit stream. The high-pass filter may be a cross-plane filter, in which a luma component, which may be an enhanced luma component, is high-pass filtered and the output of the filter is added to at least one of the chroma components to generate an enhanced chroma component.
    Type: Grant
    Filed: January 9, 2017
    Date of Patent: September 28, 2021
    Assignee: INTERDIGITAL MADISON PATENT HOLDINGS, SAS
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20210274199
    Abstract: Methods and systems are disclosed for a mobile device to decode video based on available power and/or energy. For example, the mobile device may receive a media description file (MDF) from for a video stream from a video server. The MDF may include complexity information associated with a plurality of video segments. The complexity information may be related to the amount of processing power to be utilized for decoding the segment at the mobile device. The mobile device may determine at least one power metric for the mobile device. The mobile device may determine a first complexity level to be requested for a first video segment based on the complexity information from the MDF and the power metric. The mobile device may dynamically alter the decoding process to save energy based on the detected power/energy level.
    Type: Application
    Filed: May 18, 2021
    Publication date: September 2, 2021
    Applicant: VID SCALE, Inc.
    Inventors: Yuwen He, Yan Ye, Yong He, George W. McClellan, Eun Seok Ryu