Patents by Inventor Yuwen He

Yuwen He has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20220385897
    Abstract: A video processing apparatus may comprise one or more processors that are configured to determine an interpolation filter length for an interpolation filter associated with a coding unit (CU) based on a size of the CU. The one or more processor may be configured to determine an interpolated reference sample based on the determined interpolation filter length for the interpolation filter and a reference sample for the CU. The one or more processor may be configured to predict the CU based on the interpolated reference sample. For example, if a first CU has a size that is greater than the size of a second CU, the one or more processors may be configured to use a shorter interpolation filter for the first CU than for the second CU.
    Type: Application
    Filed: September 18, 2020
    Publication date: December 1, 2022
    Applicant: VID SCALE, INC.
    Inventors: Wei Chen, Yuwen He, Hua Yang
  • Publication number: 20220385942
    Abstract: Systems, methods, and instrumentalities are disclosed for discontinuous face boundary filtering for 360-degree video coding. A face discontinuity may be filtered (e.g., to reduce seam artifacts) in whole or in part, for example, using coded samples or padded samples on either side of the face discontinuity. Filtering may be applied, for example, as an in-loop filter or a post-processing step. 2D positional information related to two sides of the face discontinuity may be signaled in a video bitstream so that filtering may be applied independent of projection formats and/or frame packing techniques.
    Type: Application
    Filed: August 9, 2022
    Publication date: December 1, 2022
    Applicant: VID SCALE, INC.
    Inventors: Philippe Hanhart, Yan Ye, Yuwen He
  • Patent number: 11516485
    Abstract: Methods and systems are disclosed for a mobile device to decode video based on available power and/or energy. For example, the mobile device may receive a media description file (MDF) from for a video stream from a video server. The MDF may include complexity information associated with a plurality of video segments. The complexity information may be related to the amount of processing power to be utilized for decoding the segment at the mobile device. The mobile device may determine at least one power metric for the mobile device. The mobile device may determine a first complexity level to be requested for a first video segment based on the complexity information from the MDF and the power metric. The mobile device may dynamically alter the decoding process to save energy based on the detected power/energy level.
    Type: Grant
    Filed: May 18, 2021
    Date of Patent: November 29, 2022
    Assignee: VID SCALE, Inc.
    Inventors: Yuwen He, Yan Ye, Yong He, George W. McClellan, Eun Seok Ryu
  • Publication number: 20220377316
    Abstract: An apparatus may be configured to determine a reference picture listed in a first reference picture list and a reference picture listed in a second reference picture list, for a coding block. The apparatus may be configured to determine whether to perform bi-directional optical flow (BDOF) for the coding block based at least in part on whether a distance between a picture associated with the coding block and the reference picture listed in the first reference picture list differs from a distance between the picture associated with the coding block and the reference picture listed in the second reference picture list. The apparatus may be configured to decode the coding block based on the determination of whether to perform BDOF for the coding block.
    Type: Application
    Filed: September 23, 2020
    Publication date: November 24, 2022
    Applicant: VID SCALE, INC.
    Inventors: Wei Chen, Yuwen He, Hua Yang
  • Publication number: 20220377324
    Abstract: A filter may be applied to a subset of components associated with a sample in a coding block. The output of the filter may be used to modify values for other component(s). For example, a filter may be applied to a selected (for example, dominant) component(s). The output of the filter may be used to modify a value for one of the other components (for example, non-dominant components). The output of the filter may be used, for example, after a weighting factor is applied to the filter output, to modify a value for another one of the other components. A joint refinement signal may be obtained, for example, as the filtered output signal minus the filter input signal of the selected component(s). A properly weighted version of the joint refinement signal may be applied to modify the other components.
    Type: Application
    Filed: September 23, 2020
    Publication date: November 24, 2022
    Applicant: VID SCALE, INC.
    Inventors: Hua Yang, Yuwen He, Wei Chen, Hongyu Li
  • Publication number: 20220377344
    Abstract: Described herein are systems, methods, and instrumentalities associated with video coding. The signaling of certain syntax elements may be moved from a slice header to a picture header and/or a layer access unit delimiter (AUD). The dependency between AUD and one or more parameter sets may be explored. Syntax elements may be signaled to enable wrap-around motion compensation for certain sub-picture(s) and specify wrap-around motion compensation offsets for the sub-picture(s).
    Type: Application
    Filed: September 18, 2020
    Publication date: November 24, 2022
    Applicant: VID SCALE, INC.
    Inventors: Yong He, Yuwen He
  • Publication number: 20220377385
    Abstract: Systems, methods, and instrumentalities may be provided for discounting reconstructed samples and/or coding information from spatial neighbors across face discontinuities. Whether a current block is located at a face discontinuity may be determined. The face discontinuity may be a face boundary between two or more adjoining blocks that are not spherical neighbors. The coding availability of a neighboring block of the current block may be determined, e.g., based on whether the neighboring block is on the same side of the face discontinuity as the current block. For example, the neighboring block may be determined to be available for decoding the current block if it is on the same side of the face discontinuity as the current block, and unavailable if it is not on the same side of the face discontinuity. The neighboring block may be a spatial neighboring block or a temporal neighboring block.
    Type: Application
    Filed: July 29, 2022
    Publication date: November 24, 2022
    Applicant: Vid Scale, Inc.
    Inventors: Philippe Hanhart, Yuwen He, Yan Ye
  • Publication number: 20220368947
    Abstract: Processing video data may include capturing the video data with multiple cameras and stitching the video data together to obtain a 360-degree video. A frame-packed picture may be provided based on the captured and stitched video data. A current sample location may be identified in the frame-packed picture. Whether a neighboring sample location is located outside of a content boundary of the frame-packed picture may be determined. When the neighboring sample location is located outside of the content boundary, a padding sample location may be derived based on at least one circular characteristic of the 360-degree video content and the projection geometry. The 360-degree video content may be processed based on the padding sample location.
    Type: Application
    Filed: July 8, 2022
    Publication date: November 17, 2022
    Applicant: VID SCALE, INC.
    Inventors: Yuwen He, Yan Ye, Philippe Hanhart, Xiaoyu Xiu
  • Publication number: 20220368943
    Abstract: A device may determine whether to enable or disable bi-directional optical flow (BIO) for a current coding unit (CU) (e.g., block and/or sub-block). Prediction information for the CU may be identified and may include prediction signals associated with a first reference block and a second reference block (e.g., or a first reference sub-block and a second reference sub-block). A prediction difference may be calculated and may be used to determine the similarity between the two prediction signals. The CU may be reconstructed based on the similarity. For example, whether to reconstruct the CU with BIO enabled or BIO disabled may be based on whether the two prediction signals are similar. It may be determined to enable BIO for the CU when the two prediction signals are determined to be dissimilar. For example, the CU may be reconstructed with BIO disabled when the two prediction signals are determined to be similar.
    Type: Application
    Filed: May 20, 2022
    Publication date: November 17, 2022
    Applicant: VID SCALE, INC.
    Inventors: Yan Ye, Xiaoyu Xiu, Yuwen He
  • Patent number: 11503328
    Abstract: Systems, methods, and instrumentalities are disclosed for motion vector clipping when affine motion mode is enabled for a video block. A video coding device may determine that an affine mode for a video block is enabled. The video coding device may determine a plurality of control point affine motion vectors associated with the video block. The video coding device may store the plurality of clipped control point affine motion vectors for motion vector prediction of a neighboring control point affine motion vector. The video coding device may derive a sub-block motion vector associated with a sub-block of the video block, clip the derived sub-block motion vector, and store it for spatial motion vector prediction or temporal motion vector prediction. For example, the video coding device may clip the derived sub-block motion vector based on a motion field range that may be based on a bit depth value.
    Type: Grant
    Filed: June 27, 2019
    Date of Patent: November 15, 2022
    Assignee: VID SCALE, Inc.
    Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye, Philippe Hanhart
  • Publication number: 20220360778
    Abstract: A method of encoding or decoding a video comprising a current picture, a first reference picture, and a weight tensor associated with a trained neural network (NN) model are provided. The method includes generating any number of kernel tensors, input channels and output channels associated with the weight tensor, each kernel tensor being associated with any of: a layer type, an input signal type, and a tree partition type, and each kernel tensor including weight coefficients, generating, for each of the any number of kernel tensors, tree partitions for any of a coding tree unit (CTU), a coding unit (CU), a prediction unit (PU), and a transform unit (TU) according to respective tree partition types associated with each of the any number of kernel tensors, and generating a compressed representation of the trained NN model by compressing and coding the any number of kernel tensors.
    Type: Application
    Filed: July 2, 2020
    Publication date: November 10, 2022
    Inventors: Hua Yang, Duanshun Li, Dong Tian, Yuwen He
  • Publication number: 20220345701
    Abstract: Systems, methods, and instrumentalities may be used for decoding and/or encoding a coding unit (CD), An intra-prediction mode for a CD may be determined. A split mode may be determined based on the intra-prediction mode, to generate a plurality of sub-partitions in the CU. A prediction for a first sub-partition of the plurality of sub-partitions in the CU may be based on a reference sample in a second sub-partition of the plurality of sub-partitions in the CU. The CU may be decoded and/or encoded, for example, based on the determined split mode.
    Type: Application
    Filed: September 17, 2020
    Publication date: October 27, 2022
    Applicant: VID SCALE, INC.
    Inventors: Hua Yang, Yuwen He, Wei Chen
  • Patent number: 11477437
    Abstract: Implementations of the present disclosure provide a solution for encoding and decoding motion information. In this solution, during a conversion between a current video block of a video and a bitstream of the video, a group type of a subgroup of candidates of motion information for the current video block is determined, wherein the group type indicating whether the subgroup of candidates are to be reordered. Further, a list of candidates are constructed based on the group type; and the motion information for the current video block is derived from the list of candidates.
    Type: Grant
    Filed: January 28, 2021
    Date of Patent: October 18, 2022
    Assignee: Lemon Inc.
    Inventors: Na Zhang, Kai Zhang, Li Zhang, Hongbin Liu, Yuwen He, Yue Wang
  • Publication number: 20220329826
    Abstract: Example implementations include a method, apparatus and computer-readable medium of video processing, including constructing, during a conversion between a current video block of a video and a bitstream of the video, at least one template set for the current video block from a plurality of sub-templates. The one or more sub-templates may be selected from a plurality of sub-templates including: a left sub-template, an above sub-template, a right-above sub-template, a left-below sub-template, and a left-above sub-template. The implementations further include deriving at least one intra-prediction mode (IPM) based on cost calculations. The implementations include determining, based on the at least one IPM, a final predictor of the current video block. The implementations include performing the conversion based on the final predictor.
    Type: Application
    Filed: June 10, 2022
    Publication date: October 13, 2022
    Inventors: Yang WANG, Kai ZHANG, Li ZHANG, Yuwen HE, Hongbin LIU
  • Publication number: 20220329816
    Abstract: A method of video processing includes performing a conversion between a current video block of a second color component of a video and a bitstream of the video using a cross-component prediction with multiple-parameter model (CCPMPM) in which samples of the current video block are predictively coded in the bitstream using a linear combination of samples of a first color component multiplied by linear coefficients and/or one or more offsets. The bitstream conforms to a format rule. The linear coefficients of the CCPMPM are determined using a first rule. The samples of the first color component are determined using a second rule.
    Type: Application
    Filed: June 24, 2022
    Publication date: October 13, 2022
    Inventors: Hongbin LIU, Li ZHANG, Kai ZHANG, Jizheng XU, Yuwen HE, Yue WANG
  • Publication number: 20220329831
    Abstract: Cross plane filtering may be used for enhanced chroma coding. An indication of a cross-plane filter associated with a current picture may be received. The current picture may include an intra-coded video block and a plurality of reference samples. The plurality of reference samples may be used to predict the intra-coded video block. A luma sample region may be determined in the current picture. The luma sample region may determined to enhance a corresponding chroma sample in the current picture. The cross-plane filter may be applied to a plurality of luma samples in the luma sample region to determine an offset. The cross-plane filter may be a high pass filter. The offset may be applied to the corresponding chroma sample to determine an enhanced chroma sample.
    Type: Application
    Filed: June 23, 2022
    Publication date: October 13, 2022
    Applicant: InterDigital Madison Patent Holdings, SAS
    Inventors: Yong He, Yuwen He, Yan Ye, Xiaoyu Xiu
  • Patent number: 11470308
    Abstract: Systems and methods are described for reducing the complexity of using bi-directional optical flow (BIO) in video coding. In some embodiments, bit-width reduction steps are introduced in the BIO motion refinement process to reduce the maximum bit-width used for BIO calculations. In some embodiments, simplified interpolation filters are used to generate predicted samples in an extended region around a current coding unit. In some embodiments, different interpolation filters are used for vertical versus horizontal interpolation. In some embodiments, BIO is disabled for coding units with small heights and/or for coding units that are predicted using a sub-block level inter prediction technique, such as advanced temporal motion vector prediction (ATMVP) or affine prediction.
    Type: Grant
    Filed: September 17, 2019
    Date of Patent: October 11, 2022
    Assignee: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye, Jiancong Luo
  • Patent number: 11470344
    Abstract: Systems and methods are described for selecting a motion vector (MV) to use in frame-rate up conversion (FRUC) coding of a block of video. In one embodiment, a first set of motion vector candidates is identified for FRUC prediction of the block. A search center is defined based on the first set of motion vector candidates, and a search window is determined, the search window having a selected width and being centered on the search center. A search for a selected MV is performed within the search window. In some embodiments, an initial set of MVs is processed with a clustering algorithm to generate a smaller number of MVs that are used as the first set. The selected MV may be subject to a motion refinement search, which may also be performed over a constrained search range. In additional embodiments, search iterations are constrained to limit complexity.
    Type: Grant
    Filed: January 28, 2019
    Date of Patent: October 11, 2022
    Assignee: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20220312001
    Abstract: Systems and methods are described for video coding using generalized bi-prediction. In an exemplary embodiment, to code a current block of a video in a bitstream, a first reference block is selected from a first reference picture and a second reference block is selected from a second reference picture. Each reference block is associated with a weight, where the weight may be an arbitrary weight ranging, e.g., between 0 and 1. The current block is predicted using a weighted sum of the reference blocks. The weights may be selected from among a plurality of candidate weights. Candidate weights may be signaled in the bitstream or may be derived implicitly based on a template. Candidate weights may be pruned to avoid out-of-range or substantially duplicate candidate weights. Generalized bi-prediction may additionally be used in frame rate up conversion.
    Type: Application
    Filed: June 13, 2022
    Publication date: September 29, 2022
    Inventors: Chun-Chi Chen, Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 11457198
    Abstract: A video coding device may be configured to periodically select the frame packing configuration (e.g., face layout and/or face rotations parameters) associated with a RAS, The device may receive a plurality of pictures, which may each comprise a plurality of faces. The pictures may be grouped Into a plurality of RASs. The device may select a frame packing configuration with the lowest cost for a first RAS. For example, the cost of a frame packing configuration may be determined based on the first picture of the first RAS. The device may select a frame packing configuration for a second RAS. The frame packing configuration for the first RAS may be different than the frame packing configuration for the second RAS. The frame packing configuration for the first RAS and the frame packing configuration for the second RAS may be signaled in the video bitstream.
    Type: Grant
    Filed: January 14, 2019
    Date of Patent: September 27, 2022
    Assignee: VID SCALE, Inc.
    Inventors: Philippe Hanhart, Yuwen He, Yan Ye