Patents by Inventor Yuwen He
Yuwen He has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20220385897Abstract: A video processing apparatus may comprise one or more processors that are configured to determine an interpolation filter length for an interpolation filter associated with a coding unit (CU) based on a size of the CU. The one or more processor may be configured to determine an interpolated reference sample based on the determined interpolation filter length for the interpolation filter and a reference sample for the CU. The one or more processor may be configured to predict the CU based on the interpolated reference sample. For example, if a first CU has a size that is greater than the size of a second CU, the one or more processors may be configured to use a shorter interpolation filter for the first CU than for the second CU.Type: ApplicationFiled: September 18, 2020Publication date: December 1, 2022Applicant: VID SCALE, INC.Inventors: Wei Chen, Yuwen He, Hua Yang
-
Publication number: 20220385942Abstract: Systems, methods, and instrumentalities are disclosed for discontinuous face boundary filtering for 360-degree video coding. A face discontinuity may be filtered (e.g., to reduce seam artifacts) in whole or in part, for example, using coded samples or padded samples on either side of the face discontinuity. Filtering may be applied, for example, as an in-loop filter or a post-processing step. 2D positional information related to two sides of the face discontinuity may be signaled in a video bitstream so that filtering may be applied independent of projection formats and/or frame packing techniques.Type: ApplicationFiled: August 9, 2022Publication date: December 1, 2022Applicant: VID SCALE, INC.Inventors: Philippe Hanhart, Yan Ye, Yuwen He
-
Patent number: 11516485Abstract: Methods and systems are disclosed for a mobile device to decode video based on available power and/or energy. For example, the mobile device may receive a media description file (MDF) from for a video stream from a video server. The MDF may include complexity information associated with a plurality of video segments. The complexity information may be related to the amount of processing power to be utilized for decoding the segment at the mobile device. The mobile device may determine at least one power metric for the mobile device. The mobile device may determine a first complexity level to be requested for a first video segment based on the complexity information from the MDF and the power metric. The mobile device may dynamically alter the decoding process to save energy based on the detected power/energy level.Type: GrantFiled: May 18, 2021Date of Patent: November 29, 2022Assignee: VID SCALE, Inc.Inventors: Yuwen He, Yan Ye, Yong He, George W. McClellan, Eun Seok Ryu
-
Publication number: 20220377316Abstract: An apparatus may be configured to determine a reference picture listed in a first reference picture list and a reference picture listed in a second reference picture list, for a coding block. The apparatus may be configured to determine whether to perform bi-directional optical flow (BDOF) for the coding block based at least in part on whether a distance between a picture associated with the coding block and the reference picture listed in the first reference picture list differs from a distance between the picture associated with the coding block and the reference picture listed in the second reference picture list. The apparatus may be configured to decode the coding block based on the determination of whether to perform BDOF for the coding block.Type: ApplicationFiled: September 23, 2020Publication date: November 24, 2022Applicant: VID SCALE, INC.Inventors: Wei Chen, Yuwen He, Hua Yang
-
Publication number: 20220377324Abstract: A filter may be applied to a subset of components associated with a sample in a coding block. The output of the filter may be used to modify values for other component(s). For example, a filter may be applied to a selected (for example, dominant) component(s). The output of the filter may be used to modify a value for one of the other components (for example, non-dominant components). The output of the filter may be used, for example, after a weighting factor is applied to the filter output, to modify a value for another one of the other components. A joint refinement signal may be obtained, for example, as the filtered output signal minus the filter input signal of the selected component(s). A properly weighted version of the joint refinement signal may be applied to modify the other components.Type: ApplicationFiled: September 23, 2020Publication date: November 24, 2022Applicant: VID SCALE, INC.Inventors: Hua Yang, Yuwen He, Wei Chen, Hongyu Li
-
Publication number: 20220377344Abstract: Described herein are systems, methods, and instrumentalities associated with video coding. The signaling of certain syntax elements may be moved from a slice header to a picture header and/or a layer access unit delimiter (AUD). The dependency between AUD and one or more parameter sets may be explored. Syntax elements may be signaled to enable wrap-around motion compensation for certain sub-picture(s) and specify wrap-around motion compensation offsets for the sub-picture(s).Type: ApplicationFiled: September 18, 2020Publication date: November 24, 2022Applicant: VID SCALE, INC.Inventors: Yong He, Yuwen He
-
Publication number: 20220377385Abstract: Systems, methods, and instrumentalities may be provided for discounting reconstructed samples and/or coding information from spatial neighbors across face discontinuities. Whether a current block is located at a face discontinuity may be determined. The face discontinuity may be a face boundary between two or more adjoining blocks that are not spherical neighbors. The coding availability of a neighboring block of the current block may be determined, e.g., based on whether the neighboring block is on the same side of the face discontinuity as the current block. For example, the neighboring block may be determined to be available for decoding the current block if it is on the same side of the face discontinuity as the current block, and unavailable if it is not on the same side of the face discontinuity. The neighboring block may be a spatial neighboring block or a temporal neighboring block.Type: ApplicationFiled: July 29, 2022Publication date: November 24, 2022Applicant: Vid Scale, Inc.Inventors: Philippe Hanhart, Yuwen He, Yan Ye
-
Publication number: 20220368947Abstract: Processing video data may include capturing the video data with multiple cameras and stitching the video data together to obtain a 360-degree video. A frame-packed picture may be provided based on the captured and stitched video data. A current sample location may be identified in the frame-packed picture. Whether a neighboring sample location is located outside of a content boundary of the frame-packed picture may be determined. When the neighboring sample location is located outside of the content boundary, a padding sample location may be derived based on at least one circular characteristic of the 360-degree video content and the projection geometry. The 360-degree video content may be processed based on the padding sample location.Type: ApplicationFiled: July 8, 2022Publication date: November 17, 2022Applicant: VID SCALE, INC.Inventors: Yuwen He, Yan Ye, Philippe Hanhart, Xiaoyu Xiu
-
Publication number: 20220368943Abstract: A device may determine whether to enable or disable bi-directional optical flow (BIO) for a current coding unit (CU) (e.g., block and/or sub-block). Prediction information for the CU may be identified and may include prediction signals associated with a first reference block and a second reference block (e.g., or a first reference sub-block and a second reference sub-block). A prediction difference may be calculated and may be used to determine the similarity between the two prediction signals. The CU may be reconstructed based on the similarity. For example, whether to reconstruct the CU with BIO enabled or BIO disabled may be based on whether the two prediction signals are similar. It may be determined to enable BIO for the CU when the two prediction signals are determined to be dissimilar. For example, the CU may be reconstructed with BIO disabled when the two prediction signals are determined to be similar.Type: ApplicationFiled: May 20, 2022Publication date: November 17, 2022Applicant: VID SCALE, INC.Inventors: Yan Ye, Xiaoyu Xiu, Yuwen He
-
Patent number: 11503328Abstract: Systems, methods, and instrumentalities are disclosed for motion vector clipping when affine motion mode is enabled for a video block. A video coding device may determine that an affine mode for a video block is enabled. The video coding device may determine a plurality of control point affine motion vectors associated with the video block. The video coding device may store the plurality of clipped control point affine motion vectors for motion vector prediction of a neighboring control point affine motion vector. The video coding device may derive a sub-block motion vector associated with a sub-block of the video block, clip the derived sub-block motion vector, and store it for spatial motion vector prediction or temporal motion vector prediction. For example, the video coding device may clip the derived sub-block motion vector based on a motion field range that may be based on a bit depth value.Type: GrantFiled: June 27, 2019Date of Patent: November 15, 2022Assignee: VID SCALE, Inc.Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye, Philippe Hanhart
-
Publication number: 20220360778Abstract: A method of encoding or decoding a video comprising a current picture, a first reference picture, and a weight tensor associated with a trained neural network (NN) model are provided. The method includes generating any number of kernel tensors, input channels and output channels associated with the weight tensor, each kernel tensor being associated with any of: a layer type, an input signal type, and a tree partition type, and each kernel tensor including weight coefficients, generating, for each of the any number of kernel tensors, tree partitions for any of a coding tree unit (CTU), a coding unit (CU), a prediction unit (PU), and a transform unit (TU) according to respective tree partition types associated with each of the any number of kernel tensors, and generating a compressed representation of the trained NN model by compressing and coding the any number of kernel tensors.Type: ApplicationFiled: July 2, 2020Publication date: November 10, 2022Inventors: Hua Yang, Duanshun Li, Dong Tian, Yuwen He
-
Publication number: 20220345701Abstract: Systems, methods, and instrumentalities may be used for decoding and/or encoding a coding unit (CD), An intra-prediction mode for a CD may be determined. A split mode may be determined based on the intra-prediction mode, to generate a plurality of sub-partitions in the CU. A prediction for a first sub-partition of the plurality of sub-partitions in the CU may be based on a reference sample in a second sub-partition of the plurality of sub-partitions in the CU. The CU may be decoded and/or encoded, for example, based on the determined split mode.Type: ApplicationFiled: September 17, 2020Publication date: October 27, 2022Applicant: VID SCALE, INC.Inventors: Hua Yang, Yuwen He, Wei Chen
-
Patent number: 11477437Abstract: Implementations of the present disclosure provide a solution for encoding and decoding motion information. In this solution, during a conversion between a current video block of a video and a bitstream of the video, a group type of a subgroup of candidates of motion information for the current video block is determined, wherein the group type indicating whether the subgroup of candidates are to be reordered. Further, a list of candidates are constructed based on the group type; and the motion information for the current video block is derived from the list of candidates.Type: GrantFiled: January 28, 2021Date of Patent: October 18, 2022Assignee: Lemon Inc.Inventors: Na Zhang, Kai Zhang, Li Zhang, Hongbin Liu, Yuwen He, Yue Wang
-
Publication number: 20220329826Abstract: Example implementations include a method, apparatus and computer-readable medium of video processing, including constructing, during a conversion between a current video block of a video and a bitstream of the video, at least one template set for the current video block from a plurality of sub-templates. The one or more sub-templates may be selected from a plurality of sub-templates including: a left sub-template, an above sub-template, a right-above sub-template, a left-below sub-template, and a left-above sub-template. The implementations further include deriving at least one intra-prediction mode (IPM) based on cost calculations. The implementations include determining, based on the at least one IPM, a final predictor of the current video block. The implementations include performing the conversion based on the final predictor.Type: ApplicationFiled: June 10, 2022Publication date: October 13, 2022Inventors: Yang WANG, Kai ZHANG, Li ZHANG, Yuwen HE, Hongbin LIU
-
Publication number: 20220329816Abstract: A method of video processing includes performing a conversion between a current video block of a second color component of a video and a bitstream of the video using a cross-component prediction with multiple-parameter model (CCPMPM) in which samples of the current video block are predictively coded in the bitstream using a linear combination of samples of a first color component multiplied by linear coefficients and/or one or more offsets. The bitstream conforms to a format rule. The linear coefficients of the CCPMPM are determined using a first rule. The samples of the first color component are determined using a second rule.Type: ApplicationFiled: June 24, 2022Publication date: October 13, 2022Inventors: Hongbin LIU, Li ZHANG, Kai ZHANG, Jizheng XU, Yuwen HE, Yue WANG
-
Publication number: 20220329831Abstract: Cross plane filtering may be used for enhanced chroma coding. An indication of a cross-plane filter associated with a current picture may be received. The current picture may include an intra-coded video block and a plurality of reference samples. The plurality of reference samples may be used to predict the intra-coded video block. A luma sample region may be determined in the current picture. The luma sample region may determined to enhance a corresponding chroma sample in the current picture. The cross-plane filter may be applied to a plurality of luma samples in the luma sample region to determine an offset. The cross-plane filter may be a high pass filter. The offset may be applied to the corresponding chroma sample to determine an enhanced chroma sample.Type: ApplicationFiled: June 23, 2022Publication date: October 13, 2022Applicant: InterDigital Madison Patent Holdings, SASInventors: Yong He, Yuwen He, Yan Ye, Xiaoyu Xiu
-
Patent number: 11470308Abstract: Systems and methods are described for reducing the complexity of using bi-directional optical flow (BIO) in video coding. In some embodiments, bit-width reduction steps are introduced in the BIO motion refinement process to reduce the maximum bit-width used for BIO calculations. In some embodiments, simplified interpolation filters are used to generate predicted samples in an extended region around a current coding unit. In some embodiments, different interpolation filters are used for vertical versus horizontal interpolation. In some embodiments, BIO is disabled for coding units with small heights and/or for coding units that are predicted using a sub-block level inter prediction technique, such as advanced temporal motion vector prediction (ATMVP) or affine prediction.Type: GrantFiled: September 17, 2019Date of Patent: October 11, 2022Assignee: VID SCALE, INC.Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye, Jiancong Luo
-
Patent number: 11470344Abstract: Systems and methods are described for selecting a motion vector (MV) to use in frame-rate up conversion (FRUC) coding of a block of video. In one embodiment, a first set of motion vector candidates is identified for FRUC prediction of the block. A search center is defined based on the first set of motion vector candidates, and a search window is determined, the search window having a selected width and being centered on the search center. A search for a selected MV is performed within the search window. In some embodiments, an initial set of MVs is processed with a clustering algorithm to generate a smaller number of MVs that are used as the first set. The selected MV may be subject to a motion refinement search, which may also be performed over a constrained search range. In additional embodiments, search iterations are constrained to limit complexity.Type: GrantFiled: January 28, 2019Date of Patent: October 11, 2022Assignee: VID SCALE, INC.Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
-
Publication number: 20220312001Abstract: Systems and methods are described for video coding using generalized bi-prediction. In an exemplary embodiment, to code a current block of a video in a bitstream, a first reference block is selected from a first reference picture and a second reference block is selected from a second reference picture. Each reference block is associated with a weight, where the weight may be an arbitrary weight ranging, e.g., between 0 and 1. The current block is predicted using a weighted sum of the reference blocks. The weights may be selected from among a plurality of candidate weights. Candidate weights may be signaled in the bitstream or may be derived implicitly based on a template. Candidate weights may be pruned to avoid out-of-range or substantially duplicate candidate weights. Generalized bi-prediction may additionally be used in frame rate up conversion.Type: ApplicationFiled: June 13, 2022Publication date: September 29, 2022Inventors: Chun-Chi Chen, Xiaoyu Xiu, Yuwen He, Yan Ye
-
Patent number: 11457198Abstract: A video coding device may be configured to periodically select the frame packing configuration (e.g., face layout and/or face rotations parameters) associated with a RAS, The device may receive a plurality of pictures, which may each comprise a plurality of faces. The pictures may be grouped Into a plurality of RASs. The device may select a frame packing configuration with the lowest cost for a first RAS. For example, the cost of a frame packing configuration may be determined based on the first picture of the first RAS. The device may select a frame packing configuration for a second RAS. The frame packing configuration for the first RAS may be different than the frame packing configuration for the second RAS. The frame packing configuration for the first RAS and the frame packing configuration for the second RAS may be signaled in the video bitstream.Type: GrantFiled: January 14, 2019Date of Patent: September 27, 2022Assignee: VID SCALE, Inc.Inventors: Philippe Hanhart, Yuwen He, Yan Ye