Patents by Inventor Yuwen He
Yuwen He has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20210274203Abstract: Systems, methods, and devices are disclosed for performing adaptive residue color space conversion. A video bitstream may be received and a first flag may be determined based on the video bitstream. A residual may also be generated based on the video bitstream. The residual may be converted from a first color space to a second color space in response to the first flag.Type: ApplicationFiled: March 24, 2021Publication date: September 2, 2021Applicant: VID SCALE, INC.Inventors: Xiaoyu Xiu, Yuwen He, Chia-Ming Tsai, Yan Ye
-
Publication number: 20210274213Abstract: Embodiments of video coding systems and methods are described for reducing coding latency introduced by decoder-side motion vector refinement (DMVR). In one example, two non-refined motion vectors are identified for coding of a first block of samples (e.g. a first coding unit) using bi-prediction. One or both of the non-refined motion vectors are used to predict motion information for a second block of samples (e.g. a second coding unit). The two non-refined motion vectors are refined using DMVR, and the refined motion vectors are used to generate a prediction signal of the first block of samples. Such embodiments allow the second block of samples to be coded substantially in parallel with the first block without waiting for completion of DMVR on the first block. In additional embodiments, optical-flow-based techniques are described for motion vector refinement.Type: ApplicationFiled: June 20, 2019Publication date: September 2, 2021Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
-
Publication number: 20210274209Abstract: Systems, methods, and instrumentalities are disclosed for motion vector clipping when affine motion mode is enabled for a video block. A video coding device may determine that an affine mode for a video block is enabled. The video coding device may determine a plurality of control point affine motion vectors associated with the video block. The video coding device may store the plurality of clipped control point affine motion vectors for motion vector prediction of a neighboring control point affine motion vector. The video coding device may derive a sub-block motion vector associated with a sub-block of the video block, clip the derived sub-block motion vector, and store it for spatial motion vector prediction or temporal motion vector prediction. For example, the video coding device may clip the derived sub-block motion vector based on a motion field range that may be based on a bit depth value.Type: ApplicationFiled: June 27, 2019Publication date: September 2, 2021Applicant: Vid Scale, Inc.Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye, Philippe Hanhart
-
Patent number: 11109044Abstract: Systems, methods, and instrumentalities are disclosed for color space conversion. A video signal associated with a first color space may be received. The first color space may comprise a point. A partition of the first color space that includes the point may be determined. The partition may be associated with a first dimensionality. The point may be converted from the first color space to a second color space using a color space conversion model that corresponds with the partition. The color space conversion model may be associated with a second dimensionality. The second dimensionality may be less than the first dimensionality.Type: GrantFiled: December 29, 2014Date of Patent: August 31, 2021Assignee: InterDigital Madison Patent Holdings, SASInventors: Yan Ye, Jie Dong, Yuwen He
-
Publication number: 20210250592Abstract: A video block of a current picture may be coded in an intra block copy (IBC) mode. Weighted prediction may be disabled for the IBC-coded screen content video block. Fractional block vectors may be used for the chroma components of the IBC-coded video block. An interpolation filter may be utilized to generate chroma prediction samples for the video block. A decoded version of the current reference picture may be added to both reference picture list L0 and reference picture list L1 that are associated with the IBC-coded video block. When constrained intra prediction is applied, reference samples that may be used to predict an intra-coded video block may be limited to those in intra-coded neighboring blocks. The range of IBC searches may be restricted by imposing a maximum absolute value for block vectors.Type: ApplicationFiled: April 26, 2021Publication date: August 12, 2021Applicant: VID SCALE, INC.Inventors: Xiaoyu Xiu, Yan Ye, Yuwen He
-
Publication number: 20210250619Abstract: Systems, methods, and instrumentalities are provided to implement video coding system (VCS). The VCS may be configured to receive a video signal, which may include one or more layers (e.g., a base layer (BL) and/or one or more enhancement layers (ELs)). The VCS may be configured to process a BL picture into an inter-layer reference (ILR) picture, e.g., using picture level inter-layer prediction process. The VCS may be configured to select one or both of the processed ILR picture or an enhancement layer (EL) reference picture. The selected reference picture(s) may comprise one of the EL reference picture, or the ILR picture. The VCS may be configured to predict a current EL picture using one or more of the selected ILR picture or the EL reference picture. The VCS may be configured to store the processed ILR picture in an EL decoded picture buffer (DPB).Type: ApplicationFiled: April 30, 2021Publication date: August 12, 2021Applicant: Vid Scale, Inc.Inventors: Yan Ye, George W. McClellan, Yong He, Xiaoyu Xiu, Yuwen He, Jie Dong, Can Bal, Eun Seok Ryu
-
Publication number: 20210203947Abstract: Systems and methods are described for video coding using affine motion models with adaptive precision. In an example, a block of video is encoded in a bitstream using an affine motion model, where the affine motion model is characterized by at least two motion vectors. A precision is selected for each of the motion vectors, and the selected precisions are signaled in the bitstream. In some embodiments, the precisions are signaled by including in the bitstream information that identifies one of a plurality of elements in a selected predetermined precision set. The identified element indicates the precision of each of the motion vectors that characterize the affine motion model. In some embodiments, the precision set to be used is signaled expressly in the bitstream; in other embodiments, the precision set may be inferred, e.g., from the block size, block shape or temporal layer.Type: ApplicationFiled: August 28, 2019Publication date: July 1, 2021Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye, Jiancong Luo
-
Publication number: 20210185353Abstract: External overlapped block motion compensation (OBMC) may be performed for samples of a coding unit (CU) located along an inter-CU boundary of the CU while internal OBMC may be performed separately for samples located along inter-sub-block boundaries inside the CU. External OBMC may be applied based on substantially similar motion information associated with multiple external blocks neighboring the CU. The external blocks may be treated as a group to provide OBMC for multiple boundary samples together in an external OBMC operation. Internal OBMC may be applied using the same sub-block size used for sub-block level motion derivation. Internal OBMC may be disabled for the CU, for example, if the CU is coded in a spatial-temporal motion vector prediction (STMVP) mode.Type: ApplicationFiled: November 1, 2018Publication date: June 17, 2021Applicant: VID SCALE, INC.Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye, Yan Zhang
-
Publication number: 20210185338Abstract: Systems, methods, and instrumentalities for sub-block motion derivation and motion vector refinement for merge mode may be disclosed herein. Video data may be coded (e.g., encoded and/or decoded). A collocated picture for a current slice of the video data may be identified. The current slice may include one or more coding units (CUs). One or more neighboring CUs may be identified for a current CU. A neighboring CU (e.g., each neighboring CU) may correspond to a reference picture. A (e.g., one) neighboring CU may be selected to be a candidate neighboring CU based on the reference pictures and the collocated picture. A motion vector (MV) (e.g., collocated MV) may be identified from the collocated picture based on an MV (e.g., a reference MV) of the candidate neighboring CU. The current CU may be coded (e.g., encoded and/or decoded) using the collocated MV.Type: ApplicationFiled: November 1, 2018Publication date: June 17, 2021Applicant: VID SCALE, INC.Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
-
Patent number: 11036459Abstract: Video data, e.g., screen content video data may be palette coded. A palette table including one or more color indices may be produced. A color index may correspond to one color. A palette index map may be created that maps one or more pixels of the video data to a color index in the palette table, or a color that may be explicitly coded. A palette index map prediction data may be generated that includes data that indicates values in the palette index map associated with at least some portions of the video data that are generated in a traverse scan order in which a scan line is scanned in an opposite direction of a preceding parallel scan line.Type: GrantFiled: June 29, 2018Date of Patent: June 15, 2021Assignee: VID Scale, Inc.Inventors: Chia-Ming Tsai, Xiaoyu Xiu, Yan Ye, Yuwen He
-
Patent number: 11039151Abstract: Methods and systems are disclosed for a mobile device to decode video based on available power and/or energy. For example, the mobile device may receive a media description file (MDF) from for a video stream from a video server. The MDF may include complexity information associated with a plurality of video segments. The complexity information may be related to the amount of processing power to be utilized for decoding the segment at the mobile device. The mobile device may determine at least one power metric for the mobile device. The mobile device may determine a first complexity level to be requested for a first video segment based on the complexity information from the MDF and the power metric. The mobile device may dynamically alter the decoding process to save energy based on the detected power/energy level.Type: GrantFiled: November 27, 2019Date of Patent: June 15, 2021Assignee: VID SCALE, Inc.Inventors: Yuwen He, Yan Ye, Yong He, George W. McClellan, Eun Seok Ryu
-
Publication number: 20210176497Abstract: A system, method, and/or instrumentality may be provided for coding a 360-degree video. A picture of the 360-degree video may be received. The picture may include one or more faces associated with one or more projection formats. A first projection format indication may be received that indicates a first projection format may be associated with a first face. A second projection format indication may be received that indicates a second projection format may be associated with a second face. Based on the first projection format, a first transform function associated with the first face may be determined. Based on the second projection format, a second transform function associated with the second face may be determined. At least one decoding process may be performed on the first face using the first transform function and/or at least one decoding process may be performed on the second face using the second transform function.Type: ApplicationFiled: December 23, 2020Publication date: June 10, 2021Applicant: VID SCALE, INC.Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
-
Patent number: 11012717Abstract: Systems, methods, and instrumentalities are provided to implement video coding system (VCS). The VCS may be configured to receive a video signal, which may include one or more layers (e.g., a base layer (BL) and/or one or more enhancement layers (ELs)). The VCS may be configured to process a BL picture into an inter-layer reference (ILR) picture, e.g., using picture level inter-layer prediction process. The VCS may be configured to select one or both of the processed ILR picture or an enhancement layer (EL) reference picture. The selected reference picture(s) may comprise one of the EL reference picture, or the ILR picture. The VCS may be configured to predict a current EL picture using one or more of the selected ILR picture or the EL reference picture. The VCS may be configured to store the processed ILR picture in an EL decoded picture buffer (DPB).Type: GrantFiled: October 24, 2019Date of Patent: May 18, 2021Assignee: VID SCALE, Inc.Inventors: Yan Ye, George W McClellan, Yong He, Xiaoyu Xiu, Yuwen He, Jie Dong, Can Bal, Eun Seok Ryu
-
Patent number: 11006130Abstract: A video block of a current picture may be coded in an intra block copy (IBC) mode. Weighted prediction may be disabled for the IBC-coded screen content video block. Fractional block vectors may be used for the chroma components of the IBC-coded video block. An interpolation filter may be utilized to generate chroma prediction samples for the video block. A decoded version of the current reference picture may be added to both reference picture list L0 and reference picture list L1 that are associated with the IBC-coded video block. When constrained intra prediction is applied, reference samples that may be used to predict an intra-coded video block may be limited to those in intra-coded neighboring blocks. The range of IBC searches may be restricted by imposing a maximum absolute value for block vectors.Type: GrantFiled: April 8, 2019Date of Patent: May 11, 2021Assignee: VID SCALE, Inc.Inventors: Xiaoyu Xiu, Yan Ye, Yuwen He
-
Publication number: 20210136413Abstract: Processing video data may include capturing the video data with multiple cameras and stitching the video data together to obtain a 360-degree video. A frame-packed picture may be provided based on the captured and stitched video data. A current sample location may be identified in the frame-packed picture. Whether a neighboring sample location is located outside of a content boundary of the frame-packed picture may be determined. When the neighboring sample location is located outside of the content boundary, a padding sample location may be derived based on at least one circular characteristic of the 360-degree video content and the projection geometry. The 360-degree video content may be processed based on the padding sample location.Type: ApplicationFiled: December 4, 2020Publication date: May 6, 2021Applicant: VID SCALE, INC.Inventors: Yuwen He, Yan Ye, Philippe Hanhart, Xiaoyu Xiu
-
Publication number: 20210120257Abstract: Inter-layer motion mapping information may be used to enable temporal motion vector prediction (TMVP) of an enhancement layer of a bitstream. For example, a reference picture and a motion vector (MV) of an inter-layer video block may be determined. The reference picture may be determined based on a collocated base layer video block. For example, the reference picture may be a collocated inter-layer reference picture of the reference picture of the collocated base layer video block. The MV may be determined based on a MV of the collocated base layer video block. For example, the MV may be determined by determining the MV of the collocated base layer video block and scaling the MV of the collocated base layer video block according to a spatial ratio between the base layer and the enhancement layer. TMVP may be performed on the enhancement layer picture using the MV of the inter-layer video block.Type: ApplicationFiled: December 23, 2020Publication date: April 22, 2021Applicant: VID SCALE, INC.Inventors: Xiaoyu Xiu, Yan Ye, Yong He, Yuwen He
-
Patent number: 10986370Abstract: A video coding system may perform inter-layer processing by simultaneously performing inverse tone mapping and color gamut conversion scalability processes on a base layer of a video signal. The video coding system may then perform upsampling on the processed base layer. The processed base layer may be used to code an enhancement layer. Bit depth may be considered for color gamut conversion modules. Luma and/or chroma bit depths may be aligned with respective larger or smaller bit depth values of luma and/or chroma.Type: GrantFiled: July 26, 2018Date of Patent: April 20, 2021Assignee: VID SCALE, Inc.Inventors: Yan Ye, Jie Dong, Yuwen He, Ralph Neff
-
Patent number: 10972728Abstract: Chroma enhancement filtering may be used in high dynamic range (HDR) video. A video coding device may identify a characteristic within a picture of a video signal. The characteristic may include a spatial region, a color, a luminance, or an edge direction. The video coding device may determine a sample set that includes one or more samples associated with the characteristic. The video coding device may generate a cross-plane filter based on the characteristic and the sample set. The cross-plane filter may include a high pass filter. The video coding device may apply the cross-plane filter to a luma plane component of a sample in the sample set to determine an offset. The video coding device may add the offset to a reconstructed chroma plane component of the sample that corresponds to the luma plane component.Type: GrantFiled: April 15, 2016Date of Patent: April 6, 2021Assignee: InterDigital Madison Patent Holdings, SASInventors: Yuwen He, Yan Ye
-
Patent number: 10972731Abstract: Systems and methods are provided for video coding (e.g., encoding and/or decoding). A coding device may include a processor. The processor may be configured to receive a video frame and partition the video frame into a super coding units (SCUs). The processor may be configured to partition each of the SCUs into coding tree units (CTUs) based on a coding mode.Type: GrantFiled: November 10, 2016Date of Patent: April 6, 2021Assignee: InterDigital Madison Patent Holdings, SASInventors: Miaohui Wang, Yuwen He, Yan Ye, Xiaoyu Xiu
-
Patent number: 10939130Abstract: Inter-layer motion mapping information may be used to enable temporal motion vector prediction (TMVP) of an enhancement layer of a bitstream. For example, a reference picture and a motion vector (MV) of an inter-layer video block may be determined. The reference picture may be determined based on a collocated base layer video block. For example, the reference picture may be a collocated inter-layer reference picture of the reference picture of the collocated base layer video block. The MV may be determined based on a MV of the collocated base layer video block. For example, the MV may be determined by determining the MV of the collocated base layer video block and scaling the MV of the collocated base layer video block according to a spatial ratio between the base layer and the enhancement layer. TMVP may be performed on the enhancement layer picture using the MV of the inter-layer video block.Type: GrantFiled: January 9, 2018Date of Patent: March 2, 2021Assignee: VID SCALE, Inc.Inventors: Xiaoyu Xiu, Yan Ye, Yong He, Yuwen He