Patents by Inventor Yuwen He

Yuwen He has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20210274203
    Abstract: Systems, methods, and devices are disclosed for performing adaptive residue color space conversion. A video bitstream may be received and a first flag may be determined based on the video bitstream. A residual may also be generated based on the video bitstream. The residual may be converted from a first color space to a second color space in response to the first flag.
    Type: Application
    Filed: March 24, 2021
    Publication date: September 2, 2021
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Chia-Ming Tsai, Yan Ye
  • Publication number: 20210274213
    Abstract: Embodiments of video coding systems and methods are described for reducing coding latency introduced by decoder-side motion vector refinement (DMVR). In one example, two non-refined motion vectors are identified for coding of a first block of samples (e.g. a first coding unit) using bi-prediction. One or both of the non-refined motion vectors are used to predict motion information for a second block of samples (e.g. a second coding unit). The two non-refined motion vectors are refined using DMVR, and the refined motion vectors are used to generate a prediction signal of the first block of samples. Such embodiments allow the second block of samples to be coded substantially in parallel with the first block without waiting for completion of DMVR on the first block. In additional embodiments, optical-flow-based techniques are described for motion vector refinement.
    Type: Application
    Filed: June 20, 2019
    Publication date: September 2, 2021
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20210274209
    Abstract: Systems, methods, and instrumentalities are disclosed for motion vector clipping when affine motion mode is enabled for a video block. A video coding device may determine that an affine mode for a video block is enabled. The video coding device may determine a plurality of control point affine motion vectors associated with the video block. The video coding device may store the plurality of clipped control point affine motion vectors for motion vector prediction of a neighboring control point affine motion vector. The video coding device may derive a sub-block motion vector associated with a sub-block of the video block, clip the derived sub-block motion vector, and store it for spatial motion vector prediction or temporal motion vector prediction. For example, the video coding device may clip the derived sub-block motion vector based on a motion field range that may be based on a bit depth value.
    Type: Application
    Filed: June 27, 2019
    Publication date: September 2, 2021
    Applicant: Vid Scale, Inc.
    Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye, Philippe Hanhart
  • Patent number: 11109044
    Abstract: Systems, methods, and instrumentalities are disclosed for color space conversion. A video signal associated with a first color space may be received. The first color space may comprise a point. A partition of the first color space that includes the point may be determined. The partition may be associated with a first dimensionality. The point may be converted from the first color space to a second color space using a color space conversion model that corresponds with the partition. The color space conversion model may be associated with a second dimensionality. The second dimensionality may be less than the first dimensionality.
    Type: Grant
    Filed: December 29, 2014
    Date of Patent: August 31, 2021
    Assignee: InterDigital Madison Patent Holdings, SAS
    Inventors: Yan Ye, Jie Dong, Yuwen He
  • Publication number: 20210250592
    Abstract: A video block of a current picture may be coded in an intra block copy (IBC) mode. Weighted prediction may be disabled for the IBC-coded screen content video block. Fractional block vectors may be used for the chroma components of the IBC-coded video block. An interpolation filter may be utilized to generate chroma prediction samples for the video block. A decoded version of the current reference picture may be added to both reference picture list L0 and reference picture list L1 that are associated with the IBC-coded video block. When constrained intra prediction is applied, reference samples that may be used to predict an intra-coded video block may be limited to those in intra-coded neighboring blocks. The range of IBC searches may be restricted by imposing a maximum absolute value for block vectors.
    Type: Application
    Filed: April 26, 2021
    Publication date: August 12, 2021
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yan Ye, Yuwen He
  • Publication number: 20210250619
    Abstract: Systems, methods, and instrumentalities are provided to implement video coding system (VCS). The VCS may be configured to receive a video signal, which may include one or more layers (e.g., a base layer (BL) and/or one or more enhancement layers (ELs)). The VCS may be configured to process a BL picture into an inter-layer reference (ILR) picture, e.g., using picture level inter-layer prediction process. The VCS may be configured to select one or both of the processed ILR picture or an enhancement layer (EL) reference picture. The selected reference picture(s) may comprise one of the EL reference picture, or the ILR picture. The VCS may be configured to predict a current EL picture using one or more of the selected ILR picture or the EL reference picture. The VCS may be configured to store the processed ILR picture in an EL decoded picture buffer (DPB).
    Type: Application
    Filed: April 30, 2021
    Publication date: August 12, 2021
    Applicant: Vid Scale, Inc.
    Inventors: Yan Ye, George W. McClellan, Yong He, Xiaoyu Xiu, Yuwen He, Jie Dong, Can Bal, Eun Seok Ryu
  • Publication number: 20210203947
    Abstract: Systems and methods are described for video coding using affine motion models with adaptive precision. In an example, a block of video is encoded in a bitstream using an affine motion model, where the affine motion model is characterized by at least two motion vectors. A precision is selected for each of the motion vectors, and the selected precisions are signaled in the bitstream. In some embodiments, the precisions are signaled by including in the bitstream information that identifies one of a plurality of elements in a selected predetermined precision set. The identified element indicates the precision of each of the motion vectors that characterize the affine motion model. In some embodiments, the precision set to be used is signaled expressly in the bitstream; in other embodiments, the precision set may be inferred, e.g., from the block size, block shape or temporal layer.
    Type: Application
    Filed: August 28, 2019
    Publication date: July 1, 2021
    Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye, Jiancong Luo
  • Publication number: 20210185353
    Abstract: External overlapped block motion compensation (OBMC) may be performed for samples of a coding unit (CU) located along an inter-CU boundary of the CU while internal OBMC may be performed separately for samples located along inter-sub-block boundaries inside the CU. External OBMC may be applied based on substantially similar motion information associated with multiple external blocks neighboring the CU. The external blocks may be treated as a group to provide OBMC for multiple boundary samples together in an external OBMC operation. Internal OBMC may be applied using the same sub-block size used for sub-block level motion derivation. Internal OBMC may be disabled for the CU, for example, if the CU is coded in a spatial-temporal motion vector prediction (STMVP) mode.
    Type: Application
    Filed: November 1, 2018
    Publication date: June 17, 2021
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye, Yan Zhang
  • Publication number: 20210185338
    Abstract: Systems, methods, and instrumentalities for sub-block motion derivation and motion vector refinement for merge mode may be disclosed herein. Video data may be coded (e.g., encoded and/or decoded). A collocated picture for a current slice of the video data may be identified. The current slice may include one or more coding units (CUs). One or more neighboring CUs may be identified for a current CU. A neighboring CU (e.g., each neighboring CU) may correspond to a reference picture. A (e.g., one) neighboring CU may be selected to be a candidate neighboring CU based on the reference pictures and the collocated picture. A motion vector (MV) (e.g., collocated MV) may be identified from the collocated picture based on an MV (e.g., a reference MV) of the candidate neighboring CU. The current CU may be coded (e.g., encoded and/or decoded) using the collocated MV.
    Type: Application
    Filed: November 1, 2018
    Publication date: June 17, 2021
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 11036459
    Abstract: Video data, e.g., screen content video data may be palette coded. A palette table including one or more color indices may be produced. A color index may correspond to one color. A palette index map may be created that maps one or more pixels of the video data to a color index in the palette table, or a color that may be explicitly coded. A palette index map prediction data may be generated that includes data that indicates values in the palette index map associated with at least some portions of the video data that are generated in a traverse scan order in which a scan line is scanned in an opposite direction of a preceding parallel scan line.
    Type: Grant
    Filed: June 29, 2018
    Date of Patent: June 15, 2021
    Assignee: VID Scale, Inc.
    Inventors: Chia-Ming Tsai, Xiaoyu Xiu, Yan Ye, Yuwen He
  • Patent number: 11039151
    Abstract: Methods and systems are disclosed for a mobile device to decode video based on available power and/or energy. For example, the mobile device may receive a media description file (MDF) from for a video stream from a video server. The MDF may include complexity information associated with a plurality of video segments. The complexity information may be related to the amount of processing power to be utilized for decoding the segment at the mobile device. The mobile device may determine at least one power metric for the mobile device. The mobile device may determine a first complexity level to be requested for a first video segment based on the complexity information from the MDF and the power metric. The mobile device may dynamically alter the decoding process to save energy based on the detected power/energy level.
    Type: Grant
    Filed: November 27, 2019
    Date of Patent: June 15, 2021
    Assignee: VID SCALE, Inc.
    Inventors: Yuwen He, Yan Ye, Yong He, George W. McClellan, Eun Seok Ryu
  • Publication number: 20210176497
    Abstract: A system, method, and/or instrumentality may be provided for coding a 360-degree video. A picture of the 360-degree video may be received. The picture may include one or more faces associated with one or more projection formats. A first projection format indication may be received that indicates a first projection format may be associated with a first face. A second projection format indication may be received that indicates a second projection format may be associated with a second face. Based on the first projection format, a first transform function associated with the first face may be determined. Based on the second projection format, a second transform function associated with the second face may be determined. At least one decoding process may be performed on the first face using the first transform function and/or at least one decoding process may be performed on the second face using the second transform function.
    Type: Application
    Filed: December 23, 2020
    Publication date: June 10, 2021
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 11012717
    Abstract: Systems, methods, and instrumentalities are provided to implement video coding system (VCS). The VCS may be configured to receive a video signal, which may include one or more layers (e.g., a base layer (BL) and/or one or more enhancement layers (ELs)). The VCS may be configured to process a BL picture into an inter-layer reference (ILR) picture, e.g., using picture level inter-layer prediction process. The VCS may be configured to select one or both of the processed ILR picture or an enhancement layer (EL) reference picture. The selected reference picture(s) may comprise one of the EL reference picture, or the ILR picture. The VCS may be configured to predict a current EL picture using one or more of the selected ILR picture or the EL reference picture. The VCS may be configured to store the processed ILR picture in an EL decoded picture buffer (DPB).
    Type: Grant
    Filed: October 24, 2019
    Date of Patent: May 18, 2021
    Assignee: VID SCALE, Inc.
    Inventors: Yan Ye, George W McClellan, Yong He, Xiaoyu Xiu, Yuwen He, Jie Dong, Can Bal, Eun Seok Ryu
  • Patent number: 11006130
    Abstract: A video block of a current picture may be coded in an intra block copy (IBC) mode. Weighted prediction may be disabled for the IBC-coded screen content video block. Fractional block vectors may be used for the chroma components of the IBC-coded video block. An interpolation filter may be utilized to generate chroma prediction samples for the video block. A decoded version of the current reference picture may be added to both reference picture list L0 and reference picture list L1 that are associated with the IBC-coded video block. When constrained intra prediction is applied, reference samples that may be used to predict an intra-coded video block may be limited to those in intra-coded neighboring blocks. The range of IBC searches may be restricted by imposing a maximum absolute value for block vectors.
    Type: Grant
    Filed: April 8, 2019
    Date of Patent: May 11, 2021
    Assignee: VID SCALE, Inc.
    Inventors: Xiaoyu Xiu, Yan Ye, Yuwen He
  • Publication number: 20210136413
    Abstract: Processing video data may include capturing the video data with multiple cameras and stitching the video data together to obtain a 360-degree video. A frame-packed picture may be provided based on the captured and stitched video data. A current sample location may be identified in the frame-packed picture. Whether a neighboring sample location is located outside of a content boundary of the frame-packed picture may be determined. When the neighboring sample location is located outside of the content boundary, a padding sample location may be derived based on at least one circular characteristic of the 360-degree video content and the projection geometry. The 360-degree video content may be processed based on the padding sample location.
    Type: Application
    Filed: December 4, 2020
    Publication date: May 6, 2021
    Applicant: VID SCALE, INC.
    Inventors: Yuwen He, Yan Ye, Philippe Hanhart, Xiaoyu Xiu
  • Publication number: 20210120257
    Abstract: Inter-layer motion mapping information may be used to enable temporal motion vector prediction (TMVP) of an enhancement layer of a bitstream. For example, a reference picture and a motion vector (MV) of an inter-layer video block may be determined. The reference picture may be determined based on a collocated base layer video block. For example, the reference picture may be a collocated inter-layer reference picture of the reference picture of the collocated base layer video block. The MV may be determined based on a MV of the collocated base layer video block. For example, the MV may be determined by determining the MV of the collocated base layer video block and scaling the MV of the collocated base layer video block according to a spatial ratio between the base layer and the enhancement layer. TMVP may be performed on the enhancement layer picture using the MV of the inter-layer video block.
    Type: Application
    Filed: December 23, 2020
    Publication date: April 22, 2021
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yan Ye, Yong He, Yuwen He
  • Patent number: 10986370
    Abstract: A video coding system may perform inter-layer processing by simultaneously performing inverse tone mapping and color gamut conversion scalability processes on a base layer of a video signal. The video coding system may then perform upsampling on the processed base layer. The processed base layer may be used to code an enhancement layer. Bit depth may be considered for color gamut conversion modules. Luma and/or chroma bit depths may be aligned with respective larger or smaller bit depth values of luma and/or chroma.
    Type: Grant
    Filed: July 26, 2018
    Date of Patent: April 20, 2021
    Assignee: VID SCALE, Inc.
    Inventors: Yan Ye, Jie Dong, Yuwen He, Ralph Neff
  • Patent number: 10972728
    Abstract: Chroma enhancement filtering may be used in high dynamic range (HDR) video. A video coding device may identify a characteristic within a picture of a video signal. The characteristic may include a spatial region, a color, a luminance, or an edge direction. The video coding device may determine a sample set that includes one or more samples associated with the characteristic. The video coding device may generate a cross-plane filter based on the characteristic and the sample set. The cross-plane filter may include a high pass filter. The video coding device may apply the cross-plane filter to a luma plane component of a sample in the sample set to determine an offset. The video coding device may add the offset to a reconstructed chroma plane component of the sample that corresponds to the luma plane component.
    Type: Grant
    Filed: April 15, 2016
    Date of Patent: April 6, 2021
    Assignee: InterDigital Madison Patent Holdings, SAS
    Inventors: Yuwen He, Yan Ye
  • Patent number: 10972731
    Abstract: Systems and methods are provided for video coding (e.g., encoding and/or decoding). A coding device may include a processor. The processor may be configured to receive a video frame and partition the video frame into a super coding units (SCUs). The processor may be configured to partition each of the SCUs into coding tree units (CTUs) based on a coding mode.
    Type: Grant
    Filed: November 10, 2016
    Date of Patent: April 6, 2021
    Assignee: InterDigital Madison Patent Holdings, SAS
    Inventors: Miaohui Wang, Yuwen He, Yan Ye, Xiaoyu Xiu
  • Patent number: 10939130
    Abstract: Inter-layer motion mapping information may be used to enable temporal motion vector prediction (TMVP) of an enhancement layer of a bitstream. For example, a reference picture and a motion vector (MV) of an inter-layer video block may be determined. The reference picture may be determined based on a collocated base layer video block. For example, the reference picture may be a collocated inter-layer reference picture of the reference picture of the collocated base layer video block. The MV may be determined based on a MV of the collocated base layer video block. For example, the MV may be determined by determining the MV of the collocated base layer video block and scaling the MV of the collocated base layer video block according to a spatial ratio between the base layer and the enhancement layer. TMVP may be performed on the enhancement layer picture using the MV of the inter-layer video block.
    Type: Grant
    Filed: January 9, 2018
    Date of Patent: March 2, 2021
    Assignee: VID SCALE, Inc.
    Inventors: Xiaoyu Xiu, Yan Ye, Yong He, Yuwen He