Patents by Inventor Yuwen He

Yuwen He has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230291927
    Abstract: Method, apparatus and systems are disclosed. In one embodiment, a method of decoding includes obtaining a sub-block based motion prediction signal for a current block of the video; obtaining one or more spatial gradients of the sub-block based motion prediction signal or one or more motion vector difference values; obtaining a refinement signal for the current block based on the one or more obtained spatial gradients or the one or more obtained motion vector difference values; obtaining a refined motion prediction signal for the current block based on the sub-block based motion prediction signal and the refinement signal; and decoding the current block based on the refined motion prediction signal.
    Type: Application
    Filed: May 18, 2023
    Publication date: September 14, 2023
    Inventors: Jiancong LUO, Yuwen HE
  • Patent number: 11722677
    Abstract: Systems and methods related to video encoding and decoding using decoder-side intra mode derivation (DIMD) are described. In an exemplary method of coding samples in a block in a video, an intra coding mode is selected based on a plurality of reconstructed samples in a template region adjacent to the block, and the samples in the block are predicted with intra prediction using the selected intra coding mode. The intra coding mode may be selected by testing a plurality of candidate intra coding modes for cost (e.g. distortion) of predicting the template region from a set of reconstructed reference samples. The mode with the lowest cost is used for prediction. In exemplary embodiments, explicit signaling of the intra mode is not required.
    Type: Grant
    Filed: May 5, 2017
    Date of Patent: August 8, 2023
    Assignee: InterDigital Madison Patent Holdings, SAS
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 11695991
    Abstract: Power aware adaptation for a power aware video streaming system may be based on the complexity information conveyed in different ways. A complexity level of a data stream, such as a video data stream, may be selected as a function of a remaining battery power of a wireless transmit/receive unit (WTRU) and on a state set of a plurality of state sets that may be stored and/or managed by the WTRU. These state sets may correspond to, for example, different content sources and/or different complexity estimation algorithms and may be used to select the complexity level of the data stream. The data stream may then be received at the selected complexity level. The complexity level and/or a bitrate of the data stream may be adapted to accommodate, for example, the remaining battery power and/or other circumstances. The adaptation may be customized according to the objectives of use cases.
    Type: Grant
    Filed: October 18, 2021
    Date of Patent: July 4, 2023
    Assignee: InterDigital Patent Holdings, Inc.
    Inventors: Yuwen He, Markus Kunstner, Yan Ye, Ralph Neff
  • Patent number: 11695950
    Abstract: Method, apparatus and systems are disclosed. In one embodiment, a method of decoding includes obtaining a sub-block based motion prediction signal for a current block of the video; obtaining one or more spatial gradients of the sub-block based motion prediction signal or one or more motion vector difference values; obtaining a refinement signal for the current block based on the one or more obtained spatial gradients or the one or more obtained motion vector difference values; obtaining a refined motion prediction signal for the current block based on the sub-block based motion prediction signal and the refinement signal; and decoding the current block based on the refined motion prediction signal.
    Type: Grant
    Filed: February 4, 2020
    Date of Patent: July 4, 2023
    Assignee: VID SCALE, INC.
    Inventors: Jiancong Luo, Yuwen He
  • Publication number: 20230199219
    Abstract: A system, method, and/or instrumentality may be provided for coding a 360-degree video. A picture of the 360-degree video may be received. The picture may include one or more faces associated with one or more projection formats. A first projection format indication may be received that indicates a first projection format may be associated with a first face. A second projection format indication may be received that indicates a second projection format may be associated with a second face. Based on the first projection format, a first transform function associated with the first face may be determined. Based on the second projection format, a second transform function associated with the second face may be determined. At least one decoding process may be performed on the first face using the first transform function and/or at least one decoding process may be performed on the second face using the second transform function.
    Type: Application
    Filed: February 21, 2023
    Publication date: June 22, 2023
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20230188748
    Abstract: A video coding device may be configured to perform directional Bi-directional optical flow (BDOF) refinement on a coding unit (CU). The device may determine the direction in which to perform directional BDOF refinement. The device may calculate the vertical direction gradient difference and the horizontal direction gradient difference for the CU. The vertical direction gradient difference may indicate the difference between the vertical gradients for a first reference picture and the vertical gradients for a second reference picture. The horizontal direction gradient difference may indicate the difference between the horizontal gradients for the first reference picture and the horizontal gradients for the second reference picture. The video coding device may determine the direction in which to perform directional BDOF refinement based on the vertical direction gradient difference and the horizontal direction gradient difference.
    Type: Application
    Filed: February 6, 2023
    Publication date: June 15, 2023
    Applicant: VID SCALE, INC.
    Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye
  • Publication number: 20230188752
    Abstract: Systems, methods, and instrumentalities are disclosed for performing horizontal geometry padding on a current sample based on receiving a wraparound enabled indication that indicates whether a horizontal wraparound motion compensation is enabled. If the horizontal wraparound motion compensation is enabled based on the wraparound enabled indication, a video coding device may determine a reference sample wraparound offset of a current sample in a picture. The reference sample wraparound offset may indicate a face width of the picture. The video coding device may determine a reference sample location for the current sample based on the reference sample wraparound offset, a picture width of the picture, and a current sample location. The video coding device may predict the current sample based on the reference sample location in a horizontal direction. Repetitive padding or clipping may be used in the vertical direction.
    Type: Application
    Filed: February 3, 2023
    Publication date: June 15, 2023
    Applicant: Vid Scale, Inc.
    Inventors: Philippe Hanhart, Yuwen He, Yan Ye
  • Publication number: 20230179795
    Abstract: Systems, methods, and instrumentalities may be provided for determining whether to bypass bi-directional optical flow (BDOF) if BDOF is used in combination with bi-prediction with coding unit (CU) weights (e.g., generalized bi-prediction (GBi)). A coding system may combine coding modes, coding techniques, and/or coding tools. The coding system may include a wireless transmit/receive unit (WTRU). For example, the coding system may combine BDOF and bi-prediction with CU weights (BCW). BDOF may include refining a motion vector associated with a current CU based at least in part on gradients associated with a location in the current CU. The coding system may determine that BDOF is enabled, and/or that bi-prediction with CU weights is enabled for the current CU. The coding system’s determination that bi-prediction with CU weights is enabled and/or that BDOF is enabled may be based on one or more indications.
    Type: Application
    Filed: January 9, 2023
    Publication date: June 8, 2023
    Applicant: VID SCALE, INC.
    Inventors: Yuwen He, Jiancong Luo, Xiaoyu Xiu, Yan Ye
  • Publication number: 20230156215
    Abstract: Systems, methods, and instrumentalities for sub-block motion derivation and motion vector refinement for merge mode may be disclosed herein. Video data may be coded (e.g., encoded and/or decoded). A collocated picture for a current slice of the video data may be identified. The current slice may include one or more coding units (CUs). One or more neighboring CUs may be identified for a current CU. A neighboring CU (e.g., each neighboring CU) may correspond to a reference picture. A (e.g., one) neighboring CU may be selected to be a candidate neighboring CU based on the reference pictures and the collocated picture. A motion vector (MV) (e.g., collocated MV) may be identified from the collocated picture based on an MV (e.g., a reference MV) of the candidate neighboring CU. The current CU may be coded (e.g., encoded and/or decoded) using the collocated MV.
    Type: Application
    Filed: January 16, 2023
    Publication date: May 18, 2023
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20230136599
    Abstract: Systems and methods are described for video coding using affine motion models with adaptive precision. In an example, a block of video is encoded in a bitstream using an affine motion model, where the affine motion model is characterized by at least two motion vectors. A precision is selected for each of the motion vectors, and the selected precisions are signaled in the bitstream. In some embodiments, the precisions are signaled by including in the bitstream information that identifies one of a plurality of elements in a selected predetermined precision set. The identified element indicates the precision of each of the motion vectors that characterize the affine motion model. In some embodiments, the precision set to be used is signaled expressly in the bitstream; in other embodiments, the precision set may be inferred, e.g., from the block size, block shape or temporal layer.
    Type: Application
    Filed: December 27, 2022
    Publication date: May 4, 2023
    Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye, Jiancong Luo
  • Patent number: 11641488
    Abstract: Systems, methods and instrumentalities are disclosed for adaptively selecting an adaptive loop filter (ALF) procedure for a frame based on which temporal layer the frame is in. ALF procedures may vary in computational complexity. One or more frames including the current frame may be in a temporal layer of a coding scheme. The decoder may determine the current frame's temporal layer level within the coding scheme. The decoder may select an ALF procedure based on the current frame's temporal layer level. If the current frame's temporal layer level is higher within the coding scheme than some other temporal layer levels, an ALF procedure that is less computationally complex may be selected for the current frame. Then the decoder may perform the selected ALF procedure on the current frame.
    Type: Grant
    Filed: October 31, 2018
    Date of Patent: May 2, 2023
    Assignee: VID SCALE, Inc.
    Inventors: Rahul Vanam, Yuwen He, Yan Ye
  • Publication number: 20230132221
    Abstract: Sketch copy mode may be used to code blocks comprising irregular lines, syntax redundancy may be removed from blocks with special characteristics, and/or run value coding may be simplified. The parsing dependencies in palette coding design may be removed. For example, the context modeling dependency of the syntax element palette_transpose_flag be removed, for example, by simplifying the corresponding context model. The context modeling of the syntax element palette mode may be removed, for example, by using run-length coding without using context. The syntax parsing dependencies and/or the syntax signaling dependencies that are related with escape color signaling may be removed. A palette table generation process may handle input screen content video with high bit depths, for example, at the encoder side.
    Type: Application
    Filed: December 22, 2022
    Publication date: April 27, 2023
    Applicant: VID SCALE, INC.
    Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye, Ralph Neff
  • Publication number: 20230125740
    Abstract: Bi-directional optical flow (BDOF) may be bypassed, for a current coding block, based on whether symmetric motion vector difference (SMVD) is used in motion vector coding for the current coding block. A coding device (e.g., an encoder or a decoder) may determine whether to bypass BDOF for the current coding block based at least in part on an SMVD indication for the current coding block. The coding device may obtain the SMVD indication that indicates whether SMVD is used in motion vector coding for the current coding block. If SMVD indication indicates that SMVD is used in the motion vector coding for the current coding block, the coding device may bypass BDOF for the current coding block. The coding device may reconstruct the current coding block without performing BDOF if it determines to bypass BDOF for the current coding block.
    Type: Application
    Filed: December 15, 2022
    Publication date: April 27, 2023
    Applicant: VID SCALE, INC.
    Inventors: Jiancong Luo, Xiaoyu Xiu, Yuwen He, Hua Yang
  • Publication number: 20230120362
    Abstract: Embodiments of video coding systems and methods are described for reducing coding latency introduced by decoder-side motion vector refinement (DMVR). In one example, two non-refined motion vectors are identified for coding of a first block of samples (e.g. a first coding unit) using bi-prediction. One or both of the non-refined motion vectors are used to predict motion information for a second block of samples (e.g. a second coding unit). The two non-refined motion vectors are refined using DMVR, and the refined motion vectors are used to generate a prediction signal of the first block of samples. Such embodiments allow the second block of samples to be coded substantially in parallel with the first block without waiting for completion of DMVR on the first block. In additional embodiments, optical-flow-based techniques are described for motion vector refinement.
    Type: Application
    Filed: December 5, 2022
    Publication date: April 20, 2023
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 11627340
    Abstract: Systems, methods, and instrumentalities are provided to implement video coding system (VCS). The VCS may be configured to receive a video signal, which may include one or more layers (e.g., a base layer (BL) and/or one or more enhancement layers (ELs)). The VCS may be configured to process a BL picture into an inter-layer reference (ILR) picture, e.g., using picture level inter-layer prediction process. The VCS may be configured to select one or both of the processed ILR picture or an enhancement layer (EL) reference picture. The selected reference picture(s) may comprise one of the EL reference picture, or the ILR picture. The VCS may be configured to predict a current EL picture using one or more of the selected ILR picture or the EL reference picture. The VCS may be configured to store the processed ILR picture in an EL decoded picture buffer (DPB).
    Type: Grant
    Filed: April 30, 2021
    Date of Patent: April 11, 2023
    Assignee: VID SCALE, Inc.
    Inventors: Yan Ye, George W. McClellan, Yong He, Xiaoyu Xiu, Yuwen He, Jie Dong, Can Bal, Eun Seok Ryu
  • Publication number: 20230107138
    Abstract: Examples of video encoding methods and apparatus and video decoding methods and apparatus are described. An example method of processing video data includes determining, for a conversion between a video block of a video and a bitstream of the video, a gradient of a prediction vector at a sub-block level for the video block according to a rule, wherein the rule specifies to use a same gradient value is assigned for all samples within a sub-block of the video block; and performing the conversion based on the determining.
    Type: Application
    Filed: December 7, 2022
    Publication date: April 6, 2023
    Inventors: Kai Zhang, Yuwen He, Li Zhang, Hongbin Liu, Tianliang Fu, Junru Li, Siwei Ma
  • Publication number: 20230108208
    Abstract: Overlapped block motion compensation (OBMC) may be performed for a current video block based on motion information associated with the current video block and motion information associated with one or more neighboring blocks of the current video block. Under certain conditions, some or all of these neighboring blocks may be omitted from the OBMC operation of the current block. For instance, a neighboring block may be skipped during the OBMC operation if the current video block and the neighboring block are both uni-directionally or bi-directionally predicted, if the motion vectors associated with the current block and the neighboring block refer to a same reference picture, and if a sum of absolute differences between those motion vectors is smaller than a threshold value. Further, OBMC may be conducted in conjunction with regular motion compensation and may use simplified filters than traditionally allowed.
    Type: Application
    Filed: November 23, 2022
    Publication date: April 6, 2023
    Applicant: VID SCALE, INC.
    Inventors: Yan Zhang, Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 11616981
    Abstract: A system, method, and/or instrumentality may be provided for coding a 360-degree video. A picture of the 360-degree video may be received. The picture may include one or more faces associated with one or more projection formats. A first projection format indication may be received that indicates a first projection format may be associated with a first face. A second projection format indication may be received that indicates a second projection format may be associated with a second face. Based on the first projection format, a first transform function associated with the first face may be determined. Based on the second projection format, a second transform function associated with the second face may be determined. At least one decoding process may be performed on the first face using the first transform function and/or at least one decoding process may be performed on the second face using the second transform function.
    Type: Grant
    Filed: December 23, 2020
    Date of Patent: March 28, 2023
    Assignee: VID Scale, Inc.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20230086192
    Abstract: Methods and systems are disclosed for a mobile device to decode video based on available power and/or energy. For example, the mobile device may receive a media description file (MDF) from for a video stream from a video server. The MDF may include complexity information associated with a plurality of video segments. The complexity information may be related to the amount of processing power to be utilized for decoding the segment at the mobile device. The mobile device may determine at least one power metric for the mobile device. The mobile device may determine a first complexity level to be requested for a first video segment based on the complexity information from the MDF and the power metric. The mobile device may dynamically alter the decoding process to save energy based on the detected power/energy level.
    Type: Application
    Filed: November 28, 2022
    Publication date: March 23, 2023
    Applicant: VID SCALE, Inc.
    Inventors: Yuwen He, Yan Ye, Yong He, George W. McClellan, Eun Seok Ryu
  • Publication number: 20230073342
    Abstract: Systems, methods, and instrumentalities are disclosed for motion vector clipping when affine motion mode is enabled for a video block. A video coding device may determine that an affine mode for a video block is enabled. The video coding device may determine a plurality of control point affine motion vectors associated with the video block. The video coding device may store the plurality of clipped control point affine motion vectors for motion vector prediction of a neighboring control point affine motion vector. The video coding device may derive a sub-block motion vector associated with a sub-block of the video block, clip the derived sub-block motion vector, and store it for spatial motion vector prediction or temporal motion vector prediction. For example, the video coding device may clip the derived sub-block motion vector based on a motion field range that may be based on a bit depth value.
    Type: Application
    Filed: November 11, 2022
    Publication date: March 9, 2023
    Applicant: Vid Scale, Inc.
    Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye, Philippe Hanhart