Patents by Inventor Yan Ye

Yan Ye has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11706439
    Abstract: The present disclosure provides a computer-implemented method for decoding video. The method includes receiving a bitstream including a coding unit of a video frame, the coding unit being coded in a skip mode or a direct mode; determining whether the coding units has a width greater than a threshold width or a height greater than a threshold height; in response to the coding unit having the width greater than the threshold width or the height greater than the threshold height, enabling a motion vector angular prediction process to store one or more motion vector angular prediction candidates in a candidate list storing a plurality of motion candidates associated with the coding unit; and performing an inter prediction to the coding unit according to the candidate list and an index parsed from the bitstream.
    Type: Grant
    Filed: March 16, 2021
    Date of Patent: July 18, 2023
    Assignee: ALIBABA GROUP HOLDING LIMITED
    Inventors: Ru-Ling Liao, Jie Chen, Yan Ye
  • Publication number: 20230217030
    Abstract: A VVC-standard encoder and a VVC-standard decoder are provided, implementing a decoder-side chroma intra prediction mode gradient-based derivation method, which improves coding efficiency of chroma intra prediction, thereby saving on signaling cost. A VVC-standard decoder configures one or more processors of a computing system to derive one of multiple possible chroma intra prediction modes by computing gradients of adjacent luma samples and chroma samples of a current chroma block. With minimal increase in signaling cost, the VVC-standard coding and decoding processes are enhanced to base intra prediction modes for chroma blocks on collocated luma block-adjacent reconstructed luma samples adjacent reconstructed chroma samples, enabling prior computational work done on coding and decoding adjacent blocks to be referenced. In this fashion, coding gains can be achieved in matching texture characteristics of the current chroma block based on a texture gradient including adjacent blocks.
    Type: Application
    Filed: December 29, 2022
    Publication date: July 6, 2023
    Inventors: Xinwei Li, Ru-Ling Liao, Jie Chen, Yan Ye
  • Publication number: 20230217026
    Abstract: The present disclosure provides methods and systems for fusing chroma intra prediction modes. An exemplary method includes: generating a plurality of predicted chroma samples associated with a pixel, by using a plurality of chroma intra prediction modes respectively; and determining a first predicted chroma sample, based on a weighted sum of the plurality of predicted chroma samples.
    Type: Application
    Filed: December 23, 2022
    Publication date: July 6, 2023
    Inventors: Xinwei LI, Ru-Ling LIAO, Jie CHEN, Yan YE
  • Patent number: 11695991
    Abstract: Power aware adaptation for a power aware video streaming system may be based on the complexity information conveyed in different ways. A complexity level of a data stream, such as a video data stream, may be selected as a function of a remaining battery power of a wireless transmit/receive unit (WTRU) and on a state set of a plurality of state sets that may be stored and/or managed by the WTRU. These state sets may correspond to, for example, different content sources and/or different complexity estimation algorithms and may be used to select the complexity level of the data stream. The data stream may then be received at the selected complexity level. The complexity level and/or a bitrate of the data stream may be adapted to accommodate, for example, the remaining battery power and/or other circumstances. The adaptation may be customized according to the objectives of use cases.
    Type: Grant
    Filed: October 18, 2021
    Date of Patent: July 4, 2023
    Assignee: InterDigital Patent Holdings, Inc.
    Inventors: Yuwen He, Markus Kunstner, Yan Ye, Ralph Neff
  • Publication number: 20230209040
    Abstract: Systems and methods described herein provide for monitoring of video streaming events. Techniques disclosed include rendering a first omnidirectional video associated with a first viewpoint. In response to a request to access an omnidirectional video associated with another viewpoint, a time of the request and information identifying the other viewpoint is recorded. Techniques disclosed also include identifying a time of switching into rendering of a second omnidirectional video associated with a second viewpoint, and reporting a viewpoint switching latency, representing a difference between the identified switching time and a recorded request time associated with information identifying the second viewpoint.
    Type: Application
    Filed: February 17, 2023
    Publication date: June 29, 2023
    Inventors: Yong HE, Yan YE, Ahmed HAMZA
  • Publication number: 20230199241
    Abstract: A secondary content such as an advertisement may be inserted based on users’ interests in 360 degree video streaming. Users may have different interests and may watch different areas within a 360 degree video. The information about area(s) of 360 degree scenes that users watch the most may be used to select an ad(s) relevant to their interests. One or more secondary content viewports may be defined within a 360 degree video frame. Secondary content viewport parameter(s) may be tracked. For example, statistics of the user’s head orientation for some time leading to the presentation of the ad(s) may be collected. Secondary content may be determined based on the tracked secondary content viewport parameter(s).
    Type: Application
    Filed: February 6, 2023
    Publication date: June 22, 2023
    Applicant: VID SCALE, INC.
    Inventors: Eduardo Asbun, Yong He, Srinivas Gudumasu, Ahmed Hamza, Yan Ye
  • Publication number: 20230199219
    Abstract: A system, method, and/or instrumentality may be provided for coding a 360-degree video. A picture of the 360-degree video may be received. The picture may include one or more faces associated with one or more projection formats. A first projection format indication may be received that indicates a first projection format may be associated with a first face. A second projection format indication may be received that indicates a second projection format may be associated with a second face. Based on the first projection format, a first transform function associated with the first face may be determined. Based on the second projection format, a second transform function associated with the second face may be determined. At least one decoding process may be performed on the first face using the first transform function and/or at least one decoding process may be performed on the second face using the second transform function.
    Type: Application
    Filed: February 21, 2023
    Publication date: June 22, 2023
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Patent number: 11683519
    Abstract: This disclosure describes filtering techniques applied by an encoder and a decoder during the prediction stage of a video encoding and/or decoding process. The filtering techniques may enhance the accuracy of predictive data used during fractional interpolation, and may improve predictive data of integer blocks of pixels. There are several aspects to this disclosure, including a useful twelve-pixel filter support that may be used for interpolation, techniques that use coefficient symmetry and pixel symmetry to reduce the amount of data needed to be sent between an encoder and a decoder to configure the filter support for interpolation, and techniques for filtering data at integer pixel locations in a manner that is similar to sub-pixel interpolation. Other aspects of this disclosure concern techniques for encoding information in the bitstream to convey the type of filter used, and possibly the filter coefficients used. Predictive coding of filter coefficients is also described.
    Type: Grant
    Filed: October 2, 2019
    Date of Patent: June 20, 2023
    Assignee: QUALCOMM Incorporated
    Inventors: Yan Ye, Marta Karczewicz
  • Publication number: 20230188748
    Abstract: A video coding device may be configured to perform directional Bi-directional optical flow (BDOF) refinement on a coding unit (CU). The device may determine the direction in which to perform directional BDOF refinement. The device may calculate the vertical direction gradient difference and the horizontal direction gradient difference for the CU. The vertical direction gradient difference may indicate the difference between the vertical gradients for a first reference picture and the vertical gradients for a second reference picture. The horizontal direction gradient difference may indicate the difference between the horizontal gradients for the first reference picture and the horizontal gradients for the second reference picture. The video coding device may determine the direction in which to perform directional BDOF refinement based on the vertical direction gradient difference and the horizontal direction gradient difference.
    Type: Application
    Filed: February 6, 2023
    Publication date: June 15, 2023
    Applicant: VID SCALE, INC.
    Inventors: Yuwen He, Xiaoyu Xiu, Yan Ye
  • Publication number: 20230188752
    Abstract: Systems, methods, and instrumentalities are disclosed for performing horizontal geometry padding on a current sample based on receiving a wraparound enabled indication that indicates whether a horizontal wraparound motion compensation is enabled. If the horizontal wraparound motion compensation is enabled based on the wraparound enabled indication, a video coding device may determine a reference sample wraparound offset of a current sample in a picture. The reference sample wraparound offset may indicate a face width of the picture. The video coding device may determine a reference sample location for the current sample based on the reference sample wraparound offset, a picture width of the picture, and a current sample location. The video coding device may predict the current sample based on the reference sample location in a horizontal direction. Repetitive padding or clipping may be used in the vertical direction.
    Type: Application
    Filed: February 3, 2023
    Publication date: June 15, 2023
    Applicant: Vid Scale, Inc.
    Inventors: Philippe Hanhart, Yuwen He, Yan Ye
  • Patent number: 11677993
    Abstract: A streaming device may request and download multi-layer video segments based on a number of factors including the artistic interest associated with the video segments and/or the status of a buffer area managed by the streaming device. The multi-layer segments may be coded using scalable coding techniques or a combination of scalable coding and simulcast coding techniques by which each of the video segments may be coded into one or more representations of different qualities and/or bitrates. When requesting the multi-layer segments, the streaming device may ensure that the fullness of the buffer area falls between a buffer underflow threshold and a backfilling threshold under various network conditions. The streaming device may estimate the available network bandwidth in order to facilitate the scheduling decisions. The streaming device may consider the artistic interest associated with the video segments during scheduling and may give priority to those segments with higher artistic interest.
    Type: Grant
    Filed: June 27, 2022
    Date of Patent: June 13, 2023
    Assignee: InterDigital Madison Patent Holdings, SAS
    Inventors: Yong He, Ahmed Hamza, Srinivas Gudumasu, Yan Ye, Eduardo Asbun
  • Patent number: 11677802
    Abstract: A client device adaptively streams a 360-degree video. A first segment is displayed based on a first viewing direction at a first time, where the first viewing direction is associated with a first viewport. The client requests a first base buffer segment based on the first viewport. The first base buffer segment has a presentation time after the first segment. At a second time, the viewing direction changes to a second viewing direction associated with a second viewport. The client requests, prior to the presentation time, a first viewport buffer segment based on the second viewport, with the same presentation time. The client device displays a second segment at the presentation time, wherein the second segment is either the first viewport buffer segment or the first base buffer segment. The client provides reports on viewport switching latency and on the most-requested segments.
    Type: Grant
    Filed: September 8, 2017
    Date of Patent: June 13, 2023
    Assignee: Vid Scale, Inc.
    Inventors: Yong He, Yan Ye, Srinivas Gudumasu, Eduardo Asbun, Ahmed Hamza, Ali C. Begen
  • Publication number: 20230179795
    Abstract: Systems, methods, and instrumentalities may be provided for determining whether to bypass bi-directional optical flow (BDOF) if BDOF is used in combination with bi-prediction with coding unit (CU) weights (e.g., generalized bi-prediction (GBi)). A coding system may combine coding modes, coding techniques, and/or coding tools. The coding system may include a wireless transmit/receive unit (WTRU). For example, the coding system may combine BDOF and bi-prediction with CU weights (BCW). BDOF may include refining a motion vector associated with a current CU based at least in part on gradients associated with a location in the current CU. The coding system may determine that BDOF is enabled, and/or that bi-prediction with CU weights is enabled for the current CU. The coding system’s determination that bi-prediction with CU weights is enabled and/or that BDOF is enabled may be based on one or more indications.
    Type: Application
    Filed: January 9, 2023
    Publication date: June 8, 2023
    Applicant: VID SCALE, INC.
    Inventors: Yuwen He, Jiancong Luo, Xiaoyu Xiu, Yan Ye
  • Publication number: 20230179832
    Abstract: Systems and methods are disclosed including techniques for rendering a 360-degree media content. Techniques disclosed include receiving a 360-degree media content and associated metadata that include a classification of a first spatial region from the received content. Techniques disclosed further include determining that a detected user movement is associated with a rendering of the first spatial region and determining whether the classification associated with the first spatial region complies with a stored user preference. If the classification violates the user preference, a path for gradually shifting the content rendering from a currently rendered spatial region to a spatial region that complies with the user preference is determined, and the received content is rendered according to the determined path.
    Type: Application
    Filed: February 6, 2023
    Publication date: June 8, 2023
    Inventors: Yong HE, Yan YE, Ali C. BEGEN, Ahmed HAMZA
  • Patent number: 11671613
    Abstract: The present disclosure provides methods for picture processing. The method can include: receiving a bitstream comprising a set of pictures; determining, according to the received bitstream, whether a virtual boundary is signaled at a sequence level for the set of pictures; in response to the virtual boundary being signaled at the sequence level, determining a position of the virtual boundary for the set of pictures, the position being bounded by a range signaled in the received bitstream; and disabling in-loop filtering operations across the virtual boundary.
    Type: Grant
    Filed: December 22, 2020
    Date of Patent: June 6, 2023
    Assignee: Alibaba Group Holding Limited
    Inventors: Jie Chen, Ru-Ling Liao, Yan Ye, Jiancong Luo
  • Publication number: 20230171415
    Abstract: The present disclosure provides a computer-implemented method for encoding or decoding video. The method includes encoding or decoding, in a plurality of picture parameter sets (PPS) associated with pictures of a coded layer video sequence (CLVS), corresponding first PPS flags indicating whether pictures are allowed to be partitioned into a plurality of tiles or slices. In a first PPS, a corresponding first PPS flag with a first value indicates a first picture of the CLVS is unpartitioned, and in a second PPS, another corresponding first PPS flag with a second value being different from the first value indicates that a second picture of the CLVS is allowed to be partitioned.
    Type: Application
    Filed: January 30, 2023
    Publication date: June 1, 2023
    Inventors: Jie CHEN, Yan YE, Ru-Ling LIAO
  • Publication number: 20230156215
    Abstract: Systems, methods, and instrumentalities for sub-block motion derivation and motion vector refinement for merge mode may be disclosed herein. Video data may be coded (e.g., encoded and/or decoded). A collocated picture for a current slice of the video data may be identified. The current slice may include one or more coding units (CUs). One or more neighboring CUs may be identified for a current CU. A neighboring CU (e.g., each neighboring CU) may correspond to a reference picture. A (e.g., one) neighboring CU may be selected to be a candidate neighboring CU based on the reference pictures and the collocated picture. A motion vector (MV) (e.g., collocated MV) may be identified from the collocated picture based on an MV (e.g., a reference MV) of the candidate neighboring CU. The current CU may be coded (e.g., encoded and/or decoded) using the collocated MV.
    Type: Application
    Filed: January 16, 2023
    Publication date: May 18, 2023
    Applicant: VID SCALE, INC.
    Inventors: Xiaoyu Xiu, Yuwen He, Yan Ye
  • Publication number: 20230156211
    Abstract: The present disclosure provides apparatus and methods for signaling sub-block transform (SBT) information. The SBT information is used for coding video data. According to certain disclosed embodiments, an exemplary method includes: signaling a first flag in a Sequence Parameter Set (SPS) of a video sequence indicating whether a sub-block transform (SBT) is enabled; and signaling a second flag indicating a maximum transform block (TB) size that allows the SBT. A maximum coding unit (CU) size that allows the SBT is determined directly based on the maximum TB size in response to the first flag indicating that the SBT is enabled.
    Type: Application
    Filed: January 19, 2023
    Publication date: May 18, 2023
    Inventors: Mohammed Golam SARWER, Jiancong LUO, Yan YE
  • Publication number: 20230156183
    Abstract: A method of encoding a video sequence into a bitstream is provided. The method includes: receiving a video sequence; and encoding one or more pictures of the video sequence; wherein the encoding includes: signaling a first flag in a slice header to indicate whether an active reference index number is present in a slice header; in response to the first flag indicating that the active reference index number is present, determining a number of entries of reference picture list 0, and signaling an active reference index number of reference picture list 0 in the slice header for P and B slices; and determining a number of entries of reference picture list 1, and signaling an active reference index number of reference picture list 1 in the slice header for B slice.
    Type: Application
    Filed: November 4, 2022
    Publication date: May 18, 2023
    Inventors: Jie CHEN, Yan YE, Ru-Ling LIAO
  • Patent number: 11647215
    Abstract: The present disclosure provides methods and apparatuses for controlling a coding mode for video data. The methods and apparatuses include receiving a bitstream of video data; enabling or disabling a coding mode for a video sequence, based on a first flag in the bitstream; and determining whether controlling of the coding mode is enabled or disabled at a level lower than a sequence level, based on a second flag in the bitstream.
    Type: Grant
    Filed: August 20, 2020
    Date of Patent: May 9, 2023
    Assignee: Alibaba Group Holding Limited
    Inventors: Jie Chen, Ru-Ling Liao, Jiancong Luo, Yan Ye