Patents by Inventor Jun Xin

Jun Xin has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 12267506
    Abstract: Techniques are disclosed for deriving prediction pixel blocks for use in intra-coding video and combined inter- and intra-coding video. In a first aspect, the techniques may include deriving value(s) for pixel location(s) of the prediction pixel block by, when a prediction direction vector assigned to the prediction vector points to quadrants I or III of a Cartesian plane, deriving the pixel location's value from pixel values in two regions of previously-decoded pixel data intercepted by extending the prediction direction vector in two opposite directions through the pixel location. When the prediction direction vector points toward quadrants II of the Cartesian plane, deriving the pixel location's value from pixel values in one region intercepted by the prediction direction vector through the pixel location, and from a second region intercepted by a vector that is orthogonal to the prediction direction vector.
    Type: Grant
    Filed: June 10, 2022
    Date of Patent: April 1, 2025
    Assignee: APPLE INC.
    Inventors: Mei Guo, Jun Xin, Hsi-Jung Wu, Alexandros Tourapis, Krishnakanth Rapaka
  • Patent number: 12267496
    Abstract: Video coders and decoders perform transform coding and decoding on blocks of video content according to an adaptively selected transform type. The transform types are organized into a hierarchy of transform sets where each transform set includes a respective number of transforms and each higher-level transform set includes the transforms of each lower-level transform set within the hierarchy. The video coders and video decoders may exchange signaling that establishes a transform set context from which a transform set that was selected for coding given block(s) may be identified. The video coders and video decoders may exchange signaling that establishes a transform decoding context from which a transform that was selected from the identified transform set to be used for decoding the transform unit. The block(s) may be coded and decoded by the selected transform.
    Type: Grant
    Filed: July 25, 2023
    Date of Patent: April 1, 2025
    Assignee: APPLE INC.
    Inventors: Hilmi Enes Egilmez, Yunfei Zheng, Alican Nalci, Yeqing Wu, Yixin Du, Guoxin Jin, Alexandros Tourapis, Jun Xin, Hsi-Jung Wu
  • Publication number: 20250080759
    Abstract: Techniques are disclosed for coding video in applications where regions of video are inactive on a frame to frame basis. According to the techniques, coding processes update and reconstruct only a subset of pixel blocks of pixels within a frame, while other pixel blocks are retained from a previously coded frame stored in a coder's or decoder's reference frame buffer. The technique is called Backward Reference Updating (or “BRU”) for convenience. At a desired pixel block granularity, based on the activity between a current frame to be coded and its reference frame(s), BRU will only perform prediction, transform, quantization, and reconstruction on selected regions that are determined to be active. The reconstructed pixels in these active regions are directly placed onto a specified reference frame in memory instead of creating a new frame. Therefore, fewer memory transfers need to be performed.
    Type: Application
    Filed: August 5, 2024
    Publication date: March 6, 2025
    Inventors: Guoxin JIN, Yeqing WU, Yunfei ZHENG, Dazhong ZHANG, Dzung T. HOANG, Felix C. FERNANDES, Van Luong PHAM, Jiancong LUO, Yixin DU, Hilmi Enes EGILMEZ, Alican NALCI, Jun XIN, Hsi-Jung WU
  • Publication number: 20250080726
    Abstract: Systems and methods are configured for accessing data representing video content, the data comprising a set of one or more symbols each associated with a syntax element; performing a probability estimation, for encoding the data, comprising: for each symbol, obtaining, based on the syntax element for that symbol, an adaptivity rate parameter value, the adaptivity rate parameter value being a function of a number of symbols in the set of one or more symbols; updating the adaptivity rate parameter value as a function of an adjustment parameter value; and generating, based on the updated adaptivity rate parameter value, a probability value; generating a probability estimation; and encoding, based on the CDF of the probability estimation, the data comprising the set of one or more symbols for transmission.
    Type: Application
    Filed: August 29, 2024
    Publication date: March 6, 2025
    Inventors: Hilmi Enes Egilmez, Alexandros Tourapis, Alican Nalci, Guoxin Jin, Hsi-Jung Wu, Jun Xin, Yeqing Wu, Yixin Du, Yunfei Zheng
  • Publication number: 20250071345
    Abstract: Disclosed is a method that includes receiving an image frame having a plurality of coded blocks, determining a prediction unit (PU) from the plurality of coded blocks, determining one or more motion compensation units arranged in an array within the PU, and applying a filter to one or more boundaries of the one or more motion compensation units. Also disclosed is a method that includes receiving a reference frame that includes a reference block, determining a timing for deblocking a current block, performing motion compensation on the reference frame to obtain a predicted frame that includes a predicted block, performing reconstruction on the predicted frame to obtain a reconstructed frame that includes a reconstructed PU, and applying, at the timing for deblocking the current block, a deblocking filter based on one or more parameters to the reference block, the predicted block, or the reconstructed PU.
    Type: Application
    Filed: November 12, 2024
    Publication date: February 27, 2025
    Inventors: Yixin Du, Alexandros Tourapis, Alican Nalci, Guoxin Jin, Hilmi Enes Egilmez, Hsi-Jung Wu, Jun Xin, Yeqing Wu, Yunfei Zheng
  • Patent number: 12192440
    Abstract: Systems and methods disclosed for video compression, utilizing neural networks for predictive video coding. Processes employed combine multiple banks of neural networks with codec system components to carry out the coding and decoding of video data.
    Type: Grant
    Filed: January 4, 2022
    Date of Patent: January 7, 2025
    Assignee: APPLE INC.
    Inventors: Jiefu Zhai, Xingyu Zhang, Xiaosong Zhou, Jun Xin, Hsi-Jung Wu, Yeping Su
  • Publication number: 20240388720
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for reducing a quantity of buffers for a video codec. One of the methods includes determining, from a plurality of prediction modes, a prediction mode for data that represents frame data in a frame; in response to determining the prediction mode, selecting, using the prediction mode, one or more buffers from a plurality of buffers, each buffer of which is for a prediction mode from the plurality of prediction modes, a first quantity of buffers in the plurality of buffers being less than a second quantity of prediction modes in the plurality of prediction modes; retrieving, from each of the one or more buffers, historical data for the frame data; and in response to retrieving the historical data, generating, using the historical data, updated data for the frame data in the frame of the video sequence.
    Type: Application
    Filed: July 25, 2024
    Publication date: November 21, 2024
    Inventors: Yunfei Zheng, Alexandros Tourapis, Alican Nalci, Felix C. Fernandes, Hilmi Enes Egilmez, Hsi-Jung Wu, Jaehong Chon, Jun Xin, Yeqing Wu, Yixin Du
  • Patent number: 12149699
    Abstract: Techniques are disclosed by which a coding parameter is determined to encode video data resulting in encoded video data possessing a highest possible video quality. Features may be extracted from an input video sequence. The extracted features may be compared to features described in a model of coding parameters generated by a machine learning algorithm from reviews of previously-coded videos, extracted features of the previously-coded videos, and coding parameters of the previously-coded videos. When a match is detected between the extracted features of the input video sequence and extracted features represented in the model, a determination may be made as to whether coding parameters that correspond to the matching extracted feature correspond to a tier of service to which the input video sequence is to be coded.
    Type: Grant
    Filed: October 4, 2021
    Date of Patent: November 19, 2024
    Assignee: APPLE INC.
    Inventors: Yeping Su, Xingyu Zhang, Chris Chung, Jun Xin, Hsi-Jung Wu
  • Patent number: 12137199
    Abstract: A system obtains a data set representing immersive video content for display at a display time, including first data representing the content according to a first level of detail, and second data representing the content according to a second higher level of detail. During one or more first times prior to the display time, the system causes at least a portion of the first data to be stored in a buffer. During one or more second times prior to the display time, the system generates a prediction of a viewport for displaying the content to a user at the display time, identifies a portion of the second data corresponding to the prediction of the viewport, and causes the identified portion of the second data to be stored in the video buffer. At the display time, the system causes the content to be displayed to the user using the video buffer.
    Type: Grant
    Filed: January 8, 2024
    Date of Patent: November 5, 2024
    Assignee: Apple Inc.
    Inventors: Fanyi Duanmu, Jun Xin, Hsi-Jung Wu, Xiaosong Zhou
  • Publication number: 20240340438
    Abstract: Techniques are proposed to improve temporal motion projection in video coding. Candidate reference frames available for use in temporal motion projection are sorted in processing order according to scores assigned based on estimates of the reference frames' suitability for prediction. Such estimates may be based on temporal distance between each candidate reference frame and that reference frame's prediction references. Estimates may be based, for each reference frame, based on an estimate of coding quality of a reference frame from which the respective candidate reference frame makes a prediction reference. Once sorted, the candidate reference frames may be processing in the sorting order to supply prediction data to a current frame that is to be coded from the candidate reference frames. Additionally, hardware friendly designs of motion field hole filling and motion vector smoothing operations are proposed.
    Type: Application
    Filed: March 1, 2024
    Publication date: October 10, 2024
    Inventors: Yeqing WU, Yunfei ZHENG, Yixin DU, Alican NALCI, Hilmi E. EGILMEZ, Guoxin JIN, Alexandros TOURAPIS, Jun XIN, Hsi-Jung WU
  • Publication number: 20240323442
    Abstract: In an example method, a decoder accesses a bitstream representing video content, and parses one or more flexible coefficient position (FCP) syntax from the bitstream, where the one or more FCP syntax indicate one or more index values. The decoder further determines side information representing one or more characteristics of an encoded portion of the video content. The decoder interprets the one or more FCP syntax based on the side information, including determining a coefficient position with respect to the encoded portion of the video content based on the one or more index values and the side information. The decoder decodes the encoded portion of the video content according to the coefficient position.
    Type: Application
    Filed: March 12, 2024
    Publication date: September 26, 2024
    Inventors: Alican Nalci, Guoxin Jin, Hilmi Enes Egilmez, Hsi-Jung Wu, Jun Xin, Yixin Du, Yunfei Zheng, Yeqing Wu
  • Publication number: 20240195978
    Abstract: The present disclosure describes techniques for efficient coding of motion vectors developed for multi-hypothesis coding applications. According to these techniques, when coding hypotheses are developed, each having a motion vector identifying a source of prediction for a current pixel block, a motion vector for a first one of the coding hypotheses may be predicted from the motion vector of a second coding hypothesis. The first motion vector may be represented by coding a motion vector residual, which represents a difference between the developed motion vector for the first coding hypothesis and the predicted motion vector for the first coding hypothesis, and outputting the coded residual to a channel. In another embodiment, a motion vector residual may be generated for a motion vector of a first coding hypothesis, and the first motion vector and the motion vector residual may be used to predict a second motion vector and a predicted motion vector residual.
    Type: Application
    Filed: December 13, 2022
    Publication date: June 13, 2024
    Inventors: Yeqing WU, Yunfei Zheng, Alexandros Tourapis, Yixin Du, Alican Nalci, Hilmi E. Egilmez, Guoxin Jin, Jun Xin, Hsi-Jung Wu
  • Publication number: 20240146892
    Abstract: A system obtains a data set representing immersive video content for display at a display time, including first data representing the content according to a first level of detail, and second data representing the content according to a second higher level of detail. During one or more first times prior to the display time, the system causes at least a portion of the first data to be stored in a buffer. During one or more second times prior to the display time, the system generates a prediction of a viewport for displaying the content to a user at the display time, identifies a portion of the second data corresponding to the prediction of the viewport, and causes the identified portion of the second data to be stored in the video buffer. At the display time, the system causes the content to be displayed to the user using the video buffer.
    Type: Application
    Filed: January 8, 2024
    Publication date: May 2, 2024
    Inventors: Fanyi Duanmu, Jun Xin, Hsi-Jung Wu, Xiaosong Zhou
  • Publication number: 20240129472
    Abstract: Improved lossless entropy coding techniques for coding of image data include selecting a context for entropy coding based on an ordered scan path of possible context locations. A symbol for a current location within a source image may be entropy coded based on a context of prior encoded symbols of other locations within source images, where the context is selected based on an ordered scan path enumerating a series of potential context locations within one or more source images. To select a context, a predetermined number of prior symbols may be selected by qualifying or disqualifying locations in the scan path, and then the current symbol may be encoded with a context based on prior symbols corresponding to the first qualifying context locations in the order of the scan path.
    Type: Application
    Filed: September 18, 2023
    Publication date: April 18, 2024
    Inventors: Yeqing WU, Yunfei ZHENG, Alican NALCI, Yixin DU, Hilmi Enes EGILMEZ, Guoxin JIN, Alexandros TOURAPIS, Jun XIN, Hsi-Jung WU
  • Patent number: 11956295
    Abstract: Techniques for multi-view video streaming are described in the present disclosure, wherein a viewport prediction may be employed at a client-end based on analysis of pre-fetched media item data and ancillary information. A streaming method may first prefetch a portion of content of a multi-view media item. The method may next identify a salient region from the prefetched content and may then download additional content of the media item that corresponds to the identified salient region.
    Type: Grant
    Filed: March 20, 2020
    Date of Patent: April 9, 2024
    Assignee: APPLE INC.
    Inventors: Fanyi Duanmu, Alexandros Tourapis, Jun Xin, Hsi-Jung Wu, Xiaosong Zhou
  • Patent number: 11924391
    Abstract: A system obtains a data set representing immersive video content for display at a display time, including first data representing the content according to a first level of detail, and second data representing the content according to a second higher level of detail. During one or more first times prior to the display time, the system causes at least a portion of the first data to be stored in a buffer. During one or more second times prior to the display time, the system generates a prediction of a viewport for displaying the content to a user at the display time, identifies a portion of the second data corresponding to the prediction of the viewport, and causes the identified portion of the second data to be stored in the video buffer. At the display time, the system causes the content to be displayed to the user using the video buffer.
    Type: Grant
    Filed: December 16, 2022
    Date of Patent: March 5, 2024
    Assignee: Apple Inc.
    Inventors: Fanyi Duanmu, Jun Xin, Hsi-Jung Wu, Xiaosong Zhou
  • Publication number: 20240073438
    Abstract: Techniques are disclosed for improved video coding with virtual reference frames. A motion vector for prediction of a pixel block from a reference may be constrained based on the reference. In as aspect, if the reference is a temporally interpolated virtual reference frame with corresponding time close to the time of the current pixel block, the motion vector for prediction may be constrained magnitude and/or precision. In another aspect, a bitstream syntax for encoding the constrained motion vector may also be constrained. In this manner, the techniques proposed herein contribute to improved coding efficiencies.
    Type: Application
    Filed: August 18, 2023
    Publication date: February 29, 2024
    Inventors: Yeqing WU, Yunfei ZHENG, Guoxin JIN, Yixin DU, Alican NALCI, Hilmi Enes EGILMEZ, Jun XIN, Hsi-Jung WU
  • Publication number: 20240048776
    Abstract: Disclosed is a method that includes receiving an image frame having a plurality of coded blocks, determining a prediction unit (PU) from the plurality of coded blocks, determining one or more motion compensation units arranged in an array within the PU, and applying a filter to one or more boundaries of the one or more motion compensation units. Also disclosed is a method that includes receiving a reference frame that includes a reference block, determining a timing for deblocking a current block, performing motion compensation on the reference frame to obtain a predicted frame that includes a predicted block, performing reconstruction on the predicted frame to obtain a reconstructed frame that includes a reconstructed PU, and applying, at the timing for deblocking the current block, a deblocking filter based on one or more parameters to the reference block, the predicted block, or the reconstructed PU.
    Type: Application
    Filed: September 29, 2022
    Publication date: February 8, 2024
    Inventors: Yixin Du, Alexandros Tourapis, Alican Nalci, Guoxin Jin, Hilmi Enes Egilmez, Hsi-Jung Wu, Jun Xin, Yeqing Wu, Yunfei Zheng
  • Publication number: 20240040151
    Abstract: Techniques are described for express and implied signaling of transform mode selections in video coding. Information derived from coefficient samples in a given transform unit (TU) or prediction unit (PU) may constrain or modify signaling of certain syntax elements at the coding block (CB), TU, or PU levels. For instance, based on the spatial locations of decoded coefficients, the spatial patterns of coefficients, or the correlation with the coefficients in neighboring blocks, various syntax elements such as the transform type and related flags/indices or secondary transform modes/flags indices, a residual coding mode, intra and inter prediction modes, and scanning order may be disabled or constrained. In another case, if the coefficient samples match a desired spatial pattern or have other desired properties then a default transform type, a default secondary transform type, a default intra and inter prediction mode or other block level modes may be inferred at the decoder side.
    Type: Application
    Filed: May 4, 2023
    Publication date: February 1, 2024
    Inventors: Alican Nalci, Yunfei Zheng, Hilmi E. Egilmez, Yeqing WU, Yixin Du, Alexis Tourapis, Jun Xin, Hsi-Jung Wu
  • Publication number: 20240040124
    Abstract: A flexible coefficient coding (FCC) approach is presented. In the first aspect, spatial sub-regions are defined over a transform unit (TU) or a prediction unit (PU). These sub-regions organize the coefficient samples residing inside a TU or a PU into variable coefficient groups (VCGs). Each VCG corresponds to a sub-region inside a larger TU or PU. The shape of VCGs or the boundaries between different VCGs may be irregular, determined based on the relative distance of coefficient samples with respect to each other. Alternatively, the VCG regions may be defined according to scan ordering within a TU. Each VCG can encode a 1) different number of symbols for a given syntax element, or a 2) different number of syntax elements within the same TU or PU. Whether to code more symbols or more syntax elements may depend on the type of arithmetic coding engine used in a particular coding specification. For multi-symbol arithmetic coding (MS-AC), a VCG may encode a different number of symbols for a syntax element.
    Type: Application
    Filed: July 25, 2023
    Publication date: February 1, 2024
    Inventors: Alican NALCI, Yunfei ZHENG, Hilmi Enes EGILMEZ, Yeqing WU, Yixin DU, Alexandros TOURAPIS, Jun XIN, Hsi-Jung WU, Arash VOSOUGHI, Dzung T. HOANG