Patents by Inventor Jun Xin

Jun Xin has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240146892
    Abstract: A system obtains a data set representing immersive video content for display at a display time, including first data representing the content according to a first level of detail, and second data representing the content according to a second higher level of detail. During one or more first times prior to the display time, the system causes at least a portion of the first data to be stored in a buffer. During one or more second times prior to the display time, the system generates a prediction of a viewport for displaying the content to a user at the display time, identifies a portion of the second data corresponding to the prediction of the viewport, and causes the identified portion of the second data to be stored in the video buffer. At the display time, the system causes the content to be displayed to the user using the video buffer.
    Type: Application
    Filed: January 8, 2024
    Publication date: May 2, 2024
    Inventors: Fanyi Duanmu, Jun Xin, Hsi-Jung Wu, Xiaosong Zhou
  • Publication number: 20240129472
    Abstract: Improved lossless entropy coding techniques for coding of image data include selecting a context for entropy coding based on an ordered scan path of possible context locations. A symbol for a current location within a source image may be entropy coded based on a context of prior encoded symbols of other locations within source images, where the context is selected based on an ordered scan path enumerating a series of potential context locations within one or more source images. To select a context, a predetermined number of prior symbols may be selected by qualifying or disqualifying locations in the scan path, and then the current symbol may be encoded with a context based on prior symbols corresponding to the first qualifying context locations in the order of the scan path.
    Type: Application
    Filed: September 18, 2023
    Publication date: April 18, 2024
    Inventors: Yeqing WU, Yunfei ZHENG, Alican NALCI, Yixin DU, Hilmi Enes EGILMEZ, Guoxin JIN, Alexandros TOURAPIS, Jun XIN, Hsi-Jung WU
  • Publication number: 20240127250
    Abstract: A resource transfer method includes obtaining, by using a target application, a resource transfer request triggered by a target user, where the resource transfer request includes verification information used to perform resource transfer processing and identity feature information of the target user. By using the target application, verifying the verification information, and invoking, if the verification succeeds by using the target application, a local device management rule, and determining, by using the local device management rule, whether the identity feature information of the target user matches identity feature information of a pre-registered user. If the identity feature information of the target user matches the identity feature information of the pre-registered user, sending the resource transfer request to a first server corresponding to the target application to trigger, based on the resource transfer request, the first server to perform resource transfer processing.
    Type: Application
    Filed: December 28, 2023
    Publication date: April 18, 2024
    Applicant: ALIPAY (HANGZHOU) INFORMATION TECHNOLOGY CO., LTD.
    Inventors: Bingying Zhu, Jun Wen, Fuping Yu, Zhi Xin, Xiaofei Wan
  • Patent number: 11956295
    Abstract: Techniques for multi-view video streaming are described in the present disclosure, wherein a viewport prediction may be employed at a client-end based on analysis of pre-fetched media item data and ancillary information. A streaming method may first prefetch a portion of content of a multi-view media item. The method may next identify a salient region from the prefetched content and may then download additional content of the media item that corresponds to the identified salient region.
    Type: Grant
    Filed: March 20, 2020
    Date of Patent: April 9, 2024
    Assignee: APPLE INC.
    Inventors: Fanyi Duanmu, Alexandros Tourapis, Jun Xin, Hsi-Jung Wu, Xiaosong Zhou
  • Publication number: 20240093348
    Abstract: A mask module includes a framework and a first strip plate fixed on the framework and extending along a first direction. The first strip plate is provided with a first concave-convex structure on one side edge along the first direction. At least one convex structure is provided in a middle of the first concave-convex structure along the first direction. The convex structure and the first stripe plate are integrally formed.
    Type: Application
    Filed: November 30, 2023
    Publication date: March 21, 2024
    Inventors: Naichao MU, Yuan Li, Yu Xin, Jun Ma, Lijing Han
  • Publication number: 20240093349
    Abstract: A mask module includes a framework, a first strip plate fixed on the framework and extending along a first direction, and a first mask. The first mask is located on a side, deviating from the framework, of the first strip plate. The first mask includes at least one preset area, and the preset area includes at least one opening area. The first strip plate is provided with a first concave-convex structure on one side edge along the first direction. At least one convex structure is provided in a middle of the first concave-convex structure along the first direction. In a direction perpendicular to a surface of the first mask, the first concave-convex structure and the convex structure cover at least a part of area of the at least one opening of the preset area. The convex structure and the first stripe plate are integrally formed.
    Type: Application
    Filed: November 30, 2023
    Publication date: March 21, 2024
    Inventors: Naichao MU, Yuan LI, Yu XIN, Jun MA, Lijing HAN
  • Publication number: 20240096344
    Abstract: An embodiment includes identifying, during a video conference attended by a first attendee, other attendees of the video conference. The embodiment renders a virtual meeting environment including virtual representations of the other attendees, where the rendering includes accessing relationship characteristic data indicative of relationships between the first attendee and other attendees. The embodiment calculates positions for virtual representations of the other attendees in the first attendee's virtual field of view based on the relationship characteristic data. The embodiment also detects simultaneous speech from two of the other attendees and, in response, directs the individual speech from each of the other attendees to respective audio channels.
    Type: Application
    Filed: September 16, 2022
    Publication date: March 21, 2024
    Applicant: International Business Machines Corporation
    Inventors: Wei Jun Zheng, Shi Kun Li, Shuai Cao, Fan Xiao Xin, Ying Li
  • Patent number: 11924391
    Abstract: A system obtains a data set representing immersive video content for display at a display time, including first data representing the content according to a first level of detail, and second data representing the content according to a second higher level of detail. During one or more first times prior to the display time, the system causes at least a portion of the first data to be stored in a buffer. During one or more second times prior to the display time, the system generates a prediction of a viewport for displaying the content to a user at the display time, identifies a portion of the second data corresponding to the prediction of the viewport, and causes the identified portion of the second data to be stored in the video buffer. At the display time, the system causes the content to be displayed to the user using the video buffer.
    Type: Grant
    Filed: December 16, 2022
    Date of Patent: March 5, 2024
    Assignee: Apple Inc.
    Inventors: Fanyi Duanmu, Jun Xin, Hsi-Jung Wu, Xiaosong Zhou
  • Publication number: 20240073438
    Abstract: Techniques are disclosed for improved video coding with virtual reference frames. A motion vector for prediction of a pixel block from a reference may be constrained based on the reference. In as aspect, if the reference is a temporally interpolated virtual reference frame with corresponding time close to the time of the current pixel block, the motion vector for prediction may be constrained magnitude and/or precision. In another aspect, a bitstream syntax for encoding the constrained motion vector may also be constrained. In this manner, the techniques proposed herein contribute to improved coding efficiencies.
    Type: Application
    Filed: August 18, 2023
    Publication date: February 29, 2024
    Inventors: Yeqing WU, Yunfei ZHENG, Guoxin JIN, Yixin DU, Alican NALCI, Hilmi Enes EGILMEZ, Jun XIN, Hsi-Jung WU
  • Publication number: 20240048776
    Abstract: Disclosed is a method that includes receiving an image frame having a plurality of coded blocks, determining a prediction unit (PU) from the plurality of coded blocks, determining one or more motion compensation units arranged in an array within the PU, and applying a filter to one or more boundaries of the one or more motion compensation units. Also disclosed is a method that includes receiving a reference frame that includes a reference block, determining a timing for deblocking a current block, performing motion compensation on the reference frame to obtain a predicted frame that includes a predicted block, performing reconstruction on the predicted frame to obtain a reconstructed frame that includes a reconstructed PU, and applying, at the timing for deblocking the current block, a deblocking filter based on one or more parameters to the reference block, the predicted block, or the reconstructed PU.
    Type: Application
    Filed: September 29, 2022
    Publication date: February 8, 2024
    Inventors: Yixin Du, Alexandros Tourapis, Alican Nalci, Guoxin Jin, Hilmi Enes Egilmez, Hsi-Jung Wu, Jun Xin, Yeqing Wu, Yunfei Zheng
  • Publication number: 20240040151
    Abstract: Techniques are described for express and implied signaling of transform mode selections in video coding. Information derived from coefficient samples in a given transform unit (TU) or prediction unit (PU) may constrain or modify signaling of certain syntax elements at the coding block (CB), TU, or PU levels. For instance, based on the spatial locations of decoded coefficients, the spatial patterns of coefficients, or the correlation with the coefficients in neighboring blocks, various syntax elements such as the transform type and related flags/indices or secondary transform modes/flags indices, a residual coding mode, intra and inter prediction modes, and scanning order may be disabled or constrained. In another case, if the coefficient samples match a desired spatial pattern or have other desired properties then a default transform type, a default secondary transform type, a default intra and inter prediction mode or other block level modes may be inferred at the decoder side.
    Type: Application
    Filed: May 4, 2023
    Publication date: February 1, 2024
    Inventors: Alican Nalci, Yunfei Zheng, Hilmi E. Egilmez, Yeqing WU, Yixin Du, Alexis Tourapis, Jun Xin, Hsi-Jung Wu
  • Publication number: 20240040120
    Abstract: Video coders and decoders perform transform coding and decoding on blocks of video content according to an adaptively selected transform type. The transform types are organized into a hierarchy of transform sets where each transform set includes a respective number of transforms and each higher-level transform set includes the transforms of each lower-level transform set within the hierarchy. The video coders and video decoders may exchange signaling that establishes a transform set context from which a transform set that was selected for coding given block(s) may be identified. The video coders and video decoders may exchange signaling that establishes a transform decoding context from which a transform that was selected from the identified transform set to be used for decoding the transform unit. The block(s) may be coded and decoded by the selected transform.
    Type: Application
    Filed: July 25, 2023
    Publication date: February 1, 2024
    Inventors: Hilmi Enes EGILMEZ, Yunfei ZHENG, Alican NALCI, Yeqing WU, Yixin DU, Guoxin JIN, Alexandros TOURAPIS, Jun XIN, Hsi-Jung WU
  • Publication number: 20240040124
    Abstract: A flexible coefficient coding (FCC) approach is presented. In the first aspect, spatial sub-regions are defined over a transform unit (TU) or a prediction unit (PU). These sub-regions organize the coefficient samples residing inside a TU or a PU into variable coefficient groups (VCGs). Each VCG corresponds to a sub-region inside a larger TU or PU. The shape of VCGs or the boundaries between different VCGs may be irregular, determined based on the relative distance of coefficient samples with respect to each other. Alternatively, the VCG regions may be defined according to scan ordering within a TU. Each VCG can encode a 1) different number of symbols for a given syntax element, or a 2) different number of syntax elements within the same TU or PU. Whether to code more symbols or more syntax elements may depend on the type of arithmetic coding engine used in a particular coding specification. For multi-symbol arithmetic coding (MS-AC), a VCG may encode a different number of symbols for a syntax element.
    Type: Application
    Filed: July 25, 2023
    Publication date: February 1, 2024
    Inventors: Alican NALCI, Yunfei ZHENG, Hilmi Enes EGILMEZ, Yeqing WU, Yixin DU, Alexandros TOURAPIS, Jun XIN, Hsi-Jung WU, Arash VOSOUGHI, Dzung T. HOANG
  • Publication number: 20230396819
    Abstract: A video delivery system generates and stores reduced bandwidth videos from source video. The system may include a track generator that executes functionality of application(s) to be used at sink devices, in which the track generator generates tracks from execution of the application(s) on source video and generates tracks having a reduced data size as compared to the source video. The track generator may execute a first instance of application functionality on the source video, which identifies region(s) of interest from the source video. The track generator further may downsample the source video according to downsampling parameters, and execute a second instance of application functionality on the downsampled video. The track generator may determine, from a comparison of outputs from the first and second instances of the application, whether the output from the second instance of application functionality is within an error tolerance of the output from the first instance of application functionality.
    Type: Application
    Filed: June 1, 2023
    Publication date: December 7, 2023
    Inventors: Ke ZHANG, Xiaoxia SUN, Shujie LIU, Xiaosong ZHOU, Jian LI, Xun SHI, Jiefu ZHAI, Albert E KEINATH, Hsi-Jung WU, Jingteng XUE, Xingyu ZHANG, Jun XIN
  • Publication number: 20230300341
    Abstract: Techniques are disclosed for generating virtual reference frames that may be used for prediction of input video frames. The virtual reference frames may be derived from already-coded reference frames and thereby incur reduced signaling overhead. Moreover, signaling of virtual reference frames may be avoided until an encoder selects the virtual reference frame as a prediction reference for a current frame. In this manner, the techniques proposed herein contribute to improved coding efficiencies.
    Type: Application
    Filed: January 20, 2023
    Publication date: September 21, 2023
    Inventors: Yeqing WU, Yunfei ZHENG, Alexandros TOURAPIS, Alican NALCI, Yixin DU, Hilmi Enes EGILMEZ, Albert E. KEINATH, Jun XIN, Hsi-Jung WU
  • Publication number: 20230188738
    Abstract: In an example method, a decoder obtains a data stream representing video content. The video content is partitioned into one or more logical units, and each of the logical units is partitioned into one or more respective logical sub-units. The decoder determines that the data stream includes first data indicating that a first logical unit has been encoded according to a flexible skip coding scheme. In response, the decoder determines a first set of decoding parameters based on the first data, and decodes each of the logical sub-units of the first logical unit according to the first set of decoding parameters.
    Type: Application
    Filed: December 6, 2022
    Publication date: June 15, 2023
    Inventors: Alican Nalci, Alexandros Tourapis, Hilmi Enes Egilmez, Hsi-Jung Wu, Jun Xin, Yeqing Wu, Yixin Du, Yunfei Zheng
  • Publication number: 20230143147
    Abstract: A cross-component based filtering system is disclosed for video coders and decoders. The filtering system may include a filter having an input for a filter offset and an input for samples reconstructed from coded video data representing a native component of source video on which the filter operates. The offset may be generated at least in part from a sample classifier that classifies samples reconstructed from coded video data representing a color component of the source video orthogonal to the native component according to sample intensity.
    Type: Application
    Filed: November 2, 2022
    Publication date: May 11, 2023
    Inventors: Yixin DU, Alexandros TOURAPIS, Yunfei ZHENG, Jun XIN, Alican NALCI, Mei T. GUO, Yeqing WU, Hsi-Jung WU
  • Publication number: 20230142771
    Abstract: A filtering system for video coders and decoders is disclosed that includes a feature detector having an input for samples reconstructed from coded video data representing a color component of source video, and having an output for data identifying a feature recognized therefrom, an offset calculator having an input for the feature identification data from the feature detector and having an output for a filter offset, and a filter having an input for the filter offset from the offset calculator and an input for the reconstructed samples, and having an output for filtered samples. The filtering system is expected to improve operations of video coder/decoder filtering systems by selecting filtering offsets from analysis of recovered video data in a common color plane as the samples that will be filtered.
    Type: Application
    Filed: November 2, 2022
    Publication date: May 11, 2023
    Inventors: Yixin DU, Alexandros TOURAPIS, Yunfei ZHENG, Jun XIN, Mukta S. Gore, Alican NALCI, Mei T. GUO, Yeqing WU, Hsi-Jung WU
  • Publication number: 20230117742
    Abstract: A system obtains a data set representing immersive video content for display at a display time, including first data representing the content according to a first level of detail, and second data representing the content according to a second higher level of detail. During one or more first times prior to the display time, the system causes at least a portion of the first data to be stored in a buffer. During one or more second times prior to the display time, the system generates a prediction of a viewport for displaying the content to a user at the display time, identifies a portion of the second data corresponding to the prediction of the viewport, and causes the identified portion of the second data to be stored in the video buffer. At the display time, the system causes the content to be displayed to the user using the video buffer.
    Type: Application
    Filed: December 16, 2022
    Publication date: April 20, 2023
    Inventors: Fanyi Duanmu, Jun Xin, Hsi-Jung Wu, Xiaosong Zhou
  • Publication number: 20230096567
    Abstract: Improved neural-network-based image and video coding techniques are presented, including hybrid techniques that include both tools of a host codec and neural-network-based tools. In these improved techniques, the host coding tools may include conventional video coding standards such H.266 (VVC). In an aspects, source frames may be partitioned and either host or neural-network-based tools may be selected per partition. Coding parameter decisions for a partition may be constrained based on the partitioning and coding tool selection. Rate control for host and neural network tools may be combined. Multi-stage processing of neural network output may use a checkerboard prediction pattern.
    Type: Application
    Filed: September 23, 2022
    Publication date: March 30, 2023
    Inventors: Alican NALCI, Alexandros TOURAPIS, Hsi-Jung WU, Jiefu ZHAI, Jingteng XUE, Jun XIN, Mei GUO, Xingyu ZHANG, Yeqing WU, Yunfei ZHENG, Jean Begaint