Patents by Inventor Hsi-Jung Wu

Hsi-Jung Wu has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20210097725
    Abstract: A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. A closed-loop color conversion process is used to improve compression while taking into consideration distortion introduced throughout the point cloud compression process.
    Type: Application
    Filed: December 11, 2020
    Publication date: April 1, 2021
    Applicant: Apple Inc.
    Inventors: Khaled Mammou, Yeping Su, David W. Singer, Alexandros Tourapis, Fabrice A. Robinet, Hsi-Jung Wu
  • Publication number: 20210099507
    Abstract: Techniques for multi-view video streaming are described in the present disclosure, wherein a viewport prediction may be employed at a client-end based on analysis of pre-fetched media item data and ancillary information. A streaming method may first prefetch a portion of content of a multi-view media item. The method may next identify a salient region from the prefetched content and may then download additional content of the media item that corresponds to the identified salient region.
    Type: Application
    Filed: March 20, 2020
    Publication date: April 1, 2021
    Inventors: Fanyi DUANMU, Alexandros TOURAPIS, Jun XIN, Hsi-Jung WU, Xiaosong ZHOU
  • Publication number: 20210084332
    Abstract: Techniques are disclosed for coding and delivering multi-view video in which the video is represented as a manifest file identifying a plurality of segments of the video available for download. The multi-view video may be partitioned spatially into a plurality of tiles that, in aggregate, encompass the entire spatial area of the video. The tiles are coded as segments contains coded video representing content contained within its respective tile. Tiles may be given different sizes based on saliency of the content within their respective regions. In this manner, tiles with high levels of interest may have relatively large spatial areas, which can lead to efficient coding in the presence of content motion.
    Type: Application
    Filed: September 13, 2019
    Publication date: March 18, 2021
    Inventors: Fanyi DUANMU, Eduardo ASBUN, Xiaosong ZHOU, Jun XIN, Hsi-Jung WU, John SU, Samir GEHANI, Christopher FLICK, Shalini SAHOO
  • Publication number: 20210076054
    Abstract: Techniques for video exchange are disclosed in which a video source sends redundant copies of video coded at a given presentation time. A decoder may determine, from metadata stored with a first decoded frame, whether other decoded frames have been designated as correlated to the first decoded frame. If so, the decoder may fetch other decoding frames using time indicator values of other decoded frames in metadata stored with the first decoded frame. When other decoded frame(s) are found, the decoder may blend content of the first coded frame and the other decoded frame(s). The decoder may render the blended frame. When one such coded frame is an intra-coded frame, the proposed technique may alleviate beating artifacts by blending recovered data of the intra-coded frame with recovered data of another frame, which may be inter-coded.
    Type: Application
    Filed: September 5, 2019
    Publication date: March 11, 2021
    Inventors: Mei GUO, Jun XIN, Yeping SU, Hsi-Jung WU, Alexandros TOURAPIS
  • Patent number: 10939139
    Abstract: In communication applications, aggregate source image data at a transmitter exceeds the data that is needed to display a rendering of a viewport at a receiver. Improved streaming techniques that include estimating a location of a viewport at a future time. According to such techniques, the viewport may represent a portion of an image from a multi-directional video to be displayed at the future time, and tile(s) of the image may be identified in which the viewport is estimated to be located. In these techniques, the image data of tile(s) in which the viewport is estimated to be located may be requested at a first service tier, and the other tile in which the viewport is not estimated to be located may be requested at a second service tier, lower than the first service tier.
    Type: Grant
    Filed: November 29, 2018
    Date of Patent: March 2, 2021
    Assignee: Apple Inc.
    Inventors: Xiaohua Yang, Alexandros Tourapis, Dazhong Zhang, Hang Yuan, Hsi-Jung Wu, Jae Hoon Kim, Jiefu Zhai, Ming Chen, Xiaosong Zhou
  • Patent number: 10924747
    Abstract: Techniques are disclosed for coding and decoding video captured as cube map images. According to these techniques, padded reference images are generated for use during predicting input data. A reference image is stored in a cube map format. A padded reference image is generated from the reference image in which image data of a first view contained in reference image is replicated and placed adjacent to a second view contained in the cube map image. When coding a pixel block of an input image, a prediction search may be performed between the input pixel block and content of the padded reference image. When the prediction search identifies a match, the pixel block may be coded with respect to matching data from the padded reference image. Presence of replicated data in the padded reference image is expected to increase the likelihood that adequate prediction matches will be identified for input pixel block data, which will increase overall efficiency of the video coding.
    Type: Grant
    Filed: February 27, 2017
    Date of Patent: February 16, 2021
    Assignee: Apple Inc.
    Inventors: Jae Hoon Kim, Chris Y. Chung, Dazhong Zhang, Hang Yuan, Hsi-Jung Wu, Jiefu Zhai, Xiaosong Zhou
  • Patent number: 10924707
    Abstract: Computing devices may implement instant video communication connections for video communications. Connection information for mobile computing devices may be maintained. A request to initiate an instant video communication may be received, and if authorized, the connection information for the particular recipient mobile computing device may be accessed. Video communication data may then be sent to the recipient mobile computing device according to the connection information so that the video communication data may be displayed at the recipient device as it is received. New connection information for different mobile computing devices may be added, or updates to existing connection information may also be performed. Connection information for some mobile computing devices may be removed.
    Type: Grant
    Filed: August 24, 2018
    Date of Patent: February 16, 2021
    Assignee: Apple Inc.
    Inventors: Xiaosong Zhou, Hsi-Jung Wu, Chris Y. Chung, James Normile, Joe S. Abuan, Hyeonkuk Jeong, Yan Yang, Gobind Johar, Thomas Christopher Jansen
  • Publication number: 20210012536
    Abstract: A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. The encoder is configured project the point cloud on to patch planes to compress the point cloud, and supports multiple layered patch planes. For example, some point clouds may have a depth, and points at different depths may be assigned to different layered patch planes.
    Type: Application
    Filed: September 25, 2020
    Publication date: January 14, 2021
    Applicant: Apple Inc.
    Inventors: Khaled Mammou, Yeping Su, David W. Singer, Alexandros Tourapis, Fabrice A. Robinet, Hsi-Jung Wu
  • Patent number: 10873763
    Abstract: Techniques are disclosed for coding high dynamic range (HDR) data. According to such techniques, HDR data may be converted to a domain of uniform luminance data. The uniform domain data may be coded by motion compensated predictive coding. The HDR data also may be coded by motion compensated predictive coding, using a coding parameter that is derived from a counterpart coding parameter of the coding of the uniform domain data. In another technique, HDR data may be coded using coding parameters that are derived from HDR domain processing but distortion measurements may be performed in a uniform domain.
    Type: Grant
    Filed: June 2, 2017
    Date of Patent: December 22, 2020
    Assignee: APPLE INC.
    Inventors: Mei Guo, Jun Xin, Yeping Su, Chris Chung, Dazhong Zhang, Xiaosong Zhou, Hsi-Jung Wu
  • Patent number: 10867413
    Abstract: A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. A closed-loop color conversion process is used to improve compression while taking into consideration distortion introduced throughout the point cloud compression process.
    Type: Grant
    Filed: February 20, 2020
    Date of Patent: December 15, 2020
    Assignee: Apple Inc.
    Inventors: Khaled Mammou, Yeping Su, David W. Singer, Alexandras Tourapis, Fabrice A. Robinet, Hsi-Jung Wu
  • Publication number: 20200389662
    Abstract: Techniques are disclosed for coding and decoding video data using object recognition and object modeling as a basis of coding and error recovery. A video decoder may decode coded video data received from a channel. The video decoder may perform object recognition on decoded video data obtained therefrom, and, when an object is recognized in the decoded video data, the video decoder may generate a model representing the recognized object. It may store data representing the model locally. The video decoder may communicate the model data to an encoder, which may form a basis of error mitigation and recovery. The video decoder also may monitor deviation patterns in the object model and associated patterns in audio content; if/when video decoding is suspended due to operational errors, the video decoder may generate simulated video data by analyzing audio data received during the suspension period and developing video data from the data model and deviation(s) associated with patterns detected from the audio data.
    Type: Application
    Filed: May 11, 2020
    Publication date: December 10, 2020
    Inventors: Xing Wen, Dazhong Zhang, Peikang Song, Xiaosong Zhou, Sudeng Hu, Hsi-Jung Wu, Jae Hoon Kim
  • Publication number: 20200382806
    Abstract: Techniques are disclosed for coding video data in which frames from a video source are partitioned into a plurality of tiles of common size, and the tiles are coded as a virtual video sequence according to motion-compensated prediction, each tile treated as having respective temporal location of the virtual video sequence. The coding scheme permits relative allocation of coding resources to tiles that are likely to have greater significance in a video coding session, which may lead to certain tiles that have low complexity or low motion content to be skipped during coding of the tiles for select source frames. Moreover, coding of the tiles may be ordered to achieve low coding latencies during a coding session.
    Type: Application
    Filed: May 26, 2020
    Publication date: December 3, 2020
    Inventors: Dazhong ZHANG, Peikang SONG, Beibei WANG, Giribalan GOPALAN, Albert E. KEINATH, Christopher M. GARRIDO, David R. CONRAD, Hsi-Jung WU, Ming JIN, Hang YUAN, Xiaohua YANG, Xiaosong ZHOU, Vikrant KASARABADA, Davide CONCION, Eric L. CHIEN, Bess C. CHAN, Karthick SANTHANAM, Gurtej Singh CHANDOK
  • Publication number: 20200382745
    Abstract: Techniques are presented for managing for visual prominence of participants in a video conference, including conferences where participants communicate visually, such as with sign language. According to these techniques, a visual prominence indication of a participant in a video conference may be estimated, a video stream of the participant may be encoded, and the encoded video stream may be transmitted along with an indication of the estimated visual prominence to a receiving device in the video conference.
    Type: Application
    Filed: November 20, 2019
    Publication date: December 3, 2020
    Inventors: Johnny Trenh, Hsi-Jung Wu, Sarah K. Herrlinger, Xiaoxia Sun, Ian J. Baird, Dazhong Zhang, Xiaosong Zhou, Christopher M. Garrido
  • Patent number: 10848762
    Abstract: Method For Implementing A Quantizer In A Multimedia Compression And Encoding System is disclosed. In the Quantizer system of the present invention, several new quantization ideas are disclosed. In one embodiment, adjacent macroblocks are grouped together into macroblock groups. The macroblock groups are then assigned a common quantizer value. The common quantizer value may be selected based upon how the macroblocks are encoded, the type of macroblocks within the macroblock group (intra-blocks or inter-blocks), the history of the motion vectors associated with the macroblocks in the macroblock group, the residuals of the macroblocks in the macroblock group, and the energy of the macroblocks in the macroblock group. The quantizer value may be adjusted in a manner that is dependent on the current quantizer value. Specifically, if the quantizer value is at the low end of the quantizer scale, then only small adjustments are made.
    Type: Grant
    Filed: September 1, 2018
    Date of Patent: November 24, 2020
    Inventors: Thomas Pun, Roger Kumar, Xiaochun Nie, Hsi-Jung Wu
  • Patent number: 10812832
    Abstract: Coding techniques for image data may cause a still image to be converted to a “phantom” video sequence, which is coded by motion compensated prediction techniques. Thus, coded video data obtained from the coding operation may include temporal prediction references between frames of the video sequence. Metadata may be generated that identifies allocations of content from the still image to the frames of the video sequence. The coded data and the metadata may be transmitted to another device, whereupon it may be decoded by motion compensated prediction techniques and converted back to a still image data. Other techniques may involve coding an image in both a base layer representation and at least one coded enhancement layer representation. The enhancement layer representation may be coded predictively with reference to the base layer representation. The coded base layer representation may be partitioned into a plurality of individually-transmittable segments and stored.
    Type: Grant
    Filed: June 5, 2015
    Date of Patent: October 20, 2020
    Assignee: APPLE INC.
    Inventors: Hang Yuan, Chris Y. Chung, Jae Hoon Kim, Yeping Su, Jiefu Zhai, Xiaosong Zhou, Hsi-Jung Wu
  • Patent number: 10789733
    Abstract: A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. The encoder is configured project the point cloud on to patch planes to compress the point cloud, and supports multiple layered patch planes. For example, some point clouds may have a depth, and points at different depths may be assigned to different layered patch planes.
    Type: Grant
    Filed: November 21, 2018
    Date of Patent: September 29, 2020
    Assignee: Apple Inc.
    Inventors: Khaled Mammou, Yeping Su, David W. Singer, Alexandros Tourapis, Fabrice A. Robinet, Hsi-Jung Wu
  • Publication number: 20200304807
    Abstract: The present disclosure describes techniques for coding and decoding video in which a plurality of coding hypotheses are developed for an input pixel block of frame content. Each coding hypothesis may include generation of prediction data for the input pixel block according to a respective prediction search. The input pixel block may be coded with reference to a prediction block formed from prediction data derived according to plurality of hypotheses. Data of the coded pixel block may be transmitted to a decoder along with data identifying a number of the hypotheses used during the coding to a channel. At a decoder, an inverse process may be performed, which may include generation of a counterpart prediction block from prediction data derived according to the hypothesis identified with the coded pixel block data, then decoding of the coded pixel block according to the prediction data.
    Type: Application
    Filed: May 20, 2020
    Publication date: September 24, 2020
    Inventors: Alexandros Michael TOURAPIS, Yeping SU, David SINGER, Hsi-Jung WU
  • Publication number: 20200304837
    Abstract: An encoding system may include a video source that captures video image, a video coder, and a controller to manage operation of the system. The video coder may encode the video image into encoded video data using a plurality of subgroup parameters corresponding to a plurality of subgroups of pixels within a group. The controller may set the subgroup parameters for at least one of the subgroups of pixels in the video coder, based upon at least one parameters corresponding to the group. A decoding system may decode the video data based upon the motion prediction parameters.
    Type: Application
    Filed: June 8, 2020
    Publication date: September 24, 2020
    Inventors: Yunfei Zheng, Dazhong Zhang, Xiaosong Zhou, Chris Y. Chung, Hsi-Jung Wu
  • Publication number: 20200296426
    Abstract: Chroma deblock filtering of reconstructed video samples may be performed to remove blockiness artifacts and reduce color artifacts without over-smoothing. In a first method, chroma deblocking may be performed for boundary samples of a smallest transform size, regardless of partitions and coding modes. In a second method, chroma deblocking may be performed when a boundary strength is greater than 0. In a third method, chroma deblocking may be performed regardless of boundary strengths. In a fourth method, the type of chroma deblocking to be performed may be signaled in a slice header by a flag. Furthermore, luma deblock filtering techniques may be applied to chroma deblock filtering.
    Type: Application
    Filed: June 2, 2020
    Publication date: September 17, 2020
    Inventors: Jiefu Zhai, Dazhong Zhang, Xiaosong Zhou, Chris Y. Chung, Hsi-Jung Wu, Peikang Song, David R. Conrad, Jae Hoon Kim, Yunfei Zheng
  • Publication number: 20200273208
    Abstract: A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. A block/sub-block organization scheme is used to encode blocks and sub-blocks of an occupancy map used in compressing the point cloud. Binary values are assigned to blocks/sub-blocks based on whether they contain patches projected on the point cloud. A traversal path is chosen that takes advantage of run-length encoding strategies to reduce a size of an encoded occupancy map. Also, auxiliary information is used to further improve occupancy map compression.
    Type: Application
    Filed: May 13, 2020
    Publication date: August 27, 2020
    Applicant: Apple Inc.
    Inventors: Khaled Mammou, Yeping Su, David W. Singer, Alexandros Tourapis, Fabrice A. Robinet, Hsi-Jung Wu