Patents by Inventor Shiqi Wang
Shiqi Wang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 12294739Abstract: A video processing method includes performing a conversion between a video including a video unit and a coded representation of the video, where, after the video unit is encoded or decoded with an intra prediction mode, one or more frequence tables and/or one or more sorted intra prediction mode (IPM) tables are selectively updated according to a rule, where the one or more frequence tables include information about frequence of the intra prediction mode used for processing the video unit in the conversion, where the frequence indicates an occurrence of the intra prediction mode used for the conversion, and where the one or more sorted IPM tables indicate the intra prediction mode used in the processing.Type: GrantFiled: February 22, 2022Date of Patent: May 6, 2025Assignees: Beijing Bytedance Network Technology Co., Ltd., Bytedance Inc., Bytedance (HK) LimitedInventors: Junru Li, Meng Wang, Li Zhang, Kai Zhang, Hongbin Liu, Yue Wang, Shiqi Wang
-
Publication number: 20250142090Abstract: An AZB detection method for video coding, which includes the steps of detecting if a residual signal includes a spatial domain GAZB; detecting if the residual signal includes a frequency domain GAZB if no spatial domain GAZB is detected in the previous step; detecting if the residual signal includes a PAZB if no frequency domain GAZB is detected in the previous step; and determining that the residual signal is a non-AZB signal if no PAZB is detected in the previous step. The proposed method achieves promising time savings for test sequences of different resolutions, with negligible rate-distortion performance loss.Type: ApplicationFiled: August 30, 2024Publication date: May 1, 2025Inventors: Sam Tak Wu KWONG, Shiqi WANG, Zhenhao SUN
-
Patent number: 12288154Abstract: Adaptive verifiable training enables the creation of machine learning models robust with respect to multiple robustness criteria. In general, such training exploits inherent inter-class similarities within input data and enforces multiple robustness criteria based on this information. In particular, the approach exploits pairwise class similarity and improves the performance of a robust model by relaxing robustness constraints for similar classes and increasing robustness constraints for dissimilar classes. Between similar classes, looser robustness criteria (i.e., smaller ?) are enforced so as to minimize possible overlap when estimating the robustness region during verification. Between dissimilar classes, stricter robustness regions (i.e., larger ?) are enforced. If pairwise class relationships are not available initially, preferably they are generated by receiving a pre-trained classifier and then applying a clustering algorithm (e.g., agglomerative clustering) to generate them.Type: GrantFiled: December 7, 2020Date of Patent: April 29, 2025Assignee: International Business Machines CorporationInventors: Kevin Eykholt, Taesung Lee, Jiyong Jang, Shiqi Wang, Ian Michael Molloy
-
Patent number: 12288367Abstract: A method for learning-based point cloud geometry compression includes: given a source point cloud, regressing an aligned mesh that is driven by a set of parameters from a deformable template mesh, quantizing the set of parameters into a parameter bitstream, generating an aligned point cloud from the quantized parameters by mesh manipulation and mesh-to-point-cloud conversion, extracting features from both the source point cloud and the aligned point cloud based on sparse tensors including coordinates and features, the coordinates being encoded into a coordinate bitstream, warping the features of the aligned point cloud onto the coordinates of the source point cloud, obtaining residual features through feature subtraction, processing the residual features using an entropy model into a residual feature bitstream, and obtaining a reconstructed point cloud by processing the parameter bitstream, the coordinate bitstream and the residual feature bitstream.Type: GrantFiled: August 1, 2023Date of Patent: April 29, 2025Assignee: City University of Hong KongInventors: Sam Tak Wu Kwong, Xinju Wu, Shiqi Wang
-
Publication number: 20250131599Abstract: The present disclosure provides a video encoding method, a decoding method, and an apparatus. The video encoding method includes: obtaining an original reference video frame and an original target video frame to be encoded; adjusting a resolution of the original target video frame to obtain an adjusted target video frame with a first preset resolution; and performing feature extraction on the adjusted target video frame to obtain a target feature through a feature extraction network corresponding to the first preset resolution; encoding the original reference video frame and the target features respectively to obtain a video bitstream, and performing video frame reconstruction based on the video bitstream to generate a reconstructed video frame with a same resolution as the original target video frame.Type: ApplicationFiled: December 20, 2024Publication date: April 24, 2025Inventors: Bolin CHEN, Zhao Wang, Yan Ye, Shiqi Wang
-
Publication number: 20250115962Abstract: The present invention provides a method for predicting whether an individual has early myocardial infarction. The method includes: providing a blood sample of the individual; testing a marker in the blood sample to obtain the quantity of the marker; comparing the obtained number with a preset threshold, and judging whether the individual will suffer from myocardial infarction by the comparison, wherein the marker is any one or more nucleotide sequences of miRNA, lncRNA or circRNA.Type: ApplicationFiled: October 9, 2024Publication date: April 10, 2025Inventors: Hanbin CUI, Pengpeng SU, Teng HU, Ning HUANGFU, Jiajun YING, Jiaxi SHEN, Jinsong CHENG, Junsong LIU, Shiqi WANG, Honglin ZHOU, Hengyi MAO, Zhenwei LI
-
Publication number: 20250117966Abstract: A method including deforming the reference sample frame through generator in an initial generative model to generate reconstructed sample frames; inputting each reconstructed sample frame and the corresponding to-be-encoded sample frame into a first discriminator in the initial generative model to obtain a first identification result; splicing the to-be-encoded sample frames in timestamp order to obtain a spliced to-be-encoded sample frame, and splicing the reconstructed sample frames to obtain a spliced reconstructed sample frame; inputting the spliced to-be-encoded sample frame and the spliced reconstructed sample frame into a second discriminator in the initial generative model to obtain a second identification result; obtaining an adversarial loss value based on the first identification result and the second identification result; and training the initial generative model based on the adversarial loss value.Type: ApplicationFiled: December 19, 2024Publication date: April 10, 2025Inventors: Bolin CHEN, Zhao Wang, Yan Ye, Shiqi Wang
-
Publication number: 20250095789Abstract: A framework that comprises a reinforcement-learning-based neural-network for compressing, and for transmitting the compressed genomes over a data network in repeated steps each of a plurality of species. The framework also takes data on inefficient transmission of compressed genome in the preceding step, and feeds this data forward to modify the selection of the compression parameter in the present step. The invention provides the possibility that the genome of any species may be compressed optimally and transmitted in optimal efficiency. That is, big genome sequence is neither over compressed, which takes a lot of processing time leading to delays, nor under compressed which will require more time to transmit.Type: ApplicationFiled: November 27, 2023Publication date: March 20, 2025Inventors: Sam Tak Wu KWONG, Shiqi WANG, Xiaona LI, Zhenhao SUN
-
Publication number: 20250088675Abstract: Methods and apparatuses are provided for performing generative face video compression by using a face feature translator. An exemplary method includes receiving a bitstream associated with a first type of facial feature data representing a facial picture; and decoding, using coded information of the bitstream, one or more pictures, wherein the decoding includes: transforming the first type of facial feature data into a second type of facial feature data; and reconstructing the facial picture based on the second type of facial feature data.Type: ApplicationFiled: August 20, 2024Publication date: March 13, 2025Inventors: Shanzhi YIN, Bolin CHEN, Yan YE, Shiqi WANG
-
Publication number: 20250088636Abstract: A method of decoding a bitstream to output one or more pictures for a video stream. The method includes receiving a bitstream comprising one or more types of facial representation parameters; and decoding, using coded information of the bitstream, one or more pictures. The decoding includes decoding the one or more types of facial representation parameters; converting the one or more types of facial representation parameters into one or more dense motion flows having a common format; and generating a facial picture based on the one or more dense motion flows and a key reference picture of the one or more pictures.Type: ApplicationFiled: August 12, 2024Publication date: March 13, 2025Inventors: Bolin CHEN, Shanzhi YIN, Yan YE, Shiqi WANG
-
Publication number: 20250078206Abstract: A system and a method for a computer implemented method of image enhancement includes the steps of receiving an input image, wherein the input image is a low light image, processing the input image, by a pre trained image enhancer, to generate an initial enhanced image, accessing, from an image memory, a response value corresponding to a sample specific property a normal image, generating an adjustment factor based on the response value from the image memory, generating a final enhanced image by applying the adjustment factor to the initial enhanced image.Type: ApplicationFiled: August 28, 2023Publication date: March 6, 2025Inventors: Tak Wu Sam Kwong, Dongjie Ye, Zhangkai Ni, Shiqi Wang
-
Publication number: 20250066212Abstract: Disclosed is a method for preparing aluminum hydroxide nanowires by a template process. The method includes: (1) dispersing nanocellulose in water to obtain a solution having a concentration of 0.1%-10% by mass; adding an aluminum salt solution to the solution obtained in step (1) and mixing; adding an alkaline solution thereto while stirring, and subjecting a resulting mixture to reactive precipitation, and washing with water and drying, to obtain the aluminum hydroxide nanowires, wherein a mass of an aluminum salt in the aluminum salt solution is in a range of 1-100% of a mass of the nanocellulose, and an alkali in the alkaline solution has a same amount in moles as the aluminum salt in the aluminum salt solution.Type: ApplicationFiled: October 21, 2022Publication date: February 27, 2025Applicant: GUILIN QIHONG TECHNOLOGY CO., LTD.Inventors: Shiqi WANG, Min LING, Huihong XIE
-
Patent number: 12235945Abstract: Two-dimensional face presentation attacks are one of most notorious and pervasive face spoofing types, causing security issues to facial authentication systems. To tackle these issues, a cost-effective face anti-spoofing (FAS) system based on acoustic modality, named as Echo-FAS, is devised, which employs a crafted acoustic signal to probe the presented face. First, a large-scale, high-diversity, acoustic-based FAS database, named as Echo-Spoof, is built. Based upon Echo-Spoof, we design a two-branch framework combining global and local frequency features of the presented face to distinguish live vs. spoofing faces. Echo-FAS has the following merits: (1) it only needs one speaker and one microphone; (2) it can capture three-dimensional geometrical information of the presented face and achieve a remarkable FAS performance; and (3) it can be handily allied with RGB-based FAS models to mitigate the overfitting problem in the RGB modality and make the FAS model more accurate and robust.Type: GrantFiled: November 21, 2022Date of Patent: February 25, 2025Assignee: City University of Hong KongInventors: Chenqi Kong, Kexin Zheng, Haoliang Li, Shiqi Wang
-
Patent number: 12234989Abstract: The present disclosure provides a central staged combustion chamber with self-excited sweeping oscillating fuel injection nozzles, including an outer housing with a cavity inside, a main stage coaxially disposed with the outer housing, and a pilot stage coaxially disposed with the outer housing. An annular fuel passage is used for connecting a plurality of self-excited sweeping oscillating fuel injection nozzles, and the self-excited sweeping oscillating fuel injection nozzles are suitable for injecting oscillating liquid fuel into a primary swirling passage. The fuel is output in a fan shape through each of the self-excited sweeping oscillating fuel injection nozzles and is dispersed by an incoming flow through the swirling passage, so that the atomization performance and spatial distribution uniformity of the fuel can be greatly improved.Type: GrantFiled: December 15, 2023Date of Patent: February 25, 2025Assignee: AERO ENGINE ACADEMY OF CHINAInventors: Shiqi Wang, Quan Wen, Xiao Han, Qian Yang
-
Publication number: 20250047864Abstract: A computer-implemented method for facilitating encoding of video data. The method includes performing an operation to determine prediction residuals associated with a unit of the video data, and, processing, using a neural network arrangement, the prediction residuals associated with the unit of the video data to determine model parameters associated with a rate-distortion model for the unit of the video data. The model parameters are arranged to facilitate encoding of at least the unit of the video data. The method can be applied to multiple ones of such unit of the video data.Type: ApplicationFiled: August 3, 2023Publication date: February 6, 2025Inventors: Sam Tak Wu Kwong, Yunhao Mao, Shiqi Wang
-
Publication number: 20250045970Abstract: A method for learning-based point cloud geometry compression includes: given a source point cloud, regressing an aligned mesh that is driven by a set of parameters from a deformable template mesh, quantizing the set of parameters into a parameter bitstream, generating an aligned point cloud from the quantized parameters by mesh manipulation and mesh-to-point-cloud conversion, extracting features from both the source point cloud and the aligned point cloud based on sparse tensors including coordinates and features, the coordinates being encoded into a coordinate bitstream, warping the features of the aligned point cloud onto the coordinates of the source point cloud, obtaining residual features through feature subtraction, processing the residual features using an entropy model into a residual feature bitstream, and obtaining a reconstructed point cloud by processing the parameter bitstream, the coordinate bitstream and the residual feature bitstream.Type: ApplicationFiled: August 1, 2023Publication date: February 6, 2025Inventors: Sam Tak Wu Kwong, Xinju Wu, Shiqi Wang
-
Publication number: 20240430463Abstract: There is provided a computer-implemented method for learned video compression, which includes processing a current frame (xt) and previously decoded frame ({circumflex over (x)}t?1) of a video data using a motion estimation model to estimate a motion vector (vt) for every pixel, compressing the motion vector (vt) and reconstructing the motion vector (vt) to a reconstructed motion vector ({circumflex over (v)}t), applying an enhanced context mining (ECM) model to obtain enhanced context ({umlaut over (C)}E) from the reconstructed motion vector ({circumflex over (v)}t) and previously decoded frame feature (x?t?1), compressing the current frame (xt) with the assistance of the enhanced context ({umlaut over (C)}E) to obtain a reconstructed frame ({circumflex over (x)}t?), and providing the reconstructed frame ({circumflex over (x)}t?) to a post-enhancement backend network to obtain a high-resolution frame ({circumflex over (x)}t).Type: ApplicationFiled: June 21, 2023Publication date: December 26, 2024Inventors: Sam Tak Wu Kwong, Haifeng Guo, Shiqi Wang, Dongjie Ye
-
Publication number: 20240428464Abstract: A method for compressing three-dimensional (3D) medical image. The method includes obtaining image data of a 3D medical image, performing a data conversion operation to convert the image data of the 3D medical image into video data of a sequence of frames each corresponding to a respective 2D image, and performing a video encoding operation to encode the video data of the sequence of frames to obtain encoded content data. The encoded content data can be used for reconstructing the 3D medical image.Type: ApplicationFiled: June 21, 2023Publication date: December 26, 2024Inventors: Sam Tak Wu Kwong, Xiangrui Liu, Meng Wang, Shiqi Wang
-
Publication number: 20240428927Abstract: A method for compressing a 3D medical image includes the steps of receiving a 3D medical image, partitioning the 3D medical image into a plurality of first slices, encoding the plurality of the first slices by a lossy codec into first bitstreams, decoding the first bitstreams by the lossy codec to obtain a plurality of second slices, computing a plurality of residues by comparing the plurality of the first slices and the plurality of the second slices, encoding the plurality of the residues by a lossless codec to obtain a plurality of encoded residues, and outputting the first bitstreams and the plurality of the encoded residues as compressed image data. Each residue corresponds to one of the first slices and its corresponding second slice. Experimental results on prevailing 3D medical image datasets demonstrate that the proposed method achieves promising compression performance and outperforms state-of-the-art methods.Type: ApplicationFiled: April 3, 2024Publication date: December 26, 2024Inventors: Sam Tak Wu KWONG, Xiangrui LIU, Shiqi WANG
-
Publication number: 20240388718Abstract: There is provided a computer-implemented method for processing a video. The computer-implemented method includes: (a) determining a target frame-level quality required for a frame of the video to be encoded, the determining of the target frame-level quality is based on, at least, a rate-quantization (R-Q) model that relates bit-rate and quantization step size and a quality-quantization model that relates quality measure and the quantization step size; and (b) determining one or more coding parameters for encoding the frame based on the determined target frame-level quality.Type: ApplicationFiled: April 30, 2024Publication date: November 21, 2024Inventors: Sam Tak Wu Kwong, Yunhao Mao, Shiqi Wang