Patents by Inventor Haozhi Huang

Haozhi Huang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11776097
    Abstract: Methods, devices, and storage medium for fusing at least one image are disclosed. The method includes obtaining a first to-be-fused image and a second to-be-fused image, the first to-be-fused image comprising first regions, and the second to-be-fused image comprising second regions; obtaining a first feature set according to the first to-be-fused image and obtaining a second feature set according to the second to-be-fused image; performing first fusion processing on the first to-be-fused image and the second to-be-fused image by using a shape fusion network model to obtain a third to-be-fused image, the third to-be-fused image comprising at least one first encoding feature and at least one second encoding feature; and performing second fusion processing on the third to-be-fused image and the first to-be-fused image by using a condition fusion network model to obtain a target fused image. Model training methods, apparatus, and storage medium are also disclosed.
    Type: Grant
    Filed: June 2, 2021
    Date of Patent: October 3, 2023
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Haozhi Huang, Senzhe Xu, Shimin Hu, Wei Liu
  • Patent number: 11610082
    Abstract: A method, apparatus, and storage medium for training a neural network model used for image processing are described. The method includes: obtaining a plurality of video frames; inputting the plurality of video frames through a neural network model so that the neural network model outputs intermediate images; obtaining optical flow information between an early video frame and a later video frame; modifying an intermediate image corresponding to the early video frame according to the optical flow information to obtain an expected-intermediate image; determining a time loss between an intermediate image corresponding to the later video frame and the expected-intermediate image; determining a feature loss between the intermediate images and a target feature image; and training the neural network model according to the time loss and the feature loss, and returning to obtaining a plurality of video frames continue training until the neural network model satisfies a training finishing condition.
    Type: Grant
    Filed: February 26, 2021
    Date of Patent: March 21, 2023
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Haozhi Huang, Hao Wang, Wenhan Luo, Lin Ma, Peng Yang, Wenhao Jiang, Xiaolong Zhu, Wei Liu
  • Patent number: 11501574
    Abstract: In a multi-person pose recognition method, a to-be-recognized image is obtained, and a circuitous pyramid network is constructed. The circuitous network pyramid includes parallel phases, and each phase includes downsampling network layers, upsampling network layers, and a first residual connection layer to connect the downsampling and upsampling network layers. The phases are interconnected by a second residual connection layer. The circuitous pyramid network is traversed, by extracting a feature map for each phase, and the feature map of the last phase is determined to be the feature map of the to-be-recognized image. Multi-pose recognition is then performed on the to-be-recognized image according to the feature map to obtain a pose recognition result for the to-be-recognized image.
    Type: Grant
    Filed: October 19, 2020
    Date of Patent: November 15, 2022
    Assignee: Tencent Technology (Shenzhen) Company Limited
    Inventors: Haozhi Huang, Xinyu Gong, Jingmin Luo, Xiaolong Zhu, Wei Liu
  • Patent number: 11417095
    Abstract: An image recognition method is provided. The method includes obtaining predicted locations of joints of a target person in a to-be-recognized image based on a joint prediction model, where the joint prediction model is pre-constructed by: obtaining a plurality of sample images; inputting training features of the sample images and a body model feature to a neural network and obtaining predicted locations of joints in the sample images outputted by the neural network; updating a body extraction parameter and an alignment parameter; and inputting the training features of the sample images and the body model feature to the neural network to obtain the joint prediction model.
    Type: Grant
    Filed: November 15, 2019
    Date of Patent: August 16, 2022
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Xiaolong Zhu, Kaining Huang, Jingmin Luo, Lijian Mei, Shenghui Huang, Yongsen Zheng, Yitong Wang, Haozhi Huang
  • Publication number: 20220223182
    Abstract: A video sound-picture matching includes: acquiring a voice sequence; acquiring a voice segment from the voice sequence; acquiring an initial position of a start-stop mark and a moving direction of the start-stop mark from an image sequence; determining an active segment according to the initial position of the start-stop mark, the moving direction of the start-stop mark, and the voice segment; and synthesizing the voice segment and the active segment to obtain a video segment. In a video synthesizing process, the present disclosure uses start-stop marks to locate positions of active segments in an image sequence, so as to match the active segments having actions with voice segments, so that the synthesized video segments are more in line with natural laws of a character during speaking, and have better authenticity.
    Type: Application
    Filed: April 1, 2022
    Publication date: July 14, 2022
    Inventors: Yonggen LING, Haozhi HUANG, Li SHEN
  • Patent number: 11356619
    Abstract: Embodiments of this application disclose methods, systems, and devices for video synthesis. In one aspect, a method comprises obtaining a plurality of frames corresponding to source image information of a first to-be-synthesized video, each frame of the source image information. The method also comprises obtaining a plurality of frames corresponding to target image information of a second to-be-synthesized video. For each frame of the plurality of frames corresponding to the target image information of the second to-be-synthesized video, the method comprises fusing a respective source image from the first to-be-synthesized video, a corresponding source motion key point, and a respective target motion key point corresponding to the frame using a pre-trained video synthesis model, and generating a respective output image in accordance with the fusing. The method further comprises repeating the fusing and the generating steps for the second to-be-synthesized video to produce a synthesized video.
    Type: Grant
    Filed: April 23, 2021
    Date of Patent: June 7, 2022
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Haozhi Huang, Kun Cheng, Chun Yuan, Wei Liu
  • Patent number: 11276207
    Abstract: An image processing method for a computer device. The method includes obtaining a to-be-processed image belonging to a first image category; inputting the to-be-processed image into a first stage image conversion model, to obtain a first intermediate image; and converting the first intermediate image into a second intermediate image through a second stage image conversion model. The method also includes determining a first weight matrix corresponding to the first intermediate image; determining a second weight matrix corresponding to the second intermediate image; and fusing the first intermediate image and the second intermediate image according to the corresponding first weight matrix and second weight matrix, to obtain a target image corresponding to the to-be-processed image and belonging to a second image category. A sum of the first weight matrix and the second weight matrix being a preset matrix.
    Type: Grant
    Filed: May 21, 2020
    Date of Patent: March 15, 2022
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Minjun Li, Haozhi Huang, Lin Ma, Wei Liu, Yugang Jiang
  • Publication number: 20220028031
    Abstract: An image processing method is provided. The method includes: encoding an input image based on an attention mechanism to obtain an encoding tensor set and an attention map set of the input image; obtaining an encoding result of the input image according to the encoding tensor set and the attention map set, the encoding result of the input image recording an identity feature of a human face in the input image; encoding an expression image to obtain an encoding result of the expression image, the encoding result of the expression image recording an expression feature of a human face in the expression image; and generating an output image according to the encoding result of the input image and the encoding result of the expression image, the output image having the identity feature of the input image and the expression feature of the expression image.
    Type: Application
    Filed: October 8, 2021
    Publication date: January 27, 2022
    Inventors: Tianyu SUN, Haozhi HUANG, Wei LIU
  • Patent number: 11200680
    Abstract: An image processing method and a related apparatus are provided. The method is applied to an image processing device, and includes: obtaining an original image, the original image including a foreground object; extracting a foreground region from the original image through a deep neural network; identifying pixels of the foreground object from the foreground region; forming a mask according to the pixels of the foreground object, the mask including mask values corresponding to the pixels of the foreground object; and extracting the foreground object from the original image according to the mask.
    Type: Grant
    Filed: November 1, 2019
    Date of Patent: December 14, 2021
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Xiaolong Zhu, Kaining Huang, Jingmin Luo, Lijian Mei, Shenghui Huang, Yongsen Zheng, Yitong Wang, Haozhi Huang
  • Publication number: 20210295483
    Abstract: Methods, devices, and storage medium for fusing at least one image are disclosed. The method includes obtaining a first to-be-fused image and a second to-be-fused image, the first to-be-fused image comprising first regions, and the second to-be-fused image comprising second regions; obtaining a first feature set according to the first to-be-fused image and obtaining a second feature set according to the second to-be-fused image; performing first fusion processing on the first to-be-fused image and the second to-be-fused image by using a shape fusion network model to obtain a third to-be-fused image, the third to-be-fused image comprising at least one first encoding feature and at least one second encoding feature; and performing second fusion processing on the third to-be-fused image and the first to-be-fused image by using a condition fusion network model to obtain a target fused image. Model training methods, apparatus, and storage medium are also disclosed.
    Type: Application
    Filed: June 2, 2021
    Publication date: September 23, 2021
    Applicant: Tencent Technology (Shenzhen) Company Limited
    Inventors: Haozhi HUANG, Senzhe XU, Shimin HU, Wei LIU
  • Publication number: 20210243383
    Abstract: Embodiments of this application disclose methods, systems, and devices for video synthesis. In one aspect, a method comprises obtaining a plurality of frames corresponding to source image information of a first to-be-synthesized video, each frame of the source image information. The method also comprises obtaining a plurality of frames corresponding to target image information of a second to-be-synthesized video. For each frame of the plurality of frames corresponding to the target image information of the second to-be-synthesized video, the method comprises fusing a respective source image from the first to-be-synthesized video, a corresponding source motion key point, and a respective target motion key point corresponding to the frame using a pre-trained video synthesis model, and generating a respective output image in accordance with the fusing. The method further comprises repeating the fusing and the generating steps for the second to-be-synthesized video to produce a synthesized video.
    Type: Application
    Filed: April 23, 2021
    Publication date: August 5, 2021
    Inventors: Haozhi HUANG, Kun CHENG, Chun YUAN, Wei LIU
  • Publication number: 20210182616
    Abstract: A method, apparatus, and storage medium for training a neural network model used for image processing are described. The method includes: obtaining a plurality of video frames; inputting the plurality of video frames through a neural network model so that the neural network model outputs intermediate images; obtaining optical flow information between an early video frame and a later video frame; modifying an intermediate image corresponding to the early video frame according to the optical flow information to obtain an expected-intermediate image; determining a time loss between an intermediate image corresponding to the later video frame and the expected-intermediate image; determining a feature loss between the intermediate images and a target feature image; and training the neural network model according to the time loss and the feature loss, and returning to obtaining a plurality of video frames continue training until the neural network model satisfies a training finishing condition.
    Type: Application
    Filed: February 26, 2021
    Publication date: June 17, 2021
    Applicant: Tencent Technology (Shenzhen) Company Limited
    Inventors: Haozhi HUANG, Hao WANG, Wenhan LUO, Lin MA, Peng YANG, Wenhao JIANG, Xiaolong ZHU, Wei LIU
  • Publication number: 20210152751
    Abstract: A model training method includes obtaining an image sample set and brief-prompt information; generating a content mask set according to the image sample set and the brief-prompt information; generating a to-be-trained image set according to the content mask set; obtaining, based on the image sample set and the to-be-trained image set, a predicted image set through a to-be-trained information synthesis model, the predicted image set comprising at least one predicted image, the predicted image being in correspondence to the image sample; and training, based on the predicted image set and the image sample set, the to-be-trained information synthesis model by using a target loss function, to obtain an information synthesis model.
    Type: Application
    Filed: December 1, 2020
    Publication date: May 20, 2021
    Inventors: Haozhi HUANG, Jiawei LI, Li SHEN, Yonggen LING, Wei LIU, Dong YU
  • Patent number: 10970600
    Abstract: A method, apparatus, and storage medium for training a neural network model used for image processing are described. The method includes: obtaining a plurality of video frames; inputting the plurality of video frames through a neural network model so that the neural network model outputs intermediate images; obtaining optical flow information between an early video frame and a later video frame; modifying an intermediate image corresponding to the early video frame according to the optical flow information to obtain an expected-intermediate image; determining a time loss between an intermediate image corresponding to the later video frame and the expected-intermediate image; determining a feature loss between the intermediate images and a target feature image; and training the neural network model according to the time loss and the feature loss, and returning to obtaining a plurality of video frames continue training until the neural network model satisfies a training finishing condition.
    Type: Grant
    Filed: April 2, 2019
    Date of Patent: April 6, 2021
    Assignee: TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED
    Inventors: Haozhi Huang, Hao Wang, Wenhan Luo, Lin Ma, Peng Yang, Wenhao Jiang, Xiaolong Zhu, Wei Liu
  • Publication number: 20210073527
    Abstract: In a multi-person pose recognition method, a to-be-recognized image is obtained, and a circuitous pyramid network is constructed. The circuitous network pyramid includes parallel phases, and each phase includes downsampling network layers, upsampling network layers, and a first residual connection layer to connect the downsampling and upsampling network layers. The phases are interconnected by a second residual connection layer. The circuitous pyramid network is traversed, by extracting a feature map for each phase, and the feature map of the last phase is determined to be the feature map of the to-be-recognized image. Multi-pose recognition is then performed on the to-be-recognized image according to the feature map to obtain a pose recognition result for the to-be-recognized image.
    Type: Application
    Filed: October 19, 2020
    Publication date: March 11, 2021
    Applicant: Tencent Technology (Shenzhen) Company Limited
    Inventors: Haozhi HUANG, Xinyu GONG, Jingmin LUO, Xiaolong ZHU, Wei LIU
  • Publication number: 20200286263
    Abstract: An image processing method for a computer device. The method includes obtaining a to-be-processed image belonging to a first image category; inputting the to-be-processed image into a first stage image conversion model, to obtain a first intermediate image; and converting the first intermediate image into a second intermediate image through a second stage image conversion model. The method also includes determining a first weight matrix corresponding to the first intermediate image; determining a second weight matrix corresponding to the second intermediate image; and fusing the first intermediate image and the second intermediate image according to the corresponding first weight matrix and second weight matrix, to obtain a target image corresponding to the to-be-processed image and belonging to a second image category. A sum of the first weight matrix and the second weight matrix being a preset matrix.
    Type: Application
    Filed: May 21, 2020
    Publication date: September 10, 2020
    Inventors: Minjun LI, Haozhi HUANG, Lin MA, Wei LIU, Yugang JIANG
  • Publication number: 20200089958
    Abstract: Embodiments of the present disclosure provide an image recognition method and apparatus, an electronic device, and a readable storage medium. The method for image recognition includes the steps of obtaining a to-be-recognized feature of a to-be-recognized image, the to-be-recognized image comprising a target person; obtaining a preset body model feature of a body frame image, the body model feature comprising locations of joints in the body frame image; inputting the to-be-recognized feature and the body model feature to a pre-constructed joint prediction model, the joint prediction model being obtained through training a neural network by using minimum respective differences between real locations of joints in a sample image and predicted locations of the corresponding joints as a training objective; and obtaining predicted locations of joints of the target person in the to-be-recognized image based on the joint prediction model.
    Type: Application
    Filed: November 15, 2019
    Publication date: March 19, 2020
    Inventors: Xiaolong ZHU, Kaining HUANG, Jingmin LUO, Lijian MEI, Shenghui HUANG, Yongsen ZHENG, Yitong WANG, Haozhi HUANG
  • Publication number: 20200082542
    Abstract: An image processing method and a related apparatus are provided. The method is applied to an image processing device, and includes: obtaining an original image, the original image including a foreground object; extracting a foreground region from the original image through a deep neural network; identifying pixels of the foreground object from the foreground region; forming a mask according to the pixels of the foreground object, the mask including mask values corresponding to the pixels of the foreground object; and extracting the foreground object from the original image according to the mask.
    Type: Application
    Filed: November 1, 2019
    Publication date: March 12, 2020
    Inventors: Xiaolong ZHU, Kaining HUANG, Jingmin LUO, Lijian MEI, Shenghui HUANG, Yongsen ZHENG, Yitong WANG, Haozhi HUANG
  • Publication number: 20190228264
    Abstract: A method, apparatus, and storage medium for training a neural network model used for image processing are described. The method includes: obtaining a plurality of video frames; inputting the plurality of video frames through a neural network model so that the neural network model outputs intermediate images; obtaining optical flow information between an early video frame and a later video frame; modifying an intermediate image corresponding to the early video frame according to the optical flow information to obtain an expected-intermediate image; determining a time loss between an intermediate image corresponding to the later video frame and the expected-intermediate image; determining a feature loss between the intermediate images and a target feature image; and training the neural network model according to the time loss and the feature loss, and returning to obtaining a plurality of video frames continue training until the neural network model satisfies a training finishing condition.
    Type: Application
    Filed: April 2, 2019
    Publication date: July 25, 2019
    Applicant: Tencent Technology (Shenzhen) Company Limited
    Inventors: Haozhi HUANG, Hao WANG, Wenhan LUO, Lin MA, Peng YANG, Wenhao JIANG, Xiaolong ZHU, Wei LIU
  • Publication number: 20140240311
    Abstract: Transitions between street view images are described. The described techniques include: obtaining an original street view image and a target street view image; constructing a three-dimensional model corresponding to the original street view image by three-dimensional modeling; obtaining matching pairs of feature points that are extracted from the original street view image and the target street view image, and simulating a virtual camera in the three-dimensional model according to matching pairs of feature points to capture street view image sequence; and switching from the original street view image to the target street view image according to the street view image sequence. Transition stability is thereby improved.
    Type: Application
    Filed: May 1, 2014
    Publication date: August 28, 2014
    Applicant: Tencent Technology (Shenzhen) Company Limited
    Inventors: Kun XU, Jianyu Wang, Baoli Li, Chengjun Li, Haozhi Huang