Patents by Inventor Liangrui Peng

Liangrui Peng has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 12080084
    Abstract: A method and a system for detecting a scene text may include extracting a first feature map for a scene image input based on a convolutional neural network, and delivering the first feature map to a sequential deformation module; obtaining sampled feature maps corresponding to sampling positions by performing iterative sampling for the first feature map, obtaining a second feature map by performing a concatenation operation in deep learning according to a channel dimension for the first feature map and the sampled feature maps; obtaining a third feature map by performing a feature aggregation operation for the second feature map in the channel dimension, and delivering the third feature map to the object detection baseline network; and performing text area candidate box extraction for the third feature map and obtaining a text area prediction result as a scene text detection result through regression fitting.
    Type: Grant
    Filed: August 20, 2021
    Date of Patent: September 3, 2024
    Assignees: TSINGHUA UNIVERSITY, HYUNDAI MOTOR COMPANY, KIA CORPORATION
    Inventors: Liangrui Peng, Shanyu Xiao, Ruijie Yan, Gang Yao, Shengjin Wang, Jaesik Min, Jong Ub Suk
  • Patent number: 11881038
    Abstract: A method of multi-directional scene text recognition based on multi-element attention mechanism include: performing normalization processing for a text row/column image output from an external text detection module by a feature extractor, extracting a feature for the normalized image by using a deep convolutional neural network to acquire an initial feature map, adding a 2-dimensional directional positional encoding P to the initial feature map in order to output a multi-channel feature map, converting the multi-channel feature map output from a feature extractor by an encoder into a hidden representation, and converting the hidden representation output from the encoder into a recognized text by a decoder and using the recognized text as the output result.
    Type: Grant
    Filed: October 15, 2021
    Date of Patent: January 23, 2024
    Assignees: TSINGHUA UNIVERSITY, HYUNDAI MOTOR COMPANY, KIA CORPORATION
    Inventors: Liangrui Peng, Ruijie Yan, Shanyu Xiao, Gang Yao, Shengjin Wang, Jaesik Min, Jong Ub Suk
  • Publication number: 20220121871
    Abstract: A method and a system of multi-directional scene text recognition based on multi-element attention mechanism are provided. The method includes: performing normalization processing for a text row/column image I output from an external text detection module by a feature extractor, extracting a feature for the normalized image by using a deep convolutional neural network to acquire an initial feature map F0, and adding a 2-dimensional directional positional encoding P to an initial feature map F0 in order to output a multi-channel feature map F; converting the multi-channel feature map F output from a feature extractor by an encoder into a hidden representation H; and converting the hidden representation H output from the encoder into a recognized text by a decoder and using the recognized text as the output result.
    Type: Application
    Filed: October 15, 2021
    Publication date: April 21, 2022
    Applicants: Tsinghua University, Hyundai Motor Company, Kia Corporation
    Inventors: Liangrui Peng, Ruijie Yan, Shanyu Xiao, Gang Yao, Shengjin Wang, Jaesik Min, Jong Ub Suk
  • Publication number: 20220058420
    Abstract: A method and a system for detecting a scene text may include extracting a first feature map for a scene image input based on a convolutional neural network, and delivering the first feature map to a sequential deformation module; obtaining sampled feature maps corresponding to sampling positions by performing iterative sampling for the first feature map, obtaining a second feature map by performing a concatenation operation in deep learning according to a channel dimension for the first feature map and the sampled feature maps; obtaining a third feature map by performing a feature aggregation operation for the second feature map in the channel dimension, and delivering the third feature map to the object detection baseline network; and performing text area candidate box extraction for the third feature map and obtaining a text area prediction result as a scene text detection result through regression fitting.
    Type: Application
    Filed: August 20, 2021
    Publication date: February 24, 2022
    Inventors: Liangrui PENG, Shanyu XIAO, Ruijie YAN, Gang YAO, Shengjin WANG, Jaesik MIN, Jong Ub SUK
  • Publication number: 20090135188
    Abstract: A method and a system of live detection based on a physiological motion on a human face are provided. The method has the following steps: in step a, a motion area and at least one motion direction in visual angle of a system camera are detected and a detected facial region is found. In step b, whether a valid facial motion exists in the detected facial region is determined. If a valid facial motion is inexistent, the object is considered as a photo of human face, otherwise, the method proceeds to step c to determine whether the facial motion is a physiological motion. If not, the object is considered as the photo of human face, yet considered as a real human face. The real human face and the photo of human face can be distinguished by the present invention so as to increase the reliability of the face recognition system.
    Type: Application
    Filed: May 30, 2008
    Publication date: May 28, 2009
    Applicant: TSINGHUA UNIVERSITY
    Inventors: Xiaoqing Ding, Liting Wang, Chi Fang, Changsong Liu, Liangrui Peng
  • Patent number: 7174044
    Abstract: In a method for character recognition based on Gabor filter group the Gabor filter's joint spatial/spatial-frequency localization and capability to efficiently extract characters' local structural features are employed to extract, from the character image, information of the stroke direction of characters as the recognition information of characters, so as to improve the capability to resist the noises, backgrounds, brightness variances in images and the deformation of characters. Using this information, a simple and effective parameter design method is put forward to optimally design the Gabor filter, ensuring a preferable recognition performance; a corrected Sigmoid function is used to non-linearly adaptively process the stroke direction information output from the Gabor filter group. When extracting the feature from blocks, Gaussian filter array is used to process the positive and negative values output from Gabor filter group to enhance the discrimination ability of the extracted features.
    Type: Grant
    Filed: May 23, 2003
    Date of Patent: February 6, 2007
    Assignee: Tsinghua University
    Inventors: Xiaoqing Ding, Xuewen Wang, Changsong Liu, Liangrui Peng, Chi Fang
  • Publication number: 20040017944
    Abstract: In a method for character recognition based on Gabor filter group the Gabor filter's joint spatial/spatial-frequency localization and capability to efficiently extract characters' local structural features are employed to extract, from the character image, information of the stroke direction of characters as the recognition information of characters, so as to improve the capability to resist the noises, backgrounds, brightness variances in images and the deformation of characters. Using this information, a simple and effective parameter design method is put forward to. optimally design the Gabor filter, ensuring a preferable recognition performance; a corrected Sigmoid function is used to non-linearly adaptively process the stroke direction information output from the Gabor filter group. When extracting the feature from blocks, Gaussian filter array is used to process the positive and negative values output from Gabor filter group to. enhance the discrimination ability of the extracted features.
    Type: Application
    Filed: May 23, 2003
    Publication date: January 29, 2004
    Inventors: Xiaoging Ding, Xuewen Wang, Changsong Liu, Liangrui Peng, Chi Fang