Patents by Inventor Juwei Lu

Juwei Lu has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11966516
    Abstract: Methods and systems for gesture-based control of a device are described. A virtual gesture-space is determined in a received input frame. The virtual gesture-space is associated with a primary user from a ranked user list of users. The received input frame is processed in only the virtual gesture-space, to detect and track a hand. Using a hand bounding box generated by detecting and tracking the hand, gesture classification is performed to determine a gesture input associated with the hand. A command input associated with the determined gesture input is processed. The device may be a smart television, a smart phone, a tablet, etc.
    Type: Grant
    Filed: May 30, 2022
    Date of Patent: April 23, 2024
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Juwei Lu, Sayem Mohammad Siam, Wei Zhou, Peng Dai, Xiaofei Wu, Songcen Xu
  • Patent number: 11954145
    Abstract: Methods, systems, and media for image searching are described. Images comprising one query image and a plurality of candidate images are received. For each candidate image, a first model similarity measure from an output of a first model configured for scene classification to perceive scenes in the images is determined. Further, for each candidate image of the plurality of candidate images, a second model similarity measure from the output of a second model configured for attribute classification to perceive attributes in the images is determined. For each candidate image of the plurality of candidate images, a similarity agglomerate index of a weighted aggregate of the first model similarity measure and the second model similarity measure is computed. The plurality of candidate images based on the respective similarity agglomerate index of each candidate image are ranked and a first ranked candidate images corresponding to the searched images are generated.
    Type: Grant
    Filed: June 22, 2021
    Date of Patent: April 9, 2024
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Varshanth Ravindra Rao, Md Ibrahim Khalil, Peng Dai, Juwei Lu
  • Patent number: 11914788
    Abstract: Methods and systems for gesture-based control of a device are described. A virtual gesture-space is determined in a received input frame. The virtual gesture-space is associated with a primary user from a ranked user list of users. The received input frame is processed in only the virtual gesture-space, to detect and track a hand. Using a hand bounding box generated by detecting and tracking the hand, gesture classification is performed to determine a gesture input associated with the hand. A command input associated with the determined gesture input is processed. The device may be a smart television, a smart phone, a tablet, etc.
    Type: Grant
    Filed: May 30, 2022
    Date of Patent: February 27, 2024
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Juwei Lu, Sayem Mohammad Siam, Wei Zhou, Peng Dai, Xiaofei Wu, Songcen Xu
  • Publication number: 20240054757
    Abstract: Systems and methods for temporal action localization of video data are described. A feature representation extracted from video data has a temporal dimension and a spatial dimension. The feature representation is self-aligned in the spatial dimension. Spatial multi-sampling is performed to obtain a plurality of sparse samples of the self-aligned representation along the spatial dimension, and the multi-sampled representation is fused with the self-aligned representation. Attention-based context information aggregation is applied on the fused representation to obtain a spatially refined representation. Local temporal information aggregation is applied on the self-aligned representation to obtain a temporally refined representation. Action localization is performed on a concatenation of the spatially refined representation and the temporally refined representation.
    Type: Application
    Filed: June 1, 2023
    Publication date: February 15, 2024
    Inventors: Yanhui GUO, Deepak SRIDHAR, Peng DAI, Juwei LU
  • Patent number: 11900260
    Abstract: Methods, devices and processor-readable media for an integrated teacher-student machine learning system. One or more teacher-student modules are trained as part of the teacher neural network training. Each student sub-network uses a portion of the teacher neural network to generate an intermediate feature map, then provides the intermediate feature map to a student sub-network to generate inferences. The student sub-network may use a feature enhancement block to map the intermediate feature map to a subsequent feature map. A compression block may be used to compress intermediate feature map data for transmission in some embodiments.
    Type: Grant
    Filed: March 5, 2020
    Date of Patent: February 13, 2024
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Deepak Sridhar, Juwei Lu
  • Patent number: 11902548
    Abstract: Systems, methods, and computer media of processing a video are disclosed. An example method may include: receiving a plurality of video frames of a video; generating a plurality of first input features based on the plurality of video frames; generating a plurality of second input features based on reversing a temporal order of the plurality of first input features; generating a first set of joint attention features based on the plurality of first input features; generating a second set of joint attention features based on the plurality of second input features; and concatenating the first set of joint attention features and the second set of joint attention features to generate a final set of joint attention features.
    Type: Grant
    Filed: March 16, 2021
    Date of Patent: February 13, 2024
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Deepak Sridhar, Niamul Quader, Srikanth Muralidharan, Yaoxin Li, Juwei Lu, Peng Dai
  • Publication number: 20230419733
    Abstract: Methods and devices are described for computer vision-based gesture detection. From a frame of image data, extracted locations of keypoints of a detected hand are obtained. The extracted locations are normalized to obtain normalized features. The normalized features are processed using a trained decision tree ensemble to generate a probability of a valid gesture for the detected hand. The generated probability is compared with a defined decision threshold to generate a binary classification to classify the detected hand as a valid gesture or invalid gesture.
    Type: Application
    Filed: June 22, 2022
    Publication date: December 28, 2023
    Inventors: Yannick VERDIE, Zi Hao YANG, Deepak SRIDHAR, Juwei LU
  • Patent number: 11809637
    Abstract: Methods, devices, and processor-readable media for adjusting the control-display gain of a gesture-controlled device are described. Adjusting the control-display gain may facilitate user interaction with content or UI elements rendered on a display screen of the gesture-controlled device. The control-display gain may be adjusted based on a property of how a mid-air dragging gesture is being performed by a user's hand. The property may be the location of the gesture, the orientation of the hand performing the gesture, or the velocity of the gesture. A hand that becomes stationary for a threshold time period while performing the dragging gesture may adjust the control-display gain to a different level. Control-display gain may be set to a different value based on the current velocity of the hand performing the gesture. The control-display gain levels may be selected from a continuous range of values or a set of discrete values. Devices for performing the methods are described.
    Type: Grant
    Filed: September 13, 2022
    Date of Patent: November 7, 2023
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Wei Li, Wei Zhou, Sachi Mizobuchi, Ghazaleh Saniee-Monfared, Juwei Lu, Taslim Arefin Khan, Rafael Veras Guimaraes
  • Patent number: 11810329
    Abstract: Methods and systems for determining a surface color of a target surface under an environment with an environmental light source. A plurality of images of the target surface are captured as the target surface is illuminated with a variable intensity, constant color light source and a constant intensity, constant color environmental light source, wherein the intensity of the light source on the target surface is varied by a known amount between the capturing of the images. A color feature tensor, independent of the environmental light source, is extracted from the image data, and used to infer a surface color of the target surface.
    Type: Grant
    Filed: November 19, 2020
    Date of Patent: November 7, 2023
    Assignee: Huawei Technologies Co., Ltd.
    Inventors: Yuanhao Yu, Shuhao Li, Juwei Lu, Jin Tang
  • Publication number: 20230350499
    Abstract: Methods and devices for machine vision-based selection of content are described. One or more hands are detected in a current frame of video data. A respective fingertip location is determined for each of up to two of the detected hands. A content selection gesture is determined corresponding to the up to two detected hands. Selected content is extracted, as indicated by the content selection gesture and based on the up to two fingertip locations. The device may be a smartphone, a tablet, a laptop, a smart light device, a reader device, etc.
    Type: Application
    Filed: June 30, 2023
    Publication date: November 2, 2023
    Inventors: Juwei LU, Sayem Mohammad SIAM, Deepak SRIDHAR, Sidharth SINGLA, Yannick VERDIE, Xiaofei WU, Srikanth MURALIDHARAN, Roy YANG, Peng DAI, Songcen XU
  • Patent number: 11778223
    Abstract: A method, device and computer-readable medium for generating a super-resolution version of a compressed video stream. By leveraging the motion information and residual information in compressed video streams, described examples are able to skip the time-consuming motion-estimation step for most frames and make the most use of the SR results of key frames. A key frame SR module generates SR versions of I-frames and other key frames of a compressed video stream using techniques similar to existing multi-frame approaches to VSR. A non-key frame SR module generates SR version of the non-key inter frames between these key frames by making use of motion information and residual information used to encode the inter frames in the compressed video stream.
    Type: Grant
    Filed: August 19, 2021
    Date of Patent: October 3, 2023
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Wentao Liu, Yuanhao Yu, Yang Wang, Juwei Lu, Xiaolin Wu, Jin Tang
  • Publication number: 20230281981
    Abstract: Method and devices for training a keypoint estimation network are described. In each training iteration, synthetic images are generated by a generator, each synthetic image being assigned respective assigned keypoints by the generator. Using a prior-iteration of the keypoint estimation network, a set of predicted keypoints is obtained for each synthetic image. Based on an error score between the predicted keypoints and the assigned keypoints, poor quality synthetic images are discarded. The remaining synthetic images, together with real world images, are used to train an updated keypoint estimation network. The performance of the updated keypoint estimation network is validated, and the training iterations are performed until a convergence criteria is satisfied.
    Type: Application
    Filed: May 11, 2023
    Publication date: September 7, 2023
    Inventors: Xin DING, Deepak SRIDHAR, Juwei LU, Sidharth SINGLA, Peng DAI, Xiaofei WU
  • Patent number: 11698926
    Abstract: Methods and systems are described for performing video retrieval together with video grounding. A word-based query for a video is and encoded into a query representation using a trained query encoder. One or more similar video representations are identified, from a plurality of video representations that are similar to the query representation. Each similar video representation represents a respective relevant video. A grounding is generated for each relevant video by forward propagating each respective similar video representation together with the query representation through a trained grounding module. The relevant videos or identifiers of the relevant videos are outputted together with the grounding generated for each relevant video.
    Type: Grant
    Filed: November 12, 2021
    Date of Patent: July 11, 2023
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Arnab Kumar Mondal, Deepak Sridhar, Niamul Quader, Juwei Lu, Peng Dai, Chao Xing
  • Patent number: 11669743
    Abstract: An adaptive action recognizer for video that performs multiscale spatiotemporal decomposition of video to generate lower complexity video. The adaptive action recognizer has a number of processing pathways, one for each level of video complexity with each processing pathway having a different computational cost. The adaptive action recognizer applies a decision making scheme that encourages using low average computational costs while retaining high accuracy.
    Type: Grant
    Filed: May 14, 2020
    Date of Patent: June 6, 2023
    Assignee: Huawei Technologies Co., Ltd.
    Inventors: Niamul Quader, Juwei Lu, Peng Dai, Wei Li
  • Publication number: 20230153352
    Abstract: Methods and systems are described for performing video retrieval together with video grounding. A word-based query for a video is and encoded into a query representation using a trained query encoder. One or more similar video representations are identified, from a plurality of video representations that are similar to the query representation. Each similar video representation represents a respective relevant video. A grounding is generated for each relevant video by forward propagating each respective similar video representation together with the query representation through a trained grounding module. The relevant videos or identifiers of the relevant videos are outputted together with the grounding generated for each relevant video.
    Type: Application
    Filed: November 12, 2021
    Publication date: May 18, 2023
    Inventors: Arnab Kumar MONDAL, Deepak SRIDHAR, Niamul QUADER, Juwei LU, Pen DAI, Chao XING
  • Patent number: 11636677
    Abstract: System and method of analyzing a video, comprising dividing the video into a set of successive basic units; generating semantic tags for the basic units using a set of hierarchical classifier nodes that comprise a parent classifier node and a plurality of child classifier nodes, wherein the basic units are each routed through selected child classifier nodes based on classification of the basic units by the parent classifier node; and generating a semantic topic for the video based on the semantic tags generated for the basic units.
    Type: Grant
    Filed: January 8, 2021
    Date of Patent: April 25, 2023
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Varshanth Ravindra Rao, Peng Dai, Hanwen Liang, Md Ibrahim Khalil, Juwei Lu
  • Publication number: 20230082789
    Abstract: Methods and systems for gesture-based control of a device are described. An input frame is processed to determine a location of a distinguishing anatomical feature in the input frame. A virtual gesture-space is defined based on the location of the distinguishing anatomical feature, the virtual gesture-space being a defined space for detecting a gesture input. The input frame is processed in only the virtual gesture-space, to detect and track a hand. Using information generated from detecting and tracking the at least one hand, a gesture class is determined for the at least one hand. The device may be a smart television, a smart phone, a tablet, etc.
    Type: Application
    Filed: September 22, 2022
    Publication date: March 16, 2023
    Inventors: Juwei LU, Sayem Mohammad SIAM, Wei ZHOU, Peng DAI, Xiaofei WU, Songcen XU
  • Publication number: 20230072445
    Abstract: This disclosure provides a training method and apparatus, and relates to the artificial intelligence field. The method includes feeding a primary video segment, representative of a concatenation of a first and a second nonadjacent video segments obtained from a video source, to a deep learning backbone network. The method further includes embedding, via the deep learning backbone network, the primary video segment into a first feature output. The method further includes providing the first feature output to a first perception network to generate a first set of probability distribution outputs indicating a temporal location of a discontinuous point associated with the primary video segment. The method further includes generating a first loss function based on the first set of probability distribution outputs. The method further includes optimizing the deep learning backbone network, by backpropagation of the first loss function.
    Type: Application
    Filed: September 7, 2021
    Publication date: March 9, 2023
    Applicant: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Hanwen LIANG, Peng DAI, Zhixiang CHI, Lizhe CHEN, Juwei LU
  • Publication number: 20230057261
    Abstract: A method, device and computer-readable medium for generating a super-resolution version of a compressed video stream. By leveraging the motion information and residual information in compressed video streams, described examples are able to skip the time-consuming motion-estimation step for most frames and make the most use of the SR results of key frames. A key frame SR module generates SR versions of I-frames and other key frames of a compressed video stream using techniques similar to existing multi-frame approaches to VSR. A non-key frame SR module generates SR version of the non-key inter frames between these key frames by making use of motion information and residual information used to encode the inter frames in the compressed video stream.
    Type: Application
    Filed: August 19, 2021
    Publication date: February 23, 2023
    Inventors: Wentao LIU, Yuanhao YU, Yang WANG, Juwei LU, Xiaolin WU, Jin TANG
  • Publication number: 20230013169
    Abstract: Methods, devices, and processor-readable media for adjusting the control-display gain of a gesture-controlled device are described. Adjusting the control-display gain may facilitate user interaction with content or UI elements rendered on a display screen of the gesture-controlled device. The control-display gain may be adjusted based on a property of how a mid-air dragging gesture is being performed by a user's hand. The property may be the location of the gesture, the orientation of the hand performing the gesture, or the velocity of the gesture. A hand that becomes stationary for a threshold time period while performing the dragging gesture may adjust the control-display gain to a different level. Control-display gain may be set to a different value based on the current velocity of the hand performing the gesture. The control-display gain levels may be selected from a continuous range of values or a set of discrete values. Devices for performing the methods are described.
    Type: Application
    Filed: September 13, 2022
    Publication date: January 19, 2023
    Inventors: Wei LI, Wei ZHOU, Sachi MIZOBUCHI, Ghazaleh SANIEE-MONFARED, Juwei LU, Taslim Arefin KHAN, Rafael VERAS GUIMARAES