Patents by Inventor Yongkang Fan

Yongkang Fan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11877084
    Abstract: A method may include obtaining a frame of a video stream of multiple video streams of a video conference, obtaining face detection information identifying a face size and a face position of at least one face detected in the frame, and cropping and scaling the frame according to at least one crop and scale parameter using the face detection information to obtain a modified first frame. The at least one crop and scale parameter is based on frames of the multiple video streams. The frames include the frame. The method may further include presenting the modified frame.
    Type: Grant
    Filed: June 29, 2020
    Date of Patent: January 16, 2024
    Assignee: Hewlett-Packard Development Company, L.P.
    Inventors: Hai Xu, Xi Lu, Yongkang Fan, Tiantian Pang, Matthew Xiang
  • Patent number: 11831984
    Abstract: A camera shooting program control method, the method is applied to terminal equipment, and the method includes obtaining motion parameters collected by at least one motion sensor in response to a situation that the terminal equipment enters a camera shooting program from a screen locking state, and determining a motion state of the terminal equipment according to the motion parameters collected by the at least one motion sensor; obtaining a mistaken touch rate of a display screen of the terminal equipment in response to the motion state of the terminal equipment being a moving state; and exiting the camera shooting program in response to a condition that the mistaken touch rate is greater than a preset first proportion threshold value.
    Type: Grant
    Filed: March 23, 2022
    Date of Patent: November 28, 2023
    Assignee: BEIJING XIAOMI MOBILE SOFTWARE CO., LTD.
    Inventors: Yongkang Fan, Suyue Liu, Xing Yu
  • Patent number: 11803984
    Abstract: A method (1000) for operating cameras (202) in a cascaded network (100), comprising: capturing a first view (1200) with a first lens (326) having a first focal point (328) and a first centroid (352), the first view (1200) depicting a subject (1106); capturing a second view (1202) with a second lens (326) having a second focal point (328) and a second centroid (352); detecting a first location of the subject (1106), relative the first lens (326), wherein detecting the first location of the subject (1106), relative the first lens (326), is based on audio captured by a plurality of microphones (204); estimating a second location of the subject (1106), relative the second lens (326), based on the first location of the subject (1106) relative the first lens (326); selecting a portion (1206) of the second view (1202) as depicting the subject (1106) based on the estimate of the second location of the subject (1106) relative the second lens (326).
    Type: Grant
    Filed: June 4, 2020
    Date of Patent: October 31, 2023
    Assignee: PLANTRONICS, INC.
    Inventors: Yongkang Fan, Hai Xu, Wenxue He, Hailin Song, Tianran Wang, Xi Lu
  • Patent number: 11805225
    Abstract: A method of tracking a presenter during a videoconference, comprising: detecting, in a data stream, data indicating a presence of one or more persons at a videoconferencing endpoint, the data stream comprising a plurality of frames; detecting tracker initiation data (e.g., that a person is actively talking, or that a person has raised their hand, or both) in the data stream; determining a subject of interest responsive to the tracker initiation data, and tracking the subject of interest as they move while making a presentation.
    Type: Grant
    Filed: June 10, 2020
    Date of Patent: October 31, 2023
    Assignee: PLANTRONICS, INC.
    Inventors: Hai Xu, Xi Lu, Yongkang Fan, Wenxue He, Hailin Song
  • Patent number: 11775834
    Abstract: A videoconferencing endpoint is described that uses a cascading sequence of convolutional neural networks to perform face detection and upper body detection of participants in a videoconference at the endpoint, where at least one member of the sequence of neural networks performs upper body detection, and where the final member of the sequence of neural networks performs face detection based on the results of the upper body detection. The models of the neural networks are trained on both large datasets of faces well as images that have been distorted by a wide-angle camera of the videoconferencing endpoint.
    Type: Grant
    Filed: November 22, 2018
    Date of Patent: October 3, 2023
    Assignee: Polycom, LLC
    Inventors: Hai Xu, Xi Lu, Yongkang Fan, Wenxue He
  • Publication number: 20230306618
    Abstract: A method may include identifying primly camera speaker coordinates of a location of a speaker relative to a primary camera, converting the primary camera speaker coordinates to secondary camera speaker coordinates, the secondary camera speaker coordinates being of the location and relative to a secondary camera, and adjusting, using the secondary camera speaker coordinates of the secondary camera, a view setting of the secondary camera to track the location of the speaker.
    Type: Application
    Filed: September 9, 2020
    Publication date: September 28, 2023
    Applicant: POLYCOM COMMUNICATIONS TECHNOLOGY (BEIJING) CO., LTD.
    Inventors: Yongkang Fan, Xi Lu, Hai Xu, Hailin Song, Wenxue He
  • Publication number: 20230245271
    Abstract: A real-time method (600) for enhancing facial images (102). Degraded images (102) of a person—such as might be transmitted during a videoconference—are rectified based on a single high definition reference image (604) of a person who is talking. Facial landmarks (501) are used to map (210) image data from the reference image (604) to an intervening image (622) having a landmark configuration like that in a degraded image (102). The degraded images (102) and their corresponding intervening images (622) are blended using an artificial neural network (800, 900) to produce high-quality images (108) of the person who is speaking during a videoconference.
    Type: Application
    Filed: July 6, 2020
    Publication date: August 3, 2023
    Inventors: Hailin Song, Hai Xu, Yongkang Fan, Tianran Wang, Xi Lu
  • Publication number: 20230186654
    Abstract: Systems and methods are provided for identifying and displaying whiteboard text and/or an active speaker in a video-based presentation, e.g., a video conference. Video images of an environment including a whiteboard may be captured by a video camera system. The video images may be analyzed to detect at least one text-containing area in the environment. Each text-containing area may be analyzed to determine whether it is an area of a whiteboard. When a text-containing area is identified as a whiteboard area, an area of view including the text-containing whiteboard area may be selected for display, e.g., a subset of the full frame captured by the video system. A video feed from the video camera system may be controlled to display the selected area of view at a client device, to provide a useful view of the whiteboard text and/or a speaking person located near the whiteboard text.
    Type: Application
    Filed: May 12, 2020
    Publication date: June 15, 2023
    Applicant: Polycom Communications Technology (Beijing) Co., Ltd.
    Inventors: Xi LU, Tianran WANG, Hailin SONG, Hai XU, Yongkang FAN
  • Publication number: 20230136314
    Abstract: A method may include calculating a color gain by applying an automatic white balance (AWB) algorithm to a video frame of a video feed, calculating an illumination color by applying a machine learning model to the video frame, transforming the illumination color into an equivalent color gain, determining that a difference between the color gain and the equivalent color gain exceeds a difference threshold, reversing an effect of the illumination color on the video frame based on the threshold being exceeded to obtain a corrected video frame, and transmitting the corrected video frame to an endpoint.
    Type: Application
    Filed: May 12, 2020
    Publication date: May 4, 2023
    Applicant: Polycom Communications Technology (Beijing) Co., Ltd.
    Inventors: Tianran WANG, Hai XU, Xingyue HUANG, Yongkang FAN, Wenxue HE
  • Publication number: 20230130129
    Abstract: A camera shooting program control method, the method is applied to terminal equipment, and the method includes obtaining motion parameters collected by at least one motion sensor in response to a situation that the terminal equipment enters a camera shooting program from a screen locking state, and determining a motion state of the terminal equipment according to the motion parameters collected by the at least one motion sensor; obtaining a mistaken touch rate of a display screen of the terminal equipment in response to the motion state of the terminal equipment being a moving state; and exiting the camera shooting program in response to a condition that the mistaken touch rate is greater than a preset first proportion threshold value.
    Type: Application
    Filed: March 23, 2022
    Publication date: April 27, 2023
    Applicant: Beijing Xiaomi Mobile Software Co., Ltd.
    Inventors: Yongkang FAN, Suyue LIU, Xing YU
  • Patent number: 11531426
    Abstract: An edge anti-false-touch method is provided. The method includes: obtaining, by a mobile terminal, a face direction of a user of a mobile terminal, and obtaining gesture data of the mobile terminal; determining, by the mobile terminal according to the face direction and the gesture data, a screen display state of the mobile terminal; and adjusting, by the mobile terminal based on the screen display state, a size of an anti-false-touch region of the mobile terminal to trigger an anti-false-touch response of the anti-false-touch region.
    Type: Grant
    Filed: March 23, 2022
    Date of Patent: December 20, 2022
    Assignee: Beijing Xiaomi Mobile Software Co., Ltd.
    Inventors: Xing Yu, Yongkang Fan, Suyue Liu
  • Publication number: 20220398864
    Abstract: A method performs zooming based on gesture detection. A visual stream is presented using a first zoom configuration for a zoom state. An attention gesture is detected from a set of first images from the visual stream. The zoom state is adjusted from the first zoom configuration to a second zoom configuration to zoom in on a person in response to detecting the attention gesture. The visual stream is presented using the second zoom configuration after adjusting the zoom state to the second zoom configuration. Whether the person is speaking is determined, from a set of second images from the visual stream. The zoom state is adjusted to the first zoom configuration to zoom out from the person in response to determining that the person is not speaking. The visual stream is presented using the first zoom configuration after adjusting the zoom state to the first zoom configuration.
    Type: Application
    Filed: September 24, 2019
    Publication date: December 15, 2022
    Inventors: Xi Lu, Tianran Wang, Hailin Song, Hai Xu, Yongkang Fan
  • Publication number: 20220329755
    Abstract: A method of tracking a presenter during a videoconference, comprising: detecting, in a data stream, data indicating a presence of one or more persons at a videoconferencing endpoint, the data stream comprising a plurality of frames; detecting tracker initiation data (e.g., that a person is actively talking, or that a person has raised their hand, or both) in the data stream; determining a subject of interest responsive to the tracker initiation data, and tracking the subject of interest as they move while making a presentation.
    Type: Application
    Filed: June 10, 2020
    Publication date: October 13, 2022
    Applicant: PLANTRONICS, INC.
    Inventors: HAI XU, XI LU, YONGKANG FAN, WENXUE HE, HAILIN SONG
  • Publication number: 20220318962
    Abstract: One illustrative method includes: (a) obtaining a video frame sequence having alternating fast and slow exposure frames; (b) applying a convolutional neural network twice to each frame in the video frame sequence, first when the frame is paired with a preceding frame, and again when the frame is paired with a subsequent frame, each time converting a pair of fast and slow exposure frames into an enhanced dynamic range video frame; and (c) outputting an enhanced video frame sequence. In another illustrative method, the convolutional neural network converts pairs of adjacent fast and slow exposure frames into corresponding pairs of enhanced dynamic range video frames. In yet another illustrative method, neighboring video frames for each given video frame are interpolated to form a fast and slow exposure frame pair, which the convolutional neural network converts into a corresponding enhanced dynamic range video frame.
    Type: Application
    Filed: June 29, 2020
    Publication date: October 6, 2022
    Applicant: PLANTRONICS, INC.
    Inventors: YONGKANG FAN, HAI XU, XI LU, HAILIN SONG
  • Publication number: 20220319034
    Abstract: A teleconferencing system (100) comprises: a first camera (202) including a first lens (326) having a first focal point (328) and a first centroid (352), and configured to capture a first view (900) corresponding to a subject (702); a second lens (326) having a second focal point (328) and a second centroid (352), and configured to capture a second view (902) corresponding to the subject (702); and a processor (206) coupled to the first camera device (202) and the second camera device (202). The processor (206) is configured to: estimate a first orientation (351) of the subject (702) relative the first lens (326) and a second orientation (351) of the subject relative the second lens (326); and determine that the first orientation (351) is more closely aligned with a first line (307) from the first centroid (352) to the first focal point (328) than is the second orientation (351) aligned with a second line (307) from the second centroid (352) to the second focal point (328).
    Type: Application
    Filed: June 4, 2020
    Publication date: October 6, 2022
    Applicant: PLANTRONICS, INC.
    Inventors: Yongkang Fan, HAI XU, Hailin Song, TIANRAN WANG, Xi Lu
  • Publication number: 20220319032
    Abstract: A method (1000) for operating cameras (202) in a cascaded network (100), comprising: capturing a first view (1200) with a first lens (326) having a first focal point (328) and a first centroid (352), the first view (1200) depicting a subject (1106); capturing a second view (1202) with a second lens (326) having a second focal point (328) and a second centroid (352); detecting a first location of the subject (1106), relative the first lens (326), wherein detecting the first location of the subject (1106), relative the first lens (326), is based on audio captured by a plurality of microphones (204); estimating a second location of the subject (1106), relative the second lens (326), based on the first location of the subject (1106) relative the first lens (326); selecting a portion (1206) of the second view (1202) as depicting the subject (1106) based on the estimate of the second location of the subject (1106) relative the second lens (326).
    Type: Application
    Filed: June 4, 2020
    Publication date: October 6, 2022
    Applicant: PLANTRONICS, INC.
    Inventors: YONGKANG FAN, HAI XU, WENXUE HE, HAILIN SONG, TIANRAN WANG, XI LU
  • Publication number: 20220303478
    Abstract: A method may include obtaining a frame of a video stream of multiple video streams of a video conference, obtaining face detection information identifying a face size and a face position of at least one face detected in the frame, and cropping and scaling the frame according to at least one crop and scale parameter using the face detection information to obtain a modified first frame. The at least one crop and scale parameter is based on frames of the multiple video streams. The frames include the frame. The method may further include presenting the modified frame.
    Type: Application
    Filed: June 29, 2020
    Publication date: September 22, 2022
    Applicant: Plantronics, Inc.
    Inventors: Hai Xu, Xi Lu, Yongkang Fan, Tiantian Pang, Matthew Xiang
  • Patent number: 11423550
    Abstract: A method for tracking movement of a presenter during a videoconference comprises tracking the location of subject of interest; quickly detecting error conditions—such as can occur when a subject of interest moves too quickly or when a subject of interest walks past another person—and quickly reinitiating the tracking sequence to help ensure a positive videoconferencing experience.
    Type: Grant
    Filed: June 17, 2020
    Date of Patent: August 23, 2022
    Assignee: PLANTRONICS, INC.
    Inventors: Xi Lu, Hai Xu, Yongkang Fan, Wenxue He, Hailin Song
  • Publication number: 20220180535
    Abstract: A method for tracking movement of a presenter during a videoconference comprises tracking the location of subject of interest; quickly detecting error conditions—such as can occur when a subject of interest moves too quickly or when a subject of interest walks past another person—and quickly reinitiating the tracking sequence to help ensure a positive videoconferencing experience.
    Type: Application
    Filed: June 17, 2020
    Publication date: June 9, 2022
    Applicant: PLANTRONICS, INC.
    Inventors: Xi LU, Hai Xu, Yongkang FAN, Wenxue HE, HAILIN SONG
  • Publication number: 20210409645
    Abstract: A videoconferencing endpoint is described that uses a cascading sequence of convolutional neural networks to perform face detection and upper body detection of participants in a videoconference at the endpoint, where at least one member of the sequence of neural networks performs upper body detection, and where the final member of the sequence of neural networks performs face detection based on the results of the upper body detection. The models of the neural networks are trained on both large datasets of faces well as images that have been distorted by a wide-angle camera of the videoconferencing endpoint.
    Type: Application
    Filed: November 22, 2018
    Publication date: December 30, 2021
    Inventors: Hai Xu, Xi Lu, Yongkang Fan, Wenxue He