Patents Examined by Feng Niu
  • Patent number: 10842680
    Abstract: The present disclosure includes methods and systems to generate compression garment fit information from 3D images taken of one or more body parts of a patient indicated for application of compression therapy. The body parts can include legs or arms or other areas. Three-dimensional (ā€œ3Dā€) imaging information can be used to derive compression garment fit information, where the shape description information includes for each body part: tissue compressibility information, outer circumference information, and length information. Comparisons between corresponding body parts on a person can also be conducted.
    Type: Grant
    Filed: August 31, 2019
    Date of Patent: November 24, 2020
    Inventors: Michael J. Weiler, Nathan Daniel Frank
  • Patent number: 10839200
    Abstract: A method for facial analytics includes capturing a series of images of individuals from a camera into a circular buffer and selecting a plurality of images from the buffer for analysis in response to a trigger event, wherein the plurality of images are chronologically proximate before and/or after the trigger event in time. The method includes analyzing the plurality of images to determine image quality and selecting one of the plurality of images based on image quality to form a cropped facial image most likely to result in positive facial recognition matching. Methods of signaling to control the pedestrian traffic flow can maximize the individuals' facial alignment to the capturing camera's field of view. Non-relevant facial images associated with individuals outside a given region of interest can be discarded. Facial recognition is run on the resultant cropped facial image. Output can be displayed with information from the facial recognition.
    Type: Grant
    Filed: May 16, 2018
    Date of Patent: November 17, 2020
    Assignee: Gatekeeper Security, Inc.
    Inventors: Jonathan Nazemi, Christopher A. Millar, Robert Rozploch
  • Patent number: 10834313
    Abstract: A system and method for generating high resolution images using a plenoptic camera having a main lens in front of an array of microlenses and an image sensor, characterized in that it comprises: capturing a first set of images in a first unexcited state of operation by using a birefringent medium disposed between a said main lens and an said array of microlenses, said unexcited state of said birefringent medium providing an ordinary ray to each pixel; causing said first unexcited state to become a second excited state by applying a voltage across said birefringent medium; capturing a second set of images in said second excited state, said excited state of said birefringent medium splitting the light from said main lens into an ordinary ray and an extraordinary ray, said extraordinary ray being shifted by a distance of one half-pixel from the ordinary ray on said image sensor; subtracting pixel value associated with said first set of images from at least two times the pixel value associated with said second se
    Type: Grant
    Filed: June 27, 2017
    Date of Patent: November 10, 2020
    Assignee: INTERDIGITAL CE PATENT HOLDINGS
    Inventors: Nicolas Le Scouarnec, Arno Schubert, Valter Drazic
  • Patent number: 10824911
    Abstract: An algorithm for performing an image or video processing task is generated that may be used to combine a plurality of different independent solutions to the image or video processing task in an optimized manner. A plurality of base algorithms may be applied to a training set of images or video and a first generation of different combining algorithms may be applied to combine the respective solutions from each of the respective base algorithms into respective combined solutions. The respective combined solutions may be evaluated to generate respective fitness scores representing measures of how well the plurality of different combining algorithms each perform the image or video processing task. The algorithms may be iteratively updated to generate an optimized combining algorithm that may be applied to an input image or video.
    Type: Grant
    Filed: May 7, 2018
    Date of Patent: November 3, 2020
    Assignee: GoPro, Inc.
    Inventors: Balineedu Chowdary Adsumilli, Ryan Lustig
  • Patent number: 10803291
    Abstract: There is provided an encoding and decoding method and an information recognition device using the same. A code block includes a center coding region and a peripheral coding region arranged around the center coding region. The encoding and decoding method uses the feature of at least one microdot included in the center coding region as codes. The encoding and decoding method uses the feature of at least one microdot included in the peripheral coding region as codes. The encoding and decoding method uses the relative feature between the center coding region and the peripheral coding region as codes. The information recognition device compares the read feature with pre-stored features to decode information such as position codes, object codes, parameter codes and control codes.
    Type: Grant
    Filed: February 2, 2018
    Date of Patent: October 13, 2020
    Assignee: PIXART IMAGING INC.
    Inventors: En-Feng Hsu, Shu-Sian Yang, Yi-Hsien Ko
  • Patent number: 10780861
    Abstract: A system includes a processor and a memory storing instructions executable by the processor to predict a path of a liquid droplet on a surface, and then, actuate one or more vehicle components based on the path.
    Type: Grant
    Filed: January 8, 2019
    Date of Patent: September 22, 2020
    Assignee: FORD GLOBAL TECHNOLOGIES, LLC
    Inventors: David Michael Herman, Ashwin Arunmozhi, Venkatesh Krishnan, Sunil Patil
  • Patent number: 10783400
    Abstract: The present disclosure relates to generating computer searchable text from digital images that depict documents utilizing an orientation neural network and/or text prediction neural network. For example, one or more embodiments detect digital images that depict documents, identify the orientation of the depicted documents, and generate computer searchable text from the depicted documents in the detected digital images. In particular, one or more embodiments train an orientation neural network to identify the orientation of a depicted document in a digital image. Additionally, one or more embodiments train a text prediction neural network to analyze a depicted document in a digital image to generate computer searchable text from the depicted document.
    Type: Grant
    Filed: December 18, 2018
    Date of Patent: September 22, 2020
    Assignee: DROPBOX, INC.
    Inventors: David J. Kriegman, Peter N. Belhumeur, Bradley Neuberg, Leonard Fink
  • Patent number: 10776646
    Abstract: An identification method includes: controlling at least one camera to acquire a face image and an eye image of a target object, wherein the eye image includes at least one of an iris feature or an eye-print feature; and identifying the target object based on the face image and the eye image.
    Type: Grant
    Filed: January 28, 2020
    Date of Patent: September 15, 2020
    Assignee: Alibaba Group Holding Limited
    Inventor: Liang Li
  • Patent number: 10771802
    Abstract: A method of color mapping a video signal represented in a first color volume from color mapping data to be applied on a video signal represented in a second color volume is disclosed. The method comprises: color mapping (14) said video signal represented in a first color volume from said first color volume into said second color volume in the case where said first and second color volumes are different; and color mapping (16) said color mapped video signal based on said color mapping data.
    Type: Grant
    Filed: June 20, 2019
    Date of Patent: September 8, 2020
    Assignee: InterDigital VC Holdings, Inc.
    Inventors: Pierre Andrivon, Sebastien Lasserre, Philippe Bordes
  • Patent number: 10748025
    Abstract: A method for a verification process that performs neighbor discovery for one or more feature points projected to an m-dimensional space (m is a natural number equal to or greater than 2), includes: acquiring a feature point group including one or more feature points projected to coordinate values of the m-dimensional space ordered in a coordinate value order on each of two or more coordinate axes that define the m-dimensional spacer (m is a natural number equal to or greater than 2); selecting a datum axis on which a comparison time number in neighbor discovery is small, the comparison time number being obtained by performing simulation of neighbor discovery.
    Type: Grant
    Filed: September 14, 2017
    Date of Patent: August 18, 2020
    Assignee: FUJITSU LIMITED
    Inventor: Takahiro Aoki
  • Patent number: 10740641
    Abstract: At least one apparatus recognizes a first object and a second object associated with the first object in a plurality of images, calculates a score for each of the plurality of images based on a result of the recognition of the first object and the second object, and selects an image concerning the first object from the plurality of images based on the score for each of the plurality of images.
    Type: Grant
    Filed: December 1, 2017
    Date of Patent: August 11, 2020
    Assignee: Canon Kabushiki Kaisha
    Inventors: Ryosuke Iguchi, Shinjiro Hori, Hiroyasu Kunieda, Masaaki Obayashi
  • Patent number: 10735826
    Abstract: An embodiment of a semiconductor package apparatus may include technology to store a block of self-contained data including one or more time sequential frames of image data from two or more camera positions, and access a portion of the image data based on a start location and offset information from a header of the block. Other embodiments are disclosed and claimed.
    Type: Grant
    Filed: December 20, 2017
    Date of Patent: August 4, 2020
    Assignee: Intel Corporation
    Inventors: Eyal Ruhm, Asaf J. Shenberg
  • Patent number: 10726539
    Abstract: An image processing apparatus having a generation unit configured to generate an aligned image by arranging a plurality of candidate images extracted from a reference image around a work inspection image extracted from an inspection target image; a unit configured to subject the aligned image to similar region extraction processing to represent a similarity between regions in the aligned image; a determination unit configured to select a candidate image and determine it as a work reference image based on the aligned image after being subjected to the similar region extraction processing; and a comparison unit to compare the work inspection image with the work reference image. The similar region extraction processing subjects each of a plurality of division regions obtained by dividing the aligned image based on predetermined division size and phase, to averaging processing, and then adds the results of the averaging processing that are obtained by varying at least one of the division size and phase.
    Type: Grant
    Filed: April 24, 2017
    Date of Patent: July 28, 2020
    Assignee: Canon Kabushiki Kaisha
    Inventors: Shinjiro Hori, Tetsuya Suwa, Tomokazu Ishikawa, Wakako Tanaka
  • Patent number: 10715803
    Abstract: Virtual boundary processing in adaptive loop filtering (ALF) requires that padded values be substituted for unavailable pixel rows outside the virtual boundaries. Methods and apparatus are provided for virtual boundary processing in ALF that allow the use of more actual pixel values for padding than in the prior art.
    Type: Grant
    Filed: March 20, 2017
    Date of Patent: July 14, 2020
    Assignee: TEXAS INSTRUMENTS INCORPORATED
    Inventor: Madhukar Budagavi
  • Patent number: 10713756
    Abstract: One aspect of the current disclosure provides a method of upscaling an image. The method includes: rendering an image, wherein the rendering includes generating color samples of the image at a first resolution and depth samples of the image at a second resolution, which is higher than the first resolution; and upscaling the image to an upscaled image at a third resolution, which is higher than the first resolution, using the color samples and the depth samples.
    Type: Grant
    Filed: May 1, 2018
    Date of Patent: July 14, 2020
    Assignee: Nvidia Corporation
    Inventors: Rouslan Dimitrov, Lei Yang, Chris Amsinck, Walter Donovan, Eric Lum, Rui Bastos
  • Patent number: 10706267
    Abstract: Methods, systems, and devices for object recognition are described. Generally, the described techniques provide for a compact and efficient convolutional neural network (CNN) model for facial recognition. The proposed techniques relate to a light model with a set of layers of convolution and one fully connected layer for feature representation. A new building block of for each convolution layer is proposed. A maximum feature map (MFM) operation may be employed to reduce channels (e.g., by combining two or more channels via maximum feature selection within the channels). Depth-wise separable convolution may be employed for computation reduction (e.g., reduction of convolution computation). Batch normalization may be applied to normalize the output of the convolution layers and the fully connected layer (e.g., to prevent overfitting). The described techniques provide a compact and efficient CNN model which can be used for efficient and effective face recognition.
    Type: Grant
    Filed: January 12, 2018
    Date of Patent: July 7, 2020
    Assignee: QUALCOMM Incorporated
    Inventors: Lei Wang, Ning Bi, Yingyong Qi
  • Patent number: 10685211
    Abstract: A head-mounted display, a display control method, and a program that facilitate a user to understand proximity between the user and an object around the user are provided. A display block (36) is arranged in front of the eyes of the user wearing a HMD (12). In accordance with proximity between the user and an object around the user, the HMD (12) controls the display block (36) so as to have the user visually recognize a forward direction of the display block (36).
    Type: Grant
    Filed: August 2, 2016
    Date of Patent: June 16, 2020
    Assignee: SONY INTERACTIVE ENTERTAINMENT INC.
    Inventors: Yuichiro Nakamura, Yasushi Okumura
  • Patent number: 10675955
    Abstract: Some implementations relate to determining whether glare is present in captured image(s) of an object (e.g., a photo) and/or to determining one or more attributes of any present glare. Some of those implementations further relate to adapting one or more parameters for a glare removal process based on whether the glare is determined to be present and/or based on one or more of the determined attributes of any glare determined to be present. Some additional and/or alternative implementations disclosed herein relate to correcting color of a flash image of an object (e.g., a photo). The flash image is based on one or more images captured by a camera of a client device with a flash component of the client device activated. In various implementations, correcting the color of the flash image is based on a determined color space of an ambient image of the object.
    Type: Grant
    Filed: November 14, 2017
    Date of Patent: June 9, 2020
    Assignee: Google LLC
    Inventors: Julia Winn, Abraham Stephens, Daniel Pettigrew, Aaron Maschinot, Ce Liu, Michael Krainin, Michael Rubinstein, Jingyu Cui
  • Patent number: 10667870
    Abstract: A method for guiding resection of local tissue from a patient includes generating at least one image of the patient, automatically determining a plurality of surgical guidance cues indicating three-dimensional spatial properties associated with the local tissue, and generating a visualization of the surgical guidance cues relative to the surface. A system for generating surgical guidance cues for resection of a local tissue from a patient includes a location module for processing at least one image of the patient to determine three-dimensional spatial properties of the local tissue, and a surgical cue generator for generating the surgical guidance cues based upon the three-dimensional spatial properties. A patient-specific locator form for guiding resection of local tissue from a patient includes a locator form surface matching surface of the patient, and a plurality of features indicating a plurality of surgical guidance cues, respectively.
    Type: Grant
    Filed: June 10, 2016
    Date of Patent: June 2, 2020
    Assignee: THE TRUSTEES OF DARTMOUTH COLLEGE
    Inventors: Venkataramanan Krishnaswamy, Richard J. Barth, Jr., Keith D. Paulsen
  • Patent number: 10653101
    Abstract: A system that includes a robotic arm, a laser, a memory, and a processor. The processor is configured to position the laser adjacent to a dairy livestock and to modify teat location information for one or more teats of the dairy livestock based on a robot position offset between the dairy livestock and the robotic arm. The processor is further configured to generate a teat position associated with an unknown teat based on a scan of the dairy livestock and to determine a position distances between the teat position and teats of the dairy livestock. The processor is further configured to identify a teat of the dairy livestock with the smallest position distance, to associate a teat identifier for the unknown teat with the identified teat, and to store the association in the memory.
    Type: Grant
    Filed: April 13, 2018
    Date of Patent: May 19, 2020
    Assignee: Technologies Holdings Corp.
    Inventors: Mark A. Foresman, Bradley J. Prevost, Marinus P. Van Aart