Patents by Inventor Zhihan Ying

Zhihan Ying has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11334376
    Abstract: A computer-implemented method of providing an emotion-aware reactive interface in an electronic device includes receiving an image of a user as an input and identifying a multi-modal non-verbal cue in the image. The method further includes interpreting the multi-modal non-verbal cue to determine a categorization and outputting a reactive interface event determined based on the categorization.
    Type: Grant
    Filed: November 8, 2018
    Date of Patent: May 17, 2022
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Jihee Kim, Abhijit Bendale, Zhihan Ying, Simon Gibbs, Lu Luo
  • Patent number: 10909371
    Abstract: A method includes retrieving, by a device, contextual information based on at least one of an image, the device, user context, or a combination thereof. At least one model is identified from multiple models based on the contextual information and at least one object recognized in an image based on at least one model. At least one icon is displayed at the device. The at least one icon being associated with at least one of an application, a service, or a combination thereof providing additional information.
    Type: Grant
    Filed: December 13, 2017
    Date of Patent: February 2, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Stanislaw Antol, Abhijit Bendale, Simon J. Gibbs, Won J. Jeon, Hyun Jae Kang, Jihee Kim, Bo Li, Anthony S. Liot, Lu Luo, Pranav K. Mistry, Zhihan Ying
  • Patent number: 10902262
    Abstract: One embodiment provides a method comprising classifying one or more objects present in an input comprising visual data by executing a first set of models associated with a domain on the input. Each model corresponds to an object category. Each model is trained to generate a visual classifier result relating to a corresponding object category in the input with an associated confidence value indicative of accuracy of the visual classifier result. The method further comprises aggregating a first set of visual classifier results based on confidence value associated with each visual classifier result of each model of the first set of models. At least one other model is selectable for execution on the input based on the aggregated first set of visual classifier results for additional classification of the objects. One or more visual classifier results are returned to an application running on an electronic device for display.
    Type: Grant
    Filed: January 19, 2018
    Date of Patent: January 26, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Stanislaw Antol, Abhijit Bendale, Simon J. Gibbs, Won J. Jeon, Hyun Jae Kang, Jihee Kim, Bo Li, Anthony S. Liot, Lu Luo, Pranav K. Mistry, Zhihan Ying
  • Publication number: 20190250934
    Abstract: A computer-implemented method of providing an emotion-aware reactive interface in an electronic device includes receiving an image of a user as an input and identifying a multi-modal non-verbal cue in the image. The method further includes interpreting the multi-modal non-verbal cue to determine a categorization and outputting a reactive interface event determined based on the categorization.
    Type: Application
    Filed: November 8, 2018
    Publication date: August 15, 2019
    Inventors: Jihee Kim, Abhijit Bendale, Zhihan Ying, Simon Gibbs, Lu Luo
  • Publication number: 20180204061
    Abstract: One embodiment provides a method comprising classifying one or more objects present in an input comprising visual data by executing a first set of models associated with a domain on the input. Each model corresponds to an object category. Each model is trained to generate a visual classifier result relating to a corresponding object category in the input with an associated confidence value indicative of accuracy of the visual classifier result. The method further comprises aggregating a first set of visual classifier results based on confidence value associated with each visual classifier result of each model of the first set of models. At least one other model is selectable for execution on the input based on the aggregated first set of visual classifier results for additional classification of the objects. One or more visual classifier results are returned to an application running on an electronic device for display.
    Type: Application
    Filed: January 19, 2018
    Publication date: July 19, 2018
    Inventors: Stanislaw Antol, Abhijit Bendale, Simon J. Gibbs, Won J. Jeon, Hyun Jae Kang, Jihee Kim, Bo Li, Anthony S. Liot, Lu Luo, Pranav K. Mistry, Zhihan Ying
  • Publication number: 20180204059
    Abstract: A method includes retrieving, by a device, contextual information based on at least one of an image, the device, user context, or a combination thereof. At least one model is identified from multiple models based on the contextual information and at least one object recognized in an image based on at least one model. At least one icon is displayed at the device. The at least one icon being associated with at least one of an application, a service, or a combination thereof providing additional information.
    Type: Application
    Filed: December 13, 2017
    Publication date: July 19, 2018
    Inventors: Stanislaw Antol, Abhijit Bendale, Simon J. Gibbs, Won J. Jeon, Hyun Jae Kang, Jihee Kim, Bo Li, Anthony S. Liot, Lu Luo, Pranav K. Mistry, Zhihan Ying