Patents by Inventor Alan Bovik

Alan Bovik has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 12058340
    Abstract: A computer-implemented method, system and computer program product for compressing video. A set of video frames is partitioned into two subsets of different types of frames, a first type and a second type. The first type of frames of videos is compressed to generate a first representation by a first stage encoder. The first representation is then decoded to reconstruct the first type of frames using a first stage decoder. The second type of frames of video is compressed to generate a second representation that only contains soft edge information by a second stage encoder. A generative model corresponding to a second stage decoder is then trained using the first representation and the reconstructed first type of frames by using a discriminator employed by a machine learning system. After training the generative model, it generates reconstructed first and second types of frames using the soft edge information.
    Type: Grant
    Filed: June 14, 2022
    Date of Patent: August 6, 2024
    Assignee: Board of Regents, The University of Texas System
    Inventors: Alan Bovik, Sungsoo Kim, Jin Soo Park, Christos G. Bampis, Georgios Alex Dimakis
  • Publication number: 20220312017
    Abstract: A computer-implemented method, system and computer program product for compressing video. A set of video frames is partitioned into two subsets of different types of frames, a first type and a second type. The first type of frames of videos is compressed to generate a first representation by a first stage encoder. The first representation is then decoded to reconstruct the first type of frames using a first stage decoder. The second type of frames of video is compressed to generate a second representation that only contains soft edge information by a second stage encoder. A generative model corresponding to a second stage decoder is then trained using the first representation and the reconstructed first type of frames by using a discriminator employed by a machine learning system. After training the generative model, it generates reconstructed first and second types of frames using the soft edge information.
    Type: Application
    Filed: June 14, 2022
    Publication date: September 29, 2022
    Inventors: Alan Bovik, Sungsoo Kim, Jin Soo Park, Christos G. Bampis, Georgios Alex Dimakis
  • Patent number: 11388412
    Abstract: A computer-implemented method, system and computer program product for compressing video. A set of video frames is partitioned into two subsets of different types of frames, a first type and a second type. The first type of frames of videos is compressed to generate a first representation by a first stage encoder. The first representation is then decoded to reconstruct the first type of frames using a first stage decoder. The second type of frames of video is compressed to generate a second representation that only contains soft edge information by a second stage encoder. A generative model corresponding to a second stage decoder is then trained using the first representation and the reconstructed first type of frames by using a discriminator employed by a machine learning system. After training the generative model, it generates reconstructed first and second types of frames using the soft edge information.
    Type: Grant
    Filed: November 24, 2020
    Date of Patent: July 12, 2022
    Assignee: Board of Regents, The University of Texas System
    Inventors: Alan Bovik, Sungsoo Kim, Jin Soo Park, Christos G. Bampis, Georgios Alex Dimakis
  • Publication number: 20210160509
    Abstract: A computer-implemented method, system and computer program product for compressing video. A set of video frames is partitioned into two subsets of different types of frames, a first type and a second type. The first type of frames of videos is compressed to generate a first representation by a first stage encoder. The first representation is then decoded to reconstruct the first type of frames using a first stage decoder. The second type of frames of video is compressed to generate a second representation that only contains soft edge information by a second stage encoder. A generative model corresponding to a second stage decoder is then trained using the first representation and the reconstructed first type of frames by using a discriminator employed by a machine learning system. After training the generative model, it generates reconstructed first and second types of frames using the soft edge information.
    Type: Application
    Filed: November 24, 2020
    Publication date: May 27, 2021
    Inventors: Alan Bovik, Sungsoo Kim, Jin Soo Park, Christos G. Bampis, Georgios Alex Dimakis
  • Patent number: 10726532
    Abstract: A method, system and computer program product for measuring non-uniformity noise produced in images or videos (e.g., infrared images or videos). Images or videos, such as infrared images or videos, are captured. A model of scene statistics (statistical model of pictures, images or videos representative of pictures, images or videos, respectively, that are captured of the physical world) is utilized to measure the non-uniformity noise in the captured images or videos by exploiting exhibited characteristics for non-uniformity noise in the captured images or videos. A number signifying a magnitude of non-uniformity for each image or video frame is then generated. In this manner, non-uniformity noise produced in images or videos is measured.
    Type: Grant
    Filed: August 31, 2016
    Date of Patent: July 28, 2020
    Assignee: Board of Regents, The University of Texas System
    Inventors: Alan Bovik, Todd Goodall
  • Patent number: 10657378
    Abstract: A method, system and computer program product for classifying an image or video. An image or video to be classified is received. Scene statistics (statistical model of pictures, images or videos representative of pictures, images or videos, respectively, that are captured of the physical world) of the image or video are captured. A model (a statistical model that describes a set of probability distributions) of the image or video is then created using the captured scene statistics. A comparison between the model of the image or video with two other models of images or videos is performed, such as a model of visible light images or videos and a model of infrared images or videos. The received image or video is then classified (e.g., classified as corresponding to a visible light image) based on the comparison.
    Type: Grant
    Filed: August 31, 2016
    Date of Patent: May 19, 2020
    Assignee: Board of Regents, The University of Texas System
    Inventors: Alan Bovik, Todd Goodall
  • Patent number: 10529066
    Abstract: A method, system and computer program product for assessing quality of images or videos. A quality assessment of an image or video to be processed is performed using a no-reference reference quality assessment algorithm. A quality measurement, such as a score, reflecting the quality of the image or video, is generated from the no-reference reference quality assessment algorithm. The image or video is then processed and a quality assessment of the processed image or video is performed using a reference quality assessment algorithm that is conditional on the quality measurement provided by the no-reference quality assessment algorithm. In this manner, a more accurate quality measurement of the image or video is provided by the reference quality assessment algorithm.
    Type: Grant
    Filed: March 8, 2018
    Date of Patent: January 7, 2020
    Assignee: Board of Regents, The University of Texas Systems
    Inventor: Alan Bovik
  • Patent number: 10497396
    Abstract: A method, system and computer program product for improving the recording of classroom lectures or other such presentations. A video frame containing a whiteboard image is converted into a black and white image for the detection of boundaries. These boundaries are classified as horizontal or vertical lines. Quadrangles are then formed using spatial arrangements of these lines. The quadrangles that are most likely to spatially coincide with the boundaries of the whiteboard image are identified. The quadrangles are then sorted (ranked) based on specific characteristics, such as size and position. The area corresponding to the identified quadrangle in the video frame is then cropped. Furthermore, the speaker in the video frame can be removed based on detecting changes that are characteristic of movements of a speaker. In this manner, the visual and educational experience involved in the recording of classroom lectures or other such presentation is improved.
    Type: Grant
    Filed: August 31, 2016
    Date of Patent: December 3, 2019
    Assignee: Board of Regents, The University of Texas System
    Inventors: Alan Bovik, Casen Hunger, Zichong Li, Mitchell Crooks, Mark Meserve, Edward Mora, Mike Webb
  • Publication number: 20190043184
    Abstract: A method, system and computer program product for measuring non-uniformity noise produced in images or videos (e.g., infrared images or videos). Images or videos, such as infrared images or videos, are captured. A model of scene statistics (statistical model of pictures, images or videos representative of pictures, images or videos, respectively, that are captured of the physical world) is utilized to measure the non-uniformity noise in the captured images or videos by exploiting exhibited characteristics for non-uniformity noise in the captured images or videos. A number signifying a magnitude of non-uniformity for each image or video frame is then generated. In this manner, non-uniformity noise produced in images or videos is measured.
    Type: Application
    Filed: August 31, 2016
    Publication date: February 7, 2019
    Inventors: Alan Bovik, Todd Goodall
  • Patent number: 10182097
    Abstract: A method, system and computer program product for predicting a viewer's quality of experience while watching mobile videos potentially afflicted with network induced impairments. The length of a stall on a video at time t is received as a first input to a model. The number of stalls up to the time t is received as a second input to the model. Furthermore, the time since a preceding rebuffering event is received as a third input to the model. Additionally, a reciprocal stall density at time t is received as a fourth input to the model. The hysteresis effect is captured using a machine-learning-based model with an input that is an aggregate of the outputs of the first, second, third and fourth inputs to nonlinear input blocks of the model, where the hysteresis effect represents an effect that a viewer's recent level of satisfaction/dissatisfaction has on their overall viewing experience.
    Type: Grant
    Filed: September 12, 2016
    Date of Patent: January 15, 2019
    Assignee: Board of Regents, The University of Texas System
    Inventors: Alan Bovik, Deepti Ghadiyaram, Janice Pan
  • Publication number: 20180286032
    Abstract: A method, system and computer program product for assessing quality of images or videos. A quality assessment of an image or video to be processed is performed using a no-reference reference quality assessment algorithm. A quality measurement, such as a score, reflecting the quality of the image or video, is generated from the no-reference reference quality assessment algorithm. The image or video is then processed and a quality assessment of the processed image or video is performed using a reference quality assessment algorithm that is conditional on the quality measurement provided by the no-reference quality assessment algorithm. In this manner, a more accurate quality measurement of the image or video is provided by the reference quality assessment algorithm.
    Type: Application
    Filed: March 8, 2018
    Publication date: October 4, 2018
    Inventor: Alan Bovik
  • Publication number: 20180268864
    Abstract: A method, system and computer program product for improving the recording of classroom lectures or other such presentations. A video frame containing a whiteboard image is converted into a black and white image for the detection of boundaries. These boundaries are classified as horizontal or vertical lines. Quadrangles are then formed using spatial arrangements of these lines. The quadrangles that are most likely to spatially coincide with the boundaries of the whiteboard image are identified. The quadrangles are then sorted (ranked) based on specific characteristics, such as size and position. The area corresponding to the identified quadrangle in the video frame is then cropped. Furthermore, the speaker in the video frame can be removed based on detecting changes that are characteristic of movements of a speaker. In this manner, the visual and educational experience involved in the recording of classroom lectures or other such presentation is improved.
    Type: Application
    Filed: August 31, 2016
    Publication date: September 20, 2018
    Inventors: Alan Bovik, Casen Hunger, Zichong Li, Mitchell Crooks, Mark Meserve, Edward Mora, Mike Webb
  • Publication number: 20180247127
    Abstract: A method, system and computer program product for classifying an image or video. An image or video to be classified is received. Scene statistics (statistical model of pictures, images or videos representative of pictures, images or videos, respectively, that are captured of the physical world) of the image or video are captured. A model (a statistical model that describes a set of probability distributions) of the image or video is then created using the captured scene statistics. A comparison between the model of the image or video with two other models of images or videos is performed, such as a model of visible light images or videos and a model of infrared images or videos. The received image or video is then classified (e.g., classified as corresponding to a visible light image) based on the comparison.
    Type: Application
    Filed: August 31, 2016
    Publication date: August 30, 2018
    Inventors: Alan Bovik, Todd Goodall
  • Publication number: 20170085617
    Abstract: A method, system and computer program product for predicting a viewer's quality of experience while watching mobile videos potentially afflicted with network induced impairments. The length of a stall on a video at time t is received as a first input to a model. The number of stalls up to the time t is received as a second input to the model. Furthermore, the time since a preceding rebuffering event is received as a third input to the model. Additionally, a reciprocal stall density at time t is received as a fourth input to the model. The hysteresis effect is captured using a machine-learning-based model with an input that is an aggregate of the outputs of the first, second, third and fourth inputs to nonlinear input blocks of the model, where the hysteresis effect represents an effect that a viewer's recent level of satisfaction/dissatisfaction has on their overall viewing experience.
    Type: Application
    Filed: September 12, 2016
    Publication date: March 23, 2017
    Inventors: Alan Bovik, Deepti Ghadiyaram, Janice Pan
  • Patent number: 9595090
    Abstract: Techniques and structures are disclosed in which one or more distortion categories are identified for an image or video, and a quality of the image or video is determined based on the one or more distortion categories. The image or video may be of a natural scene, and may be of unknown provenance. Identifying a distortion category and/or determining a quality may be performed without any corresponding reference (e.g., undistorted) image or video. Identifying a distortion category may be performed using a distortion classifier. Quality may be determined with respect to a plurality of human opinion scores that correspond to a particular distortion category to which an image or video of unknown provenance is identified as belonging. Various statistical methods may be used in performing said identifying and said determining, including use of generalized Gaussian distribution density models and natural scene statistics.
    Type: Grant
    Filed: February 24, 2014
    Date of Patent: March 14, 2017
    Assignee: Board of Regents of The University of Texas System
    Inventors: Alan Bovik, Anush Moorthy
  • Patent number: 9277148
    Abstract: A method, system and computer program product for improving the perceptual quality and naturalness of an image captured by an image acquisition device (e.g., digital camera). Statistical features of a scene being imaged by the image acquisition device are derived from models of natural images. These statistical features are measured and mapped onto the control parameters (e.g., exposure, ISO) of the digital acquisition device. By mapping these statistical features onto the control parameters, the perceptual quality and naturalness of the scene being imaged may be based on the values of these control parameters. As a result, these control parameters are modified to maximize the perceptual quality and naturalness of the scene being imaged. After modification of these control parameters, the image is captured by the image acquisition device. In this manner, the perceptual quality and naturalness of the image captured by the image acquisition device is improved.
    Type: Grant
    Filed: June 3, 2013
    Date of Patent: March 1, 2016
    Assignee: Board of Regents, The University of Texas System
    Inventors: Alan Bovik, Anush Moorthy
  • Publication number: 20150116548
    Abstract: A method, system and computer program product for improving the perceptual quality and naturalness of an image captured by an image acquisition device (e.g., digital camera). Statistical features of a scene being imaged by the image acquisition device are derived from models of natural images. These statistical features are measured and mapped onto the control parameters (e.g., exposure, ISO) of the digital acquisition device. By mapping these statistical features onto the control parameters, the perceptual quality and naturalness of the scene being imaged may be based on the values of these control parameters. As a result, these control parameters are modified to maximize the perceptual quality and naturalness of the scene being imaged. After modification of these control parameters, the image is captured by the image acquisition device. In this manner, the perceptual quality and naturalness of the image captured by the image acquisition device is improved.
    Type: Application
    Filed: June 3, 2013
    Publication date: April 30, 2015
    Applicant: Board of Regents, The University of Texas System
    Inventors: Alan Bovik, Anush Moorthy
  • Publication number: 20140169682
    Abstract: Techniques and structures are disclosed in which one or more distortion categories are identified for an image or video, and a quality of the image or video is determined based on the one or more distortion categories. The image or video may be of a natural scene, and may be of unknown provenance. Identifying a distortion category and/or determining a quality may be performed without any corresponding reference (e.g., undistorted) image or video. Identifying a distortion category may be performed using a distortion classifier. Quality may be determined with respect to a plurality of human opinion scores that correspond to a particular distortion category to which an image or video of unknown provenance is identified as belonging. Various statistical methods may be used in performing said identifying and said determining, including use of generalized Gaussian distribution density models and natural scene statistics.
    Type: Application
    Filed: February 24, 2014
    Publication date: June 19, 2014
    Applicant: BOARD OF REGENTS OF THE UNIVERSITY OF TEXAS SYSTEM
    Inventors: Alan Bovik, Anush Moorthy
  • Patent number: 8660372
    Abstract: Techniques and structures are disclosed in which one or more distortion categories are identified for an image or video, and a quality of the image or video is determined based on the one or more distortion categories. The image or video may be of a natural scene, and may be of unknown provenance. Identifying a distortion category and/or determining a quality may be performed without any corresponding reference (e.g., undistorted) image or video. Identifying a distortion category may be performed using a distortion classifier. Quality may be determined with respect to a plurality of human opinion scores that correspond to a particular distortion category to which an image or video of unknown provenance is identified as belonging. Various statistical methods may be used in performing said identifying and said determining, including use of generalized Gaussian distribution density models and natural scene statistics.
    Type: Grant
    Filed: May 10, 2011
    Date of Patent: February 25, 2014
    Assignee: Board of Regents of the University of Texas System
    Inventors: Alan Bovik, Anush Moorthy
  • Publication number: 20110274361
    Abstract: Techniques and structures are disclosed in which one or more distortion categories are identified for an image or video, and a quality of the image or video is determined based on the one or more distortion categories. The image or video may be of a natural scene, and may be of unknown provenance. Identifying a distortion category and/or determining a quality may be performed without any corresponding reference (e.g., undistorted) image or video. Identifying a distortion category may be performed using a distortion classifier. Quality may be determined with respect to a plurality of human opinion scores that correspond to a particular distortion category to which an image or video of unknown provenance is identified as belonging. Various statistical methods may be used in performing said identifying and said determining, including use of generalized Gaussian distribution density models and natural scene statistics.
    Type: Application
    Filed: May 10, 2011
    Publication date: November 10, 2011
    Applicant: Board of Regents, The University of Texas System
    Inventors: Alan Bovik, Anush Moorthy