Quality driven image processing for ocular recognition system
A method and system include receiving an image to process for an image recognition system, determining a quality of the received image, and creating a point distribution model for an active shape model, wherein the point distribution model has a number of points defining an outline of the image, the number of points being determined as a function of the quality of the image. A further method includes selecting a target local appearance model for fitting a point as a function of the determined quality of the received image to determine the location of the point. Yet a further method includes matching the probe image to a plurality of target images using a quality driven cascade classifier.
Latest Honeywell International Inc. Patents:
This invention was made with Government support under Grant Number W911NF-10-C-0027 awarded by USARMY RDECOM. The United States Government has certain rights in the invention.
BACKGROUNDAn active shape model has been used for localization of shapes in images. The active shape model represents the shape by use of a linear statistical model. The object shape is typically represented by N points, and the model is learned from manually annotated data. The number of points representing the model is fixed during a training phase. The number of points remains the same while fitting the shape to an unseen image. Each point of the model can be assigned with a weight. This weight is either entered manually or automatically. However, the number of points remains unchanged. In an active shape model, more points increase the shape fitting performance in an unseen image. On the other hand, a large number of points also accounts for high computational costs.
Biometric identification has been done using the ocular region of faces. Information captured around the ocular region is complementary to iris when the iris acquisition is unconstrained and iris quality is low such as blurred, obscure or gazed. The ocular region contains rich skin micro-pattern texture information, such as pores, spots, wrinkles, flat areas, etc. This information helps to distinguish subjects among themselves. Non skin patterns (eyebrow, tear duct, eye fold), if available, are also relevant. The feature extraction methods (e.g. local binary pattern, histogram of oriented gradients, Gabor wavelets) used for biometry identification are typically used for one fixed resolution. The parameter setting for these methods is optimized for a given image resolution. The various features are fused and form a representation for a given ocular image.
SUMMARYA method includes receiving an image to process for an image recognition system, determining a quality of the received image, and creating a point distribution model for an active shape model, wherein the point distribution model has a number of points defining an outline of the image, the number of points being determined as a function of the quality of the image.
In a further embodiment, a method includes obtaining multiple local appearance models for an image shape as a function of varying image qualities, storing the multiple local appearance models in a gallery of targets, receiving an image to process for an image recognition system, determining a quality of the received image, and selecting a target local appearance model for fitting a point as a function of the determined quality of the received image to determine the location of the point.
Yet a further method includes extracting features from a probe image corresponding to an image of an unknown subject, and matching the probe image to a plurality of target images using a quality driven cascade classifier.
A system includes a localizer to localize an object in a probe image, an evaluator to receive the localized object and perform a quality evaluation, an aligner to align the object as a function of quality of the probe image and create an active shape model for object in the probe image, and a classifier that classifies the image as a function of quality of the probe image.
In the following description, reference is made to the accompanying drawings that form a part hereof, and in which is shown by way of illustration specific embodiments which may be practiced. These embodiments are described in sufficient detail to enable those skilled in the art to practice the embodiments, and it is to be understood that other embodiments may be utilized and that structural, logical and electrical changes may be made. The following description of example embodiments is, therefore, not to be taken in a limited sense, and the scope is defined by the appended claims.
The functions or algorithms described herein may be implemented in software or a combination of software and human implemented procedures in one embodiment. The software may consist of computer executable instructions stored on computer readable media such as memory or other type of storage devices. Further, such functions correspond to modules, which are software, hardware, firmware or any combination thereof. Multiple functions may be performed in one or more modules as desired, and the embodiments described are merely examples. The software may be executed on a digital signal processor, ASIC, microprocessor, or other type of processor operating on a computer system, such as a personal computer, server or other computer system.
The active shape model consists of a point distribution model (statistical model represented by mean shape plus shape variations expressed as the transformation matrix controlled by deformation vector) and local appearance model (e.g., a few pixels around each point represented by gradient intensities).
In one embodiment, a determination of the number of points in the shape model is determined based on the quality of the image. The quality in some embodiments is represented by the spatial resolution and spatial frequencies of image (e.g. size of the image, brightness, focus). A high quality image may lead to a point distribution model with fewer points than that used for a low quality image. More points in the point distribution model are used for low quality images.
In a further embodiment, a local appearance model is created during a training phase. In the fitting phase, the intensities around each point are searched and new point location is determined by best match to the local appearance model. When the local appearance model is created for focused images and actual image is blurred, the localization of a point is less precise than if the local appearance model is created for blurred images and compared to actual blur image. Creating various local appearance models and selecting them in a fitting phase based on the image quality helps to determine the best location of points in the image.
In one embodiment, training data is first prepared. Shapes are manually annotated by selecting points in the image (e.g., 20 points on an eye contour). Several point distribution models that vary in number of points representing the shape (e.g., 20, 15, 10 points) are then trained. Taking the available training images and applying the quality metrics (e.g., blur of the image and brightness). The training images are sorted to N-groups based on their quality. For each group of images the local appearance model is trained for each annotated point.
During the fitting phase, the quality of the image is first evaluated using the same quality metrics as in the training phase. Based on the quality, the point distribution model and local appearance model are selected. If the quality metrics are applied on each part of the image separately, the local appearance model selection may vary even among the points. A classical active shape model methodology is applied to fit the model.
In a further embodiment, feature extraction for image classification after training is performed from available image information. A processed image can contain either an ocular region, face, or other biometric sample. The extracted features from the targets images (enrolled in a database), and probe image (image of unknown subject) are matched using a quality driven cascade classifier. In one embodiment, an image is rescaled to several cascades of descending image resolutions. The cascades are processed from the lowest resolution up to the highest resolution available (either in target or probe image). In each cascade the most similar targets are selected and the rest is ignored. Thus the processing is fast and doesn't require matching of every target in each cascade.
In a further embodiment, quality driven matching is performed using weighting of cascades. The similarity scores of N-best matches (most similar target images to given probe) are weighted by quality of actual probe image. For high quality images, all cascades have equal weights. For the low quality image, the higher cascades are assigned with lower weight. The sum of all the weights is equal to one in one embodiment.
In one embodiment, local binary patterns are selected as the method for feature extraction. The parameter setting for feature extraction is tuned for each resolution. At the lowest resolution, the full image is processed. At higher resolutions, the most valuable patches are selected. Patch selection may be done experimentally in some embodiments. For computation of a similarity score between two images such as a target and a probe, a cosine distance is used.
In one embodiment, a target set may be formed from high quality images. The images may be preprocessed and features extracted at three different levels of resolution, such as 64×64, 128×128, and 256×256 pixels. Given a probe image first, the quality may be evaluated, and based on the evaluations, the features are extracted from different resolutions up to the available resolution of the probe image.
The matching of a given probe image is first conducted on the lowest resolution. The most similar targets are selected and matching on higher levels is conducted only for those most similar targets. The process is repeated until the highest resolution is reached. The final scores of the similarity between the probe and selected target images are a combination of score at each cascade multiplied by the probe quality at a given resolution.
An outline of an object recognition process is illustrated in flowchart form at 100 in
At 130 quality driven object shape alignment is performed. An active shape model is used to localize the object shape. This involves identifying or localizing a boundary of the object. The shape and image are then registered to normalize the size of the object. The object is normalized with respect of objects characteristics in database.
At 140 a fine quality evaluation is performed. Fine quality measures within the object shape are evaluated to provide a fine quality. At 150 a cascade classification is performed. The final matching of the object versus the objects in the database is performed as a function of the actual object qualities such as orientation in space, brightness, focus, etc., leading to identification of an object in the database, or a conclusion that a matching object is not in the database.
With a detected eye, the ocular region is standardized for size and position (with respect to the gallery database samples) at 110. This process is called image registration, and its goal is to transform the image (realign/register) so the key facial landmarks (e.g. eye centers or corners) are located on same position, while keeping the biometric characteristic unchanged (e.g. shape of the eye).
To register a new image, the facial landmark's positions should be precisely detected. Localization of facial landmarks is performed. Without proper localization of facial landmarks, the recognition process may fail. Recognition is more effective if the probe and gallery images are aligned by size and pose. The registration algorithm realigns the image to some degree of variation (limits in one embodiment are: eyebrow moved up/down, out of plane rotation of maximal ±15 degrees). The registration may not be entirely accurate for higher degrees of the pose variation.
In one embodiment, selected datasets may be registered using three cropping methods of automatic registration including using eye centers and a full-face image. Plane rotation is removed, and fixed eye centers located. Eye corners may also be used by removing plane rotation and fixing eye corner positions. A further cropping method uses the bottom eyelid and a point on the eyebrow. The bottom eyelid is fixed to parabola and the orbital distance.
The last method aligns the bottom eyelid to approximately same position. The upper eyelid has a lot of variation (from fully open to closed eye). Additionally, the eye region may be masked out. A set of images to be matched in the gallery database also has the eye region masked out.
The illumination of the images may be normalized in order to reduce the variability of the signal and highlight the local variations of the region. Two types of photometric normalization, self quotient image (SQI) and histogram equalization (HE) may be used.
Image quality assessment is typically used for vision system processing under wide image degradation. The goal of this assessment is to estimate degradations so the best discriminative features are selected for best matching. Quality factors with respect to biometric system performance may be considered. Two main factors influencing the quality are: variation of subject characteristics and conditions during the acquisition.
Quality factors related to the subject include static anatomical characteristics, such as hair (bang on forehead), glasses, permanent jewelry, makeup, and injuries or scars. Dynamic factors corresponding to the behavior of the subject include expression and head pose toward the camera.
In one embodiment, acquisition factors that affect the matching performance include both static and dynamic factors. Static factors relate to the device properties, and dynamic factors relate to environmental conditions. Factors include spatial resolution (eye size and distance between the eyes), spatial frequencies, (brightness, contrast, and focus), dynamic range of ocular region (bits of intensity per pixel), shadows (uneven lighting), and specular reflection (saturation). The selected quality metrics are rather simple and fast. They address: spatial resolution, spatial frequencies, dynamic range of ocular region, and specular reflection. The following quality scores are defined:
Ocular Region Size Score
Dynamic Range Score
Brightness Score
Saturation Score
Out of scope score
Each quality score si is normalized to the range 0,1. The total score S is computed as weighted average:
where Wi is a normalized weight for the different scores. Most of the quality metrics for ocular region build on the cumulative histogram of ocular image.
For the ocular region score, size is indicated by numbers of row and column pixels of cropped eye before registration (number of rows is equal to columns). The eye is cropped by one of the methods described above. If the full face is available, the size of cropped eye is defined by eye distance. In the case of an ocular region, the size is defined by eye width. The mapping between the ocular size and its score is as follows:
<32, 256> pixel0 1 score
Thus, any ocular image of 32 pixels or less has a score of 0, and an ocular image of 256 or more pixels has a score of 1. The values between are uniformly distributed.
A dynamic range score may also be calculated. Dynamic range defines intensity density of the image in terms of bits required to represent the actual intensity values (e.g. 8 bits provide 256=28 intensity values). In one embodiment, a grayscale face should be at least 7 bits of intensity variation (i.e. at least 128 unique values). For the ocular region, a limit lower is set with a minimum number of bits of at least 6 (i.e. at least 64 unique intensity values). The number of intensity values (dynamic range) may be determined from cumulative histograms that illustrate dynamic range score examples with various input images.
Brightness refers to the overall lightness or darkness of an image. Brightness can be investigated with a histogram of intensities. The histogram is viewed as a probability distribution function and is typically characterized by statistical moments (mean, variance, skewness, etc.). The Brightness score in one embodiment is derived from cumulative histogram of gray intensity values. A reference value is read at 50%, if the intensity value falls in a predetermined intensity interval the brightness score equals 1. The brightness score decreases for intensity values outside of the ideal interval (bigger or smaller) and score equals 0 at a values on either side of the intensity interval.
A saturation score for saturated image intensity may be derived from a cumulative histogram. The score interval is uniformly distributed across cumulative histogram in one embodiment between zero and one.
A block diagram of a cascade system is illustrated at 400 in
In one embodiment, a method 500 in
In one embodiment, image comprises an ocular region of a person. The quality of the received image may be determined as a function of a spatial resolution and spatial frequencies of the received image. The quality of the image includes at least one of size of the image, brightness of the image and focus of the image. The number of points may be inversely related to the quality of the image such that high quality images may be represented by fewer points than images of lower quality.
In
In one embodiment the local appearance model includes a point from a point distribution model of an active shape model. The point may include a few pixels around each point represented by gradient intensities. The quality may be a function of focus of the image, of brightness of the image, or both. In some embodiments, the selected target appearance model is of the same quality as the received image quality.
In still a further embodiment, a method 700 in
Matching the probe image to a plurality of target images is shown at a method 800 in
In one embodiment, for probe images that are higher in quality, all cascades are assigned equal weights at 820. This may result in higher quality cascades being assigned a lower weight when matching to a lower quality probe image. The sum of all the weights is equal to one in one embodiment.
As shown in
The system bus 923 can be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures. The system memory can also be referred to as simply the memory, and, in some embodiments, includes read-only memory (ROM) 924 and random-access memory (RAM) 925. A basic input/output system (BIOS) program 926, containing the basic routines that help to transfer information between elements within the computer 900, such as during start-up, may be stored in ROM 924. The computer 900 further includes a hard disk drive 927 for reading from and writing to a hard disk, not shown, a magnetic disk drive 928 for reading from or writing to a removable magnetic disk 929, and an optical disk drive 930 for reading from or writing to a removable optical disk 931 such as a CD ROM or other optical media.
The hard disk drive 927, magnetic disk drive 928, and optical disk drive 930 couple with a hard disk drive interface 932, a magnetic disk drive interface 933, and an optical disk drive interface 934, respectively. The drives and their associated computer-readable media provide non volatile storage of computer-readable instructions, data structures, program modules and other data for the computer 900. It should be appreciated by those skilled in the art that any type of computer-readable media which can store data that is accessible by a computer, such as magnetic cassettes, flash memory cards, digital video disks, Bernoulli cartridges, random access memories (RAMs), read only memories (ROMs), redundant arrays of independent disks (e.g., RAID storage devices) and the like, can be used in the exemplary operating environment.
A plurality of program modules can be stored on the hard disk, magnetic disk 929, optical disk 931, ROM 924, or RAM 925, including an operating system 935, one or more application programs 936, other program modules 937, and program data 938. Programming for implementing one or more processes or method described herein may be resident on any one or number of these computer-readable media.
A user may enter commands and information into computer 900 through input devices such as a keyboard 940 and pointing device 942. Other input devices (not shown) can include a microphone, joystick, game pad, satellite dish, scanner, or the like. These other input devices are often connected to the processing unit 921 through a serial port interface 946 that is coupled to the system bus 923, but can be connected by other interfaces, such as a parallel port, game port, or a universal serial bus (USB). A monitor 947 or other type of display device can also be connected to the system bus 923 via an interface, such as a video adapter 948. The monitor 947 can display a graphical user interface for the user. In addition to the monitor 947, computers typically include other peripheral output devices (not shown), such as speakers and printers.
The computer 900 may operate in a networked environment using logical connections to one or more remote computers or servers, such as remote computer 949. These logical connections are achieved by a communication device coupled to or a part of the computer 900; other types of communication devices may also be used. The remote computer 949 can be another computer, a server, a router, a network PC, a client, a peer device or other common network node, and typically includes many or all of the elements described above I/O relative to the computer 900, although only a memory storage device 950 has been illustrated. The logical connections depicted in
When used in a LAN-networking environment, the computer 900 is connected to the LAN 951 through a network interface or adapter 953, which is one type of communications device. In some embodiments, when used in a WAN-networking environment, the computer 900 typically includes a modem 954 (another type of communications device) or any other type of communications device, e.g., a wireless transceiver, for establishing communications over the wide-area network 952, such as the internet. The modem 954, which may be internal or external, is connected to the system bus 923 via the serial port interface 946. In a networked environment, program modules depicted relative to the computer 900 can be stored in the remote memory storage device 950 of remote computer, or server 949. It is appreciated that the network connections shown are exemplary and other means of, and communications devices for, establishing a communications link between the computers may be used including hybrid fiber-coax connections, T1-T3 lines, DSL's, OC-3 and/or OC-12, TCP/IP, microwave, wireless application protocol, and any other electronic media through any suitable switches, routers, outlets and power lines, as the same are known and understood by one of ordinary skill in the art.
Claims
1. A method comprising:
- receiving, at a processor, an ocular image to process for an ocular recognition system;
- determining, via the processor, a quality of the received image;
- selecting a local appearance model from multiple local appearance models based on different image quality; and
- creating, via the processor, a point distribution model from the local appearance model, stored on a machine readable storage device, for an active shape model, wherein the point distribution model has a number of points defining an outline of the image, the number of points being determined as a function of the quality of the image.
2. The method of claim 1 wherein the quality of the received image is determined as a function of a spatial resolution and spatial frequencies of the received image.
3. The method of claim 2 wherein the quality of the image includes at least one of size of the image, brightness of the image and focus of the image.
4. The method of claim 1 wherein the number of points of the point distribution model is inversely related to the quality of the image.
6394602 | May 28, 2002 | Morrison et al. |
6714665 | March 30, 2004 | Hanna et al. |
7194393 | March 20, 2007 | Wei et al. |
7689010 | March 30, 2010 | Canzler et al. |
8055046 | November 8, 2011 | Feilkas et al. |
8295575 | October 23, 2012 | Feldman et al. |
8467627 | June 18, 2013 | Gwak et al. |
20060153430 | July 13, 2006 | Canzler et al. |
20070260157 | November 8, 2007 | Norrby |
20080170770 | July 17, 2008 | Suri et al. |
20090027558 | January 29, 2009 | Mantiuk et al. |
20100316263 | December 16, 2010 | Hamza |
20120321142 | December 20, 2012 | Trojanova et al. |
Type: Grant
Filed: Jun 15, 2011
Date of Patent: Nov 4, 2014
Patent Publication Number: 20120321142
Assignee: Honeywell International Inc. (Morristown, NJ)
Inventors: Jana Trojanova (Prague), Saad J. Bedros (West St. Paul, MN)
Primary Examiner: Gregory M Desire
Application Number: 13/161,069
International Classification: G06K 9/00 (20060101); G06K 9/62 (20060101); G06K 9/03 (20060101);