AUTO-FOCUS IMAGE SYSTEM
An auto-focus image system includes a pixel array and a focus signal generator coupled to the pixel array. The pixel array captures an image that has more than one edge. The generator generates a focus signal. Each edge has a width and contributes a quantity to compute the focus signal. The quantity may remain unchanged if all pixel values that enter a computation of the quantity are scaled up by a common multiplier. The quantity may have a dimension based solely on a dimension of length. The quantity may be the width. The generator eliminates an edge whose gradient profile has a first width at a first percentage height of the gradient profile that falls outside a tolerance region. The tolerance region varies with a second width of the gradient profile at a second percentage height. Alternatively, the edge's contribution to the focus signal is deemphasized.
This application is a continuation-in-part of International Patent Application No. PCT/IB2010/055649 filed on Dec. 7, 2010, which claims priority to U.S. Provisional Patent Application No. 61/267,436 filed on Dec. 7, 2009.
BACKGROUND OF THE INVENTION1. Field of the Invention
The subject matter disclosed generally relates to auto-focusing electronically captured images.
2. Background Information
Photographic equipment such as digital cameras and digital camcorders may contain electronic image sensors that capture light for processing into still or video images, respectively. Electronic image sensors typically contain millions of light capturing elements such as photodiodes.
Many image capturing devices such as cameras include an auto-focusing system. The process of auto-focusing includes the steps of capturing an image, processing the image to determine whether it is in focus, and if not, generating a feedback signal that is used to vary a position of a focus lens (“focus position”). There are two primary auto-focusing techniques. The first technique involves contrast measurement, the other technique looks at a phase difference between a pair of images. In the contrast method the intensity difference between adjacent pixels is analyzed and the focus is adjusted until a maximum contrast is detected. Although acceptable for still pictures the contrast technique is not suitable for motion video.
The phase difference method includes splitting an incoming image into two images that are captured by separate image sensors. The two images are compared to determine a phase difference. The focus position is adjusted until the two images match. The phase difference method requires additional parts such as a beam splitter and an extra image sensor. Additionally, the phase difference approach analyzes a relatively small band of fixed detection points. Having a small group of detection points is prone to error because noise may be superimposed onto one or more points. This technique is also ineffective if the detection points do not coincide with an image edge. Finally, because the phase difference method splits the light the amount of light that impinges on a light sensor is cut in half or even more. This can be problematic in dim settings where the image light intensity is already low.
BRIEF SUMMARY OF THE INVENTIONAn auto-focus image system includes a pixel array and a focus signal generator coupled to the pixel array. The pixel array captures an image that has more than one edge, each edge having a width. The generator generates a focus signal. Each one of a plurality of edges in the image contributes a quantity to compute the focus signal. The quantity of an edge may be such that it does not vary even if all pixel values that enter a computation of the quantity are scaled up by a common multiplier. The quantity may have a dimension based solely on a dimension of length. The quantity may be an edge width of the contributing edge. The generator eliminates an edge whose gradient profile has a first width at a first percentage height of the gradient profile that falls outside a tolerance region that varies with a second width of the gradient profile at a second percentage height of the gradient profile. Alternatively, the edge's contribution to the focus signal is deemphasized instead of being eliminated.
Disclosed is a method to generate a focus signal from a plurality of edges in an image to indicate a degree of image sharpness in the image, each edge contributing a quantity to the focus signal. The quantity may be an edge width of each edge. The quantity may have a dimension based solely on a dimension of length. The method comprises rejecting or deemphasizing, among a first plurality of edges whose first widths are same, any edge whose second width lies outside a first acceptance range. Each edge among the first plurality of edges has a gradient profile with a peak gradient, and the first width of each edge is a width of the gradient profile at a first gradient level that is a first fraction of the peak gradient. The second width of any edge is another width of the gradient profile at a second gradient level that is a second fraction of the peak gradient. The first and second fractions are common across the first plurality of edges. The rejecting or deemphasizing may involve executing computer-executable instructions.
Disclosed is also a non-transitory computer readable medium that comprises computer executable instructions and or parameter values for executing the method.
The method may further comprise rejecting or deemphasizing, among a second plurality of edges that all have a same third width across their respective gradient profiles at respective gradient levels that are at the first fraction of their respective peak gradients, any edge whose fourth width lies outside a second acceptance range. The third width and the second acceptance range are different from the first width and the first acceptance range respectively. The fourth width is a width across the gradient profile at a gradient level that is at the second fraction of the peak gradient.
The method may be performed in an Edge Detection & Width Measurement Unit housed in a package together with a pixel array. Alternatively, the Edge Detection & Width Measurement Unit may be housed in a package together with a processor that controls a focus position of a focus lens. Alternatively, the Edge Detection & Width Measurement Unit may be housed in a package that does not contain a processor that controls storing of compressed images to a removable memory card.
Another aspect of the invention is a method to generate a focus signal from a plurality of edges in an image to indicate a degree of image sharpness in the image, each edge contributing a quantity. The quantity of an edge may be such that it does not vary even if all pixel values that enter a computation of the quantity are scaled up by a common multiplier. The quantity may have a dimension based solely on a dimension of length. The quantity may be an edge width of the contributing edge. The method comprises rejecting or attenuating a contribution of the quantity of an edge towards generating the focus signal where a first parameter of a gradient profile of the edge measured at a first gradient level does not meet a predetermined criterion that varies with a second parameter of the gradient profile of the edge measured at a second gradient level.
Here, the first and second gradient levels are different gradient values defined as a first fraction and a second fraction, respectively, of a peak gradient value of the edge, and both the first and second parameters are determined by the gradient profile to both sides of a peak of the gradient profile. The predetermined criterion may depend on a slant of the edge, i.e. on a gradient in a second direction perpendicular to a first direction in which pixels of the edge are array.
The method may further comprise evaluating the first parameter and the second parameter to obtain the first and second values respectively and determining whether the first value meets the predetermined criterion given the second value.
In this method, the predetermined criterion may require the first value to be, to within a predetermined tolerance, the second value multiplied by a predetermined multiplier. The predetermined multiplier may be a function of an edge width of the edge. The edge width may be corrected for a slant of the edge for determining the predetermined criterion, i.e. being compensated for a non-zero gradient in a second direction perpendicular to a first direction in which pixels of the edge are arrayed. The predetermined multiplier may be interpolated from a sequence of multipliers, each multiplier for a different edge width.
In this method, the first parameter may be either a width of the gradient profile measured at the first gradient level or a count of pixels in the gradient profile that have gradient levels anywhere from the first gradient level to a peak gradient level of the gradient profile. The second parameter may be either a width of the gradient profile measured at the second gradient level or a count of pixels in the gradient profile that have gradient levels anywhere from the second gradient level to the peak gradient level.
In this method, the first and second gradient levels may be within a range from 10% to 90% of a peak gradient level of the gradient profile. Alternatively, the range may be from 15% to 85%. Alternatively, the range may be from 20% to 80%.
Disclosed is an auto focus image system that includes a pixel array coupled to a focus signal generator. The pixel array captures an image that has at least one edge with a width. The focus signal generator may generate a focus signal that is a function of the edge width and/or statistics of edge widths. An auto focus image system that includes a pixel array coupled to a focus signal generator. The pixel array captures an image that has at least one edge with a width. The generator generates a focus signal that is a function of the edge width and various statistics of edge width. The generator may eliminate an edge having an asymmetry of a gradient of an image signal. The generator may also eliminate an edge that fails a template for an associated peaking in the gradient. A processor receives the focus signal and/or the statistics of edge widths and adjusts a focus position of a focus lens. The edge width can be determined by various techniques including the use of gradients. A histogram of edge widths may be used to determine whether a particular image is focused or unfocused. A histogram with a large population of thin edge widths is indicative of a focused image.
ArchitectureReferring to the drawings more particularly by reference numbers,
The focus signal generator 120 receives a group of control signals 132 from the processor 112, in addition, and may output signals 134 to the processor 112. The output signals 134 may comprise one or more of the following: a focus signal 134, a narrow-edge count, and a set of numbers representing a statistics of edge width in the image. The processor 112 may generate a focus control signal 136 that is sent to the drive motor/circuit 118 to control the focus lens 104. A focused image is ultimately provided to the display 114 and/or stored in the memory card 116. The algorithm(s) used to adjust a focus position may be performed by the processor 112.
The pixel array and circuits 108, A/D Converter 110, focus signal generator 120, and processor 112 may all reside within a package. Alternately, the pixel array and circuits 108, A/D Converter 110, and focus signal generator 120 may reside within a package 142 as image sensor 150 shown in
The EDWM unit 206 may transform the input image such that the three signals of the image, red (R), green (G) and blue (B) are converted to a single image signal. Several techniques can be utilized to transform an image to a single image. RGB values can be used to calculate a luminance or chrominance value or a specific ratio of RGB values can be taken to form the single image signal. For example, the luminance value can be calculated with the equation Y=0.2126*R+0.7152*G+0.0722*B, where Y is luminance value. The single image signal may then be processed by a Gaussian filter or any lowpass filter to smooth out pixel signal values among neighboring pixels to remove a noise.
The focus signal generator 120, 120′, 120″ is not limited to grayscale signal. It may operate on any one image signal to detect one or more edges in the image signal. Or it may operate on any combination of the image signals, for example Y, R-G, or B-G. It may operate on each and every one of the R, G, B image signals separately, or any one or more combinations thereof, to detect edges. It may form statistics of edge widths for each of the R, G, B image signals, or any combination thereof. It may form a focus signal from statistics of edge widths from one or more image signals.
A gradient of the processed image is then calculated. There are various methods available to calculate the gradient, including Laplacian, and Sobel. Gradients across the columns and the rows may be calculated to detect vertical and horizontal edges respectively, for example using a Sobel-X operator and a Sobel-Y operator, respectively. Sobel X-operator at pixel location [k, q] where k is a row number and q is a column number, is given by the equation Sx[k, q]=U[k, q+1]−U[k, q−1]. Sobel Y-operator at the same location is given by the equation Sy[k,q]=U[k+1,q]−U[k−1,], where U is an image signal of the processed image.
Orientation TaggingEach pixel is tagged either a horizontal edge (‘H’) or a vertical edge (‘V’) if either vertical or horizontal gradient magnitude exceeds a predetermined lower limit (“elimination threshold”), e.g. 5 for an 8-bit image, or no edge if neither is true. This lower limit eliminates spurious edges due to gentle shading or noise. A pixel is tagged a vertical edge if its horizontal gradient magnitude exceeds its vertical gradient magnitude by a predetermined hysteresis amount or more, e.g. 2 for an 8-bit image, and vice versa. If both gradient magnitudes differ less than the hysteresis amount, the pixel gets a direction tag same as that of its nearest neighbor that has an direction tag already determined. For example, if the image is scanned from left to right in each row and from row to row downwards, a sequence of inspection of neighboring pixels may be the pixel above first, the pixel above left second, and the pixel on the left third, and the pixel above right last. Applying this hysteresis helps to ensure that adjacent pixels get similar tags if each of them has nearly identical horizontal and vertical gradient magnitudes.
The image, gradients and tags may be scanned horizontally for vertical edges, and vertically for horizontal edges. Each group of consecutive pixels in a same row, having a same horizontal gradient polarity and all tagged for vertical edge may be designated a vertical edge if no adjacent pixel on left or right of the group are likewise. Likewise, each group of consecutive pixels in a same column having a same vertical gradient polarity and all tagged for horizontal edge may be designated a horizontal edge if no adjacent pixel above or below the group satisfies the same. Thus horizontal and vertical edges may be identified.
Edge WidthEach edge may be refined by removing pixels whose gradient magnitudes are less than a given fraction of the peak gradient magnitude within the edge.
Edge width may be calculated in any one of known methods. One method of calculating edge width is simply counting the number of pixels within an edge. An alternate method of calculating edge width is shown in
Although each edge may be assigned to one prescribed direction (e.g. vertical direction or horizontal direction) or another, perpendicular, prescribed direction (e.g horizontal direction or vertical direction) and may have its edge width measured in a direction perpendicular to that assigned edge direction, the boundaries between regions of different image signal values in the image from which these edges arise may not be and usually are not aligned perfectly with either prescribed directions. In
For purposes of calculating a focus signal from edge widths, the edge widths measured in one or the other of those prescribed directions are to be corrected by reducing them down to be widths in directions perpendicular to directions of the respective edges. The Edge Detection and Width Measurement Unit 206 performs such a correction on edge widths. As shown in
Each horizontal or vertical edge's edge width may be corrected for its slant from either the horizontal or vertical orientation (the prescribed directions), respectively.
By way of example,
From step 502 to step 506, a slant angle φ is found. For each vertical edge, at step 502, locate the column position where the horizontal gradient magnitude peaks, and find the horizontal gradient x. At step 504, find where the vertical gradient magnitude peaks along the column position and within two pixels away, and find the vertical gradient y.
At step 506, find the slant angle φ=tan−1(y/x). At step 506, the slant angle may be found by looking up a lookup table. Although steps 502 to 506 present one specific procedure and method to find the slant angle, other procedures and methods known in the art may be used instead.
Finally, at step 508, scale down the edge width by multiplying with cos(φ), or with an approximation thereto as one skilled in the art usually does in practice.
A first modification of the process shown in
A second modification is to calculate a quotient y/x between a vertical gradient y and a horizontal gradient x to produce a quotient q, then use q to input to a lookup table that has entries for various values of q. For each value of q, the lookup table returns an edge width correction factor. The edge width correction factor may be an approximation to cos (tan−1(q)) to within 20%, preferably within 5%.
For finding the slant angle φ (or an approximation thereto such that the correction factor is accurate to within 20%) and subsequently the correction factor cos(φ) (or an approximation thereto), or to directly find the correction factor without finding the slant angle φ (as in the first and second modifications), the values of x and y may be obtained in steps 502 to 506, but other methods may be employed instead.
A third modification is to perform the following for each one of a plurality of pixels in the edge: (a) find horizontal gradient x and vertical gradient y both for a pixel, (b) find q=y/x for this pixel, and (c) find a correction factor that corresponds to q, for instance cos(tan−1(q)) or an approximation thereto to within 20%. Finally, find the correction factor for the edge width by averaging across the correction factor from each of the plurality of pixels. The average may be a weighted average, such as one in which a pixel that has a larger horizontal gradient is given a larger weight than another pixel that has a lesser horizontal gradient.
Other modifications are possible along these directions or other.
Screen ThresholdAdjacent edges may be prevented altogether from contributing to a focus signal, or have their contributions attenuated, if their peak gradient magnitudes are below a predetermined fraction of an adjacent wider edge's peak gradient magnitude.
The significant decline, e.g. 20% or greater, in peak gradient magnitude for a narrower edge adjacent to a wider edge having an opposite-signed gradient gives a hint that the blurred image is not well focused, and thus the narrower edge should not be relied upon as an indication that the blurred image is sharp.
Likewise, mutually adjacent edges of alternating gradient polarities should not be relied upon for such indication even if their edge width are small as long as they are in close proximity to each other, e.g. no more than 1 pixel apart (“minimum edge gap”). The minimum edge gap is in terms of a number of pixels, e.g. 1, or 2, or in between.
Furthermore, given that one edge may have been eliminated due to having a peak gradient less than the elimination threshold, two successive edges having an identical gradient polarity and spaced no more than two times the minimum edge gap plus a sharp_edge_width (sharp_edge_width is a number assigned to designate an edge width of a sharp edge) apart may be used as a condition for eliminating or demoting a contribution from one or both of the two mutually adjacent edges. either.
The Edge Detection and Width Measurement Unit 206 may execute the following algorithm for eliminating closely-packed narrower edges based on a screen threshold established from a wider edge, and a modulation screen flag that can be turned on and off.
For each edge, the screen threshold and screen flag to be used for the immediate next edge of an opposite polarity are determined according to the process of the flowchart shown in
Given the screen threshold and screen flag, an edge may be eliminated unless one of the following conditions is true: (a) the screen flag is off for this edge, (b) a peak gradient magnitude of the edge is not smaller than the screen threshold for this edge. To conditions (a) and (b) may be added condition (c) the edge width is not less than sharp_edge_width+1, where a number has been assigned for sharp_edge_width to designate an edge width of a sharp edge, and where the “+1” may be varied to set a range of edge widths above the sharp_edge_width within which edges may be eliminated if they fail (a) and (b). For the example shown in
A gradient peaking template may be specified in terms of a function, e.g. a difference or a ratio, of a width of the peaking gradient profile at an upper gradient magnitude and a width at a lower gradient magnitude that constrains one by the other. By way of example, in
Viewing it another way, the gradient peaking template's constraint stipulates that the width of a good gradient profile at a first gradient level is dependent on the width of the gradient profile at another gradient level, i.e. there is a definite relationship that constrains one by the other. Since the determination of width is done using interpolation, a certain tolerance is allowed for errors in determination of the widths. Thus the template may be expressed as F(W2)<W1<G(W2) where W1 is the width at a first gradient level and W2 the width at a second gradient level, and F(W2) and G(W2) are functions of W2 that expresses a minimum width constraint and a maximum width constraint on W1 in terms of W2.
The original of this constraint is explain with regards to
An alternate method to specify a gradient peaking template is to find a difference or ratio between the numbers of pixels above an upper gradient magnitude and those above a lower gradient magnitude, each gradient magnitude being at a certain upper fraction and lower fraction, respectively, from a peak (or interpolated peak) gradient magnitude. By way of example, in
Two different templates may be specified for two different edge widths at a predetermined fraction of a peak gradient. For example, a gradient profile having a width of 6 at a gradient level at 50% of the peak gradient value may use a different template than another having a width of 3. The predetermined fraction can be a function of a width of the gradient profile.
It should be noted that, where edges have slants, as discussed earlier in this disclosure, edge width should be corrected for the slant for generating the focus signal, and a procedure was described to perform this correction to shrink the edge widths. Before applying the gradient profile under test to select a set of constraints, the width measured from the gradient profile may be corrected for a slant before performing the lookup-table lookup and interpolation shown in
The above detection of spurious edges and solution for spurious edges may be performed in the Edge Detection & Width Measurement Unit 206.
Length FilterBelow describes a function of length filter 212. Broadly defined, length filter 212 creates a preference for edges that each connects to one or more edges of a similar orientation. A group of edges that are similarly oriented and mutually connected within the group (“concatenated edge”) is less likely to be due to noise, compared with an isolated edge that does not touch any other edge of similar orientation. The more edges of a similar orientation thus concatenated together, the lesser the chance of them being due to noise. The probability of the group being due to noise falls off exponentially as the number of edges within the group increases, and far faster than linearly. This property can be harnessed to reject noise, especially under dim-lit or short-exposure situations where the signal-to-noise ratio is weak, e.g. less than 10, within the image or within the region of interest. The preference may be implemented in any reasonable method to express such preference. The several ways described below are merely examples.
A first method is to eliminate edges that belong to vertical/horizontal concatenated edges having lengths lesser than a concatenated length threshold. The concatenated length threshold may be larger when the region of interest is dimmer. For example, the concatenated length threshold may start as small as 2, but increases to 8 as a signal-to-noise ratio within the region of interest drops to 5. The concatenated length threshold may be provided by the processor 112, 112′, 112″, for example through a ‘length command’ signal, shown in
A second method is to provide a length-weight in the length filter 212 for each edge and apply the length-weight to a calculation of focus signal in the focus signal calculator 210. An edge that is part of a longer concatenated edge receives a larger weight than one that is part of a shorter concatenated edge. For example, the length-weight may be a square of the length of the concatenated edge. Thus, a contribution of each edge towards the focus signal may be multiplied by a factor A/B before summing all contributions to form the focus signal, where B is a sum of the length-weights of all edges that enter the focus signal calculation, and A is a length-weight of the edge. Likewise, the edge-width histogram, which may be output as part of signals 134, may have edges that are members of longer concatenated edges contribute more to the bins corresponding to their respective edge width, thus preferred, instead of all edges contribute the same amount, e.g. +1. Thus, for example, each edge may contribute A/C, where C is an average value of A across the edges. Similarly, the narrow-edge count may have edges that are members to longer concatenated edges contribute more. Thus, for example, the contribution from each edge may be multiplied by A/D, where D is an average of A among edges that are counted in the narrow-edge count.
A group of N vertical (horizontal) edges where, with the exception of the top (leftmost) and the bottom (rightmost) ones, each edge touches two other vertical (horizontal) edges, one above (to the left of) itself, the other below (to the right of) itself, is a vertical (horizontal) concatenated edge of length N. The top (leftmost) edge needs only touch one edge below (to the right of) itself. The bottom (rightmost) edge needs only touch one edge above (to the left of) itself.
In a situation (not shown) where a vertical (horizontal) concatenated edge has two or more branches, i.e. having two edges in a row (column), the length may be defined as the total number of edges within the concatenated edge. Alternately, the length may be defined as the vertical (horizontal) distance from a topmost (leftmost) edge therein to a bottommost (rightmost) edge therein plus one.
There are other possible ways to define a concatenated length other than the above proposals. For example, a definition of a length for a concatenated edge shall have a property that the length is proportional to the number of member edges within the concatenated edge at least up to three. This is to be consistent with the previously stated reasoning that more edges being mutually connected by touching each other exponentially reduces a probability that the concatenated edge is caused by a noise, and as such the length should express a proportionality to the number of member edges within the concatenated edge up to a reasonable number that sufficiently enhances a confidence in the concatenated edge beyond that for a single member. The length filter 212 may de-emphasize or eliminate and thus, broadly speaking, discriminate against an edge having a concatenated length of one. The length filter 212 may discriminate against an edge having a concatenated length of two. The length filter 212 may discriminate against an edge having a concatenated length of three, to further reduce an influence of noise. The length filter 212 may do any one of these actions under a command from the processor.
Although shown in
In an alternate embodiment of a focus signal generator, the fine switch 220 may be removed so that the focus signal calculation unit 210 receives a first set of data not filtered by the width filter 209 and a second set filtered, and for each calculates a different focus signal, gross focus signal for the former, fine focus signal for the latter, and outputs both to the processor 112, 112′.
Width FilterRefer next to
In addition, the Width Filter 209 may calculate a total count of the edges whose edge widths fall within the narrow-edge range and output as part of output signals 134.
Narrow-Edge Count may be input to and used by the focus system controller (processor 112) to detect a presence of sharp image and/or for initiating tracking.
Focus SignalReferring next to the focus signal calculator 210 of
A focus control system may use the gross focus signal to search for the nearest sharp focus position in a search mode. It can move the focus position away from the current focus position to determine whether the gross focus signal increases or decreases. For example, if the gross focus signal increases (decreases) when the focus position moves inwards (outwards), there is a sharp focus position farther from the current focus position. The processor 112, 112′, 112″ can then provide a focus drive signal to move the focus lens 104 in the direction towards the adjacent sharp focus position.
A focus control system may use the fine focus signal to track an object already in sharp focus to maintain the corresponding image sharp (thus a “tracking mode”) despite changes in the scene, movement of the object, or movement of the image pickup apparatus. When an object is in sharp focus, the fine focus signal level is stable despite such changes. Hence a change in the fine focus signal suggests a change in focus distance of the object from the image pickup apparatus. By “locking” the focus control system to a given fine focus signal level near the minimum, for example between 2.0 to 2.5 in this example, in particular 2.1, any shift in the fine focus signal level immediately informs the processor 112, 112′, 112″ of a change in the focus distance of the object. The processor 112, 112′, 112″ can then determine a direction and cause the focus lens 104 to move to bring the fine focus signal level back to the “locked” level. Thus the image pickup apparatus 102, 103, 103′, 103″ is able to track a moving object.
A focus control system, e.g. as implemented in algorithm in processor 112, 112′, 112″, may use narrow-edge count to trigger a change from a search mode to a tracking mode. In the tracking mode, the focus control system uses the fine focus signal to “lock” the object. Before the focus position is sufficiently near the sharp focus position for the object, the focus control system may use the gross focus signal to identify the direction to move and regulate the speed of movement of the lens. When a object is coming into sharp focus, narrow-edge count peaks sharply. The processor 112, 112′, 112″ may switch into the tracking mode and use the fine focus signal for focus position control upon detection of a sharp rise in the narrow-edge count or a peaking or both. A threshold, which may be different for each different sharp focus position, may be assigned to each group of objects found from an end-to-end focus position “scan”, and subsequently when the narrow-edge count surpasses this threshold the corresponding group of objects is detected. For a stationary scene, e.g. for still image taking, an end-to-end focus position scan can return a list of maximum counts, one maximum count for each peaking of the narrow-edge count. A list of thresholds may be generated from the list of maximum counts, for example by taking 50% of the maximum counts.
Referring to
The processor 112′ may internally generate a focus signal and/or a narrow-edge count in addition to the functions included in the processor 112 of
The pixel array 108, A/D Converter 110, color interpolator 148, and generator 120′ may reside within a package 142, together comprising an image sensor 150′, separate from the processor 112′.
Yet another embodiment of a focus signal generator may add a census unit 240 to the generator 102 of
The main pixel array 2808 may be covered by a color filter array of a color mosaic pattern, e.g. the Bayer pattern. The optical lowpass filter 2808 prevents the smallest light spot focused on the pixel array 2808 from being too small as to cause aliasing. Where a color filter of a mosaic pattern covers the pixel array 2808, aliasing can give rise to color moiré artifacts after a color interpolation. For example, the smallest diameter of a circle encircling 84% of the visible light power of a light spot on the main pixel array 2808 (“smallest main diameter”) may be kept larger than one and a half pixel width but less than two pixel widths by use of the optical lowpass filter. For example, if the main pixel array 2808 has a pixel width of 4.5 um, whereas the smallest diameter is 2.0 um without optical lowpass filtering, the optical lowpass filter 2840 may be selected to make the light spot 6.7 um or larger in diameter.
The auxiliary pixel array 108″ may comprise one or more arrays of photodetectors. Each of the arrays may or may not be covered by a color filter array of a color mosaic pattern. The array(s) in auxiliary pixel array 108″ outputs image(s) in analog signals that are converted to digital signals 130 by A/D Converter 110. The images are sent to the focus signal generator 120. A color interpolator 148 may generate the missing colors for images generated from pixels covered by color filters. If auxiliary pixel array 108″ comprises multiple arrays of photodetectors, each array may capture a sub-image that corresponds to a portion of the image captured by the main pixel array 2808. The multiple arrays may be physically apart by more than a hundred pixel widths, and may or may not share a semiconductor substrate. Where the pixel arrays within auxiliary pixel array 108″ do not share a semiconductor substrate, they may be housed together in a package (not shown).
Main A/D Converter 2810 converts analog signals from the Main Pixel Array 2808 into digital main image data signal 2830, which is sent to the processor 112, where the image captured on the Main Pixel Array 2808 may receive image processing such as color interpolation, color correction, and image compression/decompression and finally be stored in memory card 116.
An array of photodetectors in the auxiliary pixel array 108″ may have a pixel width (“auxiliary pixel width”) that is smaller than a pixel width of the main pixel array 2808 (“main pixel width”). The auxiliary pixel width may be as small as half of the main pixel width. If an auxiliary pixel is covered by a color filter and the auxiliary pixel width is less than 1.3 times the smallest spot of visible light without optical lowpass filtering, a second optical lowpass filter may be inserted in front of the auxiliary array 108″ to increase the smallest diameter on the auxiliary pixel array 108″ (“smallest auxiliary diameter”) to between 1.3 to 2 times as large but still smaller than the smallest main diameter, preferably 1.5. The slight moiré in the auxiliary image is not an issue as the auxiliary image is not presented to the user as the final captured image.
The shaded region in
The auxiliary pixel array 108″, A/D Converter 110, focus signal generator 120 together may be housed in a package 142 and constitute an auxiliary sensor 150. The auxiliary sensor 150 may further comprise a color interpolator 148.
The auto-focus image pickup system 102, 102′, 103, 103′, 103″ may include a computer program storage medium (not shown) that comprises instructions that causes the processor 112, 112′, 112″ respectively, and/or the focus signal generator 120, 120′ to perform one or more of the functions described herein. By way of example, the instructions may cause the processor 112 or the generator 120′ to perform a slant correction for an edge width in accordance with the flowchart of
While a memory card 116 is shown as part of system 102, any nonvolatile storage medium may be used instead, e.g. hard disk drive, wherein images stored therein are accessible by a user and may be copied to a different location outside and away from the system 102.
One or more parameters for use in the system, for instance the sharp_edge_width, may be stored in a non-volatile memory in a device within the system. The device may be a flash memory device, the processor, or the image sensor, or the focus signal generator as a separate device from those. One or more formulae for use in the system, for example for calculating the concatenated length threshold, or for calculating beta may likewise be stored as parameters or as computer-executable instructions in a non-volatile memory in one or more of those devices.
While certain exemplary embodiments have been described and shown in the accompanying drawings, it is to be understood that such embodiments are merely illustrative of and not restrictive on the broad invention, and that this invention not be limited to the specific constructions and arrangements shown and described, since various other modifications may occur to those ordinarily skilled in the art.
Claims
1. A method to generate a focus signal from a plurality of edges in an image to indicate a degree of image sharpness in the image, each edge contributing a quantity, comprising:
- rejecting or attenuating a contribution of said quantity of an edge towards generating said focus signal where a first parameter of a gradient profile of said edge measured at a first gradient level does not meet a predetermined criterion that varies with a second parameter of said gradient profile of said edge measured at a second gradient level,
- wherein said first and second gradient levels are different gradient values defined as a first fraction and a second fraction, respectively, of a peak gradient value of said edge, and both said first and second parameters are determined by said gradient profile to both sides of a peak of said gradient profile.
2. The method of claim 1, further comprising:
- evaluating the first parameter and the second parameter to obtain a first value and a second value, respectively; and
- determining whether the first value meets the predetermined criterion given the second value.
3. The method of claim 1, wherein the predetermined criterion requires the first parameter to be, to within a predetermined tolerance, the second parameter multiplied by a predetermined multiplier.
4. The method of claim 3, wherein the predetermined multiplier is a function of an edge width of the edge.
5. The method of claim 3, wherein the predetermined multiplier is interpolated from a sequence of multipliers, each multiplier for a different edge width.
6. The method of claim 1, wherein the first parameter is either:
- a width of the gradient profile measured at the first gradient level; or,
- a count of pixels in the gradient profile that have gradient levels anywhere from the first gradient level to a peak gradient level of the gradient profile.
7. The method of claim 6, wherein the second parameter is either:
- a width of the gradient profile measured at the second gradient level; or,
- a count of pixels in the gradient profile that have gradient levels anywhere from the second gradient level to the peak gradient level.
8. The method of claim 6, wherein the width is measured from a gradient or interpolated gradient on one side of the peak to another gradient or interpolated gradient on the other side of the peak.
9. The method of claim 1, wherein the first and second gradient levels are within a range from 15% to 85% of the peak gradient level of the gradient profile.
10. The method of claim 1, wherein the first and second gradient levels are within a range from 20% to 80% of the peak gradient level of the gradient profile.
11. The method of claim 1, wherein the gradient profile is a sequence of consecutive all-positive or all-negative gradients across a plurality of pixels arrayed side-by-side in a first direction plotted against a sequence of consecutive integers, a peak gradient within the sequence has a peak gradient value whose magnitude is maximal among magnitudes of gradient values of all gradients within the sequence, and each side of the peak gradient includes at least a gradient.
12. The method of claim 3, wherein the edge width is corrected for a slant of the edge for determining the predetermined criterion.
13. The method of claim 1, wherein the predetermined criterion depends on a slant of the edge.
14. A method to generate a focus signal from a plurality of edges in an image to indicate a degree of image sharpness in the image, each edge contributing a quantity to the focus signal, comprising:
- rejecting or deemphasizing, among a first plurality of edges whose first widths are same, any edge whose second width lies outside a first acceptance range, wherein each edge among said first plurality of edges has a gradient profile with a peak gradient, and said first width of said each edge is a width of said gradient profile at a first gradient level that is a first fraction of said peak gradient, wherein said second width of said any edge is another width of said gradient profile at a second gradient level that is a second fraction of said peak gradient, wherein said first and second fractions are common across said first plurality of edges.
15. The method of claim 14, further comprising:
- rejecting or deemphasizing, among a second plurality of edges that all have a same third width across their respective gradient profiles at respective gradient levels that are at said first fraction of their respective peak gradients, any edge whose fourth width lies outside a second acceptance range,
- wherein said third width and said second acceptance range are different from said first width and said first acceptance range respectively,
- wherein said fourth width is a width across said gradient profile at a gradient level that is at said second fraction of said peak gradient of said any edge among said second plurality of edges.
16. The method of claim 14, wherein said rejecting or deemphasizing is performed in an Edge Detection & Width Measurement Unit housed in a package together with a pixel array.
17. The method of claim 14, wherein said rejecting or deemphasizing is performed in an Edge Detection & Width Measurement Unit housed in a package together with a processor that controls a focus position of a focus lens.
18. The method of claim 14, wherein said rejecting or deemphasizing is performed in an Edge Detection & Width Measurement Unit housed in a package that does not contain a processor that controls storing of compressed images to a removable memory card.
19. The method of claim 14, wherein said rejecting or deemphasizing involves executing computer-executable instructions.
20. A non-transitory computer readable medium, comprising computer executable instructions and or parameter values for executing the method of claim 14.
21. The method of claim 14, wherein said quantity is an edge width.
22. The method of claim 14, wherein said quantity does not vary if all pixel values of the image that enter a computation of said quantity are scaled up by a common multiplier.
23. The method of claim 14, wherein said quantity has a dimension based solely on a dimension of length.
24. The method of claim 15, wherein all widths have been corrected for slants of the respective edges from their respective edge directions in which pixels of the respective edges are arrayed.
25. The method of claim 1, wherein said quantity is an edge width.
26. The method of claim 1, wherein said quantity does not vary if all pixel values of the image that enter a computation of said quantity are scaled up by a common multiplier.
27. The method of claim 1, wherein said quantity has a dimension based solely on a dimension of length.
Type: Application
Filed: Jun 7, 2012
Publication Date: Feb 21, 2013
Inventor: Hiok Nam TAY (Singapore)
Application Number: 13/491,590
International Classification: H04N 5/232 (20060101);