Image Sensors for Robust On Chip Phase Detection, and Associated System And Methods
An image sensor for on-chip phase detection includes a pixel array for capturing an image of a scene, wherein the pixel array has a plurality of horizontal phase-detection rows, each including phase-detection pixels for detecting horizontal change in the scene, and a plurality of vertical phase-detection columns, each including phase-detection pixels for detecting vertical change in the scene, and wherein each of the horizontal phase-detection rows intersects each of the vertical phase-detection columns. A phase-detection method includes generating a pair of horizontal line profiles using one of a plurality of phase-detection rows; generating a pair of vertical line profiles using one of a plurality of phase-detection columns intersecting with the one of a plurality of phase-detection rows; and determining phase shift associated with at least one arbitrarily oriented edge in a scene, based upon the pair of horizontal line profiles and the pair of vertical line profiles.
Latest Patents:
The vast majority of electronic cameras have autofocus capability. The autofocus function automatically focuses the camera on objects in the scene viewed by the camera. Autofocus may be fully automatic such that the camera identifies objects in the scene and focuses on the objects. In some cases, the camera may even decide which objects are more important than other objects and subsequently focus on the more important objects. Alternatively, autofocus may utilize user input specifying which portion or portions of the scene are of interest. Based thereupon, the autofocus function identifies objects within the portion or portions of the scene, specified by the user, and focuses the camera on such objects.
To achieve market adoption, the autofocus function must be reliable and fast such that every time a user captures an image, the camera quickly brings the desired portion, or portions, of the scene into focus. Preferably, the autofocus function is sufficiently fast that the user does not notice any delay between pressing the trigger button and image capture. The autofocus is particularly important for cameras having no means for manual focus, such as compact digital cameras and camera phones.
Many electronic cameras use contrast autofocus, wherein the autofocus function adjusts the imaging objective to maximize contrast in at least a portion of the scene, thus bringing the portion of the scene into focus. More recently, phase-detection autofocus has gained popularity because it is faster than contrast autofocus. Phase-detection autofocus directly measures the degree of misfocus by comparing light passing through one portion of the imaging objective, e.g., the left portion, with light passing through another portion of the imaging objective, e.g., the right portion. Some digital single-lens reflex cameras include a dedicated phase-detection sensor in addition to the image sensor that captures images. However, this solution is not feasible for more compact and/or less expensive cameras. Therefore, camera manufacturers are developing image sensors with on-chip phase detection, i.e., image sensors with integrated phase detection capability. A variety of pixel layouts have been proposed for this purpose. Some of these prior art image sensors are shown in
In an embodiment, an image sensor for on-chip phase detection includes a pixel array for capturing an image of a scene. The pixel array includes a plurality of horizontal phase-detection rows and a plurality of vertical phase-detection columns. Each of the horizontal phase-detection rows has a plurality of phase-detection pixels for detecting horizontal change in the scene. Each of the vertical phase-detection columns has a plurality of phase-detection pixels for detecting vertical change in the scene. Each of the plurality of horizontal phase-detection rows intersects each of the plurality of vertical phase-detection columns.
In an embodiment, a phase-detection method uses an image sensor with on-chip phase-detection pixels. The method includes generating a pair of horizontal line profiles for light from left and right directions, respectively, using one of a plurality of horizontal phase-detection rows of the image sensor. The method further includes generating a pair of vertical line profiles for light from up and down directions, respectively, using one of a plurality of vertical phase-detection columns of the image sensor, wherein the one of a plurality of vertical phase-detection columns intersects the one of a plurality of horizontal phase-detection rows. Additionally, the method includes determining phase shift associated with at least one arbitrarily oriented edge in a scene viewed by the image sensor, based upon the pair of horizontal line profiles and the pair of vertical line profiles.
In an embodiment, an imaging system with on-chip phase-detection includes an image sensor with a pixel array for capturing an image of a scene. The pixel array has intersecting (a) horizontal phase-detection rows for measuring at least one pair of horizontal line profiles for light incident from left and right directions, respectively, and (b) vertical phase-detection columns for measuring at least one pair of vertical line profiles for light incident from up and down directions, respectively. The imaging system further includes a phase-processing module for processing the at least one pair of horizontal line profiles and the at least one pair of vertical line profiles to measure phase shift associated with an arbitrarily oriented and arbitrarily located edge in the scene.
Image sensor 100 is configured to provide robust on-chip phase detection capable of detecting edges, within scene 150, of arbitrary orientation and location. Image sensor 100 thereby enables robust autofocus for electronic camera 110. For example, through use of image sensor 100, electronic camera 110 is able to reliably autofocus on sparsely populated scenes 150. Image sensor 100 also enables a very flexible autofocus function, which may be utilized by electronic camera 110 to autofocus on objects of arbitrary location within scene 150, and/or on an arbitrary selection of portions of scene 150 that are associated with one or more edges. Herein, an “edge” in a scene refers to a spatial difference such as spatial brightness difference or a spatial color difference.
In an embodiment, image sensor 100 is a complementary metal-oxide-semiconductor (CMOS) image sensor. Image sensor 100 may be a color image sensor or a monochrome image sensor.
Imaging objective 210 has a focal length f. Assuming that imaging objective 210 is a thin lens, the thin lens equation dictates that
where DO is the distance 202 from an object to imaging objective 210 and DI is the distance 203 from imaging objective 210 to a focused image of the object. In diagram 200, imaging objective 210 is at a distance 201, denoted L, from image sensor 100, where L=DI. Therefore, object 230 is in focus of the imaging system formed by imaging objective 210 and image sensor 100, and the images formed on image sensor 100 by portions 211 and 212 coincide to yield a single image 235.
Diagram 200 (
While
Image sensor 100 includes an array of pixels 610. For clarity of illustration, individual pixels 610 are not show in
Some, but not all, of pixels 610 are masked to form phase-detection pixels: top-masked pixels 612, bottom-masked pixels 614, right-masked pixels 622, and left-masked pixels 624. For clarity of illustration, not all top-masked pixels 612, bottom-masked pixels 614, right-masked pixels 622, and left-masked pixels 624 are labeled in
Without departing from the scope hereof, top-masked pixels 612 may be configured to preferably detect light from an up direction, and bottom-masked pixels 614 may be configured to preferably detect light from a down direction. Similarly, right-masked pixels 622 may be configured to detect light from a right direction, and left-masked pixels 624 may be configured to detect light from a left direction.
The array of pixels 610 includes (a) a plurality of phase-detection columns 510 oriented along a vertical dimension of the array of pixels 610, and (b) a plurality of phase-detection rows 520 oriented along a horizontal dimension of the array of pixels 610.
Herein, “vertical” and “horizontal” refer to two orthogonal dimensions of the array of pixels 610. However, “vertical” and “horizontal” are not intended to refer to any particular directions relative to the direction of gravity. Likewise, it is understood that “up”, “down”, “left”, “right”, “top, and “bottom” do not necessarily relate to the direction of gravity. Rather, “up” is opposite “down”, “top” is opposite “bottom”, “left” is opposite “right”, “left” and “right” are associated with the “horizontal” dimension, while “top”, “bottom”, “up”, and “down” are associated with the “vertical” dimension. The top potion of pixel 610 is more towards the top of pixel 610 than the bottom portion of pixel 610, and the bottom portion of pixel 610 is more towards the bottom of pixel 610. The top portion of pixel 610 need not extend to the top extreme of pixel 610, and the bottom portion of pixel 610 need not extend to the bottom extreme of pixel 610. The top and bottom portions may overlap. Without departing from the scope hereof, light incident from the up direction may include a minority light fraction that is incident from the down direction, and light incident from the down direction may include a minority light fraction that is incident from the up direction. The left portion of pixel 610 is more towards the left of pixel 610 than the right portion of pixel 610, and the right portion of pixel 610 is more towards to right of pixel 610 than the left portion of pixel 610. The left portion of pixel 610 need not extend to the left-most extreme of pixel 610, and the right portion of pixel 610 need not extend to the right-most extreme of pixel 610. The left and right portions may overlap. Without departing from the scope hereof, light incident from the left direction may include a minority light fraction that is incident from the right direction, and light incident from the right direction may include a minority light fraction that is incident from the left direction. In an embodiment, the “vertical” and “horizontal” dimensions are parallel to orthogonal sides of the array of pixels 610, respectively, as illustrated in
Phase-detection column 510 includes a plurality of top-masked pixels 612 and a plurality of bottom-masked pixels 614. Through use of top-masked pixels 612 and bottom-masked pixels 614, phase-detection column 510 provides a measurement of the phase-shift between (a) light incident upon phase-detection column 510 from a down direction and (b) light incident upon phase-detection column 510 from an up direction. If an edge in scene 150 is imaged onto image sensor 100 at phase-detection column 510, and the image of this edge is not parallel to phase-detection column 510, a line profile taken along phase-detection column 510 indicates the edge. Following from the discussion in reference to
Phase-detection column 510 may include several vertical columns of pixels 610, as shown in
Phase-detection row 520 includes a plurality of right-masked pixels 622 and a plurality of left-masked pixels 624. Similar to the above discussion of phase-detection column 510, phase-detection row 520 enables detection, and associated phase-shift measurement, of edge images that overlap with phase-detection row 520 and are not parallel to phase-detection row 520 at the location of the overlap.
Phase-detection row 520 may include several horizontal rows of pixels 610, as shown in
In certain embodiments, top-masked pixels 612 and bottom-masked pixels 614, of each phase-detection column 510, are arranged in a vertically oriented series of phase-detection pixel pairs, wherein each phase-detection pixel pair includes one top-masked pixel 612 and one bottom-masked pixel 614. In these embodiments, right-masked pixels 622 and left-masked pixels 624, of each phase-detection row 520, are arranged in a horizontally oriented series of phase-detection pixel pairs, wherein each phase-detection pixel pair includes one right-masked pixel 622 and one left-masked pixel 624.
Each phase-detection column 510 and each phase-detection row 520 includes pixels 610 that are not phase-detection pixels. These non-phase-detection pixels provide light detection that is not reduced by phase-detection associated masks, and the non-phase-detection pixels therefore have greater light-collection efficiency than the phase-detection pixels. Furthermore, electrical signals generated by the phase-detection pixels may be corrected based upon electrical signals generated by adjacent non-phase-detection pixels, such that an electronic image generated by image sensor 100 is substantially free of artifacts associated with the phase-detection pixels. For the purpose of the present disclosure, “adjacent” pixels refer to nearest-neighbor pixels such that a pixel, located away from the perimeter of the pixel array, has eight adjacent pixels. Correction based upon adjacent pixels is more accurate than correction based upon pixels further away. Therefore, it is beneficial to space apart phase-detection pixels such that each phase-detection pixel has several adjacent non-phase-detection pixels. Preferably, each phase-detection pixel, located away from the pixel array perimeter, has adjacent non-phase-detection pixels located in several different directions from the phase-detection pixel. In one embodiment of the array of pixels 610, for each phase-detection pixel located away from the perimeter of the array of pixels 610, the majority of adjacent pixels 610 are non-phase-detection pixels.
In another embodiment, all pixels 610 adjacent to each top-masked pixel 612, all pixels 610 adjacent to each bottom-masked pixel 614, all pixels 610 adjacent to each right-masked pixel 622, and all pixels 610 adjacent to each left-masked pixel 624 are a non-phase-detection pixels. In yet another embodiment, each top-masked pixel 612, each bottom-masked pixel 614, each right-masked pixel 622, and each left-masked pixel 624 is adjacent to at least one non-phase-detection pixel.
The plurality of phase-detection columns 510 intersects with the plurality of phase-detection rows 520 to enable (a) detection of arbitrarily oriented edges in scene 150, such as a scene edge producing an arbitrarily oriented edge 560 on image sensor 100, and (b) measurement of phase shifts (as discussed in connection with
Each phase-detection column 510 and each phase-detection row 520 substantially spans the extent of the array of pixels 610 in a corresponding dimension. In an embodiment, phase-detection columns 510 and phase-detection rows 520 are arranged to provide substantially uniform coverage of the array of pixels 610. Such embodiments of image sensor 100 may be capable of measuring misfocus-induced phase shifts associated with arbitrarily oriented edges having arbitrary location within an image of scene 150 formed on image sensor 100, wherein the edge has (a) horizontal extent at least as long as the horizontal spacing between phase-detection columns 510 and (b) vertical extent at least as long as the vertical spacing between phase-detection rows 520. Smaller spacing between phase-detection columns 510 and smaller spacing between phase-detection rows 520 enable phase-shift measurement for smaller edges in scene 150. However, smaller spacing between phase-detection columns 510 and smaller spacing between phase-detection rows 520 may increase the density of phase-detection pixels in the array of pixels 610. In turn, this increased density decreases the overall amount of light collected by image sensor 100 and also increases the number of pixels, in an electronic image generated by image sensor 100, that must be corrected for the partial masking of the associated phase-detection pixel. Accordingly, there may be a trade-off between the detectable feature size in scene 150 and the non-focus related image quality achievable by image sensor 100. In one example, image sensor 100 includes at least ten phase-detection columns 510 and at least ten phase-detection rows 520, such as twelve phase-detection columns 510 and twelve phase-detection rows 520.
Without departing from the scope hereof, pixel 610 may be arranged in on a hexagonal lattice, such that each pair of neighboring rows of pixels 610 are displaced from each other, in the horizontal dimension, by half a pixel spacing.
Non-phase-detection pixel 710 includes a photosensitive element 720 and a lens 730 that focuses light incident on non-phase-detection pixel 710 onto photosensitive element 720. Photosensitive element 720 is, for example, a photodiode. Lens 730 refracts light 742 from a left direction onto a left portion of photosensitive element 720. Lens 730 refracts light 744 from a right direction onto a right portion of photosensitive element 720. Non-phase-detection pixel 710 is sensitive to both light 742 and light 744.
As compared to non-phase-detection pixel 710, right-masked pixel 722 additionally includes a mask 752. Mask 752 covers a right portion of photosensitive element 720 and thereby blocks light 744. Thus, right-masked pixel 722 is sensitive only to light 742 and not to light 744. Similarly, as compared to non-phase-detection pixel 710, left-masked pixel 724 additionally includes a mask 754 that allows only light 744 to reach photosensitive element 720. Thus, left-masked pixel 724 is sensitive only to light 744 and not to light 742. Without departing from the scope hereof, mask 752 may be located elsewhere in right-masked pixel 722, as long as mask 752 predominantly blocks light 744. Similarly, mask 754 may be located elsewhere in left-masked pixel 724, as long as mask 755 predominantly blocks light 742.
Although not illustrated in
Actual light propagation through microlens 730 may differ from that shown in
Exemplary ROIs 852 each include a portion of one phase-detection column 510 and a portion of one phase-detection row 520. Each of ROIs 852 are rectangular and include the intersection between the portion of one phase-detection column 510 and the portion of one phase-detection row 520. Exemplary ROIs 854 each include portions of two phase-detection columns 510 and portions of two phase-detection rows 520. Each of ROIs 854 are rectangular and include four intersections between the portions of phase-detection columns 510 and phase-detection rows 520. Exemplary ROIs 856 each include two intersections between phase-detection columns 510 and phase-detection rows 520. ROIs 852, 854, and 856 may be located anywhere within the grid formed by phase-detection columns 510 and phase-detection rows 520. For embodiments of image sensor 100, wherein phase-detection columns 510 and phase-detection rows 520 span the full extent of respective dimensions of the array of pixels 610 (
Image sensor 100 is configured to facilitate a great variety of phase-detection ROIs compatible with detection of edges of arbitrary orientation. This flexibility adds to the robustness of phase detection using image sensor 100 and, thus, adds to the robustness of an associated autofocus function. For example, ROIs 852, 854, and 856 may be located anywhere within the grid formed by phase-detection columns 510 and phase-detection rows 520. For embodiments of image sensor 100, wherein phase-detection columns 510 and phase-detection rows 520 span the full extent of respective dimensions of the array of pixels 610 (
The array of pixels 902 includes a plurality of phase-detection columns 910 and a plurality of phase-detection rows 920. Phase-detection column 910 and phase-detection row 920 are embodiments of phase-detection column 510 and phase-detection row 520, respectively. Phase-detection column 910 includes phase-detection pixels, top-masked pixels 912 and bottom-masked pixels 914, formed by partially masking some of pixels 902, for example as discussed in reference to
All pixels 902, adjacent to a phase-detection pixel, are non-phase detection pixels. Similar to the discussion in reference to
In the example shown in
The array of color pixel groups 1002 includes a plurality of phase-detection columns 1010 and a plurality of phase-detection rows 1020. Phase-detection column 1010 and phase-detection row 1020 are embodiments of phase-detection column 510 and phase-detection row 520, respectively. Phase-detection column 910 includes phase-detection pixels, top-masked pixels 1012 and bottom-masked pixels 1014, formed by partially masking some of pixels R, for example as discussed in reference to
Without departing from the scope hereof, top-masked pixels 1012 and bottom-masked pixels 1014 may be formed by partially masking some of pixels B, or by partially masking some of pixels G. Likewise, right-masked pixels 1022 and left-masked pixels 1024 may be formed by partially masking some of pixels R, or by partially masking some of pixels G.
All pixels R, G, and B, adjacent to a phase-detection pixel, are non-phase detection pixels. Furthermore, each color pixel group 1002, including a phase-detection pixel, is adjacent to at least one color pixel group 1002 that does not include a phase-detection pixel. Similar to the discussion in reference to
Comparing presently disclosed image sensor 100 (
Top-masked pixels 612 (
Right-masked pixels 622 and left-masked pixels 624 of a phase-detection row 520 generate electrical signals indicating horizontal line profiles 1224 and 1234 for edge 1260 along phase-detection row 520. Line profiles 1224 and 1234 are plotted as brightness and/or color measurement 1290 versus horizontal position 1284. Right-masked pixels 622 produce one of horizontal line profiles 1224 and 1234, while left-masked pixels 624 produce the other one of horizontal line profiles 1224 and 1234. Edge 1260 is apparent in each of line profiles 1224 and 1234 as a change in brightness and/or color measurement 1290. Each of line profiles 1224 and 1234 provide a measurement of the extent 1214 of edge 1260 along phase-detection row 520. Together, line profiles 1224 and 1234 provide a measurement of misfocus-induced phase shift 1204 between line profiles 1224 and 1234. If the optical system that images scene 150 onto image sensor 100 is free of astigmatism, misfocus-induced phase shift 1204 is the same as misfocus-induced phase shift 1202. If, on the other hand, the optical system is astigmatic, misfocus-induced phase shift 1204 may be different from misfocus-induced phase shift 1202.
The accuracy, with which misfocus-induced phase shift 1202 may be determined, is a function of the ratio between misfocus-induced phase shift 1202 and extent 1212. Similarly, the accuracy, with which misfocus-induced phase shift 1204 may be determined, is a function of the ratio between misfocus-induced phase shift 1204 and extent 1214. In the example of
The example of
It follows from the above discussion that phase-detection column 510 provides a better phase-shift measurement for near-horizontal edges than phase-detection row 520, while phase-detection row 520 provides a better phase-shift measurement for near-vertical edges than phase-detection column 510. It also follows that phase-detection column 510 is unable to enable measurement of the phase shift for vertical edges, and depending on non-ideal properties discussed above, may be unable to enable measurement of the phase shift for near-vertical edges. Likewise, phase-detection row 520 is unable to enable measurement of the phase shift for horizontal edges, and depending on non-ideal properties discussed above, may be unable to enable measurement of the phase shift for near-horizontal edges. Consequently, robust on-chip phase detection requires both phase-detection columns 510 and phase-detection rows 520.
For comparison, prior art image sensor 1100 (
Prior art image sensor 1310 is similar to those disclosed in U.S. Pat. No. 8,259,215 B2 and includes a plurality of dispersed phase-detection pixels 1320. However, any given edge in an image formed on prior art image sensor 1310, such as edge 1330, may coincide with no or just a few phase-detection pixels, and the generation of line profiles is unfeasible. Instead, phase detection using prior art image sensor 1310 relies on quantitative comparison of signals from a relatively small number of phase-detection pixels near an edge. Imperfections, for example the non-idealities discussed in reference to
Prior art image sensor 1350 is similar to one disclosed in U.S. Pat. No. 7,924,342 B2 and includes a plurality of dispersed, short lines of phase-detection pixels. The phase-detection pixels of prior art image sensor 1350 are arranged in horizontal lines 1360 for measurement of horizontal phase shifts, vertical lines 1370 for measurement of vertical phase shifts, and crossed vertical and horizontal lines 1380 for measurement of horizontal and vertical phase shifts. Each of horizontal lines 1360, vertical lines 1370 and crossed vertical and horizontal lines 1380 are composed entirely of phase-detection pixels. In comparison, each of phase-detection columns 510 and phase-detection rows 520 of image sensor 100 includes non-phase-detection pixels that may be used to correct image artifacts caused by the phase-detection pixels. The configuration of the array of pixels 610 provides for improved correction for each phase-detection pixel, as compared to prior art image sensor 1350. Substantial portions of prior art image sensor 1350 have no phase-detection capability. Such non-phase-detection portions include peripheral areas as well as substantially sized interior areas, such as the area around exemplary edge 1330. Thus, prior art image sensor 1350 may fail to enable autofocus on desired scene features. Comparing image sensor 100 to prior art image sensor 1350, the configuration of image sensor 100 provides superior phase-detection coverage of the full pixel array, while overcoming image artifacts attributable to phase-detection pixels by intermixing non-phase-detection pixels with phase-detection pixels in phase-detection columns 510 and phase-detection rows 520.
In comparison, in color image sensor 1000 (
Interface 1760 is an interface that handles communication between imaging system 1700 and a user and/or an external system such as a computer. Interface 1760 may include user interface devices such as a display, a touch screen, and/or a keyboard. Interface 1760 may include wired (such as Ethernet, USB, FireWire, or Thunderbolt) and/or wireless (such as Wi-Fi or Bluetooth) connections for communicating images to a user or an external system.
For each phase-detection column 510 (
For each phase-detection row 520, or each one of several portions of each phase-detection row 520, considered by phase-processing module 1720, phase-processing module 1720 processes electrical signals generated by right-masked pixels 622 and left-masked pixels 624 to determine a horizontal line profile pair 1724 consisting of a horizontal line profile 1725 and a horizontal line profile 1726. Phase-processing module 1720 determines horizontal line profile 1725 and horizontal line profile 1726 based upon electrical signals received from right-masked pixels 622 and left-masked pixels 624, respectively. Horizontal line profiles 1725 and 1726 are similar to line profiles 1224 and 1234.
Based upon vertical line profile pair 1721 and horizontal line profile pair 1724, phase-processing module 1720 detects an edge (such as edge 1260) in an image formed on the array of pixels 610 and determines an associated phase shifts 1727. The edge, thus detected by phase-processing module 1720, may have arbitrary orientation relative to the array of pixels 610.
Although image sensor 100 is shown in
In an embodiment, imaging system 1700 includes an autofocus module 1740 and an imaging objective 1710. Autofocus module 1740 adjusts imaging objective 1710 based upon phase shifts 1727 received from phase-processing module 1720.
Image sensor 100 captures image 1780 of a scene 150. Image sensor 100 may output image 1780 directly to interface 1760. In an embodiment, imaging system includes an image correction module 1750 that corrects image 1780 for contribution from top-masked pixels 612, bottom-masked pixels 614, right-masked pixels 622, and left-masked pixels 624, for example as discussed above in reference to
In an embodiment, imaging system 1700 includes a ROI selection module 1730 that selects an ROI, within the array of pixels 610, to be processed by phase-processing module 1720. ROI selection module 1730 may receive ROI specification from interface 1760. Alternatively, or in combination therewith, ROI selection module 1730 receives, from phase-processing module 1720, locations of edge(s) with respect to the array of pixels 610 and, based thereupon, determines an ROI specification.
Imaging system 1700 may further include an enclosure 1790 and/or a power supply 1770.
In a step 1810, phase-processing module 1720 uses image sensor 100 to determine a phase shift associated with an arbitrarily oriented edge in scene 150. Phase-processing module 1720 performs step 1810 using an ROI of the array of pixels 610. This ROI may include the full array of pixels 610 or a portion thereof. In one example, the ROI is composed of several non-contiguous ROIs. Exemplary ROIs are shown in
Step 1810 includes steps 1811, 1812, and 1814. In step 1811, phase-processing module 1720 generates vertical line profile pair 1721 for each of phase-detection columns 510, or portion(s) thereof, within the considered ROI. In step 1812, phase-processing module 1720 generates horizontal line profile pair 1724 for each of phase-detection rows 520, or portion(s) thereof, within the considered ROI. In step 1814, phase-detection module 1720 processes each vertical line profile pair 1721 and horizontal line profile pair 1724, generated in steps 1811 and 1812, to determine a phase shift 1727 associated with an arbitrarily oriented edge in scene 150 imaged onto the considered ROI of the array of pixels 610. Without departing from the scope hereof, phase-processing module 1720 may perform step 1814 to determine a plurality of phase shifts associated with a plurality of edges in scene 150.
In an embodiment, step 1814 includes steps 1815, 1816, and 1817. In step 1815, phase-processing module 1720 identifies a feature in at least one of the vertical line profile pairs 1721 and horizontal line profiles pairs 1724 generated in steps 1811 and 1812. The feature is, for example, a change in brightness and/or color measurement 1290 (
In certain embodiments, method 1800 includes a step 1804, wherein ROI selection module 1730 selects the ROI. In one example of such embodiments, method 1800 further includes a step 1802, wherein ROI selection module 1730 receives specification of the ROI from interface 1760. In another example, ROI selection module 1730 selects the ROI based upon data generated by phase-processing module 1720 in step 1810. In this example, phase-processing module 1720 may perform step 1810 to detect one or more edges in an image formed on the array of pixels 610, whereafter ROI selection module 1730 selects an ROI that includes one or more of such edges.
In an embodiment, method 1800 includes a step 1820, wherein autofocus module 1740 adjusts focus of imaging objective 1710 based upon phase shift 1727 determined in step 1810. Method 1800 may include one or more iterations of steps 1810 and 1820 such that autofocus module 1740 may autofocus imaging system 1700.
Computer 2010 includes a processor 2020 and a memory 2030. Processor 2020 is communicatively coupled with memory 2030, image sensor 100, interface 1760, and optionally imaging objective 1710. Memory 2030 is, for example, of type ROM, Flash, magnetic tape, magnetic drive, optical drive, RAM, other non-transitory medium, or combinations thereof. Memory 2030 includes a data storage 2032 and machine-readable instructions 2034 encoded in a non-transitory portion of memory 2030. Data storage 2032 stores vertical line profile pair(s) 1721, horizontal line profile pair(s) 1724, phase shift(s) 1727, image 1780, and optionally corrected image 1785. Computer 2010 implements phase-detection module 1720 as machine-readable instructions, within machine-readable instructions 2034, executable by processor 2020. Additionally, computer 2010 may implement one or more of ROI selection module 1730, autofocus module 1740, and image correction module 1750 as machine-readable instructions, within machine-readable instructions 2034, executable by processor 2020.
Referring now to
Referring now to
Referring now to
Combinations of Features
Features described above as well as those claimed below may be combined in various ways without departing from the scope hereof. For example, it will be appreciated that aspects of one image sensor for robust on-chip phase detection, or associated system or method, described herein may incorporate or swap features of another image sensor for robust on-chip phase detection, or associated system or method, described herein. The following examples illustrate some possible, non-limiting combinations of embodiments described above. It should be clear that many other changes and modifications may be made to the image sensors, systems, and methods herein without departing from the spirit and scope of this invention:
(A1) An image sensor for on-chip phase detection may include a pixel array for capturing an image of a scene, the pixel array including (a) a plurality of horizontal phase-detection rows, each having a first plurality of phase-detection pixels for detecting horizontal change in the scene, and (b) a plurality of vertical phase-detection columns, each having a second plurality of phase-detection pixels for detecting vertical change in the scene.
(A2) In the image sensor denoted as (A1), each of the plurality of horizontal phase-detection rows may intersect each of the plurality of vertical phase-detection columns.
(A3) In each of the image sensors denoted as (A1) and (A2), the first plurality of phase-detection pixels, for each of the horizontal phase-detection rows, may include a plurality of left-masked pixels and a plurality of right-masked pixels, for detecting the horizontal change; and the second plurality of phase-detection pixels, for each of the vertical phase-detection columns, may include a plurality of top-masked pixels and a plurality of bottom-masked pixels, for detecting the vertical change.
(A4) In the image sensor denoted as (A3), the left-masked pixels and the right-masked pixels, of each of the horizontal phase-detection rows, may be arranged in a horizontally oriented series of phase-detection pixel pairs, wherein each phase-detection pixel pair includes one of the left-masked pixels and one of the right-masked pixels; and the top-masked pixels and the bottom-masked pixels, of each of the vertical phase-detection columns, may be arranged in a vertically oriented series of phase-detection pixel pairs, wherein each phase-detection pixel pair includes one of the top-masked pixels and one of the bottom-masked pixels.
(A5) In each of the image sensors denoted as (A3) and (A4), within each of the horizontal phase-detection rows, the left-masked pixels may be vertically offset from the right-masked pixels; and, within each of the vertical phase-detection columns, the top-masked pixels may be horizontally offset from the bottom-masked pixels.
(A6) In each of the image sensors denoted as (A1) through (A5), each of the phase-detection pixels may be adjacent only to pixels, of the pixel array, that are not phase-detection pixels.
(A7) In each of the image sensors denoted as (A1) through (A6), the pixel array may be composed of a plurality of color pixel groups for generating a color image of the scene.
(A8) In the image sensor denoted as (A7), each color pixel group may have a plurality of types of color-sensitive pixels sensitive to a respective plurality of colors, and all of the phase-detection pixels may be of same one of the plurality of types.
(A9) In each of the image sensors denoted as (A7) and (A8), the plurality of color pixel groups may include (a) a plurality of phase-detection color pixel groups, each having at most one of the phase-detection pixels, and (b) a plurality of non-phase-detection color pixel groups having no phase-detection pixels.
(A10) In the image sensor denoted as (A9), each of the phase-detection color pixel groups may be adjacent to at most two other ones of the phase-detection color pixel groups.
(A11) In each of the image sensors denoted as (A1) through (A10), each of the horizontal phase-detection rows may substantially span horizontal extent of the pixel array; and each of the vertical phase-detection columns may substantially span vertical extent of the pixel array.
(A12) In each of the image sensors denoted as (A1) through (A11), the plurality of horizontal phase-detection rows may include at least ten phase-detection rows, and the plurality of vertical phase-detection columns may include at least ten phase-detection columns.
(B1) A phase-detection method, using an image sensor with on-chip phase-detection pixels, may include (a) generating a pair of horizontal line profiles for light from left and right directions, respectively, using one of a plurality of horizontal phase-detection rows of the image sensor, (b) generating a pair of vertical line profiles for light from up and down directions, respectively, using one of a plurality of vertical phase-detection columns of the image sensor, the one of a plurality of vertical phase-detection columns intersecting the one of a plurality of horizontal phase-detection rows, and (c) determining phase shift associated with at least one arbitrarily oriented edge in a scene viewed by the image sensor, based upon the pair of horizontal line profiles and the pair of vertical line profiles.
(B2) The method denoted as (B1) may further include adjusting focus of an imaging objective to reduce the phase shift.
(B3) Each of the methods denoted as (B1) and (B2) may further include capturing an image of the scene using the image sensor.
(B4) In each of the methods denoted as (B1) through (B3), the step of generating a pair of horizontal line profiles may include using a first portion of a plurality of phase-detection pixels associated with the plurality of horizontal phase-detection rows, and the steep of generating a pair of vertical line profiles may include using a second portion of a plurality of phase-detection pixels associated with the plurality of vertical phase-detection columns.
(B5) The method denoted as (B4) may further include correcting contribution to the image, from each of the phase-detection pixels associated with the plurality of horizontal phase-detection rows and each of the phase-detection pixels associated with the plurality of vertical phase-detection columns, based upon respectively adjacent non-phase-detection pixels of corresponding color-sensitivity.
(B6) In each of the methods denoted as (B1) through (B5), the step of generating a pair of horizontal line profiles may include generating (a) a line profile produced by a plurality of left-masked pixels and (b) a line profile produced by a plurality of right-masked pixels, and the step of generating a pair of vertical line profiles may include generating (a) a line profile produced by a plurality of top-masked pixels and (b) a line profile produced by a plurality of bottom-masked pixels.
(B7) In the method denoted as (B6), in the step of generating a pair of horizontal line profiles, the left-masked pixels may be arranged in a first horizontal pixel row, and the right-masked pixels may be arranged in a second horizontal pixel row; and, in the step of generating a pair of vertical line profiles, the top-masked pixels may be arranged in a first vertical pixel column, and the bottom-masked pixels may be arranged in a second vertical pixel column.
(B8) In each of the methods denoted as (B1) through (B7), the step of determining phase shift may include identifying a feature, associated with the arbitrarily oriented edge, in at least one of (a) the pair of horizontal line profiles and (b) the pair of vertical line profiles.
(B9) In the method denoted as (B8), the step of determining phase shift may further include evaluating, for at least one of (a) the pair of horizontal line profiles and (b) the pair of vertical line profiles, phase shift of the feature.
(B10) In the method denoted as (B9), the step of determining phase shift may further include defining the phase shift of the arbitrarily oriented edge as the value of the phase shift of the feature.
(B11) Each of the methods denoted as (B1) through (B10) may further include selecting a region of interest having arbitrary location and being associated with the arbitrarily oriented edge.
(B12) In the method denoted as (B11), the region of interest may have extent sufficient to include (a) at least one portion of at least one of the plurality of horizontal phase-detection rows and (b) at least one portion of at least one of the plurality vertical columns of phase-detection pixels.
(B13) In the method denoted as (B12), the step of generating a pair of horizontal line profiles may include generating the pair of horizontal line profiles for each of the at least one portion of each of the at least one of the plurality of horizontal phase-detection rows, and the step of generating a pair of vertical line profiles may include generating the pair of vertical line profiles for each of the at least one portion of each of the at least one of the plurality of vertical phase-detection columns.
(B14) In the method denoted as (B13), the step of determining phase shift may include determining the phase shift based upon one or more pairs of line profiles measured in the step of measuring a pair of horizontal line profiles and the step of measuring a pair of vertical line profiles.
(C1) An imaging system with on-chip phase-detection may include an image sensor with a pixel array for capturing an image of a scene, wherein the pixel array includes intersecting (a) horizontal phase-detection rows for measuring at least one pair of horizontal line profiles for light incident from left and right directions, respectively, and (b) vertical phase-detection columns for measuring at least one pair of vertical line profiles for light incident from up and down directions, respectively.
(C2) The imaging system denoted as (C1) may further include a phase-processing module for processing the at least one pair of horizontal line profiles and the at least one pair of vertical line profiles to measure phase shift associated with an arbitrarily oriented and arbitrarily located edge in the scene.
(C3) The imaging system denoted as (C2) may further include a region-of-interest selection module for selecting at least one of the horizontal phase-detection rows and at least one of the vertical phase-detection columns for processing by the phase-processing module to measure the phase shift.
(C4) Each of the imaging systems denoted as (C2) and (C3) may further include an autofocus module for adjusting focus of an imaging objective to reduce the phase shift.
(C5) In each of the imaging systems denoted as (C1) through (C4), each of the horizontal phase-detection rows may include a first plurality of partially masked pixels, wherein each of the first plurality of partially masked pixels detects light from one of the left and right directions; and each of the vertical phase-detection columns may include a second plurality of partially masked pixels, wherein each of the second plurality of partially masked pixels detects light from one of the up and down directions.
(C6) The imaging system denoted as (C5) may further include an image correction module for correcting contribution to the image from each of the partially masked phase-detection pixels based upon adjacent non-phase-detection pixels of the pixel array.
(C7) In each of the imaging systems denoted as (C1) through (C6), the image sensor may be any one of the image sensors denoted as (A1) through (A12).
Changes may be made in the above devices, systems and methods without departing from the scope hereof. It should thus be noted that the matter contained in the above description and shown in the accompanying drawings should be interpreted as illustrative and not in a limiting sense. The following claims are intended to cover generic and specific features described herein, as well as all statements of the scope of the present system and method, which, as a matter of language, might be said to fall therebetween.
Claims
1. An image sensor for on-chip phase detection, comprising:
- a pixel array for capturing an image of a scene, the pixel array including: a plurality of horizontal phase-detection rows, each including a first plurality of phase-detection pixels, comprising a plurality of left-masked pixels and a plurality of right-masked pixels, for detecting horizontal change in the scene, and a plurality of vertical phase-detection columns, each including a second plurality of phase-detection pixels, comprising a plurality of top-masked pixels and a plurality of bottom-masked pixels, for detecting vertical change in the scene;
- wherein each of the plurality of horizontal phase-detection rows intersects each of the plurality of vertical phase-detection columns.
2. (canceled)
3. The image sensor of claim 1,
- the left-masked pixels and the right-masked pixels, of each of the horizontal phase-detection rows, being arranged in a horizontally oriented series of phase-detection pixel pairs, each phase-detection pixel pair including one of the left-masked pixels and one of the right-masked pixels; and
- the top-masked pixels and the bottom-masked pixels, of each of the vertical phase-detection columns, being arranged in a vertically oriented series of phase-detection pixel pairs, each phase-detection pixel pair including one of the top-masked pixels and one of the bottom-masked pixels.
4. The image sensor of claim 3,
- within each of the horizontal phase-detection rows, the left-masked pixels being vertically offset from the right-masked pixels; and
- within each of the vertical phase-detection columns, the top-masked pixels being horizontally offset from the bottom-masked pixels.
5. The image sensor of claim 1, each of the phase-detection pixels being adjacent only to pixels, of the pixel array, that are not phase-detection pixels.
6. (canceled)
7. An image sensor for on-chip phase detection, comprising:
- a pixel array for capturing an image of a scene, the pixel array being composed of a plurality of color pixel groups for generating a color image of the scene, each of the color pixel groups having a plurality of types of color-sensitive pixels sensitive to a respective plurality of colors, the pixel array including: (a) a plurality of horizontal phase-detection rows, each including a first plurality of phase-detection pixels for detecting horizontal change in the scene, and (b) a plurality of vertical phase-detection columns, each including a second plurality of phase-detection pixels for detecting vertical change in the scene, each of the plurality of vertical phase-detection columns intersecting each of the plurality of horizontal phase-detection rows, all of the first and second phase-detection pixels being of same one of the plurality of types;
- the plurality of color pixel groups including: a plurality of non-phase-detection color pixel groups including no phase-detection pixels, and a plurality of phase-detection color pixel groups, each including at most one of the phase-detection pixels and being adjacent to at most two other ones of the phase-detection color pixel groups.
8. The image sensor of claim 1,
- each of the horizontal phase-detection rows spanning horizontal extent of the pixel array; and
- each of the vertical phase-detection columns spanning vertical extent of the pixel array.
9. The image sensor of claim 8,
- the plurality of horizontal phase-detection rows comprising at least ten phase-detection rows; and
- the plurality of vertical phase-detection columns comprising at least ten phase-detection columns.
10. A phase-detection method using an image sensor with on-chip phase-detection pixels, the method comprising:
- generating a pair of horizontal line profiles for light from left and right directions, respectively, using one of a plurality of horizontal phase-detection rows of the image sensor, the pair of horizontal line profiles comprising (a) a line profile produced by a plurality of left-masked pixels arranged in a first horizontal pixel row and (b) a line profile produced by a plurality of right-masked pixels arranged in a second horizontal pixel row;
- generating a pair of vertical line profiles for light from up and down directions, respectively, using one of a plurality of vertical phase-detection columns of the image sensor, the one of a plurality of vertical phase-detection columns intersecting the one of a plurality of horizontal phase-detection rows, the pair of vertical line profiles comprising (a) a line profile produced by a plurality of top-masked pixels arranged in a first vertical pixel column and (b) a line profile produced by a plurality of bottom-masked pixels arranged in a second vertical pixel column; and
- determining phase shift associated with at least one arbitrarily oriented edge in a scene viewed by the image sensor, based upon the pair of horizontal line profiles and the pair of vertical line profiles.
11. The method of claim 10, further comprising:
- adjusting focus of an imaging objective to reduce the phase shift.
12. The method of claim 11, further comprising:
- capturing an image of the scene using the image sensor.
13. The method of claim 12,
- the step of generating a pair of horizontal line profiles comprising using a first portion of a plurality of phase-detection pixels associated with the plurality of horizontal phase-detection rows;
- the step of generating a pair of vertical line profiles comprising using a second portion of a plurality of phase-detection pixels associated with the plurality of vertical phase-detection columns; and
- the method further comprising correcting contribution to the image, from each of the phase-detection pixels associated with the plurality of horizontal phase-detection rows and each of the phase-detection pixels associated with the plurality of vertical phase-detection columns, based upon respectively adjacent non-phase-detection pixels of corresponding color-sensitivity.
14-15. (canceled)
16. The method of claim 10, the step of determining phase shift comprising:
- identifying feature, associated with the arbitrarily oriented edge, in at least one of (a) the pair of horizontal line profiles and (b) the pair of vertical line profiles;
- evaluating, for at least one of (a) the pair of horizontal line profiles and (b) the pair of vertical line profiles, phase shift of the feature; and
- defining the phase shift of the arbitrarily oriented edge as the value of the phase shift of the feature.
17. A phase-detection method using an image sensor with on-chip phase-detection pixels arranged in a plurality of horizontal phase-detection rows and a plurality of vertical phase-detection columns, the method comprising:
- selecting a region of interest having arbitrary location and being associated with the arbitrarily oriented edge, the region of interest having extent sufficient to include (a) at least one portion of at least one of the plurality of horizontal phase-detection rows and (b) at least one portion of at least one of the plurality vertical columns of phase-detection pixels;
- generating a pair of horizontal line profiles for light from left and right directions, respectively, for each of the at least one portion of each of the at least one of the plurality of horizontal phase-detection rows;
- generating a pair of vertical line profiles for light from up and down directions, respectively, for each of the at least one portion of each of the at least one of the plurality of vertical phase-detection columns; and
- determining phase shift associated with at least one arbitrarily oriented edge in a scene viewed by the image sensor, based upon one or more pairs of line profiles measured in the step of generating a pair of horizontal line profiles and the step of generating a pair of vertical line profiles.
18. An imaging system with on-chip phase-detection, comprising:
- an image sensor including a pixel array for capturing an image of a scene, the pixel array having intersecting (a) horizontal phase-detection rows for measuring at least one pair of horizontal line profiles for light incident from left and right directions, respectively, and (b) vertical phase-detection columns for measuring at least one pair of vertical line profiles for light incident from up and down directions, respectively;
- a phase-processing module for processing the at least one pair of horizontal line profiles and the at least one pair of vertical line profiles to measure phase shift associated with an arbitrarily oriented and arbitrarily located edge in the scene; and
- a region-of-interest selection module for selecting at least one of the horizontal phase-detection rows and at least one of the vertical phase-detection columns for processing by the phase-processing module to measure the phase shift.
19. (canceled)
20. The imaging system of claim 18, further comprising an autofocus module for adjusting focus of an imaging objective to reduce the phase shift.
21. The imaging system of claim 18,
- each of the horizontal phase-detection rows including a first plurality of partially masked pixels, each of the first plurality of partially masked pixels detecting light from one of the left and right directions;
- each of the vertical phase-detection columns including a second plurality of partially masked pixels, each of the second plurality of partially masked pixels detecting light from one of the up and down directions; and
- the imaging system further comprising an image correction module for correcting contribution to the image from each of the partially masked phase-detection pixels based upon adjacent non-phase-detection pixels of the pixel array.
Type: Application
Filed: May 18, 2015
Publication Date: Nov 24, 2016
Applicant:
Inventors: Chengjun Li (Shanghai), Guansong Liu (San Jose, CA), Jizhang Shan (Cupertino, CA), Chin Poh Pang (Pleasanton, CA)
Application Number: 14/714,997