Abstract: An image processing apparatus includes: a plurality of micro-lenses arranged in a two-dimensional pattern so that a subject light through an image forming optical system enter there; a plurality of light receiving elements disposed in a vicinity of a focal position at rear side of the micro-lenses to correspond to the plurality of micro-lenses respectively that receive the subject light through the micro-lenses; an image synthesizing unit that synthesizes an image on a focal plane that is different from a predetermined focal plane of the image forming optical system based upon outputs from the plurality of light receiving elements; and a processing unit that, based upon at least an objective image in a vicinity of the plurality of micro-lenses, and an auxiliary image outside the vicinity of the plurality of micro-lenses, which are synthesized by the image synthesizing unit, performs a process to enhance a resolution of the objective image.
Abstract: An electronic device and a control method therefor are provided. The electronic device includes a main lens, an image sensor, and at least one processor. When an input for acquiring an image is received, the at least one processor is configured to acquire, from the at least one main lens, a first image including an object by setting the image sensor to a first position corresponding to a first focal point for the object, acquire, from the at least one main lens, a second image including the object by setting the image sensor to a second position corresponding to a second focal point for the object, and combine the acquired first image and the acquired second image to generate a combined image. The first focal point and the second focal point are positions symmetrical to each other with reference to an on-focus position for the object.
April 10, 2017
Date of Patent:
November 26, 2019
Samsung Electronics Co., Ltd.
Jae-Hyoung Park, Jeong-Won Lee, Chong-Sam Chung
Abstract: A focus detection device includes a field lens, a secondary image forming lens configured to capture a light flux from the field lens and to form a pair of images from light fluxes from different pupil positions, and a photoelectric conversion device including a pixel region including at least one focus detection region pair that detects the pair of images. The photoelectric conversion device includes pixels arranged in the pixel region in a matrix. The photoelectric conversion device includes control lines each supplying control signal to at least a part of pixels on a corresponding row, and output lines each receiving a signal from at least a part of pixels on a corresponding column. At least a pair of the pixels connected to a common control line or a common output line form the at least one focus detection region pair and output signals to be used for focus detection.
Abstract: Unit pixels that perform photoelectric conversion of light from an object are arrayed in a matrix manner. An image sensor having the unit pixels outputs a first signal for image generation based on an electric charge generated in each of the unit pixels and a second signal for phase difference detection based on an electric charge generated in a partial region of each of the unit pixels. A scanner controls scanning for reading out the first signal and the second signal for each row. The scanner performs first scanning by which the first signal is read out by being thinned out in a first period, and second scanning by which the second signal is read out in a row, in which the first signal is not read out in the first scanning, by being thinned out in each second period with a predetermined thinning rate.
Abstract: An auto-focus apparatus and an auto-focus method are provided. The auto-focus apparatus includes: a lens, a lens driving component, and a focus control module; wherein the lens is configured to guide light transmitted from a target object; wherein the lens driving component is configured to drive the lens to move to the multiple positions to obtain the original images, and drive the lens to a focusing position obtained by the focus control module; and wherein the focus control module is configured to obtain frequency respond results each of which corresponds to interested pixels in one of the original images, calculate a desired frequency respond result based on the obtained frequency respond results and take a first position of the lens corresponding to the desired frequency respond result as the focusing position. Accordingly, focus accuracy and focus speed under conditions of poor contrast texture and weak light are improved.
Abstract: An image-capturing assembly is disclosed. The image-capturing assembly includes an array of lens units and an array of optical sensor units. At least one of the lens units includes a first functional lens and a first light-blocking member having an aperture. Each of the optical sensor units includes an image-capture element corresponding to one of the lens units. The first functional lens, the aperture of the first light-blocking member, and one of the image-capture element are arranged along an optical axis, and the aperture of the first light-blocking member is asymmetrical with respect to the optical axis.
Abstract: In an image capturing apparatus, in a case where a change is made from a first magnification ratio into a second magnification ratio which is greater than the first magnification ratio in a state in which a live view is displayed at the first magnification ratio and a first AF mode is set, the live view image is enlarged and displayed at the second magnification ratio, and the AF mode is set to a second AF mode. In a case where a change to the first magnification ratio is made after the second magnification ratio is set, the first AF mode is set, and a focus adjustment position is set based on a focus adjustment position which is set in the second AF mode.
Abstract: Focus detection is to determine whether an image is in focus or not. Focus detection is able to be used for improving camera autofocus performance. Focus detection by using only one feature does not provide enough reliability to distinguish in-focus and slightly out-of-focus images. A focus detection algorithm of combining multiple features used to evaluate sharpness is described herein. A large image data set with in-focus and out-of-focus images is used to develop the focus detector for separating the in-focus images from out-of-focus images. Many features such as iterative blur estimation, FFT linearity, edge percentage, wavelet energy ratio, improved wavelet energy ratio, Chebyshev moment ratio and chromatic aberration features are able to be used to evaluate sharpness and determine big blur images.
Abstract: A phase difference AF processing unit (19) calculates, through an operation using a detection signal group SA of phase difference detection pixels (52A) in an AF area (53), a detection signal group SB of phase difference detection pixels (52B), and a detection signal group SN of G pixels 51 in a row between a row including the phase difference detection pixels (52A) and a row including the phase difference detection pixels (52B), a third correlation value corresponding to a value obtained by adding up a first correlation value between the detection signal group SA and the detection signal group SN and a second correlation value between the detection signal group SB and the detection signal group SN, and generates a defocus amount Df1 from the third correlation value. A system control unit (11) drives a focus lens based on the defocus amount Df1.
Abstract: An image pickup device includes a pixel section in which image pixels and phase difference pixels are arranged in a matrix shape, an image-signal generating section that stores pixel signals outputted from the pixel section in a memory section for one or more rows of unit rows, which are rows of unit pixels configured by a plurality of pixels adjacent to each other, combines, for each of the unit rows, the pixel signals of the unit pixels as combined pixel signals, and extracts phase difference pixel signals from the unit row, and an image-signal readout section that reads out, as signals of one row, the combined pixel signals and the phase difference pixel signals of one unit row generated by the image-signal generating section.
Abstract: An imaging apparatus including an imaging unit that obtains a plurality of photographed images of different photographing conditions includes a setting unit that sets photographing conditions of the imaging unit, an instruction unit that instructs photographing by the imaging unit, and a notification unit that notifies photographing conditions set by the setting unit before the instruction unit instructs the photographing, wherein the notification unit notifies, before the instruction unit instructs the photographing, only a photographing condition of a longest exposure time from among a plurality of photographing conditions for obtaining the plurality of photographed images or a sum of exposure times for obtaining the plurality of photographed images.
Abstract: An autofocus system includes: an imaging unit which captures an image of an object formed by an optical system; an autofocus portion which performs focus adjustment of the optical system on an autofocus area set in the image captured by the imaging unit so that the object within the autofocus area comes into focus; an AF area tracking processing portion which performs tracking processing to set a predetermined object within the autofocus area as a target to be tracked, detect a presence area of the target to be tracked in an image newly captured by the imaging unit and determine the presence area as a new autofocus area; a depth-of-field computing portion which calculates a depth of field based on information acquired from the optical system; and a control portion as defined herein.
Abstract: The camera main body 200 stores sensitivity ratio data indicating a sensitivity ratio of a pixel 51R in the position and an imaging pixel 51 which is adjacent to the pixel 51R and a sensitivity ratio of a pixel 51L in the position and a pixel 51 which is adjacent to the pixel 51L, for every information of the different incident light ray angles in the arbitrary position of a light receiving surface 50 in an X direction. The system control unit 11 obtains information of the incident light ray angle in two positions on the light receiving surface 50 corresponding to the set optical condition and corrects the level difference of the output signals of the pixels 51R and 51L using the sensitivity ratio data corresponding to the obtained incident light ray angle.
Abstract: A method for tracking collector devices in an indoor area associated with imaging devices and RF signal sources covering the area includes generating aspects of a plurality of coarse position tracks of the collector devices based on RF signal measurements obtained by the collector devices from the RF signal sources. Imaging devices capture image frames of the collector devices in the indoor area which are then processed in order to determine aspects of a plurality of fine position tracks of the collector devices. The server and the collector device communicate with each other and match aspects of at least one of the fine position tracks to the aspects of the coarse position track of the communicating collector device in order to determine the precise location of the collector device in the indoor area.
Abstract: An imaging element 5 includes an imaging pixel cell 30 and focus detecting pixel cells 31R and 31L. The digital signal processing unit 17 determines which one of interpolation processing and gain correction processing is to be performed using at least the F value at the time of imaging. The interpolation processing corrects the output signal of the focus detecting pixel cell through signal interpolation using the output signals of the surrounding imaging pixel cells, and the gain correction processing amplifies and corrects the output signal of the focus detecting pixel cell by a gain.
Abstract: An image sensor may include a pixel array having image pixels for capturing image data and phase detection pixels for gathering phase information during automatic focusing operations. Phase detection pixels may form phase detection pixel pairs having first and second pixels with different angular responses. The first and second pixels may have color filters of the same color or may have color filters of different colors. The phase detection pixel pairs may be isolated from other phase detection pixel pairs in the array or may be arranged consecutively in a line. The phase detection pixels may, for example, be provided with color filters to match the color filter pattern of the pixel array. Processing circuitry may adjust red and green pixel signals from a phase detection pixel pair having a red and green color filter and may subsequently determine a phase difference using the adjusted pixel signals.
Abstract: A system for controlling a motor may include a motor driver circuit for driving a camera motor. A memory capable of storing a plurality of parameters for controlling the camera motor may also be included. A set of parameters from the memory may be chosen to be applied to driving the motor. A motor control module may receive a signal from the control logic module, apply the chosen set of parameters to driving the camera motor, and command the motor driver circuit to drive the motor in accordance with the applied set of parameters. The parameters may be chosen based on desired behavior of the system and various other stimuli.
Abstract: There is provided an image sensor including a plurality of sensors for autofocus. The sensors are divided into a plurality of groups. Clock signals for drive in different timings for each group are supplied to the sensors.
Abstract: A focus detector for detecting a defocus amount of an image pickup optical system based on a displacement amount between two images formed by two beams divided from the image pickup optical system and passed through two pupil areas, the focus detector including: two lenses; two phase difference sensors for photoelectrically converting two subject images formed by the two lenses into two image signals; a correlation computing unit for dividing the two image signals based on a reference number-of-pixels to compute an image displacement amount for each divided area; a waveform coincidence computing unit for computing a coincidence degree of the two image signals in the each divided area in which the image displacement amount is computed by the correlation computing unit; and a defocus computing unit for computing a defocus amount based on the coincidence degree of the two image signals computed by the waveform coincidence computing unit.
Abstract: There is provided a solid state imaging apparatus, including a plurality of line sensors including a plurality of pixels arrayed in a line, each of the pixels including an amplifier which amplifies a signal corresponding to a charge accumulated in a photoelectric transducer, and signal lines each for reading a signal of each pixel of the line sensors. The plurality of line sensors are discretely arranged, and the signal lines are gathered and wired along a region in which a circuit block including the line sensors is arranged.
June 25, 2013
Date of Patent:
September 8, 2015
Kenya Kondou, Haruhisa Naganokawa, Daijiro Anai, Makoto Aoki, Syouhei Taguchi, Ken Koseki, Nobuo Nakamura
Abstract: A camera system to capture a visual image comprising of a housing that allowing a beam of light representing the visual image to be captured to enter the housing. The camera system has a dichroic beam splitter placed in the path of the beam of light to split the beam of light into a two beams of light. The first beam of light is comprised of red light and blue light while the second beam is comprised of green light and near infrared light. The camera system has two filters that will filter the two beams of light. The first filter is comprised of an array of a first and second filter elements. The first filter element allows red light to pass through it while the second filter element allows blue light to pass through it. The second filter has an array of two filter elements as well. The first and second filter elements in the second filter will allow green and near infrared light to pass through it. The amount of filtered light for each of the spectral bands is roughly equal to each other.
Abstract: A solid-state imaging apparatus, comprising a plurality of pixels arranged on a substrate, and element isolation regions formed between the plurality of pixels on the substrate, wherein the plurality of pixels include a first pixel including a first color filter for passing light having a first wavelength, a second pixel including a second color filter for passing light having a second wavelength shorter than the first wavelength, and a pixel for focus detection into which light longer than at least the second wavelength enters, and of the element isolation regions, a first region between the pixel for focus detection and the first pixel has a potential barrier against a signal charge, which is higher than that of a second region between the first pixel and the second pixel.
Abstract: An image pickup apparatus includes: an image pickup lens having an aperture stop; an image pickup device obtaining image pickup data based on light detected; one or more microlenses arranged between the image pickup lens and the image pickup device so as to correspond to a plurality of pixels in a partial region of the image pickup device; and an image processing section performing image processing based on the image pickup data obtained from the image pickup device, in which the image processing section includes: a distance measurement section measuring a distance from the image pickup lens to a measurement object based on pixel data in the partial region of the image pickup device, and an interpolation section interpolating pixel data for the partial region of the image pickup device.
Abstract: An image system for detecting chemiluminescence in a sample uses a highly binned, short exposure initial image to calculate the exposure time for a final image of the sample. After calculation of the exposure time, at least two final images are taken, with saturated pixels removed and replaced in a first image with corresponding unsaturated pixels from a second image. The corresponding pixels are adjusted to reflect the different intensity levels between the first and second images, and the first image becomes the final image reflecting the detected chemiluminescence.
March 14, 2013
Date of Patent:
June 16, 2015
Bio-Rad Laboratories, Inc.
Keith Kotchou, Kevin McDonald, James Lee
Abstract: A focus detector, which detects a defocus amount from a displacement amount between images formed by a pair of light beams split from an image pickup system so as to pass through a pair of pupil regions, includes a pair of lenses and phase difference sensors, a memory unit for storing an image displacement amount between the image signals on the phase difference sensors in an in-focus state, a waveform read out controller for setting pixels to be calculated for the phase difference sensors, respectively, based on the image displacement amount, a correlation calculator for calculating a correlation amount between the image signals from the pixels to be calculated, a waveform degree-of-conformity calculator for calculating a waveform degree-of-conformity based on the image signals obtained from the pixels to be calculated, and a defocus calculator for calculating the defocus amount based on the correlation amount and the waveform degree-of-conformity.
Abstract: The lens apparatus includes a first member having a first cam, a second member having a first cam follower engaging with the first cam and a second cam follower, which rotates in a circumferential direction and is moved in an optical axis direction by the first cam, a third member provided with a second cam engaging with the second cam follower and rotating the second member, and biasing members generating between the first and second members a biasing force in a direction oblique to the optical axis direction. The biasing force presses the first and second cam followers respectively against the first and second cams. A biasing force generation direction changes with rotation of the second member, and the biasing force presses the first and second cam followers respectively against same cam surfaces of the first and second cams in an entire second member rotation range.
Abstract: A zoom lens in which aberration variations at telephoto end during focusing are suppressed while suppressing breathing at wide angle end, which includes, from object side: a positive first unit which does not move for zooming; a negative second unit which moves during zooming; at least one zooming unit which moves during zooming; a stop; and an imaging unit which does not move for zooming. The first unit includes: a negative first sub unit which does not move for focusing; a positive second sub unit which moves to image side during focusing from infinity to proximity; and a positive third sub unit which moves to object side during focusing from infinity to proximity. Focal lengths of the first, second, first sub, and second sub units, and amounts of movement of the second and third sub units during the focusing from infinity proximity are appropriately set.
Abstract: A method, system and reference target for estimating spectral data on a selected one of three spectral information types is disclosed. Spectral information types comprise illumination of a scene, spectral sensitivity of an imager imaging the scene and reflectance of a surface in the scene. The method comprises obtaining a ranking order for plural sensor responses produced by the imager, each sensor responses being produced from a reference target in the scene, obtaining, from an alternate source, data on the other two spectral information types, determining a set of constraints, the set including, for each sequential pair combination of sensor responses when taken in said ranking order, a constraint determined in dependence on the ranking and on the other two spectral information types for the respective sensor responses and, in dependence on the ranking order and on the set of constraints, determining said spectral data that optimally satisfies said constraints.
Abstract: Subject distances of a plurality of subject areas included in a captured image are computed based on a plurality of focus evaluation values indicating in-focus positions of a plurality of focus detection areas set in the captured image. Upon detection of a change in the captured image, the subject distances of the plurality of subject areas are re-computed by re-moving a focus lens. In this case, a driving range of the focus lens includes in-focus positions corresponding to previously-calculated subject distances and corresponds to the distribution of the previously-calculated subject distances.
Abstract: A shift amount between a plurality of image signals that is obtained from pixels in each of which a plurality of photoelectric conversion units are provided is computed from an amount of correlation based on a difference value between the plurality of image signals. Also, if the difference value that is used for computing the amount of correlation is not less than a predetermined upper limit, a predetermined value that is not more than the upper limit is used as the difference value.
Abstract: An image processing apparatus includes a focus lens, an image sensor which captures an image and a controller which moves the image sensor or the focus lens. A distance measurer measures a distance to a subject based on a first image and a second image. A focal position moves for the first focal range by moving the image sensor or by moving the position of the focus lens during the first exposure time period, and the focal position moves for the second focal range by moving the image sensor or by moving the position of the focus lens during the second exposure time period.
Abstract: An imaging device of an aspect of the invention, when reading, as voltage signals, signal charges output from a first pixel receiving a light on a partial area biased to a predetermined direction from a light axis of a light flux passing an exit pupil of an imaging optical system and a second pixel arranged so as to be adjacent to the first pixel and receiving a light on a partial area biased to an opposite direction to the predetermined direction from the light axis, combines and reads the signal charges of adjacent first-number pixels with respect to the first pixel and the second pixel, and calculates an arithmetic mean of adjacent second-number voltage signals with respect to the combined and read voltage signals of the first pixel and the second pixel.
Abstract: An image capture apparatus comprises an image sensor which photo-electrically converts an object image formed by an imaging lens, the image sensor including a first pixel group having a first light-receiving area, and a second pixel group which is discretely arranged in the first pixel group and configured by dividing a light-receiving area substantially equal in area to the first light-receiving area into a second light-receiving area and a third light-receiving area different in area from the second light-receiving area, and control means for integrally controlling the second light-receiving area of the second pixel group and the first light-receiving area of the first pixel group.
Abstract: A solid-state image pickup element includes a pixel and a signal detecting unit. The pixel has at least two photoelectric conversion units including a first photoelectric conversion unit and a second photoelectric conversion unit in a semiconductor. The first photoelectric conversion unit has a higher impurity density than the second photoelectric conversion unit and is configured to allow the transfer of a charge occurring in the second photoelectric conversion unit to the first photoelectric conversion unit. The signal detecting unit commonly detects the charge amount in the first photoelectric conversion unit and the second photoelectric conversion unit.
Abstract: An image pickup apparatus of the invention includes: an optical imaging system; an image pickup device; a defocus quantity calculation section for calculating a defocus quantity based on a phase difference between signals for focus detection obtained from pixels for focus detection; a high-frequency component quantity detection section for detecting a quantity of high-frequency component contained in a pixel signal outputted from a pixel other than the pixels for focus detection; a region determination section for determining in accordance with which of the defocus quantity and the high-frequency component quantity the optical imaging system is to be driven, according to a detection result of detecting in which of a plurality of divided blocks of the image pickup device an object of interest is present; and a focusing section for driving the optical imaging system in accordance with a determination result of the region determination section.
Abstract: A method and apparatus providing an autofocus routine in a camera apparatus having a processor is disclosed. The camera apparatus is adapted to detect a number of images and communicate image signals representative thereof to the processor. The method includes determining that a degree of change between a first image signal and a second image signal is below a predetermined threshold and responsive thereto, performing the autofocus routine.
February 16, 2010
Date of Patent:
February 24, 2015
Marc Drader, James Alexander Robinson, Michael Lorne Purdy
Abstract: An image capturing apparatus includes an image sensor 104 in which at least part of pixels arranged in two dimensions are configured as focus detection pixels with divided-pupil, a memory control circuit 113 configured to read out from a memory position information for the focus detection pixels 401, 402 stored in the memory, and a correction circuit 110 configured to identify positions of the focus detection pixels 401, 402 in the image sensor 104 based on the position information for the focus detection pixels 401, 402 and to correct a defective focus detection pixel signal using defect-free focus detection pixel signals.
Abstract: An optical function device includes a base material layer, a semi-transparent layer formed on a principal plane of the base material layer, the semi-transparent layer reflecting light of incident light at a ratio determined in advance and passing remaining light; and a reflection prevention layer formed on a principal plane opposite to the principal plane of the base material layer with respect to the base material layer, the reflection prevention layer preventing reflection of the light passing through the base material layer. The image-capturing device includes an optical function device, a first light receiving device for receiving transmission light from the optical function device, and a second light receiving device for receiving reflection light from the optical function device.
Abstract: A barrel assembly includes a barrel, at least one lens group disposed in the barrel to move in an optical axis direction, an aperture disposed in the barrel that adjusts an amount of light passing through the at least one lens group; and a light adjustment unit disposed in the barrel to move in the optical axis direction and that blocks light passing through a peripheral area of the at least one lens group when the at least one lens group is in at least one part of a travel section along which the at least one lens group moves in the barrel.
Abstract: An imaging apparatus and method enables an automated extended depth of field capability that automates and simplifies the process of creating extended depth of field images. An embodiment automates the acquisition of an image “stack” or sequence and stores metadata at the time of image acquisition that facilitates production of a composite image having an extended depth of field from at least a portion of the images in the acquired sequence. An embodiment allows a user to specify, either at the time of image capture or at the time the composite image is created, a range of distances that the user wishes to have in focus within the composite image. An embodiment provides an on-board capability to produce a composite, extended depth of field image from the image stack. One embodiment allows the user to import the image stack into an image-processing software application that produces the composite image.
Abstract: An image capturing apparatus comprising an image sensor including image forming pixels each of which generates a signal for image generation, and focus detection pixels each of which generates a signal for phase difference detection, readout means for reading out the signal of each pixel of the image sensor, focus detection means for detecting a focus by a phase-difference detection method using the signals for phase difference detection from the focus detection pixels, and switching means for switching a combination of the focus detection pixels to be used for focus detection by the focus detection means between a case in which the readout means reads out the signals of the pixels of the image sensor after thinning-out and a case in which the readout means reads out the signals of the pixels of the image sensor without thinning-out.
Abstract: An automatic focusing apparatus includes: a focus lens unit; a focus driver; a first focus detector that detects an in-focus state based on a phase difference; a second focus detector that detects an in-focus state using a signal from an image pickup element; a focus controller that controls the focus driver to perform focusing based on a first focus detection result and a second focus detection result; a movement detector that detects a movement of an object with components in a direction perpendicular to an optical axis based on an image signal obtained from the first focus detector; and a re-execution determination unit that determines whether to control the focus driver to re-execute focusing based on the first focus detection result and the second focus detection result that are newly detected after the execution of focusing and based on a movement detection result detected by the movement detector.
Abstract: In an imaging apparatus having a solid-state imaging device in which a focus detection pixel (phase difference detection pixel) is mounted on a light receiving surface, when dust is attached on the light receiving surface, a phase difference amount in a dust presence region is calculated from a detection signal of a phase difference detection signal and reliability of the phase difference amount is determined and when the reliability is high, phase difference AF control is performed with the phase difference amount calculated in the dust presence region.
Abstract: In an imaging device having an objective and a stage for holding a sample to be imaged, a method for autofocusing is presented. The method includes determining a measured focus value corresponding to at least a first of a plurality of logical image segments. Further, the method includes imaging the first logical image segment using the measured focus value. The method also includes determining a predicted focus value for a second of the plurality of logical image segments using the measured focus value and a stored focus variation parameter. In addition, the method includes imaging the second logical image segment using the predicted focus value.
April 13, 2009
Date of Patent:
November 4, 2014
General Electric Company
David LaVan Henderson, Kevin Bernard Kenny, Siavash Yazdanfar
Abstract: An image capturing apparatus comprising an image sensor including a plurality of pixels each including a plurality of photoelectric conversion units, and a first holding unit and a second holding unit configured to store signals output from the plurality of pixels and a driving unit configured to perform first write processing for writing first signals supplied from a first number of photoelectric conversion elements of each pixel in the first holding unit and second write processing for writing second signals supplied from a second number of photoelectric conversion units of each pixel.
Abstract: The capacitance of a charge-accumulating layer of an imaging pixel is made different from that of a charge-accumulating layer of a focusing pixel, thereby reducing the difference in saturation capacitance due to the difference between the light-reception efficiencies of the imaging pixel and the focusing pixel. The ratio between the capacitance of the charge-accumulating layer of the imaging pixel and that of the charge-accumulating layer of the focusing pixel is determined in consideration of a variation in ratio between the light-reception efficiencies of the imaging pixel and the focusing pixel with a change in at least one of the exit pupil distance and the aperture value.
Abstract: Image data is processed to facilitate focusing and/or optical correction. According to an example embodiment of the present invention, an imaging arrangement collects light data corresponding to light passing through a particular focal plane. The light data is collected using an approach that facilitates the determination of the direction from which various portions of the light incident upon a portion of the focal plane emanate from. Using this directional information in connection with value of the light as detected by photosensors, an image represented by the light is selectively focused and/or corrected.
August 26, 2013
Date of Patent:
October 21, 2014
The Board of Trustees of the Leland Stanford Junior University
Yi-Ren Ng, Partrick M. Hanrahan, Marc A. Levoy, Mark A. Horowitz
Abstract: The solid-state image sensor includes image sensing pixels, first and second focus detection pixels configured to respectively detect lights passing through different regions of a pupil of an image sensing lens. The sensor includes a semiconductor substrate including photoelectric converters of the image sensing pixels, a photoelectric converter and a first well contact region of the first focus detection pixel, and a photoelectric converter and a second well contact region of the second focus detection pixel, a first contact plug electrically connected to the first well contact region, and a second contact plug electrically connected to the second well contact region. The relative position of the first well contact region in the first focus detection pixel differs from a relative position of the second well contact region in the second focus detection pixel.
Abstract: The focus detection apparatus includes an image pickup part 107 including first and second pupil division pixels SHA and SHB to produce first and second image signals, a processor 121 performing a restoration process on the first and second image signals to produce first and second restored image signals, and a calculating part calculating a provisional value of a defocus amount of an image-forming optical system by using the first and second image signals. When the provisional value is smaller than a predetermined value, the restoration process is performed to produce pluralities of the first and second restored image signals by using a greater number of image restoration filters than when the provisional value is greater than the predetermined value, and the defocus amount is calculated by using the pluralities of the first and second image signals.
Abstract: An imaging apparatus includes: an optical system that forms an image corresponding to subject light incident through a lens; an imaging device that produces a signal corresponding to the subject light incident through the lens and outputs the signal as a captured image; acquisition means for acquiring the distance to the subject; and correction means for correcting blur in the captured image outputted from the imaging device based on an imaging characteristic of the optical system specific to the subject distance acquired by the acquisition means.