Imaging device and analyzing apparatus using the imaging device
An imaging device includes: an optical system having a lens and a diaphragm; an image sensor having a first pixel and a second pixel which a light that has passed through the optical system enters; and an optical element array positioned between the optical system and the image sensor, the optical system has an optical filter including a first region and a second region having different optical characteristics, the optical element array makes the light that has passed through the first region enter the first pixel and makes the light that has passed through the second region enter the second pixel, and an entrance pupil of the optical system is located between the diaphragm and an object.
Latest Panasonic Patents:
1. Technical Field
The present disclosure relates to an imaging device such as a camera and an analyzing apparatus using the imaging device.
2. Description of the Related Art
There is a growing need for an imaging device that not only acquires a color image, but also has other functions. In particular, recently, there has been an increase in research and development in the field of spectroscopic imaging that performs image acquisition for each of a plurality of wavelengths.
In spectroscopic imaging, by acquiring two-dimensional brightness information of an object for a plurality of wavelengths or wavelength bands, it is possible to sense useful information that is difficult or impossible to sense only by visual inspection of the object. These information includes, for example, the degree of freshness or the sugar content of vegetables or fruits, extraction of a foreign body in various product inspection lines, and identification of diseased tissue by analysis of fluorescence that is generated as a result of an object being irradiated with excitation light.
There are mainly two methods for performing spectroscopic imaging, One is a method by which a plurality of illumination light sources, each having a specific wavelength, are prepared and an object is illuminated thereby with switching being performed between the illumination light sources and, at the same time, images of the object are taken by an imaging device at emission of each illumination light source. The other is a method by which an object is illuminated by a light source, such as a white light source, which has a wide wavelength band and an image thereof is acquired by an imaging device via a spectral filter that allows only a desired wavelength to pass therethrough.
A suitable method of the above-described two methods is selected depending on an object to be measured or an environment and circumstances in which measurement is performed. The latter has the advantage that the flexibility for the number of wavelengths or the wavelength bandwidth that can be acquired is high and spectroscopic imaging can be performed relatively easily. As a specific example, a filter wheel provided with a plurality of spectral filters having different transmission wavelength bands is positioned in front of an imaging device and switching between the spectral filters is performed, whereby a plurality of images having different wavelength bands can be sequentially acquired.
U.S. Pat. No. 7,433,042 and Japanese Unexamined Parent Application Publication No. 2011-75562 disclose examples in which a plurality of wavelength images are acquired at the same time by one imaging device.
SUMMARYIn general, a wavelength that is allowed by a spectral filter to pass therethrough varies with the angle of incidence of a light beam entering the spectral filter. This makes it difficult to perform spectroscopic imaging at a wide angle of view in a desired narrow wavelength band.
One non-limiting and exemplary embodiment provides an imaging device and an analyzing apparatus that can perform spectroscopic imaging in a narrow wavelength band at a wide angle of view.
In one general aspect, the techniques disclosed here feature an imaging device including an optical system having a lens and a diaphragm, an image sensor having a first pixel and a second pixel which a light that has passed through the optical system enters, and an optical element array positioned between the optical system and the image sensor, in which the optical system has an optical filter including a first region and a second region having different optical characteristics, the optical element array makes a light that has passed through the first region enter the first pixel and makes a light that has passed through the second region enter the second pixel, and an entrance pupil of the optical system is located between the diaphragm and an object.
With the imaging device in the present disclosure, it is possible to perform spectroscopic imaging in a narrow wavelength band at a wide angle of view.
Additional benefits and advantages of the disclosed embodiments will become apparent from the specification and drawings. The benefits and/or advantages may be individually obtained by the various embodiments and features of the specification and drawings, which need not all be provided in order to obtain one or more of such benefits and/or advantages.
(Underlying Knowledge Forming Basis of the Present Disclosure)
The inventor has found out that the following problems arise in a wavelength band and an angle of view in spectroscopic imaging described in the “Description of the Related Art” section.
As described in the “Description of the Related Art” section, U.S. Pat. No. 7,433,042 and Japanese Unexamined Patent Application Publication No. 2011-75562 disclose the techniques of acquiring a plurality of wavelength images at the same time by one imaging device.
In general, the wavelength that is allowed by a spectral filter to pass therethrough varies with the angle of incidence of a light beam entering the spectral filter. This makes it difficult to perform spectroscopic imaging in a narrow wavelength band at a wide angle of view.
As the spectral filter, a calor filter using a material having wavelength dependence as absorption characteristics and an interference filter formed of an optical multilayer film formed as stacked films made of materials with different refractive indexes are representative examples. In particular, with the interference filter formed of an optical multilayer film, it is possible to implement a narrow-band-pass filter that allows only a specific wavelength to pass therethrough. Furthermore, this filter can implement various filter characteristics such as a band-pass filter, a high-pass filter, and a low-pass filter, and the light use efficiency thereof is higher than that of an absorption-type spectral filter and the degree of flexibility in setting spectral transmission characteristics is also high. On the other hand, the wavelength separated by transmission or reflection with respect to the angle of incidence of a light beam markedly shifts. In the imaging device, light beams emitted from an object are collected and, since the greater the angle of view of the imaging device becomes, the more likely the light beams are to enter the optical system obliquely, it is impossible to obtain desired spectral characteristics. In particular, in a layout in which the interference filter is positioned in front of the imaging device, the problem of the layout making it impossible to take images at a wide angle of view has become pronounced.
In the imaging device disclosed in U.S. Pat. No. 7,433,042, an example in which a spectral filter is provided in the imaging device is disclosed. However, since a spectral filter array is disposed in the position of an entrance pupil, a light beam that enters the spectral filter has an angular width, and the problem of this making it difficult to acquire light beam information on a specific narrow wavelength width and an image arises. Moreover, also in Japanese Unexamined Patent Application Publication No. 2011-75562, no consideration is given to the angle of incidence of a light beam entering the spectral filter, and therefore the same problem arises especially when an image of an object is taken at a wide angle of view.
To solve these problems, an imaging device according to an aspect of the present disclosure includes: an optical system having a lens and a diaphragm; an image sensor having a first pixel and a second pixel which the light that has passed through the optical system enters; and an optical element array positioned between the optical system and the image sensor, the optical system has an optical filter including a first region and a second region having different optical characteristics, the optical element array makes the light that has passed through the first region enter the first pixel and makes the light that has passed through the second region enter the second pixel, and an entrance pupil of the optical system is located between the diaphragm and an object.
With this configuration, it is possible to perform spectroscopic imaging in a narrow wavelength band at a wide angle of view. More specifically, it is possible to implement an imaging device that maintains a spectral wavelength at a high degree of accuracy even at the time of acquisition of a wide-angle image. The wide angle makes it possible to increase an object target area and reduce the number of imaging devices. Moreover, since it is possible to take an image of a target object at close range, it is possible to reduce the sizes of a system and analyzing equipment which are equipped with the imaging device.
Hereinafter, embodiments of the imaging device according to the present disclosure will be described with reference to the drawings.
Incidentally, all the embodiments which will be described below are comprehensive or specific examples. The numerical values, shapes, materials, component elements, placement positions and connection configurations of the component elements, steps, order of steps, and so forth which will be described in the following embodiments are mere examples and are not meant to limit the claims. Moreover, of the component elements in the following embodiments, a component element which is not described in an independent claim describing the broadest concept is described as an arbitrary component element.
(Embodiment 1)
The lens optical system L is formed of a first element optical system Li having at least one lens, a diaphragm S, an optical filter L2, and a second element optical system L3. The light beam direction of the light that has entered the imaging device A from an object (not depicted in the drawing) is bent by the first element optical system L1, and an unnecessary light beam is removed by the diaphragm S. The optical filter L2 has regions D1 and D2 that allow respectively narrow wavelength bands having different wavelengths λ1 and λ2 as the peaks thereof to pass therethrough and is positioned near the diaphragm S.
This lens optical system L is configured in such a way that an entrance pupil E is located in a position closer to the object than the diaphragm S. Here, the entrance pupil is an image of a diaphragm obtained when a lens is viewed from the side where an object is located, that is, an image of the diaphragm obtained by a lens group that is located from the diaphragm to the side where the subject is located. To be brief, the entrance pupil is an effective aperture of a lens that is determined by a diaphragm. It is for this reason that the entrance pupil is also called an effective aperture. Moreover, the position of an entrance pupil is defined as a position in which a main light beam in an object space intersects with the optical axis after being extended as it is. The object space refers to a space from the object to the entrance to the imaging device A (in
Moreover, in
The pencil of rays B0 passes through the first element optical system L1 and reaches the diaphragm S and the optical filter L2. In
Furthermore, the optical elements M1 of the optical element array K are positioned so as to be located on the side where the imaging surface Ni is located and are configured such that one optical element M1 corresponds to two lines of pixels formed of the pixels P1 and the pixels P2 on the imaging surface Ni.
With such a configuration, most of the pencils of rays (solid lines of
Here, by the signal processing section C depicted in
The first image I1 and the second image I2 are images obtained by the passage through the region D1 and the region D2 of the optical filter L2, respectively; in this way, it is possible to acquire images having different wavelengths λ1 and λ2 at the same time.
The optical filter L2 has formed therein a filter with a narrow wavelength band. Hereinafter, as a representative example, a Fabry-Perot interference filter formed of a dielectric multilayer film, the Fabry-Perot interference filter depicted in
When light beams enter such a structure, the light beams are repeatedly reflected by the reflection interference surfaces 51a and 51b and interfere with each other, and only a light with a wavelength that resonates in the spacer layer 52 passes through the filter. If a transmission wavelength is assumed to be λ, the following Expression (1) is obtained.
kλ=2nd cos θ (1)
Here, k is a natural number and θ is an angle formed by a filter normal and an incident light beam.
in the imaging device of this embodiment, as depicted in
Incidentally, in the first element optical system L1 it is desirable that at least one lens is an optical system having negative light-collecting power. This makes it easier to make the entrance pupil E get closer to the object than the diaphragm S.
Moreover, the imaging device may be configured such that no second element optical system L3 is provided and an image is taken by the first element optical system L1 and the diaphragm S.
Furthermore, the optical filter L2 does not necessarily have to be a flat surface and may be a curved surface. The regions D1 and D2 may have different shapes. By appropriately providing the regions D1 and D2 with different shapes, it is possible to reduce aberration such as chromatic aberration.
In this embodiment, as depicted in
In this embodiment, the optical filter L2 has the regions D1 and D2 that allow respectively narrow wavelength bands having different wavelengths λ1 and λ2 as the peaks thereof to pass therethrough. However, the optical filter L2 is not limited thereto as long as the regions D1 and D2 have different optical characteristics. In particular, the greater the change in the optical characteristics for the angle of incidence of a light beam entering the optical filter L2, the more effective. For example, in a polarizing filter, the amount of transmitted light changes depending on an angle of incidence, and the amount of transmitted light is reduced as the angle of incidence gets greater and a brightness distribution is generated in a camera image. The imaging device in the present disclosure is highly effective also in a case in which such a polarizing filter is used and is effective, in addition to the spectral and polarizing filters, in all types of elements in which the optical characteristics change by the angle of incidence. Moreover, a spectral means is not limited to the Fabry-Perot interference filter.
Furthermore, the region D1 and the region D2 of the optical filter L2 have shapes symmetrical with respect to the optical axis V, but the shapes of the region D1 and the region D2 are not limited thereto. The region D1 and the region D2 of the optical filter L2 may have shapes asymmetrical with respect to the optical axis V and the areas thereof may be different from each other. In this case, image generation is performed in consideration of the correlation with each pixel on the imaging surface.
(Embodiment 2)
Embodiment 2 differs from Embodiment 1 in that an optical filter L2 is divided into four regions and a lenticular lens is replaced with a microlens as the optical element array.
Furthermore, the optical elements M2 of the optical element array K are positioned so as to be located on the side where the imaging surface Ni is located and are configured such that one optical element M2 corresponds to four pixels: the pixels P1 to P4 on the imaging surface Ni.
With such a configuration, most of the pencils of rays that have passed through the region D1 the region D2, the region D3, and the region D4 on the optical filler L2 depicted in
Here, as is the case with Embodiment 1, by a signal processing section C, a first image I1, a second image I2, a third image I3, and a fourth image I4 are output.
As a result of providing the region D1, the region D2, the region D3, and the region D4 of the optical filter L2 with different transmission wavelengths, the imaging device A can acquire four images having different wavelength information: the first image I1, the second image I2, the third image I3, and the fourth image I4 at the same time.
Also in this embodiment, as is the case with Embodiment 1, since it is possible to make the angle ωmax which a pencil of rays Bmax entering the imaging device forms with the optical axis V greater, it is possible to implement the imaging device as a wider-angle imaging device without a deviation from a desired wavelength and produce the same effect.
In this embodiment, the optical filter L2 is divided into four different regions; however, the number of divisions may be further increased and the number of pixels corresponding to each optical element M2 may be appropriately changed.
Moreover, the region D1, the region D2, the region D3, and the region D4 of the optical filter L2 have shapes symmetrical with respect to the optical axis V, but the shapes of the region D1, the region D2, the region D3, and the region D4 are not limited thereto. The region D1, the region D2, the region D3, and the region D4 of the optical filter L2 may have shapes asymmetrical with respect to the optical axis V and the areas thereof may be different from each other. The number of pixels corresponding to each optical element may be different from the number of divisions of the optical filter L2.
(Embodiment 3)
Embodiment 3 differs from Embodiments 1 and 2 in that a lenticular lens or a microlens array which is an optical element array is formed in an image sensor.
(Comparative Example)
As a comparative example for the imaging device in the present disclosure, an imaging device of
The lens optical system L is formed of a diaphragm 5, an optical filter L2, and a second element optical system L3 having at least one lens. A light that has entered the imaging device A from an object (not depicted in the drawing) enters the optical filter L2 after an unnecessary light beam is removed therefrom by the diaphragm S. The optical filter L2 has regions D1 and D2 that allow respectively narrow wavelength bands having different wavelengths λ1 and λ2 as the peaks thereof to pass therethrough and is positioned near the diaphragm S.
The imaging device of
(Embodiment 4)
An imaging device of this embodiment differs from the imaging device of Embodiment 1 in that a mirror is provided in a lens optical system. Hereinafter, a component element which is different from the counterpart thereof in Embodiment 1 will be described.
The lens optical system L is formed of the mirror Mr, a diaphragm S, an optical filter L2, and a second element optical system L3. The light beam direction of the light that has entered the imaging device A from an object (not depicted in the drawing) is bent by the mirror Mr, and an unnecessary light beam is removed by the diaphragm S. The optical filter L2 has regions D1 and D2 that allow respectively narrow wavelength bands having different wavelengths λ1 and λ2 as the peaks thereof to pass therethrough and is positioned near the diaphragm S.
Moreover, in
In the imaging device of this embodiment, as depicted in
Incidentally, it is preferable that the mirror Mr is a convex mirror. This makes it easier to make the entrance pupil E get closer to the object than the diaphragm S. Moreover, a lens or another mirror may be provided between the mirror Mr and the optical filter L2, Furthermore, the imaging device may be configured such that no second element optical system L3 is provided and an image is taken by an element optical system and the diaphragm S, the element optical system including at least the mirror Mr.
Incidentally, since the mirror is used, this embodiment is an asymmetrical imaging optical system with respect to the optical axis V. When an axisymmetric optical system is used in the second element optical system, even the pencils of rays BL and BR that have entered the imaging device at the same angle of view have different image heights on the imaging surface Ni on the image sensor N. By performing asymmetrical correction on the distortion in the image by the signal processing section C, it is possible to obtain an image having left-right or top-bottom symmetry.
(Embodiment 5)
An imaging device of this embodiment differs from the imaging device of Embodiment 1 in that light beams from an object are converted into approximately-parallel lights by a first element optical system L1, that is, collimated thereby. Hereinafter, a component element which is different from the counterpart thereof in Embodiment 1 will be described.
In this embodiment, the “approximately-parallel lights” refer to, for example, lights between which a difference in light beam direction is 5 degrees or less.
The lens optical system L is formed of a first element optical system L1 having at least one lens, a diaphragm S, an optical filter L2, and a second element optical system L3. The light beam direction of the light that has entered the lens optical system L from an object (not depicted in the drawing) is bent by the first element optical system L1, and an unnecessary light beam is removed by the diaphragm S.
Incidentally, in
Moreover, in
The pencil of rays B0 passes through the first element optical system L1 and reaches the diaphragm S and the optical filter L2 as an approximately-parallel light. In
Moreover, as depicted in
The greater the angle which a pencil of rays from the object forms with the optical axis V, the greater the angle θc. That is, the angle θc becomes the greatest in the pencil of rays Bmax, and, in the pencil of rays B0, the angle θc comes closer to a state in which it is parallel to the optical axis V. As is clear from Expression (1), the greater the angle of incidence to a filter, the wavelength λ passing through the filter shifts to the short wavelength side. In consideration of this, it is preferable to select a wavelength between a transmission wavelength of a pencil of rays entering the imaging device from the object along the optical axis V, that is, the normal of the filter and a transmission wavelength of a pencil of rays entering the imaging device at the maximum angle of view ωmax as transmission wavelength bands of the regions D1 and D2 because this makes it possible to acquire an image of a wavelength in an extremely narrow band,
The imaging device of this embodiment has an advantage that the size of the imaging device in an optical axis direction, in particular, can be reduced when, in particular, image acquisition requiring a high degree of wavelength precision is performed. For example, if the angles θc, θt, and θb of incidence to the optical filter L2 are different from one another as in
Incidentally, in the first element optical system L1, it is preferable that at least one lens has negative light-collecting power. This makes it easier to make the entrance pupil E get closer to the object than the diaphragm S.
In this embodiment, the second element optical system L3 is formed of one lens, but the second element optical system L3 may be formed of optical parts such as a plurality of lenses.
Moreover, in this embodiment, for example, an example in which conversion is performed to obtain approximately-parallel lights between which a difference in light beam direction is 5 degrees or less has been described, but the example is not limited thereto. When a pencil of rays entering the imaging device at the maximum angle of view enters the first region of the optical filter L2, the maximum value (for example, a difference between θt and θc in
(Embodiment 6)
An imaging device of this embodiment differs from the imaging device of Embodiment 5 in that no optical element array is provided, the focus of a lens optical system L is located on an imaging surface Ni of an image sensor N, an optical element is formed of two optical filters: an optical filter L2a and an optical filter L2b, and the optical filters L2a and L2b are fixed to a movable holder H.
In the imaging device of this embodiment, as depicted in
Next, by operating the holder H, the state is changed to a state in which the optical axis V passes through the optical filter L2b. The optical filter L2b allows a narrow wavelength band with a wavelength λb as the peak thereof to pass therethrough. The angle at which an arbitrary pencil of rays B entering the imaging device A from an object located in an oblique direction (an angle which the pencil of rays B forms with the optical axis V is ω) enters the optical filter L2b is the same as the angle at which the pencil of rays B enters the optical filter L2a, and the same effect is produced except for a different wavelength of an image to be obtained. A signal processing section C obtains two types of images in cooperation with the operation of the holder H and outputs these images as a first image I1 and a second image I2.
Incidentally, in this embodiment, the lens optical system L of the imaging device of Embodiment 1 is used; however, this embodiment is also effective for other lens optical systems including Embodiment 2 and the embodiments that follow. The number of types of optical elements is not limited to two. Moreover, the holder H is not limited to a sliding holder, and a rotary wheel-type holder may be adopted.
In particular, the lens optical system L of Embodiment 5 has the advantage that, since the use thereof makes the angles θc, θt, and θb at which light beams enter the optical filter L2 become almost the same angle, the size of the imaging device in an optical axis direction can be reduced when image acquisition requiring a high degree of wavelength precision is performed.
(Embodiment 7)
An analyzing apparatus using the imaging devices described in Embodiments 1 to 6 will be hereinafter described.
A food analyzing apparatus 1 has a housing 11, a table 12, an imaging device A, a processor 30, an operating section 40, and a display section 50.
The table 12 and the imaging device A are positioned in the housing 11. On the table 12, an object to be measured Sc is positioned. Incidentally, the food analyzing apparatus 1 can analyze one or a plurality of foods as the object to be measured Sa. Moreover, the food analyzing apparatus 1 can also analyze a food in a container or the like as the object to be measured Sa.
The imaging device A has a light source 21 that irradiates the object to be measured Sa positioned on the table 12 with a light. The imaging device A has the function of receiving the light reflected from the object to be measured Sa and performing imaging.
The light source 21 is placed in a position in which the light source 21 can irradiate the whole of the object to be measured Sa with a light. The light emitted from the light source 21 includes some of wavelengths of at least 700 to 2500 nm. As the light source 21, for example, a halogen lamp, an LED, a laser, or the like is used.
The imaging device A acquires an image with a first specific wavelength, an image with a second specific wavelength, and an image with a third specific wavelength.
The first to third specific wavelengths are determined by an experiment or the like based on the spectral information of a plurality of foods whose ingredients are known. Specifically, based on the relationship between the ratio of a specific ingredient in a plurality of foods and the absorbance, a wavelength well reflecting the ratio of the specific ingredient in foods is determined as a specific wavelength.
As the first specific wavelength, a wavelength having a strong correlation with protein as an ingredient is adopted. For example, as the first specific wavelength, 910 nm and a wavelength near this wavelength can be adopted.
As the second specific wavelength, a wavelength having a strong correlation with lipid as an ingredient is adopted. For example, as the second specific wavelength, 930 nm and a wavelength near this wavelength can be adopted.
As the third specific wavelength, a wavelength having a strong correlation with carbohydrate as an ingredient is adopted. For example, as the third specific wavelength, 980 nm and a wavelength near this wavelength can be adopted.
As an image sensor of the imaging device A, for example, an element using silicon having sensitivity over a wide range in a near-infrared region and indium/gallium/arsenic, the element that can convert the amount of light into an electrical signal, can be used.
With reference to
The operating section 40 has a measurement button 41 and a switching button 42. When the measurement button 41 is pressed, the operating section 40 outputs, to the processor 30, a signal indicating that the measurement button 41 has been pressed. When the switching button 42 is pressed, the operating section 40 outputs, to the processor 30, a signal indicating that the switching button 42 has been pressed.
When receiving the signal indicating that the measurement button 41 has been pressed, the processor 30 controls the imaging device A and starts the analysis of the object to be measured Sa. When receiving the signal indicating that the switching button 42 has been pressed, the processor 30 changes the contents which the display section 50 is made to display.
The processor 30 makes the light source 21 momentarily irradiate the object to be measured Sa with a light including a near-infrared light. The light scattered or reflected from the object to be measured Sa enters the imaging device A.
The imaging device A acquires an image with the first specific wavelength, an image with the second specific wavelength, and an image with the third specific wavelength without displacement. Thus, the outputs of the pixels forming the image with the first specific wavelength, the image with the second specific wavelength, and the image with the third specific wavelength reflect part of protein, lipid, and carbohydrate in the object to be measured. That is, the imaging device A outputs, to the processor 30 (see
The processor 30 performs computations to obtain the ratio and amount of protein based on the output of the image with the first specific wavelength and a previously stored relational expression.
The processor 30 performs computations to obtain the ratio and amount of lipid based on the output of the image with the second specific wavelength and a previously stored relational expression.
The processor 30 performs computations to obtain the ratio and amount of carbohydrate based on the output of the image with the third specific wavelength and a previously stored relational expression.
Incidentally, each relational expression can be determined in advance by, for example, using PLS based on the relationship between the absorbance (the amount of light) at each wavelength of a plurality of foods containing the above ingredients and the ratio of each ingredient.
The calorie of a food is obtained by multiplying each of the amount of protein, the amount of lipid, and the amount of carbohydrate by a calorie coefficient and summing the results. It is for this reason that the processor 30 performs computations to obtain the calorie of each measurement site based on the amount of protein, the amount of lipid, and the amount of carbohydrate in each measurement site. Incidentally, the ratio of protein, the amount of protein, the ratio of lipid, the amount of lipid, the ratio of carbohydrate, the amount of carbohydrate, and the calorie in each measurement site correspond to “partial nutrition information”.
The processor 30 creates, as distribution image information, distribution image information indicating the calorie and distribution image information indicating the ingredient. The processor 30 outputs the distribution image information to the display section 50 and makes the display section 50 display a distribution image P. Moreover, based on the output of the operating section 40, the processor 30 performs switching between the distribution image information indicating the calorie and the distribution image information indicating the ingredient, the distribution image information to be output to the display section 50.
With reference to
As depicted in
The calorie distribution image P visually expresses the magnitude of the calorie of each partial region PX by relating the density of a display color of each partial region PX to the magnitude of the calorie.
As depicted in
The operation of the food analyzing apparatus 1 will be described.
In a virtual food analyzing apparatus that numerically displays the calorie and the ingredients of the whole of an object to be measured Sa, for example, if the calorie or the ratio of an ingredient of the object to be measured Sa is higher than the user's target calorie or ratio of the ingredient, the user removes part of the object to be measured Sa and repeats the operation of performing the analysis by the food analyzing apparatus again. This requires a good deal of user's efforts.
Since the food analyzing apparatus 1 displays the distribution image P in the display section 50, the user can easily figure out the distribution of the calorie and ingredients of the object to be measured Sa. As a result, if the calorie or the ratio of an ingredient of the object to be measured Sa is higher than the user's target calorie or ratio of the ingredient, the user can easily figure out which part and how much part of the object to be measured Sa the user should remove to attain the target calorie or ratio of the ingredient. This improves the convenience of the user. Moreover, it is possible to attain the target calorie or ratio of the ingredient in a short time. This makes it possible to prevent the taste, shape, and so forth of the food from changing as a result of the food being altered by temperature, for example, due to a time-consuming adjustment.
In the virtual food analyzing apparatus that numerically displays the calorie and the ingredients of the whole of an object to be measured Sa, when the user performs analysis of a plurality of foods, for example, the user repeats the operation of performing the analysis for each food. This requires a good deal of user's efforts.
Since the food analyzing apparatus 1 displays the distribution image P in the display section 50, it is possible to figure out the calorie and ingredients of each of a plurality of foods by one measurement. This improves the convenience of the user.
The food analyzing apparatus 1 of this embodiment has the following advantages.
(1) The food analyzing apparatus 1 displays the distribution image P in the display section 50. This makes it easier for the user to figure out the distribution of partial nutrition information including the calorie and ingredients of an object to be measured Sa.
(2) The food analyzing apparatus 1 can obtain the partial nutrition information of the whole of the object to be measured Sa by computation by the imaging device A in a short time. As described in Embodiments 1 to 6, the imaging device A makes it possible to maintain a near-infrared spectral wavelength at a high degree of accuracy even at the time of acquisition of a wide-angle image and thereby perform a high-accuracy calorie analysis. The wide angle makes it possible to take an image of the object to be measured Sa at close range, making it possible to reduce the size of the food analyzing apparatus 1.
(3) The food analyzing apparatus 1 makes the light source 21 emit a light momentarily. This makes the object to be measured Sa less likely to be warmed by the light or changed such as undergoing degeneration.
(4) The food analyzing apparatus 1 analyzes the object to be measured Sa nondestructively. This makes it possible to use the object to be measured Sa after measurement for food as it is.
(5) The food analyzing apparatus 1 analyzes the object to be measured Sa by using a near-infrared light. This makes it possible to perform an analysis without equipment such as a centrifugal machine which is used when a chemical analysis of the object to be measured Sa is performed by crushing the object to be measured Sa, a reagent, and so forth.
In this embodiment, the food analyzing apparatus equipped with any one of the imaging devices described in Embodiments 1 to 6 has been described, but an apparatus equipped with the imaging device is not limited thereto. All the apparatuses and systems that acquire an image having a plurality of optical information for an optical element having an incident angle dependence and use the image can acquire subject information at a high degree of accuracy as small apparatuses and systems.
Moreover, if necessary, a plurality of imaging devices described in Embodiments 1 to 6 may be installed or another camera having the function of, for example, acquiring a color image may be additionally used. Alternatively, the imaging device A may be provided with the function of acquiring color spectral information such as R, G, and B.
As described above, the imaging device according to the present disclosure includes an optical system (a lens optical system L) having lenses L1 and L3 and a diaphragm S, an image sensor N having first to nth pixels which the light that has passed through the optical system L enters, and an optical element array K positioned between the optical system and the image sensor N. The optical system L has an optical filter L2 having a first region and a second region having different optical characteristics. The optical element array K makes the light that has passed through the first region enter the first pixel and makes the light that has passed through the second region enter the second pixel. An entrance pupa E of the optical system L is located between the diaphragm S and an object.
This makes it possible to perform spectroscopic imaging in a narrow wavelength band at a wide angle of view. More specifically, it is possible to implement an imaging device that maintains a spectral wavelength at a high degree of accuracy even at the time of acquisition of a wide-angle image. The wide angle makes it possible to increase an object target area and reduce the number of imaging devices. Moreover, since it is possible to take an image of a target subject at dose range, it is possible to reduce the sizes of a system and an analyzing apparatus which are equipped with the imaging device.
Here, the optical system L may have a first element optical system L1 having negative light-collecting power, the first element optical system L1 guiding the incident light to the diaphragm S and the optical filter L2.
Here, the first element optical system L1 may have a concave lens.
Here, the first element optical system L1 may have a mirror Mr with a convex surface, the mirror Mr reflecting the incident light toward the diaphragm S and the optical filter L2.
Here, when a pencil of rays entering the imaging device at the maximum angle of view enters the first region of the optical filter L2, the maximum value of differences between the angles of incidence of the light beams included in the pencil of rays may be smaller than the maximum value of the angles of incidence of the light beams included in the pencil of rays, and, when the pencil of rays entering the imaging device at the maximum angle of view enters the second region of the optical filter L2, the maximum value of differences between the angles of incidence of the light beams included in the pencil of rays may be smaller than the maximum value of the angles of incidence of the light beams included in the pencil of rays.
Here, a holder H that interchangeably holds at least one of the first region and the second region in the optical filter L2 above an optical axis of the optical system may be further provided.
Here, the optical filter L2 may be a spectral filter or a polarizing filter.
Here, a lenticular lens may be positioned in a plane of the optical element array K, the plane facing the image sensor N.
Here, a microlens array may be positioned in a plane of the optical element array K, the plane facing the image sensor N.
Here, the optical element array Ni may be positioned on the image sensor N.
Here, microlenses provided between the optical element array and the image sensor may be further provided, and the optical element array may be positioned on the image sensor with the microlenses positioned between the optical element array and the image sensor.
Moreover, an analyzing apparatus according to the present disclosure includes a light source that irradiates an object to be analyzed with a light, the above-described imaging device that receives at least one of a group of a light reflected from the object to be analyzed, a light scattered from the object to be analyzed, and a light that has passed through the object to be analyzed, and a processor that performs computation on the light received by the imaging device.
Moreover, another analyzing apparatus according to the present disclosure includes a light source that irradiates an object to be analyzed with a light including at least part of near-infrared wavelengths of 700 nm or more, the above-described imaging device that receives at least one of a light reflected from the object to be analyzed and a light that has passed through the object to be analyzed, a processor that obtains the amount of absorbed light of the light received by the imaging device by computation, and an analyzer that performs at least one of calculating the calorie of the object to be analyzed based on the correlation between the amount of absorbed light and the calorie and the amount of absorbed light obtained by the processor by computation and calculating the ingredient amount of the object to be analyzed based on the correlation between the amount of absorbed light and the ingredient amount of a food and the amount of absorbed light obtained by the processor by computation.
While the imaging device and the analyzing apparatus according to one aspect of the present disclosure has been described based on the embodiments, the present disclosure is not limited to these embodiments. What is obtained by applying various modifications conceived of by a person skilled in the art to the embodiments or any configuration obtained by combining the component elements in different embodiments is also included in the scope of one or a plurality of aspects of the present disclosure within the scope of the present disclosure.
The imaging device according to the present disclosure is especially useful for the purpose of acquiring a spectral image or a polarization image and performing sensing. The range of application of the imaging device includes a food analyzing apparatus, various product inspection lines, medical uses, and so forth, and the imaging device can sense useful information in a wide range of application area and reduce the sizes of equipment and systems. Moreover, the imaging device according to the present disclosure can also be applied to the uses such as a car-mounted camera, a security camera, biometric authentication, a microscope, and an astronomical telescope.
Claims
1. An imaging device used to take an image of an object, the imaging device comprising:
- an optical system including an optical filter, a first element optical system, a lens and a diaphragm;
- an image sensor having a first pixel and a second pixel which a light that has passed through the optical system enters; and
- an optical element array positioned between the optical system and the image sensor, wherein
- the optical filter includes a first region and a second region having different optical characteristics,
- the optical element array makes a light that has passed through the first region enter the first pixel and makes a light that has passed through the second region enter the second pixel,
- an entrance pupil of the optical system is located between the diaphragm and the object,
- the first element optical system includes a mirror with a convex surface, has negative light-collecting power and guides an incident light to the diaphragm and the optical filter, and
- the minor reflects the incident light toward the diaphragm and the optical filter.
2. The imaging device according to claim 1, wherein
- the optical filter is a spectral filter or a polarizing filter.
3. The imaging device according to claim 1, wherein
- a lenticular lens is positioned in a plane of the optical element array, the plane facing the image sensor.
4. The imaging device according to claim 1, wherein
- a microlens array is positioned in a plane of the optical element array, the plane facing the image sensor.
5. The imaging device according to claim 1, wherein
- the optical element array is positioned on the image sensor.
6. The imaging device according to claim 5, further comprising:
- microlenses positioned between the optical element array and the image sensor, wherein
- the optical element array is positioned on the image sensor with the microlenses positioned between the optical element array and the image sensor.
7. An analyzing apparatus comprising:
- a light source that irradiates an object to be analyzed with a light;
- the imaging device according to claim 1, the imaging device receiving at least one selected from the group consisting of a light reflected from the object to be analyzed, a light scattered from the object to be analyzed, and a light that has passed through the object to be analyzed; and
- a processor that performs computation on the light received by the imaging device.
8. An analyzing apparatus comprising:
- a light source that irradiates an object to be analyzed with a light including at least part of near-infrared wavelengths of 700 nm or more;
- the imaging device according to claim 1, the imaging device receiving at least one selected from the group consisting of a light reflected from the object to be analyzed and a light that has passed through the object to be analyzed;
- a processor that calculates an amount of absorbed light of the light received by the imaging device; and
- an analyzer that performs at least one selected from the group consisting of calculating a calorie of the object to be analyzed based on a correlation between an amount of absorbed light and a calorie and based on the amount of absorbed light calculated by the processor and calculating an ingredient amount of the object to be analyzed based on a correlation between the amount of absorbed light and an ingredient amount of a food and based on the amount of absorbed light calculated by the processor.
9. An imaging device used to take an image of an object, the imaging device comprising:
- an optical system including an optical filter, a lens and a diaphragm;
- an image sensor having a first pixel and a second pixel which a light that has passed through the optical system enters; and
- an optical element array positioned between the optical system and the image sensor, wherein
- the optical filter includes a first region and a second region having different optical characteristics,
- the optical element array makes a light that has passed through the first region enter the first pixel and makes a light that has passed through the second region enter the second pixel,
- an entrance pupil of the optical system is located between the diaphragm and the object,
- when a pencil of rays entering the imaging device at a maximum angle of view enters the first region of the optical filter, a maximum value of differences between angles of incidence of light beams included in the pencil of rays is smaller than a maximum value of the angles of incidence of the light beams included in the pencil of rays, and
- when a pencil of rays entering the imaging device at a maximum angle of view enters the second region of the optical filter, a maximum value of differences between angles of incidence of light beams included in the pencil of rays is smaller than a maximum value of the angles of incidence of the light beams included in the pencil of rays.
10. The imaging device according to claim 9, wherein:
- the optical system further includes a first element optical system, and
- the first element optical system includes a concave lens, has negative light-collecting power and guides an incident light to the diaphragm and the optical filter.
11. The imaging device according to claim 9, wherein
- the optical filter is a spectral filter or a polarizing filter.
12. The imaging device according to claim 9, wherein
- a lenticular lens is positioned in a plane of the optical element array, the plane facing the image sensor.
13. The imaging device according to claim 9, wherein
- a microlens array is positioned in a plane of the optical element array, the plane facing the image sensor.
14. The imaging device according to claim 9, wherein
- the optical element array is positioned on the image sensor.
15. The imaging device according to claim 14, further comprising:
- microlenses positioned between the optical element array and the image sensor, wherein
- the optical element array is positioned on the image sensor with the microlenses positioned between the optical element array and the image sensor.
16. An imaging device used to take an image of an object, the imaging device comprising:
- an optical system including a lens, a diaphragm and an optical filter that includes a first region and a second region, the first and second regions having different optical characteristics;
- a holder that interchangeably holds at least one selected from the group consisting of the first region and the second region in the optical filter on an optical axis of the optical system;
- an image sensor having a first pixel and a second pixel which a light that has passed through the optical system enters; and
- an optical element array positioned between the optical system and the image sensor, wherein
- the optical element array makes a light that has passed through the first region enter the first pixel and makes a light that has passed through the second region enter the second pixel, and
- an entrance pupil of the optical system is located between the diaphragm and the object.
17. The imaging device according to claim 16, wherein:
- the optical system further includes a first element optical system, and
- the first element optical system includes a concave lens, has negative light-collecting power and guides an incident light to the diaphragm and the optical filter.
18. The imaging device according to claim 16, wherein
- the optical filter is a spectral filter or a polarizing filter.
19. The imaging device according to claim 16, wherein
- a lenticular lens is positioned in a plane of the optical element array, the plane facing the image sensor.
20. The imaging device according to claim 16, wherein
- a microlens array is positioned in a plane of the optical element array, the plane facing the image sensor.
21. The imaging device according to claim 16, wherein
- the optical element array is positioned on the image sensor.
22. The imaging device according to claim 16, further comprising:
- microlenses positioned between the optical element array and the image sensor, wherein
- the optical element array is positioned on the image sensor with the microlenses positioned between the optical element array and the image sensor.
6326998 | December 4, 2001 | Palum |
7433042 | October 7, 2008 | Cavanaugh et al. |
20100241357 | September 23, 2010 | Chan |
20110073752 | March 31, 2011 | Berkner et al. |
20120182438 | July 19, 2012 | Berkner et al. |
20120268643 | October 25, 2012 | Imamura |
20140055661 | February 27, 2014 | Imamura et al. |
2002-287032 | October 2002 | JP |
2011-075562 | April 2011 | JP |
2012-150112 | August 2012 | JP |
2013-258647 | December 2013 | JP |
2013/114891 | August 2013 | WO |
2013/179620 | December 2013 | WO |
2014/045596 | March 2014 | WO |
Type: Grant
Filed: Mar 19, 2015
Date of Patent: Nov 29, 2016
Patent Publication Number: 20150281535
Assignee: PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD. (Osaka)
Inventors: Tsuguhiro Korenaga (Osaka), Norihiro Imamura (Osaka), Keiichi Matsuzaki (Kyoto), Kazuhiro Ochi (Kyoto)
Primary Examiner: Twyler Haskins
Assistant Examiner: Padma Haliyur
Application Number: 14/663,341
International Classification: H04N 5/225 (20060101); G02B 5/20 (20060101); G02B 5/10 (20060101); G02B 7/00 (20060101); G02B 3/00 (20060101); G02B 13/00 (20060101); G01N 21/359 (20140101); G01N 33/02 (20060101); G01N 21/31 (20060101); H04N 5/238 (20060101);