OBJECT IDENTIFICATION DEVICE, METHOD, AND STORAGE MEDIUM
Disclosed is an object identification device and the like for reducing identification error in a reference image which presents an object that is only slightly difference from an object presented in an input image. The object identification device includes a local feature quantity matching unit for calculating geometric transformation information for transformation from a coordinate in a reference image to a corresponding coordinate in an input image, and matching a local feature quantity extracted from the reference image and a local feature quantity extracted from the input image, an input image different area determination unit for transforming the different area in the reference image on a basis of the geometric transformation information about the input image determined to be in conformity by the matching, and determining a different area in the input image corresponding to the different area in the reference image, an input image different area feature quantity extraction unit for correcting a different area in the input image, and extracting a feature quantity from the corrected different area of the input image, and a feature quantity matching unit for matching a feature quantity extracted by the input image different area feature quantity extraction unit and a feature quantity extracted from the different area in the reference image, and outputting a matching result.
The present invention relates to a device, a method, and a storage medium for accurately identifying an object in an image.
BACKGROUND ARTA method has been suggested to extract feature quantities in a local area (local feature quantities) around each feature point by detecting many distinctive points in an image (feature points) in order to identify a subject in the image in a robust manner against changes in the size of image-capturing and the angle, and occlusion. A local feature quantity extraction device using SIFT (Scale Invariant Feature Transform) feature quantity is disclosed in PTL1 and NPL1 as a typical method.
First, a local feature quantity extraction device extracts information about brightness from each pixel of an image, and detects many distinctive points (feature points) from the extracted brightness information. Then, the local feature quantity extraction device outputs feature point information which is information about each of the detected feature points. In this case, for example, the feature point information indicates the coordinate position and the scale of a detected local feature point, the orientation of the feature point, and the like. Then, the local feature quantity extraction device obtains a local area, where feature quantity extraction is performed, from the feature point information, i.e., the coordinate value, the scale, the orientation, and the like of each of the detected feature points, and generates (describes) local feature quantities.
Then, an image including the same subject as the subject in the captured image is identified by comparing a local feature quantity 1 extracted from the captured image (i.e., input image) with a local feature quantity 2 generated from a referred image as described in NPL 1. More specifically, first, distances of all the combinations of the feature quantities describing information about each feature point constituting the local feature quantity 1 and the feature quantities constituting the local feature quantity 2 are calculated in the feature space. Then, a combination of the local feature quantity 1 and the local feature quantity 2 of which calculated distance is the closest is determined to be corresponding feature quantities. Further, a combination of feature points which are the sources for generating these corresponding feature quantities is also determined to be corresponding. Thereafter, a determination is made as to whether the combination of feature points determined to be corresponding moves according to particular geometric transformation information from the coordinate position of the feature point in the input image to the coordinate position of the feature point in the reference image. Whether the corresponding feature points are correct or incorrect is determined on the basis of the determination of this movement. In this case, when the number of feature points determined to be correctly corresponding is equal to or more than a preset value, the same subject is determined to be shown.
CITATION LIST Patent Literature
- [PTL 1] Specification of U.S. Pat. No. 6,711,293
- [PTL 2] Japanese Patent Application Laid-Open No. 2010-79545
- [PTL 3] WO2012/108088
- [PTL 4] Japanese Patent Application Laid-Open No. 2010-266964
- [PTL 5] Japanese Patent Application Laid-Open No. 2010-128925
- [PTL 6] Japanese Patent Application Laid-Open No. 2007-115109
- [NPL 1] David G. Lowe, “Distinctive image features from scale-invariant keypoints”, (United States), International Journal of Computer Vision, 60(2), November, 2004, p. 91-110
In an object identification method using a related local feature quantity, an object is identified on the basis of a correspondence relationship of a local feature quantity extracted from brightness information of an input image and a local feature quantity extracted from brightness information of a reference image. As a result, in a case where an image showing an object that is only slightly difference from an object shown in an input image is adopted as a reference image, there are many corresponding feature points between the input image and the reference image, and therefore, there is a problem in that both of the images are falsely identified as an image showing the same object.
It is a main object of present invention to provide a device, a method, and a storage medium for suppressing identification error in a reference image which presents an object that is only slightly difference from an object presented in an input image, and more accurately identifying a reference image showing the identical object.
Solution to ProblemAn object identification device according to the present invention includes: a local feature quantity matching unit that calculates geometric transformation information for transformation from a coordinate in a reference image to a corresponding coordinate in an input image, and matches a local feature quantity extracted from the reference image and a local feature quantity extracted from the input image; an input image different area determination unit that transforms the different area in the reference image on a basis of the geometric transformation information about the input image determined to be in conformity by the matching, and determines a different area in the input image corresponding to the different area in the reference image; an input image different area feature quantity extraction unit that corrects a different area in the input image, and extracts a feature quantity from the corrected different area of the input image; and a feature quantity matching unit that matches a feature quantity extracted by the input image different area feature quantity extraction unit and a feature quantity extracted from the different area in the reference image, and outputs a matching result.
An object identification method according to the present invention includes: calculating geometric transformation information for transformation from a coordinate in a reference image to a corresponding coordinate in an input image, and matching a local feature quantity extracted from the reference image and a local feature quantity extracted from the input image; transforming the different area in the reference image on the basis of the geometric transformation information about the input image determined to be in conformity by the matching, and determining a different area in the input image corresponding to the different area in the reference image; correcting a different area in the input image; and extracting a feature quantity from the corrected different area of the input image, and matching a feature quantity extracted from the different area of the input image and a feature quantity extracted from the different area in the reference image, and outputting a matching result.
A program according to the present invention causes a computer to execute: local feature quantity matching processing of calculating geometric transformation information for transformation from a coordinate in a reference image to a corresponding coordinate in an input image, and matching a local feature quantity extracted from the reference image and a local feature quantity extracted from the input image; input image different area determination processing of transforming the different area in the reference image on the basis of the geometric transformation information about the input image determined to be in conformity by the matching, and determining a different area in the input image corresponding to the transformed different area in the reference image; input image different area feature quantity extraction processing of correcting a different area in the input image, and extracting a feature quantity from the corrected different area of the input image; and feature quantity matching processing of matching a feature quantity extracted by the input image different area feature quantity extraction processing and a feature quantity extracted from the different area in the reference image, and outputting a matching result.
The configuration described above is employed, so that the input image different area determination unit determines the different area in the input image by transforming the different area in the reference image on the basis of the geometric transformation information calculated by the local feature quantity matching unit, and the input image different area feature quantity extraction unit extracts the feature quantity from the different area in the input image, and the feature quantity matching unit matches the feature quantity extracted from the different area in the input image and the feature quantity extracted from the different area in the reference image, and therefore, small difference, which is not able to be identified by performing only the matching based on a conventional local feature quantity, is able to be distinguished, and only an image showing the same object is able to be identified, and therefore the object of the present invention is achieved.
Further, the present invention is also able to be realized by a computer readable nonvolatile storage medium storing the program.
Advantageous Effects of InventionAccording to the present invention, a technique capable of reducing identification error in a reference image which presents an object that is only slightly difference from an object presented in an input image.
The first exemplary embodiment according to the present invention will be described with reference to drawings.
A database may be abbreviated as “DB” in the drawings and the explanation below.
The local feature quantity extraction unit 11 extracts a local feature quantity from an input image. The details of processing performed by the local feature quantity extraction unit 11 will be described later.
The local feature quantity matching unit 12 matches a local feature quantity 1, which is a local feature quantity extracted from the input image by the local feature quantity extraction unit 11, and a local feature quantity 2, which is a local feature quantity extracted from a reference image, and outputs geometric transformation information for correcting geometric deviation between the input image and the reference image. This geometric transformation information is calculated in order to determine correctness/incorrectness of correspondence of a local feature point when the local feature quantity 1 and the local feature quantity 2 are matched. Further, the local feature quantity matching unit 12 outputs an image ID (Identification) of a reference image determined to be showing the identical object (and more specifically, as a result of matching, the local feature quantity is determined to be the same) as a local feature identification image ID. As illustrated in
The input image different area determination unit 13 uses the geometric transformation information received from the local feature quantity matching unit 12 to perform geometric transformation on the reference image corresponding to the local feature identification image ID which is output from the local feature quantity matching unit 12, or different areas of a reference image group related to the local feature identification image ID, and outputs input image different area information.
In this case, the different area is an area where there is a slight difference between an object presented in the input image and the object presented in the reference image. There are multiple different areas in an image. For example, in a case where the different area is in the rectangular shape, the different area information of the reference image (or the reference image group) may be coordinate value information about the four corners of the different area in the reference image. Alternatively, the different area information may be information representing coordinate values of a pixel group constituting the different area in the reference image. In a case where the difference between an object presented in the input image and an object presented in the reference image is a character string area in the object, the different area information may be coordinate value information about the four corners of the reference image in the rectangular shape enclosing the entire character string, or may be a coordinate value information group about the four corners of the reference image in the rectangular shape enclosing each character constituting the character string, or may be information about both of them.
The input image different area information is information obtained by applying geometric transformation information to each of the coordinate values of the four corners of the different area in the reference image. The input image different area information is indicated by, for example, the coordinate values of the four corners of the different area in the input image. Alternatively, in a case where the different area information in the reference image is coordinate value information about a pixel group constituting the different area in the reference image, the geometric transformation information may be applied to each of the pixel group, and the coordinate value information about the pixel group constituting the different area in the input image may be adopted as the input image different area information. In a case where the difference between an object presented in the input image and an object presented in the reference image is a character string area in the object, the input image different area information may be coordinate value information about the four corners of the area of the input image obtained by applying the geometric transformation information to each coordinate value about the four corners of the reference image in the rectangular shape enclosing the entire character string. Alternatively, the input image different area information may be a coordinate value information group of the four corners of the area of the input image obtained by applying the geometric transformation information to each coordinate value about the four corners of the reference image in the rectangular shape enclosing each character constituting the character string. Still alternatively, the input image different area information may be information about both of them. The different area information of the reference image is registered in advance to the database. The method for realizing this may be such that, in a case where the local feature quantity 2 is stored as the database, the different area information of the reference image may be stored together in the local feature quantity DB as illustrated in
The input image different area feature quantity extraction unit 14 corrects the different area in the input image on the basis of the input image different area information determined by the input image different area determination unit 13, and extracts the feature quantity from the corrected different area. The details of processing performed by the input image different area feature quantity extraction unit 14 will be described later.
The feature quantity matching unit 15 matches a feature quantity 1, which is a feature quantity extracted from the different area in the input image by the input image different area feature quantity extraction unit 14, and a feature quantity 2, which is a feature quantity extracted from the different area in the reference image, and determines whether the input image and the reference image present the same object or not. The feature quantity matching unit 15 outputs, as a different area identification image ID, an image ID corresponding to an image determined to be presenting the same object. As illustrated in
Subsequently, the processing performed by the local feature quantity extraction unit 11 will be described in details with reference to
The brightness information extraction unit 101 receives an input image, and extracts and outputs only information about the brightness from each pixel of the input image. The input image received here is an image captured by an image-capturing device such as a digital camera, a digital video camera, or a mobile telephone, or an image captured by means of a scanner and the like. The image may be a compressed image such as JPEG (Joint Photographic Experts Group) or may be a non-compressed image such as TIFF (Tagged Image File Format). The local feature point detection unit 102 detects many distinctive points (feature points) from the image, and outputs feature point information which is information about each of the feature points. In this case, for example, the feature point information indicates the coordinate position and the scale of the detected local feature point, the orientation of the feature point, the “feature point number” which is a unique ID allocated to the feature point, and the like. It should be noted that the local feature point detection unit 102 may output the feature point information as separate feature point information for each direction of orientation of each of the feature points. For example, the local feature point detection unit 102 may output the feature point information only in the direction of the most major orientation of each of the feature points, or may also output the feature point information in the direction of the second and subsequent major orientations.
When the local feature point detection unit 102 outputs the feature point information in the direction of the second and subsequent major orientations, the local feature point detection unit 102 is able to provide a feature point number which is different for the direction of the orientation of each of the feature points. When the local feature point detection unit 102 extracts the feature point information by detecting the feature points from the image, for example, the local feature point detection unit 102 can use DoG (Difference-of-Gaussian) processing. More specifically, the local feature point detection unit 102 is able to determine the position and the scale of the feature point by performing extreme value search in the scale space using DoG processing. Further, the local feature point detection unit 102 is able to calculate the orientation of each of the feature points by using the determined position and scale of the feature point and the gradient information of the peripheral area. When the local feature point detection unit 102 extracts the feature point information by detecting the feature points from the image, the local feature point detection unit 102 may use other methods such as Fast-Hessian Detector, and the like instead of DoG. The local feature point detection unit 102 may select only important feature points from among the feature points detected inside thereof, and may output only the information about the feature points as the feature point information.
The local feature quantity generation unit 103 receives the feature point information which is output from the local feature point detection unit 102, and generates (describes) the local feature quantity which is the feature quantity of the local area for each of the feature points. It should be noted that the local feature quantity generation unit 103 may output the local feature quantity in a format compressed with a lossless compression such as ZIP and LZH. In a case where the degree of importance of the feature point detected by the local feature point detection unit 102 is determined, the local feature quantity generation unit 103 can generate and output the local feature quantity in the order of the degree of importance of the feature point. The local feature quantity generation unit 103 may generate and output the local feature quantity in the order of the coordinate position of the feature point. First, the local feature quantity generation unit 103 obtains the local area where the feature quantity extraction is performed from the coordinate value, scale, and orientation of each of the detected feature points on the basis of the feature point information. In a case where there are plural pieces of feature point information in different orientations with respect for a single feature point, the local feature quantity generation unit 103 is able to obtain the local area for each of the pieces of the feature point information. Subsequently, the local feature quantity generation unit 103 rotates and normalizes the local area according to the orientation direction of the feature point, and thereafter, divides the local area into sub-areas. For example, the local area is able to be divided into 16 blocks (4×4 blocks). Subsequently, the local feature quantity generation unit 103 generates a feature vector for each of the sub-areas of the local area. For example, a gradient direction histogram is able to be used as the feature vector of the sub-area. More specifically, the local feature quantity generation unit 103 calculates the gradient direction for each pixel in each of the sub-areas, and quantizes the gradient direction into eight directions, and calculates the frequency of eight directions quantized for each of the sub-areas, thus generating the gradient direction histogram. At this occasion, the local feature quantity generation unit 103 outputs, as a local feature quantity, the feature vector constituted by the gradient direction histogram of 16 blocks×8 directions generated for each of the feature points. The local feature quantity generation unit 103 produces output so that the coordinate position information about the feature point is included in the output local feature quantity.
Subsequently, the processing performed by the local feature quantity matching unit 12 will be described in details with reference to
The corresponding feature point determination unit 201 collects a local feature quantity 1 extracted from the input image by the local feature quantity extraction unit 11 and a local feature quantity 2 extracted from the reference image, and outputs corresponding feature point information. For example, in a case where each of the local feature quantity 1 and the local feature quantity 2 is a set of feature quantities describing the gradient histogram around the local feature point, the corresponding feature point determination unit 201 performs distance calculation in the feature quantity space for all of the combinations of the feature quantities. In this case, only in a case where the smallest distance value is sufficiently smaller than the second smallest distance value, the corresponding feature point determination unit 201 determines, with regard to a combination of feature quantities of which distance value is the minimum, that the feature quantity and a local feature point serving as the basis of the feature quantity description are corresponding. Then, the corresponding feature point determination unit 201 outputs, as corresponding feature point information, position information about the local feature point corresponding to the position information about the local feature point.
The incorrect corresponding point removing unit 202 receives the corresponding feature point information from the corresponding feature point determination unit 201, and determines correctly corresponding feature points and incorrectly corresponding feature points from among these corresponding feature points. Then, the incorrect corresponding point removing unit 202 respectively outputs the determined feature point information, and also outputs the geometric transformation information used for the determination. For example, the incorrect corresponding point removing unit 202 applies a scheme such as RANSAC (RANdom SAmple Consensus) to the corresponding feature point information received from the corresponding feature point determination unit 201, and estimates the geometric transformation information for transformation from coordinates in the reference image into coordinates in the input image. The incorrect corresponding point removing unit 202 respectively applies the geometric transformation information estimated here to the feature point, in the reference image, associated with the corresponding feature point, and when the feature point in the reference image is determined to substantially match the feature point in the input image, the incorrect corresponding point removing unit 202 determines that the feature point is a correctly corresponding feature point. On the contrary, when the feature point in the reference image is determined not to match the feature point in the input image, the incorrect corresponding point removing unit 202 determines that the feature point is an incorrectly corresponding feature point.
The identification score calculation unit 203 receives the corresponding feature point information from the incorrect corresponding point removing unit 202, and outputs an identification score. The identification score indicates the degree of similarity of the (local) feature quantity. The output identification score may be derived as follows. For example, the identification score calculation unit 203 counts the number of combinations of correctly corresponding feature points from the corresponding feature point information received from the incorrect corresponding point removing unit 202. Then, the identification score calculation unit 203 may output the identification score by referring to a table for mapping the number of combinations of correctly corresponding feature points with a score between zero and one, which is prepared in advance. In a case where the number of combinations of correctly corresponding feature points is c, the identification score calculation unit 203 may calculate m/(c+m) as the identification score, where the minimum corresponding number of the feature points defined in advance is denoted as m.
The threshold value determination unit 204 applies threshold value processing to the identification score which is output from the identification score calculation unit 203. In a case where the identification score is equal to or more than a threshold value, the threshold value determination unit 204 determines that the image is an image presenting the same object, and outputs the ID of the reference image as the local feature identification image ID. The threshold value used by the threshold value determination unit 204 may be a value determined and held inside in advance, or may be a value given from the outside.
Subsequently, the input image different area feature quantity extraction unit 14 will be described in details.
As illustrated in
The different area information correction unit 401 receives the input image different area information from the input image different area determination unit 13, and corrects the input image different area information so as to change the range of the input image different area. For example, the different area information correction unit 401 may adopt, as a corrected different area, an area enlarged by a preset ratio on the basis of any given point in the different area of the input image. Alternatively, the different area information correction unit 401 may adopt, as a corrected different area, an area widened by a preset number of pixels on the basis of any given point in the different area of the input image. In this case, the point used as the reference may be a barycenter point of the different area. In a case where the different area of the input image is coordinate values of the four corners of the different area in the input image obtained by respectively applying the geometric transformation information to the coordinate values of the four corners of the different area in the reference image, then the point used as the reference may be an intersection point of diagonal lines of the rectangular shape defined by the four corners. For example, in a case where, e.g., edges are concentrated at the end of the input image different area received from the input image different area determination unit 13, it is able to be predicted that the same type of information is included outside of the input image different area. In this case, in the correction of the input image different area information, the different area information correction unit 401 may shift the input image different area in a direction predicted to include the same type of information, or may enlarge the input image different area.
The corrected different area image generation unit 402 receives the input image and the corrected different area information from the different area information correction unit 401. In a case where the corrected different area information is coordinate value information about the four corners of the corrected different area in the input image, for example, the corrected different area image generation unit 402 successively reads pixels on which a straight line connecting two adjacent corners of the four corners, thus determining the pixels from which the values are read from the input image and the order of reading. Accordingly, the corrected different area image generation unit 402 generates and outputs the different area image. In a case where the corrected different area information received from the different area information correction unit 401 is information indicating coordinate values of a pixel group constituting the corrected different area in the input image, the corrected different area image generation unit 402 reads the input image in that order and outputs the image as the different area image.
The different area feature quantity calculation unit 403 extracts the feature quantity from the different area image generated by the corrected different area image generation unit 402, and outputs the feature quantity. The details of the different area feature quantity calculation unit 403 will be described later. When the object illustrated in the input image is bent, the different area of the input image calculated by the input image different area determination unit 13 may include incorrect difference. Even in such case, the input image different area feature quantity extraction unit 14 includes the configuration of
As illustrated in
As illustrated in
As illustrated in
As illustrated in
As illustrated in
Where the difference between the object shown in the input image and the object shown in the reference image is assumed to be the character string area in the object, this configuration is possible in a case where the different area of the input image is coordinate value information about the four corners of the area in the input image obtained by respectively applying the geometric transformation information to each coordinate value of the four corners in the reference image in the rectangular shape enclosing the entire character string. In addition, there may be a coordinate value information group of the four corners of the area in the input image obtained by respectively applying the geometric transformation information to the coordinate values of the four corners in the reference image in the rectangular shape enclosing each character constituting the character string.
The binarization processing unit 40304 performs binarization on the different area image generated by the corrected different area image generation unit 402 on the basis of a threshold value determined in advance or on the basis of a threshold value calculated in an adaptive manner, and outputs the result of the binarization as the binarization image. An example of a method for calculating the threshold value in an adaptive manner is considered to include determination binarization for automatically calculating, as the threshold value, P where the minimum within-class dispersion of two classes divided by any given point P with regard to a distribution of pixel values of the different area image becomes the minimum and inter-class dispersion is the maximum. In a case where there is a character in the different area image generated by the corrected different area image generation unit 402, a character is often written in black or white so that the contrast from the background portion increases, and when this is taken into consideration, the binarization image which is output from the binarization processing unit 40304 outputs either an image in which the character portion is black and the background portion is white, or an image in which the character portion is white and the background portion is black. In this case, in a case where the object shown in the input image is bent, the different area of the input image calculated by the input image different area determination unit 13 may include an incorrect difference. In such case, when the determination binarization is directly performed on the area of the input image obtained by respectively applying the geometric transformation information to each coordinate value of the four corners in the reference image in the rectangular shape enclosing the entire character string, the threshold value of the binarization is not correctly set, and the character area may not be detected. However, in the configuration of
The character area detection unit 40305 receives the binarization image generated by the binarization processing unit 40304, and outputs the character area information which is information about the area where a character exists in the image. In order to detect the area where a character exists, for example, the character area detection unit 40305 uses a portion where black pixels are connecting in the binarization image as a block, and outputs, as character area information, a coordinate value information group of the four corners in the rectangular shape circumscribing each connection portion, i.e., coordinate value information about the pixel group constituting each connection portion. When the connection portion is detected, and there are many black pixels in the binarization image generated by the binarization processing unit 40304, a white character may be written on a dark background, and therefore, the white pixels and the black pixels are inverted in the binarization image, and thereafter, the above processing may be performed. In the character area information which is output here, a single connection portion is considered to correspond to substantially a single character, and therefore, the character area information represents information about the four corners in the rectangular shape circumscribing each character existing in the different area image.
The character matching feature quantity calculation unit 40306 is substantially the same as the character matching feature quantity calculation unit 40303 which is a constituent element of the different area feature quantity calculation unit 403 as illustrated in
As illustrated in
As illustrated in
The configuration of
It should be noted that the different area feature quantity calculation unit 403 may have not only the configurations illustrated in
Subsequently, the feature quantity matching unit 15 will be described in details.
The different area identification score calculation unit 501 respectively receives the feature quantity extracted from the different area of the input image as the feature quantity 1, and the feature quantity extracted from the different area of the reference image as the feature quantity 2. Then, the different area identification score calculation unit 501 outputs, as the different area identification score, the identification score determined from the two feature quantities. The different area identification score is such a degree that, the more greatly the feature quantity 1 and the feature quantity 2 are similar, the higher the value of the different area identification score is. For example, the different area identification score calculation unit 501 may calculate the distance between the feature quantity 1 and the feature quantity 2 on the feature quantity space, and may output the reciprocal number thereof as a different area identification score. Alternatively, in a case where the feature quantity 1 is matched with the feature quantity 2 extracted from each of multiple reference image groups, the different area identification score calculation unit 501 may output the different area identification score as follows. First, the different area identification score calculation unit 501 finds the minimum value of the distance on the feature quantity space from among all of the combinations of the feature quantities. Subsequently, the different area identification score calculation unit 501 may output, as a different area identification score, a reciprocal number of a value obtained by dividing the distance on the feature quantity space of all the combinations of the feature quantities by the minimum value. Alternatively, the different area identification score calculation unit 501 may output the different area identification score by referring to a table for mapping the distance value on the feature quantity space between the feature quantity 1 and the feature quantity 2 to a score between zero and one, which is prepared in advance. Still alternatively, in a case where the feature quantity extracted from the different area of the input image is the brightness value as illustrated in
The threshold value determination unit 502 performs the threshold value processing on the different area identification score which is output from the different area identification score calculation unit 501. In a case where the different area identification score is equal to or more than a threshold value, the threshold value determination unit 502 determines that the image is an image showing the same object, and outputs the ID of the reference image as the different area identification image ID. The threshold value used by the threshold value determination unit 502 may be a value determined and held inside in advance, or may be a value given from the outside. Alternatively, in a case where the different area identification score which is output from the different area identification score calculation unit 501 is the degree of similarity of the feature quantity extracted from each of the multiple rectangular shape areas in the reference image and the feature quantity extracted from each of the areas in the input image corresponding to each rectangular shape area, the threshold value determination unit 502 may perform processing as follows. More specifically, the threshold value determination unit 502 may calculate the ratio of the number of corresponding areas where the degree of similarity is equal to or more than a threshold value with respect to the number of corresponding areas, and may output, as the different area identification image ID, the ID of the reference image of which ratio is equal to or more than the threshold value. Alternatively, in a case where the difference of the highest different area identification score that is output from the different area identification score calculation unit 501 and the second highest different area identification score is equal to or more than the threshold value, the threshold value determination unit 502 may output the ID of the reference image having the highest different area identification score as the different area identification image ID. Alternatively, in a case where the ratio of the second highest different area identification score with respect to the highest different area identification score which is output from the different area identification score calculation unit 501 is equal to or less than a threshold value, the threshold value determination unit 502 may output, as the different area identification image ID, the ID of the reference image having the highest different area identification score. Alternatively, first, the threshold value determination unit 502 performs the threshold value processing on the different area identification scores which are output from the different area identification score calculation unit 501, and narrows down the different area identification scores to those equal to or more than the threshold value. Thereafter, in a case where the highest different area identification score and the second highest different area identification score are sufficiently different, the threshold value determination unit 502 may output, as the different area identification image ID, the ID of the reference image having the highest different area identification score. Alternatively, in a case where the feature quantity received from the input image different area feature quantity extraction unit 14 is a feature quantity made by combining a plurality of feature quantities, the threshold value determination unit 502 may perform processing as follows. More specifically, first, the threshold value determination unit 502 performs different area identification score calculation and threshold value determination on any given feature quantity. The threshold value determination unit 502 may perform matching in a stepwise manner, e.g., the different area identification score calculation and the threshold value determination are performed on another feature quantity in the reference image corresponding to the ID which is output as a result.
As described above, the object identification devices Z1 and Z1′ according to the first exemplary embodiment of the present invention is able to reduce the identification error of the reference image showing an object that has only a slight difference from an object shown in the input image. The reason for this is as follows. The different area in the input image is determined by transforming the different area in the reference image on the basis of the geometric transformation information between the input image and the reference image calculated when the local feature quantity is calculated. Then, the feature quantity extracted from the different area of the input image and the feature quantity extracted from the different area of the reference image are matched, so that small difference, which is not able to be identified by performing only the matching based on a conventional local feature quantity, is able to be distinguished, and only an image showing the same object is able to be identified.
Second Exemplary EmbodimentThe second exemplary embodiment according to the present invention will be described with reference to drawings.
The incorrect corresponding feature point concentration degree search unit 701 receives corresponding feature point information from the local feature quantity matching unit 16, and outputs different area information which is information about a different area in the reference image. In the corresponding feature point information received from the local feature quantity matching unit 16, correctly corresponding feature points and incorrectly corresponding feature points are determined. For this reason, the different area in the reference image is able to be estimated by searching an area where incorrectly corresponding feature points are concentrated from the image. The search of the area where the incorrectly corresponding feature points are concentrated may be done according to, for example, the following method, or may be other methods. In the method, a rectangular shape window having a preset size is set, and the rectangular shape window is moved in the different image, and when the number of incorrectly corresponding feature points is equal to or more than a preset number within the rectangular shape window, the area of the rectangular shape window is adopted as the different area.
The different area image generation unit 801 is substantially the same as the different area image generation unit 404 which is a constituent element of the input image different area feature quantity extraction unit 14 illustrated in
The object identification device Z2 according to the second exemplary embodiment of the present invention is able to estimate the different area in the reference image even when the different area in the reference image is not registered to the database in advance. Therefore, the object identification device Z2 is particularly effective when information about the different area is not able to be registered in advance, e.g., when only a product having a scratch somewhere is distinguished from many products by an inspection system using object identification. The estimation of the different area in the reference image performed by the object identification device Z2 according to the second exemplary embodiment of the present invention is possible even when the different area is the entire object or when the different area is a part of the object. Therefore, the object identification device Z2 according to the second exemplary embodiment of the present invention is effective for any of the cases of
The third exemplary embodiment of the present invention will be described with reference to drawings.
The object area estimation unit 901 receives a reference image corresponding to the local feature identification image ID which is output from the local feature quantity matching unit 16, or a reference image group related to the local feature identification image ID, and outputs object area information which is information representing an area where an object exists in the reference image. The reference image received here may be stored to a database in advance as illustrated in
Since the range where the incorrect corresponding feature point concentration degree search unit 902 searches areas in which the incorrectly corresponding feature points are concentrated in the reference image, the incorrect corresponding feature point concentration degree search unit 902 is able to perform higher speed processing compared with the incorrect corresponding feature point concentration degree search unit 701 of which search range is the entire reference image.
Like the second exemplary embodiment, the object identification device Z3 according to the third exemplary embodiment of the present invention is able to estimate the different area in the reference image even when the different area in the reference image is not registered to the database in advance. Therefore, the object identification device Z3 is particularly effective when information about the different area is not able to be registered in advance, e.g., when only a product having a scratch somewhere is distinguished from many products by an inspection system using object identification. The estimation of the different area in the reference image performed by the object identification device Z3 according to the third exemplary embodiment of the present invention is possible even when the different area is the entire object and when the different area is a part of the object, and the different area is able to be estimated with a high degree of accuracy without being affected by the incorrectly corresponding feature points that appear from areas other than the object, and therefore, the estimation of the different area in the reference image performed by the object identification device Z3 according to the third exemplary embodiment of the present invention is particularly effective in the cases of
The fourth exemplary embodiment of the present invention will be described with reference to drawings.
As described above, the object identification device Z4 according to the fourth exemplary embodiment is different in that the local feature quantity matching unit 16 and the different area estimation unit 19 of the object identification device Z3 according to the third exemplary embodiment are replaced with the local feature quantity matching unit 12 and the different area estimation unit 20. The local feature quantity matching unit 12 is the same as the local feature quantity matching unit 12 of the object identification device Z1 and Z1′ according to the first exemplary embodiment, and detailed description thereabout is omitted. The details of the different area estimation unit 20 will be described later. The other constituent elements are the same as those of the third exemplary embodiment. These constituent elements are denoted with the same reference numerals, and detailed explanation thereabout is omitted.
The transformation image generation unit 2001 receives an input image and geometric transformation information which is output from the local feature quantity matching unit 12, and outputs a transformation image. For example, the transformation image generation unit 2001 applies geometric transformation information to each of the pixels of the input image, and projecting each of the pixels of the input image onto the image having the same size as the reference image, thus generating the transformation image. At this occasion, in the image onto which the pixels are projected, the pixel values of the pixels onto which the pixels in the input image are not projected are filled with zero and the like by the transformation image generation unit 2001, so that the transformation image is generated. In a case where the geometric transformation information which is output from the local feature quantity matching unit 12 is information of transformation from coordinates in the reference image into coordinates in the input image, the geometric transformation information applied by this transformation image generation unit 2001 needs to be information of performing inverse transformation thereto. More specifically, in a case where the geometric transformation information which is output from the local feature quantity matching unit 12 is a 3×3 matrix of transformation from coordinates in the reference image into coordinates in the input image, the inverse matrix thereof is used as the geometric transformation information applied by the transformation image generation unit 2001.
The different image generation unit 2002 receives the reference image corresponding to the local feature identification image ID which is output from the local feature quantity matching unit 12 or the reference image group related to the local feature identification image ID, and the transformation image which is output from the transformation image generation unit 2001, and outputs, as a different image, an image made by calculating a difference between the reference image and the transformation image. When the different image generation unit 2002 calculates the difference between the reference image and the transformation image, for example, the different image generation unit 2002 may calculate the difference after correcting the brightness of one of the images so that the average values of the brightness of both of the images are the same. The reference image received here may be stored to a database in advance as illustrated in
The object area estimation unit 2003 receives the different image from the different image generation unit 2002, and estimates and outputs the object area information. The object area information which is output here is able to be estimated by, for example, searching an area in which the difference value is small in the different image from the image. This is because the area in which the difference value is small in the different image is considered to be an area where the same object in both of the reference image and the transformation image is likely to be shown. In the estimation of the object area, for example, a rectangular shape window having a predetermined size is set, and the rectangular shape window is moved in the different image, and when the number of pixels of which pixel values are small is equal to or more than a predetermined number within the rectangular shape window, the area of the rectangular shape window is adopted as the object area. Alternatively, the estimation of the object area may be done according to other methods.
The large difference area detection unit 2004 receives the different image which is output from the different image generation unit 2002 and the object area information which is output from the object area estimation unit 2003, and outputs the different area information. The large difference area detection unit 2004 determines that a portion where the difference value is large in the object area in the different image is likely to be a portion where there is a difference in the object illustrated in the reference image and the transformation image, and searches the portion where the difference value is large from the image, and outputs the area information as the different area information. In the search of a portion where the different value is large, i.e., the search of the different area, for example, a rectangular shape window having a preset size is set, and the rectangular shape window is moved in the object area in the different image, and when the number of pixels of which pixel values are large is equal to or more than a preset number within the rectangular shape window, the area of the rectangular shape window is adopted as the different area. Alternatively, the search of a portion where the different value is large, i.e., the search of the different area, may be done according to other methods.
As described above, in the fourth exemplary embodiment of the present invention, the different area estimation unit 20 uses the input image, the reference image, and the geometric transformation information to estimate the different area in the reference image. Like the second or third exemplary embodiment, the object identification device Z4 according to the fourth exemplary embodiment of the present invention is able to estimate the different area in the reference image even when the different area in the reference image is not registered to the database in advance. Therefore, the object identification device Z4 is effective when information about the different area is not able to be registered in advance, e.g., when only a product having a scratch somewhere is distinguished from many products by an inspection system using object identification. The estimation of the different area in the reference image performed according to the fourth exemplary embodiment of the present invention is possible even when the different area is the entire object and when the different area is a part of the object, and, like the third exemplary embodiment, after the object area is first estimated so that influence of the background is eliminated, the different area is estimated again, and this enables estimating the different area with a high degree of accuracy, and therefore, the estimation of the different area in the reference image performed according to the fourth exemplary embodiment of the present invention is particularly effective in the cases of
The fifth exemplary embodiment of the present invention will be described with reference to drawings.
As described above, the object identification device Z5 according to the fifth exemplary embodiment has a configuration in which the object identification device Z3 according to the third exemplary embodiment and the object identification device Z4 according to the fourth exemplary embodiment are mixed. When the object identification device Z5 is compared with the object identification device Z3 according to the third exemplary embodiment, the object identification device Z5 is different in that the different area estimation unit 19 is replaced with the different area estimation unit 21. The details of the different area estimation unit 21 will be described later. The other constituent elements are the same as those of the third exemplary embodiment. These constituent elements are denoted with the same reference numerals, and detailed explanation thereabout is omitted.
As illustrated in
As illustrated in
The transformation image generation unit 2001, the different image generation unit 2002, and the object area estimation unit 2003 of
The different candidate area information which is output from the large difference area detection unit 2101 may be the same as the different area information which is output from the large difference area detection unit 2004, or may be area information which is grasped as an area slightly larger than the different area information. The incorrect corresponding feature point concentration degree search unit 2102 of
The different area information which is output from the incorrect corresponding feature point concentration degree search unit 2102 is obtained by further narrowing down the different area, by the incorrect corresponding feature point concentration degree search unit 2102, from the different candidate areas estimated from the four combination including the transformation image generation unit 2001, the different image generation unit 2002, the object area estimation unit 2003, and the large difference area detection unit 2101. Therefore, highly reliable different area information is output.
As illustrated in
The transformation image generation unit 2001 of
As illustrated in
As described above, in the fifth exemplary embodiment of the present invention, the different area estimation unit 21 uses the input image, the reference image, the geometric transformation information, and the corresponding feature point information to estimate the different area in the reference image. Like the second, third, and fourth exemplary embodiments, the object identification device Z5 according to the fifth exemplary embodiment of the present invention is able to estimate the different area in the reference image even when the different area in the reference image is not registered to the database in advance. Therefore, the object identification device Z5 is effective when information about the different area is not able to be registered in advance, e.g., when only a product having a scratch somewhere is distinguished from many products by an inspection system using object identification. In addition, in the fifth exemplary embodiment of the present invention, the estimation of the different area in the reference image performed according to the fifth exemplary embodiment is possible even when the different area is the entire object and when the different area is a part of the object, and as compared with the second exemplary embodiment and the like, a more highly reliable different area is able to be obtained, and therefore, highly accurate identification is able to be realized. In the fifth exemplary embodiment of the present invention, when the different area estimation unit 21 has the configuration illustrated in
The sixth exemplary embodiment of the present invention will be described with reference to drawings.
As illustrated in
As illustrated in
As described above, in the sixth exemplary embodiment of the present invention, the different area estimation unit 22 estimates the different area in the reference image by using the reference image and the template image indicating the image pattern which is seen around the different area. Like the second to fifth exemplary embodiments, the object identification device Z6 according to the sixth exemplary embodiment of the present invention does not require the different area in the reference image to be registered to a database in advance. When the image pattern typically seen in the different area is given in advance, the object identification device Z6 is able to estimate the different area by using the image pattern as the template image. For example, when only a particular mail is to be identified from among a plurality of mail images in which envelopes are the same but only the recipient names are different, the area in which the recipient name is described is able to be defined as an image pattern in which layout of character strings such as a zip code, an address, and a recipient name is somewhat fixed. For this reason, the object identification device Z6 is effective in such case. The estimation of the different area in the reference image performed according to the sixth exemplary embodiment of the present invention is possible even when the different area is the entire object and when the different area is a part of the object, and like the third to fifth exemplary embodiments, after the object area is first estimated so that influence of the background is eliminated, the different area is able to be estimated again in the case where the configuration of the different area estimation unit 22 is
The seventh exemplary embodiment of the present invention will be described with reference to drawings.
As illustrated in
As illustrated in
As illustrated in
As described above, in the seventh exemplary embodiment of the present invention, the different area estimation unit 23 uses the reference image, the template image, and the corresponding feature point information to estimate the different area in the reference image. Like the second to sixth exemplary embodiments, the object identification device Z7 according to the seventh exemplary embodiment of the present invention does not require the different area in the reference image to be registered to a database in advance. Like the sixth exemplary embodiment, when the image pattern typically seen in the different area is given in advance, the object identification device Z7 is able to estimate the different area by using the image pattern as the template image. For example, when only a particular mail is to be identified from among a plurality of mail images in which envelopes are the same but only the recipient names are different, the area in which the recipient name is described is able to be defined as an image pattern in which layout of character strings such as a zip code, an address, and a recipient name is somewhat fixed. For this reason, the object identification device Z7 is effective in such case. The estimation of the different area in the reference image performed according to the seventh exemplary embodiment of the present invention is possible even when the different area is the entire object and when the different area is a part of the object, and like the fifth exemplary embodiment, a more highly reliable different area is able to be obtained compared with the second exemplary embodiment and the like, and therefore, highly accurate identification can be realized. It should be noted that the seventh exemplary embodiment of the present invention described hereinabove is the case of the configuration in which the object identification device Z2 according to the second exemplary embodiment and the object identification device Z6 according to the sixth exemplary embodiment are mixed. In this case, the configuration of
The eighth exemplary embodiment of the present invention will be described with reference to drawings.
As illustrated in
As illustrated in
As illustrated in
As described above, in the eighth exemplary embodiment of the present invention, the object identification device Z8 uses the input image, the reference image, the geometric transformation information, and the template image to estimate information about the different area in the reference image. Like the second to seventh exemplary embodiments, the object identification device Z8 according to the eighth exemplary embodiment of the present invention does not require the different area in the reference image to be registered to a database in advance. Like the sixth and seventh exemplary embodiments, when the image pattern typically seen in the different area is given in advance, the object identification device Z8 is able to estimate the different area by using the image pattern as the template image. For example, when only a particular mail is to be identified from among a plurality of mail images in which envelopes are the same but only the recipient names are different, the area in which the recipient name is described is able to be defined as an image pattern of which layout of character strings such as a zip code, an address, and a recipient name is somewhat fixed. For this reason, the object identification device Z8 is effective in such case. The estimation of the different area in the reference image performed according to the eighth exemplary embodiment of the present invention is possible even when the different area is the entire object and when the different area is a part of the object, and like the fifth and seventh exemplary embodiments, a more highly reliable different area is able to be obtained compared with the second exemplary embodiment and the like, and therefore, highly accurate identification is able to be realized. In the eighth exemplary embodiment of the present invention, when the different area estimation unit 24 has the configuration illustrated in
The ninth exemplary embodiment of the present invention will be described with reference to drawings.
The different area estimation unit 25 is able to be configured as a combination of a configuration in which the different area is estimated only by the incorrect corresponding feature point concentration degree search unit 701 as illustrated in
As illustrated in
As illustrated in
The transformation image generation unit 2001, the different image generation unit 2002, and the object area estimation unit 2003 of
It should be noted that the different area estimation unit 25 may be configured in a manner other than
As described above, in the ninth exemplary embodiment of the present invention, the different area estimation unit 25 uses the input image, the reference image, the geometric transformation information, the corresponding feature point information, and the template image to estimate information about the different area in the reference image information. Like the second to eighth exemplary embodiments, the object identification device Z9 according to the ninth exemplary embodiment of the present invention does not require the different area in the reference image to be registered to a database in advance. Like the sixth to eighth exemplary embodiments, when the image pattern typically seen in the different area is given in advance, the object identification device Z9 is able to estimate the different area by using the image pattern as the template image. For example, when only a particular mail is to be identified from among a plurality of mail images in which envelopes are the same but only the recipient names are different, the area in which the recipient name is described is able to be defined as an image pattern of which layout of character strings such as a zip code, an address, and a recipient name is somewhat fixed. For this reason, the object identification device Z9 is effective in such case. The estimation of the different area in the reference image performed according to the ninth exemplary embodiment of the present invention is possible even when the different area is the entire object and when the different area is a part of the object, and like the fifth, seventh, and eighth exemplary embodiments, a more highly reliable different area is able to be obtained compared with the second exemplary embodiment and the like, and therefore, highly accurate identification is able to be realized. In the ninth exemplary embodiment of the present invention, for example, when the different area estimation unit 25 has the configuration illustrated in
The tenth exemplary embodiment of the present invention will be described with reference to drawings.
As illustrated in
As illustrated in
Unlike the first to ninth exemplary embodiments, the tenth exemplary embodiment of the present invention uses the feature quantity generated by cutting out a portion of the local feature quantity in order to identify the different area. For this reason, in the tenth exemplary embodiment, when the feature quantity is generated by the input image different area feature quantity extraction unit, the input image different area feature quantity extraction unit may receive the local feature quantity extracted from the input image, and does not require the input image itself. Therefore, in a case where the object identification device Z10 is configured as a server client system in which only the extraction of the local feature quantity is performed by a client side and the other processing is performed by a server side, only the local feature quantity lighter than the input image may be transmitted to the server side. Therefore, the object identification device Z10 is able to reduce the processing time until an identification result is obtained. The feature quantity matching unit according to the tenth exemplary embodiment of the present invention performs substantially the same processing as the local feature quantity matching unit, but is able to perform matching only with the different areas to exclude influence of the correspondence of the feature point detected from an area other than the different area. As compared with a conventional method using all of the local feature quantities extracted from the entire image, the feature quantity matching unit according to the tenth exemplary embodiment of the present invention is able to distinguish the difference in the object, and is able to realize highly accurate identification as a result.
The eleventh exemplary embodiment of the present invention will be described with reference to drawings.
As illustrated in
As illustrated in
Like the first exemplary embodiment, the eleventh exemplary embodiment of the present invention has information about the different area registered to a database in advance. As the registered different area in the reference image, the different area, in the input image, determined using the geometric transformation information and the different area information, and the area obtained by further correcting the different area in the input image, the minimum necessary areas where there are differences are able to be extracted from the reference image and the input image. Therefore, even when the different area is the entire object and when the different area is only a part of the object, the object identification device Z11 according to the eleventh exemplary embodiment of the present invention is able to accurately identify an object shown in an image, and is able to suppress identification error which is the problem that occurred when only a conventional local feature quantity is used. Further, in contrast to the first exemplary embodiment in which identification is done using only the feature quantity extracted from any one of the non-corrected different area of the input image and the corrected different area thereof, the object identification device Z11 according to the eleventh exemplary embodiment of the present invention performs identification using both of the feature quantities extracted from the non-corrected different area of the input image and the feature quantity extracted from the corrected different area thereof, so that the object identification device Z11 is able to perform identification in a more accurate manner. It should be noted that the
The twelfth exemplary embodiment of the present invention will be described with reference to drawings.
Unlike the first to eleventh exemplary embodiments, the object identification device Z12 according to the twelfth exemplary embodiment of the present invention does not determine the ultimate identification result with the different area identification scores only, but determines the ultimate identification result from a score obtained by integrating them with the identification score based on the local feature quantity. For example, in a case where images showing the same object are captured in a bad environment (for example, dark environment) and another similar object is captured in an ideal environment, and in a case where the similar object is similar not only in the texture but also in the color tone, then it is not possible to perform correct identification with only the feature quantity extracted from the different area. However, when combined with the identification result based on the local feature quantity, the identification result for the same object is able to be relatively increased. It should be noted that the
The thirteenth exemplary embodiment of the present invention will be described with reference to drawings.
As described above, the object identification device Z13 according to the thirteenth exemplary embodiment is different in that the different area estimation unit 20 of the object identification device Z4 according to the fourth exemplary embodiment is replaced with the different area estimation unit 33. The other constituent elements are the same as those of the fourth exemplary embodiment. These constituent elements are denoted with the same reference numerals, and detailed explanation thereabout is omitted.
The different area estimation unit 33 compares two reference images at a time, so that the different area estimation unit 33 estimates the different area in the reference image. More specifically, the different area estimation unit 33 performs the following processing. First, the different area estimation unit 33 extracts the local feature quantities from the two reference images, respectively. Subsequently, the different area estimation unit 33 uses the geometric transformation information calculated by matching these local feature quantities to adjust the positions of the two reference images. Subsequently, the different area estimation unit 33 derives the difference of the two reference images which are position-adjusted. Then, the different area estimation unit 33 estimates the different area on the basis of the difference, so that the different area estimation unit 33 outputs the different area information corresponding to the two reference images. The different area estimation unit 33 may output the different area information for each combination of the two reference images selected from the reference images stored in the reference image DB. For example, in a case where the reference image DB stores five reference images, i.e., reference images A, B, C, D, and E, the different area estimation unit 33 may output ten types of different area information, which is as many as the number of combinations of two reference images selected from five reference images.
In the processing of estimating the different area, the different area estimation unit 33 may extract the local feature quantities from the two reference images, respectively, and match the local feature quantities, so that the different area estimation unit 33 calculates the corresponding feature points. Since the calculated corresponding feature points indicate the areas where the two reference images are in conformity (the same area), the different area estimation unit 33 may estimate an area excluding the area circumscribing the corresponding feature points so that the area is determined to be the different area in each reference image, and may output the area as the different area information.
Like the feature quantity matching unit 15 according to the first exemplary embodiment of
An example of specific processing according to the above tournament method will be described. The reference image DB is assumed to be storing five reference images, i.e., the reference images A, B, C, D, and E. First, in the different areas corresponding to the reference image A and the reference image B, the feature quantity matching unit 15 calculates the different area identification score from the feature quantity of the input image and the feature quantity of each of the reference image A and the reference image B. In this case, in a case where the reference image A has a different area identification score higher than that of the reference image B, the feature quantity matching unit 15 calculates the feature quantity of the input image and the different area identification scores from the feature quantities of the reference image A and the reference image C in the different areas corresponding to the reference image A and the reference image C which is a subsequent reference image. In a case where the reference image C has a different area identification score higher than that of the reference image A, the feature quantity matching unit 15 performs the same processing in the different areas corresponding to the reference image C and the reference image D which is a subsequent reference image. In a case where the reference image D has a different area identification score higher than that of the reference image C, the feature quantity matching unit 15 performs the same processing in the different areas corresponding to the reference image D and the reference image E which is a subsequent reference image. In a case where the reference image D has a different area identification score higher than that of the reference image E, all the reference images is processed, and therefore, the feature quantity matching unit 15 determines that the reference image D is an image showing the same object as the input image. Then, the feature quantity matching unit 15 outputs the ID of the reference image D as the different area identification image ID.
Alternatively, the feature quantity matching unit 15 may not necessarily output the reference image ultimately selected in the processing of the tournament method. For example, the feature quantity matching unit 15 may determine whether the average value or the minimum value of the different area identification scores, which are calculated with the combinations of the ultimately selected reference image and another reference image, of the reference image ultimately selected are equal to or more than a threshold value or not, so that the feature quantity matching unit 15 may ultimately determine whether to output or not.
Alternatively, processing different from the processing of the above-described tournament method may be performed. For example, in all the combinations of the reference image, the feature quantity matching unit 15 matches the feature quantity extracted from the different area of the input image and the feature quantity extracted from each of the different areas of the two reference images, and calculates the different area identification score of each of the reference image. Then, instead of selecting one of the reference images in an alternative manner, the feature quantity matching unit 15 selects all the reference images of which different area identification scores are equal to or more than the preset threshold value. Alternatively, the feature quantity matching unit 15 may select all the reference images of which the different area identification scores are equal to or more than the preset threshold value, and of which differences of the different area identification scores from the different area identification scores of other the reference images are equal to or more than a preset value. Then, the feature quantity matching unit 15 may determine that the reference image thus selected is the image showing the same object as the input image, and may output the IDs of the reference images as the different area identification image IDs.
Alternatively, the following processing may be performed. The feature quantity matching unit 15 holds the different area identification score calculated in each combination of the reference images. At the point in time when the different area identification scores is calculated for all the combinations of the reference images, the feature quantity matching unit 15 calculates the average value of the different area identification scores for each of the reference images, and adopts the calculated value as the ultimate different area identification score of each of the reference images. The feature quantity matching unit 15 may output the different area identification image ID on the basis of the ultimate different area identification scores.
In the processing of estimating the different area, the different area estimation unit 33 may estimate the area where the color is not similar in the two reference images so that the area is determined to be the different area. For example, this processing is performed as follows. First, the different area estimation unit 33 changes the sizes of the two reference images so as to make the sizes of the two reference images be the same. Subsequently, the different area estimation unit 33 calculates the degree of similarity of the colors of the pixels corresponding to the two reference images. Subsequently, the different area estimation unit 33 estimates the area where the colors are not similar so that the area is determined to be the different area, and outputs the area as the different area information. In the area detected as the area where the colors are not similar in the comparison of the reference images, the feature quantity matching unit 15 performs the matching processing between the reference image and the input image of which size is changed to the same image size as the reference image. In a case where this method is used, when the degree of similarity of the colors of the reference images or of the reference image and the input image is calculated, the similar colors not only in the corresponding pixels but also in the surrounding pixels may be searched. As a method of calculating the degree of similarity of the colors, for example, a method of calculating a sum of squares of differences of brightness values of each of R (red), G (green), and B (blue) of the pixels to be compared may be used. Alternatively, it may be possible to use a method of calculating, after normalizing each of the brightness values of R, G, and B by using the total summation of the brightness values of R, G, and B, a summation of squares of differences of the normalized values in the pixels to be compared.
In the processing of estimating the different area, the different area estimation unit 33 may estimate the area where not the colors but the edge components are not similar in the two reference images so that the area is determined to be the different area. More specifically, in the two reference images of which sizes have been changed into the same image size, the different area estimation unit 33 extracts the edge components from the reference images, and calculates the degree of similarity of the edge components extracted. Subsequently, the different area estimation unit 33 estimates the area where the edge components are not similar so that the area is determined to be the different area, and outputs the area as the different area information. In the area detected as the area where the edge components are not similar in the comparison between the reference images, the feature quantity matching unit 15 performs matching processing between the input image and the reference image.
The similar processing may be performed by using the feature quantities other than the colors and the edges.
As described above, the object identification device Z13 according to the thirteenth exemplary embodiment of the present invention does not estimate the different area by dynamically comparing the input image and the reference image as is done in the second to twelfth exemplary embodiments, and instead, the object identification device Z13 according to the thirteenth exemplary embodiment of the present invention estimates the different area by comparing the reference images in advance. By using the different area estimated in advance, the object identification device Z13 according to the thirteenth exemplary embodiment of the present invention is able to obtain an identification result in a shorter time compared with the second to twelfth exemplary embodiments.
Fourteenth Exemplary EmbodimentThe fourteenth exemplary embodiment of the present invention will be described with reference to drawings.
As described above, the object identification device Z14 according to the fourteenth exemplary embodiment is different in that the local feature quantity matching unit 12 and the different area estimation unit 33 of the object identification device Z13 according to the thirteenth exemplary embodiment are replaced with the local feature quantity matching unit 16 and the different area estimation unit 34. The other constituent elements are the same as those of the thirteenth exemplary embodiment. These constituent elements are denoted with the same reference numerals, and detailed explanation thereabout is omitted.
Like the different area estimation unit 33 according to the thirteenth exemplary embodiment, the different area estimation unit 34 compares two reference images at a time, so that the different area estimation unit 34 estimates the different area in the reference images. The different area estimation unit 34 is different from the thirteenth exemplary embodiment in that, in the processing of estimating the different area, the different area estimation unit 34 uses the corresponding feature point information which is output from the local feature quantity matching unit 16 and the corresponding feature point information of the two reference images (hereinafter referred to as “reference image corresponding feature point information”).
More specifically, the different area estimation unit 34 performs the following processing. First, the different area estimation unit 34 extracts the local feature quantities from the two reference images, respectively. Subsequently, the different area estimation unit 34 matches the local feature quantities of the two reference images, thus calculating the reference image corresponding feature point information. Subsequently, the different area estimation unit 34 excludes the corresponding feature point information that matches the calculated reference image corresponding feature point information from the corresponding feature point information which is output from the local feature quantity matching unit 16. Subsequently, the different area estimation unit 34 selects, in the two reference images, the reference image having more corresponding feature point information that remains without being excluded, and in the selected reference image, the different area estimation unit 34 outputs the area circumscribing the remaining corresponding feature points as the different area information corresponding to the two reference images.
Alternatively, instead of the information about the area circumscribing the corresponding feature points, the different area estimation unit 34 may output, as the different area information, the corresponding feature point information itself that is obtained by excluding the corresponding feature point information that matches the reference image corresponding feature point information from the corresponding feature point information. In this case, the feature quantities used by the input image different area feature quantity extraction unit 14 are the local feature quantities.
In the processing of matching, two images are compared at a time like the thirteenth exemplary embodiments.
Like the different area estimation unit 33 according to the thirteenth exemplary embodiment, the method, by the different area estimation unit 34, of estimating the different area in the reference image by comparing two reference images at a time may be the method of adopting the area where the colors and the edge components are not similar as the different area. In this method, the different area estimation unit 34 may additionally perform the processing using the corresponding feature point information of the input image and the reference image. For example, the different area estimation unit 34 may set a high degree of reliability of the calculated degree of similarity of the colors and edges in an area where there is a corresponding feature point in proximity, and may set a low degree of reliability of the calculated degree of similarity of the colors and edges in an area where there is no corresponding feature point in proximity. It should be noted that “high degree of reliability” means the calculated degree of similarity is estimated to be higher (for example, a larger coefficient is multiplied), and “low degree of reliability” means the calculated degree of similarity is estimated to be lower (for example, a smaller coefficient is multiplied).
As described above, in the estimation processing of the different area in advance, the object identification device Z14 according to the fourteenth exemplary embodiment of the present invention performs not only the comparison processing between the reference images but also the processing based on the corresponding feature point information of the input image and the reference image. Therefore, the object identification device Z14 is also able to perform more highly accurate identification than the thirteenth exemplary embodiment.
The constituent elements of the object identification devices Z1, Z1′, Z2 to Z14 described above may be realized by causing the CPU 9010 of the computer 9000 to execute a program. More specifically, these constituent elements may be realized by causing the CPU 9010 to read the program from the ROM 9030, the hard disk drive 9040, or the detachable storage medium 9060, and causing the CPU 9010 to execute the read program according to a procedure of a flowchart as illustrated in
Alternatively, these constituent elements may be realized by dedicated hardware. The object identification devices Z1, Z1′, Z2 to Z14 may be dedicated hardware having the constituent elements thereof.
The invention of the present application is hereinabove described with reference to the exemplary embodiments, but the invention of the present application is not limited to the above exemplary embodiments. The structure and the details of the invention of the present application may be changed in various manners that can be understood by a person skilled in the art within the scope of the invention of the present application.
This application claims priority based on Japanese Patent Application No. 2012-288397 filed on Dec. 28, 2012, and the entire disclosure thereof is incorporated herein by reference.
A part or all of the above exemplary embodiments may be described as in the following Supplementary Notes, but it is to be understood that the present invention is not limited to what is described below.
(Supplementary Note 1) An object identification device characterized by including:
-
- a local feature quantity matching unit that calculates geometric transformation information for transformation from coordinates in a reference image to corresponding coordinates in an input image, and matches a local feature quantity extracted from the reference image and a local feature quantity extracted from the input image;
- an input image different area determination unit that transforms the different area in the reference image on the basis of the geometric transformation information about the input image determined to be in conformity by the matching, and determines a different area in the input image corresponding to the different area in the reference image;
- an input image different area feature quantity extraction unit that corrects a different area in the input image, and extracts a feature quantity from the corrected different area of the input image; and
- a feature quantity matching unit that matches a feature quantity extracted by the input image different area feature quantity extraction unit and a feature quantity extracted from the different area in the reference image, and outputs a matching result.
(Supplementary Note 2) The object identification device according to Supplementary Note 1, wherein the input image different area feature quantity extraction unit corrects the different area of the input image into an area enlarged by a preset pixel, and extracts a feature quantity from the corrected different area of the input image.
(Supplementary Note 3) The object identification device according to Supplementary Note 1, wherein the input image different area feature quantity extraction unit corrects the different area of the input image into an area enlarged by a preset ratio on the basis of a reference point in the different area, and extracts a feature quantity from the corrected different area of the input image.
(Supplementary Note 4) The object identification device according to Supplementary Note 1, wherein in a case where there is an area where edges are concentrated at an end in the different area of the input image, the input image different area feature quantity extraction unit corrects the different area of the input image into an area enlarged in a direction where the edges are concentrated, and extracts a feature quantity from the corrected different area of the input image.
(Supplementary Note 5) The object identification device according to Supplementary Note 1, further including: a different area estimation unit which compares two reference images at a time, and calculates, as the different area in the reference image, an area where there is a difference between compared reference images.
(Supplementary Note 6) The object identification device according to Supplementary Note 5, wherein
-
- the local feature quantity matching unit outputs geometric transformation information and corresponding feature point information including information indicating a correctly corresponding feature point and an incorrectly corresponding feature point, and
- the different area estimation unit calculates the different area in the reference image on the basis of corresponding feature point information which is output from the local feature quantity matching unit.
(Supplementary Note 7) The object identification device according to any one of Supplementary Notes 1 to 6, wherein
-
- the input image different area feature quantity extraction unit outputs a first feature quantity which is a feature quantity extracted from a different area in the input image determined by the input image different area determination unit, and a second feature quantity which is a feature quantity extracted by the input image different area feature quantity extraction unit, and
- the feature quantity matching unit matches the first feature quantity and a feature quantity extracted from the different area in the reference image, and in a case where it is determined that there is no sameness, the feature quantity matching unit matches the second feature quantity and the feature quantity extracted from the different area in the reference image, and outputs a result of the matching.
(Supplementary Note 8) The object identification device according to Supplementary Note 7, wherein the feature quantity matching unit executes, in parallel, matching of the first feature quantity and the feature quantity extracted from the different area in the reference image and matching of the second feature quantity and the feature quantity extracted from the different area in the reference image, and outputs, as a matching result, a result obtained by integrating the results of the two matchings.
(Supplementary Note 9) An object identification method including:
-
- a local feature quantity matching step of calculating geometric transformation information for transformation from coordinates in a reference image to corresponding coordinates in an input image, and matching a local feature quantity extracted from the reference image and a local feature quantity extracted from the input image;
- an input image different area determination step of transforming the different area in the reference image on the basis of the geometric transformation information about the input image determined to be in conformity by the matching, and determining a different area in the input image corresponding to the different area in the reference image;
- an input image different area feature quantity extraction step of correcting a different area in the input image, and extracting a feature quantity from the corrected different area of the input image; and
- a feature quantity matching step of matching a feature quantity extracted in the input image different area feature quantity extraction step and a feature quantity extracted from the different area in the reference image, and outputting a matching result.
(Supplementary Note 10) A program for causing a computer to function as:
-
- a local feature quantity matching unit that calculates geometric transformation information for transformation from coordinates in a reference image to corresponding coordinates in an input image, and matches a local feature quantity extracted from the reference image and a local feature quantity extracted from the input image;
- an input image different area determination unit that transforms the different area in the reference image on the basis of the geometric transformation information about the input image determined to be in conformity by the matching, and determines a different area in the input image corresponding to the different area in the reference image;
- an input image different area feature quantity extraction unit that corrects a different area in the input image, and extracts a feature quantity from the corrected different area of the input image; and
- a feature quantity matching unit that matches a feature quantity extracted by the input image different area feature quantity extraction unit and a feature quantity extracted from the different area in the reference image, and outputs a matching result.
By a conventional object identification using only a local feature quantity, it is difficult to perform accurate identification of products which are of the same brand but are different only in the color of the package or some of the characters, and it is difficult to perform accurate identification of mails which are of the same envelopes but are different only in the recipient names. According to the present invention, small difference, which could not be identified by performing only the matching based on a conventional local feature quantity, is able to be distinguished, and only an image showing the same object is able to be identified. The present invention is able to be applied to a barcodeless POS (Point of Sale) system, an inspection system, a mail automatic sorting system, and the like.
REFERENCE SIGNS LIST
- Z1, Z1′, Z2, Z3, Z4, Z5, Z6, Z7, Z8, Z9, Z10, Z11, Z12, Z13, Z14 object identification device
- 11 local feature quantity extraction unit
- 12 local feature quantity matching unit
- 13 input image different area determination unit
- 14 input image different area feature quantity extraction unit
- 15 feature quantity matching unit
- 16 local feature quantity matching unit
- 17 different area estimation unit
- 18 different area feature quantity extraction unit
- 19 different area estimation unit
- 20 different area estimation unit
- 21 different area estimation unit
- 22 different area estimation unit
- 23 different area estimation unit
- 24 different area estimation unit
- 25 different area estimation unit
- 26 input image different area feature quantity extraction unit
- 27 feature quantity matching unit
- 28 input image different area feature quantity extraction unit
- 29 feature quantity matching unit
- 30 local feature quantity matching unit
- 31 feature quantity matching unit
- 32 identification score integrated determination unit
- 33 different area estimation unit
- 34 different area estimation unit
- 10 brightness information extraction unit
- 102 local feature point detection unit
- 103 local feature quantity generation unit
- 201 corresponding feature point determination unit
- 202 incorrect corresponding point removing unit
- 203 identification score calculation unit
- 204 threshold value determination unit
- 401 different area information correction unit
- 402 corrected different area image generation unit
- 403 different area feature quantity calculation unit
- 404 different area image generation unit
- 501 different area identification score calculation unit
- 502 threshold value determination unit
- 701 incorrect corresponding feature point concentration degree search unit
- 801 different area image generation unit
- 901 object area estimation unit
- 902 incorrect corresponding feature point concentration degree search unit
- 2001 transformation image generation unit
- 2002 different image generation unit
- 2003 object area estimation unit
- 2004 large difference area detection unit
- 2101 large difference area detection unit
- 2102 incorrect corresponding feature point concentration degree search unit
- 2103 incorrect corresponding feature point concentration degree search unit
- 2104 different image generation unit
- 2105 large difference area detection unit
- 2106 different candidate area overlapping detection unit
- 2201 template matching unit
- 2202 template matching unit
- 2301 template matching unit
- 2302 template matching unit
- 2501 template matching unit
- 2502 different candidate area overlapping detection unit
- 2601 different area local feature quantity extraction unit
- 2602 different area local feature quantity extraction unit
- 2701 incorrect corresponding point removing unit
- 2702 threshold value determination unit
- 2901 different area identification score calculation unit
- 2902 threshold value determination unit
- 2903 different area identification score calculation unit
- 2904 different area identification score calculation unit
- 2905 threshold value determination unit
- 3001 threshold value determination unit
- 3101 threshold value determination unit
- 3201 identification score integration unit
- 3202 threshold value determination unit
- 9000 computer
- 9010 CPU
- 9020 RAM
- 9030 ROM
- 9040 hard disk drive
- 9050 communication interface
- 9060 detachable storage medium
- 40301 color configuration ratio feature quantity calculation unit
- 40302 color arrangement feature quantity calculation unit
- 40303 character matching feature quantity calculation unit
- 40304 binarization processing unit
- 40305 character area detection unit
- 40306 character matching feature quantity calculation unit
- 40307 image value extraction unit
- 40308 shape feature quantity calculation unit
Claims
1. An object identification device comprising:
- a local feature quantity matching unit that calculates geometric transformation information for transformation from coordinates in a reference image to corresponding coordinates in an input image, and matches a local feature quantity extracted from the reference image and a local feature quantity extracted from the input image;
- an input image different area determination unit that transforms the different area in the reference image on a basis of the geometric transformation information about the input image determined to be in conformity by the matching, and determines a different area in the input image corresponding to the different area in the reference image;
- an input image different area feature quantity extraction unit that corrects a different area in the input image, and extracts a feature quantity from the corrected different area of the input image; and
- a feature quantity matching unit that matches a feature quantity extracted by the input image different area feature quantity extraction unit and a feature quantity extracted from the different area in the reference image, and outputs a matching result.
2. The object identification device according to claim 1, wherein
- the input image different area feature quantity extraction unit corrects the different area of the input image into an area enlarged by a preset pixel, and extracts a feature quantity from the corrected different area of the input image.
3. The object identification device according to claim 1, wherein
- the input image different area feature quantity extraction unit corrects the different area of the input image into an area enlarged by a preset ratio on a basis of a reference point in the different area, and extracts a feature quantity from the corrected different area of the input image.
4. The object identification device according to claim 1, wherein
- in a case where there is an area where edges are concentrated at an end in the different area of the input image, the input image different area feature quantity extraction unit corrects the different area of the input image into an area enlarged in a direction where the edges are concentrated, and extracts a feature quantity from the corrected different area of the input image.
5. The object identification device according to claim 1 further comprising:
- a different area estimation unit that compares two reference images at a time, and calculates, as the different area in the reference image, an area where there is a difference between compared reference images.
6. The object identification device according to claim 5, wherein
- the local feature quantity matching unit outputs geometric transformation information and corresponding feature point information including information indicating a correctly corresponding feature point and an incorrectly corresponding feature point, and
- the different area estimation unit calculates the different area in the reference image on a basis of corresponding feature point information which is output from the local feature quantity matching unit.
7. The object identification device according claim 1, wherein
- the input image different area feature quantity extraction unit outputs a first feature quantity which is a feature quantity extracted from a different area in the input image determined by the input image different area determination unit, and a second feature quantity which is a feature quantity extracted by the input image different area feature quantity extraction unit, and
- the feature quantity matching unit matches the first feature quantity and a feature quantity extracted from the different area in the reference image, and in a case where it is determined that there is no sameness, the feature quantity matching unit matches the second feature quantity and the feature quantity extracted from the different area in the reference image, and outputs a result of the matching.
8. The object identification device according to claim 7, wherein
- the feature quantity matching unit executes, in parallel, matching of the first feature quantity and the feature quantity extracted from the different area in the reference image and matching of the second feature quantity and the feature quantity extracted from the different area in the reference image, and outputs, as a matching result, a result obtained by integrating the results of the two matchings.
9. An object identification method comprising:
- calculating geometric transformation information for transformation from coordinates in a reference image to corresponding coordinates in an input image, and matching a local feature quantity extracted from the reference image and a local feature quantity extracted from the input image;
- transforming the different area in the reference image on a basis of the geometric transformation information about the input image determined to be in conformity by the matching, and determining a different area in the input image corresponding to the different area in the reference image;
- correcting a different area in the input image, and extracting a feature quantity from the corrected different area of the input image; and
- matching a feature quantity extracted from the different area of the input image and a feature quantity extracted from the different area in the reference image, and outputting a matching result.
10. A non-transitory computer readable storage medium storing a computer program causing a computer to perform:
- local feature quantity matching processing of calculating geometric transformation information for transformation from coordinates in a reference image to corresponding coordinates in an input image, and matching a local feature quantity extracted from the reference image and a local feature quantity extracted from the input image;
- input image different area determination processing of transforming the different area in the reference image on a basis of the geometric transformation information about the input image determined to be in conformity by the matching, and determining a different area in the input image corresponding to the transformed different area in the reference image;
- input image different area feature quantity extraction processing of correcting a different area in the input image, and extracting a feature quantity from the corrected different area of the input image; and
- feature quantity matching processing of matching a feature quantity extracted by the input image different area feature quantity extraction processing and a feature quantity extracted from the different area in the reference image, and outputting a matching result.
Type: Application
Filed: Dec 25, 2013
Publication Date: Dec 10, 2015
Patent Grant number: 9633278
Inventor: Ryota Mase (Tokyo)
Application Number: 14/655,836