IMAGE PROCESSING APPARATUS AND IMAGE PROCESSING METHOD
An image processing apparatus, includes: a transforming area setter, operable to set at least a part of an area of a target image as a transforming area; a transforming area divider, operable to arrange a plurality of dividing points in the transforming area and to divide the transforming area into a plurality of small areas by using a line connecting the dividing points; and a transforming processor, operable to move a position of at least one of the dividing points to transform at least one of the small areas, thereby transforming an image in the transforming area.
1. Technical Field
The present invention relates to an image processing technique for transforming an image.
2. Related Art
There is a related-art image processing technique for transforming an image to be intended for a digital image (for example, JP-A-2004-318204). JP-A-2004-318204 has disclosed an image processing of setting a part of areas on an image of a face (an area representing an image of a cheek) as a correcting area, dividing the correcting area into a plurality of small areas in accordance with a predetermined pattern and enlarging or reducing an image in a magnification set every small area, thereby transforming a shape of the face.
In the related-art image processing for transforming an image, an image is enlarged or reduced in a magnification set every small area. For this reason, the processing is complicated. Moreover, the related-art image processing for transforming an image is specialized for correcting a line of a cheek and copes with other various transforming manners with difficulty.
SUMMARYAn advantage of some aspects of the invention is to provide a technique capable of easily and efficiently implementing an image processing for transforming an image which corresponds to various transforming manners.
According to an aspect of the invention, there is provided an image processing apparatus, comprising:
a transforming area setter, operable to set at least a part of an area of a target image as a transforming area;
a transforming area divider, operable to arrange a plurality of dividing points in the transforming area and to divide the transforming area into a plurality of small areas by using a line connecting the dividing points; and
a transforming processor, operable to move a position of at least one of the dividing points to transform at least one of the small areas, thereby transforming an image in the transforming area.
The present disclosure relates to the subject matter contained in Japanese patent application No. 2007-082325 filed on Mar. 27, 2007, which is expressly incorporated herein by reference in its entirety.
The invention will be described with reference to the accompanying drawings, wherein like numbers reference like elements.
Next, an embodiment according to the invention will be described in the following order based on examples.
A. First Example:
A-1. Structure of Image Processing Apparatus:
A-2. Face Shape Correction Print Processing:
A-3. Variant of First Example:
B. Other Variants:
A. First Example A-1. Structure of Image Processing ApparatusThe printer engine 160 is a printing mechanism for carrying out a print based on print data. The card interface 170 serves to transfer data together with a memory card MC inserted into a card slot 172. In the example, image data to be RGB data are stored in the memory card MC and the printer 100 acquires the image data stored in the memory card MC through the card interface 170.
A face shape correcting portion 200, a displaying portion 310 and a printing portion 320 are stored in the internal memory 120. The face shape correcting portion 200 is a computer program for executing a face shape correction processing which will be described below under a predetermined operating system. The displaying portion 310 is a display driver for controlling the display portion 150 to display a processing menu or a message on the display portion 150. The printing portion 320 is a computer program for generating print data from image data and controlling the printer engine 160 to execute a print of an image based on the print data. The CPU 110 reads and executes the programs from the internal memory 120, thereby implementing a function of each of the portions.
The face shape correcting portion 200 includes, as a program module, a transforming manner setting portion 210, a face area detecting portion 220, a face area regulating portion 230, a transforming area setting portion 240, a transforming area dividing portion 250 and a transforming portion 260. The transforming manner setting portion 210 includes a designation acquiring portion 212, and the face area regulating portion 230 includes a specific area setting portion 232, an evaluating port ion 234 and a determining portion 236. A function of each of the portions will be explained in detail in the following description of a face shape correction print processing.
A dividing point arranging pattern table 410 and a dividing point movement table 420 are also stored in the internal memory 120. The contents of the dividing point arranging pattern table 410 and the dividing point movement table 420 will be explained in detail in the following description of the face shape correction print processing.
A-2. Face Shape Correction Print ProcessingThe printer 100 serves to print an image based on the image data stored in the memory card MC. When the memory card MC is inserted into the card slot 172, a user interface including a list display of an image stored in the memory card MC is displayed on the display portion 150 through the displaying portion 310.
When an image (or a plurality of images) is selected and a normal print button is selected by a user in the user interface shown in
At Step S120 (
Moreover, the user interface shown in
In the example, it is possible to designate the details of the transforming manner through the user as will be described below. In the case in which a check box having a purport that a request for designating the details is given is checked by the user in the user interface shown in
Subsequently, description will be given on the assumption that the transforming type “type A” for causing the shape of the face to be sharp is set as the image transforming type and the degree of the extent “middle” is set as the image transforming degree, and the request for designating the details is not given by the use.
At Step S130 (
In the detection of the face area FA at the Step S130, if the face area FA is not detected, a notice of the purport is given to the user through the display portion 150. In this case, a normal print having no face shape correction may be carried out or a processing of detecting the face area FA again using another face detecting method may be carried out.
In general, a well-known face detecting method such as a pattern matching method utilizing a template does not serve to detect a position and a tilt (an angle) for a whole face or a face portion (an eye or a mouse) in detail but to set, as the face area FA, an area supposed to include an image of the face from the target image TI on the whole. On the other hand, as will be described below, the printer 100 according to the example sets an area over which the image transformation processing for correcting the shape of the face is carried out based on the face area FA which is detected (a transforming area TA which will be described below). In general, the image of the face has a high degree of attention of an observer. Depending on a relationship of a position and an angle between the transforming area TA which is set and the image of the face, therefore, there is a possibility that an image obtained after correcting the shape of the face might be unnatural. In the example, therefore, it is assumed that positioning and a tilt regulation which will be described below are carried out for the face area FA detected at the Step S130 in such a manner that a more natural and preferable face shape correction can be implemented.
At Step S140 (
As shown in
At Step S142 (
As shown in
The evaluating portion 234 selects a pixel to be used in the calculation of the evaluation value (which will be hereinafter referred to as an “evaluating target pixel TP”) from pixels constituting the target image TI for each of the target pixel specifying lines PL1 to PLn.
On the other hand, depending on the method of detecting the face area FA and the method of setting the specific area SA, the target pixel specifying line PL is not parallel with the row direction (X direction) of the pixel in the target pixel TI in some cases as shown in
In the case in which a tilt of the target pixel specifying line PL exceeds 45 degrees with respect to the X direction, a relationship between the column and the row in the pixel matrix is reversed in the description so that only one of the pixels is selected as the evaluating target pixel TP from one of the rows in the pixel matrix. In some cases, moreover, one of the pixels is selected as the evaluating target pixel TP for a plurality of target pixel specifying lines PL depending on a relationship between the sizes of the target image TI and the specific area SA.
The evaluating portion 234 calculates, as the evaluation value, a mean value of the R value of the evaluating target pixel TP for each of the target pixel specifying lines PL. In the example, it is assumed that a part of the pixels having great R values in the evaluating target pixels TP which are selected are excluded from the calculating targets of the evaluation value for each of the target pixel specifying lines PL. More specifically, in the case in which k evaluating target pixels TP are selected for any of the target pixel specifying lines PL, for example, the evaluating target pixels TP are divided into two groups including a first group constituted by 0.75 k pixels having comparatively great R values and a second group constituted by 0.25 k pixels having comparatively small R values and only the pixels belonging to the second group are calculating targets of the mean value of the R value to be the evaluation value. The reason why a part of the evaluating target pixels TP is thus excluded from the calculating target of the evaluation value will be described below.
As described above, in the example, the evaluation value is calculated for each of the target pixel specifying lines PL through the evaluating portion 234. The target pixel specifying line PL is the straight line which is orthogonal to the reference line RL. Therefore, the evaluation value can be expressed to be calculated for a plurality of positions (evaluating positions) provided along the reference line RL. Moreover, the evaluation value can be represented as a value indicative of a feature of a distribution of pixel values in the orthogonal direction to the reference line RL for each of the evaluating positions.
At Step S143 (
In case of Mongoloid, it can be supposed that a portion representing an image of a skin in the divided specific area has a great R value, while a portion representing an image of an eye (in more detail, a black portion provided on a center of the eye) has a small R value. As described above, therefore, it is possible to decide, as the eye position Eh, the position placed along the reference line RL in which the evaluation value (the mean value of the R value) takes a minimal value.
As shown in
Even if the curve is placed on a lower side of the position of the image of the eye (a position which mainly corresponds to an image of a skin), moreover, there is a possibility that the curve might take a great evaluation value and a minimal value. For this reason, any of the minimal values which is greater than a predetermined threshold may be disregarded. Alternatively, a position of the target pixel specifying line PL corresponding to a minimum value in an evaluation value calculated for each of the target pixel specifying lines PL may be simply set to be the eye position Eh.
In the example, the eye (the black portion provided on the center of the eye) to be a portion which is supposed to have a comparatively great difference in a color from surroundings in the face is used as the reference object for positioning the face area FA. However, the mean value of the R value to be the evaluation value is calculated by setting, as a target, the evaluating target pixels TP on the target pixel specifying line PL. For example, therefore, there is a possibility that precision in the detection of the black portion might be reduced by the influence of an image of a white portion provided on a peripheral edge of the black portion. In the example, as described above, a part of the evaluating target pixels TP supposed to have the great difference in a color from the reference object (for example, the pixel belonging to the first group and having the comparatively great R value) is excluded from the calculating target of the evaluation value to enhance precision in the detection of the reference object.
Next, the determining portion 236 determines the height reference point Rh based on the eye position Eh which is detected.
In the example, the determining portion 236 serves to calculate an approximate tilt angle of the face image (which will be hereinafter referred to as an “approximate tilt angle RI”) based on the eye position Eh which is detected. The approximate tilt angle RI of the face image is obtained by estimating an approximate tilt of the image of the face in the target image TI with respect to the reference line RL of the face area FA.
At Step S144 (
After the positioning of the face area FA, at Step S150 (
At Step S152 (
In the example, a predetermined range for the angle formed by each of the evaluating direction lines EL and the reference line RL is set to be −±20 degrees. In the specification, a rotating angle at which the reference line RL is rotated clockwise is expressed in a positive value and a rotating angle at which the reference line RL is rotated counterclockwise is expressed in a negative value. The specific area setting portion 232 rotates the reference line RL counterclockwise and clockwise while increasing the rotating angle within such a range as not to exceed 20 degrees, for example, α degrees, 2α degrees . . . , and sets the evaluating direction lines EL.
The evaluating specific area ESA corresponding to the evaluating direction line EL representing each of the evaluating directions is obtained by rotating the initial evaluating specific area ESA(0) around the central point CP at an equal angle to a rotating angle in the operation for setting the evaluating direction line EL. The evaluating specific area ESA corresponding to the evaluating direction line EL(φ) is represented as an evaluating specific area ESA(φ).
At Step S153 (
A method of calculating the evaluation value is the same as the method of calculating the evaluation value in the positioning of the face area FA. More specifically, as shown in
A method of setting the target pixel specifying line PL and a method of selecting the evaluating target pixel TP in the evaluating specific area ESA have a difference as to whether they transversely divide the area but are the same as the method of positioning the face area FA shown in
The target pixel specifying line FL is the straight line which is orthogonal to the evaluating direction line EL. Therefore, the evaluation value can be expressed to be calculated for a plurality of positions (evaluating positions) placed along the evaluating direction line EL. Moreover, the evaluation value can be expressed as a value representing a feature of a distribution of a pixel value in an orthogonal direction to the evaluating direction line EL for each of the evaluating positions.
At Step S154 (
Description will be given to the reason why the angle corresponding to the evaluating direction in which the value of the variance of the evaluation value is the maximum is determined as the regulating angle to be used for regulating the tilt. As shown in a second stage from a top in
On the other hand, as shown in uppermost, third and fourth stages in
As described above, the value of the variance of the evaluation value along the evaluating direction line EL is increased when the evaluating direction is close to the direction of the tilt of the image of the face, and is reduced when the evaluating direction is distant from the direction of the tilt of the image of the face. By determining, as a regulating angle to be used for the tilt regulation, an angle corresponding to the evaluating direction in the case in which the value of the variance of the evaluation value is a maximum, accordingly, it is possible to implement the regulation of the tilt of the face area FA in such a manner that the tilt of the face area FA is adapted to the tilt of the image of the face.
In the example, in the case in which a maximum value is taken with a critical value in a range of an angle, that is, at −20 or 20 degrees as a result obtained by calculating the variance of the evaluation value, it can be supposed that there is a high possibility that the tilt of the face might not be evaluated accurately. For this reason, it is assumed that the tilt of the face area FA is not regulated.
In the example, moreover, the determined regulating angle is compared with the approximate tilt angle RI calculated in the positioning of the face area FA. In the case in which a difference between the regulating angle and the approximate tilt angle RI is greater than a predetermined threshold, it can be supposed that some error is made in the evaluation or determination in the positioning and tilt regulation of the face area FA. For this reason, it is assumed that the positioning and tilt regulation of the face area FA is not carried out.
At Step S155 (
At Step S160 (
When the transforming area TA is set, thus, the reference line RL to be a parallel straight line with a contour line in the vertical direction of the face area FA is also a straight line which is parallel with a contour line in the vertical direction of the transforming area TA. Moreover, the reference line RL is a straight line for dividing a width of the transforming area TA into halves.
As shown in
At Step S170 (
A manner for arranging the dividing points D (the number and positions of the dividing points D) is defined corresponding to the transforming type set at the Step S120 (
As shown in
The horizontal dividing line Lh1 is disposed below the image of the chin in the transforming area TA and the horizontal dividing line Lh2 is disposed in the vicinity of a just lower part of the image of the eye. Moreover, the vertical dividing lines Lv1 and Lv4 are disposed on an outside of the image of the line of the cheek, and the vertical dividing lines Lv2 and Lv3 are disposed on an outside of an image of a corner of the eye. The horizontal dividing line Lh and the vertical dividing line Lv are disposed in accordance with a corresponding relationship with the preset size of the transforming area TA in such a manner that the positional relationship between the images of the horizontal dividing line Lh and the vertical dividing line Lv is consequently obtained as described above.
In accordance with the arrangement of the horizontal dividing line Lh and the vertical dividing line Lv, the dividing point D is disposed on the intersection of the horizontal dividing line Lh and the vertical dividing line Lv and the intersection of the horizontal dividing line Lh and vertical dividing line Lv and the outer frame of the transforming area TA. As shown in
As shown in
The transforming area dividing portion 250 divides the transforming area TA into a plurality of small areas through straight lines connecting the arranged dividing points D (that is, the horizontal dividing line Lh and the vertical dividing line Lv). In the example, the transforming area TA is divided into 15 small rectangular areas as shown in
In the example, the arrangement of the dividing point D is determined by the number and positions of the horizontal dividing lines Lh and the vertical dividing lines Lv. Therefore, it is also apparent that the dividing point arranging pattern table 410 defines the number and positions of the horizontal dividing lines Lh and the vertical dividing lines Lv.
At Step S180 (
A manner for moving the position of each of the dividing points D to carry out the transformation processing (a moving direction and a moving distance) is preset through the dividing point movement table 420 (
In the example, as described above, the transforming “type A” for causing the face to be sharp (see
In the example, it is assumed that a position of the dividing point ID positioned on the outer frame of the transforming area TA (for example, a dividing point D10 shown in
In
In the example, the moving manner is defined in such a manner that all of combinations of the two dividing points ID having a symmetrical positional relationship with respect to the reference line RL (a combination of the dividing points D11 and D41, for example) also maintain the symmetrical positional relationship with respect to the reference line RL after the movement of the dividing point D.
The transforming portion 260 carries out a processing of transforming an image in such a manner that an image of each small area constituting the transforming area TA in a state set before the movement of the position of the dividing point D is newly defined by the movement of the position of the dividing point D. For example, in
In the example, the small rectangular area is divided into four triangular areas by using a center of gravity CG of the small area and a processing of transforming an image is carried out on a unit of the triangular area. In the example of
For example, in
{right arrow over (s′p′)}=m1·{right arrow over (s′t′)}+m2·{right arrow over (s′u′)} [Equation 1]
By using the coefficients m1 and m2 thus calculated, next, a sum of vectors st and su in the triangular area stu before the transformation is calculated by the following Equation (2) so that the position p is obtained.
{right arrow over (sp)}=m1·{right arrow over (st)}+m2·{right arrow over (su)} [Equation 2]
In the case in which the position p in the triangular area stu before the transformation is coincident with a pixel center position of the image which has not been transformed, a pixel value of the pixel is set to be a pixel value of the image obtained after the transformation. On the other hand, in the case in which the position p in the triangular area stu which has not been transformed is shifted from the pixel center position of the image which has not been transformed, a pixel value in the position p is calculated by an interpolating calculation such as a bicubic using a pixel value of a pixel around the position p, and the pixel value thus calculated is set to be a pixel value of the image obtained after the transformation.
By calculating the pixel value for each of the pixels in the image of the triangular area s′t′u′ obtained after the transformation as described above, it is possible to carry out the image transformation processing from the image of the triangular area stu to that of the triangular area s′t′u′. The transforming portion 260 defines the triangular area to carry out the transformation processing as described above for each of the small areas constituting the transforming area TA shown in
A manner for correcting a face shape according to the example will be described in more detail.
As shown in
On the other hand, referring to the orthogonal direction to the reference line RL (the H direction), the position of the dividing point D (D11, D12) disposed on the vertical dividing line Lv1 is moved in a rightward direction and the position of the dividing point D (D41, D42) disposed on the vertical dividing line Lv4 is moved in a leftward direction (see
As described above, the vertical dividing lines Lv1 and Lv4 are disposed on an outside of the image of the cheek line and the vertical dividing lines Lv2 and Lv3 are disposed on an outside of the images of the corners of the eyes. In the face shape correction according to the example, therefore, images in outside parts of the corners of both eyes in the image of the face are wholly reduced in the H direction. In particular, a reduction ratio is increased in the vicinity of the chin. As a result, the shape of the face in the image is wholly thinned in the transverse direction.
When the transforming manners in the H and V directions are integrated, the shape of the face in the target image TI is made sharp through the face shape correction according to the example. The sharpness of the face shape can also be expressed to be a so-called “small face”.
Small areas (shown in hatching) using the dividing points D22, D32, D33 and D23 as apexes shown in
At Step S190 (
At Step S200 (
In the first example, the description has been given to the face shape correction processing in the case in which the transforming “type A” (see
As described above, the manner for moving the position of the dividing point D for the transformation processing (the moving direction and the moving distance) is determined corresponding to the combination of the transforming type and the transforming degree through the dividing point movement table 420 (
As described above, moreover, the manner for arranging the dividing point D in the transforming area TA (the number and the positions of the dividing points D) is defined corresponding to the set transforming type through the dividing point arranging pattern table 410 (
When the dividing point D is moved in the manner shown in
As described above, in the example, in the case in which a request is given through the user interface shown in
In the user interface shown in
As described above, in the face shape correction print processing to be carried out by the printer 100 according to the example, a plurality of dividing points D is arranged in the transforming area TA set onto the target image TI, and the transforming area TA is divided into a plurality of small areas by using the straight lines for connecting the dividing points D (the horizontal dividing line Lh and the vertical dividing line Lv). Moreover, there is executed the processing of transforming an image in the transforming area TA by moving the position of the dividing point D and transforming the small area. In the face shape correction print processing to be carried out by the printer 100 according to the example, thus, it is possible to transform an image by simply arranging the dividing point D in the transforming area TA and moving the dividing point D thus arranged. Thus, the transformation of the image corresponding to various transforming manners can be implemented easily and efficiently.
In the face shape correction print processing to be carried out by the printer 100 according to the example, moreover, the dividing point D is arranged in accordance with the arranging pattern corresponding to any of the transforming types which is selected and set. For this reason, there is carried out the arrangement of the dividing point D, that is, the division of the transforming area TA which is suitable for respective transforming types, for example, the transforming type for causing the face to be sharp and the transforming type for enlarging the eye. Thus, it is possible to implement the image transformation of each of the transforming types more easily.
In the face shape correction print processing to be carried out by the printer 100 according to the example, moreover, the dividing point D is moved in the moving manner (the moving direction and the moving amount) corresponding to the combination of the transforming type and the transforming degree which are selected and set. If the transforming type and the transforming degree are set, therefore, the image transformation is executed depending on their combination. Thus, the image transformation can be implemented more easily.
In the face shape correction print processing to be carried out by the printer 100 according to the example, moreover, the arrangement of the dividing point D in the transforming area TA is symmetrical with respect to the reference line RL, and the moving manner of the dividing point D is determined in such a manner that all of the combinations of two dividing points D having a symmetrical positional relationship with respect to the reference line RL maintain the symmetrical positional relationship with respect to the reference line RL after the movement of the dividing point D. In the face shape correction print processing according to the example, therefore, the symmetrical image transformation with respect to the reference line RL is carried out. Consequently, it is possible to implement the image transformation of the face image which is more natural and preferable.
In the face shape correction print processing to be carried out by the printer 100 according to the example, furthermore, it is possible to prevent the transformation from being carried out for a part of the small areas constituting the transforming area TA. More specifically, as shown in
In the face shape correction print processing to be carried out by the printer 100 according to the example, moreover, in the case in which a request for designating the details of the transforming manner is given by the user, the moving amounts in the H and V directions are designated for each of the dividing points D through the user interface and the position of the dividing point D is moved in accordance with the designation. Therefore, it is possible to easily implement the image transformation in a closer manner to the request of the user.
In the face shape correction print processing to be carried out by the printer 100 according to the example, furthermore, the positioning in the vertical direction of the face area FA which is detected is executed (the Step S140 in
Moreover, the positioning of the face area FA according to the example is executed by referring to the position along the reference line RL of the image of the eye to be a reference object. In the example, an evaluation value representing a feature of the distribution of the pixel value in the orthogonal direction to the reference line RL is calculated for a plurality of evaluating positions along the reference line RL in the specific area SA set as the area including the image of the eye. Therefore, it is possible to detect a position along the reference line RL of the image of the eye based on the evaluation value thus calculated.
More specifically, it is possible to detect the position of the image of the eye by selecting the evaluating target pixel TP for each of the target pixel specifying lines PL which are orthogonal to the reference line RL and using, as the evaluation value, the mean value of the R value of the evaluating target pixel TP.
Moreover, the position of the image of the eye is detected individually for the left dividing specific area SA(l) and the right dividing specific area SA(r) which are set to include an image of one of the eyes, respectively. As compared with the case in which the position of the image of the eye is detected by setting the whole specific area SA as a target, therefore, it is possible to eliminate the influence of a positional shift along the reference line RL for left and right eyes, thereby enhancing precision in the detection.
In the calculation of the evaluation value for detecting the position of the image of the eye, furthermore, it is assumed that a part of the selected evaluating target pixels TP which has a great R value is excluded from the calculating target of the evaluation value for each of the target pixel specifying lines PL. By excluding, from the calculating target of the evaluation value, a part of the evaluating target pixels TP which is supposed to have a great difference in a color from the image of the eye to be the reference object, therefore, it is possible to enhance precision in the detection of the position of the image of the eye more greatly.
In the face shape correction print processing to be carried out by the printer 100 according to the example, moreover, the regulation of the tilt of the face area FA is executed (the Step S150 in
In addition, the regulation of the tilt of the face area FA according to the example is executed by referring to the tilt of the images of both eyes to be the reference objects. In the example, the area including the images of both eyes is set to be the evaluating specific area ESA corresponding to each of the evaluating direction lines EL obtained by rotating the reference line RL at various angles. In each of the evaluating specific areas ESA, the evaluation value representing the feature of the distribution of the pixel value in the orthogonal direction to the evaluating direction is calculated for the evaluating positions in the evaluating direction. Based on the evaluation value thus calculated, therefore, it is possible to detect the tilt of the images of both eyes.
More specifically, referring to each of the evaluating specific areas ESA, by selecting the evaluating target pixel TP for the target pixel specifying lines PL which are orthogonal to the evaluating direction line EL, calculating the mean value of the R value of the evaluating target pixel TP as the evaluation value and determining an evaluating direction in which a variance of the evaluation value is a maximum, it is possible to detect the tilt of the images of both eyes.
In the calculation of the evaluation value for detecting the tilt of the images of both eyes, moreover, it is assumed that a part of the selected evaluating target pixels TP which has a great R value is excluded from the calculating target of the evaluation value for each of the target pixel specifying lines PL. By excluding a part of the evaluating target pixels TP which is supposed to have a greater difference in a color from the images of both eyes to be the reference objects from the calculating target of the evaluation value, therefore, it is possible to enhance the precision in the detection of the tilt of the images of both eyes more greatly.
In the face shape correction print processing to be carried out by the printer 100 according to the example, furthermore, a plurality of small areas constituting the transforming area TA is divided into four triangular areas and the image transformation processing is carried out on a unit of the triangular area. At this time, the division of the small area into four triangles is carried out by using a segment connecting each of the apexes of the small area to the center of gravity CG (CG′) before and after the transformation, respectively. The position of the center of gravity of the small area can be calculated from coordinates of the four apexes. As compared with the case in which the transforming area TA is divided into the small triangular areas from the beginning, therefore, it is possible to decrease the number of the coordinates to be designated, thereby increasing a speed of the processing. In the case in which the image is transformed without the division of the small area into the triangles, moreover, there is a possibility that the small area might take a shape having an interior angle exceeding 180 degrees, resulting in a hindrance to the transformation processing depending on the moving direction and amount of each of the apexes (the dividing points D) of the small area. In the example, the transformation processing is carried out through the division of the small area into the triangles. Therefore, it is possible to prevent the drawback from being generated and to carry out the processing smoothly and stably.
B. Other VariantsThe invention is not restricted to the examples and the embodiment but can be carried out in various manners without departing from the scope thereof and the following transformation can also be performed, for example.
B1. Other Variant 1:
While the mean value of the R value for each of the target pixel specifying lines PL is used as the evaluation value in the positioning or tilt regulation of the face area FA in the example (see
For the values, moreover, it is also possible to use the number of pixels having a cumulative value or a value which is equal to or smaller than (or is equal to or greater than) a threshold in place of a mean value of the pixel to be the evaluating value calculating target. For example, it is also possible to use, as the evaluation value, the cumulative value of the R value for each of the target pixel specifying lines PL or the number of pixels having an R value which is equal to or smaller than a threshold. Although a part of the evaluating target pixels TP is not used for calculating the evaluation value for each of the target pixel specifying lines PL in the example, moreover, all of the evaluating target pixels TP may be used to calculate the evaluation value.
Although the mean value of the R value is used as the evaluation value on the premise that the Mongoloid is a target in the example, furthermore, other evaluation values (for example, a luminance, a brightness and a B value) may be used in the case in which another race (a white race or a black race) is intended.
B2. Other Variant 2:
In the example, in the positioning or tilt regulation of the face area FA, n target pixel specifying lines PL are set to the specific area SA or the evaluating specific area ESA, and the evaluation value is calculated in the position of the target pixel specifying line PL (see
B3. Other Variant 3:
While the evaluating direction is set within a range of 20 degrees clockwise and counterclockwise around the direction of the reference line RL in the regulation of the tilt of the face area FA in the example (see
Although the evaluating direction is set at the pitch of the certain angle α in example, moreover, the pitches of the evaluating directions do not need to be constant. For example, it is also possible to reduce the pitch and to thus set the evaluating direction within a close range to the direction of the reference line RL, and to increase the pitch and to thus set the evaluating direction within a distant range from the reference line RL.
Although the specific area SA corresponding to the face area FA subjected to the positioning is set to be the initial evaluating specific area ESA(0) in the regulation of the tilt of the face area FA in the example, furthermore, the initial evaluating specific area ESA(0) may be set independently of the specific area SA.
B4. Other Variant 4:
In the example, in the regulation of the tilt of the face area FA, the evaluating directions are set and the evaluating specific area ESA corresponding to the evaluating direction line EL representing each of the evaluating directions is set. The evaluating specific area ESA is obtained by rotating the initial evaluating specific area ESA(0) at an equal angle to the rotating angle from the reference line RL of the evaluating direction line EL (see
B5. Other Variant 5:
In the example, in the positioning and tilt regulation of the face area FA, the position and tilt of the image of the eye to be the reference object is detected. By using the position and the tilt which are thus detected, the positioning and tilt regulation of the face area FA is executed. However, another image, for example, an image of a nose or a mouth may be used as the reference object.
Moreover, the detection of the position and tilt of the image of the reference object according to the example is not restricted to the case in which the positioning and tilt regulation of the face area FA is intended but can be widely applied to the case in which the position and tilt of the image of the reference object in the target image TI is detected. In this case, the reference object is not restricted to the face portion but an optional object can be employed as the reference object.
B6. Other Variant 6:
Although the transforming area TA (see
Moreover, the method of dividing the transforming area TA into small areas according to the example (see
B7. Other Variant 7:
In the example, a part of the transforming area TA is protruded from the target image TI in some cases. In those cases, a part of the dividing point D cannot be disposed on the target image TI. In the case in which a part of the dividing points D cannot be disposed on the target image TI, the horizontal dividing line Lh and the vertical dividing line Lv for defining the position of the dividing point D (see
B8. Other Variant 8:
In the example, the contents of the face shape correction print processing (
Moreover, the order of the positioning of the face area FA (the Step S140 in
While the detection of the face area FA (the Step S130 in
B9. Other Variant 9:
While the description has been given to the face shape correction print processing (
B10. Other Variant 10:
In the example, a part of the structure implemented in hardware may be replaced with software. To the contrary, a part of the structure implemented in the software may be replaced with the hardware.
According to an aspect of the invention, a plurality of dividing points is arranged in the transforming area set onto the target image and the transforming area is divided into a plurality of small areas by using the straight line connecting the dividing points. Moreover, the position of the dividing point is moved and the small area is transformed so that the processing of transforming the image in the transforming area is executed. Thus, the dividing point is arranged in the transforming area and the arranged dividing point is simply moved so that the image can be transformed. Thus, the image processing of transforming an image corresponding to various transforming manners can be implemented easily and efficiently.
The image processing apparatus may further include a transforming manner setting portion for selecting one of a plurality of predetermined transforming types and setting the type as a transforming type to be applied to a transformation of an image in the transforming area. The transforming area dividing portion may arrange the dividing points in accordance with a predetermined arranging pattern corresponding to the set transforming type.
Thus, the arrangement of the dividing points, that is, the division of the transforming area which is suitable for respective transforming types such as a transforming type for causing a face to be sharp and a transforming type for enlarging eyes is carried out. Therefore, it is possible to implement a further easiness of the image processing for transforming an image corresponding to each of the transforming types.
Moreover, the transforming manner setting portion may select one of a plurality of predetermined transforming degrees and may set the transforming degree as a transforming degree to be applied to a transformation of an image in the transforming area. The transforming portion may move a position of the dividing point in accordance with a predetermined moving direction and moving amount corresponding to a combination of the transforming type and the transforming degree which are set.
If the transforming type and the transforming degree are set, thus, the image transformation corresponding to their combination is executed. Therefore, it is possible to implement a further easiness of the image processing for transforming an image.
Furthermore, the transforming manner setting portion may include a designation acquiring portion for acquiring a user designation related to a moving direction and a moving amount of the dividing point for at least one of the dividing points. The transforming portion may move a position of the dividing point in accordance with the acquired user designation.
Thus, it is possible to easily implement the image processing for transforming an image in a manner which is closer to a demand of a user.
Moreover, the transforming area setting portion may set the transforming area in such a manner that at least a part of an image of a face is included in the transforming area.
Consequently, it is possible to easily and efficiently implement the image processing for transforming an image corresponding to various transforming manners with an image of a face set to be a target.
Furthermore, the transforming area dividing portion may arrange the dividing points in such a manner that at least one pair of dividing points is mutually arranged in symmetrical positions with respect to a predetermined reference line. The transforming portion may move the at least one pair of dividing points while maintaining a positional relationship in which they are mutually symmetrical with respect to the predetermined reference line.
Thus, the symmetrical image transformation with respect to the predetermined reference line is carried out. Consequently, it is possible to implement an image processing for transforming an image of a face which is more natural and preferable.
Moreover, the transforming portion may not carry out the transformation for at least one of the small areas.
Consequently, it is possible to carry out a desirable image transformation without greatly changing an impression of a face. Thus, it is possible to implement an image processing for transforming a face image which is more natural and preferable.
Furthermore, the transforming portion may not carry out the transformation for the small areas including an image of an eye.
Thus, the small area including the image of the eye is not transformed. Consequently, it is possible to implement an image processing for transforming a face image which is more natural and preferable.
In addition, the image processing apparatus may further include a face area detecting portion for detecting a face area representing an image of a face on the target image. The transforming area setting portion may set the transforming area based on the face area thus detected.
For the image transformation in the transforming area set based on the face area detected from the target image, thus, it is possible to easily and efficiently implement the image processing for transforming an image corresponding to various transforming manners.
Moreover, the image processing apparatus may further include a printing portion for printing the target image subjected to a transformation of an image in the transforming area.
Thus, it is possible to easily and efficiently print an image obtained after the image transformation corresponding to various transforming manners.
The invention can be implemented in various manners, for example, an image processing method and apparatus, an image transforming method and apparatus, an image correcting method and apparatus, a computer program for implementing functions of the methods or apparatuses, a recording medium recording the computer program thereon, and a data signal including the computer program and materialized in a carrier.
Claims
1-12. (canceled)
13. An apparatus comprising:
- a detecting unit configured to detect a face in an image; and
- a processing unit configured to perform a transforming process to a predetermined area that includes at least a part of the face, wherein
- the processing unit performs the transforming process when the predetermined area is disposed inside of the image, and
- the processing unit does not perform the transforming process when at least a part of the predetermined area is disposed outside of the image.
14. The apparatus according to claim 13, further comprising:
- a transforming area dividing unit configured to arrange a plurality of dividing points in the predetermined area to divide the transforming area into a plurality of small areas by a line connecting the dividing points; and
- a transforming manner setting unit configured to select one of a plurality of predetermined transforming types to set the one of the predetermined transforming types as a transforming type to be used for transforming an image in the transforming area, wherein
- the transforming area dividing unit arranges the dividing points in accordance with a predetermined arranging pattern corresponding to the set transforming type, and
- the processing unit moves a position of at least one of the dividing points to perform the transforming process to the predetermined area.
15. An apparatus comprising:
- a detecting unit configured to detect a face in an image; and
- a processing unit configured to perform a transforming process to a predetermined area that includes at least a part of the face, wherein
- the processing unit performs the transforming process when the predetermined area can be set on the image, and
- the processing unit does not perform the transforming process when the predetermined area cannot be set on the image.
16. The apparatus according to claim 15, further comprising:
- a transforming area dividing unit configured to arrange a plurality of dividing points in the predetermined area to divide the transforming area into a plurality of small areas by a line connecting the dividing points; and
- a transforming manner setting unit configured to select one of a plurality of predetermined transforming types to set the one of the predetermined transforming types as a transforming type to be used for transforming an image in the transforming area, wherein
- the transforming area dividing unit arranges the dividing points in accordance with a predetermined arranging pattern corresponding to the set transforming type, and
- the processing unit moves a position of at least one of the dividing points to perform the transforming process to the predetermined area.
17. A method comprising:
- a detecting step for detecting a face in an image; and
- a processing step for performing a transforming process to a predetermined area that includes at least a part of the face, wherein
- in the processing step, the transforming process is performed when the predetermined area is disposed inside of the image, and
- in the processing step, the transforming process is not performed when at least a part of the predetermined area is disposed outside of the image.
18. The method according to claim 17, further comprising:
- a transforming area dividing step for arranging a plurality of dividing points in the predetermined area to divide the transforming area into a plurality of small areas by a line connecting the dividing points; and
- a transforming manner setting step for selecting one of a plurality of predetermined transforming types to set the one of the predetermined transforming types as a transforming type to be used for transforming an image in the transforming area, wherein
- in the transforming area dividing step, the dividing points are arranged in accordance with a predetermined arranging pattern corresponding to the set transforming type, and
- in the processing step, a position of at least one of the dividing points is moved to perform the transforming process to the predetermined area.
19. A method comprising:
- a detecting step for detecting a face in an image; and
- a processing step for performing a transforming process to a predetermined area that includes at least a part of the face, wherein
- in the processing step, the transforming process is performed when the predetermined area can be set on the image, and
- in the processing step, the transforming process is not performed when the predetermined area cannot be set on the image.
20. The method according to claim 19, further comprising:
- a transforming area dividing step for arranging a plurality of dividing points in the predetermined area to divide the transforming area into a plurality of small areas by a line connecting the dividing points; and
- a transforming manner setting step for selecting one of a plurality of predetermined transforming types to set the one of the predetermined transforming types as a transforming type to be used for transforming an image in the transforming area, wherein
- in the transforming area dividing step, the dividing points are arranged in accordance with a predetermined arranging pattern corresponding to the set transforming type, and
- in the processing step, a position of at least one of the dividing points is moved to perform the transforming process to the predetermined area.
21. A non-transitory computer-readable recording medium in which a computer program causing a computer to execute a method is stored, the method comprising:
- a detecting step for detecting a face in an image; and
- a processing step for performing a transforming process to a predetermined area that includes at least a part of the face, wherein
- in the processing step, the transforming process is performed when the predetermined area is disposed inside of the image, and
- in the processing step, the transforming process is not performed when at least a part of the predetermined area is disposed outside of the image.
22. The non-transitory computer-readable recording medium according to claim 21, the method further comprising:
- a transforming area dividing step for arranging a plurality of dividing points in the predetermined area to divide the transforming area into a plurality of small areas by a line connecting the dividing points; and
- a transforming manner setting step for selecting one of a plurality of predetermined transforming types to set the one of the predetermined transforming types as a transforming type to be used for transforming an image in the transforming area, wherein
- in the transforming area dividing step, the dividing points are arranged in accordance with a predetermined arranging pattern corresponding to the set transforming type, and
- in the processing step, a position of at least one of the dividing points is moved to perform the transforming process to the predetermined area.
23. A non-transitory computer-readable recording medium in which a computer program causing a computer to execute a method is stored, the method comprising:
- a detecting step for detecting a face in an image; and
- a processing step for performing a transforming process to a predetermined area that includes at least a part of the face, wherein
- in the processing step, the transforming process is performed when the predetermined area can be set on the image, and
- in the processing step, the transforming process is not performed when the predetermined area cannot be set on the image.
24. The non-transitory computer-readable recording medium according to claim 23, the method further comprising:
- a transforming area dividing step for arranging a plurality of dividing points in the predetermined area to divide the transforming area into a plurality of small areas by a line connecting the dividing points; and
- a transforming manner setting step for selecting one of a plurality of predetermined transforming types to set the one of the predetermined transforming types as a transforming type to be used for transforming an image in the transforming area, wherein
- in the transforming area dividing step, the dividing points are arranged in accordance with a predetermined arranging pattern corresponding to the set transforming type, and
- in the processing step, a position of at least one of the dividing points is moved to perform the transforming process to the predetermined area.
Type: Application
Filed: Jun 10, 2014
Publication Date: Oct 2, 2014
Inventors: Ikuo HAYAISHI (Matsumoto-shi), Akio YAMAZAKI (Shiojiri-shi)
Application Number: 14/301,171
International Classification: G06T 3/60 (20060101); G06K 9/00 (20060101);