Method for calibrating a vehicular camera system

A method of calibrating a vehicular multi-camera system includes equipping a vehicle with a plurality of cameras wherein each camera of the plurality of cameras captures image data, equipping the vehicle with an image processor, inputting image data from each of the plurality of cameras to the image processor, the image processor processing input image data in order to calibrate the vehicular multi-camera system, and wherein calibration of the vehicular multi-camera system is achieved independently of a model of the real world.

Skip to: Description  ·  Claims  · Patent History  ·  Patent History
Description

This application is a continuation of U.S. patent application Ser. No. 12/604,432, filed Oct. 23, 2009, now U.S. Pat. No. 8,169,480, which claims the benefits of German Application No. 102008053047.6, filed Oct. 24, 2008.

FIELD OF THE INVENTION

The present invention relates to a method for automatically calibrating a virtual camera and to a virtual camera apparatus which is set up for carrying out the method according to the invention. In particular, the present invention relates to a virtual camera for producing a view of the surroundings of a motor vehicle from a bird's eye perspective on a motor vehicle.

BACKGROUND OF THE INVENTION

What is known as a virtual camera refers to sets comprising a real recording camera and a habitually electronic image data processing device, which produce an output signal with a coded image or a coded image sequence, wherein the perspective of the coded images does not match the perspective of the recording camera. On account of the loss of information when a real three-dimensional object is mapped into a two-dimensional image data model by a real recording camera, the virtual camera is able to reproduce non-moving objects correctly particularly when they are approximately flat.

Virtual cameras have been proposed as driver assistance devices in motor vehicles. These are particularly what are known as top view systems or omnidirectional cameras. These typically comprise a plurality of real recording cameras which are arranged in or on a vehicle and which are used to produce a chronological sequence of image data records. The image data in the image data records are subjected to different transformations in a typically electronic image data processing device and are mixed to form a chronological sequence of overall image data. This makes it possible to obtain, by way of example, a view of the surroundings of the vehicle from a perspective above the vehicle roof. This chronological sequence of overall image data can be continuously displayed to the driver of the motor vehicle on a display apparatus in order to simplify shunt or parking manoeuvres.

It is evident that in a virtual camera with a plurality of real recording cameras the quality of the overall image delivered is distinctly dependent on the exact knowledge of the positions and directions of the real recording cameras. The more accurately that these data are known, the easier that it is possible to determine the transformation that determines the best possible image quality in the area of adjacent recording areas. Against this background, there have been a series of proposals involving either the automatic determination of the positions and directions of the recording cameras or the correction of errors in these variables in relation to initially stored values.

An omnidirectional camera with automatic calibration is revealed by the officially published document DE 10 2007 043 905 A1. Said document proposes identifying an object element in the image data for the purpose of the calibration. The recognition of mapped areas of the outer vehicle skin is recognized as particularly advantageous. The proposed approach is intended to allow compensation for, inter alia, situational changes in the real recording cameras on the vehicle on account of vibrations, ageing and thermal or mechanical stresses.

However, the known method is based on the processing of information from a model of the real world and particularly on the processing of information regarding the shape of the outer vehicle skin.

SUMMARY OF THE INVENTION

Against this background, the present invention is based on the object of providing a method for the automatic calibrating of a virtual camera which manages independently of a model of the real world.

This object is achieved by the present invention by means of a method having the features specified in claim 1.

Advantageous refinements and developments of the method according to the invention are specified in the subclaims.

BRIEF DESCRIPTION OF THE DRAWINGS

A preferred manner of carrying out the method according to the invention and a virtual camera apparatus which is set up to do so are described below, reference being made to the appended drawings, in which:

FIG. 1 shows a schematic illustration of a typical arrangement of real recording cameras on a motor vehicle and of the situation of the virtual perspective; and

FIG. 2 shows a schematic illustration of the functional units and data streams in a preferred apparatus for carrying out a method according to the invention; and

FIG. 3 shows a schematic synopsis of an algorithm for implementing a method according to the invention.

DETAILED DESCRIPTION OF THE INVENTION

As FIG. 1 shows, an expedient apparatus for carrying out a preferred method according to the invention for a driver assistance system on a motor vehicle first of all comprises a plurality of real recording camera devices 1, 2, 3. A fourth recording camera device 4 is located in the area which cannot be seen on the opposite side of the vehicle from the recording camera device 2. In one expedient implementation, the real recording camera devices 1, 2, 3, 4 are in the form of a CCD or CMOS array with a wide-angle or fisheye lens in a manner which is known per se and are arranged at different points on the motor vehicle 4 behind the glazing and in suitable pockets on the exterior mirrors. In the fitting positions which are possible in practice, the real perspectives P1 . . . 4 of the real recording camera devices 1, 2, 3, 4 are naturally offset and/or tilted with respect to the virtual perspective Pv, that is to say the viewing point of the virtual viewer. The greater that the difference between the real perspective P1 . . . 4 of a real recording camera device 1, 2, 3, 4 and the virtual perspective Pv turns out to be, the worse the quality of the image data with the transformed perspective under realistic conditions. This will be discussed later. Conversely, some fields of vision would not be visible to a viewer with a perspective corresponding to the virtual perspective Pv, because they are situated behind portions of the outer vehicle skin. In this respect, the selection of the camera positions is a trade-off between these advantages and disadvantages.

As further components, the apparatus described which is shown in FIG. 2 comprises an image data processing device 5 and a display device 6. In this case, the display device 6 is typically arranged in the area of the instrument panel or the central console. In a typical implementation, the image data processing device 5 is a digital signal processor or a powerful microprocessor for general applications with adequate equipment in terms of main memory and non-volatile program memory.

In line with the schematic illustration of a preferred algorithmic implementation of the method according to the invention which is shown in FIG. 3, the real recording camera devices 1, 2, 3, 4 respectively deliver a chronological sequence of N×4 raw subimage data items DR1 . . . N, 1 . . . 4 to the image data processing device 5 in the operating period under consideration. Said image data processing device combines the data streams to form a chronological sequence of subimage data records DR1 . . . N. Each subimage data record DRi in the sequence contains the subimage data DRi, 1 . . . 4 from the real recording camera devices 1, 2, 3, 4. In the image data processing device 5, a respective transformation T1 . . . 4 is applied to the subimage data DRi, 1 . . . 4 which a subimage data record DRi contains when it is present. These transformations T1 . . . 4 are respectively determined such that the relevant image 7, 8, 9, 10 of a predetermined planar mapping area Aj of the real recording camera device j with the perspective Pj is transformed into the image from the virtual camera with the perspective Pv. For the case of an ideal recording camera device, this transformation would be linear and could be compiled from a planar perspective extension of the image by means of rotation and displacement. In comparison with the aforementioned ideal recording camera device, however, the real recording camera devices 1, 2, 3, 4 deliver maps with nonlinear distortions. The primary cause of these nonlinear distortions is inadequacies in the real mapping lenses. This relates particularly distinctly to the proposed lenses with a strong wide-angle or fisheye characteristic. From this point of view, positions for the real recording camera devices 1, 2, 3, 4 which are further away from the road are naturally preferable. However, the nonlinear distortions can usually be attributed to a satisfactory measure for observation of the environment using what are known as inverse nonlinear transformations, which are known per se. Consequently, a person skilled in the art will, according to the respective situation, select a chain comprising a nonlinear transformation for the purpose of equalization and an ideal perspective extension for the described transformation T1 . . . 4. The application of the transformations T1 . . . 4 to the subimage data DRi, 1 . . . 4 results in transformed subimage data DTi, 1 . . . 4. In the present case, it is assumed that the transformations T1 . . . 4 transform the subimage data DTi, 1 . . . 4 directly into the coordinate system of the overall image 11. Accordingly, the overall image data DGi can be produced by simply combining the data from the transformed subimage data DTi, 1 . . . 4.

Since, in the ideal case described, the positions and recording directions of the recording camera devices 1, 2, 3, 4 did not change, the information about the orientation and situation of the subimages in the overall image would need to be set only once for the relevant vehicle geometry as a parameter for the transformations T1 . . . 4 prior to startup. In the course of such initial calibration, the parameters can be determined either by calculating or calibrating the apparatus on a test bench. Even with optimum initial calibration, ambiguities can arise in the area of the overlaps 12 in the subimages 7, 8, 9, 10 on account of the described nonlinear distortions in the real recording camera devices 1, 2, 3, 4. Through suitable selection of the transformations and/or expedient stipulation of the subimage boundaries, however, it is possible to attain entirely satisfactory results in practice. In this regard, an appropriately adjusted image data processing device 5 produces as an output signal a sequence of overall image data DG1 . . . N which are displayed in their chronological order on a display device 6 and give the driver an impression of the immediate surroundings of the motor vehicle.

Effects of ageing, overloads, accidents and the like may result in the position and/or orientation of the recording camera devices 1, 2, 3, 4 being changed. If such a change is followed by the subimages continuing to be assembled to form an overall image in the originally stipulated manner, the result is poorer quality for the overall image. To counteract this drawback, the image data processing device 5 recurrently performs calibration with the aim of optimization using a prescribed quality criterion Q for the overall image. In this case, the quality criterion Q is a scalar value which is dependent on the data from the subimages and on the stored information relating to the positions and directions of the real recording camera device 1, 2, 3, 4. Expediently, the quality criterion Q is stipulated such that it reflects the quality of the overall image, as subjectively perceived by an average viewer. In one advantageous refinement, the quality criterion Q will also relate to the correlations of the subimages 7, 8, 9, 10 in the areas of overlap 12. The quality criterion Q is optimized for a firmly prescribed subimage data record DRi by varying the parameters of the transformations. The parameters varied to the optimum replace the values originally stored in the apparatus for the purpose of further operation of the apparatus.

This calibration is recurrently performed over time whenever a selection criterion C flags a subimage data record DRk for this purpose. In the present case, the selection criterion C is defined such that the flagged subimage data record DRk means that the calibration provides the best possible result. Intuitively, a good result will be assumed if the application of the quality criterion Q to the overall images 11 which follow the calibration provides the best possible result overall. Since the quality criterion Q in the present case relates only to an individual image data record, the quality of a sequence naturally requires appropriate definition. To this end, it is possible to use the generally known statistical functions, such as the mean value. Expediently, the selection criterion C processes the subimage data DRk, 1 . . . 4 anyway in order to assess the suitability of the flagged subimage data record DRk for the calibration. This complies with the insight that not all subimage data records DRi are equally good for the calibration in practice. By way of example, calibration must obviously remain undone in the event of a lack of contrast, under exposure, defocusing or motion blurring. Equally disadvantageous are image data records with periodic structures, which can be recognized by means of frequency analysis of the subimage data DRk, 1 . . . 4, for example. In addition, the subimage data DRk, 1 . . . 4 can be examined to determine whether they contain maps of three-dimensional objects above the road level. Typical objects of this kind are high kerbstones, crash barriers, marker posts, for example. A subimage data record DRi with maps of such objects in the area of the overlaps 12 should not be used for the calibration.

In addition, the selection criterion C flags a subimage data record DRk for calibration only if there was a particular state of the motor vehicle at the time at which said subimage data record was recorded, and the driving situation was within prescribed limits at this moment. To this end, the image data processing device 5 also derives, collects and assesses vehicle state variables and driving state variables from detection devices on the motor vehicle. In this context, preferred vehicle state variables are the operating period and mileage of the vehicle, the number of vehicle starts and the operating period and also the mileage since the vehicle was last started. By including these variables in the selection criterion C, it is particularly possible to take account of thermal changes, mechanical settling or ageing effects and wear. Preferred driving state variables selected are the speed of travel, the acceleration, the steering angle, the angle of inclination and the loading of the vehicle. If available, it is also possible to include data about the tyre pressure and the setting of the suspension. When these data are included, it is possible to take account of the dynamic differences in the vehicle situation in relation to the road surface when deciding about calibration.

Further preferred variables for inclusion in the selection criterion C could be the GPS position of the vehicle, the exterior light conditions and signals from proximity sensors for the near field of the vehicle. By including such variables, it is possible to base the decision about calibration on considerations concerning whether and to what extent the current vehicle surroundings favour or impede calibration.

On the basis of a subimage data record DRk flagged by the selection criterion C, the calibration can preferably be performed by calculating a correlation between the subimage data DTi, 1 . . . 4 in a manner which is known per se. In this case, the areas of overlap 12 are identified and the situation and orientation of the image sections coded in the subimage data DTi, 1 . . . 4 relative to one another are determined. The quality criterion Q is optimized for the flagged subimage data record DRk by varying the parameters of the transformations T1 . . . 4.

In one refinement of the method described above, it is also possible to include the history of the calibrations performed in the past in the selection criterion. For example, this history could be used to determine the time for the next calibration. It is also possible to determine the parameters of the transformations not exclusively on the basis of the result of the last calibration, but rather to perform historical averaging. Yet another option is to anticipate adaptation of the parameters without calibration by extrapolation on the basis of the already collected historical data from past calibrations.

Claims

1. A method of calibrating a vehicular multi-camera system, said method comprising:

equipping a vehicle with a plurality of cameras wherein each camera of said plurality of cameras captures image data;
wherein said plurality of cameras comprises a first camera mounted at a forward portion of the vehicle and having a field of view that includes a ground surface forward of the vehicle, a second camera mounted at a rearward portion of the vehicle and having a field of view that includes a ground surface rearward of the vehicle, a third camera mounted at a sideward portion of the vehicle and having a field of view that includes a ground surface sideward of the vehicle at one side of the vehicle, and a fourth camera mounted at another sideward portion of the vehicle and having a field of view that includes a ground surface sideward of the vehicle at the other side of the vehicle;
equipping said vehicle with an image processor;
inputting image data from each of said plurality of cameras to said image processor;
said image processor processing input image data in order to calibrate said vehicular multi-camera system;
wherein calibration of said vehicular multi-camera system is achieved independently of a model of the real world;
producing a chronological sequence of image data by recurrently applying a selection criterion to select from a sequence of subimage data records recorded by said first, second, third and fourth cameras;
determining parameters of transformations of said subimage data records via an optimization method that uses a quality criterion for an overall image; and
assembling a sequence of transformed subimage data records to form a sequence of overall image data.

2. The method of claim 1, wherein equipping a vehicle with a plurality of cameras comprises equipping a vehicle with a plurality of CMOS cameras.

3. The method of claim 1, wherein calibration of said vehicular multi-camera system is achieved independently of the shape of the outer vehicle skin.

4. The method of claim 1, comprising an initial calibration before producing a chronological sequence of image data.

5. The method of claim 1, wherein said quality criterion comprises a scalar value.

6. The method of claim 5, wherein said scalar value is dependent on said subimage data.

7. The method of claim 6, wherein said scalar value is dependent on stored information relating to at least one of (i) a position of at least one of said first, second, third and fourth cameras and (ii) a direction of said field of view of at least one of said first, second, third and fourth cameras.

8. The method of claim 6, wherein said scalar value is dependent on stored information relating to (i) a position of at least one of said first, second, third and fourth cameras and (ii) a direction of said field of view of at least one of said first, second, third and fourth cameras.

9. The method of claim 6, wherein said scalar value is dependent on stored information relating to at least one of (i) respective positions of said first, second, third and fourth cameras and (ii) respective directions of said fields of view of said first, second, third and fourth cameras.

10. The method of claim 1, wherein said quality criterion reflects a quality of said overall image data.

11. The method of claim 1, wherein said calibration is recurrently performed over time in response to said selection criterion selecting a particular subimage data record.

12. The method of claim 11, wherein said selection criterion selects a particular subimage data record when a particular state of the vehicle at the time at which said particular subimage data record was recorded is within a prescribed limit.

13. A method of calibrating a vehicular multi-camera system, said method comprising:

equipping a vehicle with a plurality of CMOS cameras wherein each camera of said plurality of CMOS cameras captures image data;
wherein said plurality of CMOS cameras comprises a first camera mounted at a forward portion of the vehicle and having a field of view that includes a ground surface forward of the vehicle, a second camera mounted at a rearward portion of the vehicle and having a field of view that includes a ground surface rearward of the vehicle, a third camera mounted at a sideward portion of the vehicle and having a field of view that includes a ground surface sideward of the vehicle at one side of the vehicle, and a fourth camera mounted at another sideward portion of the vehicle and having a field of view that includes a ground surface sideward of the vehicle at the other side of the vehicle;
equipping said vehicle with an image processor;
inputting image data from each of said plurality of cameras to said image processor;
said image processor processing input image data in order to calibrate said vehicular multi-camera system;
wherein calibration of said vehicular multi-camera system is achieved independently of a model of the real world, and wherein calibration of said vehicular multi-camera system is achieved independently of the shape of the outer vehicle skin;
producing a chronological sequence of image data by recurrently applying a selection criterion to select from a sequence of subimage data records recorded by said first, second, third and fourth cameras;
determining parameters of transformations of said subimage data records via an optimization method that uses a quality criterion for an overall image; and
assembling a sequence of transformed subimage data records to form a sequence of overall image data.

14. The method of claim 13, comprising an initial calibration before producing a chronological sequence of image data.

15. The method of claim 13, wherein said quality criterion comprises a scalar value, and wherein said scalar value is dependent on stored information relating to at least one of (i) a position of at least one of said first, second, third and fourth cameras and (ii) a direction of said field of view of at least one of said first, second, third and fourth cameras.

16. The method of claim 13, wherein said calibration is recurrently performed over time in response to said selection criterion selecting a particular subimage data record, and wherein said selection criterion selects a particular subimage data record when a particular state of the vehicle at the time at which said particular subimage data record was recorded is within a prescribed limit.

17. A method of calibrating a vehicular multi-camera system, said method comprising:

equipping a vehicle with a plurality of CMOS cameras wherein each camera of said plurality of CMOS cameras captures image data;
wherein said plurality of CMOS cameras comprises a first camera mounted at a forward portion of the vehicle and having a field of view that includes a ground surface forward of the vehicle, a second camera mounted at a rearward portion of the vehicle and having a field of view that includes a ground surface rearward of the vehicle, a third camera mounted at a sideward portion of the vehicle and having a field of view that includes a ground surface sideward of the vehicle at one side of the vehicle, and a fourth camera mounted at another sideward portion of the vehicle and having a field of view that includes a ground surface sideward of the vehicle at the other side of the vehicle;
equipping said vehicle with an image processor;
inputting image data from each of said plurality of cameras to said image processor;
said image processor processing input image data in order to calibrate said vehicular multi-camera system;
wherein calibration of said vehicular multi-camera system is achieved independently of a model of the real world;
producing a chronological sequence of image data by recurrently applying a selection criterion to select from a sequence of subimage data records recorded by said first, second, third and fourth cameras;
determining parameters of transformations of said subimage data records via an optimization method that uses a quality criterion for an overall image;
assembling a sequence of transformed subimage data records to form a sequence of overall image data; and
initially calibrating said vehicular multi-camera system before producing a chronological sequence of image data.

18. The method of claim 17, comprising an initial calibration before producing a chronological sequence of image data.

19. The method of claim 17, wherein said quality criterion comprises a scalar value, and wherein said scalar value is dependent on stored information relating to at least one of (i) a position of at least one of said first, second, third and fourth cameras and (ii) a direction of said field of view of at least one of said first, second, third and fourth cameras.

20. The method of claim 17, wherein said calibration is recurrently performed over time in response to said selection criterion selecting a particular subimage data record, and wherein said selection criterion selects a particular subimage data record when a particular state of the vehicle at the time at which said particular subimage data record was recorded is within a prescribed limit.

Patent History
Patent number: 8421865
Type: Grant
Filed: May 1, 2012
Date of Patent: Apr 16, 2013
Patent Publication Number: 20120224064
Assignee: Magna Electronics Europe GmbH & Co. KG (Holland, MI)
Inventors: Christian Euler (Haibach), Markus Adameck (Roedermark)
Primary Examiner: Andy Rao
Application Number: 13/460,896
Classifications
Current U.S. Class: Vehicular (348/148); Navigation (348/113); Direction Finding Or Location Determination (348/116)
International Classification: H04N 7/18 (20060101);