The calibration problem of these cameras relies on formulating functions that describe the relationships between the camera model parameters and the lens settings. This is usually achieved by calibrating a conventional static camera model at a number of lens settings which span the lens control space using traditional calibration techniques. The calibrated model parameters at each lens setting are then stored in lookup tables [4],[10], or polynomials (or perhaps other functions) are formulated to model the parameters [7], [13],[14]. The result is a predictive camera model that can interpolate between the original sampled lens settings to produce a set of values for the terms in the static camera model for any lens setting. The approach can be summarized in:
Often polynomials are used to model the parameters variations versus lens setting [7], [13],[14], which may fail to capture complex variations in the model parameters. That is why these variations are sometimes stored in lookup tables [4],[10].
To be able to capture complex variations in the camera model parameters,
we propose a complete neural framework for zoom-lens calibration. This
task can be looked at as a combination of camera calibration and function
interpolation over a large collection of data. Due to the proven power
of neural networks as universal approximators [3],
function interpolation is best attacked via MLFNs. Thus after we develop
a neural network capable of calibrating the fixed parameters of a camera,
we can combine it with these MLFNs into an efficient all-neural framework
for the optimization of the calibration error over all data across continuous
ranges of the lens control space.