Camera resectioning is the process of estimating the parameters of a pinhole camera model approximating the camera that produced a given photograph or video; it determines which incoming light ray is associated with each pixel on the resulting image. Basically, the process determines the pose of the pinhole camera.
Usually, the camera parameters are represented in a 3 × 4 projection matrix called the camera matrix.The extrinsic parameters define the camera pose (position and orientation) while the intrinsic parameters specify the camera image format (focal length, pixel size, and image origin).
This process is often called geometric camera calibration or simply camera calibration, although that term may also refer to photometric camera calibration or be restricted for the estimation of the intrinsic parameters only. Exterior orientation and interior orientation refer to the determination of only the extrinsic and intrinsic parameters, respectively.
The classic camera calibration requires special objects in the scene, which is not required in camera auto-calibration.Camera resectioning is often used in the application of stereo vision where the camera projection matrices of two cameras are used to calculate the 3D world coordinates of a point viewed by both cameras.
The camera projection matrix is derived from the intrinsic and extrinsic parameters of the camera, and is often represented by the series of transformations; e.g., a matrix of camera intrinsic parameters, a 3 × 3 rotation matrix, and a translation vector. The camera projection matrix can be used to associate points in a camera's image space with locations in 3D world space.
In this context, we use
[u v 1]T
[xw yw z
T | |
w 1] |
M
\begin{bmatrix} wu\\ wv\\ w\end{bmatrix}=K\begin{bmatrix} R&T\end{bmatrix}\begin{bmatrix} xw\\ yw\\ zw\\ 1\end{bmatrix} =M\begin{bmatrix} xw\\ yw\\ zw\\ 1\end{bmatrix}
where
M=K\begin{bmatrix}R&T\end{bmatrix}
u,v
K
RT
xw,yw,zw
w
K=\begin{bmatrix} \alphax&\gamma&u0\\ 0&\alphay&v0\\ 0&0&1\end{bmatrix}
The
K
\alphax=f ⋅ mx
\alphay=f ⋅ my
mx
my
f
\gamma
u0
v0
Nonlinear intrinsic parameters such as lens distortion are also important although they cannot be included in the linear camera model described by the intrinsic parameter matrix. Many modern camera calibration algorithms estimate these intrinsic parameters as well in the form of non-linear optimisation techniques. This is done in the form of optimising the camera and distortion parameters in the form of what is generally known as bundle adjustment.
{}\begin{bmatrix}R3&T3\\ 01&1\end{bmatrix}4
R,T
T
T
C
C=-R-1T=-RTT
R
Camera calibration is often used as an early stage in computer vision.
When a camera is used, light from the environment is focused on an image plane and captured. This process reduces the dimensions of the data taken in by the camera from three to two (light from a 3D scene is stored on a 2D image). Each pixel on the image plane therefore corresponds to a shaft of light from the original scene.
There are many different approaches to calculate the intrinsic and extrinsic parameters for a specific camera setup. The most common ones are:
Zhang's method[2] [3] is a camera calibration method that uses traditional calibration techniques (known calibration points) and self-calibration techniques (correspondence between the calibration points when they are in different positions). To perform a full calibration by the Zhang method, at least three different images of the calibration target/gauge are required, either by moving the gauge or the camera itself. If some of the intrinsic parameters are given as data (orthogonality of the image or optical center coordinates), the number of images required can be reduced to two. In a first step, an approximation of the estimated projection matrix
H
n
K
n
R
T
bf{H}
x\pi
\pi
x
The circular points
I,J=\begin{bmatrix}1&\pmj&0\end{bmatrix}T
\pi
\Omegainfty
\Omegainfty
\omega
T | |
x | |
1 |
\omegax1=0
T | |
x | |
2 |
\omegax2=0
\begin{align} x1&=bf{H}I=\begin{bmatrix} h1&h2&h3 \end{bmatrix} \begin{bmatrix} 1\\ j\\ 0 \end{bmatrix} =h1+jh2 \\ x2&=bf{H}J= \begin{bmatrix} h1&h2&h3 \end{bmatrix} \begin{bmatrix} 1\\ -j\\ 0 \end{bmatrix} =h1-jh2 \end{align}
We can actually ignore
x2
x1
T | |
\begin{align} x | |
1 |
\omegax1&=\left(h1+jh2\right)T\omega\left(h1+jh2\right)\\ &=\left(
T | |
h | |
1 |
+j
T | |
h | |
2 |
\right)\omega\left(h1+jh2\right)\\ &=
T | |
h | |
1 |
\omegah1+j\left(
T | |
h | |
2 |
\omegah2\right)\\ &=0 \end{align}
Tsai's algorithm, a significant method in camera calibration, involves several detailed steps for accurately determining a camera's orientation and position in 3D space. The procedure, while technical, can be generally broken down into three main stages:
The process begins with the initial calibration stage, where a series of images are captured by the camera. These images, often featuring a known calibration pattern like a checkerboard, are used to estimate intrinsic camera parameters such as focal length and optical center.[6]
Following initial calibration, the algorithm undertakes pose estimation. This involves calculating the camera's position and orientation relative to a known object in the scene. The process typically requires identifying specific points in the calibration pattern and solving for the camera's rotation and translation vectors.
The final phase is the refinement of parameters. In this stage, the algorithm refines the lens distortion coefficients, addressing radial and tangential distortions. Further optimization of internal and external camera parameters is performed to enhance the calibration accuracy.
This structured approach has positioned Tsai's Algorithm as a pivotal technique in both academic research and practical applications within robotics and industrial metrology.
Selby's camera calibration method[7] addresses the auto-calibration of X-ray camera systems.X-ray camera systems, consisting of the X-ray generating tube and a solid state detector can be modelled as pinhole camera systems, comprising 9 intrinsic and extrinsic camera parameters.Intensity based registration based on an arbitrary X-ray image and a reference model (as a tomographic dataset) can then be used to determine the relative camera parameters without the need of a special calibration body or any ground-truth data.