Distortion
Characterization
Lens distortions in an imaging system are a result of a camera's lens assembly not perfectly representing the pinhole camera model that we use to relate our world (object space) to our image coordinate system. Lens distortions can be broken into roughly three effects:
- Symmetric radial distortion models a kind of distortion (such as barrel or pincushion distortion) that applies across the image extent radially and is irrespective of the distance of any objects in the scene.
- Tangential or decentering distortion models non-orthogonalities (decentering) in the lens-assembly of a camera, and quantifies how that deviates from a pinhole model. As decentering errors in the lens-assembly gets smaller, these errors can often correlate highly with small errors in \(c_x\) and \(c_y\). As a result, these effects sometimes go uncharacterized as their observability can be low.
- Asymmetric radial distortion comprises effects in the radial direction of our observations that are dependent on the relative distances between objects in the scene. In many cases, asymmetric radial distortion effects are modeling refraction of light rays in our object space or "world," and not necessarily lens effects. As a result, these effects are difficult to observe in close-range applications, and we often choose not to model them so as not to over-fit our model.
Most models operate as a function of the distance to the principal point of the lens, modeled by r, as well as angles between the principal point and image space.
Further Reading
We've written in-depth about camera distortions on the Tangram Vision Blog. See Part 1 and Part 2 of our distortion series for more information.
Supported Distortion Models
Brown-Conrady
Brown-Conrady is the typical distortion model used by default in most calibration pipelines. With it, we can model both the symmetric radial distortion and the tangential distortion of the lens:
Note
Radial distortion terms can expand beyond \(k_3\). This would just mean that there are major distortion artifacts near the edges of the field of view. However, This is not usually done in practice; most lenses are manufactured well nowadays, and more terms aren't necessarily needed for an accurate model.
Based on these definitions, the radial distortion terms are often simplified into \(x\) and \(y\) components:
The Brown-Conrady distortion model is useful for cameras that have a standard field-of-view lens. The tangential components of this distortion can often correlate with \(c_x\) and \(c_y\), and so may be left out. However, in cameras with complex lens assemblies (i.e. a compound lens with many parts), decentering distortion is very likely to be a problem and should be measured.
Further Reading
The Brown-Conrady distortion model is originally described in two papers. Read them here and here.
Kannala-Brandt
The Kannala-Brandt model is an empirically derived model that is parameterized slightly differently than Brown-Conrady. Instead of parameterizing the radial distortion according to the radial component of the polar image point \(r\), it is instead parameterized in terms of \(\theta\), the angle between the principal axis and an extended ray to the image point.
The Kannala-Brandt distortion model is most useful when used with high field-of-view lenses (typically greater than 90° FOV). Smaller field-of-view lenses are better served by the Brown-Conrady model, as it is conceptually simpler and tends to reduce projective compensation.
The complete Kannala-Brandt model contains separate terms for all three distortion types; however, the Tangram Vision Platform only supports symmetric distortion. Thus, the Kannala-Brandt model can be summarized as:
Note
Only the first four terms of this expansion are modeled. Symmetric terms beyond \(k_4\) are seldom used and are typically insignificant.
As with Brown-Conrady, this is often broken into:
Further Reading
The Kannala-Brandt distortion model used in the Tangram Vision Platform is originally described in this paper.
Note
Why only symmetrical distortion? We do this merely because the full 23-parameter model is often overkill (and the original paper states as much!).