Common Vision Blox 15.0
Foundation Package

Common Vision Blox Foundation Package (Windows only!)


C-Style

C++

 .Net API (C#, VB, F#) 

Python
 CVCFoundation.dll   Cvb::Foundation   Stemmer.Cvb.Foundation   cvb.foundation 

Introduction

The Common Vision Blox Foundation Package is a complete easy to use comprehensive collection of optimised image processing functions.
Based on the CVB Image Manager that offers unique functionality in image acquisition, image handling and image processing, it contains an extensive set of image processing functionality allowing you to control many different types of image acquisition hardware, as well as providing an optimised image display.
Basic functionality such as abstract image access is possible using predefined tools as well as user specific algorithms.
It is also easily possible to create special algorithms for your specific application based on the samples delivered for all supported compilers.

In addition to core machine vision technologies the CVB Foundation Package contains an extensive set of optimised algorithms for

  • edge detection,
  • blob analysis,
  • segmentation of objects,
  • functions for statistical image analysis or correlation of images
  • functions for filtering of images or superimposing destructive text overlays.

It even contains an extensive range of

  • arithmetic and logical functions,
  • functions for binarization using dynamic thresholding,
  • different colour conversion functions,
  • Bayer to RGB decoding functions,
  • algorithm for 2D spatial calibration of image data and
  • metric conversion functions.

Beside the Foundation Tools, CVB Foundation offers binaries, header files and tutorials under:

> Common Vision Blox > Tutorial > Foundation  (CVBTutorial/Foundation)

Note, that for some tools, ActiveX Controls are available, too:

Overview

The separately available components that are bundled in the Common Vision Blox Foundation Package are

  • Common Vision Blox Image Manager
  • Arithmetic (1)
  • BayerToRGB
  • Curves (2)
  • Edge
  • LightMeter & DrawGraph
  • Metric
  • TextOut

(1) Note that the functions of CVB Arithmetic have been superseded by new functions introduced in the Common Vision Blox Foundation Package. Nevertheless, Arithmetic is part of the Package to facilitate an easier migration for existing customers for Arithmetic. However we strongly recommend new customers to use the new arithmetic functions in the Foundation Package.

(2) There is a separate theory of operations chapter available for this topic.

Additional functionality that comes with the Common Vision Blox Foundation Package may roughly be split into the following groups:

  • Functions for Non-linear Transformations and Coordinate Calibration
  • Affine Linear image Transformations
  • Arithmetic and Logic Operations
  • Color Space Conversions
  • LUT functionality
  • Thresholding
  • Morphological Filters
  • Linear and Non-linear Filters
  • Higher Order Evaluation of Images
  • Correlation
  • Blob detection and Measurement (1)
  • Utility functions for 2D geometric operations
  • FFT functions
  • Wavelet Transformation
  • Image Segmentation
  • Intrinsic and extrinsic calibration
  • Other utility functions

(1) The Blob implementation in the Common Vision Blox Foundation Package does not supersede the CVB Blob tool. It is more of a complement to the CVB Blob tool, in that it implements the standard functionality like measurement of center, area and 2nd order moments, but it lacks the ability to measure perimeters, holes or convex projections. It is to a high degree code compatible with CVB Blob, allowing you easy migration to that tool if you should need to make use of the more advanced features of CVB Blob.

For a list of the executables that form the Common Vision Blox Foundation Package, please see CVBTutorial/Foundation.

Foundation API

The bundled tools mentioned in the overview come with all their binaries, header files and tutorials.

Common Vision Blox Foundation Package contains resources for:

  • Image Manager
  • Arithmetic
  • BayerToRGB
  • Edge
  • Lightmeter
  • Polarization
  • TextOut

CVFoundation DLL

with functions for

Examples can be found in Common Vision Blox Tutorial ˃ Foundation directory.

To use any of the functions in the Foundation Package Libraries it is necessary to include libraries and header files to the development project. These libraries and headers are different for each supported development language.

Find here an example description for a programming approach:

CSharp:

Libraries CVFoundation.dll
.Net Wrapper Dlls ../Lib/Net/ iCVFoundation.dll
.Net Code documentation ./Lib/Net/ iCVFoundation.xml
Examples ../Tutorial/Foundation/VB.Net, CSharp,....

C++ ( Visual C, VC.Net unmanaged)

Libraries CVFoundation.dll and ../Lib/C/CVFoundation.lib
Function declarations: ./Lib/C/ iCVFoundation.h
Examples ../Tutorial/Foundation/VC

FArithmetic

Image Processing Tool for Arithmetic and Logical Operations

There are arithmetic and logical operators as a part of the Common Vision Blox Foundation Package available.

  • Available in CVFoundation.dll
  • Functions for arithmetic operations on images
  • Functions for arithmetic operations with constant values
  • Focused on providing high processing speed
  • Works on monochrome and color images
Examples
VC Common Vision Blox > Tutorial > Foundation > VC > VCArithmeticSingleImg
  Common Vision Blox > Tutorial > Foundation > VC > VCArithmeticDualImg

FBlob

Image Processing Tool for FBlob analysis

  • Foundation Blob Functions available in CVFoundation.dll
  • Supports rectangular AOIs
  • Functions for Initialization, Parameter Inquiry, Execution, Filtering and Sorting of Blob results
  • Focused on providing high processing speed
  • Works on single planes

FBlob is a part of the Common Vision Blox Foundation Package for measuring various morphometric parameters in objects of any shape (contiguous pixel ranges) which are defined by means of a binary threshold (blobs).
In this process, FBlob does not analyze individual pixels but operates with a representation of contiguous object ranges in an image row – the run length code.
Each image row is coded in such a way that the start address and length in the X direction are stored for every contiguous object range.
This approach speeds up operations considerably compared with a pixel-based or recursion-based algorithm.
Then the adjacency of the object coordinates in the current row to the existing objects is analyzed, and the measurement parameters of these objects are updated accordingly.
Objects can be of any shape and complexity.

It is possible to make the return of measured values dependent on whether an object satisfies certain criteria (filters).
For instance, a size threshold can be set below so that small objects that are frequently generated by noise are not measured, or objects that touch an image border can be suppressed.

The FBlob implementation bears a lot of similarity to the CVB Blob tool that has been available in the Common Vision Blox tool series for many years now.
However the FBlob implementation - compared to CVB Blob - is limited in its capabilities.
While CVB Blob returns a large number of results for the individual blobs, FBlob is limited to the following set of Blob parameters:

  • area (measured without the holes)
  • center of mass
  • 2nd order moments, including the axis of minimum moments
  • bounding boxes

If you need more information about your objects we recommend that you use the CVB Blob tool.
The usage and function signatures of FBlob and CVC Blob are built in such a way that those two implementations are almost interchangeable (with the exception of those features that are not common to both implementations).

Examples
CSharp CSharp FBlob Example (needs .NET Runtime): Common Vision Blox > Tutorial > Foundation > CSharp > CSFBlobExample
VC  VC FBlob Example: Common Vision Blox > Tutorial > Foundation > VC > VCFBlobDemo

Colorspace Conversion

The CVB Foundation package provides a set of functions for colour space conversion.

Overview

  • Available in CVFoundation.dll
  • Functions for color space conversion
  • Supported color modes: RGB, YUV, YCbCr, XYZ, LUV, Lab, YCC, HLS, HSV
  • Focused on providing high processing speed

It supports conversions between the following color modes:

RGB, YUV, YCbCr, XYZ, LUV, Lab, YCC, HLS, HSV

It supports also the conversion from any of the color modes to gray scale and to a custom model by using the following color twist matrix, too.

Related Topics

More about Color Spaces

Examples
CSharp Common Vision Blox > Tutorial > Foundation > CSharp > CSColorSpaceConversion
VC  Common Vision Blox > Tutorial > Foundation > VC > VCColourSpaceConv

About Color Spaces

Description

Color spaces are mathematical models that help us describe colors (e.g. in an image) through a tuple of n numbers (in most color spaces n = 3, sometimes n = 4).
Any color space or color model defines color relative to a certain reference color space.
For example the chromaticity diagrams of the CIE (Commission Internationale de l'Éclairage) that have been designed to encompass all the colors that may be distinguished by the human eye may serve as such a reference (note that these chromaticity diagrams evolved over the years, so it is always necessary to specify to which of the diagrams you actually refer):

CIE chromaticity chart issued 1931

In such a chromaticity chart, the principal components of a color space represent a distinct location, and the colors that may be described by a given color model are represented by an area spanned by these principal component points.

One of the most common color models, and certainly one with which any one working with computer is to some degree familiar, is the RGB color model, where a color is derived through the additive mixing of the three principial components red, green and blue.
In PAL broadcast and video compression, the color space YUV or YCbCr is commonly used. Many more models exist, and a subset of them has become accessible in Common Vision Blox through the Color Space Conversion functions found in the CVFoundation.
A more detailed description of the supported color spaces may be found in the discriptions of the respective transformation functions

Recommended Reading:

  • R. W. G. Hunt, The Reproduction of Colour in Photography, Printing & Television, 5th Ed. Fountain Press, England, 1995. ISBN 0863433812
  • Mark D. Fairchild, Color Appearance Models, Addison-Wesley, Reading, MA (1998). ISBN 0-201-63464-3
  • http://en.wikipedia.org/wiki/Color_space

Color Spaces in CVB

With the inclusion of a set of Color Space conversion functions in the Common Vision Blox Foundation Package, it has become necessary to add to an image object the information about which color space has been used to describe the colors in the image.

Therefore the CVCImg.dll exports the function ImageColorModel, which returns an enumerated value describing the color space that is being used.
Unfortunately, only few acquisition systems and image formats actually give this kind of information.
Therefore a way had to be found, how to treat images that do not have a color space information attached to them, and the following simple heuristic has been chosen:

  • If the image supports a color space information, the color space number is returned by ImageColorModel (have a look at the Image Dll header file to get the list of possible return values).
    Currently the only examples for such images are the images produced by the color space conversion routines of the CVFoundation.dll.
  • If the image does not have a color space information, but the number of planes is 1, then the enum value CM_Guess_Mono (-1) is returned.
  • If the image does not have a color space information, but the number of planes is 3, then the enum value CM_Guess_RGB (-2) is returned.
  • In all other cases, the enum value CM_Unknown (0) is returned.

Note that the guessed values have been chosen such that abs( CM_Guess_Mono ) = CM_Mono and abs( CM_Guess_RGB ) = CM_RGB.

Whenever an image is passed to one of the color space conversion functions exported by the CVFoundation.dll, it first checks for the appropriate input format. CM_Guess_RGB is always believed to be an appropriate input format.
For example an image passed to ConvertYUVToRGB must either have the color model CM_YUV (produced by ConvertRGBToYUV) or CM_Guess_RGB (if loaded from a file or from a frame grabber) - otherwise the image will be rejected and the return code 27 (CVC_E_INVALIDCOLORMODEL) will be given to indicate the mistake.

Correlation

CVFoundation.dll provides functions for image correlation.

Foundation Correlation functions in  CVFoundation.dll  provides correlation functions such as:

  • FindCorrelationMatches
  • CalculateCorrelationCoefficents
  • CalculateCrossCorrelations
  • CalculateSumOfSquareDistances
Examples
VC  Common Vision Blox > Tutorial > Foundation > VC > VCCorrelation

Fast Fourier Transformation

CVFoundation.dll provides functions for Foundation Fast Fourier Transform Functions

  • Available in CVFoundation.dll
  • Functions for - Forward and inverse FFT Transformation - Filtering - FFT Spectrum
  • Optimized high accuracy FFT Algorithm
  • Functions to generate high pass, low pass, band pass and band stop filter in fourier space
  • Supports odd resolution
  • Supports 8 to 16 bit unsigned integers and float datatypes
Examples
CSharp  Common Vision Blox > Tutorial > Foundation > CSharp > CSFoundationFFT
VC  Common Vision Blox > Tutorial > Foundation > VC > VCFFTFilter
VC  Common Vision Blox > Tutorial > Foundation > VC > VCFFTScaling

Curves

Image Processing functions for Curves

Curves

A curve is a function that maps from a 1D space into a higher dimensional space. For CVB this space is restricted to 2D. Curves are often drawn in 2D and can be confused with graphs that map from 1D to 1D but are also drawn in 2D. These graphs however, map from 1D to 1D. Drawing a circle on an image can be described as a curve, but no graph exists that can do the same. Curves are used to restrict areas or describe general shapes. Curves can be rotated, translated and more while keeping properties of a curve.  As such they are an elementary tool in image processing and for describing shapes.

Curves in CVB are represented by a list of ordered points with linear interpolation between them. This is also called a "discrete representation". An example of a spiral curve is depicted in the figure below. On the left side the points are stored in the correct order and the curve is as expected. On the right side the points where shuffled randomly and stored as a list again. When interpreting this shuffled point set as a curve the curve would look like the right image.

For handling point clouds CVB includes functions to

  • re-sample curves
  • align curves
  • measure distances to curves
  • intersect curves and lines
  • merge two curves and
  • calculate the enclosed area.

These functions and their underlying theory is explained in the corresponding subsections.

Some general advice when working with curves:

  • Use the re-sampling method if you need equidistant curves. This can be advantageous if two curves are from different sources such as CAD and camera. A CAD-constructed curve will have a higher point density in areas with high curvature while the measured curve will have a more uniformly distributed density along the curve.
  • CVB does not account for different speed along different curves. This means only the geometry is considered and constant speed along the curve for all curves is assumed, i.e. s(t) = s(2*t).
  • If you have a graph and rotate it, it may not be a graph anymore. Consider interpreting it as a curve to be able to rotate and transform it without any restriction. This is usually the case when working with laser triangulation where a measured profile can be interpreted as graph or curve.
  • A polygon is a closed curve, i.e. starting point is equal to end point. Any curve function can therefore be used when working with polygons. Of special interest in this case is alignment, re-sampling and enclosed area.
  • Merging curves is not a Boolean operator. It is applied on the given curves, not any possible underlying volume element.
  • If the order of points representing the curve was altered, the algorithm will not be able to detect this nor is there any way to automatically reorder the points. The user must take care to not alter any order. All provided curve algorithms will not alter the order, even if all points are transformed.

Distances and Intersection

In this section different distances and how they are defined as well as calculated are presented.

There are three different distances that need to be clarified when considering the distance of a point from a curve. For the remainder of this text the curve is denoted with s(t) and assumed to be represented by an ordered point set. The distance is being calculated with respect to point p (denoted "sample point" in the picture above).

  1. Point to point
  2. Point to curve with interpolation
  3. Point to curve with interpolation and left / right information

All three different distances are depicted above. A more detailed information about the different distances is given in the text below.

The first distance is a direct result of the way how curves are represented in CVB, which is as a list if points. It is natural to use this representation to directly calculate distances between points and curves. This, however does not account for the linear interpolation between the points in the curve and thus is only an approximation.

A more accurate distance does not only consider the point to point distance between curve s(t) and point p, but interpolates the curve correctly to find the normal distance to the curve. To calculate this distance, first the closest point in the point set representing the curve is found. We denote this point as q1. Then, the neighbor of q1 which is closest to p is selected . We denote this point with q2. The final distance is then calculated between the line going though the selected points, q1 and q2,  and the point p.

Since a curve is ordered and there is a direction of traversal from start to end an additional information is available and may be desired: On which side of the curve s(t) the point p is situated. When calculating the point to curve distance with interpolation a hessian normal form is used for calculating the desired distance. This distance is always positive but during the calculation the necessary information about left and right is present. When calculating the distance with side information we denote point on the left with a positive distance and point on the right with a negative distance. Technically speaking, the result is not a distance any more since it doesn't fulfill the mathematical properties of a distance. Using the absolute value of the "signed" distance will resolve this issue. The "signed" distance now includes on which side of the curve the point p is, denoted by its sign, and its distance to the curve.

Intersections

After this thorough examination of distances between points and curves the next topic is how to find intersections between lines and curves. Finding the intersection between a curve and a line can be done by calculating the signed "distance" of each point representing the curve to the given line. The curve and line intersect at a point with distance zero or between two consecutive points with changing sign. Since linear interpolation is applied between two consecutive points in a curve, the exact intersection point can be calculated mathematically by intersecting the two lines.
In CVB all possible intersections are found and stored in a list.

Resampling

To speed up computation times it can be beneficial to only work with the point set representing a curve. Depending on the distances between points this approximation can be accurate enough while giving a significant performance increase.  In some instances points are not distributed uniformly along the curve, this can result from the acquisition or if the curve is generated artificially from a CAD file or mathematical formula. In this case re-sampling the curve to have equidistant points can be helpful. Another reason for re-sampling may be to reduce the overall number of points that need to be stored and processed.

To resample a given curve into n equidistant points, first the overall length of the curve is calculated and divided into n-1 equal parts. Then the curve is re-sampled at the beginning and the calculated length. This ensures that the new points represent the given curve while at the same time being uniformly distributed. An example of a curve with 50 points and its re-sampled counterpart, again with 50 points is shown below.

While this approach is fast and  yields good results it does have two major downsides. Firstly, the new curve was generated from equidistant points but since now the re-sampled curve has different distances between adjacent points due to the linear interpolation, it may not be equally distributed in itself. The figure below illustrates the changes within a curve after resampling (note that the given example will be equidistant after resampling). This effect could be minimized by iterating inversely how the points should be distributed in the original curve. Since in practice the difference should be small, the faster approach was chosen for CVB.

Secondly, this method does not use any type of smoothing kernel or functional approximation. It will suppress high frequency changes within the curve if the new number of points is lower than the original number of points (see example below), but does not provide any further parameters for fine tuning.
For the denoising example a curve is generated from 500 points equally spaced between 0 and 2π in x and y = sin(x)+δ. The noise δ is generated from a normal distribution with center 0 and scale .08. The resulting curve is re-sampled with 250 points and 50 points. The resulting curves are shown below

To see the effect on the high frequency noise the three violin plots for each curve and its difference to an ideal sinus curve are shown.

Alignment

To align two curves an enhanced ICP (iterative closest point) method is implemented. This method allows for fast alignment for curves that may only partly overlap. The input curves are aligned using only the points in the lists without any interpolation. This will ensure that "holes" in the data are not improperly closed which would lead to false information being processed. A typical use case is aligning data from two sensors while the object is only partly visible and/or partly occluded for each sensor. In this case, the data will only represent curves with "holes".

The implemented method is an iterative method which consists of the following steps

  1. find matching point pairs
  2. calculate matching score
  3. check if any stopping criteria is fulfilled
  4. find a rigid body transformation between the found point pairs
  5. start over at 1.

More details about the steps and all possible input parameters are listed below. And an example alignment is shown in the following image.

Finding corresponding pairs and calculate the matching score

Let P1 be the point set representing the first curve and point set P2 represent the second curve. For each point w in P2 the closest neighbor in P1 is found. The found neighbor to w in P1 is denoted v. If the euclidean distance is less than a given threshold, the point pair is considered corresponding to each other. If the value is greater, the closest neighbor of v in P2 is found. This neighbor is denoted w2. If

the point pair v, w is considered corresponding to each other.
The set C is the set of found corresponding point pairs. The matching score between the two sets is only calculated over the set of corresponding point pars. Let the set P3 be the subset of P2 that includes all the points that have a corresponding neighbor in P1. The matching sore is then calculated via

And if no stopping criteria is met, the rigid body transformation between the pairs in C is calculated in a way that minimizes the current matching error. In each iteration of the algorithm the corresponding pairs and thus C is recalculated following the method described above.

Matching score tolerance

If the matching score falls below this threshold, the matching is considered successful and will stop. The default value is 1e-6.

Early stopping criteria

If the matching score does not improve by at least the given tolerance, and early stopping criterion is triggered and the found solution is returned. The default value for this tolerance is 1e-5.

Shape sensitivity

The shape sensitivity is defined by a float value between 0 and 1. A higher value means higher sensitivity. A value of 0 corresponds to the classical ICP without any shape sensitivity. A value between .8-.9 is recommended as default value when aligning curves. This value is not very sensitive, meaning a value of .88 yields similar results to when using .9 as shape sensitivity. This value helps find stable parameters without the need to re-parametrize the function constantly.

Maximum number of iterations

The implemented algorithm is based on the classical ICP algorithm and thus iteratively minimizes the matching discrepancy. The maximum number of iterations restricts the iterations that the algorithm can perform. In most applications the algorithm converges within 30 iterations and thus 40 maximum iterations should be sufficient. The larger the maximum iterations the larger the computation time differences between successful and failed alignments become. If consistency is desired, a small test set should be used for finding a good limit for when convergence should be reached and the maximum number of iterations subsequently be set accordingly.

Restrictions of the method

ICP based methods are local optimizations. As such the alignment depends on the starting positions and rotations of the curves. Local optimization methods have the advantage of being fast and precise but do fail to find a global optimum if the minimization function is non-convex. By introducing shape sensitivity to the classical ICP the provided algorithm is able to converge faster and is less dependent on the starting position and rotation. If multiple matches are possible, this method finds only one, depending on the starting position. See below figure for an example.

Merging

To merge two curves that may overlap we need to have a starting- and end-point for each curve. Just combining the two point lists into one long list will not work if the curves overlap. In this case an average at the overlap or suitable interleaving has to be found. See the image below as an example for a failed merged using the naive approach.

Instead, given two curves for merging, the outcome is supposed to be a single curve that keeps the order in each point set as given while combining both sets with a shortest path. This will ensure that overlapping areas are merged even if they don't match perfectly. To find this shortest path, the implemented algorithm uses nearest neighbor information for each curve and adds interpolation when altering between the two given curves. It is capable of handling curves with holes, and multiple overlaps.

For a successful merge it is important that the order of the given curves is matching, the matching assumes that the first point in each curve has to be merged before the second point. Detecting the best order (front to back or back to front) is not part of the implemented algorithm. If the two curves were generated from different cameras or sensors, the caller has to make sure that left to right is the same for each camera, or save the curves accordingly before calling the merge function. Below is an example of a merge of two operation where the order was incorrect. A successful merge is shown in the example above on the same data.

One issue when merging curves is how to handle diverging curves, meaning two curves that overlap in the middle and then diverge into different directions. The given algorithm will use the closest neighbor information to find a suitable interpolation. Each point of the to curves will be considered and this the divergent path with the most points will ultimately be dominant. Below are some examples of merged diverging curves.

If it is unclear or undefined which one of the two given curves is the starting curve, it is determined by calculating the distance of both possible starting points to the other curve. The one starting point with the larger distance is taken as the starting point for the merged curve. This calculation is optional and can be switched on / off by the user when calling the merge curve function.

Enclosed Area

Calculating the enclosed area of a curve is equal to calculating the area of a polygon. This equality holds, since the curves in CVB are represented by points and linear interpolation between them. As a side note, the approaches described in this section do not work for other representations such as Bezier-Curves.

Below are some examples of different types of curves, sorted from simple to difficult to calculate the enclosed area.

Basic geometric shapes have defined formulas to calculate their area and are therefore easy to calculate. Convex curves can be simply divided into triangles and the area then calculated as the sum of the inner triangles. This approach does not work for non-convex curves however.

Searching for area of a polygon or non-convex polygon one may come across the definition of a simple polygon. This is a polygon that is not self intersecting and not self overlapping. In CVB only the area of simple polygons can be properly calculated. The implemented algorithm is the shoelace algorithm which allows to calculate the area fast, but does not check if the polygon is actually a simple polygon. To check if a curve is a simple polygon the sum of inner angles can be calculated. If and only if the sum of the inner angles is equal to (p-2)*180, where p is the number of points in the polygon. This check is not part of the enclosed area function to save computation time.

The general idea of the shoelace algorithm is to calculate the area of triangles between two consecutive points of the curve and the origin (which is the point (0,0)). While adding up the sum of these triangles would be incorrect, the algorithm does consider sweeping direction between the two points of the curve. and thus does add and subtract triangles to find the correct enclosed area of any curve, as long as the curve does not intersect itself.

In CVB, the implemented algorithm will close the given curve, meaning that if the first and last point in the provided curve do not match, the curve will be closed between first and last point.

Filter

Image Processing functions for Filtering

Overview

The Foundation package provides functionality for various frequently used filters in image processing.
These filters are normally used for preprocessing of images to achieve better results in the image analysis.

  • Available in CVFoundation.dll
  • Morphological filters
  • other frequent filters like Laplace, Gauss and other edge, lowpass or highpass filters
  • Focused on providing high processing speed

Introduction to the Morphological Operators

What is a morphological operator?

Morphology is the study of form. With morphology it is possible to use knowledge about the form of images in order to detect several parts of the image.
For example the knowledge of the form of interference can support efficiently the elimination of fault.
The local erosion and dilation operators are the base morphological operators.
All morphological operators are local operations, that have to work with a structuring element.
The structuring element is a user-defined operator mask, that slides over every pixel of the image and determines the region of influence of the operator.

When to use morphological operators

Erosion is useful for

  • deleting several object pixels
  • smoothing the borders
  • separating objects in the image (objects become smaller, holes become bigger) Dilation is useful for
  • filling several backround pixels
  • smoothing the borders
  • connecting several elements in the image (object become bigger, holes become smaller) Opening is useful for
  • separating objects
  • smoothing the borders (object size nearly unchanged) Closing is useful for
  • connecting objects
  • smoothing the borders (object size nearly unchanged)

Theory of Morphological Operators

The structuring element (image 3 x 3) is dragged pixel per pixel over the image and the input neighborhood, that is covered by the mask, determines the new pixel in the output image.

1. Erosion

When the erosion operator is used the current pixel will be set to true only if the mask completely overlaps with the " 1 - region" of the input image.
Otherwise the current pixel is set to "0".
The image below shows the deletion of the region's boundary.

(a) Input image, (b) Structuring element, (c) Output image
The erosion operator chooses the smallest value from a sorted sequence of neighbouring pixels.
Therefore the darker regions become larger at the expense of the brighter regions (please see the image below).

2. Dilation

The dilation operator adds pixels to the borders of the regions.
Consequently the image object gets bigger.

(a) Input image, (b) Structuring element, (c) Output image
The erosion operator chooses the maximum value from a sorted sequence of neighboring pixels.
Therefore the dark regions become bigger at the expense of the brighter regions (please see the image below).

3. Opening

The opening operator consists of two morphological operations, erosion followed by a dilation (please see the images below).

(a) Input image, (b) Erosion from (a), (c) Result: Dilation from (b)
Frayed object borders are smoothened and small structures like pixel noise or small lines are supressed (please see the images below).

4. Closing

The closing operator consists of two morphological operations, dilation followed by erosion.

(a) Input image (b) Erosion of (a), (c) Result: Dilation of (b)
Small holes in the object's borders are closed (please see the images below).

Morphological Operation examples

Description

Here you can see the effects of the different morphological filter operations:

  • ErodeImage,
  • DilateImage,
  • OpeningImage,
  • ClosingImage and the different settings of  the morphology mask parameter (MM_Custom, MM_Square, MM_Rect, MM_Cross, MM_Circle, MM_Ellipse).

For that purpose the different mask types have been applied under the different operations on the same image:

The custom mask that has been used for the MM_Custom setting is this:

The following table presents the results:

  Erode Dilate Open Close
MM_Custom
Size 16x16
Offset (8,8)
MM_Square
Size 15x15
Offset (7,7)
MM_Rect
Size 7x15
Offset (3,7)
MM_Cross
Size 15x15
Offset (7,7)
MM_Circle
Size 15x16
Offset (7,7)
MM_Ellipse
Size 7x15
Offset (3,7)

Image Manipulation

CVFoundation DLL provides Foundation Image Manipulation functions for

  • converting and scaling images
  • manipulating Image contents
  • image segmentation
  • geometrical image transformation

Image Conversion

CV Foundation provides functions for converting and scaling images.

There are two flavors of functions that perform conversions of (multi-planar) images between two data formats:
The ConvertToXXX and the ScaleToXXX functions (where XXX stands for a short description of the target format).

Both flavors differ fundamentally in the way the conversions are performed (which in turn depends on whether conversion is done into a format with higher bit-depth or lower bit-depth):

  • Converting to higher bit-depth:
    The data bits of the source image are simply copied into the target image and padded with zeroes.
    In other words: no conversion of values is being done. When converting into floating-point format, the numeric values of the input image are simply converted to 32 bit float numbers and written into the destination image.
  • Converting to lower bit-depth:
    When converting to lower bit-depths, the pixel value from the source image is simply copied into the target image, if it is within the dynamic range; otherwise the pixel value is replaced by the allowable maximum of the respective datatype. Operation is likewise for conversion from floating point data to integer data, with the additional conversion step in between.
  • Scaling between integer formats:
    When scaling between integer formats occurs, the following formula is applied:
    Dst = Dstmin + n * (Src - Srcmin) with n = (Dstmax - Dstmin)/(Srcmax - Srcmin)

For scaling from floating point to integer formats, the full range of values in the floating point image is automatically determined before the scaling takes place.
Scaling to floating point format converts the input image to a floating point image with pixel values ranging from the specified minimum to the specified maximum value.
Not all combinations of source and target format are currently possible!
The following table gives an overview of the currently supported scaling and conversion combinations:

Function 8 bits per pixel unsigned 16 bits per pixel unsigned 16 bits per pixel signed 32 bits per pixel signed 32 bits per pixel float
ConvertTo8BPPUnsigned (X) X X X X
ConvertTo16BPPUnsigned X (X) - - X
ConvertTo16BPPSigned X - (X) - X
ConvertTo32BPPSigned X - X (X) -
ConvertTo32BPPFloat X X X - (X)
ScaleTo8BPPUnsigned (X) X X X X
ScaleTo16BPPUnsigned X (X) - - -
ScaleTo16BPPSigned X - (X) - -
ScaleTo32BPPSigned X - - - -
ScaleTo32BPPFloat X - - - -
Examples
CSharp  Common Vision Blox > Tutorial > Foundation > CSharp > CSFoundationFFT ConvertToXXX
VC  Common Vision Blox > Tutorial > Foundation > VC > VCCorrelation and VCFFTFilter ScaleToXXX
VC  Common Vision Blox > Tutorial > GPUprocessing > VC > VCBenchmark
VC  Common Vision Blox > Tutorial > GPUprocessing > VC > VC32bitFloatingPoint

Image Manipulation

CV Foundation provides functions for manipulating Image contents.

Examples
Delphi  Common Vision Blox > Tutorial > Foundation > Delphi > ImageManipulation

Image Segmentation

CV Foundation provides functions for Image Segmentation based on region growing with markers or watershed.
These functions only support 8-Bit images.

Examples
VC  Common Vision Blox > Tutorial > Foundation > VC > VCSegmentation

Image Transformation

CV Foundation provides functions for  Image Segmentation based on region growing with markers or watershed.
These functions only support 8-Bit images.

Examples
CSharp  Common Vision Blox > Tutorial > Foundation > CSharp > CSGeoTransform

LUT

Foundation LUT functions for adjusting the LookUp Table adjust an image intensity values by offering four different LUT modi.

  • Available in CVFoundation.dll
  • Functions for adjusting an image intensity values by offering four different LUT modi
  • Focused on providing high processing speed

What is a Lookup Table?

A Look Up Table (LUT) is an array indexed by the intensity value of the input pixel, containing in its entry the new intensity value of the output pixel.
The formula below describes this relationship: Dst = Values[Src] Src stands for the intensity value of the input pixel, and is used as the index in the LUT array Values for determining the output intensity Dst.

Two approaches

The Lookup Table (LUT) in the CVB Foundation Package is described in two ways:

  • One dimension
    The LUT is described as an Array of 256 intensity values (similar to hardware LUTs in framegrabbers).
    The function ApplyLUT8Bit is available.
  • Two dimensions
    The LUT is defined by a number of points (nodes) in a two-dimensional diagram.
    Between those nodes either the intensity of the previous node is used, or a linear or cubic interpolation is applied. The functions ApplyLUTGeneric, ApplyLUTLinear and ApplyLUTCubic are available.

The image below shows a sample input image converted with the three LUT functions mentioned above.

MathHelpers

The CVFoundation.dll contains a group of functions that may be properly described by the term "Math Helpers".
This set of functions provides functionality to easily perform calculations involving several mathematical entities that exist in 2-dimensional vector space such as angles, vectors, lines and circles, allowing you to easily calculate e.g. normal vectors, intersections and even regressions.

Statistics

These functions provide statistical information on the gray value distribution in an image.
For simple information such as average, minimum and maximum brightness and standard deviation please refer to the Lightmeter Tool.

Thresholding

CVFoundation.dll contains functionality for Thresholding.

  • Available in  CVFoundation.dll
  • Functions for thresholding using images, constant values or performing dynamic thresholding
  • Focused on providing high processing speed
Examples
VC  Common Vision Blox > Tutorial > Foundation > VC > VCThresholding

Transformations

CVB Foundation Package includes several useful Geometric Image Transformations and Non-Lineare Coordinate Transformations.

Geometrical Image Transformations

CVB Foundation Package includes several useful geometric image transformations for rotating, mirroring, resizing images as well as Matrix Transformations and Perspective Transformations.

Perspective Transformation:

This Perspective Transformation transforms the source image pixel coordinates (x,y) according to the following formula:

  • x' = (c00*x + c01*y + c02)/(c20*x + c21*y + c22)
  • y' = (c10*x + c11*y + c12)/(c20*x + c21*y + c22)

where x' and y' denote the pixel coordinates in the transformed image, and cij are the perspective transform coefficients passed in the array coeffs.

Example of a Perspective Transformation, Transformation of one cube side surface

Source image and result of the transformation:

As you can see, the surface "plane" of the right cube side was transformed as if a virtual camera would be look perpendicular onto it.
The resulting image may now be subject to further processing.
The function CalcPerspectiveTransformation is used to obtain the transform coeffcients C00 - C22 first.
In order to get these coefficients, you have to specify vertex coordinates of a quadrangle which will be mapped to a destination ROI.
After the coefficients have been calculated it is possible to transform an image with the function PerspectiveTransformImage.
In addition to the height and width of the destination image you can also set a certain offset to place the section of the transformed space you would like to see.

Non-Linear Transformations

Introduction

Generally speaking, Coordinate Transformations are useful whenever certain characteristics of an image taken by an optical system need to be corrected.
Probably the two most common uses are

  • Correction for distortions that arise from non-perpendicular viewing angles (either because the camera was mounted at an odd angle with respect to the scenery, or because the object being inspected by the camera has uneven surfaces.
  • Correction for optical aberrations that are introduced through the lens being used.
    Standard lenses tend to distort the image more and more heavily the shorter their focal length is (in extreme cases known as "Fisheye effect").

In most scenarios the total distortion is a mixture of both of these effects. For example the following image has been taken by looking sideways at the slab of an optical stand through an 8 mm C-Mount lens.
Of course the perspective distortion is obvious, but if you have a close look at the lines close to the image boundaries you may notice that the lines of the grid on the slab are not straight but bent, due to the aberration of the lens being used here.

In most cases it is possible to find a coordinate transformation that allows us to transform an image as the one seen above into one that is seen as if the camera was mounted directly atop the slab.

Modeling the Transformation

There are basically two approaches to find such a transformation:

  1. The more "scientific" approach is to derive a more or less elaborate model of the distortions bein introduced through the optical system, then determine the model parameters by fitting the model to data derived from a calibration measurement.
  2. The "heuristic" approach is to try to describe the unknown exact transformation as a linear combination of suitable (orthogonal) base functions by fitting such a linear combination to calibration measurement data. The most obvious example of such base functions would of course by polynomials, but generally any set of orthogonal functions will do.

Both approaches have their advantages and disadvantages. The "scientific" approach is the more correct one, but the models necessary to properly describe the transformation are either very complex or lack portability to other problems.
As a side-effect of their complexity, they are normally numerically expensive because they take lots of calculations to perform the transformation.
On the other hand, when using a "heuristic" approach one needs to be aware, that it is not the true distortion towards which you are fitting, but just an approximation, and the quality of the fit does not really tell about the quality of the approximation - what this means will become more obvious throughout this description.

We have chosen to introduce in Common Vision Blox a non-linear transformation (as opposed to the affine linear transformations that are possible through the Coordinate System) based on a 2-dimensional polynomial model, therefore we will now have a look at the characteristics and shortcomings of the 2nd method.

Non-Linear Transformations Through 2 Dimensional Polynomials

As pointed out in the previous section, we are using in Common Vision Blox a 2 dimensional polynomial model to perform Non-Linear Coordinate Transformations.
The selection of polynomials is of course justified by the fact that polynomials are known to be suitable to approximate any continuous and differentiable function and that we expect the unknown exact transformation function ƒ(x,y) -> (x',y') to fall into this category.
The order of the polynomial is selectable, giving us a possibility to select between the quality of the fit and the speed of the transformation.
Of course there are always two such polynomials: One transforming the x coordinate and one transforming the y coordinate.

An example for a 2nd order transformation looks like this:

x' = a1 * x² + a2 * x * y + a3 * y² + a4 * x + a5 * y + a6

y' = b1 * x² + b2 * x * y + b3 * y² + b4 * x + b5 * y + b6

All that is left to do to is to find the actual values of a1, a2, a3, a4, a5, a6, b1, b2, b3, b4, b5 and b6 by performing a fit to calibration data that minimizes the error of the polynomial transformation on the calibration data.
Now mathematics tells us that the higher the order of the polynomials we use for the approximation of ƒ, the smaller the deviation between our polynomial and ƒ becomes.
While globally speaking this is really true, we should nevertheless not take this as an indication to choose the polynomial order as high as possible.
There are numerical and logical constraints that need to be taken into account.
irst of all, the order cannot be chosen arbitrarily, because we need at least as many calibration points as we have variables to fit to.
This is kind of trivial, but nevertheless worth mentioning (by the way: The number of variables for a 2 dimensional polynomial of Order n is n+1 added to the number of variables for order n-1).
The second reason is more difficult to understand: Even though higher order polynomials give a better global approximation of ƒ, the quality of the approximation may locally deteriorate!
Why ist that? Now, polynomials of higher order tend to oscillate (in fact they do oscillate so well that even sine and cosine functions may be approximated using polynomials).
And if the length of these oscillations is about the same order of magnitude as the distance between our calibration points, all we gain from higher order polynomials is a more pronounced oscillation between our calibration points - up to the point where the transformation becomes useless because it introduces too many errors between the calibration points - even though the error on the calibration points themselves is zero or at least lower than with lower order polynomials!

To illustrate this, let's have a look at the following case: Again a camera is looking sideways at the slab of a camera stand, this time with a 12 mm lens that does not introduce such a pronounced fish-eye effect.
Thanks to the grid we can easily define a great number of calibration points on the slab - we have chosen 256, arranged in a square of 16 by 16 crossing points on the grid.
With that number of coefficients we can easily calculate polynomials of up to 14th order and have a look at the transformation:
This is the untransformed image with the calibration points shown:

The 1st order transformation looks like this. Of course this is a linear transformation that would also have been possible to describe through an appropriate coordinate system of the image.
We clearly see that the linear transformation is not sufficient to describe the distortions in the image:

The 2nd order transformation already looks considerably better, especially near the center.
There is some distortion left in the left top and left bottom corner, which we could not get rid of with a (still quite simple) 2nd order polynomial...

... but already with a 3rd Order polynomial even these flaws are gone. Judging from the looks of this result, a 3rd order polynomial would be enough to get a good transformation result.
Of course the look alone is not always sufficient to judge if a calibration is good or not - we will come back to that thought a few lines later...

We may now continue increasing the order of the polynomial without seeing much of a change at first. However at some point (about 8th or 9th order) the transformed image starts to deviate from what we have seen so far.
With 11th order, the problem already is clearly visible: The lines forming the grid have started to bend between the calibration points, which is best seen, but not limited to the squares near the image border.

With 14th order, the highest order we may achieve with our 256 calibration points, things get even weirder:

Now this looks ugly. Heavily bent lines near the borders, one of the broader outside lines has even been split in two.
But believe it or not: The error on the calibration points is not higher than it was with the 3rd order transformation!
This clearly illustrates that looking at the residual error of the polynomial fit of the calibration points does not tell us enough about the correspondence between the transformation and its polynomial approximation. To determine a numerical value for this correspondence we would need to introduce some kind of norm (e. g. sum up the difference image between the original and the result of transformation and inverse transformation...).
In the absence of such a norm, the choice of the actual order of the transformation is left to our subjective impression which order solves the task best (and of course the timing needs of your application, because with increasing order, of course also the transformations will take more time to calculate, therefore we always will have a tendency to prefer lower order polynomials).

Note: Of course it would be possible to chose a regularization approach to circumvent these problems by introducing a penalty term that penalizes higher order polynomials in the regression.
While this approach evades the problem of overfitting quite elegantly it leaves us with another problem: We would need to always calculate higher order polynomials, even if they hardly contribute.
Therefore we have chosen to leave the choice of the order to be used to the user, instead of giving the user a regularization parameter to play with...

Non-Linear Transformations in Common Vision Blox

After introducing the background information necessary to use the non-linear transformations, it's about time to talk about implementation details.
In Common Vision Blox, a Non-Linear Transformation is always represented by an object whose handle is accessible through the data type NLTRANSFORMATION.

Just like the IMG object, NLTRANSFORMATION objects are reference counted object, which means that you may use the functions ShareObject, ReleaseObject and RefCount to manage the lifetime of these transformation objects.
They may be loaded from mass storage devices or stored thereon by calling LoadNLTransformFile and WriteNLTransformFile.
The functions IsNLTransform, NLTransformOrder, NLTransformNumCoefficients and NLTransformCoefficients give you further informations about a specific transformation object.

A transformation object is always generated by a call to CreateNLTransform in which the caller provides two pixel lists that give the calibration points in untransformed and transformed 2D space.
Note that along with the transformation, the inverse transformation is at the same time, so an NLTRANSFORMATION object holds both aspects of a transformation.
By the way: Of course those calibration points should not all be aligned on a single line (in other words: They should not all be linear dependent) otherwise it may not be possible to calculate a proper transformation from them).
To apply a transformation that has been loaded or created, the functions ApplyNLTransform, ApplyInverseNLTransform (both to transform a single coordinate) and CreateNLTransformedImage and CreateInverseNLTransformedImage (to transform a part of an image) exist.

Limitations

One thing to keep in mind is certainly, that transforming complete images is a time-intensive operation.
Even with only 2nd ord 3rd order polynomials you can easily spend several 100 ms in functions like CreateNLTransformedImage.
Therefore you should consider well if you really need a transformation of the whole image, or if you can do with transformations of a few coordinates only.
The first is normally only necessary, if you need the undistorted image for further processing (like doing OCR on it).
If the aim is just the transformation of a processing result in real-world coordinates, it's normally sufficient to transform a single point.

It should also be pointed out that this kind of non-linear transformation is only a 2-dimensional one.
Let's have one more look at the slab used in the previous examples: We have performed a calibration by using calibration points that were distributed over the (2-dimensional!) surface of the slab.
If we now put a 3-dimensional object on the slab...

... and then apply the coordinate transformation that has been derived from the flat calibration points, then the result looks somehow unnatural:

This is of course because the transformation does not know or care about the 3-dimensional nature of our object on the slab and just treats the image provided for transformation as if everything on the image were the same height as the slab.
If we would have wanted the 3-dimensional object to be transformed "properly", it would have been necessary to adjust the calibration points accordingly (by using calibration points taken on the object's surface rather than on the ground on which the object is lying).

Tutorial Programs

For an easy access and overview and for feasibility there are several Tutorial programs available.
All of the provided with Source code so that the user could try out first if the functionality serves for his special purpose and to see the processing time of the function.

Geometrical Transformations

Non Linear Transformation

Wavelet Transformations

The CVFoundation.dll contains a group of functions for Wavelet Transformations.

Calibration of Linescan Cameras

Introduction

For high precision applications (e.g. pick and place), images acquired by linescan cameras have to be calibrated. Therefore the sensor direction (including the lens distortion errors) and the direction of movement (including the encoder step) need to be considered.

The following sections describe the theoretical aspects of calibration and provide step-by-step guidance on calibrating using CVB.

Setup and Prerequisites

Before starting the software calibration the camera hardware needs to be setup correctly and the device temperature should already be in the working point. Once these prerequisites are met, images containing a calibration target can be acquired. The main part of the linescan calibration is the correction of errors induced by lens distortion. These errors can be approximated by a 3rd order polynomial as described in detail in section Lens Distortion Calibration. For the estimation of the correction parameters a pattern with alternating black and white stripes where the width of the stripes is precisely known can be used. Ensure that the stripes are oriented vertically to the scan direction. The provided reference width corresponds to the stripe width as seen in the image scan direction. However, if the stripes are tilted, it's necessary to recalculate the reference width based on the tilt angle.

After applying the correction for the lens distortion, coordinates along the sensor line become metric. In order to get square pixels, where x- and y-axis share the same units, also the direction of movement has to be calibrated by a scaling factor. Therefore two circular markers with known position can be used (see dots in the figures below). Note, that for the calibration two separate acquisitions can be used: one with the markers and one with the stripe pattern. However, it is essential to ensure that both images have identical widths.

There are two distinct configurations based on whether the scan line of the sensor is stored as a row in the image, spanning all columns (indicating a scan direction along x-axis) or stored as a column, spanning all rows (indicating a scan direction is along y-axis).

ATTENTION: Please note that in some linescan cameras, the user can set the parameter "Scan Direction", which indicates the "direction of movement"! The "Scan Direction" parameter of the algorithm in Common Vision Blox is not related to the "Scan Direction" of the linescan camera.


Scan direction in X Scan direction in Y

The functionality of the linescan camera calibration is implemented in the CVCFoundation.dll. To execute the code snippets provided in this documentation, following DLLs are additionally required:

To utilize this functionality, the following namespaces and include files are required:

Namespace: Cvb::Foundation::CalibrationLineScan
#include <cvb/foundation/calibration_line_scan.hpp>
Namespace: cvb.foundation
import cvb
import numpy as np

Calibration

The results of the calibration will be stored as a 3rd order polynomial in the CVB transformation object:

The coefficients stored in this object represent a 3rd order polynomial defined as follows:

‍x' = a1 * x3 + a2 * x2y + a3 * xy2 + a4 * y3 + a5 * x2 + a6 * x * y + a7 * y2 + a8 * x + a9 * y + a10

‍y' = b1 * x3 + b2 * x2y + b3 * xy2 + b4 * y3 + b5 * x2 + b6 * x * y + b7 * y2 + b8 * x + b9 * y + b10

where
x: pixel position in x (column),
y: pixel position in y (row),
a1-a10: coefficients in x,
b1-b10: coefficients in y,
x' and y': transformed x and y coordinates.

Note, that most of the coefficients are zero for the linescan calibration. They are described in detail in the following two sections.

Lens Distortion Calibration

The main part of the linescan calibration is the correction of errors induced by lens distortion. These errors can be approximated by a 3rd order polynomial, neglecting mixed coefficients. Depending on the scan direction depicted in the figures above, the coefficients correcting the lens distortion are stored to the x or y coefficients in the CVB transformation object. From the previously described equation, it follows for the corresponding scan direction:

Scan direction in X:     x' = a1 * x3 + a5 * x2 + a8 * x + a10
Scan direction in Y:     y' = b4 * y3 + b7 * y2 + b9 * y + b10

For the estimation of the correction parameters a pattern with alternating black and white stripes where the width of the stripes is known can be used. The pattern should be designed as described in section Setup and Prerequisites. The stripes of the target must be detected in the image first, which will be described in the following section.

Stripe Target Detection

For the detection of the stripes, the CVB Edge Tool is internally used. The edges are detected using the contrast method. The following code example detects the stripes from an image. The parameters to be set are described in detail below.

// load image with stripes
auto imgfile = ...
auto image = Cvb::Image::Load(imgfile);
// configure detection
int numStripes = ...
double threshold = ...
// define AOI including only areas with stripes
auto area = Cvb::Rect<int>(0, 0, image->Width()-1, image->Height()-1); // left, top, right, bottom
// detect edges
auto detectedEdges = Cvb::Foundation::CalibrationLineScan::DetectEdgesOfStripeTarget(*image, area, numStripes, scanDirection, threshold);
static std::unique_ptr< Image > Load(const String &fileName)
EdgeDetectionResult DetectEdgesOfStripeTarget(const Image &imageStripes, const Rect< int > &aoi, int numStripes, const ScanDirection &scanDirection, double threshold)
// load image with stripes
var image = Image.FromFile(...);
// configure detection
var scanDirection = ScanDirection.X;
int numStripes = ...
double threshold = ...
// define AOI including only areas with stripes
var area = new Rect(0, 0, image.Width - 1, image.Height - 1);
// detect edges
var detectedEdges = CalibrationLineScan.DetectEdgesOfStripeTarget(image, area, numStripes, scanDirection, threshold);
static EdgeDetectionResult DetectEdgesOfStripeTarget(Image imageStripes, Rect aoi, int numStripes, ScanDirection scanDirection, double threshold)
static FileImage FromFile(string fileName)
# load image with stripe target
imgfile = ...
image = cvb.Image.load(imgfile)
# configure detection
scan_direction = cvb.foundation.ScanDirection.X
num_stripes = ...
threshold = ...
# define AOI including only areas with stripes
area = cvb.Rect.create(0, 0, image.width-1, image.height-1) # left, top, right, bottom
# detect edges
detected_edges = cvb.foundation.detect_edges_of_stripe_target(image, area, num_stripes, scan_direction, threshold)
cvb.Image load(str file_name)
cvb.Rect create(float left=0.0, float top=0.0, float right=0.0, float bottom=0.0)
cvb.foundation.EdgeDetectionResult detect_edges_of_stripe_target(cvb.Image image_stripes, cvb.Rect aoi, int num_stripes, int scan_direction, float threshold)

After loading the image with the stripe target, the following input parameters need be set:

  • The area of interest must only cover areas with stripes. Since the reference width corresponds to the stripe width as seen in the image scan direction, the stripes have to be oriented vertically to the scan direction. If they are tilted, the reference width has to be recalculated considering the tilt angle φ: ref_width_tilted = ref_width / cos(φ) (where ref_width represents the actual width of stripes).
  • The number of stripes included by the target must be specified. Note, that this value is only used for memory allocation. It does not necessarily have to precisely match the number of stripes, but it must be equal to or greater than the actual number of stripes.
  • The scan direction must be set correctly. It can be along the x- or the y-axis as depicted in the figures above.
  • Threshold for gray value change: Detailed information can be found here. A good value is 20. Experiment on your own to determine the value that yields satisfactory results.

The edge detection is done line by line. Only lines, where the same number of edges are found, are stored to the result object.

Direction of Movement Calibration

If square pixel with accurate metric values are needed, the direction of movement has to be calibrated, too. If the encode step (velocity of camera or object movement) is precisely known, you can use this value for the calibration as outlined in section Special Case: Known Encoder Step, which represents the simpler scenario.

In some cases, metric values along the y-axis might not be necessary. In such cases, users can also follow the steps outlined in Special Case: Known Encoder Step using a fictive encoder step.

But in most cases, you will have to calibrate the direction of movement. Therefore an image with two circular markers has to be acquired. Note, that it is essential to ensure that the marker image and the stripe image have identical widths. The distance between these two markers has to be precisely known. Section Calibration Point Detection describes how to detect the calibration points in the image.

The calibration of the direction of movement represents only a scaling factor. From the equation that describes the calibration polynomial, we derive the following for the corresponding scan direction:

Scan direction in X:     y' = b9 * y
Scan direction in Y:     x' = a8 * x

Calibration Point Detection

For the detection of the calibration points a blob search is conducted. If more than two blobs are found, the outermost blobs (most proximate to the edges of the image) are used. The following code snippet, shows how to detect the calibration points with CVB:

// load image with two calibration points
auto imgfile = ...
auto image = Cvb::Image::Load(imgfile);
// define AOI
auto areaPoints = Cvb::Area2D(Cvb::Rect<double>(0, 0, image->Width() - 1, image->Height() - 1)); // left, top, right, bottom
// extract
auto pointSize = Cvb::ValueRange<double>(..., ...);
image->Plane(0), areaPoints, Cvb::Foundation::CalibrationPatternContrast::WhiteOnBlack, 80,
pointSize, scanDirection);
auto p1 = calPoints.first;
auto p2 = calPoints.second;
std::cout << "extracted top point(x,y) : " << p1.X() << "," << p1.Y() << "\n";
std::cout << "extracted bottom point(x,y): " << p2.X() << "," << p2.Y() << "\n";
std::pair< Point2D< double >, Point2D< double > > CalculateTwoPointsForCalibrationOfMovement(const ImagePlane &imagePlane, const Area2D &aoi, Foundation::Transform2D::CalibrationPatternContrast contrast, int minContrast, const ValueRange< double > &pointSizeRange, const ScanDirection &scanDirection)
// load image with two calibration points
var image = Image.FromFile(...);
// define AOI
var areaPoints = new Area2D(new Rect(0, 0, image.Width - 1, image.Height - 1));
// extract
var scanDirection = ScanDirection.X;
var pointSize = new ValueRange<double>(..., ...);
image.Planes[0], areaPoints, CalibrationPatternContrast.WhiteOnBlack, 80,
pointSize, scanDirection);
var p1 = calibrationPoints[0];
var p2 = calibrationPoints[1];
Console.Write("extracted top point(x,y) : " + p1.X + "," + p1.Y + "\n");
Console.Write("extracted bottom point(x,y): " + p2.X + "," + p2.Y + "\n");
static Point2Dd[] CalculateTwoPointsForCalibrationOfMovement(ImagePlane imagePlane, Area2D aoi, CalibrationPatternContrast contrast, int minContrast, ValueRange< double > pointSizeRange, ScanDirection scanDirection)
# load image with calibration points
imgfile = ...
image = cvb.Image.load(imgfile)
# define AOI.
area_points = cvb.Area2D.create(cvb.Rect.create(0, 0, image.width-1, image.height-1)) # left, top, right, bottom
# detect calibration points
scan_direction = cvb.foundation.ScanDirection.X
point_size = cvb.NumberRange(...,...) # minimum and maximum point size in [number of pixels]
image.planes[0],
area_points,
cvb.foundation.CalibrationPatternContrast.BlackOnWhite, 80,
point_size,
scan_direction
)
P1 = cal_points[0]
P2 = cal_points[1]
print(f"extracted top or left point(x,y) : {P1.x},{P1.y}")
print(f"extracted bottom or right point(x,y): {P2.x},{P2.y}")
cvb.Area2D create()
Tuple[cvb.Point2D, cvb.Point2D] calculate_two_points_for_calibration_of_movement(cvb.ImagePlane image_plane, cvb.Area2D aoi, int contrast, int min_contrast, cvb.NumberRange point_size_range, int scan_direction)

After loading the image with the calibration points, the following input parameters have to be set:

  • Area of interest containing the calibration points: In this example the whole image is used. If the image includes disturbing background elements, consider narrowing the AOI to focus solely on the area containing the calibration points.
  • The type of the calibration pattern, which can be "black markers on white background" or vice versa.
  • The minimum gray value contrast between the object and the background of the calibration points. The optimal value depends on the quality of the image taken. A good value could be 80.
  • The minimum and maximum size of the markers in the image in [number of pixels].
  • The scan direction must be set correctly. It can be along the x- or the y-axis as depicted in the figures above.

Special Case: Known Encoder Step

If the encoder step of your setup is precisely known (e.g. in [mm/scanline]), you do not need to acquire an image with calibration points. You may manually define fictive calibration points and the corresponding reference distance. Be mindful to consistently utilize the same units (in this example [mm]).

double encoderStep = ... // mm/scanline
auto p1 = Cvb::Point2D<double>(0, 0); // first point
auto p2 = Cvb::Point2D<double>(0, 1); // second point
auto refDist = encoderStep; // reference distance between calibration points in [mm]
double encoderStep = ... // mm/scanline
var p1 = new Point2Dd(0, 0); // first point
var p2 = new Point2Dd(0, 1); // second point
var refDist = encoderStep; // reference distance between calibration points in[mm]
encoder_step = ... # in mm/scanline
P1 = cvb.Point2D(0,0) # first point
P2 = cvb.Point2D(0,1) # second point
ref_dist = encoder_step # reference distance between calibration points in [mm]

Calibration Coefficients Estimation

After you successfully created an object containing the result of the stripe detection and calculated the position of the calibration points in the image, you can start the calibration:

// distance between calibration points in [mm]
double refDist = ...
// with of stripes in [mm]
double refWidth = ...
// configure calibration
auto scanDirection = ...
configuration.SetScanDirection(scanDirection);
// estimate calibration
p1, p2, refDist, detectedEdges, refWidth, configuration);
std::cout << "--- Results linescan calibration: ---\n";
std::cout << "mean error and stdev: " << linescanCalibrator.MeanError() << " mm / " << linescanCalibrator.StandardDeviation() << " mm\n";
std::cout << "pixel_size after calibration: " << linescanCalibrator.PixelSize() << "\n";
// save calibration
linescanCalibrator->Transformation()->Save("linescan_calibrator.nlt");
std::unique_ptr< LineScanCalibrator > CreateLineScanCalibration(const Point2D< double > &calibrationPoint1, const Point2D< double > &calibrationPoint2, double referenceDistanceCalibrationPoints, const EdgeDetectionResult &edgeDetectionResult, double referenceWidthStripes, const LineScanCalibrationConfiguration &configuration)
// distance between calibration points in[mm]
double refDist = ...
// with of stripes in[mm]
double refWidth = ...
// configure calibration
var scanDirection = ...
var configuration = new LineScanCalibrationConfiguration();
configuration.ScanDirection = scanDirection;
// estimate calibration
p1, p2, refDist, detectedEdges, refWidth, configuration);
Console.Write("--- Results linescan calibration: ---\n");
Console.Write("mean error and stdev: " + linescanCalibrator.MeanError + " mm / " + linescanCalibrator.StandardDeviation + " mm\n");
Console.Write("pixel_size after calibration: " + linescanCalibrator.PixelSize + "\n");
// save calibration
linescanCalibrator.Transformation.Save("linescan_calibrator.nlt");
static LineScanCalibrator CreateLineScanCalibration(Point2Dd calibrationPoint1, Point2Dd calibrationPoint2, double referenceDistanceCalibrationPoints, EdgeDetectionResult edgeDetectionResult, double referenceWidthStripes, LineScanCalibrationConfiguration configuration)
# reference distance between calibration points in [mm]
ref_dist = ...
# reference width of stripes in [mm]
ref_width = ...
# configure calibration
configuration.scan_direction = ...
# estimate calibration
P1, P2, ref_dist, detected_edges, ref_width, configuration)
print("--- Calibration results:---")
print(f"mean error and stdev: {linescan_calibrator.mean_error} mm / {linescan_calibrator.standard_deviation} mm")
print(f"pixel_size after calibration: {linescan_calibrator.pixel_size}")
print(f"coeff X: {linescan_calibrator.transformation.coefficients_x}")
print(f"coeff Y: {linescan_calibrator.transformation.coefficients_y}")
# save calibration
linescan_calibrator.transformation.save("linescan_calibrator.nlt")
cvb.foundation.LineScanCalibrator create_line_scan_calibration(cvb.Point2D calibration_point_1, cvb.Point2D calibration_point_2, float reference_distance_calibration_points, cvb.foundation.EdgeDetectionResult edge_detection_result, float reference_width_stripes, cvb.foundation.LineScanCalibrationConfiguration configuration)

If you additionally like to fix the pixel size of the transformed image, you can configure that via the linescan calibration object. You have to set the "predefined pixel size mode" to "use" and specify the desired value for the "pixel size".

The reference distance between calibration points and the width of stripes have to be precisely known in metric units. They must be provided in the same units, as well as the pixel size (if specified). Again the scan direction must be set as depicted in the figures above. The resulting linscan calibrator object includes some error statistics, the pixel size after the calibration and the coefficients in X and Y.

Calibrated Image Creation

As the calibrated image should only encompass areas with values, suitable target dimensions have to be calculated. To determine the new width, height and offset for the calibrated image, you should transform the original aoi (defined by the width and height of the original/uncalibrated image). With the following code the target dimensions and offset can be calculated:

auto image = ... // original image
// original image dimensions (AOI)
auto aoiOrg = Cvb::Rect<double>(0, 0, static_cast<double>(image->Width() - 1), static_cast<double>(image->Height() - 1));
// transform AOI (double)
auto aoiD = linescanCalibrator.Transformation()->Transform(aoiOrg);
// round AOI (int)
auto aoi = Cvb::Rect<int>(
Cvb::Point2D<int>(static_cast<int>(round(aoiD.Location().X())), static_cast<int>(round(aoiD.Location().Y()))),
Cvb::Size2D<int>(static_cast<int>(round(aoiD.Width())), static_cast<int>(round(aoiD.Height()))));
std::cout << "--- Dimensions of calibrated image:---\n";
std::cout << "pixel_size after calibration: " << linescanCalibrator.PixelSize() << "\n";
std::cout << "w x h of original image : " << image->Width() << " x " << image->Height() << "\n";
std::cout << "w x h of calibrated image : " << aoi.Width() << " x " << aoi.Height() << "\n";
std::cout << "offset x,y : " << aoi.Left() << "," << aoi.Top() << "\n\n";
var image = ... // original image
// original image dimensions (AOI)
var aoiOrg = new RectD(0, 0, image.Width - 1, (image.Height - 1));
// transform AOI (double)
var aoiD = linescanCalibrator.Transformation.Transform(aoiOrg);
// round AOI (int)
var aoi = new Rect(
new Point2D((int) Math.Round(aoiD.Location.X), (int)Math.Round(aoiD.Location.Y)),
new Size2D((int)Math.Round(aoiD.Width), (int) Math.Round(aoiD.Height)));
Console.Write("--- Dimensions of calibrated image:---\n");
Console.Write("pixel_size after calibration: " + linescanCalibrator.PixelSize + "\n");
Console.Write("w x h of original image : " + image.Width + " x " + image.Height + "\n");
Console.Write("w x h of calibrated image : " + aoi.Width + " x " + aoi.Height + "\n");
Console.Write("offset x,y : " + aoi.Location.X + "," + aoi.Location.Y + "\n");
image = ... # original image
# original image dimensions (AOI)
aoi_org = cvb.Rect(0, 0, image.width-1, image.height-1) # left, top, right, bottom
# transform AOI
aoi = linescan_calibrator.transformation.transform_rect(aoi_org)
# target dimensions (rounded to integer values)
target_size = cvb.Size2D(round(aoi.width), round(aoi.height))
target_offset = cvb.Point2D(round(aoi.location.x),round(aoi.location.y))
print("--- Dimensions of calibrated image:---")
print(f"pixel_size after calibration: {linescan_calibrator.pixel_size}")
print(f" w x h: {width} x {height}")
print(f"offset (x,y): {target_offset.x},{target_offset.y}")

Finally the image can be simply transformed as follows:

auto calibratedImage = linescanCalibrator->Transformation()->Transform(*image, aoi.Size(), aoi.Location());
calibratedImage->Save("calibrated_image.bmp");
var calibratedImage = linescanCalibrator.Transformation.Transform(image, aoi.Size, aoi.Location);
calibratedImage.Save("calibrated_image.bmp");
image = ... # original image
calibratedImage = linescan_calibrator.transformation.transform_image(image, target_size, target_offset)
calibratedImage.save("calibrated_image.bmp")