Georeferencing

Introduction

In the early days of geoinformation science, spatially referenced data usually originated within national boundaries, i.e. these data were derived from printed maps published by national mapping organizations. Nowadays, users of geoinformation are combining spatial data from a given country with global spatial data sets, reconciling spatial data from published maps with coordinates established by satellite positioning techniques, and integrating their spatial data with that from neighbouring countries.To perform these kinds of tasks successfully, we need to understand basic spatial referencing concepts.

Explanation

The simplest way to link image coordinates to map coordinates is to use a transformation formula. A geometric transformation is a function that relates the coordinates of two systems. A transformation relating (x, y) to (i, j) is commonly defined by linear equations, such as:  x = 3 + 5i, and y = −2 + 2.5j.

Using the above transformation, for example, the image position (i = 3, j = 4) corresponds to map coordinates (x = 18, y = 8). Once the transformation parameters have been determined, the map coordinates for each pixel can be calculated. This implies that we can superimpose data that are given in the map coordinate system on the image vector, or that we can store features by map coordinates when applying on-screen digitizing. Note that the image in the case of georeferencing remains stored in the original (i, j) raster structure and that its geometry is not altered. As we will see in Geocoding, transformations can also be used to change the actual shape of an image and thus make it geometrically equivalent to the map.

Examples

The effect of georeferencing is illustrated in the Figure below along with geocoding.  

Figure: Original (top), georeferenced (middle) and geocoded (bottom) image of a part of Flevoland.

How to

The process of georeferencing involves two steps: (1) selection of the appropriate type of transformation, and (2) determination of the transformation parameters. The type of transformation depends mainly on the sensor–platform system used. For aerial photographs (of flat terrain) what is known as “projective transformation” models well the effect of pitch and roll. Polynomial transformation, which enables 1st, 2nd to nth order transformations, is a more general type of transformation. In many situations a 1st order transformation is adequate. Such transformation relates map coordinates (x, y) with image coordinates (i, j) as follows:

x%3Da%2Bbi%2Bcj
Equation 1

 

y%3Dd%2Bei%2Bfj
Equation 2


Equations 1 and 2 require that six parameters (a to f) be determined. The transformation parameters can be determined by means of ground control points (GCPs). GCPs are points that can be clearly identified in the image and on the target map. Once a sufficient number of GCPs have been specified, software is used to determine the parameters a to f of the Equations 1 and 2 and quality indications.

To solve the 1st order polynomial equations, only three GCPs are required; nevertheless, you should use more points than the strict minimum. Using merely the minimum number of points for solving the system of equations would obviously lead to a wrong transformation if you made an error in one of the measurements, whereas including more points for calculating the transformation parameters enables software to also compute the error of the transformation. Table 1 gives an example of the input and output of a georeferencing computation in which five GCPs have been used. Each GCP is listed with its image coordinates (i, j) and its map coordinates (x, y).

Table 1: A set of five ground control points, which are used to determine a 1st order transformation. xc and ycare calculated using the transformation, dx and dy are the residual errors.
GCP i j x y xc yc dx dy
1 254 68 958 155 958.552 154.935 0.552 -0.065
2 149 22 936 151 934.576 150.401 -1.424 -0.599
3 40 132 916 176 917.732 177.087 1.732 1.087
4 26 269 923 206 921.835 204.966 -1.165 -1.034
5 193 228 954 189 954.146 189.459 0.146 0.459

Software performs a “least-squares adjustment” to determine the transformation parameters. The least squares adjustment ensures an overall best fit of the GCPs. We then use the computed parameter values to calculate coordinates (xc, yc) for any image point (pixel) of interest:

x_c%20%3D%20902.76%20%2B%200.206%20i%20%2B%200.051%20j%2C%5C
Equation 3

and

y_c%20%3D%20152.579%20-%200.044%20i%20%2B%200.199%20j.%5C
Equation 4

For example, for the pixel corresponding to GCP 1 (i = 254, j = 68) we can calculate the transformed image coordinates xc and yc as 958.552 and 154.935, respectively. These values deviate slightly from the input map coordinates (as measured on the map). Discrepancies between measured and transformed coordinates of GCPs are called residual errors (residuals for short). The residuals are listed in the table as dx and dy. Their magnitude is an indicator of the quality of the transformation. Residual errors can be used to analyse whether all GCPs have been correctly determined.

GCP Accuracy

The overall accuracy of a transformation is either stated in the accuracy reportusually provided by software in terms of variances or as Root Mean Square Error (RMSE), which calculates a mean value from the residuals (at check points). The RMSE in the x-direction, mx, is calculated using the following equation:

m_x%20%3D%20%5Csqrt%7B%5Cfrac%20%7B1%7D%20%7Bn%7D%20%5Csum_%7Bi%3D1%7D%5En%20%5Cdelta%7Bx_i%7D%5E2%7D.
Equation 5

For the y-direction, a similar equation can be used to calculate my. The overall error, mp, is calculated by:

%20m_%7Bp%7D%20%3D%20%5Csqrt%7B%7Bm_x%7D%5E2%20%2B%20%7Bm_y%7D%5E2%7D.
Equation 6

For the example data set given in Table 1, the residuals mx, my and mp are 1.159, 0.752 and 1.381, respectively. The RMSE is a convenient measure of overall accuracy, but it does not tell us which parts of the image are accurately transformed and which parts are not. Note also that the RMSE is only valid for the area that is bounded by the GCPs. In the selection of GCPs, therefore, points should be well distributed and include locations near the edges of the image.

Learning outcomes

Prior knowledge

Outgoing relations

Incoming relations

Learning paths