Talk:Geometric corrections

From AWF-Wiki
(Difference between revisions)
Jump to: navigation, search
(Blanked the page)
 
(3 intermediate revisions by one user not shown)
Line 1: Line 1:
{{author|name=[[User:Pmagdon|Paul Magdon]]}}{{Rscontent}}
 
  
= Background =
 
 
All remote sensing images of the Earth surface contain a significant amount of distortions. The distortions are inherent in the acquisition geometry which is determined by the technical design of the observing unit (e.g sensor, platform) and of the observed object (e.g. Earth surface) which can be summarize as sensor geometry. A spatial analysis of the images which includes the assessment of areas, distances, locations, angles and combination with other spatial information by integrating the images into a geographic information system (GIS) is not ad hoc possible. Consequently, for a spatial image analysis a transformation of the the sensor geometry into a meaningfully geometry with known geometric properties which is usually a form of a [[map projection]] is required. This transformation from sensor geometry to a map projection is referred to as geometric correction and is based on sensor models that reflect the nature and magnitude of distortions caused by the various sources such as Earth rotation, platform movement, and sensor errors.
 
 
= Image Coordinate Systems =
 
Digital images store the information taken by a sensor into a format which can be read and analyzed by computers. The information is stored in an regular array of pixels which can be mathematically expressed as an <math>N\times M</math> matrix with <math>x</math> columns and <math>y</math> rows (see Fig.1). For each pixel one digital number (DN) is saved in the matrix. The image coordinate system is then defined by the origin <math>x=0, y=0</math> and the spacing of the pixels. The row indices start at the origin <math>(0,0)</math> and increase downward, while the column indices increases to the right. Depending on the image format used the origin can be located either in the center of the upper left pixel or at the upper left edge of that pixel. For multi-spectral images a third index <math>b</math> is introduced to represent the number of bands. Depending on the image format this starts either with <math>0</math> or <math>1</math>.
 
[[File:ImageCoordinateSystem.png|right|thumb|350px|Figure 1: Image Coordinate System]]
 
 
 
= Acquisition Geometry of Satellites =
 
Different from geo-stationary orbits where the satellites stay over the same location on Earth permanently (e.g. TV-broadcasting satellites), earth observation satellites should cover large areas of the Earth. Therefore, many of the multispectral satellite systems currently used for earth observation (e.g. Landsat, Rapideye) operate in a polar sun-synchronous orbit as shown in Fig. 2. In a polar orbit the satellites pass above or nearly above both pols of the Earth on each revolution. Due to the rotation of the earth the longitude at which the satellites cross over the equator shifts with each orbit, so that the full earth surface can be covered. Additionally many of the optical earth observing satellites operate in sun-synchronous orbit, meaning that each successive orbital occurs at the same local time of day. Under this condition the illumination angle of the sun is kept constant.
 
[[File:satellite_orbit.png|right|thumb|350px|Figure 2: Satellite Acquisition Pattern]]
 
 
Depending on the purpose, satellite systems are designed to operate in various acquisition modes. Two of the most common modes for optical sensors are along-track and cross-track acquisition.
 
 
*Cross-track scanners are based on one detector unit. The spatial coverage is achieved by using a rotating mirror which scans the surface perpendicular to the flight direction. Thus the angular field of view is determined by the deflection of the mirror on the sensor altitude. Well known system using Cross-Track scanners are the [http://en.wikipedia.org/wiki/Landsat_program Landsat] family (MSS, TM, ETM+), [http://de.wikipedia.org/wiki/Moderate-resolution_Imaging_Spectroradiometer MODIS] and [http://en.wikipedia.org/wiki/AVHRR NOAA/AVHRR].
 
 
*Due to the rapid technical development of the detector units (CCD=Charge Couple Device) it became possible to combine many detectors in an array which resulted in a new generation of sensors called push-broom sensors. These sensors scan the surface parallel to the flight direction using a linear array of CCD&rsquo;s aligned across-track. The image is build up by the movement of the platform. Using this technique, theoretically endless acquisition lines can be captured, thought data storage still limits the length of lines. Prominent along-track scanners are [http://en.wikipedia.org/wiki/RapidEye RapidEye], [http://en.wikipedia.org/wiki/SPOT_(satellites) SPOT], [http://en.wikipedia.org/wiki/Advanced_Spaceborne_Thermal_Emission_and_Reflection_Radiometer ASTER], [http://en.wikipedia.org/wiki/Ikonos IKONOS],[http://en.wikipedia.org/wiki/Quickbird Quickbird] and [http://en.wikipedia.org/wiki/Geoeye Geoeye].
 
 
The acquisition geometry of the satellite systems can be described by the position of the sensor (x,y,z) the rotation parameters of the sensor (<math>\phi,\omega,\kappa</math>) and the surface coordinates given in a project map system (X,Y,Z) as depicted in Fig.3.
 
 
[[File:acquisitionGeometry.png|right|thumb|250px|Figure 3: Satellite Acquisition Geometry]]
 
 
= Sources of Image Distortion =
 
 
Depending on the acquisition geometry and mode different distortion are inherent in the satellite raw data. Toutin (2003) categorized them as being either related to the observer or to the observed object. On the observer side the geometry is influences by the platform position (x,y,z) , the earth surface location and the sensor orientation (<math>\phi,\omega,\kappa</math>). A full nadir view is given when all rotation parameters are zero which minimize the images distortions. On the observed side distortions are caused by the earth rotation and the relief displacement due to topography effects. Additionally the Earth atmosphere can cause distortions by refraction of the electromagnetic waves. Table 1 summarizes the sources of image distortion.
 
 
<table style="border-collapse: collapse;">
 
Table 1. Sources of image distortion
 
<tr>
 
    <th>Category</th>
 
    <th>Subcategory</th>
 
    <th>Description</th>
 
    <th>Error Type</th>
 
</tr>
 
 
<tr>
 
    <td rowspan="7"> Observer</td>
 
    <td rowspan="2"> Platform</td>
 
    <td>Variations in platform movement</td>
 
    <td>Systematic </td>
 
</tr>
 
 
<tr>
 
    <td> Variations in platform attitude</td>
 
    <td>Systematic </td>
 
</tr>
 
 
<tr>
 
    <td rowspan="3"> Sensor </td>
 
    <td> Variations in sensor mechanics</td>
 
    <td>Systematic </td>
 
</tr>
 
 
<tr>
 
    <td> View/look angles </td>
 
    <td>Systematic </td>
 
</tr>
 
 
<tr>
 
    <td> Panorama effect with field of view </td>
 
    <td>Systematic </td>
 
</tr>
 
 
<tr>
 
    <td rowspan="2"> Onboard Measuring Instruments </td>
 
    <td> Time-variations or drift </td>
 
    <td> Random </td>
 
</tr>
 
<tr>
 
    <td> Clock-synchrony </td>
 
    <td> Random </td>
 
</tr>
 
 
<tr>
 
    <td rowspan="3"> Observed</td>
 
    <td  style="border-top: 1px solid #000;"> Earth</td>
 
    <td  style="border-top: 1px solid #000;"> Curvature, rotation, topography</td>
 
    <td  style="border-top: 1px solid #000;"> Systematic </td>
 
</tr>
 
 
<tr>
 
    <td> Atmosphere</td>
 
    <td> Refraction</td>
 
    <td> Random </td>
 
</tr>
 
 
<tr>
 
    <td> Map </td>
 
    <td> Geoid to ellipsoid, Ellipsoid to map</td>
 
    <td> Systematic </td>
 
</tr>
 
</table>
 
 
= Geometric Corrections =
 
As described above there are various sources of distortion in raw satellite imagery which depend on the platform, sensor and the surface to be imaged. To allow an spatial analysis of the images (e.g. the transformation of image coordinates to map coordinates) these distortions need to be corrected. This correction is done by describing the impacts of the distortions using sensor models. The correction is done in five step.
 
 
== Acquisition of ground control points (GCPs)==
 
 
GCP's are either two (X,Y) or three (X,Y,Z) dimensional points for which both, the image coordinates (x,y,z) and the map coordinates (X,Y,Z) are known
 
 
*ideally as small as a single pixel
 
*high contrast to the surroundings
 
*constant location over time
 
 
Typical examples of GCP's are intersections of roads, distinctive water bodies, edges of land
 
cover parcels and, corners of buildings.
 
 
== Building a Sensor Model ==
 
 
Sensor models formulate the relation between the images pixels coordinates (x, y) and map coordinates (X, Y ). = vector of parameters describing the image acquisition geometry which can be obtained by: -&gt; A physical sensor model -&gt; A empirical sensor model If no model is available an approximation of the sensor model can be calculated using reference points (GCPs) Sensor models formulate the relation between the images pixels coordinates (<math>c,r</math>) and map coordinates (<math>X,Y</math>). For the models it can be distinguished between ''image to ground''- forward models and ''ground to image''- inverse models. They can be expressed as: <math>X=f_x(c,r,h,\vec{\Theta})\\
 
  Y=f_y(c,r,h,\vec{\Theta})
 
</math>
 
 
In this equation <math>\Theta</math> represents a vector of parameters describing the image acquisition geometry and <math>h</math> gives the height typically derived from an elevation model.<br />
 
 
 
Basically two types of sensor models can be differentiated: ''mechanistic'' and ''empirical'' models. The first are rigorous, complex models where the parameters have a physical meaning (e.g distances, angles)which reconstruct the image geometry. This approach is used for example for SAR orthorectification based on the range-doppler model. The later approach empirically approximate the physical conditions. This is often done by expressing the parameters as polynomials or rations of polynomials the so-called ratio polynomial coefficients '''RPC''' which are sometimes also called ''replacement models''. As this models are approximation they are less accurate then the physical models. But they can be standardized and thus are easier to implement in remote sensing software.
 
 
=== Rigorous Sensor Models ===
 
 
The physical model should model all distortions of platform, sensor, Earth and cartographic projection based on measurable physical units. (degree, time, distance). This is typically achieved by using collinearity functions for optical data and the Rang-Doppler equation for RADAR images. Rigorous sensor models provide information about the sensors orbital movements. They describe the relationship between the image points and homologous ground points, with parameters which have understandable physical meanings using a collinear equation.
 
 
'''Adavantages''':
 
 
* Useful when it is not possible to identify control points (e.g. to coarse resolution, little contrast,high cloud cover)
 
* Requires no human interaction
 
 
'''Disadavantages''':
 
 
* Can only model systematic distortions
 
* Require complicate platform movement and position models (e.g Ikonos Sensor Model is XXX padges)
 
 
=== Non-Rigorous Models ===
 
 
Empirical approaches are used when no physical model or no a priori information for parametrization (e.g. about sensor, platform , Earth) is available. This empirical approaches are either based on polynomial models or rational functions. Empirical approaches are used when: no physical model is available no a priori information for parameterization (e.g. about sensor, platform , Earth) is available. {Limitations and Drawbacks of sensor models}There are two critical elements in orthorectification based on sensor modeling. First, the validity of the sensor models hast to be questioned. But often these models are provided by the sensor operators and are thus generally trustworthy. Second, the accuracy of the available elevation models is critical to the modeling as accurate models can only be based on high quality elevation information.
 
 
== Coordinate Transformation ==
 
== Resampling of Pixel Values ==
 

Latest revision as of 12:49, 11 November 2013

Personal tools
Namespaces

Variants
Actions
Navigation
Development
Toolbox
Print/export