FUNDAMENTALS OF REMOTE
SENSING AND GIS
Shunji Murai Professor and Doctor of Engineering
Institute of Industrial Science
University of Tokyo, Japan
Chair Professor, STAR Program
Asian Institute of Technology, Thailand
1
Preface
Geographic Information System (GIS) has undergone a sort of boom all around the world
in this decade as personal computers and engineering workstations have become
available at reasonable prices. There were and are so many applications of GIS on
various levels of central governments, local governments, utility service corporations,
distribution service companies, car navigation systems, marketing strategies etc., but
unfortunately not all them are successful. For successful GIS one of the keys is education
and training, particularly with well organized teaching materials.
When I was teaching at the Asian Institute of Technology (AIT), Bangkok, Thailand for
three years between 1992 and 1995, and also when I organized an international
symposium on AM/FM GIS ASIA'95 in Bangkok, Thailand, August 1995, I was strongly
requested by many people in the developing countries in Asia to publish a GIS text book,
which is easily understandable in not only theory and principle but also in planning and
application for a successful GIS. If you look at the exiting text books, most of them are
not very much unified because some are collection of articles written by multiple authors,
some are too thick and too expensive for educational purposes, some are too conceptual
and theoretical background. Thus I have attempted to write an easily understood text with
short explanation in not more than a page for each item on the left together with another
page of only figures, tables and/or pictures on the right page, in organized manner.
In 1996 and 1997, I published GIS Work Book- Fundamental Course and Technical
Course respectively with bi-lingal of English and Japanese. As some readers request me
to publish only English version, I reedited the two volumes into a book with only English
version.
I believe that this text book with its two parts ; "fundamental course" and "technical course"
would be useful and helpful to not only students, trainees, engineers, salesmen but also
to top managers or decision makers.
I would like to thank Mr. Minoru Tsuzura, Japan Association of Surveyors for his
administrative support to make this English version possible.
August, 1998
Tokyo, Japan
2
CONTENTS
Chapter 1 Fundamentals of Remote Sensing
1.1 Concept of remote sensing
1.2 Characteristics of electro-magnetic radiation
1.3 Interactions between matter and electro-magnetic radiation
1.4 Wavelength regions of electro-magnetic radiation
1.5 Types of remote sensing with respect to wavelength regions
1.6 Definition of radiometry
1.7 Black body radiation
1.8 Reflectance
1.9 Spectral reflectance of land covers
1.10 Spectral characteristics of solar radiation
1.11 Transmittance of the atmosphere
1.12 Radioactive transfer equation
Chapter 2 Sensors
2.1 Types of sensors
2.2 Characteristics of optical sensors
2.3 Resolving power
2.4 Dispersing element
2.5 Spectroscopic filter
2.6 Spectrometer
2.7 Characteristics of optical detectors
2.8 Cameras for remote sensing
2.9 Film for remote sensing
2.10 Optical mechanical scanner
2.11 Pushbroom scanner
2.12 Imaging spectrometer
3
2.13 Atmospheric sensors
2.14 Sonar
2.15 Laser radar
Chapter 3 Microwave Remote Sensing
3.1 Principles of microwave remote sensing
3.2 Attenuation of microwave
3.3 Microwave radiation
3.4 Surface scattering
3.5 Volume scattering
3.6 Types of Antenna
3.7 Characteristics of Antenna
Chapter 4 Microwave Sensors
4.1 Types of microwave sensor
4.2 Real aperture radar
4.3 Synthetic aperture radar
4.4 Geometry of radar imagery
4.5 Image reconstruction of SAR
4.6 Characteristics of radar image
4.7 Radar images of terrains
4.8 Microwave radiometer
4.9 Microwave scatterometer
4.10 Microwave altimeter
4.11 Measurement of sea wind
4.12 Wave measurement by radar
Chapter 5 Platforms
5.1 Types of platform
5.2 Atmospheric condition and altitude
5.3 Attitude of platform
5.4 Attitude sensors
4
5.5 Orbital elements of satellite
5.6 Orbit of satellite
5.7 Satellite positioning systems
5.8 Remote sensing satellites
5.9 Landsat
5.10 SPOT
5.11 NOAA
5.12 Geostationary meteorological satellites
5.13 Polar orbit platform
Chapter 6 Data used in Remote Sensing
6.1 Digital data
6.2 Geometric characteristics of image data
6.3 Radiometric characteristics of image data
6.4 Format of remote sensing image data
6.5 Auxiliary data
6.6 Calibration and validation
6.7 Ground data
6.8 Ground positioning data
6.9 Map data
6.10 Digital terrain data
6.11 Media for data recording,storage and distribution
6.12 Satellite data transmission and reception
6.13 Retrieval of remote sensing data
Chapter 7 Image Interpretation
7.1 Information extraction in remote sensing
7.2 Image interpretation
7.3 Stereoscopy
7.4 Interpretation elements
5
7.5 Interpretation keys
7.6 Generation of thematic maps
Chapter 8 Image Processing Systems
8.1 Image processing in remote sensing
8.2 Image processing systems
8.3 Image input systems
8.4 Image display systems
8.5 Hard copy systems
8.6 Storage of image data
Chapter 9 Image Processing - Correction
9.1 Radiometric correction
9.2 Atmospheric correction
9.3 Geometric distortions of the image
9.4 Geometric correction
9.5 Coordinate transformation
9.6 Collinearity equation
9.7 Resampling and interpolation
9.8 Map projection
Chapter 10 Image Processing - Conversion
10.1 Image enhancement and feature extraction
10.2 Gray scale conversion
10.3 Histogram conversion
10.4 Color display of image data
10.5 Color representation -color mixing system
10.6 Color representation -color appearance system
10.7 Operations between images
10.8 Principal component analysis
10.9 Spatial filtering
6
10.10 Texture analysis
10.11 Image correlation
Chapter 11 Image Processing - Classification
11.1 Classification techniques
11.2 Estimation of population statistics
11.3 Clustering
11.4 Parallelpiped classifier
11.5 Decision tree classifier
11.6 Minimum distance classifier
11.7 Maximum likelihood classifier
11.8 Applications of fuzzy set theory
11.9 Classification using an expert system
Chapter 12 Applications of Remote Sensing
12.1 Land cover classification
12.2 Land cover change detection
12.3 Global vegetation map
12.4 Water quality monitoring
12.5 Measurement of sea surface temperature
12.6 Snow survey
12.7 Monitoring of atmospheric constituents
12.8 Lineaments extraction
12.9 Geological interpretation
12.10 Height measurement(DEM generation)
Chapter 13 Geographic Information System (GIS)
13.1 GIS and remote sensing
13.2 Model and data structure
13.3 Data input and editing
13.4 Spatial query
13.5 Spatial analysis
7
13.6 Use of remote sensing data in GIS
13.7 Errors and fuzziness of geographic data and their influences on GIS products
8
Chapter 1 Fundamentals of Remote
Sensing
1.1 Concept of Remote Sensing
Remote Sensing is defined as the science and technology by which the characteristics of
objects of interest can be identified, measured or analyzed the characteristics without direct
contact.
Electro-magnetic radiation which is reflected or emitted from an object is the usual source
of remote sensing data. However any media such as gravity or magnetic fields can be
utilized in remote sensing.
A device to detect the electro-magnetic radiation reflected or emitted from an object is
called a "remote sensor" or "sensor". Cameras or scanners are examples of remote sensors.
A vehicle to carry the sensor is called a "platform". Aircraft or satellites are used as
platforms.
The technical term "remote sensing" was first used in the United States in the 1960's, and
encompassed photogrammetry, photo-interpretation, photo-geology etc. Since Landsat-1,
the first earth observation satellite was launched in 1972, remote sensing has become
widely used.
The characteristics of an object can be determined, using reflected or emitted electromagnetic radiation, from the object. That is, "each object has a unique and different
characteristics of reflection or emission if the type of deject or the environmental condition
is different."Remote sensing is a technology to identify and understand the object or the
environmental condition through the uniqueness of the reflection or emission.
This concept is illustrated in figure 1.1.1 while figure 1.1.2 shows the flow of remote
sensing, where three different objects are measured by a sensor in a limited number of
bands with respect to their, electro-magnetic characteristics after various factors have
9
affected the signal. The remote sensing data will be processed automatically by computer
and/or manually interpreted by humans, and finally utilized in agriculture, land use,
forestry, geology, hydrology, oceanography, meteorology, environment etc.
In this chapter, the principles of electro-magnetic radiation are described in sections1.21.4, the types of remote sensing with respect to the spectral range of the electro-magnetic,
radiation in section 1.5, the definition of radiometry in section 1.6, black body radiation in
section 1.7, electro-magnetic characteristics in sections 1.8 and 1.9, solar radiation in
section 1.10 and atmospheric behavior in sections 1.11 and 1.12.
1.2 Characteristics of Electro-Magnetic Radiation
Electro-magnetic radiation is a carrier of electro-magnetic energy by transmitting the
oscillation of the electro-magnetic field through space or matter. The transmission of
electro-magnetic radiation is derived from the Maxwell equations. Electro-magnetic
radiation has the characteristics of both wave motion and particle motion.
(1) Characteristics as wave motion
Electro-magnetic radiation can be considered as a transverse wave with an electric field
and a magnetic field. A plane wave for an example as shown in Figure 1.2.1 has its electric
field and magnetic field in the perpendicular plane to the transmission direction. The two
fields are located at right angles to each other. The wavelength , frequency
and the velocity have the following relation.
Electro-magnetic radiation is transmitted in a vacuum of free space with the velocity of
light c, ( = 2.998 x 108 m/sec) and in the atmosphere with a reduced but similar velocity to
that in a vacuum. The frequency n is expressed as a unit of hertz (Hz), that is the number
of waves which are transmitted in a second.
(2) Characteristics as particle motion
Electro-magnetic can be treated as a photon or a light quantum. The energy E is expressed
as follow.
E = hµ
10
Where h: Plank's constant
µ: frequency
The photoelectric effect can be explained by considering the electro-magnetic radiation as
composed of particles. Electro-magnetic radiation has four elements of frequency (or
wavelength), transmission direction, amplitude and plane of polarization. The
amplitude is the magnitude of oscillating electric field. The square of the amplitude is
proportional to the energy transmitted by electro-magnetic radiation. The energy radiated
from an object is called radiant energy. A plane including electric field is called a plane of
polarization. When the plane of polarization forms a uniform plane, it is called linear
polarization.
The four elements of electro-magnetic radiation are related to different information content
as shown in Figure 1.2.2. Frequency (or wavelength) corresponds to the color of an object
in the visible region which is given by a unique characteristic curve relating the wavelength
and the radiant energy. In the microwave region, information about objects is obtained
using the Doppler shift effect in frequency that is generated by a relative motion between
an object and a platform. The spatial location and shape of objects are given by the linearity
of the transmission direction, as well as by the amplitude. The plane of polarization is
influenced by the geometric shape of objects in the case of reflection or scattering in the
microwave region. In the case of radar, horizontal polarization and vertical polarization
have different responses on a radar image.
1.3 Interactions between Matter and Electro-magnetic Radiation
All matter reflects, absorbs, penetrates and emits electro-magnetic radiation in a unique
way. For example, the reason why a leaf looks green is that the chlorophyll absorbs blue
and red spectra and reflects the green spectrum (see 1.9). The unique characteristics of
matter are called spectral characteristics (see 1.6). Why does an object have a peculiar
characteristic of reflection, absorption or emission? In order to answer the question, one
has to study the relation between molecular, atomic and electro-magnetic radiation. In this
section, the interaction between hydrogen atom and absorption of electro-magnetic
radiation is explained for simplification.
11
A hydrogen atom has a nucleus and an electron as shown in Figure 1.3.1. The inner state
of an atom depends on the inherent and discrete energy level. The electron's orbit is
determined by the energy level. If electro-magnetic radiation is incident on an atom of H
with a lower energy level (E1), a part of the energy is absorbed, and an electron is induced
by excitation to rise to the energy level (E2) resulting in the upper orbit.
The electro-magnetic energy E is given as follow.
E = hc /
where h: Plank's constant
c: velocity of light
: wavelength
The difference of energy level
E = E2 - E1 = hc / H is absorbed.
In other words, the change of the inner state in an H-atom is only realized when electromagnetic radiation at the peculiar wavelength lH is absorbed in an H-atom. Conversely
electro-magnetic radiation at the wavelength H is radiated from an H-atom when the
energy level changes from E2 to E1.
All matter is composed of atoms and molecules with a particular composition. Therefore,
matter will emit or absorb electro-magnetic radiation at a particular wavelength with
respect to the inner state.
The types of inner state are classified into several classes, such as ionization, excitation,
molecular vibration, molecular rotation etc. as shown in Figure 1.3.2 and Table 1.3.1,
which will radiate the associated electro-magnetic radiation. For example, visible light is
radiated by excitation of valence electrons, while infrared is radiated by molecular
vibration or lattice vibration.
1.4 Wavelength Regions of Electro-magnetic Radiation
Wavelength regions of electro-magnetic radiation have different names ranging from ray,
X ray, ultraviolet (UV), visible light, infrared (IR) to radio wave, in order from the
shorter wavelengths. The shorter the wavelength is, the more the electro-magnetic radiation
is characterized as particle motion with more linearity and directivity. (See 1.2).
12
Table 1.4.1 shows the names and wavelength region of electro-magnetic radiation. One has
to note that classification of infrared and radio radiation may vary according to the
scientific discipline. The table shows an example which is generally used in remote
sensing.
The electro-magnetic radiation regions used in remote sensing are near UV(ultra-violet)
(0.3-0.4 m), visible light(0.4-0.7 m), near shortwave and thermal infrared (0.7-14 m)
and micro wave (1 mm - 1 m).
Figure 1.4.1 shows the spectral bands used in remote sensing. The spectral range of near
IR and short wave infrared is sometimes called the reflective infrared (0.7-3 m)
because the range is more influenced by solar reflection rather than the emission from the
ground surface (see 1.5). In the thermal infrared region, emission from the ground's surface
dominates the radiant energy with little influence from solar reflection (see 1.5 and 1.7).
Visible light corresponds to the spectral colors. They are, in order from the longer
wavelengths in the visible region, the so called rainbow colors; red, orange, yellow, green,
blue, indigo and violet are located with respect to the wavelength.
Short wave infrared has more recently been used for geological classification of rock types.
Thermal infrared is primarily used for temperature measurement (see 1.7), while micro
wave is utilized for radar and micro wave radiometry. A special naming of k band, X band,
C band, L band etc. is given to the micro wave region as shown in Figure 1.4.1.
1.5 Types of Remote Sensing with Respect to Wavelength Regions
Remote sensing is classified into three types with respect to the wavelength regions; (1)
Visible and Reflective Infrared Remote Sensing, (2) Thermal Infrared Remote
Sensing and (3) Microwave Remote Sensing, as shown in Figure 1.5.1.
The energy source used in the visible and reflective infrared remote sensing is the sun.
The sun radiates electro-magnetic energy with a peak wavelength of 0.5 m (see 1.7 and
1.10). Remote sensing data obtained in the visible and reflective infrared regions mainly
depends on the reflectance of objects on the ground surface (see 1.8). Therefore,
information about objects can be obtained from the spectral reflectance. However laser
13
radar is exceptional because it does not use the solar energy but the laser energy of the
sensor.
The source of radiant energy used in thermal infrared remote sensing is the object itself,
because any object with a normal temperature will emit electro-magnetic radiation with a
peak at about 10 m (see 1.7), as illustrated in Figure 1.5.1.
One can compare the difference of spectral radiance between the sun (a) and an object with
normal earth temperature (about 300 K), as shown in Figure 1.5.1. However it should be
noted that the figure neglects atmospheric absorption (see 1.11), for simplification, though
the spectral curve varies with respect to the reflectance, emittance and temperature of the
object.
The curves of (a) and (b) cross at about 3.0 m. Therefore in the wavelength region shorter
than 3.0 m, spectral reflectance is mainly observed, while in the region longer than 3.0 m,
thermal radiation is measured.
In the microwave region, there are two types of micro wave remote sensing, passive
microwave remote sensing and active remote sensing. In passive microwave remote
sensing, the microwave radiation emitted from an object is detected, while the back
scattering coefficient is detected in active micro wave remote sensing. (see 3.4).
Remarks: the two curves (a) and (b) in Figure 1.5.1 show the black body's spectral
radiances of the sun at a temperature of 6,000 K and an object with a temperature of 300
K, without atmospheric absorption.
1.6 Definition of Radiometry
In remote sensing, electro-magnetic energy reflected or emitted from objects is measured.
The measurement is based on either radiometry or photometry, with different technical
terms and physical units.
Radiometry is used for physical measurement of a wide range of radiation from x-ray to
radio wave, while photometry corresponds to the human perception of visible light based
on the human eye's sensitivity as shown in Figure 1.6.1.
14
Figure 1.6.1shows the rdoiometric definitions of radiant energy, radiant fiux, radiant
intensity, irradiance, raiant emittance and radiance.
Table 1.6.1 show the comparision with respect to the techical terms, symbols and units
between radiometry and photometry.
One can add an adjective "Spectral" before the technical terms of radiometry when defined
as per unit of wavelength. For example, one can use spectral radiant flux ( W m ) or
spectral radiance (Wm sr
m ).
Radiant energy is defined as the energy carried by electro- magnetic radiation and
expressed in the unit of joule (J).
Radiant flux is radiant energy transmitted as a radial direction per unit time and expressed
in a unit of watt (W). Radiant intensity is radiant flux radiated from a point source per
unit solid angle in a radiant direction and expressed in the unit of Wsr . Irradiance is
radiant flux incident upon a surface per unit area and expressed in the unit of Wm .
Radiant emittance is radiant flux radiated from a surface per unit area, and expressed in
a unit of Wm . Radiance is radiant intensity per unit projected area in a radial direction
and expressed in the unit of Wm sg .
1.7 Black Body Radiation
An object radiates unique spectral radiant flux depending on the temperature and emissivity
of the object. This radiation is called thermal radiation because it mainly depends on
temperature. Thermal radiation can be expressed in terms of black body theory.
A black body is matter which absorbs all electro-magnetic energy incident upon it and does
not reflect nor transmit any energy. According to Kirchhoff's law the ratio of the radiated
energy from an object in thermal static equilibrium, to the absorbed energy is constant and
only dependent on the wavelength and the temperature T. A black body shows the
maximum radiation as compared with other matter. Therefore a black body is called a
perfect radiator.
15
Black body radiation is defined as thermal radiation of a black body, and can be given by
Plank's law as a function of temperature T and wavelength as shown in Figure 1.7.1 and
Table 1.7.1.
In remote sensing, a correction for emissivity should be made because normal observed
objects are not black bodies. Emissivity can be defined by the following formula-
Emissivity ranges between 0 and 1 depending on the dielectric constant of the object,
surface roughness, temperature, wavelength, look angle etc. Figure 1.7.2 shows the spectral
emissivity and spectral radiant flux for three objects that are a black body, a gray body
and a selective radiator.
The temperature of the black body which radiates the same radiant energy as an observed
object is called the brightness temperature of the object.
Stefan-Boltzmann's law is obtained by integrating the spectral radiance given by Plank's
law, and shows in that the radiant emittance is proportional to the fourth power of absolute
temperature (T ). This makes it very sensitive to temperature measurement and change.
Wien's displacement law is obtained by differentiating the spectral radiance, which shows
that the product of wavelength (corresponding to the maximum peak of spectral radiance)
and temperature, is approximately 3,000 ( m K). This law is useful for determining the
optimum wavelength for temperature measurement of objects with a temperature of T. For
example, about 10 m is the best for measurement of objects with a temperature of 300 K.
1.8 Reflectance
Reflectance is defined as the ratio of incident flux on a sample surface to reflected flux
from the surface as shown in Figure 1.8.1. Reflectance ranges from 0 to 1. Reflectance was
originally defined as a ratio of incident flux of white light to reflected flux in a hemisphere
direction. Equipment to measure reflectance are called spectrometers (see 2.6).
Albedo is defined as the reflectance using the incident light source from the sun.
Reflectance factor is sometime used as the ratio of reflected flux from a sample surface to
16
reflected flux from a perfectly diffuse surface. Reflectance with respect to wavelength is
called spectral reflectance as shown for a vegetation example in Figure 1.8.2. A basic
assumption in remote sensing is that spectral reflectance is unique and different from one
object to an unlike object.
Reflectance with a specified incident and reflected direction of electro-magnetic radiation
or light is called directional reflectance. The two directions of incident and reflection have
can be directional, conical or hemispherical making nine possible combinations.
For example, if incident and reflection are both directional, such reflectance is called
bidirectional reflectance as shown in Figure 1.8.3. The concept of bidirectional reflectance
is used in the design of sensors.
Remarks; A perfectly diffuse surface is defined as a uniformly diffuse surface with a
reflectance of 1, while the uniformly diffused surface, called a Lambertian surface, reflects
a constant radiance regardless of look angle.
The Lambert cosine law which defines a Lambertian surface is as follows:
I ( ) = In .cos
where I(
): luminous intensity at an angle of
from the normal to the surface.
In : luminous intensity at the normal angle
1.9 Spectral Reflectance of Land Covers
Spectral reflectance is assumed to be different with respect to the type of land cover, as
explained in 1.3 and 1.8. This is the principle that in many cases allows the identification
of land covers with remote sensing by observing the spectral reflectance or spectral
radiance from a distance far removed from the surface.
Figure 1.9.1 shows three curves of spectral reflectance for typical land covers; vegetation,
soil and water. As seen in the figure, vegetation has a very high reflectance in the near
infrared region, though there are three low minima due to absorption.
Soil has rather higher values for almost all spectral regions. Water has almost no reflectance
in the infrared region.
17
Figure 1.9.2 shows two detailed curves of leaf reflectance and water absorption.
Chlorophyll, contained in a leaf, has strong absorption at 0.45 m and 0.67 m, and high
reflectance at near infrared (0.7-0.9 m). This results in a small peak at 0.5-0.6 (green color
band), which makes vegetation green to the human observer.
Near infrared is very useful for vegetation surveys and mapping because such a steep
gradient at 0.7-0.9 m is produced only by vegetation.
Because of the water content in a leaf, there are two absorption bands at about 1.5 m and
1.9 m. This is also used for surveying vegetation vigor.
Figure 1.9.3 shows a comparison of spectral reflectance among different species of
vegetation.
Figure 1.9.4 shows various patterns of spectral reflectance with respect to different rock
types in the short wave infrared (1.3-3.0 m). In order to classify such rock types with
different narrow bands of absorption, a multi-band sensor with a narrow wavelength
interval is to be developed. Imaging spectrometers (see 2.12) have been developed for rock
type classification and ocean color mapping.
1.10 Spectral Characteristics of Solar Radiation
The sun is the energy source used to detect reflective energy of ground surfaces in the
visible and near infrared regions.
Sunlight will be absorbed and scattered by ozone, dust, aerosols, etc., during the
transmission from outer space to the earths surface (see 1.11 and 1.12). Therefore, one has
to study the basic characteristics of solar radiation.
The sun is considered as a black body with a temperature of 5,900 K. If the annual average
of solar spectral irradiance is given by FeO( ), then the solar spectral irradiance Fe(l) in
outer space at Julian day D, is given by the following formula.
Fe( ) = FeO( ){1 + cos (2 (D-3)/365)}
where : 0.167 (eccentricity of the Earth orbit) : wavelength
D-3: shift due to January 3 as apogee and July 2 as perigee
The sun constant that is obtained by integrating the spectral irradiance for all wavelength
regions is normally taken as 1.37Wm . Figure 1.10.1 shows four observation records of
18
solar spectral irradiance. The values of the curves correspond to the value at the surface
perpendicular to the normal direction of the sun light. To convert to the spectral irradiance
per m on the Earth surface with a latitude of , multiply the following coefficient by the
observed values in Figure 1.10.1.
= (L0 / L) cos z cosz = sin sin + cos cos cos h
where z : solar zenith angle
: declination
h : hour angle,
L : real distance between the sun and the earth
L0: average distance between the sun and the earth
The incident solar radiation at the earth's surface is very different to that at the top of the
atmosphere due to atmospheric effects, as shown in Figure 1.10.2, which compares the
solar spectral irradiance at the earth's surface to black body irradiance from a surface of
temperature 5900 K.
The solar spectral irradiance at the earth's surface is influenced by the atmospheric
conditions and the zenith angle of the sun. Beside the direct sunlight falling on a surface,
there is another light source called sky radiation, diffuse radiation or skylight, which is
produced by the scattering of the sunlight by atmospheric molecules and aerosols.
The skylight is about 10 percent of the direct sunlight when the sky is clear and the sun's
elevation angle is about 50 degree. The skylight has a peak in its spectral characteristic
curve at a wavelength of 0.45 m
1.11 Transmittance of the Atmosphere
The sunlight's transmission through the atmosphere is affected by absorption and
scattering of atmospheric molecules and aerosols. The reduction of sunlight intensity is
called extinction. The rate of extinction is expressed as extinction coefficient (see 1.12).
19
The optical thickness of the atmosphere corresponds to the integrated value of the
extinction coefficient at each altitude by the atmospheric thickness. The optical thickness
indicates the magnitude of absorption and scattering of the sunlight. The following
elements will influence the transmittance of the atmosphere.
a. Atmospheric molecules(smaller size than wavelength):
carbon dioxygen, ozone, nitrogen gas, and other molecules
b. Aerosols (larger size than wavelength):
water drops such as fog and haze, smog, dust and other particles with a bigger size
Scattering by atmospheric molecules with a smaller size than the wavelength of the sunlight
is called Rayleigh scattering. Raleigh scattering is inversely proportional to the fourth
power of the wavelength.
The contribution of atmospheric molecules to the optical thickness is almost constant
spatially and with time, although it varies somewhat depending on the season and the
latitude.
Scattering by aerosols with larger size than the wavelength of the sunlight is called Mie
scattering. The source of aerosols will be suspended particles such as sea water or dust in
the atmosphere blown from the sea or the ground, urban garbage, industrial smoke,
volcanic ashes etc., which varies to a great extent depending upon the location and the time.
In addition, the optical characteristics and the size distribution also changes with respect to
humidity, temperature and other environmental conditions. This makes it difficult to
measure the effect of aerosol scattering.
Scattering, absorption and transmittance of the atmosphere are different for different
wavelengths. Figure 1.11.1 shows the spectral transmittance of the atmosphere. The low
parts of the curve show the effect of absorption by the molecules described in the figure.
Figure 1.11.2 shows the spectral transmittance, or conversely absorption, with respect to
various atmospheric molecules. The open region with higher transmittance in called "an
atmospheric window".
As the transmittance partially includes the effect of scattering, the contribution of scattering
is larger in the shorter wavelengths. Figure 1.11.3 shows a result of simulation for resultant
20
transmittance multiplied by absorption and scattering which would be produced for a
standard "clean atmospheric model" in the U.S.A. The contribution by scattering is
dominant in the region less than 2mm and proportional according to the shorter
wavelength. The contribution by absorption is not constant but depends on the specific
wavelength.
1.12 Radiative Transfer Equation
Radiative transfer is defined as the process of transmission of the electro-magnetic
radiation through the atmosphere, and the influence of the atmosphere. The atmospheric
effect is classified into multiplicative effects and additive effects as shown in Table 1.12.1.
The multiplicative effect comes from the extinction by which incident energy from the
earth to a sensor will reduce due to the influence of absorption and scattering. The additive
effect comes from the emission produced by thermal radiation from the atmosphere and
atmospheric scattering, which is incident energy on a sensor from sources other than the
object being measured.
Figure 1.12.1 shows a schematic model for the absorption of the electro-magnetic radiation
between an object and a sensor, while Figure 1.12.2 shows a schematic model for the
extinction. Absorption will occur at specific wavelengths (see 1.11) when the electromagnetic energy converts to thermal energy. On the other hand, scattering is remarkable in
the shorter wavelength region when energy conversion does not occur but only the
direction of the path changes.
As shown in Figures 1.12.3 and 1.12.4, additional energy by emission and scattering of the
atmosphere is incident upon a sensor. The thermal radiation of the atmosphere which is
characterized by Plank's law (see 1.7), is uniform in all directions. The emission and
scattering of the atmosphere incident on the sensor, is indirectly input from other energy
sources of scattering than those on the path between a sensor and an object.
The scattering depends on the size of particles and the direction of incident light and
scattering.
Thermal radiation is dominant in the thermal infrared region, while scattering is dominant
in the shorter wavelength region.
21
Generally, as extinction and emission occur at the same time,both effects should be
considered together in the radiative transfer equation as indicated in the formula in Table
1.12.2.
Chapter 2 Sensor
2.1 Types of Sensor
Figure 2.1.1 summarizes the types of sensors now used or being developed in remote
sensing. It is expected that some new types of sensors will be developed in the future.
22
Passive sensors detect the reflected or emitted electro-magnetic radiation from natural
sources, while active sensors detect reflected responses from objects which are irradiated
from artificially generated energy sources, such as radar.. Each is divided further in to nonscanning and scanning systems.
A sensor classified as a combination of passive, non-scanning and non-imaging method
is a type of profile recorder, for example a microwave radiometer. A sensor classified as
passive, non-scanning and imaging method, is a camera, such as an aerial survey camera
or a space camera, for example on board the Russian COSMOS satellite.
Sensors classified as a combination of passive, scanning and imaging are classified further
into image plane scanning sensors, such as TV cameras and solid state scanners, and
object plane scanning sensors, such as multispectral scanners (optical-mechanical
scanner) and scanning microwave radiometers.
An example of an active, non-scanning and non-imaging sensor is a profile recorder such
as a laser spectrometer and laser altimeter. An active, scanning and imaging sensor is a
radar, for example synthetic aperture radar (SAR), which can produce high resolution,
imagery, day or night, even under cloud cover.
The most popular sensors used in remote sensing are the camera, solid state scanner, such
as the CCD (charge coupled device) images, the multi-spectral scanner and in the future
the passive synthetic aperture radar.
Laser sensors have recently begun to be used more frequently for monitoring air pollution
by laser spectrometers and for measurement of distance by laser altimeters.
Figure 2.1.2 shows the most common sensors and their spectral bands.
Those sensors which use lenses in the visible and reflective infrared region, are called
optical sensors
2.2 Characteristics of Optical Sensors Radiation
Optical sensors are characterized specified by spectral, radiometric and geometric
performance. Figure 2.2.1 summarizes the related elements for the three characteristics of
optical sensor. Table 2.2.1 presents the definitions of these elements.
23
The spectral characteristics are spectral band and band width, the central wavelength,
response sensitivity at the edges of band, spectral sensitivity at outer wavelengths and
sensitivity of polarization.
Sensors using film are characterized by the sensitivity of film and the transmittance of the
filter, and nature of the lens. Scanner type sensors are specified by the spectral
characteristics of the detector and the spectral splitter. In addition, chromatic aberration is
an influential factor. The radiometric characteristics of optical sensors are specified by
the change of electro-magnetic radiation which passes through an optical system. They are
radiometry of the sensor, sensitivity in noise equivalent power, dynamic range, signal to
noise ratio (S/N ratio) and other noises, including quantification noise.
The geometric characteristics are specified by those geometric factors such as field of view
(FOV), instantaneous field of news (IFOV), band to band registration, MTF (see 2.3),
geometric distortion and alignment of optical elements.
IFOV is defined as the angle contained by the minimum area that can be detected by a
scanner type sensor. For example in the case of an IFOV of 2.5 milli radians, the detected
area on the ground will be 2.5 meters x 2.5 meters,if the altitude of sensor is 1,000 m above
ground.
2.3 Resolving Power Radiation
Resolving power is an index used to represent the limit of spatial observation. In optics,
the minimum detectable distance between two image points is called resolving limit, and
the reverse is defined as the resolving power.
There are several methods to measure the resolving limit or resolving power. Two such
methods, (1) resolving power by refraction and (2) MTF, are introduced below.
(1) Resolving limit by refraction
Theoretically an object point will be projected as a point on an image plane if the optical
system has no aberration. However, because of diffraction the image of a point will be
a circle with a radius of about one wavelength of light, which in called the Airy
pattern, as shown in Figure 2.3.1. Therefore there exists a limit to resolve the distance
between two points even though there is no aberration.
24
The resolving limit depends on how the minimum distance between two Airy images is
defined. There are two definitions, as follows.
a. Rayleigh's resolving power: the distance between the left Airy peak and the right Airy
peak when it coincides with the zero point of the left peak, that is 1.22u in Figure 2.3.2.
b. Sparrow's resolving limit: the distance between the two peaks when the central gap
fades away, that is 1.08u in Figure 2.3.3 .
(2) MTF (modular transfer function )
The resolving power measured on a resolving test chart by human eyes, depends on
individual ability and the shape or contrast of the chart. On the other hand, MTF has no
such problems because MTF comes from a scientific definition in which the response of
spatial frequency, with respect to the amplitude, considers the optical imaging system as a
spatial frequency filter.
As the spatial frequency is defined as the frequency of a sine wave, MTF shows how much
the ratio of the amplitude decreases before and after an optical imaging system with respect
to the spatial frequency as shown in Figure 2.3.4.
MTF coincides with the power spectrum which is obtained by Fourier transformation of a
point image. Generally speaking, an optical imaging system will give a low pass filter as
shown in Figure 2.3.5.
Modulation (M), contrast (K) and density (D) have the following relations.
= max / min, D = log( max / min), M = ( max - min ) / ( max + min ) = ( - 1) / (
+ 1)
' = 'max / 'min, D' = log( 'max / min), M = ( 'max - 'min ) / ( 'max + 'min ) = ( ' 1) / ( ' + 1)
The resolving power (or spatial frequency) is obtained from the MTF curve with a given
contrast, which can be converted to the modulation.
2.4 Dispersing Element
An array of light arranged by order of wavelength is called a spectrum. Spectroscopy is
defined as the study of the dispersion of light into its spectrum. There are two types of
dispersing elements, the prism and the diffraction grating.
25
Figure 2.4.1 shows the types of dispersing elements. The optical mechanism of prisms and
diffraction gratings are shown in Figure 2.4.2 and Figure 2.4.3 respectively.
(1) Prism
A prism designed for spectroscopy is called a dispersing prism, which is based on the
theory that refractive index is different depending on the wavelength, as shown in Figure
2.4.4. The spectral resolution of a prism is much lower than that of a diffraction grating. If
higher spectral resolution is required, a layer prism should be produced. This can be a
problem, because it is rather difficult to prepare homogeneous material and to keep the
weight low.
(2) Diffraction grating
A diffraction grating is a diffraction element which utilizes the theory that incident light to
a grating is dispersed in multiple different directions depending on the difference of light
path length or phase difference between two neighboring gratings. Multiple spectra are
generated in integer order direction in which multiplication by the wavelength corresponds
to the light path difference as shown in Figure 2.4.5. Most diffraction gratings are reflection
type rather than transparency type. Though the specular diffraction gives the maximum
intensity as 0 order diffraction, it cannot be utilized because 0 order diffraction does not
produce a spectrum. Therefore a reflecting plane is adjusted to have a proper angle to obtain
a strong enough spectrum at a certain order. Such an adjusted grating is called a blazed
grating.
2.5 Spectroscopic Filter
A filter can transmit or reflect a specified range of wavelength. A filter designed for
spectroscopy is called a spectroscopic filter.
Filters are classified into three types - long wave pass filters, short wave pass filters and
band pass filters from the viewpoint of function, as shown in Figure 2.5.1. A cold mirror
which transmits thermal infrared and reflects visible light is a long wave pass filter, while
a hot mirror which reflects thermal infrared and transmits visible light is a short wave filter.
Figure 2.5.2 shows the types of filter from the viewpoint of function.
(1) Absorption filter :
26
a filter which absorbs a specific range of wavelengths, for examples, colored filter
glass and gelatin filter.
(2) Interference filter :
a filter which transmits a specific range of wavelengths by utilizing the interference effect
of a thin film. When light is incident on a thin film, only a specific range of wavelengths
will pass due to the interference by multiple reflection in a thin film as shown in Figure
2.5.3 and 2.5.4. The higher the reflectance of the thin film, the narrower the width of the
spectral band becomes. If two of these films, with different refractive indexes, are
combined, the reflectance becomes very high which results in a narrow spectral band, for
example of the order of several nanometers. In order to obtain a band pass filter which
transmits a single spectral band, a short wave pass filter and long wave pass filter should
be combined. A dichroic mirror ,which is used for three primary color separation, is a kind
of multiple layer interference filter, as shown in Figure 2.5.5 and 2.5.6. It utilizes both
functions of transparency and reflection.
(3) Diffraction grating filter :
a reflective long wave pass filter utilizing the diffraction effect of a grating, which reflects
all light of wavelengths longer than the wavelength determined by the grating interval and
the oblique angle of the incident radiation.
(4) Polarizing interference filter :
a filter with birefringent crystallinity plate between two polarizing plates, which pass a very
narrow spectral band, for example less than 0.1 mm. This utilizes the interference by two
rays of light ; a light following Snell's law and the other not following Snell's law, which
pass a light with a narrow band of wavelength determined by the thickness of the
birefringent crystallinity plate .
2.6 Spectrometer
There are many kinds of spectral measurement devices ,for example, spectroscopes for
human eye observation of the spectrum, spectrometer to record spectral reflectance,
monochro meter to read a single narrow band, spectro photometer for photometry,
27
spectro radiometer for measurement of spectral radiation etc. However, in this section
only optical spectrometers are of interest.
Figure 2.6.1 shows a classification of spectrometers, which are divided mainly into
dispersing spectrometers and interference spectrometers. The former utilizes prisms or
diffraction gratings, while the latter the interference of light. (1) Dispersing spectrometer
A spectrum is obtained at the focal plane after a light ray passes through a slit and
dispersing element as shown in Figure 2.6.2. Figure 2.6.3 and Figure 2.6.4 are typical
dispersing spectrometers ; Littnow spectrometer and Czerny - Turner spectrometer
respectively.
(2) Twin beam interference spectrometer
A distribution of the spectrum is obtained by cosine Fourier transformation of the
interferogram which is produced by the inference between two split rays. Figure 2.6.5
shows Michelson interferometer which utilizes a beam splitter.
(3) Multi-beam interference spectrometer
The interference of light will occur if oblique light is incident on two parallel semitransparent plane mirrors. A different spectrum is obtained depending on incident angle,
interval of the two mirrors and the refraction coefficient.
2.7 Characteristics of Optical Detectors
An element which converts the electro-magnetic energy to an electric signal is called a
detector. There are various types of detectors with respect to the detecting wavelength.
Figure 2.7.1 shows three types of detectors; photo emission type, optical excitation type
and thermal effect type. Photo tube and photo multiplier tubes are the examples of the
photo emission type which has sensitivity in the region from ultra violet to visible light.
Figure 2.7.2 shows the response sensitivity of several photo tubes.
Photodiode, phototransistor, photo conductive detectors and linear array sensors (see
2.11), are examples of optical excitation types, which have sensitivity in the infrared
region. Photo diode detectors utilize electric voltage from the excitation of electrons, while
photo transistor and photo conductive detector utilize electric current. Table 2.7.1 shows
28
the characteristics of these optical detectors with respect to type, temperature, range of
wavelength, peak wavelength, sensitivity in term of D* and response time.
Thermocouple barometers and pyroelectric barometers are examples of the thermal
effect type, which has sensitivity from near infrared to far infrared regions. However the
response is not very high because of the thermal effect. Figure 2.7.3 shows the detectivity
of the pyroelectric barometer.
Detectivity denoted as D* (termed D star) is usually related to the sensitivity, expressed as
NEP (noise equivalent power). D* is used for comparison between different detectors.
NEP is defined as the signal input identical to the noise output. NEP depends on the type
of detector, surface of detector or band of frequency. D* is inversely proportional to NEP,
and is given as follows.
D* : (Ad f ) / NEP
D* : detectivity (cm Hz / W)
NEP : noise equivalent power (W)
Ad : surface area of detector ( cm )
f : band of frequency ( Hz )
2.8 Cameras for Remote Sensing
Aerial survey cameras, multispectral cameras, panoramic cameras etc. are used for
remote sensing.
Aerial survey cameras, sometimes called metric cameras are usually used on board aircraft
or space craft for topographic mapping by taking stereo photographs with overlap. A
typical aerial survey camera is RMK made by Carl Zeiss or RC series made by Leica
company. Figure 2.8.1 shows the mechanics of the Zeiss RMK aerial survey camera.
Typical well known, examples of space cameras, are the Metric Camera on board the
Space Shuttle by ESA, the Large Format Camera also on board the Space Shuttle by
NASA, and the KFA 1000 on board COSMOS by Russia. Figure 2.8.2 shows the LFC
system and its film size. Figure 2.8.3 shows a comparison of photographic coverage on the
ground between LFC (173 km x 173 km) and KFA (75 km x 75 km).
29
As the metric camera is designed for very accurate measurement of topography, the
following requirements in optics as well as geometry should be specified and fulfilled.
(1) Lens distortion should be minimal
(2) Lens resolution should be high and the image should be very sharp even in the corners
(3) Geometric relation between the frame and the optical axis should be established, which
is usually achieved by fiducial marks or reseau marks
(4) Lens axes and film plane should be vertical to each other.
(5) Film flatness should be maintained by a vacuum pressure plate
(6) Focal length should be measured and calibrated accurately
(7) Successive photographs should be mode with high speed shutter and film winding
system
(8) Forward Motion Compensation (FMC) to prevent the image motion of high speed
moving objects during shutter time, should be used, particularly in the case of space
cameras
Multispectral cameras with several separate film scenes in the visible and reflective IR, are
mainly used for photo-interpretation of land surface covers.
Figure 2.8.4 shows a picture taken by the MKF-4, with 4 bands, on board the Russian
Soyuz 22.
Panoramic cameras are used for reconnaissance surveys, surveillance of electric
transmission lines, supplementary photography with thermal imagery, etc., because the
field of view is so wide.
2.9 Film for Remote Sensing
Various type of films are used in cameras for remote sensing. Film can record the
electromagnetic energy reflected from objects in the form of optical density in an emulsion
placed on a film base of polyester. There are panchromatic (black and white film),
infrared film, color film, color infrared film, etc.
The spectral sensitivity of film is different depending on the film type. Black and white
infrared film has wider sensitivity up to near infrared as compared with panchromatic film.
30
Color film has three different spectral sensitivities according to three layers of primary
color emulsion (B,G, R). Color infrared film has sensitivity up to 900 nm. Kodak aerial
color film SO-242 has high resolution and is specially ordered for high altitude
photography.
Generally film is composed of a photographic emulsion which records various gray levels
from white to black according to the reflectivity of objects.
A curve which shows the relationship between the exposure E (meter caudera second) and
the photographic density is called the "characteristic curve". Usually the horizontal axis
of the curve is log E, while the vertical axis is D (density) which is given as follow.
D = log (1 / T ) where T : transparency of film
The characteristic curve is composed of three parts of toe, straight line and shoulder.
Gamma is defined as the gradient of the straight line part, which is an index of contrast. If
is given by D / logE which gives high contrast in the case of gamma larger than 1.0,
and low contrast in the case of gamma smaller than 1.0.
The sensitivity of a photographic emulsion is defined as the minimum exposure to give the
minimum recognizable density. In the definition of JIS (Japan Industrial Standard), the
sensitivity is given as log (1 / EA ) under the conditions of exposure and development
density as denoted ( EA, EB ) and ( A, B ) respectively,
where A : the gross fog + 0.1 EB : EA + 1.5 0.75 < B - A < 0.9
The spectral sensitivity of photographic emulsion (S ) represents the sensitivity with
respect to each wavelength, which is usually given in the form of a spectral sensitivity
curve. In the spectral sensitivity curve log S is used instead of S as the vertical axis, while
sometimes relative sensitivity is used.
Figure 2.9.3 (a) - (d) show the spectral sensitivity curves corresponding to panchromatic,
infrared,color and color infrared films respectively.
2.10 Optical Mechanical Scanner
An optical mechanical scanner is a multispectral radiometer by which two dimensional
imagery can be recorded using a combination of the motion of the platform and a rotating
or oscillating mirror scanning perpendicular to the flight direction. Optical mechanical
31
scanners are composed of an optical system, spectrographic system, scanning system,
detector system and reference system.
Optical mechanical scanners can be carried on polar orbit satellites or aircraft.
Multispectral scanner (MSS) and thematic mapper (TM) of LANDSAT, and Advanced
Very High Resolution Radiometer (AVHRR) of NOAA are the examples of optical
mechanical scanners. M2S made by Daedalus Company is an example of an airborne type
optical mechanical scanner.
Figure 2.10.1 shows the concept of optical mechanical scanners, while Figure 2.10.2 shows
a schematic diagram of the optical process of an optical mechanical scanner.
The function of the elements of an optical mechanical scanner are as follows.
a. Optical system: Reflective telescope system such as Newton, Cassegrain or RitcheyChretien is used to avoid color aberration.
b. Spectrographic system: Dichroic mirror, grating, prism or filter are utilized.
c. Scanning system: rotating mirror or oscillating mirror is used for scanning perpendicular
to the flight direction.
d. Detector system: Electro magnetic energy is converted to an electric signal by the optical
electronic detectors. Photomultiplier detectors utilized in the near ultra violet and visible
region, silicon diode in the visible and near infrared, cooled ingium antimony (InSb) in the
short wave infrared, and thermal barometer or cooled Hq Cd Te in the thermal infrared.
e. Reference system: The converted electric signal is influenced by a change of sensitivity
of the detector. Therefore light sources or thermal sources with constant intensity or
temperature should be installed as a reference for calibration of the electric signal.
Compared to the pushbroom scanner, the optical mechanical scanner has certain
advantages. For examples, the view angle of the optical system can be very narrow, band
to band registration error is small and resolution is higher, while it has the disadvantage
that signal to noise ratio (S/N) is rather less because the integration time at the optical
detector cannot be very long due to the scanner motion.
2.11 Pushbroom Scanner
32
The pushbroom scanner or linear array sensor is a scanner without any mechanical
scanning mirror but with a linear array of solid semiconductive elements which enables
it to record one line of an image simultaneously, as shown is Figure 2.11.1.
The pushbroom scanner has an optical lens through which a line image is detected
simultaneously perpendicular to the flight direction. Though the optical mechanical
scanner scans and records mechanically pixel by pixel, the pushbroom scanner scans and
records electronically line by line.
Figure 2.11.2 shows an example of the electronic scanning scheme by switching method.
Because pushbroom scanners have no mechanical parts, their mechanical reliability can be
very high.
However, there will be some line noise because of sensitivity differences between the
detecting elements.
Charge coupled devices, called CCD, are mostly adopted for linear array sensors.
Therefore it is sometime called a linear CCD sensor or CCD camera. HRV of SPOT,
MESSR of MOS-1, and OPS of JERS-1 are examples of linear CCD sensors as is the Itres
CASI airborne system. As an example, MESSR of MOS-1 has 2048 elements with an
interval of 14 mm. However CCD with 5,000 - 10,000 detector elements have been
developed and available recently made available.
2.12 Imaging Spectrometer
Imaging spectrometers are characterized by a multispectral scanner with a very large
number of channels (64-256 channels) with very narrow band widths, though the basic
scheme is almost the same as an optical mechanical scanner or pushbroom scanner.
The optical system of imaging spectrometers are classified into three types; dioptic system,
dio and catoptic system and catoptic system which are adopted depending on the scanning
system. Table 2.12.1 shows a comparison of the three types. In the case of object plane
scanning, the catoptic system is best chosen because the linearity of the optical axis is very
good due to the narrow view angle and the observation wave range is so wide. However in
33
the case of image plane scanning, the dioptic system or dioptic and catoptic system is best
suited because the view angle should be wider.
Figure 2.12.1 shows four different types of multispectral scanner. The left upper
(multispectral imaging with discrete detectors) corresponds to the optical mechanical
scanner on the using the object plane scanning method used in the LANDSAT. The right
upper (multispectral imaging with line arrays) corresponds to pushbroom scanner using the
image plane scanning method with a linear CCD array.
The left lower (imaging spectrometry with line arrays) shows a similar scheme to the right
upper system but with an additional dispersing element (grating or prism) to increase the
spectral resolution. The right lower (imaging spectrometry with area arrays) shows an
imaging spectrometer with area arrays.
Table 2.12.2 shows the optical scheme of the Moderate Resolution Imaging SpectrometerTilt (MODIS-T) which is scheduled to be carried on EOS-a (US Earth Observing
Satellite). MODIS-T has an area array of 64 x 64 elements which enables 64 multispectral
bands from 0.4 mm to 1.04 mm with a 64 km swath. The optical path is guided from scan
mirror to Schmitt type off axis parabola of dio and catoptic system. The the light is then
dispersed into 64 bands by a grating and is detected by an area CCD array of 64 x 64
elements.
As imaging spectrometer provides multiband imagery with a narrow wave length range,
and is useful for rock type classification and ocean color analysis.
2.13 Atmospheric Sensors
Atmospheric sensors are designed to provide measures of air temperature, vapor,
atmospheric constituents, aerosols the etc. as well as wind and earth radiation budget.
Figure 2.13.1 shows the important atmospheric constituents for green house effect gases,
ozone layer and acid rain.
As remote sensing techniques cannot meet the direct measurement of these physical
magnitude, it is necessary to estimate them from spectral measurement of atmospheric
scattering, absorptance or emission.
34
The spectral wave length range is very wide from the near ultraviolet to the millimeter
radio wave depending on the objects to be measured (See 1.4).
There are two types of atmospheric sensor, that is, active and passive . Because the active
sensor is explained in section 2.15 "Laser Radar" or Lidar,only the passive type sensors
will be introduced here.
Two directions of atmospheric observation are usually adopted; one is nadir observation
and the other is limb observation as shown in Figure 2.13.1. The nadir observation is
superior in the horizontal resolution compared to vertical resolution. It is mainly useful in
the troposphere but not in the stratosphere where the atmospheric density is very low.
The limb observation method is to measure the limb of the earth with an oblique angle.
In this case, not only atmospheric emission but also atmospheric absorption of the light of
the sun, the moon and the stars are measured, as shown in Figure 2.13.1. Compared with
the nadir observation, the limb observation has higher vertical resolution and higher
measurability in the stratosphere. The absorption type of limb observation has rather high
S/N but observation direction or area is limited except for the stars.
There are two types of atmospheric sensors, that is, sensors with a fixed observation
direction, called sounders and scanners.
The main element of optical sensor is a spectrometer with a very high spectral resolution
such as the Michelson spectrometer, Fabry-Perot spectrometer and other spectrometers
with grating and prism.
Figure 2.13.2 shows the structure of Michelson spectrometer called IMG which will be
borne on ADEOS (Advanced Earth Observing Satellite to be launched in 1995 by Japan).
2.14 Sonar
Sound waves or ultrasonic waves are used underwater to obtain imagery of geological
features at the bottom of the sea or lakes because radio waves are not usable in water.
Sound waves have many characteristics such as reflection, refraction, interference,
diffraction etc. similar to radio waves, though it is an elastic wave different from the radio
wave. Sound waves have a form of longitudinal wave in water along the direction of the
wave. Generally, sound waves transmitting in water have a higher resolution according to
35
higher frequency but also higher attenuation. The detectability depends on S/N ratio when
receiving the sound signal after loss by noises in water.
The velocity of sound is approximately 1,500 meters/second which varies depending on
temperature, water pressure, and salinity of the medium.
As shown in Figure 2.14.1, there are side scan sonar and multi-beam echo sounder by
which the sea bottom is scanned and imaged. These sensors are kinds of active sensors
which record the sound intensity reflected from the projected sound wave onto the bottom.
Because sonar is an active sensor, it generates image distortion from the effects of
foreshortening, layover and shadow, with respect to incident angle at the bottom, in the
same manners as radar.
As shown in Figure 2.14.2, a side scan wave is produced from a transducer borne on a
towfish connected by tow cable to a tug boat. The incident sound wave on the sea bottom
will produce sound pressure on the bottom materials causing back scattering to return to
the receiver, after attenuation, according to the shape and density of the bottom. The sonar
acquires the backscattering in the time sequence to a form of image.
Figure 2.14.3 shows a multi narrow beam sounder with a transmitting transducer and a
receiving transducer in a T shape at the bottom of the boat. A receiving transduce has 20
to 60 elements which receive the sound signal reflected from the sea bottom, which is
usually converted to an image as for the side scan sonar.
2.15 Laser Radar
Devices which measures the physical characteristics such as distance, density, velocity,
shape etc., using scattering, returned time, intensity, frequency and/or polarization of light
are called optical sensors. However as the actual light used by the optical sensor is mostly
laser, it is usually called laser radar or lidar (light detection and ranging).
Laser radar is an active sensor which is used to measure air pollution, physical
characteristics of atmospheric constituents in the stratosphere and its spatial distribution.
The theory of laser radar is also utilized to measure distance, so that for this application it
is called laser distancemeter or laser altimeter.
36
The main measurement object is the atmosphere although laser radar is also used to
measure water depth, thickness of oil film or vividness of chlorophyll in vegetation.
[ Theory of Lidar ]
Figure 2.15.1 shows a schematic diagram of a lidar system. The power of the received light
Pr (R) reflected from a distance of R can be expressed as follows.
Pr (R) = Po K Ar q (R) T (R) Y (R) / R + Pb
where Po : intensity of transmitted light
K : efficiency of optical system
Ar : aperture
q : half wavelength
b (R) : backscattering coefficient
T (R) : transmittance of atmosphere
Y (R) : geometric efficiency
Pb : light noises of background
Received light is converted to an electric signal which is displayed or recorded after A/D
conversion. The effective distance of lidar depends on the relationship between the
received light intensity and the noise level.
Lidar can be classified with respect to its physical characteristics, interactive effects,
physical quantities etc., as shown in Table 2.15.1. In this table, Mie Lidar is the most
established sensor,with a which signal intensity large enough to measure Mie scattering
due to aerosols.
Fluorescence lidar, Roman lidar and differential absorption lidar Figure 2.15.3 are utilized
for measurement of density of gaseous body, while Doppler lidar is used for measurement
of velocity. Polarization effects of lidar is utilized for measurement of shape.
There are several display modes for example, a scope with horizontal axis of distance and
with vertical axis of intensity, PPI (plane position indication ) with gray level in polar
coordinate system, RHI (range high indication) with a display of the vertical profile, THI
(time height indication) with a horizontal axis of elapsed time and with a vertical axis of
altitude.
37
Chapter 3 Microwave Remote
Sensing
3.1 Principles of Microwave Remote Sensing
Microwave remote sensing, using microwave radiation using wavelengths from about one
centimeter to a few tens of centimeters enables observation in all weather conditions
without any restriction by cloud or rain. This is an advantage that is not possible with the
visible and/or infrared remote sensing. In addition, microwave remote sensing provides
38
unique information on for example, sea wind and wave direction, which are derived from
frequency characteristics, Doppler effect, polarization, back scattering etc. that cannot be
observed by visible and infrared sensors. However, the need for sophisticated data analysis
is the disadvantage in using microwave remote sensing.
There are two types of microwave remote sensing; active and passive. The active type
receives the backscattering which is reflected from the transmitted microwave which is
incident on the ground surface.
Synthetic aperture radar (SAR), microwave scatterometers, radar altimeters etc. are active
microwave sensors. The passive type receives the microwave radiation emitted from
objects on the ground. The microwave radiometer is one of the passive microwave sensors.
The process used by the active type, from the transmission by an antenna, to the reception
by the antenna is theoretically explained by the radar equation as described in Figure
3.1.1.
The process of the passive type is explained using the theory of radiative transfer based on
the law of Rayleigh Jeans as explained in Figure 3.1.2 (see 1.7, 1.12 and 3.2) In both active
and passive types, the sensor may be designed considering the optimum frequency needed
for the objects to be observed. (see 4.1)
In active microwave remote sensing, the characteristics of scattering can be derived from
the radar cross section calculated from received power Pr and antenna parameters (At ,
Pt , Gt ) and the relationship between them, and the physical characteristics of an object.
For example, rainfall can be measured from the relationship between the size of water drops
and the intensity of rainfall.
In passive microwave remote sensing, the characteristics of an object can be detected from
the relationship between the received power and the physical characteristics of the object
such as attenuation and/or radiation characteristics. (see 3.2 and 3.3)
3.2 Attenuation of Microwave
Attenuation results from absorption by atmospheric molecules or scattering by aerosols
in the atmosphere between the microwave sensor on board a spacecraft or aircraft and the
target to be measured. The attenuation of the microwave will take place as a function of
39
the exponential with respect to the transmitted distance mainly due to absorption and
scattering. Therefore the attenuation will increase in proportion to the distance, under
homogeneous atmospheric conditions. The attenuation per unit of distance is called
specific attenuation. Usually the loss due to attenuation can be expressed in the units of
dB (decibel) as follows.
dB = K e B dr
where Ke: specific attenuation (dBkm )
B : brightness temperative ( Wm sr ) in the distance of dr
dr : incremental distance
Figure 3.2.1 shows the attenuation characteristics of atmospheric molecules with respect
to frequency. From this figure it can be seen that the influence of atmospheric attenuation
occurs in the region greater than 10GHz. The intensity of attenuation depends on the
specific frequency (absorption spectrum) of the corresponding molecule. This is the reason
why the energy of the microwave is absorbed by the molecular motion of the atmospheric
constituent. However, if proper frequencies are carefully selected, the attenuation can be
minimized because the composition of the atmospheric constituent is almost homogeneous.
In the case of satellite observation, the optical path is usually long in distance, so that
attenuation can be influenced by the change in atmospheric conditions. Particularly
because the attenuation of vapor (H2O) is very strong in the specific frequencies, the
change of vapor can be detected by a microwave radiometer.
The most remarkable scattering in the atmosphere is due to rain drops. Figure 3.2.2 shows
the attenuation characteristics due to scattering of rain drops and mist. The attenuation
increases if the intensity of rainfall increases, and the frequency increases until about 40
GHz. However, over 40 GHz the attenuation does not depend on the frequency.
Remarks 1) dB is 1 / 10 bel.
"bel" is logarithmic ratio of two powers P1 and P2 .
N = log10 ( P1 / P2 ) [bel] or
n = 10 log10 ( P1 / P2 ) [dB]
40
2) Specific attenuation Ke is originally expressed as Np m or neepers m . But K e is
converted to dB km for convenience by multiplying 10 log e = 4.34 x 10 by Np m .
3.3 Microwave Radiation
The earth surface radiates a little microwave energy as well as visible and infrared because
of thermal radiation. The thermal radiation of a black body depends on Plank's law in the
visible and infrared region, while the thermal radiation in the microwave region is given
by the Rayleigh Jeans radiation law .
Real objects, the so called gray bodies are not identical to a black body but have constant
emissivity which is less than a blackbody . The brightness temperature TB is expressed
as follows.
TB = T
where T : physical temperature
: emissivity (0 < < 1)
Emissivity of an object changes depending on the permittivity, surface roughness,
frequency, polarization, incident angle, azimuth etc., which influence the brightness
temperature.
Figure 3.3.1 shows the characteristics of emissivity for 3.5 % density salt water with respect
to incident angle, polarization and frequency. Figure 3.3.2 shows the emissivity of
horizontal polarization (eh) and vertical polarization (ev) for two clay soils with different
soil moisture, and a sea water with respect to incident angle.
Table 3.3.1 shows the emissivity of typical land surface covers for two different grazing
angles of 30 and 45 .
Most users would like to get the physical temperature T instead of brightness temperature
TB , which is measured by microwave radiometers.
Therefore emissivity should be measured or algorithms should be developed to identify the
component of atmospheric radiation.
In the case of the receiving antenna of the microwave radiometer, all radiation from various
angles are input to the antenna, which needs a correction of the received temperature with
41
respect to directional properties of the antenna. The corrected temperature is called the
antenna temperature.
3.4 Surface Scattering
Surface scattering is defined as the scattering which takes place only on the border surface
between two different but homogeneous media, from one of which electro-magnetic energy
is incident on to the other. Scattering of microwave on the ground surface increases
according to the increase of complex permittivity, and the direction of scattering depends
on the surface roughness, as shown in Figure 3.4.1.
In the case of a smooth surface as shown in Figure 3.4.1 (a), there will be a specular
reflection with a symmetric angle to the incident angle. The intensity of specular
reflection is given by Fresnel reflectivity, which increases in accordance with the increase
of the ratio of complex permittivity.
When the surface roughness increases a little as shown in Figure 3.4.1 (b), there exists a
component of specular reflection and a scattering component. The component of specular
reflection is called the coherent component, while that of scattering is called diffuse or the
incoherent component.
When the surface is completely rough, that is diffuse, only diffuse components will remain
without any component of specular reflection as shown in Figure 3.4.1 (c). Such surface
scattering depends on the relationship between the wavelength of electro-magnetic
radiation and the surface roughness which is defined by the Rayleigh Criterion or
Fraunhofer Criterion.
Rayleigh Criterion : if h < / 8 cos , the surface is smooth
Fraunhofer Criterion : if h < / 32 cos , then the surface is smooth
where h : standard deviation of surface roughness
: wavelength
: incident angle
Generally the scattering coefficient, that is scattering area per unit area, is a function of
incident angle and the scattering angle. However in the case of remote sensing, the
scattering angle is identical to the incident angle because the receiving antenna of radar or
42
scatterometer is located at the same place as the transmitting antenna. Therefore, in remote
sensing only back- scattering may be taken into account. The radar sectional area
i Ai is
given as follows.
where P t : transmitting power
G : antenna gain
: wavelength
Pr : receiving power
R : distance between radar and object
Ai : differential area of surface scattering
Scattering area per a unit area
is called the backscattering coefficient.
= i / Ai
The backscattering coefficient depends on the surface roughness and incident angle as
shown in Figure 3.4.2.
3.5 Volume Scattering
Volume scattering is defined as the scattering occurring in a medium when electromagnetic radiation transmits from one medium to another medium. Figure 3.5.1 shows the
schematic model of volume scattering for two examples; (a)scattering by widely distributed
particles such as rain drops and (b)scattering in uneven media with different permittivities.
Scattering by trees or branches, subsurface or soil layers, snow layers etc. are examples of
volume scattering.
Volume scattering can be observed if microwave radiation penetrates into a medium. The
penetration depth is defined as the distance when the incident power attenuates to 1/e
(exponential coefficient).
The intensity of volume scattering is proportional to the discontinuous inductivity in a
medium and the density of the heterogeneous medium. The scattering angle depends on
surface roughness, average relative permittivity and wavelength.
43
The receiving intensity is proportional to multiplication of the intensity and the volume
involved in the region of range gate and beam width as shown in the example in Figure
3.5.3. Volume scattering in the case of rainfall, shown in Figure 3.5.2, is represented as a
function of wavelength and Z factor as follows.
where : wavelength
D : diameter of rain drop
k : constant (k = ( -1) / ( +2))
Z : Z factor (Z= Di )
In the case of soil and snow, volume scattering occurs together with surface scattering,
although the surface scattering is small as shown in Figure 3.5.3. There will exist an error
for the measurement of surface scattering coefficient because of the effect of volume
scattering.
In the case of forest as shown in Figure 3.5.4, it is necessary to introduce a model of the
volume scattering by leaves and branches as well as surface scattering by the crown of
trees, and corner reflection effects due to the soil and vertical tree trunks.
3.6 Types of Antenna
An antenna is a transducer to transform from a high frequency electric current to radio
waves and vice versa. An antenna is used to transmit and receive radio waves. There are
many kinds of antenna ranging from very small size (such as a monopoly antenna in a
cordless telephone) to very large antenna reflectors of 100 meters in diameter for radio
wave astronomy. In this section, antennas used for microwave remote sensing are
introduced.
Typical antennas in microwave remote sensing are those of the passive type of microwave
radiometer, active types of microwave altimeter, scatterometer and imaging radar.
44
There are three major types of antenna; horn antenna, reflector mirror antenna and array
antenna.
The horn antenna such as the conical horn or rectangular horn is used for power supply to
the reflector antenna, calibration of low temperatures for the microwave radiometer in the
form of a sky horn looking upward, and calibration for active radar as shown in Figure
3.6.1.
Reflector antenna such as parabolic antenna and Cassegrainian antenna are composed of
primary radiator and a reflective mirror as shown in Figure 3.6.2. The reflector antenna is
used for microwave radiometers, altimeters and scatterometers. In case of wide angle
scanning, all antenna will be controlled, while in the case of narrow beam scanning only
the radiometer or reflective mirror will be controlled.
An array antenna is composed of multiple element arrays for example, linear array, area
array or nonformal array. The element antennas are half-wavelength dipoles, microstrip
patches and wave guide slot. The advantages of array antenna are to enable beam scanning
without changing the looking angle of each array antenna and to generate an appropriate
beam shaping by selective excitation of current distribution of each element.
The array antenna is used for synthetic aperture radar (SAR) and real aperture radar. Figure
3.6.3 shows a wave guide slot array antenna designed for real aperture radar.
3.7 Characteristics of Antenna
An ordinal antenna is used for transmitting radio waves in a specific direction or for
receiving radio waves from a specific direction. Therefore it can be said that the antenna is
a spatial filter for radio wave.
Relative power, given as a function of the beam angle is called the radiation pattern or
beam pattern. Usually the beam pattern is given in an orthogonal coordinate system or polar
coordinate system, as shown in Figure 3.7.1.
The characteristics of the beam pattern can be determined by making a Fourier
transformation of the aperture distribution. If the size of antenna aperture is infinite, the
45
beam pattern should be an impulse pattern. But as actual antenna are limited in size, the
beam pattern has several lobes with respect to beam angles, as shown in Figure 3.7.2.
The point with zero power is called the null and the pattern between two nulls is called the
lobe. The central biggest lobe is called the main lobe, while the other robes are called
sidelobes.
The beam width of the antenna is defined as the beam width at the power level of 3dB
downward from the peak of the main lobe (equivalent to the half power beam width). The
difference between the peaks of the main lobe and the biggest sidelobe is called the sidelobe
level. Antenna gain is given as the ratio of power density of an antenna to the reference
antenna with a given constant power at a specific angle. The antenna gain that is obtained
by an isotropic antenna as the standard antenna is called standard gain. The ratio of the
power density at a specific angle to the average power density determined from all radiative
power is called directivity and is given as follows.
where E ( , ) : field strength at the direction of and (horizontal and vertical angles)
Usually characteristics of transmitting antenna and receiving antenna are identical to each
other.
Chapter 4 Microwave Sensors
4.1 Types of Microwave Sensor
There are two types of microwave sensors, passive and active. Table 4.1.1 shows the
typical microwave sensors and targets to be measured. Table 4.1.2 shows the frequency of
passive microwave sensor for monitoring major targets. Table 4.1.3 shows the frequency
of active microwave sensor for monitoring major targets.
46
Many of the earth observation satellites to be launched after 1992 are planned to have
microwave sensors onboard. Active sensors will be classified into more types in terms of
the target with respect to horizontal or vertical polarization.
4.2 Real Aperture Radar
Imaging radar as shown in Table 4.1.1 is classified further into Real Aperture Radar
(RAR) and Synthetic Aperture Radar (SAR). In this section RAR is explained.
RAR transmits a narrow angle beam of pulse radio wave in the range direction at right
angles to the flight direction (called the azimuth direction) and receives the backscattering
from the targets which will be transformed to a radar image from the received signals, as
shown in Figure 4.2.1.
Usually the reflected pulse will be arranged in the order of return time from the targets,
which corresponds to the range direction scanning.
The resolution in the range direction depends on the pulse width, as shown in Figure 4.2.2.
However if the pulse width is made small, in order to increase the resolution, the S/N ratio
of the return pulse will decrease because the transmitted power also becomes low.
Therefore, the transmitted pulse is modulated to chirp with a high power but wide band,
which is received through a matched filter, with reverse function of transmission, to make
the pulse width very narrow and high power as shown in Figure 4.2.3. This is called pulse
compression or de-chirping. By making the pulse compression, with an increase of
frequency f in transmission, the amplitude becomes
times bigger, and the pulse width
becomes 1/TDf narrower. This method is sometime called range compression.
The resolution in the azimuth direction is identical to the multiplication of beam width and
the distance to a target. As the resolution of azimuth direction increases with shorter wave
length and bigger antenna size, a shorter wavelength and a bigger antenna is used for higher
azimuth resolution, as shown in Figure 4.2.4.
However as it is difficult to attach such a large antenna, requiring for example a 1 km
diameter antenna in order to obtain 25 meters resolution with L band ( =25 cm) and 100
km distance from a target, a real aperture radar therefore has a technical limitation for
improving the azimuth resolution.
47
4.3 Synthetic Aperture Radar
Compared to real aperture radar, Synthetic Aperture Radar (SAR) synthetically increases
the antenna's size or aperture to increase the azimuth resolution though the same pulse
compression technique as adopted for range direction. Synthetic aperture processing is a
complicated data processing of received signals and phases from moving targets with a
small antenna, the effect of which is to should be theoretically convert to the effect of a
large antenna, that is a synthetic aperture length, as shown as Figure 4.3.1.
The synthetic aperture length is the beam width by range which a real aperture radar of the
same length, can project in the azimuth direction.
The resolution in the azimuth direction is given by half of real aperture radar as shown as
follows.
Real beam width : = /D
Real resolution: L= R=Ls (synthetic aperture length)
Synthetic beam width : s = / 2Ls= D / 2R
Synthetic resolution : Ls = sR = D / 2
where :wavelength D: aperture of radar R: slant range
This is the reason why SAR has a high azimuth resolution with a small size of antenna
regardless of the slant range, or very high altitude of a satellite.
Figure 4.3.2 shows the basic theory of SAR or synthetic aperture processing including the
Doppler effect, matched filter and azimuth compression.
SAR continues to receive return pulses from a target during the time the radar projects the
beam to the target. In the meanwhile the relative distance between the radar and the target
changes with the movement of the platform, which produces a Doppler effect to modulate
a chirp modulation of received pulse. A matched filter corresponding to the reverse
characteristics of chirp modulation will increase the azimuth resolution of azimuth
direction. This is called azimuth compression.
In the case of SAR, unsuitability of satellite velocity and attitude will reduce the effect of
the Doppler effect. Therefore the satellite with SAR is required to be high, because the
48
correction for synthetic aperture processing due to instability at lower altitudes is very
difficult.
.4 Geometry of Radar Imagery
The incident angle of microwave to a target is the angle from the normal line while the
aspect angle is the supplementary angle as shown in Figure 4.4.1. The smaller the incident
angle , the larger the back scattering intensity .
Off nadir angle is the angle between the microwave and the nadir, while the depression
angle is the angle from the horizon, as shown in Figure 4.4.2. There occurs a geometric
distortion or shadow depending on the relationship between the off nadir angle and the
terrain relief, as shown in Figure 4.4.3 and Figure 4.4.4. Foreshortening occurs when the
ground range difference (horizontal distance X ) is reduced to the slant range difference
DR, because the slant range to the top and the bottom are not proportional to the horizontal
distances, as shown in the right hand example of Figure 4.4.5.
When the shortening becomes greater, the image of the top of a feature such as a mountain
will be closer to the antenna than the bottom, which causes loss of, the slope image, as
shown in the left hand example of Figure 4.4.5. This is called layover.
Such phenomena will occur when the terrain relief is greater and the off nadir angle is
smaller.
However for large off nadir angles, there will be a shadow area, as called radar shadow,
behind a hill or mountain, which makes the image very dark as shown in Figure 4.4.5.
The radar shadow will occur if the following relationship is satisfied.
+ > 90
where : slope angle of back slope
: off nadir angle
The slant range of the shadow is given as sec as shown in Figure 4.4.4.
4.5 Image Reconstruction of SAR
Raw data from SAR are records of backscattering in time sequence which are returned
from the ground targets. Return signal from a point P is recorded in the expanded range in
the range direction which is identical to the pulse width. In addition, the return signal from
49
a point P is also expanded in the azimuth direction because the point P continues to be
radiated by microwave pulses during the flight motion(see Figure 4.5.1).
Data processing to generate an image of gray tone corresponding to the backscattering
intensity of each point on the ground is called image reconstruction of synthetic aperture
radar (SAR). Figure 4.5.3 shows the flow of image construction of SAR.
The image reconstruction is divided into range compression and azimuth compression
which make compression of expanded signals in both range and azimuth directions into a
point signal. The compression is usually carried out by adopting Fourier transformation to
achieve convolution of received signals and a reference function.
The reference function of range compression is the complex conjugate of the transmitted
signal, while the reference of azimuth compression is a complex conjugate of the
modulated signal by chirp modulation.
The slant range to a point on the ground is expressed as the quadratic function of time with
respect to the movement of the platform. The change of the slant range is called range
migration. The first order term is called range walk resulting from the earth rotation,
while the second order term is called range curvature.
The range migration correction is to relocate the quadratic distribution (see Figure 4.5.4
(c)), in which range walk and range curvature may be separately processed.
In image reconstruction, there is a major problem, called speckle which is due to high
frequency noise, as seen in the example of Figure 4.5.2. In order to reduce the speckle,
mulch-look processing is applied in which range compression and azimuth compression
with respect to subdivided frequency domains are independently overlaid three or four
times termed the number of looks. Sometimes a median filter or local averaging may be
applied to reduce the speckle. The speckle will be reduced by the square root of the number
of looks, although the spatial resolution declines in proportion to the number of looks.
4.6 Characteristics of Radar Image
The main objective of microwave remote sensing is to estimate the property of objects by
interpreting the features of the radar image. Typical objects to be measured by microwave
remote sensing are mountainous land forms, subsurface geology, sea wind and waves etc.
50
In order to estimate these properties, it is very important to understand the effects of
microwave backscattering on the objects.
Two factors of microwave characteristics are of importance; frequency (or wavelength)
and polarization. In microwave remote sensing, various wavelengths (or frequency) such
as L band, C band, X band, P band etc. will be used ranging from millimeter wavelengths
(1 mm - 1cm) up to about 30 cm. According to the wavelength or frequency, specular
reflection will occur, so that the surface roughness can be detected if multi-frequency radar
images are compared.
Figure 4.6.1 and 4.6.3 are radar images of P band, L band and C band respectively with
HH polarization, in which the difference of wavelength or frequency will provide different
images.
Polarization is defined as the oscillating direction involved in an electric field. Usually
transmitted microwave and received microwave will have a choice between horizontal
polarization and vertical polarization. Therefore four combinations; HH, HV, VH and VV
can be used for SAR. The backscattering characteristics are also different with respect to
polarization. Figure 4.6.2 and 4.6.4 are HH, HV, and VV combinations respectively with
L band in which one can see different features.
In future, SAR systems with functions of multi-frequency and multi- polarization will be
onboard earth observation satellites.
4.7 Radar Images of Terrains
The biggest effect of microwave backscattering on variations in the radar image is due to
terrain features. It is larger than the effect of permittivity. Particularly the effect of incident
beam angle in terms of off nadir angle and terrain slope will produce various effects such
as foreshortening, layover and shadow as already explained in section 4.4.
Normally in the closer range from SAR, called near range, layover may occur while in
the far range more shadow may be seen. This means that care should be taken with the
flight direction and range direction in interpretation of terrain features.
Figure 4.7.1 shows an example of a SAR image of SEASAT around the mountainous areas
of Geneva, Switzerland.
51
Figure 4.7.2 shows the effect of terrain and off nadir angle on foreshortening, layover and
shadow.
Usually foreshortening and layover appear as a bright response around the summit or ridge,
while shadow appears black without any information in the shadow area.
As seen in the figure, the effect of microwave back scattering can be better seen along the
track or azimuth direction than in the cross track direction.
By interpreting the radar image, land form classification, lineament analysis, mineral
resources exploration, monitoring of active volcanoes, land slide monitoring, geological
structure analysis and so on can be carried out.
Two parallel flights may produce a stereo pair which will offer the elevation information
on terrain features. Recent work is Canada has demonstrated that terrain elevation
information can also be derived from the use of interferometry with a single flight line.
4.8 Microwave Radiometer
As indicated in section 3.4, a part of the microwave is also radiated by thermal radiation
from the objects on the earth. Microwave radiometers or passive type microwave sensors
are used to measure the thermal radiation of the ground surface and/or atmospheric
condition.
Brightness temperature measured by a microwave radiometer is expressed by ReyleighJean's law (see 1.7), which is the resultant energy of thermal radiation from the ground
surface and the atmospheric media. Multi-channel radiometers with multi- polarization are
used to avoid the influences of unnecessary factors to measure the specific physical
parameter.
Figure 4.8.1 shows the sensitivity of physical parameters in oceanography with respect to
frequency and the optimum channels as arrow symbols.
Figure 4.8.2 shows two typical microwave scanning radiometers; the conical scanning
type and the cross track scanning type. The former is used for the microwave channel
which is influenced by the ground surface, while the latter is used for the channel which
can be neglected by the influence of the ground surface.
52
The most simple radiometer is the total power radiometer, as shown in Figure 4.8.3. This
system has a mixer to enable it to mix high frequency of a local oscillator in order to
amplify the high signal after transforming to a low frequency. However the influence of
system gain variation cannot be neglected in this system.
The Dicke radiometer can reduce the influence of system gain variation by introducing a
switch generator which allows it to receive the antenna signal and noise source of constant
temperature, alternatively of which antenna signal can be detected later on, synchronously
with the switch generator.
The zero-balance Dicke radiometer can reduce the influence of system gain variation and
increase the sensitivity further by adding a noise generator to the Dicke radiometer in order
to increase the sensitivity about two times higher than total power radiometer.
4.9 Microwave Scatterometer
Microwave scatterometers can measure the received power of surface backscattering
reflected from the surface of objects. According to a narrow definition, a microwave
scatterometer may be a space borne sensor to measure the two dimensional velocity vectors
of the sea wind, while according to the wider definition, it also involves air-borne sensors,
as well as ground based sensors to measure the surface backscattering as well as volume
scattering, such as rain radar.
Microwave scatterometers are classified as two types, pulse type and continuous wave
type (CW). The pulse type uses wide band which has restrictions in obtaining a license to
operate and in avoid obstructions. CW type has the advantage that the band width can be
reduced to 1/100 times that of the pulse type and the price becomes cheaper.
SEASAT-SASS (Seasat-A Satellite Scatterometer) is one of the typical scatterometers.
SASS has four fixed antennas to transmit the pulse in a fan beam of 14.5 GHz to four
different angles, and to receive the backscattering in subdivided cells through a Doppler
filter. Figure 4.9.1 shows the four beam patterns and the incident angles. In accordance
with the satellite flight, the same cell of sea area can be observed from both fore beam and
aft beam with 90 different angle, which enables the determination of wind direction and
wind velocity.
53
Figure 4.9.2 shows a ground based microwave scatterometer with a rotation system.
ERS-1-AMI Wind Mode (European Remote Sensing Satellite-1-Active Microwave
Instrument) and ADEOS-NSCAT (Advanced Earth Observation System-1 NASA
Scatterometer) will be available for measurement of velocity vectors of the sea wind with
three antennas looking fore, side and aft directions.
Table 4.9.1 compares the basic functions of ERS-AMI Wind Mode and ADEOS-NSCAT.
Both have functions to measure wind velocity of the sea wind with an accuracy of 2 m/s or
10 % of the waveheight , and wind direction with an accuracy of 20 .
4.10 Microwave Altimeter
Microwave altimeters or radar altimeters are used to measure the distance between the
platform (usually satellite or aircraft) and the ground surface. Some applications of
microwave altimetry are in ocean dynamics of the sea current, geoid surveys, and sea ice
surveys. Therefore, precise measurement of the satellite orbit and the geoid should be
carried out.
The principle of satellite altimetry is shown in Figure 4.10.1. The ground height or the sea
surface height is measured from the reference ellipsoid. If the altitude of the satellite Hs
is given as the height from the reference ellipsoid, the sea surface height HSSH is
calculated as follows.
HSSH = Hs - Ha where Ha : measured distance between satellite and the sea surface.
The sea surface height is also represented by the geoid height Hg that is measured between
the geoid surface and the reference ellipsoid and the sea surface topography DH is given
as follows.
HSSH = Hg + ΔH
The sea surface topography results from ocean dynamics such as sea current, wave height,
tidal flow etc., which can be determined if the geoid height Hg is given. The distance
between a satellite and the ground surface or sea surface; Ha, is measured on the basis of
the travel time of the transmitted microwave pulses. From the time (t=0), when the first
edge of pulse arrives at the surface, to the time (t= ) when the end edge of a pulse with a
width of arrives at the surface, the received power increases linearly as shown in Figure
54
4.10.2. The received pulses are composed of echoes from various parts of the sea surface.
Therefore the travel time from a satellite to the sea surface can be calculated by averaging
the received pulses. Pulse compression techniques will be also applied (see 4.2) in order to
obtain a high frequency pulse for improvement of the resolution.
Ocean wave height can be estimated if the relation between the average scattering
coefficient and elapsed time as seen in the different gradients in Figure 4.10.3 is taken into
account.
4.11 Measurement of Sea Wind
Measurement of the sea wind is not made directly but indirectly from two processes, that
is, the microwave scattering from the sea surface and relationship between sea wind and
wave height.There are two methods for measurement of sea wind.
a. to estimate the backscattering coefficient
using a microwave scatterometer
b. to measure the brightness temperature using a microwave radiometer
The following three models are used to estimate the backscattering coefficient in the case
of a microwave scatterometer.
a. specular point model
b. Bragg model
c. composite surface model
Figure 4.11.1 shows the relationship between backscattering cross section and the incident
angle which has been obtained from the actual measurement. The specular point model can
be applied to the region A in the figure, with the incident angle from 0 to 25 , where the
sea clutter or sea roughness is much larger than the microwave wavelength. In this case,
the backscattering coefficient is proportional to the resultant probability density of x and
y components of the gradient.
The Bragg model can be applied to the region B in the figure with the incident angle larger
than 25 , where ¡ decreases very gently except near 90 . This is called Bragg scattering
which can be seen under the condition when the wavelengths of microwave and the sea
wave have a similar spectrum. The ideal condition for Bragg scattering has the range from
25 to 65 with the wavelength of the capillary wave being from 1 cm to a few cm.
55
However, the actual sea surface is a composite of capillary waves and gravity waves for
which the composite surface model has been developed but not yet verified theoretically.
The second procedure is to estimate the sea surface condition from the backscattering
coefficient . Figure 4.11.2 shows the correlation between variance of wave slope, S , and
the wind velocity measured by Cox and Munk, which can be applied for the specular point
model when is given as a function of S . In the case of >25, it is found that the sea wind
is proportional to the spectral density, but the models are still under development. Figure
4.11.3 shows the distribution of wind velocity and wind direction that was measured by
SASS (Seasat-A Satellite Scatterometer) for the typhoon on Oct. 2, 1978. In the case of the
microwave radiometer, the sea wind can be computed from the brightness temperature
using the fact that the emissivity is a function of complex permittivity with parameters of
salinity, sea clutter, sea temperature and bubbles. Two algorithms have been developed by
Wentz and Wilheit for Seasat- SMMR for the sea wind velocity.
4.12 Wave Measurement by Radar
While ordinary measurements of waves by a wave gauge are used for a time variation of
wave height at a point, remote sensing techniques gives information over a broader area.
There are two methods of wave measurement used in remote sensing. a. space borne
sensors such as SAR and microwave altimeter b. ground base radar such as X band radar
and HF Doppler radar
As space borne sensors have a lower resolution (of the order of a few ten meters to a few
kilometers), large size currents, typhoons over wide areas, global wave distribution, etc.,
will be better monitored by these systems.
On the other hand, ground based radar is suitable for monitoring the waves in near offshore
or shallow zones, with the wave field of a few ten centimeters to a few hundred meters.
Figure 4.12.1 shows the distribution of average significant wave height which was
measured by GEOSAT-ALT, with an accuracy of +-0.5 m or 10 % of the wave height.
Space borne SAR and ground based X band radar are used for measurement of reflectivity
from the sea clutter at similar wavelengths to the sensors based on Bragg scattering (see
4.11).
56
Figure 4.12.2 explains the effect of capillary waves with respect to slope change or
incident angle for Bragg scattering and wave height.
Figure 4.12.3 shows the sea surface conditions measured by SAR and X band radar. The
measurement of wave direction and wave length is already operational as a ship borne radar
but the measurement of wave height is still being researched. HF Doppler radar using high
frequency band (10-100 m) , which is longer than microwave can measure the Bragg
scattering from the sea clutter with wave lengths of 5-50 m.
Wave conditions such as wave direction, significant wave height, predominant wave,
current etc. of wind waves with longer wave lengths, can be measured by using the Doppler
effect with phase velocity of the wave crest. HF Doppler radar is already operational.
Chapter 5 Platforms
5.1 Types of Platform
The vehicle or carrier for remote sensors are borne is called the platform. Typical
platforms are satellite and aircraft, but they can also include radio controlled airplanes,
balloons, kites for low altitude remote sensing, as well as ladder trucks or "cherry pickers"
for ground investigations.
Table 5.1.1 shows various platforms, altitudes and objects being sensed. Platforms with the
highest altitude are geosynchronous satellites such as the Geosynchronous Meteorological
Satellite (GMS), which has an altitude of 36,000 km at the Equator. Most of the earth
observation satellites, such as Landsat, SPOT, MOS etc. are at about 900 km altitude with
a sun synchronous orbit.
From lower orbit, there are space shuttle (240-280 km), radio sonde ( - 100 km), high
altitude jet-plane ( 10,000 m), low or middle altitude plane (500-8,000 m), radio controlled
plane ( - 500 m) and so on.
57
The key factor for the selection of a platform is the altitude which determines the ground
resolution if IFOV (instantaneous field of view) of the sensor is constant, where
=H
The selection of platform also depends on the purpose which is sometime requested for
example a constant altitude is required for aerial surveys, while various altitudes are needed
to survey vertical atmospheric distribution, for example.
For aerial photogrammetry, the flight path is strictly controlled to meet the requirement of
geometric accuracy. However, helicopter or radio controlled planes are used for a free path
approach, for example in disaster monitoring.
5.2 Atmospheric Condition and Altitude
Atmospheric condition is different depending on the altitude. This factor must be
considered in the selection of platforms or sensors. In this section, air pressure, air density
and temperature are considered.
Dependence of air pressure on altitude is based on the hydro-static equilibrium of balance
between the vertical pressure of the atmosphere and gravity.
The atmospheric constituents without water vapor are assumed constant in volume ratio
with 78.08 % nitrogen, 20.95 % oxygen, and argon 0.93 % up to about 100 km regardless
of time and place. It gives an average molecular weight at 28.97 for the atmosphere and
the average molecular mass of 4.810 x 10
kg.
When temperature is constant with respect to altitude, the air pressure decreases as an
exponential function, which gives about an 8 km altitude for a decrease of air pressure to
1/e, as shown in Figure 5.2.1.
However, since the actual atmosphere varies in temperature with altitude as shown in
Figure 5.2.2, the air pressure can be calculated from the hydro-static equilibrium with a
given temperature.
58
For general purposes, the standard model atmosphere has been specified with respect to
the average temperature distribution and the vertical air pressure. Also the average model
with respect to latitude and season has been specified, although the actual temperature
sometimes has a difference of 10 - 20 K. Therefore the measurement of temperature using
radio-sonde is necessary for high accuracy. The vertical structure of the atmosphere is
composed of the following layers.
Troposphere : from the ground surface to 10 - 17 km,
Stratosphere : from 10 - 17 km to about 50 km
Mesosphere : from about 50 km to about 90 km
Thermosphere : from about 90 km to 500 km
The classification of the above layers depends on the distribution of thermal energy and
thermal transportation. The vertical decrease of temperature in the troposphere is 9.8 K km
for dry atmosphere, but 6.5 K km for the actual atmosphere because of water vapor. The
boarder between the troposphere and the stratosphere is called tropopause. The tropics
tropopause is rather constant at 17 km in altitude while the middle latitude tropopause
depends on seasonal change and jet stream with 10 - 17 km in altitude as shown in Figure
5.2.3.
5.3 Attitude of Platform
The geometric distortion depends on not only the geometry of the sensor but also the
attitude of the platform. Therefore it is very important to measure the attitude of the
platform for the consequent geometric correction.
The attitude of the platform is classified by the following two components.
a. Rotation angles around the three axes ; roll, pitch and yaw b. Jitter ; random and
unsystematic vibration which cannot be measured.
The rotation angles ; roll ( ), pitch ( ) and yaw ( ) are defined as the rotation angles around
the flight direction, the main wing and the vertical line respectively, as shown in Figure
5.3.1. Figure 5.3.2 show the satellite attitude parameters.
For a frame camera, the rotation angles are single values common to a full scene of aerial
photograph, while for a line scanner the attitude changes as a function of line number or
59
time. In the case of satellites, the variation of the position and the attitude will be
continuous, though in case of aircraft, the variation will not always be smooth, which
makes the geometric correction more difficult.
The typical attitude sensors for aircraft are as follows.
- Speedometer
- altimeter
- gyro compass (for attitude measurement)
- Doppler radar (for measurement of altitude)
- GPS (for positioning )
- gyro horizon
- TV camera
- flight recorder
The attitude sensors for satellites are introduced in section 5.4.
5.4 Attitude Sensors
Attitude control of a satellite is classified by two methods ; spin control and three axis
control. The former method is usually adopted for geosynchronous meteorological
satellites which itself rotates itself together with rotating scanner. The latter method is
mainly adopted for earth observation satellites such as Landsat which needs accurate look
angle in the direction of the earth.
The spin control is rather simple but has a low S/N ratio, while the three axis control is
more complex, but has a high S/N.
Figure 5.4.1 shows the typical types of attitude measurement sensors, which are used for
different purposes.
A gyro-compass is used for measurement of attitude variation over a short interval. Earth
sensor detects the radiation of CO2 with in the wavelength range of 14 - 16 mm emitted
from the rim of the earth, from which two axis attitude of roll and pitch can be measured
with an accuracy of 0.3 - 1 degree, as shown in Figure 5.4.2. If the earth sensor is combined
with a sun sensor and gyro-compass, the three axis attitude can be measured with higher
accuracy of 0.1 - 0.3 degree. Magnetic sensors can measure the three axis attitude but with
60
a slightly low an accuracy. The responsivity of the above sensors is 2 Hz at maximum. If
the high frequency attitude such as jitter is to be measured, the angular displacement
sensor (ADS) is necessary.
The angular displacement sensor of Landsat 4 and 5 has a responsivity of 2 - 18 Hz. The
highest accuracy of attitude can be achieved by the star sensor. For example, the standard
star tracker (SST) on board Landsat 4 and 5 will measure an accurate attitude from the
image of stars acquired by an image dissector with a reference of about 300 star catalogue
up to the sixth grade stars stored in an on board computer. The accuracy of SST is about
+- 0.03 degree (3 ).(3 standard deviations).
In case of the space shuttle, the star sensor has a lower accuracy with only about a 50 star
catalog compared to the SST, because the space shuttle does not need the higher attitude
control when it returns to the troposphere.
5.5 Orbital Elements of Satellite
A set of numerical values to define an orbit of a satellite or planet are called orbital
elements. The independent orbital elements of the earth observation satellite are six
elements of the Keplerian orbit.
A satellite can be considered to rotate around the earth in a plane, called the orbital plane,
because the influence of gravity of the moon and the sun can be neglected as compared
with the gravity of the earth.
A point in space can be expressed in the equatorial coordinate system as follows. The
origin of equatorial coordinate system is the center of the earth.
The reference great circle : the equatorial plane
The origin of astronomical longitude (right ascension) : the vernal equinox
The astronomical longitude (right ascension) : 0 - 24 hours to the east from the vernal
equinox
The astronomical latitude (declination) : angle from the equatorial plane ( +90 degree in
the north pole ; -90 degree in the south pole)
The six elements of Keplerian orbit are ;
61
(1) The semi-major axis (A):
(2) Eccentricity of orbit (e) :
(3) Inclination angle (i) :
(4) Right ascension of ascending node (h)
(5) Argument of perigee (g)
(6) Time of passage of the perigee ( )
Figure 5.5.1 shows the above elements. The shape and size of an orbit can be defined by A
and e, while the orbit plane can be defined by i and h. The longer axis of the orbit ellipse
can be determined by g. The position of a satellite can be located by T.
Sometime the orbital elements are replaced by three dimensional geocentric coordinates
and velocity for numerical analysis instead of the elements of the the Keplerian orbit.
Table 5.5.1 shows the relationship between the elements of the Keplerian orbit and the
geocentric.
5.6 Orbit of Satellite
The orbit of a satellite is referred to by several names with respect to orbit figure,
inclination, period and recurrence as shown in Figure 5.6.1.
The circular orbit is the most basic orbit and is explained as follows. The orbit can be
expressed as the polar coordinates (r, ).
r=re+hs = 0t
where re : radius of the earth 6,378,160 m
hs : altitude of satellite
t : time
0 : angular velocity
The angular velocity and the period are expressed as follows.
where : gravity constant ; 3.986005
a. Geosynchronous orbit
62
The orbit with the same earth rotation rate (h24 = the sidereal day ; 86164.1 sec) is called
an earth synchronous orbit or geosynchronous orbit. The geosynchronous orbit with an
inclination of i = 0 is called a geostationary orbit because the satellite looks stationary over
the equator from a ground surface view. As such, a geostationary satellite is useful for
covering wide areas. Many meteorological satellites and communication satellites are
geosynchronous types.
b. Sun synchronous orbit
Most earth observation satellites, such as Landsat, with lower altitudes have sun
synchronous and semi-recurrent orbits. The sun synchronous orbit can be defined as the
orbit in which the orbital plane rotates in a year in unison with the one revolution / year
apparent motion of the sun. The model precession rate , is a function of inclination i, orbit
altitude hs and orbital period T as shown in Figure 5.6.2. As seen in the figure, the sun
synchronous orbit has W=1 (revolution / year). For example, in case of i = 100 degree, the
altitude of the sun synchronous orbit is about 1,200 km with about 108 minutes of orbit
period. The advantage of the sun synchronous orbit is that the observation conditions can
be kept with a constant solar incident angle.
c. Semi-recurrent orbit
While the recurrent orbit can be defined as the orbit which returns to the same nadir point
in a day, the semi-recurrent orbit returns to the same nadir point in N days repetition (N>1),
which is much better for covering all of the earth than the recurrent orbit.
5.7 Satellite Positioning Systems
There are two methods for positioning a satellite: distance measurement from the ground
station, and GPS as shown in Figure 5.7.1. As GPS is explained in 6.8, the former method
will be explained here. The measurement of distance for satellite positioning is called the
range and range rate system, by which the time of a radio wave transmitted between the
ground station and a transponder onboard the satellite is measured with the Doppler
frequency. It enables the distance or range and the range rate to be measured. The accuracy
of the range and range rate is a few meters/second, respectively. The accuracy depends on
63
the parameters of frequency, signal, the location of the ground station, coordinate system,
time measurement system, reflection in the troposphere and the ionosphere etc.
The position of a satellite by the range and range rate method is limited only near the
ground station and is also discrete. In order to determine the satellite orbit in a time
sequential function, it is necessary to construct a model.
A primitive model is a parabolic curve based on the third theory of Kepler that defines
the motion of two bodies in space, termed the two body problem, under the law of
universal gravitation. The parabolic curve can be expressed with the six elements of the
Keplerian orbit.
However in reality there are influences from other planets, which results in departure from
the parabolic curve. This is called the perturbation.
There are two methods to determine the precision for an body problem (n>2); the numerical
integral calculation, which has high accuracy but is time consuming and the analytical
method with a lower accuracy but faster calculation, as shown in Table 5.7.1.
The satellite position in an orbit will be determined at certain time intervals, from which
the orbit at an arbitrary time can be interpolated by the least square method for high order
polynomials or by Lagrange's interpolation method.
5.8 Remote Sensing Satellites
A satellite with remote sensors to observe the earth is called a remote sensing satellite or
earth observation satellite. Meteorological satellites are sometimes discriminated from the
other remote sensing satellites.
Remote sensing satellites are characterized by their altitude, orbit and sensors. The main
purpose of the geosynchronous meteorological satellite (GMS) with an altitude of 36,000
km is meteorological observations, while Landsat with an altitude of about 700 km, in a
polar orbit, is mainly for land area observation.
NOAA AVHRR with an altitude of 850 km in a polar orbit is mainly designed for
meteorological observation but is also successfully used for vegetation monitoring.
In future some remote sensing satellites will have large payloads with many kinds of multipurpose sensors, such as the polar orbit platform (POP) project under the international
64
cooperation of US, EEC, Japan and Canada. As well, there will be more specialized
missions using small satellites.
Appendix-1 shows the Plan of Earth Observation Satellites up to the year 2,000. The details
of major satellites are shown in appendix-2.
Figure 5.8.1 shows the JERS-1(Japanese Earth Resource Satellite-1) spacecraft with SAR,
Visible and Near Infrared Radiometer (VNIR) and Short Wavelength Infrared Radiometer
(SWIR).
The important functions of a remote sensing satellite system include the following three
major systems.
a. Tracking and control system: determination of satellite orbit, orbital control, processing
of housekeeping data etc.
b. Operation control system: planning of mission operation, evaluation of observed data,
data base of processed data etc.
c. Data acquisition system: receiving, recording, processing, archiving and distribution of
observed data.
Figure 5.8.2 shows the total system of the JERS-1.
5.9 Landsat
Landsat-1 was launched by the USA in 1972, and was the first earth observation satellite
in the world, which initiated the remarkable advance of remote sensing. To date, five
Landsat's (Landsat 1-5) have been launched, with only Landsat 5 still in operation.
Figure 5.9.1 shows the general configuration of Landsat 4 and 5.
a. Orbit of Landsat 4,5 and 6
Altitude; 705 km, Inclination; 98 ,
Sun synchronous and semi-recurrent orbit,
Time of passage of the equator; 9:39a.m.,
Recurrent: 17 days
Swath: 185 km
65
b. Sensors
(1) MSS (multispectral scanner)
(2) TM (thematic mapper)
Both the sensors are optical-mechanical scanners.
Table 5.9.1 shows the bands, wavelength and resolution of MSS and TM. Landsat 6 will
have only ETM (enhanced thematic mapper) with an additional panchromatic mode with
15 meter resolution.
c. Data
MSS and TM data are composed in a unit of scene with a size of 185 x 170 km. Each scene
is coded with path number and row number, based on what is called WRS (world reference
system). For example, Japan is covered with about 63 scenes of path number 104 - 114 and
row numbers 28 - 42. Image data are recorded with respect to each pixel with a numerical
value (V) of 8 bits (0 - 255). The absolute radiance R (mW / cm .sr ) can be computed by
the following formula.
R = V[ ( Rmax - Rmin ) / Dmax ] + Rmin
where Rmax : maximum recorded radiance
Rmin : minimum recorded radiance
Dmax: 255 for TM
127 for MSS
Table 5.9.2 and Table 5.9.3 show Rmin and Rmax of TM and MSS respectively. One
should note that the radiances Rmax and Rmin are measured onboard but not on the ground.
Therefore they include atmospheric influences.
d. Data Utilization
There are 15 Landsat receiving stations in the world from which Landsat data are
distributed to users for resources management and environmental monitoring.
5.10 SPOT
SPOT was first launched in February, 1986 by the French Government. SPOT-2 was
launched in February, 1990 and is now in operation. SPOT-3 will be launched in 1993.
66
SPOT has two HRV (High Resolution Visible imaging system) sensors with stereoscopic
and oblique pointing functions. Figure 5.10.1 shows the general configuration of SPOT.
a. Orbit
Altitude; 830 Km, Inclination; 98.7°,
Sun synchronous and semi-recurrent orbit,
Time of passage of the equator; 10:30a.m.,
Recurrent : 26 days nominally but 4 - 5 days
if observed with oblique pointing.
b. Sensors
HRV is not an optical-mechanical sensor but a linear CCD (charge coupled device ) camera
with an electronic scanning system. Table 5.10.1 shows the HRV characteristics for the
three multi-spectral bands with 20 m IFOV, and a panchromatic mode with 10 m IFOV.
HRV can change the look angle by changing the pointing mirror angle by up to +-27
degrees, as shown in Figure 5.10.2. The enables it to look at the same position from two
different orbits as shown in Figure 5.10.3. Such a sidelooking function produces
stereoscopic images, with a baseline to height ratio (B/H ratio) of up to 1, for measurement
of topographic elevation.
c. Data
A scene of HRV has a nadir coverage of 60 x 60 km, but an oblique coverage of 81 km
square, at maximum look angle of 27 . Each scene is coded with column number (K) and
row number (J), termed the GRS (SPOT Grid Reference System).
Each node is basically given for a nadir observation with odd numbers of K for the coverage
of the first HRV sensor. For the oblique scene, the nearest node to the center of the scene
is assigned to that scene.
d. Data Utilization
SPOT data are received at 14 ground receiving stations. The main purpose of data
utilization is for land area observation as well as for topographic mapping at scales
1/50,000 and smaller.
67
Sometimes SPOT HRV Panchromatic band (10 m IFOV) and Landsat TM (30 m IFOV)
are combined into a color composite for better image interpretation. SPOT panchromatic
and multispectral modes are also often overlaid to aid in interpretation.
5.11 NOAA
The NOAA satellite series are the third generation of meteorological satellites operated by
the National Oceanic and Atmospheric Administration (NOAA), USA (see Figure 5.11.1).
The first generation was the TIROS series (1960 - 1965), while the second generation was
ITOS series (1970 - 1976). The NOAA series, the third generation, are listed in Appendix
2.
NOAA has a circular and sun synchronous orbit. The altitude is 870 km (NOAA-11) and
833 km (NOAA-12) with inclination of 98.7 degree (NOAA-11) and 98.9 degree (NOAA12)to the equator. The orbital period is 101.4 minutes.
As the NOAA series are operational for meteorological observation, two NOAA satellites
(currently NOAA-11 and NOAA-n) are in operation. A NOAA satellite can observe the
same are area twice a day (day and night), so that the two satellite can cover the same area
four times a day. Figure 5.11.2 shows the flyover times of NOAA-11 and NOAA-12 over
Japan.
The major sensors of NOAA are AVHRR/2 (Advanced Very High Resolution Radiometer;
model 2) with a 1.1 km IFOV for a swath of 2,800 km, and TOVS (TIROS Operational
Vertical Sounder ) including HIRS/2 (High Resolution Infrared Sounder; model 2) with
20 km IFOV, for a 2,200 km swath, SSU ( Stratospheric Sounding Unit ) with 147 km
IFOV, for a 736 km swath and MSU (Microwave Sounding Unit) with 110 km IFOV, for
a 2,347 km swath.
Table 5.11.1 shows the characteristics of AVHRR/2, while Table 5.11.2 shows those of the
TOVS
5.12 Geostationary Meteorological Satellites
68
Geostationary meteorological satellites are launched under the WWW (World Weather
Watch) project organized by the WMO (World Meteorological Organization), which will
cover all the earth with five satellites as shown in Figure 5.12.1.
The five geostationary meteorological satellites are METEOSAT (ESA), INSAT (India),
GMS (Japan), GOES-E (USA) and GOES-W (USA). The schedules for these satellites
are shown in Appendix 1.
As of 1991, METEOSAT-5, INSAT-1D, GMS-4 and GOES-7 are in operation.
GMS-4 has a sensor called VISSR (Visible and Infrared Spin Scan Radiometer) with two
bands of visible and thermal infrared. The VISSR scans four lines for the visible band and
a line for the thermal band simultaneously from the north to the south, which takes 25
minutes to cover the semi-sphere as shown in Figure 5.12.2.
The total scan lines are 10,000 lines for the visible band and 2,500 lines for the thermal
band.
GMS has a data collection platform (DCP) system to collect various information, not only
from the ground station, but also from the stations on the sea as shown in Figure 5.12.3.
The image data are transmitted to the ground station in a high resolution mode of S-VISSR
signals, and also in a low resolution mode of WEFAX, which can be received by cheaper
and simpler receiving facilities. Some statistical data such as histograms, cloud volumes,
sea surface temperatures, wind distribution and so on, are recorded in the archives
including the ISCCP (International Satellite Cloud Climatology Project) data set.
5.13 Polar Orbit Platform
The Polar orbit platform (POP) is a newly designed system for the 21st century, intended
to establish a longer life space infrastructure with multiple sensors as well as multiple uses,
when compared with the existing satellites which are used for a limited period and purpose.
POP is composed of a main space station, a space shuttle and an inter-orbital vehicle as
shown in Figure 5.13.1, by which exchange of mission equipments and repair of the form
will be possible.
69
POP is made of a module structure with ORU (Orbital Replacement Unit) for replacement
of mission parts and battery. Such functions of POP will make the system large in size and
payload, but long in life.
Japanese ADEOS (Advanced Earth Observation Satellite) as shown in Table 5.13.1 is not
POP a but is designed for a future type platform of earth observation satellite with the
function of data relay. ADEOS is characterized by the multiple sensors of OCTS (Ocean
Color and Temperature Scanner by NASDA), AVNIR (Advanced Visible and Near
Infrared Radiometer by NASDA), and AO (Applications of Opportunity) sensors such as
NSCAT, TOMS, IMG, POLDER, and ILAS .
At present the space station project is delayed and cut back to a smaller system without
orbital services. For example, NASA EOS-a and b, ESA POEM-1 and 2 will be launched
at the end of the 20th century for earth environmental monitoring.
Chapter 6 Data used in Remote
Sensing
6.1 Digital Data
Images with a continuous gray tone or color, like a photograph are called analog images.
On the other hand, a group of divided small cells, with integer values of average intensity,
the center representing the cell's value, is called a digital image. The spatial division into
a group of cells is called sampling as illustrated in Figure 6.1.1, while conversion of analog
images into integer image data is called quantization as illustrated in Figure 6.1.2 and
6.1.3.
70
An individual divided cell is called a pixel (picture cell). The shape of the cell is usually
square for easy use in a computer, though triangular or hexagonal can also be considered.
A digital image has coordinates of pixel number, normally counted from left to right, and
line number, normally counted from top to bottom.
The most important factor in sampling is pixel size or sampling frequency. If the pixel size
is large or the sampling frequency is long, the appearance of the image becomes worse,
while in the reverse case the data volume becomes very large. Therefore the optimum
sampling should be carefully considered.
Shannon's sampling theorem, for specifying the optimum sampling, is given as follows.
"There will be no loss of information if sampling is taken with a half frequency of the
maximum frequency involved in the original analog frequency wave."
Let the analog intensity be f and the unit intensity v(>0) as divider in quantization. Let the
quantized intensity be fd, fd is given by n as illustrated in Figure 6.1.2. The difference
between f and fd is called quantization error.
The question is how to determine the number of quantization levels or the unit intensity as
divider. If the number of levels is too small, the quantization error will increase. In the
reverse, the data volume increases with informationless data because of the noise level, as
shown in Figure 6.1.3.
For example in Figure 6.1.3, the quantization should be divided by a level larger than that
of the noise. In this example, four levels would be an appropriate quantization.
6.2 Geometric Characteristics of Image Data
Remote sensing data are data digitized by a process of sampling and quantization of the
electro-magnetic energy which is detected by a sensor. In this section, geometric
characteristics of sampling are described, while radiometric characteristics by quantization
are explained in 6.3.
IFOV (Instantaneous Field Of View) is defined as the angle which corresponds to the
sampling unit as shown in Figure 6.2.1. Information within an IFOV is represented by a
pixel in the image plane.
71
The maximum angle of view which a sensor can effectively detect the electro magnetic
energy, is called the FOV (Field Of View). The width on the ground corresponding to the
FOV is called the swath width.
The minimum detectable area, or distance on the ground is called the ground resolution.
Sometimes the projected area on the ground corresponding to a pixel or IFOV is also called
the ground resolution.
In remote sensing, the data from a multiple number of channels or bands which divide the
electromagnetic radiation range from Ultra Violet to Radio Waves are called multichannel data, multi-band data or multi- spectral data.
In general, multi-channel data are obtained by different detectors as shown in Figure 6.2.2.
Because the detectors are located at slightly different positions, and the light path of
different wavelengths is a little different from each other, the images of multi-channel data
are not identical in geometric position. To correct such geometric errors between channels
is called registration.
The term registration is also used for registration of multi-temporal (or multi-date) images.
6.3 Radiometric Characteristics of Image Data
Electromagnetic energy incident on a detector is converted to an electric signal and then
digitized. In this quantization process, the relationship between the input signal and the
output signal is generally represented as shown in Figure 6.3.1. In this curve the left part
corresponds to the insensitive area, with less response, while the right part is the saturated
area with almost constant output regardless of the input intensity.
In the central part, there is almost a linear relationship between the input and the output.
The approximation to a linear relationship is called linearity. The range of the linear part
or the ratio of maximum input to minimum input is called the dynamic range, which is
usually expressed in dB (see 2.2).
One should be careful of the noise level in the case of quantization, as explained in 6.1.
The ratio of effective input signal S to the noise level N is called the S/N ratio (signal to
noise ratio), which is given as follows.
S / N ratio = 20 log10 (S/N) [dB]
72
In conclusion, quantization is specified by the dynamic range and the S/N ratio.
Information contained in digitized image data are expressed by bit (binary digit) per pixel
per channel.
A bit is a binary number, that is 0 or 1. Let the quantization level be n, then the information
in terms of bits is given by the following formula.
log2 n (bit)
In remote sensing, the quantization level is normally 6, 8 or 10 bits as shown in Table 6.3.1.
For computer processing, the unit of byte (1 byte = 8 bits;integer value 0-255 ; 256 gray
levels) is much more convenient. Therefore remote sensing data will be treated as one or
two byte data.
The total data volume of multi-channel data per scene is computed as follows.
Data Volume(byte) = (line number) x (pixel number) x (channel number) x (bits) /8
Output data usually corresponds to the observed radiance detected by the sensor. The
absolute radiance is converted by a linear formula from the observed radiance (see 9.1).
The parameters are usually listed in the User's Manual for the particular remote sensing
system.
6.4 Format of Remote Sensing Image Data
Multi-band image data are represented by a combination of spatial position (pixel number
and line number) and band.
The data format for multi-band images is classified into the following three type, as shown
in Figure 6.4.1.
a) BSQ format (band sequential) image data (pixel number and line number) of each band
are separately arranged.
b) BIL format (band interleaved by line) line data are arranged in the order of band number
and repeated with respect to line number.
c) BIP format (band interleaved by pixel) A set of multi-band data with respect to each
pixel arranged spatially by pixel number and line number.
73
For color image output, BSQ format would be convenient because three bands will be
assigned to R(red), G(green) and B(blue). However BIP format would be better for
classification by a maximum likelihood classifier because multi-band data are required
pixel by pixel for the multi-variable processing. BIL would be a compromise between BSQ
and BIP.
Remote sensing data usually includes various annotation data in addition to image data.
Since 1982, satellite image data have been provided in a standard format called World
Standard Format, or LTWG format (specified by Landsat Technical Working Group).
The World Standard Format has the data structure called super structure with three
records of volume descriptor, file pointer and file descriptor which describe the contents of
the data (see 6.5).
Either BSQ or BIL format is chosen in the World Standard Format.
6.5 Auxiliary Data
An image scene is composed of multiple files, each of which is composed of multiple
records. Data other than image data in the files, are called auxiliary data.
The auxiliary data involves descriptions of file, image data, platform, sensor, data
processing and other data, including telemetry.
Figure 6.5.1 shows the basic configuration of the World Standard Format or LTWG format
with the files and the records. The files and the records in the figure are as follows.
Reader file: header record, ancillary record, annotation record etc.
Image file: image record (line information and spectral data)
Trailer file: trailer record (quality of data)
Supplemental file: information on satellite, detector, data correction etc.
Table 6.5.1 shows the contents of auxiliary data. There is a text record in which any
comment can be described. In the LTWG format, the text record is located in the volume
directory file.
The LTWG format has no fixed specification on the content of each record, while the
CEOS Format (Committee on Earth Observation Satellite) specifies the standard content
of a record, which will be utilized more than LTWG in the future.
74
6.6 Calibration and Validation
Remote sensing data involves many radiometric errors resulting from sensitivity of
detector, atmospheric condition, alignment of detectors and so on. Calibration is defined
as the correction of the observed data, or relationship, into physically meaningful data, or
relationship, by using a reference. For example, calibration involves the correction of
observed data into absolute irradiance, reflectance or actual temperature.
Calibration can be classified into two types; ground calibration and on-board calibration,
as shown in Table 6.6.1. The ground calibration data are measured before launch with a
halogen lamp for visible and reflective infrared, and a black body for thermal infrared,
which are normally described in the User's Manual.
The on-board calibration data are obtained on board after launch with on board references
such as lamp and blackbody as well as physically known or constant objects such as
sunlight, shadows on the ground and space with low temperature. The on board calibration
data are transmitted from the satellite to ground receiving stations, together with the image
data.
Table 6.6.2 shows the three calibration levels; interband calibration, band-to-band
calibration, and absolute calibration.
In the case of NOAA AVHRR data, the ground calibration data are used for calibrating
visible and near infrared data, while on-board calibration data are used for calibrating
thermal data. Twelve lamps provide the ground calibration data, by which image data can
be converted to Albedo. Thermal data can be converted to brightness temperature with the
two reference temperature data of space (-270 C) and black body (l5 C) measured by a
platinum resistance thermometer (see Figure 2.10.1).
The brightness temperature obtained after calibration involves atmospheric influences.
Therefore atmospheric correction is necessary. There are two types of atmospheric
correction; a theoretical approach using an atmospheric model, and an empirical approach
with ground truth data which are measured simultaneously with the satellite orbit.
In the latter case, so called validation data should be collected as ground data, for example,
observed sea surface temperature from boats and buoys.
75
Validation is classified into three types; instrument specification, physical quantities and
practical usages as shown in Table 6.6.3. Nevertheless, validation should be linked with
ground data as explained in the next section 6.7.
6.7 Ground Data
Ground data, in some cases called ground "truth" is defined as the observation,
measurement and collection of information about the actual conditions on the ground in
order to determine the relationship between remote sensing data and the object to be
observed. Investigation on the sea is sometimes called sea truth. Generally ground data
should be collected at the same time as data acquisition by the remote sensor, or at least
within the time that the environmental condition does not change. It should not be inferred
that the use of the word "truth" implies that ground truth data is not without error. Ground
data is used as for sensor design, calibration and validation, and supplemental use, as shown
in Figure 6.7.1.
For the sensor design, spectral characteristics are measured by a spectrometer to determine
the optimum wavelength range and the band width.
For supplemental purposes, there are two applications; analysis and data correction. The
former case, for example, is ground investigation, at a test area, to collect training sample
data for classification. The latter case, for example, is a survey of ground control points for
geometric correction.
The items to be investigated by ground data are as follows.
a. Information about the object type, status, spectral characteristics, circumstances, surface
temperature etc.
b. Information about the environment, the sun azimuth and elevation, irradiance of the sun,
atmospheric clarity, air temperature, humidity, wind direction, wind velocity, ground
surface condition, dew, precipitation, etc.
Depending on the purpose, the above items and the time of ground investigation should be
carefully selected.
76
Ground data will mainly include identification of the object to be observed, and
measurement by a spectrometer, as well as visual interpretation of aerial photographs and
survey by existing maps, and a review of existing literature and statistics.
Figure 6.7.2 shows data collection from various altitudes including ground data.
As the collection of ground data is time consuming as well as expensive, it is best to
establish a test site for sensor design, calibration and validation, and data correction. The
test area should be carefully selected with respect to ease of survey, variety of features
present, weather condition and so on.
6.8 Ground Positioning Data
In order to achieve accurate geometric correction, ground control points with known
coordinates are needed. The requirements of ground control points are that the point should
be identical and recognizable both on the image and on the ground or map, and its image
coordinates (pixel number and line number) and geographic coordinates (latitude,
longitude and height), should be measurable.
Use of a topographic map is the easiest way to determine the position of ground control
point. However maps are not always available, especially in developing countries. In such
cases, control surveys had previously been required.
Today, however GPS (global positioning system) can provide geographic coordinates in a
short time using a GPS receiver to measure time information from multiple navigation
satellites.
GPS is a technique, used to determine the coordinates of a GPS receiver which receives
radio signals from more than four navigation satellites. The received navigation message
includes exact time and orbit elements which can be converted into the satellite position.
Two methods can be used for positioning; single point positioning and relative positioning.
The single point positioning method determines the coordinates with the use of a single
GPS receiver, as shown in Figure 6.8.1. The geodetic accuracy achieved is about 10-30
meters. The unknown variables are four; X0, Y0, Z0 and t (clock-timing error of a
receiver). Therefore at least four navigation satellites are necessary. GPS has 18 satellites
77
in total, at an altitude of 20,000 km, with three satellites each in six different orbits, which
enable any point on the earth to view at least four satellites.
The relative positioning method determines the relative relationship between a known point
and an unknown point to the measured (see Figure 6.8.2). In this case, at least two GPS
receivers should be located at the same time. The accuracy is 0.1-1 ppm of the base length
between a known point and an unknown point. It is about 2-5 cm in planimetric accuracy
and 20-30 cm in height accuracy
6.9 Map Data
In remote sensing, the following maps are needed for particular objectives. Given below
are requirements for satellite remote sensing. For airborne sensing one usually requires the
larger scaled maps.
a. Topographic map:
1.25,000 or 1/50,000 topographic map will be best used to select ground control points and
to extract DEM (digital elevation model) for digital rectification or the generation of a three
dimensional view.
b. Thematic maps;
Land use, forest, soil, and geological maps etc. are used to collect training data for
classification. A map scale of 1/50,000 - 1/250,000 is best for this purpose. The thematic
maps can be digitized to permit integration of remote sensing data into geographic
information systems (GIS) containing the thematic information.
c. Socio-economic maps;
Political units, transportation network, population distribution, agricultural and industrial
census, tax or land price, and so on, are important factors for remote sensing applications
and GIS.
Table 6.9.1 summarizes the required maps for remote sensing and GIS. Global change
monitoring with the use of NOAA AVHRR, Nimbus CZCS or geosynchronous
meteorological satellites is important for earth environmental analysis. In such cases, world
maps which cover the whole earth may be necessary as a reference. Up to now, United
78
Nations organizations such as UNESCO, UNEP, UNFAO etc. as well as NASA, NOAA
and other international organizations, have produced various world maps.
Table 6.9.2 shows the UN statistics for world topographic mapping as of 1987 with respect
to the map scale and region. As seen in this table, topographic maps of 1:50,000 scale have
been completed for only about 60 % of the total earth's land area, with especially low
coverage in Africa, South America and Oceania. This is one reason why high resolution
stereo image data are required for topographic mapping from space, and why radar imagery
is attractive.
6.10 Digital Terrain Data
Digital terrain data are topographic data, including ground height or elevation, slope
(gradient and slope aspect), types of slope etc., which are called DTM (Digital Terrain
Model) or DEM (Digital Elevation Model).
Terrain features can be expressed using the following four methods.
1) Contour Lines.
Usually elevations on a topographic map are represented as a group of contour lines with
a discrete and constant contour interval.
2) Grid data.
For convenience of computer processing, a set of grid data with elevation are acquired from
contour maps, aerial photographs or stereo satellite image data, as shown in Figure 6.10.1.
Terrain data other than the grid data are interpolated from the surrounding grid data.
3) Random point data.
Terrain features are sometimes represented by a group of randomly located terrain data
with three dimensional coordinates. For computer processing, random point data are
converted to triangulated irregular network (TIN) as shown in Figure 6.10.2. TIN has the
advantage of easy control of point density according to the terrain feature, though it has the
disadvantage of being time consuming in the random search for the terrain point.
4) Surface function.
79
Terrain surface can be expressed mathematically as a surface function, for example, a
Spline function.
A DEM can be generated by the following two methods.
1) Survey and photogrammetry
A ground survey is implemented using a total station, with a function of digital output,
giving a high accuracy over a comparatively narrow area. Aerial photogrammetry can be
executed by a digital plotter with a function of automated image matching. The digital 3D
coordinates will be automatically generated. Stereo remote sensing data from space will be
a powerful tool to produce 1:50,000 topographic maps in the near future.
2) DTM generation from contour maps
Contour lines are measured by a tablet digitizer manually, or by a scanner automatically or
semi-automatically, to generate the DEM.
The DEM is used for generating a digital orthophotomap and a 3-D view as well, for terrain
analysis in geomorphology and geological studies.
6.11 Media for Data Recording, Storage and Distribution
Generally satellite data received at a ground station are recorded in real time into HDDT
(high density digital tape) with 14 or 28 tracks. Depending on the requests, HDDT data
will be transferred to CCT (computer compatible tape) with 9 tracks and/or other media
for distribution. Recently optical disks for examples, WORM (write once read many ),
MO disk (magneto-optical disk with erasable function ) and CD-ROM (compact disk read only memory ) are becoming popular.
These media are characterized by the following factors.
a. Memory capacity: total memory in byte
b. Cost: cost of media, reader and unit data volume (cost per 1 MB)
c. Compatibility: for data formats and computer systems
d. Portability: size and weight e. Durability: years of life
The type of media should be selected depending on the purpose, in consideration of the
above items. Table 6.11.1 shows the characteristics of major media used in remote sensing.
80
For use in data centers the factors of data storage, portability, cost and durability are more
important than compatibility. Recently DAT (digital audio tape) or 8 mm cartridge tape is
replacing HDDT and CCT because of its compact size.
For distribution to public users, compatibility is most important, which makes CCT and
floppy disk more popular. CD-ROM is very convenient and also low cost as mass media,
similar to a music record disk. Optical disks such as MO and WORM are very attractive,
though standardization and compatibility are not yet fully implemented. However the
optical disk has the big advantages of low cost of both media and driver, and a large
memory capacity, especially as a large auxiliary supplementary memory for personal
computers or work stations.
6.12 Satellite Data Transmission and Reception
Transmitted data from remote sensing satellites involve not only image data but also
telemetry data including temperature, electric voltage and electric current of various
onboard equipment. Such data are usually transmitted as a digital signal in the form of
PCM (pulse code modulation ) with a binary pulse because the digital signal has the
advantages of being noise proof, requiring less electric power and having available narrow
radio bands. As the data volume or rate of transmission is very high, high frequency bands,
such as S band or X band ranging from several GHz to several tens of GHz, are used to
achieve the high rate of transmission.
These data are generally received by direct reception at a ground station. However this
direct method is limited to reception only when the satellite is in view, nominally several
degrees over the horizon, but usually above the horizons will suffice.
There are two methods used to record the satellite data at other areas outside the look angle;
these are MDR (mission data recorder) and TDRS (tracking and data relay satellite).
MDR can record other data from areas other than the covering area of the ground station,
and replay the data when the satellite flies over the ground station. For examples, NOAA,
SPOT, JERS-1 etc. have the MDR system.
81
TDRS's have been launched by NASA over the equator at 41 W and 171 W , which can
cover the whole of the earth tracking a lower altitude satellite and relaying the data to the
ground station, located at White Sands, United States, as shown in Figure 6.12.1.
Landsat 4 and 5 are linked to TDRS. Table 6.12.1 shows the reception method for areas
outside those covered by receiving stations.
6.13 Retrieval of Remote Sensing Data
Data received by remote sensing satellites are normally purchased or made available at the
operating space agencies, receiving stations or data distribution centers. Appendix Table 4
shows the main data distributors in the world. Searching for a satellite scene from a number
of scenes is so complicated that retrieval systems with key words of satellite name, sensor,
observed data, path and row number, cloud coverage etc. are made available for users.
In addition to such retrieving systems, which have been developed by each center, a
comprehensive and international directory system being developed by the CEOS
(Committee on Earth Observation Satellite ). CEOS-PID (prototype international
directory) is a world wide database, which is based on the master directory developed by
NASA. This is a directory database to indicate what kind of satellite data are available at
which center. Some data can be directly accessible on line to the data inventory of the
center of interest.
Figure 6.13.1 shows the main menu of CEOS-PID. At present CEOS-PID has an
international network, as shown in Figure 6.13.2. The users can utilize CEOS-PID by
telephone line from any node of the network
82
Chapter 7 Image Interpretation
7.1 Information Extraction in Remote Sensing
Information extraction in remote sensing can be categorized into five types as shown in
Table 7.1.1. Classification is a type of categorization of image data using spectral, spatial
and temporal information. Change detection is the extraction of change between multidate images. Extraction of physical quantities corresponds to the measurement of
temperature, atmospheric constitutients, elevation and so on, from spectral or stereo
information. Extraction of indices is the computation of a newly defined index, for
example, the vegetation index (see 10.6) from satellite data. Identification of specific
83
features is the identification, for example, of disaster, lineament, archaeological and other
features, etc.
Information extraction can be made by human or computer methods. Information
extraction by human interpretation will be described in the next sections, while
information extraction by computer will be explained in chapter 8.
Table 7.1.2 provides a comparison between human and computer information extraction.
As seen in the table, human and computer methods supplement each other, so that they
both may offer better results when combined. For example in geology, computers will
produce an enhanced image, from which humans can interpret the geological features.
A computer system with an interactive graphic display through which humans and
computers can interactively work together is called "a man-machine interactive system".
Because human interpretation is time consuming, as well as expensive, a special computer
technique, with the ability of human interpretation, is being developed. For example, an
expert system is a computer software system with a training ability to use the interpreter's
knowledge for information extraction.
7.2 Image Interpretation
Image interpretation is defined as the extraction of qualitative and quantitative
information in the form of a map, about the shape, location, structure, function, quality,
condition, relationship of and between objects, etc. by using human knowledge or
experience. As a narrow definition, " photo-interpretation " is sometimes used as a
synonym of image interpretation.
Image interpretation in satellite remote sensing can be made using a single scene of a
satellite image, while usually a pair of stereoscopic aerial photographs are used in photointerpretation to provide stereoscopic vision using, for example, a mirror stereoscope. Such
a single photo-interpretation is discriminated from stereo photo-interpretation (see 7.3).
Figure 7.2.1 shows a typical flow of the image interpretation process.
Image reading is an elemental form of image interpretation. It corresponds to simple
identification of objects using such elements as shape, size, pattern, tone, texture, color,
84
shadow and other associated relationships. Image reading is usually implemented with
interpretation keys with respect to each object, as explained in 7.4 and 7.5.
Image measurement is the extraction of physical quantities, such as length, location,
height, density, temperature and so on, by using reference data or calibration data
deductively or inductively.
Image analysis is the understanding of the relationship between interpreted information
and the actual status or phenomenon, and to evaluate the situation.
Extracted information will be finally represented in a map form called an interpretation
map or a thematic map.
Generally the accuracy of image interpretation is not adequate without some ground
investigation. Ground investigations are necessary, first when the keys are established and
then when the preliminary map is checked.
7.3 Stereoscopy
A pair of stereoscopic photographs or images can be viewed stereoscopically by looking
at the left image with the left eye and the right image with the right eye. This is called
stereoscopy. Stereoscopy is based on Porro-Koppe's Principle that the same light path will
be generated through an optical system if a light source is projected onto the image taken
by an optical system. The principle will be realized in a stereo model if a pair of
stereoscopic images are reconstructed using the relative location or tilt at the time the
photography was taken. Such an adjustment is called relative orientation in
photogrammetric terms. The eye-base and the photo-base must be parallel in order to view
at a stereoscopic model, as shown in Figure 7.3.1.
Usually a stereoscope is used for image interpretation. There are several types of
stereoscope, for example, portable lens stereoscope, stereo mirror scope (see Figure 7.3.2),
stereo zoom transfer scope etc.
The process of stereoscopy for aerial photographs is as follows. At first the center of both
aerial photographs, called the principal point, should be marked. Secondly the principal
point of the right image should be plotted in its position on the left image. At the same time
the principal point of the left image should be also plotted on the right image. These
85
principal points and transferred points should be aligned along a straight line, called the
base line, with an appropriate separation (normally 25-30 cm in the case of a stereo mirror
scope ) as shown in Figure 7.3.3. By viewing through the binoculars a stereoscopic model
can now be seen.
The advantage of stereoscopy is the ability to extract three dimensional information, for
example, classification between tall trees and low trees, terrestrial features such as height
of terraces, slope gradient, detailed geomorphology in flood plains, dip of geological layers
and so on.
The principle of height measurement by stereoscopic vision is based on the use of parallax,
which corresponds to the distance between image points, of the same object on the ground,
on the left and right image. The height difference can be computed if the parallax
difference is measured between two points of different height, using a parallax bar, as
shown in Figure 7.3.3.
7.4 Interpretation Elements
The following eight elements are mostly used in image interpretation; size, shape, shadow,
tone, color, texture, pattern and associated relationship or context.(see Figure 7.4.1[Size,
Shape, Shadow, Tone], Figure 7.4.2[Texture, Pattern]
(1) Size:
A proper photo-scale should be selected depending on the purpose of the interpretation.
Approximate size of an object can be measured by multiplying the length on the image by
the inverse of the photo-scale.
(2) Shape:
The specific shape of an object as it is viewed from above will be imaged on a vertical
photograph. Therefore the shape looking from a vertical view should be known. For
example, the crown of a conifer tree looks like a circle, while that of a deciduous tree has
86
an irregular shape. Airports, harbors, factories and so on, can also be identified by their
shape.
(3) Shadow:
Shadow is usually a visual obstacle for image interpretation. However, shadow can also
give height information about towers, tall buildings etc., as well as shape information from
the non-vertical perspective-such as the shape of a bridge.
(4) Tone:
The continuous gray scale varying from white to black is called tone. In panchromatic
photographs, any object will reflect its unique tone according to the reflectance. For
example dry sand reflects white, while wet sand reflects black.In black and white near
infrared infrared photographs, water is black and healthy vegetation white to light gray.
(5) Color:
Color is more convenient for the identification of object details. For example, vegetation
types and species can be more easily interpreted by less experienced interpreters using
color information. Sometimes color infrared photographs or false color images will give
more specific information, depending on the emulsion of the film or the filter used and the
object being imaged.
(6) Texture:
Texture is a group of repeated small patterns. For example homogeneous grassland exhibits
a smooth texture, coniferous forests usually show a coarse texture. However this will
depend on the scale of the photograph or image.
(7) Pattern:
Pattern is a regular usually repeated shape with respect to an object. For example, rows of
houses or apartments, regularly spaced rice fields, interchanges of highways, orchards etc.,
can provide information from their unique patterns.
(8) Associated relationships or context:
A specific combination of elements, geographic characteristics, configuration of the
surroundings or the context of an object can provide the user with specific information for
image interpretation
87
7.5 Interpretation Keys
The criteria for identification of an object with interpretation elements is called an
interpretation key. The image interpretation depends on the interpretation keys which an
experienced interpreter has established from prior knowledge and the study of the current
images. Generally, standardized keys must be established to eliminate the differences
between different interpreters.
The eight interpretation elements (size, shape, shadow, tone, color, texture, pattern and
associated relationship), as well as the time the photograph is taken, season, film type and
photo-scale should be carefully considered when developing interpretation keys. Keys
usually include both a written and image component.
Table 7.5.1 shows an example of interpretation keys for forestry mapping which have been
standardized by the Japan Association for Forestry.
The keys are specified with respect to the crown's shape, rim shape of the crown, tone,
shadow, projected tree shape, pattern, texture and other factors.
Table 7.5.2 shows an example of an interpretation key for land cover mapping with Landsat
MSS images in the case of single band and false color images.
7.6 Generation of Thematic Maps
An image interpretation map is usually produced by transferring the interpreted
information to a base map which has been prepared in advance. The requirements of the
base map should be as follows.
(1) Proper map scale to enable appropriate presentation of interpreted information
(2)
Geographic
coordinate
system
to
establish
the
geographic
reference
(3) Basic map information to be printed in light tones as background which results in
enhancement of interpreted information
Normally a topographic map, plan map or orthophotomap is used as a base map.
A topographic map with a scale of 1:50,000, 1:100,000 or 1:250,000 is usually the
preferable base map for higher resolution satellite image interpretation.
For oceanographic purposes or marine science, charts with a scale of 1:50,000 to 1:500,000
should be used as the base map.
88
Orthophotomaps are more easily used by cartographers for the transfer of interpreted
information, particularly in the case of forest classification.
The methods of transfer of information to a base map, are as follows.
(1) Tracing
The interpreted image is traced on to a base map by overlaying on a light table.
(2) Optical projection
The interpreted image is projected via a lens and a mirror onto a base map. The optical
zoom transferscope or mirror projector is very useful for image interpretation. (see Figure
7.6.1)
(3) Grid system
Grid lines are drawn on both an image and a base map. Then the interpreted information in
a grid on the image is transferred to the corresponding grid on the map.
(4) Photogrammetric plotting
Aerial photographs are interpreted into a thematic map using a photogrammetric plotter.
(see Figure 7.6.2)
89
Chapter 8 Image Processing
Systems
8.1 Image Processing in Remote Sensing
Remotely sensed data are usually digital image data. Therefore data processing in remote
sensing is dominantly treated as digital image processing.
Figure 8.1.1 shows the data flow in remote sensing. Figure 8.1.2 shows the major data
processing techniques in remote sensing.
(1) Input data
There are two data sources; analog data and digital data. Digital data, for example
multispectral scanner data, is converted from HDDT (high density digital tape) to CCT
(computer compatible tape) for ease of computer analysis. Analog data for example, film
must be digitized by an image scanner or drum scanner into digital image data.
(2) Reconstruction / Correction
90
Reconstruction, restoration and/or correction of radiometry and geometry should be
undertaken in the process of preprocessing.
(3) Transformation
Image enhancement, spatial and geometric transformation and/or data compression is
normally required to generate a thematic map or database.
(4) Classification
Image features are categorized, which is called labeling in image processing, using those
techniques of learning, classification, segmentation and/or matching.
(5) Output
There are two output methods; analog output such as film or color copy, and digital output
in the form of a database, which is usually used as one of the layers of geographic data in
GIS (geographic information system).
8.2 Image Processing Systems
a. Hardware
There are two types of image processing hardware.
(1) Image processing system with specific image processor
An image processor with frame buffer is connected with a host computer as shown in
Figure 8.2.1. The image processor has the function of high speed image processing and
image input / output. The hardware system depends on what type of host computer
(personal computer, work station, mini-computer, general purpose computer etc.) is
selected and what the computer is used for.
(2) General purpose computer
The host computer has only a frame buffer, as shown in Figure 8.2.2. Therefore the image
processing is implemented by software developed or purchased by the users. Though the
transportability of the system is flexible, the size of software naturally becomes very large.
Usually a personal computer or a work station is selected as the host computer.
91
Recently network systems connecting server computers and front end computers, as
shown in Figure 8.2.3 have become popular.
b. Peripherals
Image processing systems need various peripherals, such as image input devices, to enable
A/D conversion, image output devices for image display, image recorder to produce
hardcopy, and image recording equipment to establish data archives (see 8.3, 8.4, 8.5, 8.6).
c. Software
The software of image processing has the following basic subsystems.
(1) Data input/output (reading and writing CCT etc.)
(2) Image display and operation (color output, image operation, image enhancement etc.)
(3) Reconstruction and correction (geometric correction, radiometric correction etc.)
(4) Image analysis (image transformation, classification etc.)
(5) Image output (hard copy, film recording etc.)
8.3 Image Input Systems
Image input systems are defined in this section as analog to digital (A/D) converters of
analog images. The image input system provides digital data which are the converted tone
or color of a film or photograph. In the case of a color image, the components of the three
primary colors (Red, Green and Blue) are digitized by using three color filters.
The function of an image input system depends on the following factors.
(1) Film size: allowable maximum size
(2) Spatial resolution: pixel size or dot per inch (DPI)
(3) Gray scale : number of bits or bytes
(4) Speed: speed of A/D conversion and data
(5) Environment : dark room or illumination
(6) Accuracy : allowable error of coordinates
(7) Type of image : transparency or reflective
Table 8.3.1 shows the comparison between five image input devices with respect to spatial
resolution, density resolution, positioning accuracy etc. Figure 8.3.1 shows the typical
mechanism of the five image input systems.
92
a. Mechanical scanner:
An image placed around a drum is scanned using the rotation of the drum and a shift of a
light source. Though the speed of scanning is not very high, it is widely used because the
spatial resolution and density resolution are very high. Recently laser beams have been
used as the light source which enables a faster speed.
b. Electronic image tube:
Electronic image tube such as a TV camera is used for A/D conversion of an image.
However, the spatial resolution, density resolution and positioning accuracy are low. The
advantages are its low price and ease of use.
c. CCD camera:
The electronic image tube is now being replaced by CCD cameras with higher spatial
resolution and positioning accuracy. These systems are compact and lightweight.
d. Linear array CCD camera:
A linear array CCD with very high resolution, for example 409 pixels per line is drived
mechanically to enable line, scanning on a flat table. The spatial resolution, density
resolution and positioning accuracy are very high, so that desk top scanners are becoming
popular.
e. Flying spot:
An illuminated spot on a CRT is projected onto a film, at a given coordinate, with high
speed. The density of the film can be digitized regularly as well as randomly depending on
the input coordinates. The disadvantage is that a dark room is required.
8.4 Image Display Systems
Image display is used for displaying digital image data into a visual color image as a tool
for real time "man-machine interface". An image display system consists of a frame buffer,
look up table, D/A converter and a display, as shown in Figure 8.4.1.
Frame buffer is an image memory to allow high speed reading of digital image data. The
size of the image memory is usually 512 x 512 to 2048 x 2048 picture elements.
93
Look up table is a pre-set function to enable conversion from an input signal to an output
signal in real time. Linear function, contrast enhancement function, gamma function, log
function etc. are mostly used, as shown in Figure 8.4.2. D/A converter converts digital
image data in a frame buffer to an analog video signal.
Figure 8.4.3 shows R, G, B separate type as the D/A output system, while Figure 8.4.4
shows the color map type. The former system has an independent frame buffer and look
up table with respect to R, G, B, which enables individual color control. Thus full color
images (256 x 256 x 256 = 16,787,216 colors) can be generated. The latter system has a
unified frame buffer and a R, G, B separate look up table, which only allows generation of
a limited number of colors, for example 256 colors in the case of a combination of 8 bits
of frame buffer and 8 bits to each of R, G, B.
There are several types of display; CRT, liquid crystal display, plasma display, for
example.
8.5 Hard Copy Systems
A hard copy system or image output system is used to produce an analog image on paper
or a film from digital image data. Depending on the system, the recording media,
resolution, gray level, output size, output rate, cost and stability, are different as compared
and shown in Table 8.5.1.
a. Silver halide photography
This is a so-called film recorder which enables the production of film products from digital
image data with a light source such as CRT and laser beam. There are two types; the drum
type and the flat bed type. The resolution and gray level are excellent. Recently thermal
developing systems have become operational replacing the chemical developing system.
b. Electro photographyThe negative image is firstly generated on a photo-sensitive drum.
Secondly toner is electronically placed according to the negative image. Thirdly the toner
is transcribed onto plane paper. The advantage is that the running cost is low. It is widely
used as an ordinary hard copy machine.
c. Electro static recorder
94
Electronically coated paper is firstly given an electric charge in a form of a negative, in
accordance with the dot pattern of the image. Toner is secondly placed electro-statically
dot by dot. The advantage is that a large size of output can be obtained in a short time with
a moderate cost. It is sometimes called an electro static dot printer or simply a dot printer.
d. Thermal transcripter
There are two types; the melting type and the sublimation type. The melting type
transcribes a melted ink layer using a thermal head onto plane paper with the use of a coated
wax type ink ribbon. As the gray levels, are few only a limited number of color outputs,
for example in the case of a classified color map, are available rather than continuous color
tones. The sublimation type heats a coated ink sheet ,using a thermal head into vaporized
ink, which is transcribed onto a coated paper. The gray level is so many that the image
quality is similar to that of a film. The disadvantage of both types is that the paper size is
limited because of the site of the ink sheet.
e. Ink jet recorder
A water drop of ink is ejected from a nozzle and is transcribed onto plane paper pixel by
pixel. The advantage is that the ink volume can be controlled to produce a continuous gray
level onto a large size of paper. The disadvantage is that nozzle maintenance is a problem,
because the nozzle hole is sometimes blocked due to irregular ink particles.
8.6 Storage of Image Data
As the volume of image data is generally large, a storage device with a large volume is
needed to record the original image data, as well as the results of image processing. The
volume capacity of the recording media has increased year by year, because of industrial
development. Table 8.6.1 shows the characteristics of eight different media (see 6.11).
a. Magnetic tape
Magnetic tape is most widely used with general purpose computers or minicomputers. The
data format is well standardized so that transportability is also guaranteed. The
disadvantage is that the size of the magnetic tape, as well as the magnetic tape unit, is so
big that the storage space becomes bulky.
b. Streamer Tape
95
Is a small sized of cartridge tape popular in personal computers (PC) or work stations (WS).
The disadvantage is its low speed of data transfer rate.
c. Digital audio tape (DAT)
Because its size is smaller than the streamer, and the capacity is bigger and the price is
lower, the DAT is becoming popular for PC and WS. The disadvantage is its low data
transfer rate.
d. 8 mm video tape
It is cheaper in price and bigger in storage capacity than DAT. The data transfer rate is not
very fast but is a little faster than streamer and DAT.
e. Optical tape
As the capacity is about 1 terabyte, the data transfer rate is more than ten times faster than
DAT (faster than magnetic disk), rewriting is possible and the device is exchangeable. The
optical tape is expected to be the new media for the next generation. While only a few
manufacturers can produce the device and the price is very expensive, data capacity and
the life of the tape make it economic for all large volume users since for less standard tape
is used.
f. Magneto-optical disk (MO-DISK)
The size is compact and the capacity is also large, similar to an ordinary hard disk. Because
rewriting is possible, exchange is available, the data transfer rate is much faster than tape
media, and the price is lower, this media is very popular for PC and WS.
g. Write once and read many optical disk (WORM)
As rewriting is impossible, the users are decreasing. However the capacity is a little
larger than a MO-DISK and the storage life is longer.
h. Floppy disk
Is the most popular storage for PC. The disadvantage is that the capacity is limited to a
few M bytes and the data transfer rate is slow. The advantages are its low price and data
exchangeability.
96
Chapter 9 Image Processing –
Correction
9.1 Radiometric Correction
As any image involves radiometric errors as well as geometric errors, these errors should
be corrected. Radiometric correction is to avoid radiometric errors or distortions, while
geometric correction is to remove geometric distortion.
When the emitted or reflected electro-magnetic energy is observed by a sensor on board an
aircraft or spacecraft, the observed energy does not coincide with the energy emitted or
reflected from the same object observed from a short distance. This is due to the sun's
azimuth and elevation, atmospheric conditions such as fog or aerosols, sensor's response
etc. which influence the observed energy. Therefore, in order to obtain the real irradiance
or reflectance, those radiometric distortions must be corrected.
Radiometric correction is classified into the following three types (see Figure 9.1.1.)
97
(1)
Radiometric
correction
of
effects
due
to
sensor
sensitivity
In the case of optical sensors, with the use of a lens, a fringe area in the corners will be
darker as compared with the central area. This is called vignetting. Vignetting can be
expressed by cos , where is the angle of a ray with respect to the optical axis. n is
dependent on the lens characteristics, though n is usually taken as 4. In the case of electrooptical sensors, measured calibration data between irradiance and the sensor output signal,
can be used for radiometric correction.
(2) Radiometric correction for sun angle and topography
a. Sun spot
The solar radiation will be reflected diffusely onto the ground surface, which results in
lighter areas in an image. It is called a sun spot. The sun spot together with vignetting
effects can be corrected by estimating a shading curve which is determined by Fourier
analysis to extract a low frequency component (see Figure 9.1.2).
b.Shading
The shading effect due to topographic relief can be corrected using the angle between the
solar radiation direction and the normal vector to the ground surface.
(3) Atmospheric correction
Various atmospheric effects cause absorption and scattering of the solar radiation.
Reflected or emitted radiation from an object and path radiance (atmospheric scattering)
should be corrected for.(see 9.2).
9.2 Atmospheric Correction
The solar radiation is absorbed or scattered by the atmosphere during transmission to the
ground surface, while the reflected or emitted radiation from the target is also absorbed or
scattered by the atmosphere before it reaches a sensor. The ground surface receive not only
the direct solar radiation but also sky light, or scattered radiation from the atmosphere. A
sensor will receive not only the direct reflected or emitted radiation from a target, but also
the scattered radiation from a target and the scattered radiation from the atmosphere, which
is called path radiance. Atmospheric correction is used to remove these effects(see
Figure 9.2.1 and Figure 9.2.2).
98
The atmospheric correction method is classified into the method using the radiative transfer
equation, the method using ground truth data and other methods.
a. The method using the radiative transfer equation
An approximate solution is usually determined for the radiative transfer equation. For
atmospheric correction, aerosol density in the visible and near infrared region and water
vapor density in the thermal infrared region should be estimated. Because these values
cannot be determined from image data, a rigorous solution cannot be used.
b. The method with ground truth data
At the time of data acquisition, those targets with known or measured reflectance will be
identified in the image. Atmospheric correction can be made by comparison between the
known value of the target and the image data (output signal). However the method can only
be applied to the specific site with targets or a specific season.
c. Other method
A special sensor to measure aerosol density or water vapor density is utilized together with
an imaging sensor for atmospheric correction. For example, the NOAA satellite has not
only an imaging sensor of AVHRR (Advanced Very high Resolution Radiometer) but also
HIRS (High Resolution Infrared Radiometer Sounder) for atmospheric correction.
9.3 Geometric Distortions of the Image
Geometric distortion is an error on an image, between the actual image coordinates and the
ideal image coordinates which would be projected theoretically with an ideal sensor and
under ideal conditions.
Geometric distortions are classified into internal distortion resulting from the geometry
of the sensor, and external distortions resulting from the attitude of the sensor or the shape
of the object.
Figure 9.3.1 schematically shows examples of internal distortions, while Figure 9.3.2
shows examples of external distortions.
Table 9.3.1 shows the causes of internal and external distortions and the types of
distortions.
99
9.4 Geometric Correction
Geometric correction is undertaken to avoid geometric distortions from a distorted image,
and is achieved by establishing the relationship between the image coordinate system and
the geographic coordinate system using calibration data of the sensor, measured data of
position and attitude, ground control points, atmospheric condition etc.
The steps to follow for geometric correction are as follows (see Figure 9.4.1)
(1) Selection of method
After consideration of the characteristics of the geometric distortion as well as the available
reference data, a proper method should be selected.
(2) Determination of parameters
Unknown parameters which define the mathematical equation between the image
coordinate system and the geographic coordinate system should be determined with
calibration data and/or ground control points.
(3) Accuracy check
Accuracy of the geometric correction should be checked and verified. If the accuracy does
not meet the criteria, the method or the data used should be checked and corrected in order
to avoid the errors.
(4) Interpolation and resampling
Geo-coded image should be produced by the technique of resampling and interpolation.
There are three methods of geometric correction as mentioned below.
a. Systematic correction
b. When the geometric reference data or the geometry of sensor are given or measured,
the geometric distortion can be theoretically or systematically avoided. For example, the
geometry of a lens camera is given by the collinearity equation with calibrated focal length,
parameters of lens distortions, coordinates of fiducial marks etc. The tangent correction for
an optical mechanical scanner is a type of system correction. Generally systematic
correction is sufficient to remove all errors.
c. Non-systematic correction
100
d. Polynomials to transform from a geographic coordinate system to an image coordinate
system, or vice versa, will be determined with given coordinates of ground control
points using the least square method. The accuracy depends on the order of the
polynomials, and the number and distribution of ground control points(see Figure
9.4.2).
e. Combined method
f. Firstly the systematic correction is applied, then the residual errors will be reduced
using lower order polynomials. Usually the goal of geometric correction is to obtain an
error within plus or minus one pixel of its true position(see Figure 9.4.3).
9.5 Coordinate Transformation
The technique of coordinate transformation is useful for geometric correction with ground
control points (GCP). The key points are contained in the following two selections.
a. Selection of transform formula
Depending on the geometric distortions, the order of polynomials will be determined.
Usually a maximum of a third order polynomials will be sufficient for existing remote
sensing images, such as LANDSAT. Table 9.5.1 shows the examples of available formulas.
b. Selection of ground control points
The number and distribution of ground control points will influence the accuracy of the
geometric correction. The number of GCP's should be more than the number of unknown
parameters as shown in Table 1, because the errors will be adjusted by the least square
method.
The distribution of GCP's should be random, but almost equally spaced including corner
areas. About ten to twenty points which are clearly identified both on the image and the
map should be selected depending on the order of the selected formula or the number of
unknown parameters. Figure 9.5.1 shows the comparison of accuracy with respect to
number and distribution of GCP's. The accuracy of geometric correction is usually
represented by the standard deviation (RMS), in pixel units, in the image plane as follows.
101
u : standard deviation in pixel number
v: standard deviation in line number
where
u = {ui-f(xi,yi)} /n
v = {vi-g(xi,yi)} /n
(ui, vi) : image coordinates of the i th ground control point
(xi , yi) : map coordinates of the i th ground control point
f ( xi , yi) : coordinate transformation from map coordinates to pixel number
g ( xi , yi) : coordinate transformation from map coordinate to line number
The accuracy should be usually within +- one pixel. If the error is larger than the
requirement, the coordinates on the image or map should be rechecked, otherwise the
formula should be reselected.
9.6 Collinearity Equation
The Collinearity equation is a physical model representing the geometry between a sensor
(projection center), the ground coordinates of an object and the image coordinates, while
the coordinate transformation technique as mentioned in 9.5 can be considered as a black
box type of correction. The collinearity equation gives the geometry of a bundle of rays
connecting the projection center of a sensor, an image point and an object on the ground,
as shown in Figure 9.6.1.
For convenience, an optical camera system is described to illustrate the principle. Let the
projection center or lens be 0 (X0, Y0, Z0), with rotation angles , , around X, Y and Z
axis respectively (roll, pitch and yaw angles), the image coordinates be p (x,y) and the
ground coordinates be P(X,Y, Z). The collinearity equation is given as follows-
102
where f: focal length of lens, and a1 to a9 are given by the following matrix relationship.
In the case of a camera, the previous formula includes six unknown parameters (X0,Y0,Z0
; , , ) which can be determined with the use of more than three ground control points
(Xi,Yi; Xi,Yi,Zi). The collinearity equation can be inversed as follows-
In the case of a flat plane (Z: constant), the formula coincides with the two dimensional
projection as listed in Table 9.5.1. The geometry of an optical mechanical scanner and a
CCD linear array sensor is a little different from the one of a frame camera. Only the cross
track direction is a central projection similar to a frame camera, while along track direction
is almost parallel (y=0) with a slight variation of orbit and attitude, as a function of time or
line number, of not more than a third order as follows.
X0 = X0(l) = X0 + X1 l+ X2 l + X3 l
Y0 = Y0(l) = Y0 + Y1 l+ Y2 l + Y3 l
Z0 = Z0(l) = Z0 + Z1 l+ Z2 l + Z3 l
0 = 0(l) = 0 + 1 l+ 2 l + 3 l
0 = 0(l) = 0 + 1 l+ 2 l + 3 l
0 = 0(l) = 0 + 1 l+ 2 l + 3 l
where l is line number.
9.7 Resampling and Interpolation
In the final stage of geometric correction a geo-coded image will be produced by
resampling. There are two techniques for resampling as shown in Figure 9.7.1, and given
as follows(1) Projection from input image to output image
103
Each pixel of the input image is projected to the output image plane. In this case, an image
output device with random access such as flying spot scanner is required.
(2) Projection from output image to input image
Regularly spaced pixels in the output image plane are projected into the input image plane
and their values interpolated from the surrounding input image data. This is a more general
method.
Usually the inverse equation to transform from the output image coordinate system to the
input image coordinate system, is not possible to determine because the geometric equation
is very complex. In such a case, the following methods can be adopted(1) Partition into small areas
As a small area can be approximated by the lower order polynomials, such as affine or
pseudo affine transformation, the inverse equation can be easily determined. Resampling
can be undertaken for each small area, one by one.
(2) Line and pixel functions
A line function can be determined approximately to search for a scan line number which is
closest to the pixel to be resampled, while a pixel function can be determined to search for
the pixel number.
In resampling as shown in Figure 9.7.1(b), a projected point in an input image plane does
not coincide with the input image data. Therefore the spectral data should be interpolated,
and the following methods can be used(1) Nearest neighbor (NN)
As shown in Figure 9.7.2 (a), the nearest point will be sampled. The geometric error will
be a half pixel at maximum. It has the advantage of being easy and fast.
(2) Bi-linear (BL)
As shown in Figure 9.7.2 (b), the bi-linear function is applied to the surrounding four
points. The spectral data will be smoothed after the interpolation.
(3) Cubic convolution (CC)
As shown in Figure 9.7.2 (c), the spectral data will be interpolated by a cubic function
using the surrounding sixteen points.The cubic convolution results in sharpening as well
104
as smoothing, though the computation takes a longer time when compared with the other
methods
9.8 Map Projection
A map projection is used project the rotated ellipse representing the earth's shape, to a
two-dimensional plane. However there will remain some distortions because the curved
surface of the earth cannot be projected precisely on to a plane.
There are three major map projection techniques; perspective projection, conical projection
and cylindrical projection, which are used in remote sensing. There are described as
follows.
a. Perspective projection
The perspective projection projects the earth from a projection center to a plane as shown
in Figure 9.8.1. The Polar stereo projection is a perspective projection, as shown in Figure
9.8.2, which projects the northern or southern hemisphere from a projection center at the
opposite pole to a vertical plane tangent at the pole. The NOAA Global Vegetation Index
(GVI) data are edited in the polar stereo projection.
b. Conical projection
The conical projection projects the earth from the center of the earth to a conical body
which envelops the earth. The Lambertian conical projection is a typical conical
projection with the axis of the conical body identical to the axis of the earth. Aerial
navigation charts are drawn using this projection for mid-latitudes, with wider areas from
the west to the east.
c. Cylindrical projection
The cylindrical projection projects the earth from the center of the earth to a cylinder which
envelops or intersects the earth. The Mercator projection, as shown in Figure 9.8.3, is a
typical cylindrical projection with the equator tangent to the cylinder. The Universal
Transverse Mercator (UTM) is also an internationally popular map projection. UTM is
a type of Gauss-Kruger projection, with the meridian tangent to the cylinder, as shown
in Figure 9.8.4. The UTM has an origin point at every six degrees of longitude with a scale
105
factor of 0.9996 at the origin and 1.0000 at a distance of 90 kilometers from the central
meridian.
d. Other projections
For computer processing, a grid coordinate system with equal intervals of latitude and
longitude, is often more convenient.
Chapter 10 Image Processing –
Conversion
10.1 Image Enhancement and Feature Extraction
Image enhancement can be defined as conversion of the image quality to a better and
more understandable level for feature extraction or image interpretation, while radiometric
correction is to reconstruct the physically calibrated value from the observed data.
On the other hand, feature extraction can be defined as the operation to quantify the image
quality through various parameters or functions, which are applied to the original image.
These processes can be considered as conversion of the image data. Image enhancement is
applied mainly for image interpretation in the form of an image output, while feature
extraction is normally used for automated classification or analysis in a quantitative form
(see Figure 10.1.1).
106
a. Image Enhancement
Typical image enhancement techniques include gray scale conversion, histogram
conversion, color composition, color conversion between RGB and HSI, etc., which
are usually applied to the image output for image interpretation.
b. Feature Extraction
Features involved in an image are classified as follows.
(1) Spectral features
special color or tone, gradient, spectral parameter etc.
(2) Geometric features
edge, linearment, shape, size, etc.
(3) Textural features
pattern, spatial frequency, homogeneity, etc.
Figure 10.1.2 shows three examples of spectral, geometric and textural feature extraction
10.2 Gray Scale Conversion
Gray scale conversion is one of the simplest image enhancement techniques. Gray scale
conversion can be performed using the following function.
y = f (x)
where x : original input data
y : converted output data
In this section, the following five typical types are introduced, though there are many more
functions that could be used. ( see Figure 10.2.1)
a. Linear conversion
y = ax + b a : gain , b : offset
contrast stretch is one of linear conversion as follows.
Statistical procedures can be also applied in two ways as follows.
107
(1) Conversion of average and standard deviation
where xm : average of input image
Sx : standard deviation of input image
ym : average of output image
Sy : standard deviation of output image
(2) Regression
In such cases as multi-date images for producing a mosaic or radiometric adjustment, a
selected image can be related to other images using regression technique. Line noise due
to different detectors, for example Landsat MSS, can be eliminated by using the regression
technique between different detectors.
Figure 10.2.2 shows various examples of gray scale conversion.
b. Fold conversion
Multiple linear curves are applied in order to enhance only a part of the gray scale.
c. Saw conversion
Where a discontinuous gray scale, occurs, drastic contrast stretch can be made.
d. Continuous function
Function such as exponential, logarithm, polynomials etc. may be applied.
e. Local gray scale conversion
Instead of the conversion being applied to the whole scene by a single formula, parameters
are changed with respect to small local areas.
10.3 Histogram Conversion
Histogram conversion is the conversion of the histogram of original image to an other
histogram. Histogram conversion can be said to be a type of gray scale conversion.
There are two typical histogram conversion techniques.
a. Histogram equalization
108
b. Histogram equalization is to convert the histogram of an original image to equalized
histogram as shown in Figure 10.3.1. As a first step, an accumulated histogram should be
made. Then the accumulated histogram should be divided into a number of equal regions.
Thirdly , the corresponding gray scale in each region should be assigned to a converted
gray scale.
The effect of histogram equalization is that parts of the image with more frequency
variation will be more enhanced, while parts of an image with less frequency will be
neglected.
Figure 10.3.2 shows a comparison between the original image and the converted image,
after histogram equalization.
c. Histogram normalization
Generally a normal distribution of the density in an image would create an image that is
natural for a human observation. In this sense the histogram of the original image may be
sometimes converted to the normalized histogram. However in this conversion, pixels with
same gray scale should be reallocated to other pixels with a different gray scales, in order
to form a normalized histogram.
Therefore such a gray scale conversion is not a 1:1 conversion and thus enables no reverse
conversion. Histogram normalization may be applied, for example, to an unfocused image
of a planet with a low dynamic range, though it is not be very much popular for ordinary
remote sensing data
10.4 Color Display of Image Data
Color display of remote sensing data is of importance for effective visual interpretation.
There are two color display methods; color composite, to generate color with multi-band
data and pseudo-color display, to assign different colors to the gray scale of a single image.
a. Color Composite
A color image can be generated by composing three selected multi-band images with the
use of three primary colors. Different color images may be obtained depending on the
selection of three band images and the assignment of the three primary colors.
109
There are two methods of color composite; an additive color composite and a subtractive
color composite, as shown in Figure 10.4.1. Additive color composite uses three light
sources of three primary colors (Blue, Green and Red) for example, in a multispectral
viewer or color graphic display. The subtractive color composite, uses three pigments of
three primary color (Cyan, Magenta and Yellow), for example, in color printing.
When three filters of B, G and R are assigned to the same spectral regions of blue, green
and red as shown in Figure 10.4.2, almost the same color as the natural scale, can be
reproduced, and is called a natural color composite.
However in remote sensing multi-band images are not always divided in to the same
spectral regions as the three primary color filters. In addition invisible regions, such as
infrared, are often used, which are required to be displayed in color. As a color composite
with an infrared band is no longer natural color, it is called a false color composite.
In particularly the color composite with the assignment of blue to the green band, green to
the red band and red to the near infrared band is very popular, and is called an infrared
color composite, which is the same as found in color infrared film (see Figure 10.4.2).
In the case of digital data, three values corresponding to R, G and B will make various
color combinations, as listed in Table 10.4.1.
b. Pseudo Color Display
Different colors may be assigned to the subdivided gray scale of a single image. Such a
color allocation is called pseudo-color. For example, a pseudo-color image of a thermal
infrared image will give a temperature map. If one wishes to produce a continuous color
tone, three different functions of three primary colors should be applied. Figure 10.4.3 is
an example of a pseudo-color display with continuous color tone.
10.5 Color Representation - Color Mixing System
Light is perceived as color by the human eye, termed color stimulus, and corresponds to
the visible region of electro-magnetic spectrum, with a specific spectral curve of radiance
as shown in Figure 10.5.1.
110
However as the physical value such as a spectral curve, is not convenient for representing
color in daily life, a psychological representation or sensitivity expression are more
practical.
Color representation can be classified into two types; a color mixing system using a
quantitative and physical approach, and a color appearance system using a qualitative
approach, by color code or color sample.
The color mixing system can generate any color by mixing of the three primary colors. The
RGB color system specified by CIE, uses three primary color stimuli; blue of 435.8 nm(B),
green of 546.1 nm(G) and red of 700.0 nm(R) by which all spectral colors ranging from
380 nm to 780 nm, can be generated with the mixing combinations (termed a color
matching function or spectral tristimulus values) as shown in Figure 10.5.2.
As a part of the three spectral stimuli
and
includes a negative region, a
coordinate transformation is applied to generate the virtual three spectral stimuli
and
,and
and
with positive values, as shown in Figure 10.5.3. This is called
the XYZ color system. The three stimuli, X, Y and Z, can be computed as follows.
where K: constant
L( ) : spectral irradiance of standard illumination
( ) : spectral reflectance of sample
Trichromatic coordinates (x,y) can be computed as follows.
111
The value of Y corresponds to brightness while the coordinates of (x,y) represent hue and
saturation (or chrome) .
The fringe of the bell shape corresponds to the spectral color with high chrome, while the
inside corresponds with a low chrome.
10.6 Color Representation - Color Appearance System
TheMunsell color system is a typical color appearance system, in which color is
represented with hue (H), saturation(S) and intensity (I) as a psychological response. Hue
is composed of the five basic color; red (R), yellow (Y), green(G), blue (B) and purple (P)
which are located along a hue ring with intervals of 72 degrees as shown in Figure 10.6.1.
Intermediate colors between the above five basic colors; YR, GY, BG, PB and RP are
located in between each other. Finally each hue is divided into ten but actually four. For
example 1R, 5R, 10R, 1YR, 5YR, 10YR, 1Y, ...... are a series along the hue ring.
Intensity is an index of brightness with 11 ranks from 0 (dark) to 10 (light). Saturation is
an index of pureness ranging from 0 to 16 depending on the hue and intensity. Color in the
Munsell color system is identified as a combination of hue, intensity / saturation, for
example 5R4 / 10, which means 5R (hue), 4 (intensity) and 10 (saturation).
Figure 10.6.2 shows a three dimensional color solid as called the Munsell's solid, with the
40 panels with color samples of intensity and saturation with respect to the hue. Munsell
color samples are available in the commercial market. Any user can identity arbitrary
colors by comparison with the Munsell's color samples. Psychologically defined HSI has
been correlated with physically defined RGB or Yxy as mentioned in 10.5. Therefore
conversion between RGB and HSI can be made mathematically. In the case of a color
display using a digital image processing device, the RGB signal has to be input, though
color control is much easier using HSI indices. Figure 10.6.3 shows the relationship
between RGB space and HSI space.
112
The following are conversions from RGB to HSI, and from HSI to RGB. The ranges of
R,G,B,S,I are [0,1] :, the range of H is [0,2p].
(1) from RGB to HIS
I = Max. (R,G,B)
1) I = 0 ; S = 0, H= indeterminate
S = (I-i)/I , where i = min. {R, G, B}
Let r = (I-R) / (I-i), g = (I-G) / (I-i), b = (I-B) / (I-i), then
if R = I H = (b-g) / 3
if G = I H = (2+r-b) / 3
if B = I H = (4+g-r) / 3
(2) from HSI to RGB
1) S = 0 ; R = G = B = I regardless of value of H
H' = 3H / h = floor(H') If H = 2 , then H = 0
(floor (x): the function of getting the truncated value of x)
P = I(1-S), Q = I{1-S (H' - h)} , T = I {1-S(1-H'+h)} , then
h = 0 R = I, G = T, B = P
h = 1 R = Q, G = I, B = P
h = 2 R = P, G = I, B = T
h = 3 R = P, G = Q, B = I
h = 4 R = I, G = P, B = Q
h = 5 R = I, G = P, B = Q
10.7 Operations between Images
Operations between multi-spectral images or multi-date images are very useful for image
enhancement and feature extraction.
Operations between images include two techniques; arithmetic operation and logical
operation.
a. Arithmetic Operations
113
Addition, subtraction, multiplication, division and their combinations, can be applied for
many purposes, including noise elimination. As the results of the operation can sometimes
negative or small values between 0 and 1, they should be adjusted to a range, usually in
eight bits or 0 to 255 for image display.
Typical operations are ratioing, for geological feature extraction, and normalized
difference vegetation index, for vegetation monitoring with NOAA AVHRR data or other
visible near infrared sensors.
(1) Ratioing
Ratio = Xi / Xj
Ratioing may be useful for geological feature extraction. Such ratioing can be applied to
multi-temporal thermal infrared data for extraction of thermal inertia.
(2) Normalized Difference Vegetation Index(NDVI)
where ch.1 : red band
ch.2 : infrared band
NDVI shows as a high value for denser vegetation, while the NDVI is very low in desert,
or non-vegetation regions.
Figure 10.7.1 shows two examples of arithmetic operations.
c. Logical Operation
Logical addition (OR set ), logical multiplication (AND set), true and false operations etc.
can be applied to multi-date images or a combination of remote sensing images and
thematic map images.
For example a remote sensing image or the classified result can be overlaid on map data,
such as political boundaries.
Figure 10.7.2 shows an example of forest land change by overlaying a remote sensing
image on the forest land which has been mapped from the old map. Such an overlay will
be very useful for change detection.
114
10.8 Principal Component Analysis
Principal component analysis is used to reduce the dimensions of measured variables (p
dimension) to the representative principal components (m dimension , m
Let the measured p dimensional variables be {xi } i = 1,p, the principal components{zk }
k = 1, m can be expressed as the linear combination as follows.
zk = a1k x1 + a2k x2 + ...... + apk xp
The coefficients (a1k - apk ) are determined under the following constrains.
(1) aik = 1
(2) Variance zk should be maximum
(3) zk and z k+1 should be independent of each other
The solution of the above problem can be obtained by determining the unique values and
the unique vectors which correspond to the variance and vector of the principal components
respectively.
The unique value represents the contribution ratio which indicates how much percentage
the principal component represents of the total tendency of the variables. The
accumulative contribution ratio percentage all the principal components represent of the
total tendency of the variables. Using an accumulative contribution ratio of 80 - 90 percent,
will indicate how many principal components should be adopted to effectively represent
the major variations in the image data.
Graphically speaking, the first principal component for example in the case of two
dimensional variables (see Figure 10.8.1) will be the principal axis which gives the
maximum variance. The principal component analysis can be used for the following
applications.
(1) Effective classification of land use with multi-band data
(2) Color representation or visual interpretation with multi-band data
(3) Change detection with multi-temporal data
In the case of multi-band data with more than four bands, all bands cannot be assigned to
R, G or B at the same time. However the first three principal components can represent up
to five spectral variables with little information loss.
115
Figure 10.8.2 show the principal components and their color composite of Landsat TM (6
bands). Generally the first principal component corresponds to the total radiance
(brightness), while the second principal component represents the vegetation activity
(greenness).
10.9 Spatial Filtering
Spatial filtering is used to obtain enhanced images or improved images by applying, filter
function or filter operators in the domain of the image space (x,y) or spatial frequency (x,h).
Spatial filtering in the domain of image space aims at image enhancement with so-called
enhancement filters, while in the domain of spatial frequency it aims at reconstruction with
so-called reconstruction filters.
a. Filtering in the Domain of Image Space
In the case of digital image data, spatial filtering in the domain of image space is usually
achieved by local convolution with an n x n matrix operator as follows.
where f: input image
h: filter function
g: output image
The convolution is created by a series of shift-multiply-sum operators with an n x n matrix
(n: odd number). Because the image data are large, n is usually selected as 3, although n is
sometimes selected as 5, 7, 9 or 11.
Figure 10.9.1 shows typical 3 x 3 enhancement filters. Figure 10.9.2 shows the input image
and several output images with various 3 x 3 operators.
b. Filtering in the domain of Spatial Frequency
Filtering in the domain of spatial filtering uses the Fourier transformation to convert from
image space domain to spatial frequency domain as follows.
G(u,v) = F(u,v) H(u,v)
116
F: Fourier transformation of input image
H: filter function
An output image from filtering of spatial frequency, can be obtained by using an inverse
Fourier transformation of the above formula.
Low pass filters, high pass filters, band pass filters etc., are typical filters with a criterion
of frequency control. Low pass filters which out puts only lower frequency image data, less
than a specified threshold, can be applied to remove high frequency, noise,while high pass
filter are used for removing, for example, stripe noise of low frequency.
10.10 Texture Analysis
Texture is a combination of repeated patterns with a regular frequency. In visual
interpretation texture has several types, for example, smooth, fine, coarse etc., which are
often used in the classification of forest types. Texture analysis is defined as the
classification or segmentation of textural features with respect to the shape of a small
element, density and direction of regularity.
Figure 10.10.1 (a) shows two different textures of density, while Figure 10.10.1 (b) shows
two different textures with respect to the shape of the elements.
In the case of digital image, it is difficult to treat the texture mathematically because texture
cannot be standardized quantitatively and the data volume is so huge.
However texture analysis has been made with statistical features which are combined with
spectral data for improving land cover classification. Power spectrum analysis is another
form of textural analysis in which direction and wavelength or frequency can be determined
for regular patterns of , for example, sea waves and sand waves in the desert.
a. Use of Statistical Features
The following statistical values of an n x n window can be used as textural information
(1) Gray level histogram
(2) Variance - co-variance matrix
(3) Run-length matrix
These values are used for classification together with the spectral data.Figure 10.10.2 (a)
shows the land cover classification using only spectral data while Figure 10.10.2 (b) shows
117
the result of classification with spectral data as well as textural information. The result
shows a better classification for the urban area which has a higher frequency and variance
of image density.
b. Analysis using Power Spectrum
Power spectrum analysis is useful for those images which have regular wave patterns with
a constant interval, such as glitter image of the sea surface or wave patterns of sand dunes.
Fourier transformation is applied to determine the power spectrum which gives the
frequency and direction of the pattern.
10.11 Image Correlation
Image correlation is a technique by which the conjugate point of a slave image (right)
corresponding to the master image (left) will be searched for the maximum correlation
coefficient. Image correlation is applied to stereo images for DEM (digital elevation model)
generation or multi-date images for automated recognition of ground control points.
As shown in Figure 10.11.1, the master window in the left image is fixed, while the slave
window in the right image is moved to search for the maximum image correlation as
computed from the following formula.
or
where ai : image data of the master window
bi : image data of the slave window
n : total number of image data
Because the above two correlations show almost no difference, the first correlation is
preferred to save computing time.
118
The size of the window should be selected depending on the image resolution and feature
size. For example, 5 x 5 to 9 x 9 windows might be selected for SPOT stereo images, while
9 x 9 to 21 x 21 would be better used for digitized aerial photographs.
When the conjugate points of stereo images are determined, the corresponding digital
elevation can be computed using collinearity equations based on photogrammetric theory.
Figure 10.11.2 shows the conjugate points as white dots in a pair of SPOT stereo images,
which were automatically recognized by image correlation techniques.
Chapter 11 Image Processing –
Classification
11.1 Classification Techniques
Classification of remotely sensed data is used to assign corresponding levels with respect
to groups with homogeneous characteristics, with the aim of discriminating multiple
objects from each other within the image.
119
The level is called class. Classification will be executed on the base of spectral or spectrally
defined features, such as density, texture etc. in the feature space. It can be said that
classification divides the feature space into several classes based on a decision rule. Figure
11.1.1 shows the concept of classification of remotely sensed data
. In many cases, classification will be undertaken using a computer, with the use of
mathematical classification techniques. Classification will be made according to the
following procedures as shown in Figure 11.1.2.
Step 1: Definition of Classification Classes
Depending on the objective and the characteristics of the image data, the classification
classes should be clearly defined.
Step 2: Selection of Features
Features to discriminate between the classes should be established using multi-spectral
and/or multi-temporal characteristics, textures etc.
Step 3: Sampling of Training Data
Training data should be sampled in order to determine appropriate decision rules.
Classification techniques such as supervised or unsupervised learning will then be selected
on the basis of the training data sets.
Step 4: Estimation of Universal Statistics
Various classification techniques will be compared with the training data, so that an
appropriate decision rule is selected for subsequent classification.
Step 5: Classification
Depending up on the decision rule, all pixels are classified in a single class. There are two
methods of pixel by pixel classification and per-field classification, with respect to
segmented areas.
Popular techniques are as follows.
a. Multi-level slice classifier
b. Minimum distance classifier
c. Maximum likelihood classifier
d. Other classifiers such as fuzzy set theory and expert systems
120
Step 6: Verification of Results
The classified results should be checked and verified for their accuracy and reliability.
11.2 Estimation of Population Statistics
a. Supervised classification
In order to determine a decision rule for classification, it is necessary to know the spectral
characteristics or features with respect to the population of each class. The spectral features
can be measured using ground-based spectrometers. However due to atmospheric effects,
direct use of spectral features measured on the ground are not always available. For this
reason, sampling of training data from clearly identified training areas, corresponding to
defined classes is usually made for estimating the population statistics (see Figure 11.2.1).
This is called supervised classification. Statistically unbiased sampling of training data
should be made in order to represent the population correctly.
b. Unsupervised Classification
In the case where there is less information in an area to be classified, only the image
characteristics are used as follows.
(1) Multiple groups, from randomly sampled data, will be mechanically divided into
homogeneous
spectral
classes
using
a
clustering
technique
(see
11.3).
(2) The clustered classes are then used for estimating the population statistics. This
classification technique is called unsupervised classification (see Figure 11.2.2).
c. Estimation of Population Statistics
Maximum likelihood estimation is the most popular method by which the population
statistics such as mean and variance, are estimated to maximize the probability or
likelihood from a defined probability density function within the feature space.
In most cases, the probability density function is selected to be a multiple normal
distribution. The multiple normal distribution gives the following the maximum likelihood
estimator.
121
Variance - covariance matrix
where m: number of bands
n: number of pixels
Before adopting the maximum likelihood classification, it should be checked to determine
if the distribution of training data will fit the normal distribution or not.
11.3 Clustering
Clustering is a grouping of data with similar characteristics. Clustering is divided into
hierarchical clustering and non-hierarchical clustering as mentioned as follows.
a. Hierarchical Clustering
b. The similarity of a cluster is evaluated using a "distance" measure. The minimum
distance between clusters will give a merged cluster after repeated procedures from a
starting point of pixel-wise clusters to a final limited number of clusters.
Figure 11.3.1 shows the general procedure of hierarchical clustering.
The distances to evaluate the similarity are selected from the following methods.
(1) Nearest neighbor method
Nearest neighbor with minimum distance will form a new merged cluster.
(2) Furthest neighbor method
Furthest neighbor with maximum distance will form a new merged cluster.
(3) Centroid method
Distance between the gravity centers of two clusters is evaluated for merging a new
merged cluster.
(4) Group average method
Root mean square distance between all pairs of data within two different clusters, is used
for clustering.
(5) Ward method
Root mean square distance between the gravity center and each member is minimized.
b. Non-hierarchical Clustering
122
At the initial stage, an arbitrary number of clusters should be temporally chosen. The
members belonging to each cluster will be checked by selected parameters or distance and
relocated into the more appropriate clusters with higher separability. The ISODATA
method and K-mean method are examples of non-hierarchical clustering.
The ISODATA method is composed of the following procedures (see Figure 11.3.2).
(1) All members are relocated into the closest clusters by computing the distance between
the member and the clusters.
(2) The center of gravity of all clusters is recalculated and the above procedure is repeated
until convergence.
(3) If the number of clusters is within a certain specified number, and the distances between
the clusters meet a prescribed threshold, the clustering is considered complete.
11.4 Parallelpiped Classifier
The parallelpiped classifier (often termed multi-level slicing) divides each axis of multispectral feature space, as shown in an example in Figure 11.4.1. The decision region for
each class is defined on the basis of a lowest and highest value on each axis. The accuracy
of classification depends on the selection of the lowest and highest values in consideration
of the population statistics of each class. In this respect, it is most important that the
distribution of population of each class is well understood.
The parallelpiped classifier is very simple and easy to understand schematically. In addition
the computing time will be a minimum, when compared with other classifiers.
However the accuracy will be low especially when the distribution in feature space has
covariance or dependency with oblique axes. Orthogonalization should be undertaken
using principal component analysis, for example, before adopting the parallelpiped
classifier.
Figure 11.4.2 shows an example of classification with the use of the parallelpiped classifier
11.5 Decision Tree Classifier
The decision tree classifier is an hierarchically based classifier which compares the data
with a range of properly selected features. The selection of features is determined from an
assessment of the spectral distributions or separability of the classes. There is no generally
123
established procedure. Therefore each decision tree or set of rules should be designed by
an expert. When a decision tree provides only two outcomes at each stage, the classifier is
called a binary decision tree classifier (BDT).
Figure 11.5.1 shows the spectral characteristics of ground truth data for nine classes and
the corresponding decision tree classifier to classify the nine classes using their spectral
characteristics.
Generally a group of classes will be classified into two groups with the highest separability
with respect to a feature.
Features often used are as follows.
(1) Spectral values
(2) An index which is computed from spectral values. For example, the vegetation index
is a popular indices.
(3) any arithmetic value such as addition, subtraction or ratioing.
(4) Principal components.
The advantages of the decision tree classifier are that computing time is less than the
maximum likelihood classifier and by comparison the statistical errors are avoided.
However the disadvantage is that the accuracy depends fully on the design of the decision
tree and the selected features.
Figure 11.5.2 shows an example of classification with a decision tree classifier.
11.6 Minimum Distance Classifier
The minimum distance classifier is used to classify unknown image data to classes which
minimize the distance between the image data and the class in multi-feature space. The
distance is defined as an index of similarity so that the minimum distance is identical to the
maximum similarity. Figure 11.6.1 shows the concept of a minimum distance classifier.
The following distances are often used in this procedure.
(1) Euclidian distance
124
Is used in cases where the variances of the population classes are different to each other.
The Euclidian distance is theoretically identical to the similarity index.
(2) Normalized Euclidian distance
The Normalized Euclidian distance is proportional to the similarity in dex, as shown in
Figure 11.6.2, in the case of difference variance.
(3) Mahalanobis distance
(4) In cases where there is correlation between the axes in feature space, the Mahalanobis
distance with variance-covariance matrix, should be used as shown in Figure 11.6.3.
where X : vector of image data (n bands)
X = [ x1, x2, .... xn]
k : mean of the kth class
k = [ m1, m2, .... mn]
k : variance matrix
k : variance-covariance matrix
Figure 11.6.4 shows examples of classification with the three distances.
11.7 Maximum Likelihood Classifier
The maximum likelihood classifier is one of the most popular methods of classification
in remote sensing, in which a pixel with the maximum likelihood is classified into the
125
corresponding class. The likelihood Lk is defined as the posterior probability of a pixel
belonging to class k.
Lk = P(k/X) = P(k)*P(X/k) / P(i)*P(X/i)
where P(k) : prior probability of class k
P(X/k) : conditional probability to observe X from class k, or probability density function
Usually P(k) are assumed to be equal to each other and P(i)*P(X/i) is also common to all
classes. Therefore Lk depends on P(X/k) or the probability density function.
For mathematical reasons, a multivariate normal distribution is applied as the probability
density function. In the case of normal distributions, the likelihood can be expressed as
follows.
where n: number of bands
X: image data of n bands
Lk(X) : likelihood of X belonging to class k
k : mean vector of class k
k : variance-covariance matrix of class k
In the case where the variance-covariance matrix is symmetric, the likelihood is the same
as the Euclidian distance, while in case where the determinants are equal each other, the
likelihood becomes the same as the Mahalanobis distances. Figure 11.7.1 shows the
concept of the maximum likelihood method.
The maximum likelihood method has an advantage from the view point of probability
theory, but care must be taken with respect to the following items.
(1) Sufficient ground truth data should be sampled to allow estimation of the mean
vector and the variance-covariance matrix of population.
(2) The inverse matrix of the variance-covariance matrix becomes unstable in the case
where there exists very high correlation between two bands or the ground truth data are
126
very homogeneous. In such cases, the number of bands should be reduced by a principal
component analysis.
(3) When the distribution of the population does not follow the normal distribution, the
maximum likelihood method cannot be applied.
Figure 11.7.2 shows an example of classification by the maximum likelihood method.
11.8 Applications of Fuzzy Set Theory
Fuzzy set theory, to treat fuzziness in data, was proposed by Zadeh in 1965. In Fuzzy set
theory the membership grade can be taken as a value intermediate between 0 and 1 although
in the normal case of set theory membership the grade can be taken only as 0 or 1.
Figure 11.8.1 shows a comparison between the normal case of set theory and fuzzy set
theory. The function of the membership grade is called its "membership function" in Fuzzy
theory. The membership function will be defined by the user in consideration of the
fuzziness.
In remote sensing it is often not easy to delineate the boundary between two different
classes. For example, there are transitive vegetation or mixed vegetation between forest
and grass land. In such cases as unclearly defined class boundaries, Fuzzy set theory can
be usefully applied, in a qualitative sense.
The following shows how the maximum likelihood method with Fuzzy set theory. Let the
membership function be Mf( ) of class k (k=1,n), the likelihood Lf of fuzzy class f can be
defined as follows.
Fuzzy set theory can be also extended to clustering. Figure 11.8.2 shows an example of
land cover classification using Fuzzy set theory. In this classification, the concrete structure
(code 90), with clearly defined characteristics, was first classified using the ordinary
maximum likelihood method, while the loosely defined urban classes were classified by
the fuzzy based maximum likelihood method.
11.9 Classification using an Expert System
127
Experts interpret remote sensing images with knowledge based on experience. However
computer assisted classification utilizes only very limited expert knowledge. The expert
system, therefore, is a problem solving system which supports expert knowledge in a
computer based system.
The following two types of knowledge are required for an expert system in remote sensing.
(1) Knowledge about image analysis
Procedures for image analysis can be made only with adequate knowledge about image
processing and analysis. A feedback system should be introduced for checking and
evaluating the objectives and the results.
(3) Knowledge about the objects to be analyzed
Knowledge about the objects to be recognized or classified should be introduced in addition
to the ordinary classification method. The fact that forest does not exist over 3,000 meters
above sea level, is one example of the type of knowledge that can be introduced.
Table 11.9.1 shows a list of knowledge required for delineating a tidal front in sea surface
condition mapping. Figure 11.9.1 shows the sea surface condition map that was interpreted
by an expert. Such knowledge will assure an increase in the accuracy or reliability of
classification.
In many cases, knowledge can be represented as "if A is ..., then B becomes...." which is
called the IF/THEN rule or production rule.
If the IF/THEN rule is fuzzy, then Fuzzy set theory can be also introduced to the expert
system.
Figure 11.9.2 shows an example of the delineation of a tidal front using the expert system.
The expert system can be integrated with a geographic information system (GIS). It is
necessary to accumulate experiences and to evaluate the knowledge for an expert system
to be operationally applied.
128
Chapter 12 Applications of Remote
Sensing
12.1 Land Cover Classification
Land cover mapping is one of the most important and typical applications of remote
sensing data. Land cover corresponds to the physical condition of the ground surface, for
129
example, forest, grassland, concrete pavement etc., while land use reflects human activities
such as the use of the land, for example, industrial zones, residential zones, agricultural
fields etc.
Generally land cover does not coincide with land use. A land use class is composed of
several land covers. Remote sensing data can provide land cover information rather than
land use information.
Initially the land cover classification system should be established, which is usually defined
as levels and classes. The level and class should be designed in consideration of the purpose
of use (national, regional or local), the spatial and spectral resolution of the remote sensing
data, user's request and so on.
The definition should be made as quantitatively clear as possible. Figure 12.1.1 shows an
example of land cover classes for land cover mapping in the Sagami River Basin, Japan,
for use with Landsat MSS data.
The classification was carried out as follows.
a. Geometric correction (see 9.4)
A geo-coded Landsat image was produced.
b. Collection of the ground truth data (see 6.7)
A ground investigation was made to identify each land cover class on the geo-code
Landsat image as well as on topographic maps.
c. Classification by Maximum Likelihood Method (see 11.5 )
The Maximum Likelihood Method was adopted using the training samples obtained from
the ground truth.
Figure 12.1.1 shows the classified land cover map.
Generally Landsat MSS imagery can provide about about ten land cover classes,
depending upon the size and complexity of the classes.
12.2 Land Cover Change Detection
Land cover change detection is necessary for updating land cover maps and the
management of natural resources. The change is usually detected by comparison between
130
two multi-date images, or sometimes between an old map and an updated remote sensing
image.
The method of change detection is divided into two;
a. comparison between two land cover maps which are independently produced
b.change enhancement by integrating two images into a color composite or principal
component image.
Figure 12.2.1 shows the changes over a 5 year period, which were detected by using a color
composite with blue assigned to an old image of Landsat TM and red assigned to a new
image of Landsat TM.
Such detection is very useful for updating "vegetation maps" of 1:50,000 to 1:100,000 scale
with Landsat TM or SPOT, and of 1:250,000 scale with Landsat MSS.
The land cover change can also be divided into two;
a. seasonal change
agricultural lands and deciduous forests change seasonally
b. annual change
land cover or land use changes, which are real changes, for example deforested areas or
newly built towns.
Usually seasonal change and annual change are mixed within the same image. However
only the real change should be detected, so that two multi-data images of almost same
season should be selected to eliminate the effects of seasonal change. One should note that
a cycle of seasonal change can be rather complex as shown in Figure 12.2.2. Sometimes
seasonal change rate is very high, for example in spring time in cold area.
12.3 Global Vegetation Map
NOAA AVHRR data (see 5.1) are very useful for producing a global vegetation maps
which cover the whole world, because NOAA has edited global cloud free mosaics in the
form of a GVI(global vegetation index) on a weekly basis since April of 1982.
131
The GVI data include information about NDVI (normalized difference vegetation index)
as computed as follows
Ch.1 : visible band
Ch.2 : near infrared band
NDVI is sometimes simply called. NVI (normalized vegetation index).
NDVI or NVI are indicators of the intensity of biomass.
The larger the NVI is, the denser the vegetation.
Though the original resolution of NOAA AVHRR is 1.1 km per pixel of the Equator, the
GVI has a low resolution of 16 km x 16 km per pixel at the Equator. In spite of the low
resolution, the GVI is useful for producing a global vegetation map.
As much noise is involved in weekly data, noise free GVI compiled on a monthly base
should be used. Figure 12.3.1 shows six categories out of a total of 13 categories obtained
from cluster analysis (see 11.2).
Figure 12.3.2 shows the result of a cluster analysis applied to GVI data from 1987.
Though the clustered map in Figure 12.3.2 has not yet been verified, it shows the possibility
of using remote sensing data for global map production.
12.4 Water Quality Monitoring
Water pollution has become a very serious problem in big cities and in offshore areas along
industrial zones. Water quality monitoring is one of the typical applications of remote
sensing.
Figure 12.4.1 shows the characteristics of reflection, absorption and scattering on the water
surface, beneath the surface water and from the bottom. The sea color depends on the
absorption and scattering due to water molecules and suspended particles or plankton.
Figure 12.4.2 shows various curves of spectral attenuation with respect to various types of
water. As seen in the figure, clear water has a peak of minimum attenuation around 0.5
m, while turbid water with suspended solid (SS) has larger attenuation with a minimum
peak around 0.55 m. In other word, radiation can penetrate into deep clear water and is
132
scattered by the water volume, causing the typical bluish color. Turbid waters cannot be
penetrated and radiation is scattered near the surface, giving a greenish or yellowish color.
The sea color depends on not only on suspended solids but also the chlorophyll of plankton
within the water body.
Figure 12.4.3 shows an example of the measurement for spectral reflectance of various
amounts of chlorophyll. As seen in the figure, chlorophyll in the sea can be detected in the
region of 0.45 - 0.65 m.
12.5 Measurement of Sea Surface Temperature
Satellite remote sensing can provide thermal information in a short time over a wide area.
Temperature measurement by remote sensing is based on the principle that any object emits
electro-magnetic energy corresponding to the temperature, wavelength and emissivity.
The temperature detected by a thermal sensor is called the "brightness temperature" (see
1.7). Though the brightness temperature coincides with the real temperature if the objects
is a black body, the actual object on the earth has a different emissivity e (e < 1) which
emits electro-magnetic energy of e.I, where I indicates the radiance of a black body with
the same temperature.
Thus the value of e as well as the emitted radiance should be measured in order to compute
the exact temperature, as explained in Figure 12.5.1.
However the value of e for sea water is very nearly equal to 1 and also comparatively
constant, while the value e for ground surfaces is not homogeneous. Thus sea surface
temperature can be estimated more accurately than ground surface temperature.
As the actual brightness temperature includes emitted radiance from the atmosphere, this
will cause a temperature error ranging 2-3 degrees Centigrade between the actual sea
surface temperature and calculated brightness temperature from satellite data. Thus
atmospheric correction (see 9.2) is very important for accurate sea surface temperature
measurement.
Figure 12.5.2 shows the sea surface temperature in pseudo color in Northern Japan using
NOAA AVHRR data which were atmospheric ally as well as geometrically corrected with
overlays of sea coast lines and latitude and longitude grid lines.
133
Using the most recent technology, the estimated accuracy of sea surface temperature is
claimed to be about +-0.5 C on a global scale and about +-0.3 C on a regional scale.
12.6 Snow Survey
As snow cover has a very high reflectance, the aerial distribution of snow can be identified
very easily from satellite remote sensing data. Several models to estimate water resources
in units of snow water equivalent have been proposed with use of the estimated snow
cover area.
Figure 12.6.1 shows a conceptual diagram of the estimation of basin-wide snow water
equivalent with the three parameters of elevation h, latitudinal distribution of snow water
equivalent S(h) and hydrometric curve A(h).
Snow water equivalent Ss in a river basin can be computed as follows.
where hH : maximum elevation
hL: minimum elevation
If snow appears over the average elevation of snow line ho, hL should be replaced by ho.
From the above formula, the latitudinal distribution of snow water equivalent S(h) and the
hydrometric curve A(h) should be determined in order to estimate the snow water
equivalent.
It is known that the snow water equivalent increases linearly proportional to the elevation,
which can ve obtained from the existing snow survey. On the other hand the catchment
area can be expressed in lower order of polynomials as a function of elevation. Therefore
the snow water equivalent can be estimated as a function of percentage of snow cover area
in a river basin.
Figure 12.6.2 shows an estimated curve of snow water equivalent which was obtained from
sample data in the three years of 1979, 1982 and 1983 in the Takaragawa River Basin,
Japan. From the curve, the snow water equivalent can be estimated if the snow cover area
is detected from remote sensing imagery (see a series of examples, Figure 12.6.3(a), (b)
and (c)).
134
Recently microwave remote sensing has been applied to estimate snow volume. Passive
microwave radiometers can provide snow surface temperature with respect to inductivity
of the snow, which may provide snow information. Active microwave radar can provide
reflectivity or scattering of snow with respect to snow density, snow temperature, snow
temperature, size of snow particles etc. It is still difficult to estimate snow volume from
microwave data but several research projects are currently being carried out on this topic.
12.7 Monitoring of Atmospheric Constituents
Each atmospheric constituent such as water vapor, carbon-dioxidide, ozone, methane etc.
has its own unique spectral characteristic of emission and absorption. With the use of these
characteristics, the density of these atmospheric molecules can be monitored by measuring
the spectral energy which transmits from the sun, the moon or the stars through the
atmosphere, the scattering energy from the atmosphere or the clouds, the reflected energy
from the earth surface and/or the thermal radiation emitted from the atmosphere and the
earth surface. The spectral energy can be measured by two methods; absorption
spectroscopy and emission spectroscopy.
These methods have been applied for many years ago for the measurement of the upper
atmosphere from the ground. Recently the methods have been extended for measurements
from aircraft, balloon and satellite. In addition, multi-spectral laser with variable
wavelength, called laser rader or lidar, has been developed for the measurement of the
spatial distribution of the atmospheric constituents.
Figure 12.7.1 shows the spectral transmittance of H2O, CO2, O3, N2O and CH4 in the
infrared region.
Figure 12.7.2 shows the spectral attenuation of water vapor(H2O), and oxygen with a
number of channels of the AMSU (Advanced Microwave Sounding Unit) instrument.
There are three methods used to measure the vertical distribution of atmospheric
constituents; the occultation method which measures the attenuating light of the sun light
at the sun rise and the sun sets, from a satellite, the limb scan method which measures the
spectrum of atmosphere around the limb of the earth and the vertical viewing method,
which measures the atmospheric emission from various altitudes and contribution ratio are
135
analyzed with respect to the spectral absorption coefficient by the inversion method. The
vertical look down method is operationally applied for carbon-dioxidide and water vapor
in the infrared region and for ozone in the ultra-violet region.
Figure 12.7.3 shows the normalized contribution function at various wavelengths in the
ultra-violet region where the vertical distribution of ozone is measured from the back
scattering of blue ultra-violet (BUV) radiation.
Figure 12.7.4 shows the distribution of the integrated ozone which was measured with the
TOMS (Total Ozone Mapping Spectrometer) on board Nimbus 7.
12.8 Lineament Extraction
Lineament is defined as a line feature or pattern interpreted on a remote sensing image.
The lineament reflects the geological structure such as faults or fractures. In this sense, the
lineament extraction is very important for the application of remote sensing to geology.
However the real meaning of lineament is still unclear. It should be discriminated from
other line features that are not due to geological structures. Therefore the lineament
extraction should be carefully interpreted by geologists.
Computer generated lineament would involve all linear features of natural terrain as well
as artificial structures which have to be removed by interpretation.
Figure 12.8.1 shows an example of computer generated lineaments in the Southern Brazil
region.
As lineaments can be interpreted very well on satellite images, geological survey methods
have been advanced, particularly over large areas.
Lineament extraction is useful for geological analysis in oil exploration in which oil flow
along faults, oil storage within faults and the oil layer can be estimated.
Lineament information can even allow analysis of the geological structure and history.
12.9 Geological Interpretation
The applicability of remote sensing data increases according to the improvement in the
spatial resolution as well as the spectral resolution , for example as from Landsat MSS to
Landsat TM and SPOT HRV.
136
The advantage of satellite remote sensing in its application to geology is the wide coverage
over the area of interest, where much useful information such as structural patterns and
spectral features can be extracted from the imagery.
There are two ways of information extraction; geometric feature extraction with the use of
geomorphologic patterns and radiometric feature extraction using the unique
characteristics of spectral absorption corresponding to the rock type.
Generally visual image interpretation is most widely used in order to extract geological
information from remote sensing images.
A comprehensive analysis can be carried out with geomorphologic information such as
land form and slope, drainage pattern and density, and land cover.
Figure 12.9.1 shows a Landsat TM image of the oil deposit basin in California, USA.
Figure 12.9.2 shows the tectonic analysis of the same basin.
Radiometric interpretation of multi-spectral features is mainly applied to rock type
classification.
Figure 12.9.3 shows a color composite of bands 4, 5 and 7 of Landsat TM in Gold Field
Nevada, USA, in which the light green color shows the hydrothemal zones. Because each
rock has its own spectral absorption band in the region of the short wave infrared, data
from multi-spectral scanners or imaging spectrometers with multi channels is very useful
for rock type classification. Thus the OPS data of JERS-1 will be useful in geology because
of the shortwave infrared bands.
12.10 Height Measurement (DEM Generation)
Topographic mapping or DEM (Digital Elevation Model) generation is possible with a pair
of stereo images.
The height accuracy h depends on the parameters of base-height ratio (B/H) and the
accuracy of parallax which may be approximated by ground resolution G, as indicated as
follows.
h = H/B G
137
Table 12.10.1 shows the theoretical accuracy of height determination for Landsat MSS,
Landsat TM and SPOT HRV (panchromatic). In the case of SPOT HRV, with maximum
base length and a B/H ratio of about 1, the height accuracy will be about 10 m the same as
the ground resolution, which will be suffcient to produce topographic maps with contour
lines of 40 meters interval.
There are two methods of topographic mapping or DEM generation; using operator based
analytical plotters with special software, and automated DEM generation by stereo
matching.
Usually rectified images for a pair of stereo images are initially pre-processed using ground
control points.
Then stereo matching is applied to determine the conjugate points, which give x-parallax
or difference of height to be converted to the height or elevation.
Figure 12.10.1 shows a pair of stereo images of SPOT HRV panchromatic data. Figure
12.10.2 shows a three dimensional view with the use of a DEM generated by stereo
matching (see Figure 12.10.3 as an example of a bird's eye view image).
Chapter 13 Geographic Information
System (GIS)
13.1 GIS and Remote Sensing
138
a. GIS in remote sensing
For the users of remote sensing, it is not sufficient to display only the results obtained from
image processing. For example, to detect land cover change in an area is not enough,
because the final goal would be to analyse the cause of change or to evaluate the impact of
change. Therefore the result should be overlaid on maps of transportation facilities and
land use zoning as shown in Figure 13.1.1. In addition, the classification of remote sensing
imagery will become more accurate if the auxiliary data contained in maps are combined
with the image data.
In order to promote the integration of remote sensing and geographic data, geographic
information system (GIS) should be established in which both the image and graphic data
are stored in a digital form, retrieved conditionally, overlaid on each other and evaluated
with the use of a model.
Figure 13.1.2 shows a comparison between the computer assisted GIS and the conventional
analog use of maps.
b. Function of GIS
The following three functions are very important in GIS.
(1) To store and manage geographic information comprehensively and effectively
(2) To display geographic information depending on the purpose of use
(3) To execute query, analysis and evaluation of geographic information effectively
At present, the following research and development have been undertaken. In this book
the following technologies , a part from, visualization will be described.
(1) Model and data structure for GIS
(2) Data input and edition
(3) Spatial query
(4) Spatial analysis
(5) Visualization
13.2 Model and Data Structure
a. Requirement of the Model and Data Structure
139
In order to process and manage geographic information by computers, it is necessary to
describe the spatial location and distribution, as well as the attributes and characteristics,
according to a specified form, termed a spatial representation model with a standardized
data structure.
c. Modeling and Data Structure
Geographic information can be represented with geometric information such as location,
shape and distribution, and attribute information such as characteristics and nature, as
shown in Figure 13.2.1.
Vector and raster forms are the major representation models for geometric information.
(1)
Vector form and its data structure
Most objects on a map can be represented as a combination of a point (or node), edge (or
arc) and area (or polygon). The vector form is provided by the above geometric factors.
The attributes are assigned to points, edges and areas.
The data structure is specified for the vector form as follows.
A point is represented by geographic coordinates. An edge is represented by a series of line
segments with a start point and an end point. A polygon is defined as the sequential edges
of a boundary. The inter-relationship between points, edges and areas is called a topological
relationship. Any change in a point, edge or area will influence other factors through the
topological relationship. Therefore the data structure should be specified to fulfill the
relationship, as for the example as shown in Figure 13.2.2.
(2) Raster form and its data structure
In the raster form, the object space is divided into a group of regularly spaced grids
(sometimes called pixels) to which the attributes are assigned. The raster form is basically
identical to the data format of remote sensing data.
As the grids are generated regularly, the coordinates correspond to the pixel number and
line number, which is usually represented in a matrix form as shown in Figure 13.2.3.
13.3 Data Input and Editing
a. Role of data input and editing
140
Data acquisition occupies about 80 percent of the total expenditure in GIS. Therefore data
input and editing are very important procedures for the use of GIS.
b. Initial data input
Geometric data as well as attribute data are input by the following methods.
(1)
Direct
data
acquisition
by
land
surveying
or
remote
sensing
Vector data can be measured with digital survey equipment such as total stations or
analytical photogrammetric plotters. Raster data are sometime obtained from remote
sensing data.
(2) Digitization of existing maps (see Figure 13.3.1)
Existing maps can be digitized with a scanner or tablet digitizer. Raster data are obtained
from a scanner while vector data are measured by a digitizer. In GIS, raster data and vector
data are frequently converted to vector data and raster data respectively, which are called
raster/vector conversion and vector/raster conversion respectively.
c. Editing
Editing is needed to correct, supplement and add to the initial input data through interactive
communication on a graphic display using the following procedures.
(1) to input manually or interactively those complicated attributes which are not effectively
digitized in the initial input stage.
(2) to correct errors of input data or to supplement with other data.
d. Problems in Data Input and Editing
There are two main problems.
(1) Manual operations
It is difficult to automate data input and editing because of unremovable noise and
incomplete original maps, which result in a large amount of manual work with resultant
inefficiencies in time and cost.
(2) Unreliability of input data
As the input involve many kinds of errors, mistakes and misregistration because of the
manual input, further effort should be applied to obtain data high quality and reliability.
13.4 Spatial Query
141
a. Types of spatial query
Spatial query is a search of the data to satisfy a given condition. There are two types of
spatial query.
(1) Query of attribute data
A spatial distribution or an area will be searched with respect to a given attribute of interest.
(2) Query of geometric data
With a given geometric condition for example location, shape or intersection, all data that
satisfy the condition will be searched. In the case of a vector data form, to search an area
which includes a given point, and to find all line segments which intersect a given line
would be a typical query of geometric data.
In the case of raster form of data, it will be easier to search any attribute and geometric data
based on a given grid.
Figure 13.4.1 shows an example of a query of attribute data in the aster form, in which the
areas with slope gradient of greater than 30 degrees are located.
Figure 13.4.2 shows an example of a query of geometric data in which the area was
searched that includes a point, as given by a cursor.
d. Data Structure for High Speed Query
It is important to develop a data structure which allows for high speed query, because the
data volume is usually very huge.
For example, in order to search all points which are included in an area, it is necessary to
check many points whether those points are included in the area or not.
Tree structure and block structure are typical data structures used to save time of query.
Figure 13.4.3 shows the block structure for solving a point-in-polygon problem, where only
the block that includes a polygon should be checked and searched instead of all other
blocks.
The Quadtree structure has been proposed and used not only for high speed query but also
for data compression.
13.5 Spatial Analysis
142
a. Concept of Spatial Analysis
Spatial analysis is used to produce additional geographic information using existing
information or to enhance the spatial structure or relationship between geographic
information. Many techniques have been proposed, as follows.
b. Production of Additional Geographic Information
The following three techniques are very often used in GIS.
(1) Overlay technique (see Figure 13.5.1)
Various geographic data comprised of multiple layers are overlaid with logical operations
including logical addition or logical multiplication. For example, a hazard risk area of soil
erosion can be estimated by overlaying deforested and slope gradient maps in a
mountainous area.
(2) Buffering technique (see Figure 13.5.2 )
Buffering is to find an area the within a certain distance from a given point or a line. For
example noise polluted areas will be extracted by buffering an area within 30 meter
distance from a trunk road.
(3) Volonoi tessellationAn area may be divided in a group of "influential areas" termed
Volonoi tessellation, that can be formed by bisectors between spatially distributed points.
For example, a school zone can be drawn by Volonoi tessellation between differently
located schools.
c. Statistical Analysis for Spatial Structure
Spatial auto-correlation is one of the statistical techniques to find the spatial structure of
geographic information. Spatial auto-correlation is a correlation factor between two
differently located events. High accuracy spatial interpolation can be executed with a lower
density of samples in the case of high spatial auto-correlation.
d. Combined Technique
Figure 13.5.3 shows an example of a combined technique using remote sensing, buffering
and overlay. In this example, the land use change ratio is tabulated with respect to
accessibility to a railway station and land use zoning.
143
13.6 Use of Remote Sensing Data in GIS
Remote sensing data after geometric correction, can be overlaid on other geographic data
in a raster form. In GIS, there are two uses of use of remote sensing data; as classified data
and as image data.
a. Use of classified data
Land cover maps or vegetation maps classified from remote sensing data can be overlaid
onto other geographic data, which enables analysis for environmental monitoring and its
change.
Figure 13.6.1 shows a case study in which statistical data with lower spatial resolution are
reallocated with a higher spatial resolution using the fact that the remotely sensed data have
higher resolution than the statistical data.
b. Use of image data
Remote sensing data will be classified or analyzed with other geographic data to obtain a
higher accuracy of classification. Figure 13.6.2 shows a comparison between two results
of classification without the use of map data and with the use of map data. If ground height
and slope gradient are given as map data, rice fields, for eg., can be checked and located
only in flat and low land areas. Forest areas and mangrove area are also classified with less
errors if map data are combined with remote sensing data.
Image data are sometimes also used as image maps, with an overlay of political boundaries,
roads, railways etc. Such an image map can be successfully used for visual interpretation.
If a digital elevation model (DEM) is used with remote sensing data, shading corrections
in mountainous areas can be made by dividing by cos q (where q : angle between sun light
and the normal to the sloping surface ).
13.7 Errors and Fuzziness of Geographic Data and their Influences on GIS Products
a. Errors and Fuzziness of Geographic Data
There are various errors in geographic data with respect to error sources. Of those errors,
errors due to the data input method can be avoided by a proper control and check system,
while errors due to measurement methods are difficult to avoid completely. It is necessary
144
for users to evaluate the data errors and their influences by sensitivity analysis. Quality
control of geographic data in also very essential in GIS.
b. Influences of Errors and Fuzziness
Influences of errors and fuzziness are explained in the following two examples.
(1) Influence on spatial query
Consider a case to check any underground pipe which may be damaged by excavation at a
road construction site, as shown in Figure 13.7.1. If only the geometric relationship is
checked, there will be no problem as shown in Figure 13.7.1 (a). However, one should
consider the uncertainly or error of excavation as well as the pipe, which will make possible
the problem as shown in Figure 13.7.1 (b).
(2) Influence on spatial analysis (See Figure 13.7.2)
Consider a case to select suitable land for rice paddy fields by overlaying a slope gradient
map, soil map and irrigation areas in Jogjakarta, Indonesia. The area of suitable land will
change depending it that there is uncertainly along the boundary of the overlaid areas. If
there is 120 meter width of uncertainly along the boundary, the area of suitable land will
reduce by about 50 per cent. Thus this uncertainly or error should always be considered.
FUNDAMENTALS OF GEOGRAPHICAL
INFORMATION SYSTEMS
145
CONTENTS
Chapter 1 What is the GIS?
1.1 Defination of GIS
1.2 Why is a GIS needed?
1.3 Required Functions for GIS
1.4 Computer System for GIS
146
1.5 GIS as a Multidisciplinary Science
1.6 Areas of GIS Applications
1.7 GIS as an Information Infrastructure
1.8 GIS for Decision Support
Chapter 2 Data Model and Structure
2.1 Data Model
2.2 Geometry and Topology of Vector Data
2.3 Topological Data Structure
2.4 Topological Relationships between Spatial Objects
2.5 Geometry and Topology of Raster Data
2.6 Topological Features of Raster Data
2.7 Thematic Data Modeling
2.8 Data Structure for Continous Surface Model
Chapter 3 Input of Geospatial Data
3.1 Required Data Sources for GIS
3.2 Digitizers for Vector Data Input
3.3 Scanner for Raster Data Input
3.4 Digital Mapping by Aerial Photogrammetry
3.5 Remote Sensing with Satellite Imagery
3.6 Rasterization
3.7 Vectorization
3.8 Advanced Technologies for Primary Data Acquisition
Chapter 4 Spatial Database
4.1 Concept of Spatial Database
4.2 Design of Spatial Database
4.3 Database Management System
4.4 Hierachical Model
4.5 Relational Database
4.6 Object Oriented Database
147
Chapter 5 Required Hardware and Software for GIS
5.1 Required Computer System
5.2 Required Functions of GIS Software
5.3 PC Based GIS for Education
5.4 Image Diplay
5.5 Color Hard Copy Machine
5.6 Pen Computer
Chapter 6 Installation of GIS
6.1 Plan for GIS Installation
6.2 Considerations for Installation of GIS
6.3 Keys for Successful GIS
6.4 Reasons for Unsuccessful GIS
6.5 Required Human Resource for GIS
6.6 Cost Analysis of GIS Project
Chapter 1 what is GIS?
1-1 Definition of GIS
Geographic Information System (GIS) is defined as an information system that is used to
input, store, retrieve, manipulate, analyze and output geographically referenced data or
geospatial data, in order to support decision making for planning and management of land
148
use, natural resources, environment, transportation, urban facilities, and other
administrative records.
The key components of GIS are a computer system, geospatial data and users, as shown in
Figure 1.1.
A computer system for GIS consists of hardware, software and procedures designed to
support the data capture, processing, analysis, modeling and display of geospatial data.
The sources of geospatial data are digitized maps, aerial photographs, satellite images,
statistical tables and other related documents.
Geospatial data are classified into graphic data (or called geometric data) and attributes (or
called thematic data) as shown in Figure 1.2. Graphic data has three elements ; point (or
called node), line (or called arc) and area (or called polygon) in either vector or raster form
which represent a geometry of topology, size, shape, position and orientation.
The roles of the user are to select pertinent information, to set necessary standards, to
design cost-efficient updating schemes, to analyze GIS outputs for relevant purpose and
plan the implementation.
1-2 Why is a GIS needed?
These are the following reasons why a GIS is needed.
- geospatial data are poorly maintained
- maps and statistics are out of date
- data and information are inaccurate
- there is no data retrieval service
- there is no data sharing
Once a GIS is implemented, the folllwing benefits are expected.
- geospatial data are better maintained in a standard format
- revision and updating are easier
- geospatial data and information are easier to search, analyze and represent
- more value added product
- geospatial data can be shared and exchanged freely
- productivity of the staff is improved and more efficient
149
- time and money are saved
- better decisions can be made
Table 1.1 shows the advantages of GIS and the disadvantages of conventional manual
works without GIS.
Figure 1.3 shows a comparison between geospatial information management with and
without GIS.
1-3 Required Functions for GIS
The questions that a GIS is required to answer are mainly as follows :
What is at......? (Locational question ; what exists at a particular location)
Where is it.....? (Conditional question ; which locations satisfy certain conditions)
How has it changed........? (Trendy question ; identifies geographic occurrence or trends
that have changed or in the process of changing)
Which data are related ........? (Relational question : analyzes the spatial relationship
between objects of georaphic features)
What if.......? (Model based question ; computers and displays an optimum path, a
suitable land, risky area against disasters etc. based on model)
Figure 1.4 shows examples of questions to be answered by GIS.
In order to meet the above requirements, the following functions are necessary for GIS
(see Table 1.2)
- data acquisition and pre-processing
- data based management and retrieval
- spatial measurement and analysis
- graphic output and visualization
1-4 Computer System for GIS
A Computer system is mainly composed of hardware and software.
a. Hardware system
A hardware system is supported by several hardware components.
Central processing unit (CPU)
150
CPU executes the programs and controls the operation of all components.
Usually a personal computer (PC) or a work station is selected for the required CPU or as
a server computer.
Memory
Main memory :essential for the operation of the computer because all data and program
must be in main memory for fastest access. More than 64 M bytes are at least necessary for
PC based GIS.
Auxiliary memory : is used for large permanent or semi-permanent files with slower
access. Harddisks, floppy disks, magnetic tapes, or optical compact disks (CD-ROM) are
used. At least more than 1 G bytes is required for hard disk in GIS.
Peripherals
Input devices : key board, mouse, digitizers, image scanners, digital cameras, digital
photogrammetric workstations etc.
Output devices : color displays, printers, color plotters, film recorders etc. Figure 1.5
shows an exammples of components of a GIS hardwaare system.
b. Software System
A software system is composed of programs including operating system, compilers and
application programs.
Operating System (OS) : controls the operation of the programs as well as all input and
output.
For Pcs : MS-DOS (IBM PCs) and WINDOWS is the dominant OS.
For Workstations : UNIX and VMS are the dominant OSs.
Compilers : convert a program written in a computer language to machine code so that
CPU can execute binary operation. Commonly used languages include C, Pascal and
FORTRAN and BASIC etc.
Application Programs : Many vendors are providing GIS software systems as listed in
Table 1.3.
1-5 GIS as Multidisciplinary Science
151
GIS in an intergrated multidisciplinary science consisting of the following
traditional disciplines.
Geography
Statistics
Cartography
Operations Research
Remote Sensing
Computer Science
Photogrammetry
Mathematics
Surveying
Civil Engineering
Geodesy
Urban Planning etc.
Table 1.4 summarizes how the above disciplines make up GIS with respect to the
functions. GIS has many alternative names used over the years with respect to the
range of applications and emphasis as listed below.
- Land Information System (LIS)
- AM/FM-Automated Mapping and Facilities Management
- Environmental Information System (EIS)
- Resources Information System
- Planning Information System
- Spatial Data Handling System
GIS is now becoming an indepedent discipline in the name of "Geomatic",
"Geoinformatics"; or "Geospatial Information Science" that is used in many departments
of government and university.
1-6 Area of GIS Applications
Major areas of GIS application can be grouped into five categories as follows.
Facilities Management
Large scale and precise maps and network analysis are used mainly for utility management.
AM/FM is frequently used in this area.
Environment and Natural Resourrces Management
152
Medium or small scale maps and overlay techniques in combination with aerial
photographs and satellite images are used for management of natural resources and
environmental impact analysis.
Street Network
Large or medium scale maps and spatial analysis are used for vehicle routing, locating
house and streets etc.
Planning and Engineering
Large or medium scale maps and engineering models are used mainly in civil enginerring.
Land Information System
Large scale cadastre maps or land parcel maps and spatial analysis are used for cadastre
administration, taxation etc.
Table 1.5 summarizes the major areas of GIS applications.
1-7 GIS as an Information Infrastructure
Information has become a key issue in the age of computer, space technology and
multimedia, because the information infrastructure contributes to the quality of life as in
the follwings infrastructure.
Social infrastructure...better society
Environmental infrastructure....better management
Urban infrastructure.....better life
Economic infrastructure.......better business
Educational infrastructure......better knowledge
Figure
1.6
shows
major
components
of
GIS
information
infrastructure.
In order to achieve the GIS information infrastructure, the following issues should be
solved and promoted (see Figure 1.7)
Open data policy
GIS data and information should be accessible by any user, freely or at inexpensive costs
and without restriction.
153
Standardization
Standards for data format and structure should be developed to enable transfer and
exchange of geospatial data.
Data/Information sharing
In order to save cost and time for dizitization, data sharing should be promoted. In order to
foster operational use of geospatial data, information and experience should be shared
among users.
Networking
Distributed computer systems as well as databases should be linked to each other to a
network for better access as well as better service.
Multi-disciplinary approach
Because GIS is a multi-disciplinary science, scienctists, engineers, technicians and
administrators of different fields of study should cooperate with each other to achieve the
comman goals.
Interoperable procesure
GIS should be interwoven with other proceduces such as CAD, computer graphics, image
processing, DEM etc.
1-8 GIS for Decision Support
GIS can be a very important tool in decision making for sustainable development, because
GIS can provide decision makers with useful information by means of analysis and
assessment of spatial database as shown in Figure 1.8.
Decision making including policy making, planning and management can be interactively
implemented taking into consideration human driving forces through public consensus.
Driving forces include population growth, health and wealth, technology, politics,
economics etc. by which human society will set up targets and goals on how to improve
the quality of life.
Thus human driving forces, the key elements of human dimensions, will give impacts on
the environment such as development of natural resources, urbanization, industrializations,
construction, energy consumption etc. These human impacts will accordingly induce
154
environmental changes such as land use change, change of life style, land degradation,
pollution, climate change etc. Such environmental change should be timely monitored to
increase public awareness. Remote sensing can be very useful for better understanding of
relationship between human impacts and the environmental change as well as for building
databases.
Physical dimensions monitored by remote sensing can be fed back to human dimensions
through analysis and assessment by GIS in order to support better decision. In this sense,
remote sensing should be integrated with GIS.
Chapter 2 Data Model and
Structure
2-1 Data Model
155
The data model represents a set of guidelines to convert the real world (called entity) to the
digitally and logically represented spatial objects consisting of the attributes and geometry.
The attributes are managed by thematic or semantic structure while the geometry is
represented by geometric-topological structure.
There are two major types of geometric data model ; vector and raster model, as shown in
Figure 2.1
a. Vector Model
Vector model uses discrete points, lines and/or areas corresponding to discrete objects with
name or code number of attributes.
b. Raster Model
Raster model uses regularly spaced grid cells in specfic sequence. An element of the grid
cell is called a pixel (picture cell). The conventional sequence is row by row from the left
to the right and then line by line from the top to bottom. Every location is given in two
dimensional image coordinates ; pixel number and line number, which contains a single
value of attributes.
2-2 Geometry and Topology of Vector Data
Spatial objects are classfied into point object such as meteorological station, line object
such as highway and area object such as agricultural land, which are represented
geometrically by point, line and area respectively. For spatial analysis in GIS, only the
geometry with the position, shape and size in a coordinate system is not enough but the
topology is also required.
Topology refers to the relationships or connectivity between spatial objects.
The geometry of a point is given by two dimensional coordinates (x, y), while line, string
and area are given by a series of point coordinates, as shown in Figure 2.2 (a). The topology
however defines additional structure as follows (see Figure 2.2 (b)).
Node : an intersect of more than two lines or strings, or start and end point of string with
node number
Chain : a line or a string with chain number, start and end node number, left and right
neighbored polygons
156
Polygon : an area with polygon number, series of chains that form the area in clockwise
order (minus sign is assigned in case of anti-clockwise order).
2-3 Topological Data Structure
In order to analyze a network consisting of nodes and chains, the following topology should
be built.
Chain : Chain ID, Start Node ID, End Node ID, Attributes
Node: Node ID, (x, y), adjacent chain IDs (positive for to node, negative for from node)
In order to analyze not only a network but also relationships between polygons, the
following addtional geometry and topology are required as shown in an example of Figure
2.3.
Chain geometry : Chain ID, Start Coordinates, Point Coordinates, End Coodinates
Polygon topology : Polygon ID, Series of Chain ID, in clockwise order (Attritutes)
Chain topology : Chain ID, Start Node ID, End Node ID, Left Polygon ID, Right Polygon
ID, (Attributes)
The advantages of the topological data model are to avoid duplication in digitizing common
boundaries of two polygons and to solve problems when the two versions of the common
boundary do not coincide.
The disadvantages are to have to build very correct topological data sets without any single
error and to be unable to represent islands in a polygon.
2-4 Topological Relationships between Spatial Objects
In practical appliations of GIS, all possible relationships in spatial data should be used
logically with more complicated data structures.
The following topology relationships are commonly defined.
a. Point-Pont Relationship
"is within" : within a certian distance
"is nearest to" : nearest to a certain point
b. Point-Line Relationships
"on line" : a point on a line
"is nearest to" : a point nearest to a line
157
c. Point-area Relationships
"is contained in’’ : a point in an area
"on border of area" : a point on border of an area
d. Line-Line Relationships
"intersects" : two lines intersect
"crosses" : two lines cross without an intersect
"flow into" : a stream flows into the river
e. Line-Area Relationship
"intersects" : a line intersects an area
"borders" : a line is a part of border of an area
f. Area-Area Relationships
"overlaps" : two areas overlap
"is within" : an island within an area
"is adjacent to" : two area share a common boudary
Figure 2.4 shows the several topological relationships between spatial objects.
Figure 2.5 shows geometric and topological modeling between point, line and area.
2-5 Geometry and Topology of Raster Data
The geometry of raster data is given by point, line and area objects as follows (see Figure
2.6(a)).
a. Point objects
A point is given by point ID, coordinates (i, j) and the attributes
b. Line object
A line is given by line ID, series of coordinates forming the line, and the attributes
c. Area objects
An area segment is given by area ID, a group of coordinates forming the area and the
attributes. Area objects in raster model are typically given by "Run Length" that rearranges
the raster into the sequence of length (or number of pixels) of each class as shown in Figure
2.6 (a).
158
The topology of raster model is rather simple as compared with the vector model as shown
in Figure 2.6 (b).
The topology of line objects is given by a sequence of pixels forming the line segments.
The topology of an area object is usually given by "Run Length" structure as follows.
- start line no., (start pixel no., number of pixels)};
- second line no., (start pixel no., number of pixels)};
-6 Topological Features of Raster Data
One of the weak points in raster model is the difficulty in network and spatial analysis as
compared with vector model.
For example, though a line is easily identified as a group of pixels which form the line, the
sequence of connecting pixels as a chain would be a little difficult in tracing. In case of
polygons in raster model, each polygon is easily indentified but the boundary and the node
(when at least more than three polygons intersect) should be traced or detected.
a. Flow directions
A line with directions can be represented by four directions called as the Rook's move in
the chess game or eight directions called as the Queenís move, as shown in Figure 2.7 (a)
and (b).
Figure 2.7 (c) shows an example of flow directions in the Queen's move. Water flow, links
of a network, roads etc. can be represented by the flow directions (or called Freeman chain
code).
b. Topological Features of Raster Data
Boundary is defined as 2 x 2 pixel window that has two different classes as shown in Figure
2.8 (a). If a window is traced in the direction shown in Figure 2.8 (a), the boundary can be
indentified.
c. Node
A node in polygon model can be defined as a 2 x 2 window that has more than three
different classes as shown in
Figure 2.8 (b).
Figure 2.8 (c) and (d) show an example of identification of pixels on boundary and node.
159
2-7 Thematic Data Modeling
The real world entities are so complex that they should be classified into object classes
with some similarity through thematic data modeling in a spatial database.
The objects in a spatial database are defined as representations of real world entities with
associated attributes.
Generally, geospatial data have three major components; position, attributes and time.
Attributes are often termed "thematic data" or "non-spatial data", that are linked with
spatial data or geometric data.
An attribute has a defined characteristic of entity in the real world.
Attribute can be categorized as normal, ordinal, numerical, conditional and other
characteristics. Attribute values are often listed in attribute tables which will establish
relationships between the attributes and spatial data such as point, line and area objects,
and also among the attributes.
Figure 2.9 shows a schematic diagram of thematic data modeling.
Spatial objects in digital representation can be grouped into layers as shown in Figure 2.10.
For example, a map can be divided into a set of map layers consisting of contours,
boundaries, roads, rivers, houses, forests etc.
2-8 Data Structure for Continuous Surface Model
In GIS, continuous surface such as terrain surface, meteorolgical observation (rain fall,
temperature, pressure etc.) population density and so on should be modeled. As sampling
points are observed at discrete interval, a surface model to present the three dimensional
shape ; z = f (x, y) should be built to allow the interpolation of value at arbitrary points of
interest.
Usually the following four types of sampling point structure are modeled into DEM.
Grid at regular intervals :
Bi-linear surface with four points or bi-cubic surface with sixteen points is commonly used
Random points :
160
Triangulated irregular network (TIN) is commonly used. Interpolation by wieghted
polynomails is also used.
Contour lines :
Interpolation based on proportional distance between adjacent conmtours is used. TIN is
also used.
Profile :
Profiles are observed perpendicular to an alignment or a curve such as high ways. In case
the alignment is a straight line, grid points will be interpolated. In case the alignment is a
curve, TIN will be generated.
Figure 2.11 shows different types of DEMs.
Chapter 3 Input of Geospatial
Data
3-1 Required Data Sources for GIS
As data acquisition or data input of geospatial data in digital format is most expensive
(about 80% of the total GIS project cost) and procedures are time consuming in GIS, the
data sources for data acquisitions should be carefully selected for specific purposes.
161
The following data sources are widely used.
Analog maps
Topographic maps with contours and other terrain features and thematic maps with respect
to defined object classes are digitized by digitizers manually or by scanners semiautomatically. Problems of analog map are lack of availability, out of date, inconsistency
in map production time, inaccurate etc.
Aerial photographs
Analytical or digital photogrammetry is rather expensive but the best method for updating
Satellite image
Satellite images or data are available for land use classification, digital elevation model
(DEM), updating highway network etc. But the image map scale would be around 1:50,000
to 1:100,000.
High resolution satellite image with ground resolution of 1~3 meters will produce 1:25,000
topomaps in near future.
Ground survey with GPS
Total station together with GPS (Global Positioning System) will modernize the ground
survey. It is very accurate but too expensive to cover wide areas.
Reports and publications
Social economic data are usually listed in the reports of statistics and census with respect
to administration units.
Figure 3.1 summarizes major data sources for GIS.
Table 3.1 shows method, equipments, accuracy and cost for different data sources.
3-2 Digitizers for Vector Data Input
Tablet digitizers with a free cursor connected with a personal computer are the most
common device for digitizing spatial features with the planimetric coordinates from analog
maps. The analog map is placed on the surface of the digitizing tablet as shown in Figure
3.2. The size of digitizer usually ranges from A3 to A0 size.
The digitizing operation is as follows.
Step 1 : a map is affixed to a digitizing table.
162
Step 2 : control points or tics at four corners of this map sheet should be digitized by the
digitizer and input to PC together with the map coordinates of the four corners.
Step 3 : map contents are digitized according to the map layers and map code system in
either point mode or stream mode at short time interval.
Step 4 : editing errors such as small gaps at line junctions, overshoots, duplicates etc.
should be made for a clean dataset without errors.
Step 5 : conversion from digitizer coordinates to map coordinates to store in a spatial
database.
Major problems of map digitization are :
- the map will stretch or shrink day by day which makes the newly digitized points
slightly off from the previous points.
- the map itself has errors
- discrepancies across neighboring map sheets will produce disconnectivity.
operators will make a lot of errors and mistakes while digitizing as shown in Figure 3.3.
3-3 Scanners for Raster Data Input
Scanner are used to convert from analog maps or photographs to digital image data in raster
format. Digital image data are usually integer-based with one byte gray scale (256 gray
tones from 0 to 255) for black and white image and a set of three gray scales of red (R),
green (G) and blue(B) for color image.
The following four types of scanner are commonly used in GIS and remote sensing.
a. Mechanical Scanner
It is called drum scanner since a map or an image placed on a drum is digitized
mechanically with rotation of the drum and shift of the sensor as shown in Figure 3.4(a). It
is accurate but slow.
b. Vedeo Camera
Video camera with CRT (cathode ray tube) is often used to digitize a small part of map of
firm. This is not very accurate but cheap. (see Figure 3.4(b))
163
c. CCD Camera
Area CCD camera (called digital still camera) instead of video camera will be also
convenient to acquire digital image data (see Figure 3.4 (c)). It is more stable and accurate
than video camera.
d. CCD Scanner
Flat bed type or roll feed type scanner with linear CCD (charge coupled device) is now
commonly used to digitize analog maps in raster format, either in mono-tone or color mode.
It is accurate but expensive.
Table 3.2 shows the performance of major scanners.
3-4 Digital Mapping by Aerial Photogrammetry
Though aerial photogrammetry is rather expensive and slow in air flight as well as
subsequent photogrametric plotting and editing, it is still very important to input accurate
and up-to-date spatial information. Aerial photogrammetry needs a series of the procedures
including aerial photography, stereo-plotting, editing and output as shown in Figure 3.5.
There are two types of aerial photogrammetry.
a. Analytical photogrammetry
Though computer systems are used for aerial triangulation, measuring map data, editing
and output with pen plotter, a stereo pair of analog films are set up in a stereo plotter and
the operator will manually read terrain features through stereo photogrammetric plotter
called analytical plotter.
b. Digital Photogrammetry
In digital photogrammetry, aerial films are converted into digital image data with high
resolution (5stereo matching using digital photogrammetric workstation. Digital ortho photo and 3D
bird's eye view using DEM will be also automatically created as bi-products. It is still very
expensive but only a method for automated mapping. There is a need for futher research
for identifying the patterns of houses, roads, structures and other terrain features
automatically, that is so called image understanding.
Figure 3.6 shows a digital photogrammetric workstation.
164
3-5 Remote Sensing with Satellite Imagery
Satellite remote sensing is a modern technology to obtain digital image data of the terrain
surface in the eletro-magnetic region of visible, infrared and microwave.
Multi-spectral bands including visible, near-infrared and/or thermal infrared are most
commonly used for production of land use map, soil map, geological map, agricultureal
map, forest map etc. at the scale of 1:50,000 ~ 250,000. A lot of earth observation satellites
for example landsat, SPOT, ERS-1, JERS-1, IRS, Radarsat etc. are available.
Synthetic aperture radar (SAR) is now becoming a new technology in remote sensing
because SAR can penetrate through clouds, which enables cloud free imagery in all
weather conditions.
Satellite images have different ground resolutions depending on the sensors used as listed
in Table 3.3.
Since the cold war was over in 1990's, very high resolution satellite imagery with ground
resolution of 1 to 3 meters will become available from 1998. Such high resolution satellite
images are expected to identify individual houses in urban area. Table 3.4 shows high
resolution satellites proposed to be launched by three US commercial companies.
The high resolution satellite images are highly expected to apply to urban GIS.
3-6 Rasterization
Conversion between raster and vector data is very useful in practical applications of GIS.
Rasterization refers to conversion from vector to raster data. Raster format is more
convennient to produce color coded polygon maps such as color coded land use map, while
map digitizing in vector format is more easier to trace only the boundary. Rasterization is
also useful to integrate GIS with remote sensing becaues remote sensing images are in
raster format.
A simple algorithm for calculation of trapezoid area can be applied to convert vectorized
polygon to rasterized polygon with grid cells as shown in Figure 3.7. If vertical lines are
drpped to the x axis from two adjacent vertices, a trapezoid will be formed as shown in
Figure 3.7.
The area of trapezoid is given by
165
Ai = (xi+1- xi) (yi + yi+1) / 2
The sum of all trapezoids will give the area of the original polygon as shown in Figure
3.7. Using this algorithm, the grid cells in the polygon are easily identified as shown in
the upper right of Figure 3.7.
3-7 Vectorization
Vectorization refers to conversion from raster to vector data, which is often called raster
vector conversion. Vectorization is not very easy as compared with rasterization, because
vector format needs topological structure, for example, direction of line or chain, boundaris
and nodes of polygons, order of series of chains that form a polygon, left and right polygons
ID of a chain and so on.
A simple algorithm of vectorization is explained in Figure 3.8, in which the original image
in raster format is converted to vector data through thinning and chain coding (see 2.6).
This algorithm is useful to convert raster image to vector data with the coordinates, but it
is not sufficient because the agorithm will not build topological structure.
Raster vector conversion with automatic building of topology is possible if 2 x 2 window
is continuously moved along the boundary from a node. The boundary and node can be
identitified by the method as described in 2.6 and Figure 2.8.
Figure 3.9 shows schematically the raster vector conversion by which left and right
polygons are identified.
In order to automate raster vector conversion as much as possible, a clean image without
noises or unnecessary marks should be scanned in the beginning.
3-8 Advanced Technologies for Primary Data Acquisition
Several advanced technologies have become available for primary data acquisition of
geospatial data as well as digital elevation model (DEM).
Following advanced technologies will be useful for future GIS.
a. Electronic Plane Surveying System
An integrated system of total station with automated tracking function, kinematic global
positioning system (GPS) and a pen computer (see Figure 3.10 (a)) will replace the
conventional plane surveying. Direct data acquisition in digital form, at the field site will
166
be very useful for large scale GIS data set, for example in the application to cadastre, utility
facilities, urban structures etc.
b. Mobile Mapping System
Different sensors such as GPS, INS (inertia navigation system), more than two digital
cameras, voice recorder etc. are fixed on a vehicle as shown in Figure 3.10 (b) in order to
map objects in close range. For example center line of highways, utility lines, railways etc.
as well as to determine trajectory of the moving vehicle.
c. Laser Scanner
Airborne laser scanner together with GPS and INS will measure directly the terrain releif
or DEM as shown in Figure 3.10 (c) with the height accuracy of 10 cm up to the altitude
of 1,000 m.
d. SAR Interferometry
SAR (synthetic apperture radar) inter-ferometry is a new technology to produce DEM
automatically by special interferometric processing of a pair of SAR images. Airborne
and space borne SAR interferometry are now available if the inter-ferometric condition
meets the standard.
Chapter 4 Spatial Database
4-1 Concept of Spatial Database
A spatial database is defined as a collection of inter-related geodspatial data, that can
handle and maintain a large amout of data which is shareable between different GIS
applications.
Required functions of a spatial database are as follows.
- consistency with little or no redundancy
- maintenance of data quality including updating
167
- self descriptive with metadata
- high performance by database mangement system with database langauage
- security including access control
In 1980's, GIS institution was centralized with a centralized spatial database. But in 1990's,
the network concept arose, whice will be more convenient to meet user needs with
distributed databases as shown in Figure 4.1. Such distributed databases in a network
structure have the following benefits.
- better data storage and updating
- more efficient retrieval
- more efficient output
4-2 Design of Spatial Database
The design of spatial database will be made by the database manager who is responsible
for the following issues.
- definition of database contents
- selection of database structure
- data distribution to users
- mainteneance and updating control
- day-to-day operation
For the design of detail items, the following parameters should be well designed.
Storage media
Volume, access speed and on line service should be considered. Table 4.1 shows the
different types of storage media
partition of data
Choice of administrative boundaries, map sheets, watersheds etc. will be made in
consideration of GIS applications (see Figure 4.2)
standards
Format, accuracy and quality should be standardized.
change and updating
168
Add, delete, edit and update should be well controlled by the database manager.
scheduling
Data availability, priorities, data acquisition etc. should be well scheduled.
security
Copyright, back up system and responsibilities should be well managed.
4-3 Database Management System
A database management system (DBMS) provides a number of functions to create, edit,
manipulate and analyse spatial and non-spatial data in the applications of a GIS.
Major functions of a database are as follows :
- creating records of various data types; integer, real, character, data, image etc.
- operations ; sort, delete, edit, select etc.
- manipulation ; input, analysis, output, reformatting etc.
- query ; will be made by a standardized language such as SQL (Satandard Query
Language)
- programming ; will be useful for application programs
- documentation ; metadata or description of the contents of the database should be
complied.
There are four types of database models :
- hierarchical model
- network model
- relational model
- object oriented model
Although all four types are used, the relational model has been most successful in GIS.
Well known relational databases include dBase, Oracle and Info. Object oriented model is
a new concept that has been recently developed.
Threre has been debate on which of the two-layers or object oriented is efficient in GIS.
Layers may be efficient for natural resources management, for example with different
layers of land use, soil, geology, agriculture, forests etc.
169
On the other hand object orientation may be more convenient for facility management with
grouped attributes.
Figure 4.3 shows the concept of four types of database model.
4-4 Hierarchical Model
Several records or files are hierachically related with each other. For example, an
organization has several departments, each of which has attributes such as name of director,
number of staffs, annual products etc.
Each department has several divisions with attributes of name of manager, number of
staffs, annual products etc.
Then each division has several sections with attributes such as name of head, number of
staff, number of PCs etc.
Hierachical model is a type of tree structure as shown already in Figure 4.3 (a). A set of
links connect all record types in a tree structure.
The advantages of hierachical model are high speed of access to large datasets and eases
of updating. However the disadvantage is that linkages are only possible vertically but not
horizontally or diagonally, that means there is no relation between different trees at the
same level unless they share the same parent.
The Quadtree, that is used to access a small part of a large raster image or map area, is a
type of hierachical model. Quadtree first divides a total map area into 4, 16, 32, .... step by
step as shown in Figure 4.4 (a).
Secondly a quadtree is built as shown in Figure 4.4 (b), that makes the access to a particular
area at high speed. Numbering of 0, 1, 2 and 3 known as Morton order, makes effective
coding of a block or a pixel in a raster model. For example the block of 211 in Figure 4.4
(a) can be expressed 100101 in a pair of base 2 digits, while the conventional block number
(4, 3), line and row number needs more bits in a computer.
4-5 Relational Database
Relational database is the most popular model for GIS. For example, the following
relational database softwares are widely used.
170
- INFO in ARC/INFO
- DBASE III for several PC-based GIS
- ORACLE for several GIS uses
Relational database is a model to link the complex spatial relationships between objects.
The spatial objects are tabulated in tables consisting of records with a set of attributes as
shown already in Figure 4.3 (c). Each table (called relation) consists of different number
of attributes, which is called the degree. The degree of an attributes refers to n-ary (ex.
unary, binary etc.) relation.
In a relational model, the following two important concepts should be defined.
Key of relation ; a subset of attributes
Unique identification ; e.g. the key attributes is a phone directory in a set of last name, first
name and address.
non redundancy ; any key attribute selected and tabulated should keep the key's uniqueness.
e.g. address can not be dropped from telephone address, because there may be many with
the same names.
Prime attribute : an atribute listed in at least one key.
The most important point of the relational database design is to build a set of key attributes
with a prime attribute, so as to allow dependence between attributes as well as to avoid loss
of general information when records are inserted or deleted.
Table 4.2 shows how to build the relational database by normalizing an unstructured table.
The advantages of the relational database are as follows.
- there is no redundancy.
- type of building of an owner can be changed without destroying the relation between type
and rate.
- a new type of building for example "Clay" can be inserted.
4-6 Object Oriented Database
An Object Oriented model uses functions to model spatial and non-spatial relationships of
geographic objects and the attributes.
171
An object is an encapsulated unit which is characterized by attributes, a set of orientations
and rules.
An object oriented model has the following characteristics.
generic properties : there should be an inheritance relationship.
abtraction : objects, classes and super classes are to be generated by classification,
generalization, association and aggregation.
adhoc queries : users can order spatial operations to obtain spatial relationships of
geographic objects using a special language.
For example, let us try to represent a thought : "Hawaii is an island that is a state of USA"
in GIS. In this case, we don’t mind the geographic location with latitude and longtitude in
the conventional GIS model. This is not appropriate to use the layers. In an object oriented
model, we are more careful with spatial relationships for example, "is a" (island is a land)
and "part of" (state is a part of country). In addition Hawii (state) has Honolulu City and
also is in Pacific Region. Figure 4.5 (a) shows "is a" inheritance for the super class of land,
while Figure 4.5 (b) shows the spatial relationships for the object of state.
An object oriented database is based on a semantic model as shown in Figure 4.6, which is
usually manged by a spacial language although the language has not yet been fully
completed.
172
Chapter 5 Required Hardware and
Software for GIS
5-1 Required Computer System
In 1990’s, a distributed processing system having the functions of networking with other
computer systems has been a technical trend, particularly in GIS. This is often called
"Client Server Architecture" or "Computing Network". Networks can be linked with LAN
(Local Area Network) using optical fiber or coaxial cables, WAN (Wide Area Netrwork)
or Internet.
Users can select an optimal combination of computers such as personal computers and
UNIX workstations which can be connected to each other even at different locations.
173
There should be a rather powerful computer (usually UNIX workstation) as so called
"client server" with big memory and disk capacity, that can be shared with many other
computers connected in the network. LAN can be also connected with input and output
machines as well as with public telephone lines, mobile telephones, microwave and private
telephone lines.
Although personal computers (PCs) had a lot of limitation in memory capacity, processing
speed and functions , PCs have become powerful with 32 - bit microprocessor, bigger
memory size (64 MB) and bigger hard disk (1 GB) available at very reasonable costs (about
2,000 US dollars depending on the configuration). PCs are useful to control printers,
digitizers and color plotters.
UNIX workstations with the function of multi-processing are of course more powerful than
PCs in memory size, processing speed and other functions.
A UNIX workstation as a client server can be connected with PCs or X terminals as a
controller of input and output devices in GIS.
Figure 5.1 shows a typical GIS computer system.
5-2 Required Functions of GIS Software
In practical GIS applications, a lot of software are required for input, manipulation,
processing, analysis and output of spatial data, both in vector and raster format.
The following functions are required for rather wide GIS applications (see Table 5.1)
Operating System (OS)
Unix for workstation, MS-DOS
WINDOWS for PC
Data Input
Map digitizing and editing for vector based GIS
Map/Photo scanning for raster based GIS
Color separation
Database Management
Relational database software
Database integrator for data exchange through network
174
Spatial Analysis
Vector Data Analysis
building topology
spatial query
buffering
mixing layers
overlay of layers
network analysis (route finding, tracing etc.)
Digital Terrain Model (DTM)
Overlays of layers
buffering
raster vector conversion
TIN
Grid based DEM
Drainage Analysis
Shading
Oblique Views of Bird’s Eye Views
Image Proceessing
Image Enhancement
Color Manipulation
Classification
Image Analysis/Measurement
Mathematical Morphology
Mapping System/Data Output
Map Projection
Graphic Representation
Cartographic Output
Vector raster conversion
Raster Data Analysis
175
Table 5.1 shows a comparison between two major GIS softwares ; MGE of Intergragh
and ARC/INFO of ESRI.
5-3 PC Based GIS for for Education
PC based GIS is necessary for education and training with minimum functions at
inexpensive rate, particularly in developing countries.
The following PC based GIS will be recommendable.
Hardware
- PC : Pentium : 32 bits microprocessor
- 64 M bytes memory
- 1 G bytes and more hard disk
- Floppy disk
- CD-ROM driver
- Color Graphic Monitor (17 inches is preferable)
- Digitizer (A3 size at minimum)
- Laser Printer
- Color Ink Jet Plotter
- UPS (power stabilizer)
Software
- OS : WINDOWS
- Compiler : C, FORTRAN etc.
- Public domain or cheaper GIS software ; GRASS (free), IDRISI (560 US$), PC
ARC/INFO (500 US$ only for education)
For education and training, a PC will be better, shared by two students or trainees at
maximum. Therefore 10 PC’s are required for a maximum capacity of 20 students or
trainees.
- Teaching materials such as text books, manual of software, educational dataset for map
digitizing and analysis etc. should be well organized and prepared. For hands-ontraining (HOT) with PC based GIS, at least two- week long courses (three weeks are
preferable) should be planned.
176
Table 5.2 shows a typical curriculum of short course for elementary GIS.
5-4 Image Display
An image display is to display a vector map or image data on a color monitor. An image
display consists of frame buffer, look up table, D/A converter and monitor (or called
display) as shown in Figure 5.2.
Frame Buffer
Frame buffer is a memory device that stores the image data and reads at memory board.
The memory capacity ranges 512 x 512 x 3 bytes to 2,048 x 2,048 x 3 bytes.
Look Up Table
User can assign a look up table that makes image transformation in real time. The table
includes linear function, gamma function, log function etc. as shown in Figure 5.3.
D/A Converter
D/A converter will convert digital image data to analog video signals of red (R), green(G)
and blue(B) for image display on an analog color monitor.
The video signals will be usually NTSC in interlace mode or RGB independent in noninterlace mode, with 525 scan lines or pixels typically. High vision video with high
resolution of 1,024 scan lines is now becoming available.
Monitor (Display)
Converted analog video signals are displayed on a monitor of CRT (cathode ray tube) or
liquid crystal monitor.
Image display is classified into two types:
RGB type
Independent frame buffer and look up table is installed to each primary color of R, G and
B. Full color means that 8 bits are assigned to R, G and B which makes 256 x 256 x 256 =
16,777,216 colors. (see Figure 5.4)
Color Map type
A limited and integrated frame buffer and look up table are installed in rather cheaper
display. For example, in case of only 8 bits for frame buffer and look up table, only 256
colors will be generated (see Figure 5.5).
177
5-5 Color Hard Copy Machine
A color hard copy machine is a data output equipment to produce color hard copies in the
GIS applications.
There are several different color hard copy machines as listed below.
Pen Plotter
Usually four different colors of black, blue, green and red are available to draw line maps
or diagrams in vector mode. In case of color painted polygon map, hatching pattern with
fine and parallel lines will be used. The advantage is cheap cost, while the disadvantages
are limited color selections, slow output and limited only to vector map.
Color Ink Jet Recorder
A color ink jet recorder is to produce a raster image with fine and color ink dots with the
resolution of 200~400 dpi (dots per inch). Uaually toners and three kinds of yellow, cyan
and magenta inks are available, the mixture of which makes a variety of colors with dot
matrix patterns. The advantages are cheap cost and rapid color output, while the
disadvantage is color degradation with time elapse and/or wet hands.
Color Dot Printer
A color dot printer is based on electrostatic theory, that can produce very fast color vector
maps as well as raster images up to the size of AO. The advantage is high speed output,
while the disadvantage is high cost.
Film Recorder
There are two types of color films recorder, drum scanner with ordinary film, and thermal
electronic color printer with special coated paper and toner.
The advantages are fine resolution of 300~1,000 dpi and continuous full color tones, while
the disadvantage is high cost and rather slow speed.
Figure 5.6 shows major color hard copy machines.
Table 5.3 shows a comparison between different types of color hard copy machines.
5-6 Pen Computer
178
A pen computer is a portable type of personal computer as shown in Figure 5.7, which are
very useful for on spot GIS data acquisition at the site.
It can be connected with a mobile telephone (or called cellular telephone) or a public
telephone through which the stored data can be tranmitted to the host computer with GIS
database.
Because of its light weight, it will be vey useful to carry a pen computer together with a
kinematic GPS system to measure geospatial objects which is so called GPS survey, as
shown in Figure 5.8.
The pen computer can be also connected with a GPS based total station as shown in Figure
5.9, which has the function of geodetic surveying for spatial objects as well as data input
of the attributes at the site in digital form.
The pen computer is designed against water, dust and shock as well as with working range
from -5° to 50° C in temperature, which is stronger than the ordinary laptop computer in
the open field. This system will replace the conventional plane surveying of analog
mapping, which is commonly used in the field.
Chapter 6 Installation of GIS
6-1 Plan for GIS Installation
Installion planning is divided into three phases; planning, anlysis and implementation.
The First Phase : Planning
Step 1 : Proposal of Plan
The objectives, rationale, system configuration, database, budget, scheduling etc. should
be proposed.
Step 2 : Review of Plan
The proposed plan should be circulated and explained at the related departments or
divisions to obtain the consensus in the bottomup approach.
179
Step 3 : Approval of Plan
The plan should be approved by the top-manager particularly with respect to the policy and
strategy of GIS installation.
Step 4 : Organization of Project Team
A project with well defined terms of reference should be organized.
The Second Phase : Analysis
Step 5 : Pilot Scale Feasibility Study
A pilot scale feasibilitiy Study should be made by the project team in consultation with
GIS experts.
Step 6 : Approval of Pilot Project
The pilot project will be approved particularly with respect to the budget.
Step 7: Drafting of Specification
Specifications for hardware and software, as well as database structure should be made.
Step 8 : Selection of Vendor
A vendor will be selected through bidding.
The Third Phase : Implementation
Step 9 : Design of Database
Data acquisitions, maintenance and updating with well defined data format and database
model should be designed.
Step 10 : Implementation of Pilot Project
A small scale GIS project should be tested as a pilot project in coorperation with the
selected vendor
Step 11 : Review of Pilot Project
Database design, data input cost, performances of hardware and software and so on
should be reviewed and improved.
Step 12 : Purchase Order of Systems
Hardware and software should be purchased.
The following steps are common flow of the last stage.
180
Step 13 : Training
Step 14 : Data Input
Step 15 : Daily Operation
Figure 6.1 summarieze the installation planning.
6-2 Considerations for Installation of GIS
The following items should be taken into consideration for installation of GIS
a) Cost
The cost of GIS is very important with regard to the installation cost and operational cost
as enumerated.
Installation Cost
Cost of hardware and software with respect to the requirements
Cost of Data Input
Cost of Database Management
Training Cost
Cost of Application Software
Version Up of Hardware and Software
Other Necessary Facilities and Equipments
Operation Cost
Maintenance of Hardware
Cost for updating of Database
Cost for Data Analysis
Cost for Data Output
Cost for Archiving System/Back Up System
b) Functions of GIS
Data Input Selection of Geospatial Data
Data Model and Strucutre
Digitizing Methods and Tools
Error Check and Correction
Database Management system
181
Data Processing
Map Projection/Map Production
Map Mozaicing
Topological Structure
Raster-Vector, Vector-Raster Conversion
Spatial Analysis
Overlay
Query of Spatial Data and the Attributes
Measurement of spatial Data
DEM
Network Analysis
c) Support by Vendor
Maintenance of Hardware and Software
After Services
New Products
Service Persons
d) Support for Users
Training
Provision of Metadata
On Line Help Service
Data Access/Exchange
Application Package
Figure 6.2 summariezes the considerations for installation of GIS.
6-3 Keys for a Successful GIS
The following six keys are the most important factors for a sucessful GIS.
Data Input
As the cost of data input will occupy about 80 percent of the total cost in GIS, the first key
should be data input. More attention should be given to selection and classification of
required geospatial data for the project taking into consideration of the digitizing method.
182
Maintenance of Database
The second key is the maintenance of database, particularly maintaining data quality and
routinely updating the system.
Consensus of Supporters
Not only top managers but also other administrative staffs and engineers should support
the GIS project.
Customizing Sotfware
As the existing GIS softwares provided by vendors are not enoght for practical
applications, the users should develop customized software or solution for the problem by
building a model and programming an application package.
Data Sharing
Data sharing is one of the important keys to minimize the total cost of data input and also
to maximize the use of the database. Political and administrative problems should be solved
to promote the data sharing for a successful GIS.
Education and Training
Edcudation and training are also very important for understanding GIS concept, goals and
techniques. They should be organized in three levels for makers, professionals and
technicians.
Figure 6.3 summarizes the keys for a successful GIS
6-4 Reasons for an Unsuccessful GIS
These are the following six major reasons that may lead to a unsuccessful GIS.
Lack of Vision
The objectives, targets and goals of a GIS project were not definded by the top manager,
who just purchased GIS hardware and software only by the name of GIS. In such case, a
GIS is only a toy for the top manager.
Lack of Long Term Planning
183
One should note that GIS projects are long term projects that run for about ten years at
least. The budget for version up and updating the database is sometimes not prepared, and
as such cannot keep the GIS project running.
Lack of Support by Decsion Makers
On some occasion, the top manager in charge of GIS is replaced by an other person who is
not very supportive to the GIS project.
Lack of System Analysis
Digital approach with GIS in replacement of the existing analog approach based on manual
works is sometimes not acceptable in the existing conventional system. The restructuring
of the organizationo and reducation/restraining is not implemented.
Lack of Expertise
Imprper selection and misuse of GIS hardware and software very often occur due to lack
of experstise. Professional consultants or experts should be invited to evaluate the plan.
Lack of Access for User’s
There would be very few user if the training for users is not well organized and not provided
with a well organized manual. Sometimes users are not responsible after the installation as
they did not participate in the project at the initial stage.
Figure 6.4 Shows the six reasons for unsuccessful GIS.
6-5 Required Human Resource for GIS
Required managers and staffs for the operation of GIS are listed below with the terms of
reference.
GIS Project manager
Planning of implementation of GIS applications
Planning of GIS products
Selection of hardware and software
Consultation with users
184
Communication with users
Management of personnel
Budgeting and fund raising
Report to advisory board and top manager
Database Manager
Design of GIS database
Maintenance and updating of database
Plan of data output and map production
Production of GIS database
Quality control of geospatial data
Plan of data acquisition
Digital Map Maker
Compiling of the existing map sources
Map digitization
Data input of attributes
Data acquisition with aerial photogerammetry and remote sensing
Design of Digital maps
Production of digital maps
System Operator
Operation of hardware, software and other peripherals (input/output devices).
Management of materials
Back up of programs and data file
Management of software library with the manuals
Support to user’s request
Prioritization of user’s access
Programmer
Programming for data conversion/reformatting
Programming of applicatrion software
Development of custom command menu
185
Solution to problems with respect of programs and data files
Figure 6.5 Summarizes the required human resources for the operation of GIS to the
institutional scheme.
6-6 Cost Analysis of GIS Project
The major cost required for a GIS project is classified into three categories as follows.
Cost for Hardware and Software
PC based GIS system will range from 10,000 to 30,000 US dollars for a PC, a CD ROM
driver, a digitizer (A3 size), a color ink jet recorder and a public domain software or
discounted software for educational purpose.
UNIX workstation based system will range from 50,000 to 300,000 US dollars for an
UNIX workstation, input/output devices and comercialized GIS software.
Cost for Establishment of Database
Map digitization, scanning, error check, updating and database management are the most
expensive, with a share of about 60~80 percent of the total cost.
Cost of Mainteneance and Daily Operation
Personnel, material, electric power, training and so on are necessary.
The question is how to demonstrate to the decision maker or financial administrator the
possibility of cost saving if GIS is imlemented instead of the conventional analog system.
The justification to promote a GIS project is to emphasize the following three point ;
- Much better decisions can be made with the help of more information provided by GIS,
which will save the unnecessary costs due to mismanagement.
- Higher productivity can be expected because of implementation of a more systematized
and standardized management of
geospatial datra and information.
- More saving of personnel cost will be achieved as the total productivity will raise under
a restructured scheme.
Table 6.1 shows approximate prices of hardware, software and input/output devices.
Figure 6.6 shows the balance between the cost and saving, which implies that the total cost
will be compensated in about ten years.
186