Ganesh College of Engineering, Salem: Ec8093-Digital Image Processing Notes
Ganesh College of Engineering, Salem: Ec8093-Digital Image Processing Notes
Ganesh College of Engineering, Salem: Ec8093-Digital Image Processing Notes
Prepared By
Mr.B.VINOD B.E.,M.E.,(Ph.D)
LECTURE NOTES
PART - A
1. Define Image
Image may be defined as two dimensional light intensity function f(x, y) where x
and y denote spatial co-ordinate and the amplitude or value of f at any point(x, y) is called
intensity or grayscale or brightness of the image at that point.
2. What is Dynamic Range?
The range of values spanned by the gray scale is called dynamic range of an image.
Image will have high contrast, if the dynamic range is high and image will have dull washed
out gray look if the dynamic range is low.
3. Define Brightness
Brightness of an object is the perceived luminance of the surround. Two objects
with different surroundings would have identical luminance but different brightness.
4. What do you meant by Gray level?
Gray level refers to a scalar measure of intensity that ranges from black to grays and
finally to white.
5. Define Digital Image.
An image may be defined as a two dimensional function, f(x,y), where x and y are
spatial co-ordinates and f is the amplitude. When x,y and the amplitude values of f are all
finite, discrete quantities such type of image is called Digital Image.
6. What is Digital Image processing?
The term Digital Image Processing generally refers to processing of a two
dimensional picture by a digital computer. i.e., processing of any two dimension data
7. Define Intensity (or) gray level.
An image may be defined as a two-dimensional function, f(x,y). the amplitude of f
at any pair off co-ordinates (x,y) is called the intensity (or) gray level off the image at that
point.
8. Define Pixel
A digital image is composed of a finite number of elements each of which has a
particular location value. The elements are referred to as pixels, picture elements, pels, and
image elements. Pixel is the term most widely used to denote the elements of a digital
image.
9. What do you meant by Color model?
A Color model is a specification of 3D-coordinates system and a subspace within
that system where each color is represented by a single point.
10. List the hardware oriented color models
1. RGB model,2. CMY model, 3. YIQ model,4. HSI model
PART – B
1. Explain the fundamental steps in digital image processing
1. Image acquisition is the first process in the digital image processing. Note that
acquisition could be as simple as being given an image that is already in digital form.
Generally, the image acquisition stage involves pre-processing, such as scaling.
2. The next step is image enhancement, which is one among the simplest and most
appealing areas of digital image processing. Basically, the idea behind enhancement
techniques is to bring out detail that is obscured, or simply to highlight certain features of
interest in an image. A familiar example of enhancement is when we increase the contrast
of an image because “it looks better.” It is important to keep in mind that enhancement is a
very subjective area of image processing.
3. Image restoration is an area that also deals with improving the appearance of an image.
However, unlike enhancement, which is subjective, image restoration is objective, in the
sense that restoration techniques tend to be based on mathematical or probabilistic models
of image degradation. Enhancement, on the other hand, is based on human subjective
preferences regarding what constitutes a “good” enhancement result.i.e remove noise and
restores the original image
4. Color image processing is an area that has been gaining in importance because of the
significant increase in the use of digital images over the Internet. Color image processing
involves the study of fundamental concepts in color models and basic color processing in a
digital domain. Image color can be used as the basis for extracting features of interest in an
image.
5. Wavelets are the foundation for representing images in various degrees of resolution. In
particular, wavelets can be used for image data compression and for pyramidal
representation, in which images are subdivided successively into smaller regions.
6. Compression, as the name implies, deals with techniques for reducing the storage
required saving an image, or the bandwidth required transmitting it. Although storage
technology has improved significantly over the past decade, the same cannot be said for
transmission capacity. This is true particularly in uses of the Internet, which are
characterized by significant pictorial content. Image compression is familiar (perhaps
inadvertently) to most users of computers in the form of image file extensions, such as the
jpg file extension used in the JPEG (Joint Photographic Experts Group) image compression
standard.
7. Morphological processing deals with tools for extracting image components that are
useful in the representation and description of shape. The morphological image processing
is the beginning of transition from processes that output images to processes that output
image attributes.
8. Segmentation procedures partition an image into its constituent parts or objects. In
general, autonomous segmentation is one of the most difficult tasks in digital image
processing. A rugged segmentation procedure brings the process a long way toward
successful solution of imaging problems that require objects to be identified individually.
On the other hand, weak or erratic segmentation algorithms almost always guarantee
eventual failure. In general, the more accurate the segmentation, the more likely recognition
is to succeed.
9. Representation and description almost always follow the output of a segmentation
stage, which usually is raw pixel data, constituting either the boundary of a region (i.e., the
set of pixels separating one image region from another) or all the points in the region itself.
10. Recognition is the process that assigns a label (e.g., “vehicle”) to an object based on its
descriptors. Recognition topic deals with the methods for recognition of individual objects
in an image.
2. Remote Sensing -
For this application, sensors capture the pictures of the earth’s surface in remote
sensing satellites or multi – spectral scanner which is mounted on an aircraft. These pictures
are processed by transmitting it to the Earth station. Techniques used to interpret the objects
and regions are used in flood control, city planning, resource mobilization, agricultural
production monitoring, etc.
3. Moving object tracking -
This application enables to measure motion parameters and acquire visual record of
the moving object. The different types of approach to track an object are:
· Motion based tracking
· Recognition based tracking
4. Defense surveillance –
Aerial surveillance methods are used to continuously keep an eye on the land and
oceans. This application is also used to locate the types and formation of naval vessels of
the ocean surface. The important duty is to divide the various objects present in the water
body part of the image. The different parameters such as length, breadth, area, perimeter,
compactness are set up to classify each of divided objects. It is important to recognize the
distribution of these objects in different directions that are east, west, north, south, northeast,
northwest, southeast and south west to explain all possible formations of the vessels. We
can interpret the entire oceanic scenario from the spatial distribution of these objects.
5. Biomedical Imaging techniques -
For medical diagnosis, different types of imaging tools such as X- ray, Ultrasound,
computer aided tomography (CT) etc are used. The diagrams of X- ray, MRI, and computer
aided tomography (CT) are given below.
· Digital mammograms – This is used to detect the breast tumour. Mammograms can be
analyzed using Image processing techniques such as segmentation, shape analysis, contrast
enhancement, feature extraction, etc.
6. Automatic Visual Inspection System –
This application improves the quality and productivity of the product in the
industries.
· Automatic inspection of incandescent lamp filaments – This involves examination of the
bulb manufacturing process. Due to no uniformity in the pitch of the wiring in the lamp, the
filament of the bulb gets fused within a short duration. In this application, a binary image
slice of the filament is created from which the silhouette of the filament is fabricated.
Silhouettes are analyzed to recognize the non uniformity in the pitch of the wiring in the
lamp. This system is being used by the General Electric Corporation.
not only used in the clinical setting, it is increasingly playing a role in many research
investigations. Magnetic Resonance Imaging (MRI) utilizes the principle of Nuclear
Magnetic Resonance (NMR) as a foundation to produce highly detailed images of the
human body. The patient is firstly placed within a magnetic field created by a powerful
magnet.
Image File Formats:
Image file formats are standardized means of organizing and storing digital .
Image fil are composed of digital data in one of these formats that can be for use on a
computer display or printer. An image file format may store data in uncompressed,
compressed, or vect formats.
Image file
sizes:
In raster images, Image file size is positively correlated to the number of pixels in
an image and the color depth, or bits per pixel, of the image. Images can be compressed in
various ways, however.
Major graphic file formats
The two main families of graphics Raster and Vector.
Raster formats
1. JPEG/JFIF
JPE (Joint Photographic Experts Group) is a compression method; JPEG-
compressed images are usually stored in the JFIF (JPEG File Interchange Format) file
format. JPEG compression is (in most cases) lossy compression . The JPEG/JFIF filenam
extension is JPG or JPEG.
2. JPEG 2000
JPEG is a compression standard enabling both lossless and lossy storage. The
compression methods used are different from the ones in standard JFIF/JPEG; they improve
quality and compression ratios, but also require more computational power to process.
JPEG 2000 also adds features that are missing in JPEG.
3. TIFF
The TIFF (Tagged Image File Format) format is a flexible format that normally
saves 8 bits or 16 bits per color (red, green, blue) for 24-bit and 48-bit totals, respectively,
usually using either the TIFF or TIF filename extension.TIFFs can be lossy and lossless;
4. GIF
GIF (Graphics
format Interchange Format) is limited to an 8-bit palette, or 256 colors. This
makes the GIF format suitable for storing graphics with relatively few colors such as simple
diagrams, shapes, logos and cartoon style images.
5. BMP
The BMP file (Windows bitmap) handles graphics files within the
Microsoft Windows OS. Typically, BMP files are uncompressed, hence they are large; the
advantage is their simplicity and wide acceptance in Windows programs.
6. PNG
The PNG (Portable Network Graphics) file format was created as the free, open-
source successor to GIF. The PNG file format supports 8 bit paletted images (with optional
transparency for all palette colors) and 24 bit truecolor (16 million colors).
3. Explain the principle and working of Vidicon digital camera with neat diagram
(Image sensing and Acquisition)
A digital camera is used to convert the optical information into a corresponding
electrical signal, the amplitude which varies in accordance with the variations of brightness.
Photoelectric Effects
The two photoelectric effects used for converting variations of light
intensity into electrical variations are (i) photoemission and (ii) photoconductivity.
Photo emission: Certain metals emit electrons when light falls on their surface.
These emitted electrons are called photoelectrons and the emitting surface a photocathode.
Light consists of small bundles of energy called photons. When light is made incident on a
photocathode, the photons give away their energy to the outer valence electrons to allow
them to overcome the potential-energy barrier at the surface. The number of electrons
which can overcome the potential barrier and get emitted depends on the light intensity.
Alkali metals are used as photocathode because they have very low work-function. Cesium-
silver or bismuth-silver-cesium oxides are preferred as photo emissive surfaces because
they are sensitive to incandescent light and have spectral response very close to the human
eye.
Photo conductivity:-The second method of producing an electrical image is by
photoconduction, where the conductivity or resistivity of the photosensitive surface varies
in proportion to the intensity of light focused on it. In general the semiconductor metals
including selenium, tellurium and lead with their oxides have this property known as
photoconductivity. The variations in resistance at each point across the surface of the
material are utilized to develop a varying signal by scanning it uniformly with an electron
beam
Fig-Photo emission
Fig-Photo conductivity
Types of camera tubes: 1) orthicon 2) vidicon 3) Plumbicon.
electrons hitting the screen may cause stray electron field &may grid tangentially which
may cause defocusing and loss resolution. The beam should strike at right angles with
screen.
3. Electron gun multiplier. When the surface of a metal is bombarded by incident
electrons having high velocities secondary emissions takes place. Camera tubes make use of
secondary emissions to amplify small amount of photoelectric current that is later employed
to develop video signal. The electron multiplier is a collection of anode cathode electrodes
called dynodes. The electrons are accelerated by the positive potential of the dynode. The
number of electrons available is multiplied each time the secondary electron strikes the
emitting surface of the next positive dynode. The current amplification is noise free as it
does not have any active or passive components.
Fig-Dgital camera(VIDICON)
4. Explain in Detail the structure of the Human eye and also explain the
image formation in the eye. (OR) Explain the Elements of Visual Perception.
Visual Perception: Means how an image is perceived by a human observer
The Shape of the human eye is nearly a sphere with an Average diameter = 20mm.It
has 3 membranes:
1. Cornea and Sclera – outer cover
2. Choroid
3. Retina -enclose the eye
Cornea: tough, transparent tissue covers the anterior surface of the eye.
Sclera: Opaque membrane, encloses the remainder of the optic globe
Choroid: Lies below the sclera .It contains a lot of blood vessels providing nutrition to the
eye. The Choroid coat is heavily pigmented and hence helps to reduce the amount of
extraneous light entering the eye and the backscatter within the optical globe. At the
anterior part, the choroid is divided into Ciliary body and the Iris diaphragm.
The Iris contracts or expands according to the amount of light that enters into the
eye. The central opening of the iris is called pupil, which varies from 2 mm to 8 mm in
diameter.
Lens is made of concentric layers of fibrous cells and is suspended by fibers that attach to
the Ciliary body. It contains 60-70% of water, 6 % fat and more protein. Lens is colored by
yellow pigment which increases with eye. Excessive clouding of lens can lead to poor or
loss of vision, which is referred as Cataracts.
Retina: Innermost membrane of the eye which lines inside of the wall’s entire posterior
portion. When the eye is properly focused, light from an object outside the eye is imaged on
the retina. It consists of two photo receptors.
Blind spot: the absence of receptors in the retina is called as Blind spot. From the diagram,
1. Receptor density is measured in degrees from the fovea.
2. Cones are most dense in the center of the retina (in the area of the fovea).
3. Rods increase in density from the center out to approx.20° off axis and then
decrease in density out to the extreme periphery of the retina
5. Illustrate the concept of Brightness adaptation.(or)write short notes on 1. Mach
band effect 2. Weber’s Ratio 3. Simultaneous Contrast (May 2007)
Subjective Brightness: is defined as the intensity as preserved by the human visual system
(HVS).
Brightness Adaptation means the HVS can operate only from scotopic to glare limit. It
cannot operate over the range simultaneously.
From the diagram it is seen that the brightness is a logarithmic function of light intensity.Ba
is a brightness adaptation level. Below Ba, the shades of colors cannot be discriminated.
Contrast sensitivity: The ability of the eye to discrimination b/w changes in brightness at
any specific adaptation level is of considerable interest
Weber’s Ratio: The ratio of increment of illumination to background of illumination is
called as Weber ratio.(ie) ∆Ic/I
At low levels of illumination, when the Weber ratio is large, brightness discrimination is
poor .Here vision is carried out by the Rods.
At high levels of illumination, when the Weber ratio is small, brightness discrimination is
improved and better .Here vision is carried out by the Cones.
Mach Band: Mach bands are an optical illusion consisting of an image of two wide bands,
one light and one dark, separated by a narrow strip with a light – to – dark gradient.
Thus mach band effect means the intensity of the stripes is constant. Therefore it
preserves the brightness pattern near the boundaries and these bands are called as mach
band.
Simultaneous Contrast: Two colors side by side interact with one another and change our
perception accordingly. The effect of this interaction is called as Simultaneous Contrast
Here all the inner squares have the same intensity, but they appear progressively
brighter as the background becomes lighter
5. Write short notes on (i) luminance,(ii) hue or tint, and (iii) saturation,iv) Brightness
v) Contrast
Any color has three characteristics to specify its visual information. These are (i) luminance,
(ii) hue or tint, and (iii) saturation. These are defined as follows:
(i) Luminance
This is the amount of light intensity as perceived by the eye regardless of the color.
In black and white pictures, better lighted parts have more luminance than the dark areas.
Different colours also have shades of luminance in the sense that though equally
illuminated appear more or less bright. Thus on a monochrome TV screen, dark red colour
will appear as black, yellow as white and a light blue colour as grey.
(ii) Hue
This is the predominant spectral colour of the received light. Thus the colour of any
object is distinguished by its hue or tint. The green leaves have green hue and red tomatoes
have red hue. Different hues result from different wavelengths of spectral radiation and are
perceived as such by the sets of cones in the retina.
(iii) Saturation
This is the spectral purity of the colour light. Since single hue colours occur rarely
alone, this indicates the amounts of other colours present. Thus saturation may be taken as
an indication of how little the colour is diluted by white. A fully saturated colour has no
white. vivid green is fully saturated and when diluted by white it becomes light green. The
hue and saturation of a colour put together is known as chrominance. Note that it does not
contain the brightness information. Chrominance is also called chroma.
(iv) Brightness:
Brightness is the overall or average intensity of illumination and it determines
background light level in the reproduced picture. The brightness control can be varied to get
average illumination of the scene.
(ii) Contrast:
Contrast is the difference in light intensity between black and white parts of the
picture and above the average brightness level. Too much contrast makes the picture hard
white little contrast gives the impression of a washed out picture.
Magenta
White
(0,1,0)
Black G
Green
(1,0,0)
Red Yellow
R
Figure -RGB color cube.
Mixing all the three primary colors results in white. Color television reception is based
on this three color system with the additive nature of light.
There are several useful color models: RGB, CMY, YUV, YIQ, and HSI.
1. RGB color model
The colors of the RGB model can be described as a triple (R, G, B), so that
R, G, B .The RGB color space can be considered as a three-dimensional unit cube, in which
each axis represents one of the primary colors, see Figure. Colors are points inside the cube
defined by its coordinates. The primary colors thus are red=(1,0,0), green=(0,1,0), and
blue=(0,0,1). The secondary colors of RGB are cyan=(0,1,1), magenta=(1,0,1) and
yellow=(1,1,0).
The nature of the RGB color system is additive in the sense how adding colors
makes the image brighter. Black is at the origin, and white is at the corner where
R=G=B=1. The gray scale extends from black to white along the line joining these two
points. Thus a shade of gray can be described by (x,x,x) starting from black=(0,0,0) to
white=(1,1,1).
The colors are often normalized as given in equation . This normalization guarantees
that r+g+b=1.
r R
R B G
g G
R B G
B
b
R B G
The scale of C, M and Y also equals to unit: C, M, Y .The CMY color system is used in
offset printing in a subtractive way, contrary to the additive nature of RGB. A pixel of color
Cyan, for example, reflects all the RGB other colors but red. A pixel with the color of
magenta, on the other hand, reflects all other RGB colors but green. Now, if we mix cyan
and magenta, we get blue, rather than white like in the additive color system.
In order to generate suitable colour signals it is necessary to know definite ratios in which
red, green and blue combine form new colours. Since R,G and B can be mixed to create any colour
including white, these are called primary colours.
R + G + B = White colour
R – G – B = Black colour
The primary colours R, G and B combine to form new colours
R+G=Y
R + B = Magenta (purplish)
G + B = Cyan (greenish blue)
GRASSMAN’S LAW:
Our eye is not able to distinguish each of the colours that mix to form a new colour but
instead perceives only the resultant colour. Based on sensitivity of human eye to various colours,
reference white for colour TV has been chosen to be a mixture of colour light in the ratio
Similarly yellow can be produced by mixing 30% of red and 59% of green, magenta by
mixing 30% of red and 11% of blue and cyan by mixing 59% of green to 11% of blue. Base on this
it is possible to obtain white light by mixing 89% of yellow with 11% of blue or 70% of cyan with
30% of red.
Thus the Eye perceives new colours based on the algebraic sum of red, green and blue light
fluxes. This forms the basic of colour signal generation and is known as GRASSMAN’S LAW.
3. YUV color model
The basic idea in the YUV color model is to separate the color information apart
from the brightness information. The components of YUV are:
Y 0.3 R 0. 6 G 0.1 B
U B
YV R
Y
Y represents the luminance of the image, while U,V consists of the color
information, i.e. chrominance. The luminance component can be considered as a gray-scale
version of the RGB image, The advantages of YUV compared to RGB are:
1. The brightness information is separated from the color information
2. The correlations between the color components are reduced.
3. Most of the information is collected to the Y component, while the information
content in the U and V is less. where the contrast of the Y component is much greater
than that of the U and V components.
The latter two properties are beneficial in image compression. This is because the
correlations between the different color components are reduced, thus each of the
components can be compressed separately. Besides that, more bits can be allocated to the Y
component than to U and V. The YUV color system is adopted in the JPEG image
GANESH COLLEGE OF ENGINEERING 2
EC8093-
compression standard.
1 1 R G R B
2
H cos 1 , if B G
360 ∘
R G R BG B
2
1 R G R B
1 2
H 1 cos1 , otherwise
360 ∘
R G R BG B
2
3
S=1- min R, G, B
R G B
1
I RG
B
3
Table -Summary of image types.
Image type Typical bpp No. of colors Common file formats
Binary image 1 2 JBIG, PCX, GIF, TIFF
Gray-scale 8 256 JPEG, GIF, PNG, TIFF
6
Color image 24 16.6 10 JPEG, PNG, TIFF
Color palette 8 256 GIF, PNG
image
7. Discuss the role of sampling and quantization in the context of image encoding
applications.
Image Sampling and Quantization:
To generate digital images from sensed data. The output of most sensors is a
continuous voltage waveform whose amplitude and spatial behavior are related to the
physical phenomenon being sensed. To create a digital image, we need to convert the
continuous sensed data into digital form. This involves two processes; sampling and
quantization.
Figure: Generating a digital image. (a) Continuous image. (b) A scan line from A to B
in the continuous image, used to illustrate the concepts of sampling and quantization.
(c) Sampling and quantization. (d) Digital scan line.
When a sensing strip is used for image acquisition, the number of sensors in the
strip establishes the sampling limitations in on image direction. Mechanical motion in the
other direction can be controlled more accurately, but it makes little sense to try to achieve
sampling density in one direction that exceeds the sampling limits established by the
number of sensors in the other. Quantization of the sensor outputs completes the process of
generating a digital image.
Quantization
This involves representing the sampled data by a finite number of levels based on some
criteria such as minimization of the quantizer distortion, which must be meaningful.
Quantizer design includes input (decision) levels and output (reconstruction) levels as well
as number of levels. The decision can be enhanced by psychovisual or psychoacoustic
perception.
Quantizers can be classified as memory less (assumes each sample is quantized
independently or with memory (takes into account previous sample)
Alternative classification of quantisers is based on uniform or non- uniform quantization.
They are defined as follows.
Uniform quantizers Non-uniform quantizers
They are completely defined by (1) the The Step sizes are not constant. Hence non-
number of levels it has (2) its step size and uniform quantization is specified by input
whether it is midriser or midtreader. We and output levels in the 1st and 3rd
will consider only symmetric quantizers quandrants.
i.e. the input and output levels in the 3 rd
quadrant are negative of those in 1st
quandrant.
f ( x , y ) f ( x , y ), x m x , y n y,
s
0 , otherwise
m,n Z.
Sampling conditions for no information loss – derived by examining the spectrum of the
image by performing the Fourier analysis:
Saveetha Engineering 29
EC8093- Dr.M.Moorthi,Professor-
Saveetha Engineering 30
EC8093- Dr.M.Moorthi,Professor-
Since the sinc function has infinite extent => it is impossible to implement in practice the
ideal LPF it is impossible to reconstruct in practice an image from its samples without error
if we sample it at the Nyquist rates
9. Explain Uniform Quantization and Non-Uniform Quantization
1. Non-Uniform Quantization:
Max Lloyd Quantizer
Given the range of input u as from aL to a u , and the number of output levels as L,
we design the quantizer such that MSQE is minimum.
Saveetha Engineering 31
EC8093- Dr.M.Moorthi,Professor-
MSQE E
u u' (u u')
tk 1u
pudu 1
2
2
tk
u r pudu 2
k 1 tk
k
u r pudu tk r
k 1 tk 1
k k 1
ptk dtk
tk r k
ptk dtk 3
tk tk 1 tk
2
2
tk 1
Saveetha Engineering 32
EC8093- Dr.M.Moorthi,Professor-
tk rk 1 p (tk ) dtk 4
t k1
tk
dtk tk tk rk p tk
k
2 t k rk2 pt r pt 0
1 t k k k
k
t k rk 1 t k rk
Since t k rk 1 0 and t k rk 0 solution that is valid is
t k rk 1 t k rk or t k rk rk 1 / 2 5
i.e. input level is average of two adjacent output levels.
Saveetha Engineering 33
EC8093- Dr.M.Moorthi,Professor-
Find rk
tk 1
d u r pudu 0 6
rk r
tk
k 2
k
tk 1
2 u rk p u du 0 7
rk tk
Hence
k 1
tk 1
rk u p u du p u du 8
tk tk
rk tk 1
tk 9
2
Output level is the centroid of adjacent input levels. This solution is not closed form. To
find input level t k , one has to find rk and vice versa. However by iterative techniques both
rk and tk can be found using Newton’s method.
i.e. each reconstruction level lies mid-way between two adjacent decision levels.
2.Uniform Quantizer:
When 1 1
pu , p(e) 1/ q , as shown
au aL A
Uniform quantizer transfer function
r4=224
r3=160
Reconstruction
r2=96
r1=32
Since quantization error is distributed uniformly over each step size (q)with zero mean i.e.
over q / 2, q / 2, the variance is,
e q/2
2
e 2 p(e)de 1
q / 2
2 q/2
1 e 2 de q 2 /12 2
e q q /2
Saveetha Engineering 34
EC8093- Dr.M.Moorthi,Professor-
Let range of u be A and 2 the variance;
u
A/2
u
2
u
2
p(u)du 3
A/2
1 A/ 2
A u 2du A2 /12 4
2u
Then A/2
Saveetha Engineering 35
EC8093- Dr.M.Moorthi,Professor-
A
q .
2b
1
Hence 2 A2 / 22b ; 5
e 12
SNR for uniform quantizer in dB =10 log(2 /2 ) 22b 6bdB-----6
u e
4. It is sufficient to design mean square quantizers for zero mean and unity variance
distributions.
10. Explain in detail the ways to represent the digital image.
Assume that an image (x,y) is sampled so that the resulting digital image has M
rows and N columns. The values of the coordinates (x,y) now become discrete quantities.
For notational clarity and convenience, we shall use integer values for these discrete
coordinates. Thus, the values of the coordinates at the origin are (x,y) = (0,0). The next
coordinate values along the first row of the image are represented as (x,y) = (0,1 Figure
shows the coordinate convention used.
Saveetha Engineering 37
EC8093- Dr.M.Moorthi,Professor-
and pixel will be used throughout the rest of our discussions to denote a digital image and
its elements.
In some discussions, it is advantageous to use a more traditional matrix notation to denote a
digital image and its elements:
a ao,1 a0,N1
0,0
a a a
A 1,0 1,1 1,N1
⁝ ⁝ ⁝
a a a
M1,0 M1,1 M1,N1
11. Write short notes on: (a) Neighbors of Pixels (b) Distance Measures c.
Connectivity d. Adjacency
1. Neighbors of a pixels:
A pixel p at coordinates (x, y) has four horizontal and vertical neighbors whose coordinates
are given by (x + 1, y), (x -1, y), (x, y +1), (x, y -1)
This set of pixels, called the 4-neighbors of p, is denoted by N4(p). Each pixel is a
unit distance from (x, y), and some of the neighbors of p lie outside the digital image if (x, y)
is on the border of the image.
2. Distance Measures;
For pixels p, q, and z, with coordinates (x, y), (s, t), and (, w), respectively, D is a distance
function or metric if
(a) D(p, q) 0 (D (p, q) = 0 iff p = q),
(b) D(p, q) = D (p, q), and
(c) D(p, z) D (p, q) +D( q, z).,
For this distance measure the pixels having a distance less than or equal to some
value r from (x, y) are the points contained in a disk of radius r centered at (x, y).
D4(p,q)
xs yt …………2
.
In this case, the pixels having a D4 distance from (x, y) less than or equal to some
value r form a diamond centered at (x, y). For example, the pixels with D4 distance 2
from (x, y) (the center point) form the following contours of constant distance:
2
2 1 2
2 1 0 1 2
2 1 2
2
Saveetha Engineering 40
EC8093- Dr.M.Moorthi,Professor-
2 2 2 2 2
2 1 1 1 2
2 1 0 1 2
2 1 1 1 2
2 2 2 2 2
3.Connectivity
0 1 0 0 1 0 0 1 0
0 0 1 0 0 1 0 0 1
V{1}: The set V consists of pixels with value 1.
Two image subsets S1, S2 are adjacent if at least one pixel from S1 and one pixel from S2 are
adjacent.
12. Write short notes on two dimensional Orthogonal and Unitary transforms
Two dimensional unitary transforms play an important role in image processing. The
term image transform refers to a class of unitary matrices used for representation of
images.
In analogy, with I-D signals that can be represented by an orthogonal series of basis
functions, we can similarly represent an image in terms of a discrete set of basis arrays
called “basis images”. These are generated by unitary matrices.
Alternatively an ( N N ) image can be represented as ( N 2 1) vector. An image
transform provides a set of coordinates or basis vectors for the vector space.
I-D-Transforms:
For a one dimensional sequence u( n ), n 0,1.....N 1 representing a
→ → →
vector u of size N , a unitary transform is : v = A u
Saveetha Engineering 41
EC8093- Dr.M.Moorthi,Professor-
N 1
v(k) = a( k ,n )u( n ) , for 0 K N 1
(1)
n
0
where T *T
A1 = A* (unitary)
→ *T →
This implies , u=A v
N
or, u(n) = 1 v( k )a* ( k,n) , for 0 n N (2) Equation
k 0 1
(2) can be viewed as a series representation of sequence u(n) . The columns of
*T →* *
A i.e the vectors a a ( k ,n ) 0 n N 1 are called the “basis vectors” of
k
, T
A.
The series coefficients v(k) give a representation of original sequence u(n) and are useful in
compression , filtering , feature extraction and other analysis.
Two dimensional Orthogonal and Unitary transforms:
As applied to image processing, a general orthogonal series expansion for an
N N image is a pair of transformations of the form :
N 1
v(k,l) = u( m,n )ak ,l ( m,n ) , 0 k ,l N (3)
m,n 1
0
* N 1 v( k ,l
u(m,n) =
)a ( m,n ) , 0 m,n N 1 (4)
k ,l
k ,l 0
where ak,l ( m,n ) is called an ” image transform.”
It is a set of complete orthogonal discrete basis functions satisfying the properties:-
1) Orthonormality: N 1
ak ,l ( m,n )a*k/ ( m,n )
m,n /
0
= δ( Nk 1k ,l l )
k
2) Completeness : a ( m,n )a ( m,n )
k ,l k ,l
k ,l 0
= δ( m m,n n )
The elements v ( k ,l ) are transform coefficients and V v( k,l is the transformed
)
image.
Q 1 property assures that any truncated series expansion of the form
The orthonomality
U ( m,n ) P 1 * v( k ,l ( m,n ) , for P N ,
)a
P,Q
Saveetha Engineering 42
EC8093- Dr.M.Moorthi,Professor-
QN k ,l
k 0 l
0
will minimize the sum of squares
2
error
σ 2
N u( m,n ) U ( m,n )
1
e P,Q
m,n 0
where coefficients v( k ,l ) are given by (3).
The completeness property assures that this error will be zero for P Q N
Separable Unitary Transforms:
Saveetha Engineering 43
EC8093- Dr.M.Moorthi,Professor-
i.e *
T
* T * T
AA = I = A A and BB = I = BT B*
Often one chooses B same as A
N 1
v( k ,l ) a( k ,m )u( m,n )a( l ,n )
= m,n
0
T
V = AU A (5)
And u( m,n ) N 1
* *
= a ( k ,m )v( k ,l )a ( l ,n )
k ,l 0
U =
* *
T
(6)
A VA
T T
Eqn (5) can be written as V = A( AU )
Eqn (5) can be performed by first transforming each column of U and then
transforming each row of the result to obtain rows of V .
T
Basis Images : Let denote kth column of A* .
→* k
a
and v( k ,l ) ,u *k ,
= l
Saveetha Engineering 45
EC8093- Dr.M.Moorthi,Professor-
Example: Let A = 1 1 1 ; U = 1 2
2 1 1
3 4
T
Transformed image V = A U A 5 1
=
2 0
T
*
And Basis images are found as outer product of columns of A i.e
0,0 = 1 1
* (1 1)
A
2
=
1 1
0,1
(1 1)
*
A
= 1 21 1 =
2 1 1 *T
A1,0
*
A = 1 1 1 1 1)
(1
1,1
2 1 1
The inverse transformation
T
* 1 1 1 5 1 1 1
*
A V A = 2 1 1 2 0 1 1
1 2
= =U
3 4
Saveetha Engineering 46
1 N 1 N 1EC8093- Dr.M.Moorthi,Professor-
F(u,
v)e j 2(uxvy )/ N
N u0 v0
It is straightforward to prove that the two dimensional Discrete Fourier Transform is
separable, symmetric and unitary.
Saveetha Engineering 47
EC8093- Dr.M.Moorthi,Professor-
14. Explain in detail the discrete Cosine Transform (DCT) and also explain its
properties.
The N X N cosine transform matrix C = {c (K, n)}, also called the discrete cosine transform
(DCT), is defined as
1
, k 0, 0 n N 1
N
C (k,n) =
(2n 1)k
2 cos , 1 k N 1,0 n N 1
N 2N
Where
(0) 1, 2
(k) N for 1 k N 1
N
The inverseNtransformation
1 is
given
(2n by
1)k
U (n) = (k)(k) cos , 0nN1
2N
k0
The two-dimensional cosine transform pair is obtained by substituting A = A* = C
in (5.11) and (5.12). The basis images of the 8 x 8 two-dimensional cosine transform are
shown in figures shows examples of the cosine transform of different images.
Properties of the Cosine Transform
1. The cosine transform is real and orthogonal, that is, C = C* C-1 = CT
2. The cosine transform is not the real part of the unitary DFT. This can be seen by
inspection of C and the DFT matrix F. (Also see Problem) However, the cosine transform
of a sequence is related to the DFT of its symmetric extension (see Problem)
3. The cosine transform is a fast transform. The cosine transform of a vector of N elements
can be calculated in O (N log 2N) operations via an N-point FFT [19]. To show this we
define a new sequence u (n) by reordering the even and odd elements of u (n) as
-
u (n) u (2n)
, 0 N n
2
u (N n 1) u (2n 1)
Now, we split the summation term in into even and odd terms and use to obtain.
u(2n 1) cos
Saveetha Engineering 48
EC8093- Dr.M.Moorthi,Professor-
n0 2N
Saveetha Engineering 49
EC8093- Dr.M.Moorthi,Professor-
Saveetha Engineering 50
EC8093-
Dr.M.Moorthi,Professor-
0 1
This gives the approximation 2 R-1 QC for 1
Saveetha Engineering 51
EC8093- Dr.M.Moorthi,Professor-
Hence the eigenvectors of R and the eigenvectors of Q c, that is, the cosine transform,
will be quite close. These aspects are considered in greater depth in Section on sinusoidal
transforms.
This property of the cosine transform together with the fact that it is a fast transform
has made it a useful substitute for the KL transform of highly correlated first order Markov
sequences.
Saveetha Engineering 52
EC8093- Dr.M.Moorthi,Professor-
PART A
H (u,v) 1 2n
D(u,v)
1
D o
The transfer function for a Butterworth High pass filter is,
H (u 1
, v)
D o 2
1
D (u , v )
Where , n = filter order, Do = cutoff frequency and D(u,v) =
(u2+v2)1/2 = distance from point (u,v) to the origin
4. Specify the objective of image enhancement technique. (Or)Why the
image enhancement is needed in image processing technique. (May / June
2009)
1. The objective of enhancement technique is to process an image so that the
result is more suitable than the original image for a particular application.
2. Image Enhancement process consists of a collection of technique that seek to
improve the visual appearance of an image. Thus the basic aim is to make the image look
better.
3. Image enhancement refers to accentuation or sharpening of image features such
as edges, boundaries or contrast to make a graphic display more useful for display and
analysis.
4. Image enhancement includes gray level and contrast manipulation, noise
Saveetha Engineering 53
EC8093- Dr.M.Moorthi,Professor-
reduction, edge crispening & sharpening, filtering, interpolation & magnification, pseudo
coloring, so on…
Saveetha Engineering 54
EC8093- Dr.M.Moorthi,Professor-
Saveetha Engineering 56
EC8093- Dr.M.Moorthi,Professor-
2. Median filters are useful for removing isolated lines or points (pixels) while
preserving spatial resolutions. They perform very well on images containing binary (salt
and pepper) noise but perform poorly when the noise is Gaussian.
9. What do you mean by Point processing?
When enhancement is done with respect to a single pixel, the process is called point
processing.ie., enhancement at any point in an image depends only on the gray level at that
point.
10. What is Image Negatives?
A negative image can be obtained by reversing the intensity levels of an image, according
to the transformation,
s = L – 1 -r
, when the gray level range is [0,L-1]
1. Digital negatives are useful in the display of medical images and to produce
negative prints of image.
2. It is also used for enhancing white or gray details embedded in dark region of
an image.
11. Write the steps involved in frequency domain filtering.
1. Multiply the input image by (-1)x+y to center the transform.
2. Compute F(u,v), the DFT of the image from (1).
3. Multiply F(u,v) by a filter function H(u,v).
4. Compute the inverse DFT of the result in (3).
Saveetha Engineering 57
EC8093- Dr.M.Moorthi,Professor-
f 2
f ( x 1, y) f ( x 1, y) f ( x, y 1) f ( x, y 1) 4 f ( x, y)
The above digital laplacian equation is implemented using filter mask,
0 1 0
1 -4 1
0 1 0
S=T(r)
23. Define contrast stretching.
The contrast stretching is a process of to increase the dynamic range of the gray
levels in the image being processed.
24. Define Thresholding.
Thresholding is a special case of clipping where a=b and output becomes binary.
For examples, a seemingly binary image, such as printed page, does not give binary output
when scanned because of sensor noise and background illumination variations.
Thresholding is used to make such an image binary.
25. What are the types of functions of gray level Transformations for Image
Enhancement?
Three basic types of functions used frequently for image enhancement are
1. Linear (negative and identity transformations
2. Logarithmic Transformations
3. Power law Transformations.
26. Give the expression of Log Transformations?
The general form of the log Transformation is given
S= C log (1+r).
Where C is a construct and it is assumed that r 0. this transformation enhances the small
magnitude pixels compared to those pixels with large magnitudes.
27. Write the expression for power-Law Transformations.
Saveetha Engineering 59
EC8093- Dr.M.Moorthi,Professor-
Power-law transformations have the basic form
Saveetha Engineering 60
EC8093- Dr.M.Moorthi,Professor-
S=Cr
Where c and r are positive constants.
28. What is meant by gamma and gamma correction?
Power law transformation is given by
S=Cr
The exponent in the power-law equation is referred to as gamma. The process used to
correct the power law response phenomena is called gamma corrections.
29. What is the advantage & disadvantages of piece-wise Linear Transformations?
The principal advantage of piecewise Linear functions over the other types of
functions is that piece wise functions can be arbitrarily complex.
The main disadvantage of piece wise functions is that their specification requires
more user input.
30. What is meant by gray-level slicing?
Gray-level slicing is used in areas where highlighting a specific range of gray level
in a image. There are 2 basic approaches gray level slicing.
1. Highlights range {A,B] of gray levels and reduces all other to a constant level.
2. Second approach is that highlights range [A,B] but preserves all other levels.
31. What is meant by Bit-plane slicing?
Instead of high lighting grey-level ranges, highlighting the contributes made to total
image appearance by specific bits is called Bit-plane slicing.
use of bit-plane slicing: This transformation is useful in determining the visually
significant bits.
32. What is histogram and write its significance?
The histogram of an image represents the relative frequency of occurrence of the
various gray levels in the image. They are the basis for numerous spatial domain processing
techniques. Histogram manipulation can be used effectively for image enhancement.
33. Define histogram Equalization.
Histogram Equalization refers to the transformation (or) mapping for which the
processed (output) image is obtained by mapping each pixel into the input image with the
corresponding pixel in the output image.
34. Define Histogram matching (or) histogram specification.
The method used to generate a processed image that has a specified histogram is
called histogram specification (or) matching.
35. What is meant by Image subtraction?
The difference between two images f(x,y) and h(x,y) expressed as
Y(x,y)=f(x,y)-h(x,y)
is obtained by computing the difference all pairs off corresponding pixels from f and h.
the key usefulness of subtraction is the enhancement of differences between images.
36. Define mask & what is meant by spatial mask?
Same neighborhood operations work with the values of eh image pixels in the
neighborhood and the corresponding values of a sub image that has the same dimension as
Saveetha Engineering 61
EC8093- Dr.M.Moorthi,Professor-
the neighborhood. The sub image is called filter (or) mask (or) kernel (or) template (or)
window.
When the image is convolved with a finite impulses response filter it is called as spatial
mask.
Compute the average value of the corrupted image g(x,y) in the area defined
The value of the restored image at any point (x,y)
1
fˆ(x, y) g(s, t)
mn ( s,t )S x , y
Note: Using a convolution mask in which all coefficients have value 1/mn. Noise is
reduced as a result of blurring.
43. What is meant by geometric mean filer?
Geometric mean filter is given by the expression
1
fˆ ( x , y ) g ( s , t )
mn
( s , t ) S xy
44. What is meant by harmonic mean filer?
The harmonic mean filter operation is given by the expression
( s,t )S
fˆ(x, y)
Saveetha Engineering 62
EC8093- Dr.M.Moorthi,Professor-
mn
1
g(s, t)
x
Saveetha Engineering 63
EC8093- Dr.M.Moorthi,Professor-
g(s, t) Q1
fˆ(x, y)
( s,t )Sxy
g(s, t)
( s,t )Sxy
Q
Where Q is called the order of the filter. This filter is well suited for reducing or virtually
eliminating the effects of salt-and-pepper noise.
PART – B
Saveetha Engineering 64
EC8093- Dr.M.Moorthi,Professor-
S=T(f(x,y)
The center of the sub image is moved from pixel to pixel starting, say, at the top left
Saveetha Engineering 65
EC8093- Dr.M.Moorthi,Professor-
corner. The operator T is applied at each location (x, y) to yield the output, g, at that
location. The process utilizes only the pixels in the area of the image spanned by the
neighborhood.
Where f(x, y) is the input image, g(x, y) is the processed image, and T is an operator
on f, defined over some neighborhood of (x, y). In addition, T can operate on a set of
input images, such as performing the pixel-by-pixel sum of K images for noise
reduction
Since enhancement is done with respect to a single pixel, the process is called point
processing.ie., enhancement at any point in an image depends only on the gray level at that
point.
Applications of point processing:
To change the contrast or dynamic range of an image
To correct non linear response of a sensor used for image enhancement
1. Basic Gray Level Transformations
The three basic types of functions used frequently for image enhancement
( a ) Image Negatives: or Digital Negatives:
A negative image can be obtained by reversing the intensity levels of an image,
according to the transformation,
, when the gray level range is [0,L-1]
s = L – 1 -r
Digital negatives are useful in the display of medical images and to produce
negative prints of image.It is also used for enhancing white or gray details embedded in
dark region of an image.
( b ) Log Transformations:
The dynamic range of gray levels in an image can be very large and it can be
compressed. Hence, log transformations are used to compress the dynamic range of the
image data, using the equation,
s = c log10 (1+│r│)
Saveetha Engineering 66
EC8093- Dr.M.Moorthi,Professor-
From the above diagram, it can be seen that this transformation maps narrow range of
low gray-level values in the input image into a wider range of output levels. The
opposite is true which is used to expand the values of dark pixels in an image while
compressing the higher-level values. This done by the inverse log transformation
( c ) Power Law Transformations:
The power law transformation equation is given as,
Where c and γ are the constants
s = c rγ
Saveetha Engineering 67
EC8093- Dr.M.Moorthi,Professor-
This transformation maps a narrow range of dark input values into a wider range of
output values, for fractional values of γ.
For c = γ = 1, identity transformation is obtained.
For γ >1, the transformation is opposite for that obtained for γ < 1
The process which is used to convert this power law response phenomenon is called
gamma correction.
> 0 Compresses dark values and Expands bright values
< 0 Expands dark values and Compresses bright values
2.Piecewise-linear transformations:
( a ) Contrast Stretching:
Contrast stretching is used to increase the dynamic range of the gray levels in the
image being processed.
Low contrast images occur often due to poor / non-uniform lighting conditioning or due
to non-linearity or small dynamic range of the image sensor.
Saveetha Engineering 68
EC8093- Dr.M.Moorthi,Professor-
Produce an image of higher contrast than the original by darkening the levels
below m and brightening the levels above m in the original image. In this technique,
known as contrast stretching.
( b ) Clipping:
A special case of Contrast stretching, where α = γ = 0 is called clipping.This
transformation is used for noise reduction, when the input signal is said to lie in the range
[a,b].
( c ) Thresholding:
Saveetha Engineering 69
EC8093- Dr.M.Moorthi,Professor-
s =L , A ≤ r ≤ B 0, else
b) To brighten the desired range of gray levels while preserving the background of
the image.
i.e., with background,
s =L , A ≤ r ≤ B r, else
In terms of 8 bit bytes, plane0 contains all the lowest order bits in the bytes
comprising the pixels in the image and plane7 contains all the high order bits.
Thus, the high-order bits contain the majority of the visually significant data while
other bits do not convey an information about the image structure
Bit Extraction:
If each pixel of an image is uniformly quantized to B bits, then the nth MSB can be
extracted and displayed as follows,
Saveetha Engineering 70
EC8093- Dr.M.Moorthi,Professor-
The histogram of a digital image with gray levels in the image [0,L-1] is a discrete function
given as,
h(rk ) nk , where rk is the kth grey level, k 0,1,…, L 1
nk is the number of pixels in the image with grey level rk
Normalized histogram is given as,
p (rk ) n k
n
Where , n is the total number of pixels in the image and n ≥ nk
The Normalized histogram is the histogram divided by the total number of pixels in the
source image.
1. The sum of all values in the normalized histogram is 1.
2. The value given by the normalized histogram for a certain gray value can be read as the
probability of randomly picking a pixel having that gray value
3. The horizontal axis of each histogram plot corresponds to grey level values rk in the
nk
image and the vertical axis corresponds to h(rk ) nk or p(rk ) , if the values are
n
normalized.
Histogram of four basic gray level characteristics:
Dark image: (under exposed image): Here components of the histogram are
concentrated on the low or dark side of the gray scale.
Bright image: (over exposed image): Here components of the histogram are
Saveetha Engineering 71
EC8093- Dr.M.Moorthi,Professor-
Low-contrast image: Here components of the histogram at the middle of the gray
scale and are narrow.
High-contrast image: Here components of the histogram cover a broad range of the
gray scale and they are almost uniformly distributed.
S = T(r), 0 ≤ r ≤ 1
This transformation produces a level s for every pixel value r in the original image.
The transformation function T(r) satisfies the following conditions,
(a) T(r) is single –valued and monotonically increases in the interval 0 ≤ r ≤ 1 (ie.,
preserves the order from black to white in the output image) and
(b) 0 ≤ T(r)) ≤ 1 for 0 ≤ r ≤ 1 (ie., guarantees that the output gray levels will be in
the same range as the input levels)
The following shows the transformation function satisfying the above two conditions.
Saveetha Engineering 72
EC8093- Dr.M.Moorthi,Professor-
If the gray levels in an image can be viewed as random variables in the interval [0,1], then
Pr(r) and Ps(s) denote the pdfs of the random variables r and s.
If Pr(r) and T(r) are known and T-1(s) satisfies the condition (a) , then
Ps(s) = Pr(r) dr
------- (1)
ds
r = T-1(s)
Thus , the pdf of the transformed variables is determined by the gray level pdf of the input
image and by the chosen transformation function.
For Continuous gray level values:
Consider the transformation,
r
s T (r) p (w)dw (2)
0 r
where w is a dummy variable of integration and
r
p r (w)dw cummulative distribution function [CDF] of random variable ' r' con
0
dition (a) and (b) are satisfied by this transformation as CDF increases
monotonically from 0 to 1 as a function of r.
When T(r) is given, we can find Ps(s) using equation (1),
We know that,
s T (r)
ds dT ( r ) d r
p (w)dw p (r)
dr dr dr 0 r r
Saveetha Engineering 73
EC8093- Dr.M.Moorthi,Professor-
dr 1
ds Pr(r) (3)
Thus the processed output image is obtained by mapping each pixel with level rk in
the input image into a corresponding pixel with level sk in the output image.
The inverse transformation is given as ,
r T 1 (s ), k 0,1,2,...L 1
k k
Saveetha Engineering 74
EC8093- Dr.M.Moorthi,Professor-
Let r be an input image, Z be an output image
pr (r) is the original probability density function
Saveetha Engineering 75
EC8093- Dr.M.Moorthi,Professor-
s G(z) p z (z)dz
0
Where G is the transformation of z = s .
From equation (1) and (2),
G(z) = s = T(r)
The inverse z G1(s) G1[T(r)]
process
Therefore, the process of histogram specification can be summarised in the following steps.
i) Obtain the transformation function T(r) using equation (1)
ii) Obtain the transformation function G(z) using equation (2)
iii) Obtain the Inverse transformation function G-1
iv) Obtain the output image by applying using equation (3) to all pixels in the input image
This procedure yields an image whose gray levels z have the specified pz ( z)
pdf
For Discrete gray level values:
The probability of occurrence of gray level rk in an image is given as,
n
p (r ) k , k 0,1,.....L 1
r k
n
Where, n = total no.of pixels in the image
nk = No.of pixels that have gray level rk
L = total no.of possible gray levels in the image
The discrete form of the transformation function is ,
k k n
j
sk T (rk ) pr (rj ) , 0 k 0,1,…, L 1
n 1,
r k
j 0 j 0
sk G(z k ) p z (z i ),
i 0
From (1) and (2), the inverse transformation is given as ,
G(z k ) sk T (rk )
z G 1 (s ) G 1[T (r )]
k k k
5. Write short notes on the various Noise models (Noise distribution) in detail. (Nov /
Dec 2005)
The principal noise sources in the digital image arise during image acquisition or
digitization or transmission. The performance of imaging sensors is affected by a variety of
factors such as environmental conditions during image acquisition and by the quality of the
Saveetha Engineering 76
EC8093- Dr.M.Moorthi,Professor-
sensing elements.
Assuming the noise is independent of spatial coordinates and it is uncorrelated with
respect to the image itself.
The various Noise probability density functions are as follows,
Saveetha Engineering 77
EC8093- Dr.M.Moorthi,Professor-
1. Gaussian Noise: It occurs due to electronic circuit noise, sensor noise due to poor
illumination and/or high temperature
The PDF of Gaussian random variable, z, is given by
p( z) 1 / 22
e ( z z )
2
2
p(z) 1 / 22
e( z z )
2
2
/b for z a
p(z)
0 for z a
Saveetha Engineering 79
EC8093- Dr.M.Moorthi,Professor-
ae
az
for z 0
p( z)
0 for z a
The mean and variance of this density ar e given by
z 1 / a
2 1 / a2
5. Uniform Noise
It is a Least descriptive and basis for numerous random number generators
The PDF of uniform noise is given by
1
for a z b
p(z) b a
0 otherwise
The mean and variance of this density are given by
z (a b) / 2
Saveetha Engineering 80
EC8093- Dr.M.Moorthi,Professor-
2 (b a)2 / 12
Saveetha Engineering 81
EC8093- Dr.M.Moorthi,Professor-
If b > a, gray level b will appear as a light dot and level a will appear like a dark dot.
If either Pa or Pb is zero, then the impulse noise is called Unipolar
When a = minimum gary level value = 0, denotes negative impulse appearing black (pepper)
points in an image and when b = maximum gray level value = 255, denotes positive
impulse appearing white (salt) noise, hence it is also called as salt and pepper noise or
shot and spike noise
7. Periodic Noise
Periodic noise in an image arises typically from electrical or electromechanical
interference during image acquisition.It is a type of spatially dependent noise
Periodic noise can be reduced significantly via frequency domain filtering
Saveetha Engineering 82
EC8093- Dr.M.Moorthi,Professor-
Where,
x 0,1,2,....., M 1
y 0,1,2,......., N 1
Saveetha Engineering 83
EC8093- Dr.M.Moorthi,Professor-
From the general expression, it is seen that the linear spatial filtering is just
“convolving a mask with an image”. Also the filter mask is called as Convolution mask
or Convolution kernel.
The following shows the 3 x 3 spatial filter mask,
R w1 z1 w2 ....... w9 z9
z2
9
wi zi
In general
i1
R w1z1 w2 z2 . . wmnzmn
mn
w z
i1
i i
Where, w’s are mask coefficients, z’s are values of the image gray levels corresponding to
the mask coefficients and mn is the total number of coefficients in the mask.
Non-Linear Spatial filter:
This filter is based directly on the values of the pixels in the neighborhood and they
don’t use the coefficients.These are used to reduce noise.
Saveetha Engineering 84
EC8093- Dr.M.Moorthi,Professor-
1.Neighborhood Averaging:
1. The response of a smoothing linear spatial filter is simply the average of the pixels in the
neighborhood of the filter mask,i.e., each pixel is an image is replaced by the average of the
gray levels in the neighborhood defined by the filter mask. Hence this filter is called
Neighborhood Averaging or low pass filtering.
2. Thus, spatial averaging is used for noise smoothening, low pass filtering and sub
sampling of images.
3. This process reduces sharp transition in gray level
4. Thus the averaging filters are used to reduce irrelevant detail in an
image. Consider a 3 x 3 smoothing or averaging filter
1 1 1
1
1 1 1
9
1 1 1
The response will be the sum of gray levels of nine pixels which could cause R
to be out of the valid gray level range.
R w1z1 w 2 z 2 ...... w 9 z 9
Thus the solution is to scale the sum by dividing R by 9.
R w1 z1 w2 z2 ....... w9 z9
9
wi z i
i1
9
1 9 i
z i , when
1
Saveetha Engineering 85
EC8093- Dr.M.Moorthi,Professor-
.. w i 1
Saveetha Engineering 86
EC8093- Dr.M.Moorthi,Professor-
This is nothing but the average of the gray levels if the pixels in the 3 x 3 neighborhood
defined by the pixels.
Other spatial low pass filters of various sizes are ,
1 1 1 1 1
1 1 1 1 1 1 1
1
1 1 1 1 1 1 1
2 1 1 1 1 1 4
1 1 1 1 1
Thus a m x n mask will have a normalizing constant equal to 1/mn. A spatial average filter
in which all the coefficients are equal is called as a Box Filter.
2. Weighted Average Filter: For filtering a M x N image with a weighted
averaging filter of size m x m is given by,
a b
w(s,t) f (x s, y t)
g(x, y) satb
a b
w(s,t)
sat
b
Example:
Here the pixel at the centre of the mask is multiplied by a higher value than any
other, thus giving this pixel more importance.
The diagonals are weighed less than the immediate neighbors of the center pixel.
Saveetha Engineering 87
EC8093- Dr.M.Moorthi,Professor-
This reduces the blurring.
Saveetha Engineering 88
EC8093- Dr.M.Moorthi,Professor-
3. Uniform filtering
The most popular masks for low pass filtering are masks with all their coefficients
positive and equal to each other as for example the mask shown below. Moreover, they sum
up to 1 in order to maintain the mean of the image.
1 1 1
1 1
9 1 1 1
4. Gaussian filtering
The two dimensional Gaussian mask has values that attempts to approximate the
continuous function
2 2
x y
1 2
G(x, y) e
2
2
. The following shows a suitable integer-valued convolution kernel that approximates a
Gaussian with a of 1.0.
1 4 7 4 1
4 16 26 16 4
1
273 7 26 41 26 7
4 16 26 16 4
1 4 7 4 1
Saveetha Engineering 89
EC8093- Dr.M.Moorthi,Professor-
fˆ(x, y) mediang(s, t)
( s,t )Sxy
Saveetha Engineering 90
EC8093- Dr.M.Moorthi,Professor-
V(m,n) = median{y(m-k,n-l),(k,l)εw}
Isolated
point
0 0 0 0 0 0
Median filtering
0 1 0 0 0 0
0 0 0 0 0 0
6. Directional smoothing
To protect the edges from blurring while smoothing, a directional averaging filter
can be useful. Spatial averages g (x, y :) are calculated in several selected directions (for
example could be horizontal, vertical, main diagonals)
g ( x , y : ) 1
f (x k,y l)
N (k ,l)W
and a direction
is found such that f ( x, y) g( x, y : ) is minimum. (Note that
W is the neighbourhood along the direction and N is the number of pixels within this
neighbourhood).Then by replacing g( x, y :) with g( x, y : ) we get the desired result.
Saveetha Engineering 91
EC8093- Dr.M.Moorthi,Professor-
0 0 0 0 0
0 0 0 0 0
W0 0 0 0 0 0 l
0 0 0 0 0
0 0 0 0 0
K
7. Max Filter & Min Filter:
The 100th percentile filter is called Maximum filter, which is used to find the
brightest points in an image.
R max{zk ,k 1,2,3,....,9}, for..3x3matrix
The 0th percentile filter is called Minimum filter, which is used to find the
darkest points in an image.
R min{zk ,k 1,2,3,....,9} for.....x3matrix
(i.e)Using the 100th percentile results in the so-called max filter, given by
This filter is useful for finding the brightest points in an image. Since pepper noise has very
low values, it is reduced by this filter as a result of the max selection processing the sub
image area Sxy.
The 0th percentile filter is min filter:
This filter is useful for finding the darkest points in an image. Also, it reduces salt noise as
a result of the min operation
8. Mean Filters
This is the simply methods to reduce noise in spatial domain.
1. Arithmetic mean filter
2.Geometric mean filter
3.Harmonic mean filter
4.Contraharmonic mean filter
Let Sxy represent the set of coordinates in a rectangular sub image window of size mxn,
centered at point (x,y).
1. Arithmetic mean filter
Compute the average value of the corrupted image g(x,y) in the area defined by
Sx,y.The value of the restored image at any point (x,y)
1
fˆ(x, y) fˆ g(s, t)
Saveetha Engineering 92
EC8093- Dr.M.Moorthi,Professor-
mn ( s,t )S x , y
Saveetha Engineering 93
EC8093- Dr.M.Moorthi,Professor-
Note: Using a convolution mask in which all coefficients have value 1/mn. Noise is
reduced as a result of blurring.
2. Geometric mean filter
Geometric mean filter is given by the expression
fˆ(x, y)
1
mn
g(s, t)
( s,t )S xy
3. Harmonic mean filter
The harmonic mean filter operation is given by the expression
fˆ(x, y) mn
1
g(s, t)
( s,t )
S x
g(s, t) Q1
fˆ(x, y)
( s,t )Sxy
g(s, t) Q
( s,t )Sxy
Where Q is called the order of the filter. This filter is well suited for reducing or virtually
eliminating the effects of salt-and-pepper noise.
8. Explain the various sharpening filters in spatial domain. (May / June 2009)
sharpening filters in spatial domain:
This is used to highlight fine details in an image, to enhance detail that has been
blurred and to sharpen the edges of objects
Smoothening of an image is similar to integration process,
Sharpening = 1 / smoothening
Since it is inversely proportional, sharpening is done by differentiation process.
Derivative Filters:
1. The derivatives of a digital function are defined in terms of differences.
2. It is used to highlight fine details of an image.
Properties of Derivative:
1. For contrast gray level, derivative should be equal to zero
2. It must be non-zero, at the onset of a gray level step or ramp
Saveetha Engineering 94
EC8093- Dr.M.Moorthi,Professor-
3. It must be non zero along a ramp [ constantly changes ]
Ist order derivative:
Saveetha Engineering 95
EC8093- Dr.M.Moorthi,Professor-
The first order derivative of one dimensional function f(x) is given as,
f
f (x 1)
f (x)
x
II nd order derivative:
The second order derivative of one dimensional function f(x) is given as,
2f
f 1) f (x 1) 2 f (x)
x 2
(x
Conclusion:
I. First order derivatives produce thicker edges in an image.
II. Second order derivatives have a stronger response to fine details , such as thin lines
and isolated points
III. First order derivatives have a stronger response to a gray level step
IV. Second order derivatives produces a double response at step changes in gray level
Laplacian Operator:
It is a linear operator and is the simplest isotropic derivative operator, which is
defined as,
(1)
Where,
(2)-
(3)
Sub (2) & (3) in (1), we get,
f 2
f (x 1, y) f (x 1, y) f (x, y 1) f (x, y 1) 4 f (x, ---------------(4)
y)
The above digital laplacian equation is implemented using filter mask,
0 1 0
1 -4 1
0 1 0
The digital laplacian equation including the diagonal terms is given as,
2 f (x 1, y) f (x 1, y) f (x, y 1) f (x, y 1) f (x 1, y 1) f (x 1, y 1)
f
f (x 1, y 1) f (x 1, y 1) 8 f (x, y)
the corresponding filter mask is given as,
Saveetha Engineering 96
EC8093- Dr.M.Moorthi,Professor-
1 1 1
1 -8 1
1 1 1
0 -1 0 -1 -1 -1
-1 4 -1 -1 -8 -1
0 -1 0 -1 -1 -1
g(x, y)
f (x, y) 2 f (x, y),
Substitute equation (4) in (6),
g(x, y) f (x, y) f (x 1, y) f (x 1, y) f (x, y 1) f (x, y 1) 4 f (x, y)
g(x, y) 5 f (x, y) f (x 1, y) f (x 1, y) f (x, y 1) f (x, y 1)
The above digital laplacian equation is implemented using filter mask,
0 -1 0
-1 5 -1
0 -1 0
Saveetha Engineering 97
EC8093- Dr.M.Moorthi,Professor-
f (x 1, y) f (x 1, y) f (x, y 1) f (x, y 1) f (x 1, y 1) f (x 1, y 1)
g(x, y) 9 f (x, y)
f (x 1, y 1) f (x 1, y 1)
-1 -1 -1
-1 -1 -1
-1 -8 -1
-1 A+8 -1
-1 -1 -1
-1 -1 -1
Gradient operator:
For a function f (x, y), the gradient f at co-ordinate (x, y) is defined as the 2-dimesional
column vector
∆f = Gx
Gy
= ∂f/∂x
∂f/∂y
2 2 1/2
∆f = mag (∆f) = [Gx2 + Gy2] ½= {[(∂f/∂x) +(∂f/∂y) ]}
Gx Gy
1. Roberts operator
consider a 3 3 mask is the following.
y
z1z2z3
z4z5z6
z7z8z9
Saveetha Engineering 98
EC8093- Dr.M.Moorthi,Professor-
Gx f ( x, y )
f ( x 1) f ( x) z6 z5
x
Gy f (x, y)
f ( y 1) f ( y) z8 z5
y
f Gx Gy
z6 z z z
5 8 5
1-1 1 0
0 0 -10
Roberts operator
Z8Z9
This approximation is known as the Roberts cross gradient operator, and is expressed
mathematically as follows
f z9 z8 z6
The above Equation can be z5
implemented using the following mask
-1 0 0 -1
.
0 1 1 0
Roberts operator
Saveetha Engineering 99
EC8093- Dr.M.Moorthi,Professor-
The original image is convolved with both masks separately and the absolute values of the
two outputs of the convolutions are added.
Saveetha Engineering 10
EC8093- Dr.M.Moorthi,Professor-
2. Prewitt operator
consider a 3 3 mask is the following.
y
z1 z4 z2 z5 z3 z6
z7 z8 z9
x
Gxf (x, y) r row1strow (z z z ) 1 z2 z3 )
x 3
d
7 8 (z
Gy f (x, y)
3rd col1st col ) z7 )
y
(z z6 z9 (z1 z4
3
y
-1 0 1 -1 -1 -1
-1 0 1 0 0 0
-1 0 1 1 1 1
Prewitt operator
This approximation is known as the x
op ator.
Prewitt
Saveetha Engineering 10
EC8093- Dr.M.Moorthi,Professor-
1.Multiply the input image by (-1)x+y to center the transform, image dimensions M x N
2.Compute F(u, v) DFT of the given image ,DC at M/2, N/2.
3. Multiply F(u, v) by a filter function H(u, v) to get the FT of the output image,
Saveetha Engineering 10
EC8093- Dr.M.Moorthi,Professor-
Definition: The filter which controls both high frequency and low frequency
components are called Homomorphic filtering.
Images normally consist of light reflected from objects. The basic nature of the image f(x,y)
may be characterized by two components:
(1) The amount of source light incident on the scene being viewed, &
These portions of light are called the illumination and reflectance components, and are
denoted i(x,y) and r(x,y) respectively. The functions i and r combine multiplicatively to give
the image function F:
f(x,y) = i(x,y)r(x,y),
where 0 < i(x,y) < 1-----indicates perfect black body -----indicates perfect absorption ,
and 0 < r(x,y) < 1 -----indicates perfect white body----indicates perfect reflection.
Since i and r combine multiplicatively, they can be added by taking log of the image
intensity, so that they can be separated in the frequency domain.
Saveetha Engineering 10
EC8093- Dr.M.Moorthi,Professor-
illumination component tends to vary slowly across the image and the reflectance
tends to vary rapidly. Therefore, by applying a frequency domain filter the intensity
variation across the image can be reduced while highlighting detail. This approach is shown
below.
BLOCK DIAGRAM:
Enhanced Image
g(x,y)
Analysis:
= DFT{ln[i(x,y)]} + DFT{ln[r(x,y)]}-----(2)
Saveetha Engineering 10
EC8093- Dr.M.Moorthi,Professor-
If we now apply a filter with a transfer function that suppresses low frequency
components and enhances high frequency components, then we can suppress the
illumination component and enhance the reflectance component.
Thus ,the Fourier transform of the output image is obtained by multiplying the DFT of
the input image with the filter function H(u,v).
where, io(x,y) = e i’(x,y) , ro(x,y) = e r’(x,y) are the illumination and reflection
components of the enhanced output image.
Saveetha Engineering 10
EC8093- Dr.M.Moorthi,Professor-
Input image
R Inverse
color
Color space transform Monochrome image enhancement algorithm
G space Output
transfor image
B m renderi
ng
Saveetha Engineering 10
EC8093- Dr.M.Moorthi,Professor-
PART – A
1. Define Image Restoration.
Image restoration refers to removal or minimization of known degradations in an
image i.e. Restoration attempts to reconstruct or recover an image that has been degraded.
2. Compare Image Enhancement with image restoration?
3. Show the block diagram of a model of the image degradation /restoration process.
Original g(x,y) Restored
image Degradation Restoration Image
f(x,y) function + filter f(x,y)
H R
Noise η(x,y)
Degradation Restoration
M
e
t
h
o
d
Blin
d
deco
nvol
utio
n
Saveetha Engineering 10
EC8093- Dr.M.Moorthi,Professor-
e2 E (f f )2
Where E{.} is the expected value off the argument.
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
g(x, y)
f (, )H[(x , y )]dd
But the impulse response of H is given as, h(x,α,y,β) = H[δ(x-α,y-β)]
Where h(x,α,y,β) is called Ponit spread function (PSF).
Therefore equation becomes,
Noise η(x,y)
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
PART B
1. Explain the image degradation model and its properties. (May / June 2009)
(or) Explain the mathematical model of Image degradation process
Noise η(x,y)
Degradation Restoration
[Blurring Filter]
[Deblurring /
Deconvolution filter]
The image f(x,y) is degraded with a function H and also a noise signal η(x,y) is added. The
restoration filter removes the degradation from the degraded image g(x,y) and restores the
original image f(x,y).
The Degradation model is given as,
where
g(x, y) is the degraded image, f (x, y) is the original image
H is an operator that represents the degradation process
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
Noise η(x,y)
Consider the general degradation model
g(x, y) H[ f (x, y)] (x, y)--------------------(1)
a) Property of Linearity:
If the external (x, y) is assumed to be zero, then equation (1) becomes,
noise
g(x, y) H[ f (x, y)]
H is said to be linear if
H af 1 (x, y) bf 2 (x, y) aH f 1 (x, y) bH f 2 (x, y) - (2)
Where, a and b are scalars, and f1 (x, y)and f 2 (x, y)are any two input images.
The linear operator posses both additivity and homogeneity property.
b)Property of Additivity:
If a= b= 1, then equation (2) becomes,
H f1 (x, y) f 2 (x, y) H f1 (x, y) H f 2 (x, y)
Is called the property of additivity. It states that , if H is a linear operator , then the response
to a sum of two inputs is equal to the sum of two responses.
c)Property of Homogeneity:
If f2(x,y) = 0, then equation (2) becomes,
H af 1 (x, y) aH f 1 (x, y) is called the property of homogeneity. It states that
the response to a constant multiple of any input is equal to the response of that input
multiplied by the same constant.
d)Property of Position or space invariant:
An operator having the input – output relationship as g(x, y) H[ f (x, y)] is said to be
position or space invariant if
H f (x , y ) g(x , y )
For any f(x,y) and any α and β.It states that the response at any point in the image
depends only on the value of the input at that point and not on its position.
2. Explain the degradation model for continuous function and for discrete function.
(May / June 2009)
a. degradation model for continuous function
An image in terms of a continuous impulse function is given as,
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
g(x, y)
H[ f (, )(x , y )dd]
Since f(α,β) is independent of x & y , using homogeneity property, we get,
g(x, y)
f (, )H[(x , y )]dd--------------------------------(3)
But the impulse response of H is given as, h(x,α,y,β) = H[δ(x-α,y-β)]
Where h(x,α,y,β) is called Ponit spread function (PSF).
Therefore equation (3) becomes,
g(x, y)
f (, )h(x,, y, )dd------------------------------------(4)
This equation is called as Superposition or Fredholm Integral of first kind.
It states that if the response of H to an impulse is known, then the response to any input
f(α,β) can be calculated using the above equation.
Thus, a linear system H is completely characterized by its impulse response.
If H is position or space invariant, that is,
H (x , y ) h(x , y )
Then equation (4) becomes,
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
g e (x)
m0
f e (m)he (x m)
In matrix form
g Hf ,
where g and h are M dimensional column vectors
fe (0) ge (0)
e
f f (1 , g g (1 , and H is a M x M matrix given as,
e
⁝ ) )
⁝
f (M 1) g (M 1)
he … he (M 1)
(0) he
(1)
he (1) … h (M 2)
H he (0) e
(M M) ⁝ ⁝ ⋱ ⁝
h (M 1) h (M 2) … h (0)
e e e
Assuming he (x) is periodic, then he (x) = he (M x) and hence,
he he (M … he (1)
(0)
1) … h (2)
H he (1) he (0) e
(M M) ⁝ ⁝ ⋱ ⁝
h (M h (M 2) … h (0)
e e e
1)
g e (x, y)
m0 n0
f e (m, n)he (x m, y n) ne (x, y)
In matrix form,
g Hf
where f , g and are MN dimensional column vectors and H is MN x MN dimension. It
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
consists of M2 partition , each partition having dimension of size M x N and ordered
accordingly and is given as ,
H HM1 … H1
0
1 H0 … H2
H H
⁝ ⁝ ⋱ ⁝
H H … H
M M 0
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
Each partition Hj is constructed from the jth row of the extended function he (x, is,
y)
he ( he ( j, N … he ( j,1)
j,0)
1) … h ( j,2)
H h e( j,1) he ( j,0) e
j ⁝ ⁝ ⋱ ⁝
h ( j, N 1) h ( j, N 2) … h ( j,0)
e e e
3. Write notes on inverse filtering as applied to image restoration. (May / June 2009)
(or) Write short notes on Inverse filter (Nov / Dec 2005)
Inverse filter:
The simplest approach to restoration is direct inverse filtering; an estimate ^
F (u, v) of
the transform of the original image is obtained simply by dividing the transform of the
degraded image G(u, v) by the degradation function.
^
G(u, v)
F (u, v)
H (u, v)
Problem statement:
Given a known image function f(x,y) and a blurring function h(x,y), we need to recover
f(x,y) from the convolution g(x,y) = f(x,y)*h(x,y)
Solution by inverse filtering:
a.Without Noise:
F (u, v) G(u, v)
H (u,
----------------------------(1)
(2), ge F (u,
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
v) F (u, v) (u, v)G(u, v) -----------(2)
H (u,
v)
Hence exact recovery of the original image is possible.
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
b. With Noise:
Noise η(x,y)
^
F (u, v) F (u, v) (u, v)
H (u, v) (5)
Here, along with the original image , noise is also added. Hence the restored image consists
of noise.
Drawbacks:
1. Doesnot perform well when used on noisy images
1 1
2. In the absence of noise, If H(u,v)=0,then R(u, v) H 1
(u, v) =Infinite,
H (u, 0
from equation (2) v)
3. In the presence of noise, when H(u,v) is small, then the noise η(u,v) dominates(from
equation(5)
Pseudo inverse filtering.:The solution is to carry out the restoration process in a limited
neighborhood about the origin H (u, is not very small. This procedure is called
where v)
Pseudo inverse filtering. In this case,
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
1
H (u, v)
2
H (u, v)
Fˆ (u, v)
H (u, v)
0
The threshold is defined by the user. In general, the noise may possess large components
at high frequencies (u, v) , H (u, and G(u, v) normally will be dominated by low
while frequency components. v)
4. Write short notes on the Wiener filter characteristics. (Nov / Dec 2005) (or)
Derive the Wiener filter equation. Under what conditions does it become the pseudo
inverse filter?
OBJECTIVE:
To find an estimate of the uncorrupted image f such that the mean square error
between them is corrupted. This error measure is given as,
e 2 2
E f f
ASSUMPTION:
1. Noise and Image are uncorrelated
2. Then either noise or image has a zero mean
3. Gray levels in the estimate are a linear function of the levels in the degraded
image
Let, H (u, = Degradation function
v)
H * (u, v) = Complex conjugate of H(u,v)
H (u, v) 2
H * (u, v).H (u, v)
S(u, v) N (u, v) 2
Power spectrum of the noise
S (u, v) F (u, v) 2 Power spectrum of the undegraded image
f
Wiener Model with Noise:
Original g(x,y) Restored
image Degradation Wiener Image
f(x,y) function + filter f(x,y)
H W= H-1
Noise η(x,y)
-------------------(2) ---
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
1 .G(u, v)
S (u, v)
H (u, v) S f (u, v) xH * (u, v)
Thus when the filter consists of the terms inside the bracket, it is called as the minimum
mean square error filter or the least square error filter.
Case 1:If there is no noise, Sη(u,v) = 0, then the Wiener filter reduces to pseudo Inverse
filter.
From equation (2),
1
H (u, v)
0
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
H * (u, v).S f (u, v) H *(u, v)
W (u, v)
| H (u, v) |2 Sf (u, v) 0 H * (u, v)H (u, v)
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
,when H(u,v)≠0
, when H(u,v)=0
H * (u, v)
S (u, v)
H * (u, v).H (u, v)
S f (u, v)
1
S ( u , v )
1
S( fu , v )
1
1 K
Where
S (u, v)
K
S f (u, v)
W (u, v) 1 S f (u, v)
1 K S (u, v) S (u, v)
and f
S f (u,v) (SNR)
W (u,v)
S f (u,v) Sn (u,v) (SNR) 1
(i)
( SN ) 1 W ( u , v ) 1
R
(ii) ( SN ) 1 W ( u , v ) ( SNR )
R
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
(SNR) is high in low spatial frequencies and low in high spatial frequencies so
W (u, v) can be implemented with a low pass (smoothing) filter.
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
Wiener filter is more robust to noise and preserves high frequency details SNR is low,
Estimate is high
5. Explain in detail the constrained least squares restoration. (May / June 09).
In Wiener filter, the power spectra of the undegraded image and noise must be
known. Although a constant estimate is sometimes useful, it is not always suitable.
Constrained least squares filtering just require the mean and variance of the noise.
With Noise:
Noise η(x,y)
The blurred image is given as,
η = g - Hf
----- ---(1)
Problem statement:
To find the minimum of a criterion function C defined as,
M 1 N 1
2
C 2
f (x, y) ,
x 0 y 0
Wher w 2
wT w is the Euclidean vector norm,
e ^
f is the estimate of the undegraded image, and
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
0 1 0
p(x, y) 1 4 1
0 1 0
Here the noise is replaced by P(u,v)…i.e., Noise is approximated to Laplace such that mean
= 0.If mean is zero, the amount of DC component is zero ….i.e, overall brightness of image
^
is zero.The value of γ is selected by iteration such that F (u, is a good approximation of
F(u,v). v)
One Approach to find the value of γ,
A residual vector r is defined r g H f ,---------------(5)
as,
^
Since F (u, v) is a function of γ, then γ is also a function of r,
i.e., r 2
(),
2
from equation (3) and (5), r 2
, then f^ =f ia a good estimation
if
Therefore,
r 2
2 a ,
Steps:
I. Specify an initial value of “γ”
II. Compute r 2 , which is given as a function of γ,
i.e., 2
r (),
2
III. Check whether r
2 2
a, is satisfied , else increase γ if r 2
a, or
decrease γ if r 2
2 a,
^
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
Another approach to find the value of γ is Newton – raphson algorithm:
By knowing the mean and variance of noise, we can estimate the restored image.
We know that the residual vector is defined as,
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
rgH
f
M 1 N 1
Where 2
r r
x 0 y 0
2
(x, y),
The variance of the noise over the entire image is given by the sample average method,
M 1 N 1
1
(x, ,
2
2
MN
y) m
x 0 y 0
M 1 N 1
Where sample mean is m 1 (x, y),
MN
x0 y 0
MN 2 m ,
2
From the above expressions,
Thus the value for the constraint is determined in terms of noise mean and variance.
6. Discuss about the Lagrange Multipliers.
Lagrange Multipliers
Lagrange multiplier is a method to find maximum or minimum value of a
function subject to some constrains (side conditions) on the domain of the function.
Examples:
Maximize or minimize f(x,y) subject to condition g(x,y) = c.
Maximize or minimize f(x,y,z) subject to condition g(x,y,z) = c.
Maximize or minimize f(x,y,z) subject to conditions g(x,y,z) = c1,and h(x,y,z) = c2.
Lagrange Theorem:
Let f and g have continuous first partial derivatives such that f has a local
maximum or local minimum at (x0 , y0 ) on the smooth constraint curve g( x, y) = c. If
g(x0 , y0 ) = 0, then there is a real number λ such that
f (x0 , y 0 ) g(x0 , y 0 )
. λ is Lagrange multiplier.
Lagrange Method:
To find extrema of f( x ,y ) subject to constraint g( x, y ) = 0:
I. Solve the equations
f (x, y) g (x, y) and g( x, y ) = 0, by solving
f x (x, y) g x (x, y)
f y (x, y) g y (x, y)
g( x, y) 0.
Find all solutions ( x1 , y1 ( x 2 , y 2 , 2 ( x 3 , y 3 , 3 ), . . .
, 1 ), ),
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
Discard the Lagrange 1 , 2 , 3 ,. . .
multipliers
II. Evaluate f( x, y ) at each solution point obtained above.
The largest value yields the maximum and the smallest value yields the
minimum subject to constraint g( x, y ) = 0.
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
The conditions in 3 are called first-order conditions for the problem of maximizing or
minimizing a function f ( x, y) subject to g x, y c
Example: Solve the f x, y x2 y2 subject to
problem:Maximize
g x, y x2 xy y2 3
The Lagrangian
L x, y x y x xy y 3
2 2 2 2
is
Lx x, y 2x 2x y
0 Ly x, y 2 y x 2 y
0 x2 xy y2 3 0
2
2 and
2x y if 2x y 0 if 2 y x 0
y2y
2x 2
Combining them together we get 2 2
and x y or x y or x y
y
2x y y2y
Suppose y x From 3 we get x2 1 , so x or x 1and it gives two solutions
1
2
x, y 1,1 and 1, 1 with
3
Suppose y x From 3 we get x 3 , so x 3 or x 3 and it gives two solutions
2
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
x, y 3, 3 and
3, 3 with 2
Two more cases 2x y 0 and 2 y x 0 yield no solutions since they result
in x or y 0 which contradicts the constraint x2 xy y2 3 0 .
0
Two solutions x, y 1,1 and 1, 1 lead to f 1,1 1, 1 2 while the solutions
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
x, y
3, 3
3, 3 lead to f 3, 3 f 3, 3 6
The first two solutions solve minimization problem while the last two solve maximization
problem.
a. Unconstrained Reconstruction
If we know very little about the noise , then we try to find an estimate image fr,
such that Hfr approximates g in a least-squares manner. This can be accomplished by
minimizing the norm of the noise . Squaring the norm of both sides of Eq. (6) after
substituting for f by the estimate vector fr and moving Hfr to the other side of the equation,
we get
Where ||a||2 is the square of the norm of vector a and is given by ||a||2 = aa’, where a’ is the
transpose of the vector a.
Consider the error function E, where
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
Then our goal is to minimize E with respect to fr. This can be accomplished by taking the
derivative of E with respect to fr, and equating the result to zero, that is
Assuming that H-1 exists, the solution for this equation is given as
b.Constrained Reconstruction
To account for the noise term in Eq. (6), we introduce the square matrix Q (M2 × M2)
to represent some linear operator on f. By selecting different Q's, we are able to set the goal
of restoration as desired. Equation (8) is now modified to
Where constant is called the LaGrange multiplier. Again we try to minimize the error
function E, by taking its derivative with respect to fr and equating the result to zero, that is
x1 = r(x, y)..............1
where r(x, y) and s(x, y) are the spatial transformations that produced the
geometrically distorted image g(x1, y1). For example, if r(x, y) = x /2 and s (x, y) = y/2, the
“distortion” is simply a shrinking of the size of f (x, y) by one half in both spatial directions.
If r(x, y) and s(x, Y) were known analytically, recovering f (x, y) from the distorted image
g(x1, y1) by applying the transformation in reverse might be possible theoretically. In
practice, however, formulating a single set of analytical functions r(x, y) and s(x, y) that
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
describe the geometric distortion process over the entire image plane generally is not
possible. The method used most frequently to overcome this difficulty is to formulate the
spatial relocation of pixels by the use of tiepoints, which are a subset of pixels whose
location in the input (distorted) and output (corrected) images is known precisely.
Figure shows quadrilateral regions in a distorted and corresponding corrected image.
The vertices of the quadrilaterals are corresponding tiepoints.
Suppose that the geometric distortion process within the quadrilateral regions is
modeled by a pair of bilinear equations so that
Since there are a total of eight known tiepoints, these equations can be solved for the
eight coefficients ci, I = 1,2, …….8. The coefficients constitute the geometric distortion
model used to transform all pixels within the quadrilateral region defined by the tiepoints
used to obtain the coefficients. In general, enough tiepoints are needed to generate a set of
quadrilaterals that cover the entire image, with each quadrilateral having its own set of
coefficients.
Once we have the coefficients, the procedure used to generate the corrected (i.e.,
restored) image is not difficult. If we want to find the value of the undistorted image at any
point (x0, y0), we simply need to know where in the distorted image f(x 0, y0) was mapped.
This we find out by substituting (x0, y0) into Equations (5) and (60 to obtain the
1 1
geometrically distorted coordinates (x0 , y ) . The value of the point in the undistorted
image
1 1 1 1
that was mapped to (x , y ) is g (x , y ) . So we obtain the restored image value simply by
0 0 0 0
Saveetha Engineering 13
EC8093- 1 1 Dr.M.Moorthi,Professor-
‸
‸
letting f (x0 , y 0 ) = g (x , y ) . For example, to generate f (0,0) , we substitute (x, y) = (0, 0)
0
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
into Equations (5) and (6) to obtain a pair of coordinates (x1, y1) from those equations. Then
we let f‸(0,0) = g (x1, y1), where x1 and y1 are the coordinate values just obtained. Next, we
substitute (x, y) = (0, 1) into Equations (5) and (6), obtain another pair of values (x1, y1),
and let f(0, 1) = g(x1, y1) for those coordinate values. The procedure continues pixel by
pixel and row by row until an array whose size does not exceed the size of image g is
obtained. A column (rather than a row) scan would yield identical results. Also, a
bookkeeping procedure is needed to keep track of which quadrilaterals apply at a given
pixel location in order to use the proper coefficients.
Tie points are established by a number of different techniques, depending on the
application. For instance, some image generation systems having physical artifacts (such as
metallic points) embedded on the imaging sensor itself. These produce a known set of
points (called reseau marks) directly on the image as it is acquired. If the image is distorted
later by some other process (such as an image display or image reconstruction process),
then the image can be geometrically corrected using the technique just described.
Gray-Level Interpolation
The method discussed in the preceding section steps through integer values of the
coordinates (x, y) to yield the restored image. However, depending on the values of the
coefficients CI , Equations (5) and (6) can yield no integer values
for x1 and y1. Because the distorted images g is digital, its pixel values are defined only at
integer coordinates. Thus using no integer values for x 1 and y1 causes a mapping into
locations of g for which no gray levels are defined. inferring what the gray-level values at
those locations should be, based only on the pixel values at integer coordinate locations,
then becomes necessary. The technique used to accomplish this is called gray-level
interpolation.
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
images of high resolution. For general-purpose image processing a bilinear interpolation
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
approach that uses the gray levels of the four nearest neighbors usually is adequate. This
approach is straightforward. Because the gray level of each of the four integral nearest
neighbors of a nonintegral pair of coordinates (x1, y1) is known, the gray-level value at
these coordinates, denoted (x1, y1) , can be interpolated from the values of its neighbors by
using the relationship
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
PART A
1. Define Image Segmentation (May / June 2009)
Segmentation is a process of subdividing an image in to its constituted regions or
objects or parts. The level to which the subdivision is carried depends on the problem being
solved, i.e., segmentation should be stopped when object of interest in an application have
been isolated.
Image segmentation techniques are used to isolate the desired object from the scene
so that measurements can be made on it subsequently.
Applications of segmentation.
* Detection of isolated points.
* Detection of lines and edges in an image
2. What are the different image segmentation techniques?
Image segmentation is classified as follows:
Template matching
Thresholding
Boundary detection
Clustering
Quad- trees
Texture matching
3. What are the two properties that are followed in image segmentation?
Image segmentation algorithms generally are based on one of the basic properties of
intensity values
1. discontinuity
2. similarity
4. Explain the property of discontinuity?
The principal approach of this property is to partition an image based on abrupt
changes in intensity such as edges in an image.
5. What are the three types of discontinuity in digital image?
Points, lines and edges.
6. What is the idea behind the similarity property?
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
The principal approach in the property is based on portioning an image into regions
that are similar according to a set of predefined criteria. Thresholding, region growing and
region splitting and merging are examples of this method.
7. What is edge?
An edge is set of connected pixels that lie on the boundary between two regions edges
are more closely modeled as having a ramp like profile. The slope of the ramp is inversely
proportional to the degree of blurring in the edge.
8. What is edge detection? Explain. (May / June 2009)
Edge direction is fundamental importance in image analysis. Edges characterize
object boundaries and are therefore useful for segmentation, registration an identification of
objects in scenes. Edge points can be thought of as pixel locations of abrupt gray- level
change.
Edge detection is used for detecting discontinuities in gray level. First and second
order digital derivatives are implemented to detect the edges in an image.
9. Define zero crossing property of edge detection.
An imaginary straight line joining the extreme positive and negative values of the
second derivative could cross zero near the midpoint of the edge. Zero crossing property is
useful for locating the centers of thick edges.
10. What is meant by gradient operators?
First order derivatives in an image are computed using the gradient operators. The
gradient of an image f(x,y) at location (x,y) is defined as the vector.
Gx fx
f
Gy
f
y
The f is called as the gradient. Magnitude of the vector is
2 2 l/2
∆f=mag f=[Gx +Gy ]
-l
α(x,y)=tan (Gy/Gx),α(x,y) is the direction angle of vector ∆f
11. Define Laplacian.
The Laplacian of a 2 D function f(x,y) is a second order derivative defined as
2 2f
f 2
f
2
x y2
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
18. What is the role of gradient operator and laplacian operator in segmentation?
The first and second derivative at any point in an image can be obtained by using
the magnitude of the gradient operators at that point and laplacian operators
19. What is the role of first derivative and second derivative in segmentation?
The magnitude of the first derivative is used to identify the presence of an edge in the
image.
The second derivative is positive for the part of the transition which is associated with the
dark side of the edge. It is zero for pixel which lies exactly on the edge.
It is negative for the part of transition which is associated with the light side of the edge.
The sign of the second derivative is used to find that whether the edge pixel lies on the light
side or dark side.
20. What are the types of thresholding?
There are three different types of thresholds. They are
1. global thresholding
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
2. local thresholding
3. dynamic (or) Adaptive
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
Catchment basins
Local minima
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
P Ri P R j , Ri , R j are neighbors.
2. Explain the various detection of discontinuities in detail.(or)Explain Edge detection
in detail
Detection of discontinuities
Discontinuities such as isolated points, thin lines and edges in the image can be
detected by using similar masks as in the low- and high-pass filtering. The absolute value of
the weighted sum given by equation (2.7) indicates how strongly that particular pixel
corresponds to the property described by the mask; the greater the absolute value, the
stronger the response.
1. Point detection
The mask for detecting isolated points is given in the Figure. A point can be defined
to be isolated if the response by the masking exceeds a predefined threshold:
f x T
-1 -1 -1
-1 8 -1
-1 -1 -1
-1 -1 -1 -1 -1 2 -1 2 -1 2 -1 -1
2 2 2 -1 2 -1 -1 2 -1 -1 2 -1
-1 -1 -1 2 -1 -1 -1 2 -1 -1 -1 2
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
THICK EDGE:
The slope of the ramp is inversely proportional to the degree of blurring in the edge.
We no longer have a thin (one pixel thick) path.
Instead, an edge point now is any point contained in the ramp, and an edge would
then be a set of such points that are connected.
The thickness is determined by the length of the ramp.
The length is determined by the slope, which is in turn determined by the degree of
blurring.
Blurred edges tend to be thick and sharp edges tend to be thin
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
The first derivative of the gray level profile is positive at the trailing edge, zero in
the constant gray level area, and negative at the leading edge of the transition
The magnitude of the first derivative is used to identify the presence of an edge in
the image
The second derivative is positive for the part of the transition which is associated
with the dark side of the edge. It is zero for pixel which lies exactly on the edge.
It is negative for the part of transition which is associated with the light side of the
edge
The sign of the second derivative is used to find that whether the edge pixel lies on
the light side or dark side.
The second derivative has a zero crossings at the midpoint of the transition in gray
level. The first and second derivative t any point in an image can be obtained by
using the magnitude of the gradient operators at that point and laplacian operators.
1. FIRST DERIVATE. This is done by calculating the gradient of the pixel relative to its
neighborhood. A good approximation of the first derivate is given by the two Sobel
operators as shown in the following Figure , with the advantage of a smoothing effect.
Because derivatives enhance noise, the smoothing effect is a particularly attractive feature
of the Sobel operators.
First derivatives are implemented using the magnitude of the gradient.
i. Gradient operator:
For a function f (x, y), the gradient f at co-ordinate (x, y) is defined as the 2-
dimesional column vector
∆f = Gx Gy
∂f/∂x
∂f/∂y
=
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
2 2 1/2
∆f = mag (∆f) = [Gx2 + Gy2] ½= {[(∂f/∂x) +(∂f/∂y) ]}
Gx Gy
Let (x,y) represent the direction angle of the vector f at (x,y)
(x,y) = tan-1(Gy/Gx)
The direction of an edge at (x,y) is perpendicular to the direction of the gradient
vector at that point
1 Gx
The direction of the gradient is: tan G
y
a) Roberts operator
consider a 3 3 mask is the following.
y
z1z2z3
z5
z4 z6
z7z8z9
Gx f ( x, y )
x f ( x 1) f ( x) z6 z5
Gy f ( y)
f (x, y) f ( y 1) z8 z5
y
f Gx Gy
z6
z5 z8 z5
1-1 1 0
0 0 -10
EC8093- Dr.M.Moorthi,Professor-
Roberts operator
Save etha Engineering 112
EC8093- Dr.M.Moorthi,Professor-
Z8Z9
This approximation is known as the Roberts cross gradient operator, and is expressed
mathematically as follows
f z9 z5 z8 z6
-1 0 0 -1
0 1 1 0
Roberts operator
The above Equation can be implemented using the following mask.
The original image is convolved with both masks separately and the absolute values of the
two outputs of the convolutions are added.
b) Prewitt operator
consider a 3 3 mask is the following.
y
z1z2z3
z4z5z6
z7z8z9
x
Gx f ( x, y ) 3 rd row 1 srow ( z 7 z 8 z 9 ) ( z z3)
x
t
z1 2
Gy f ( x , y )
y 3 rd col 1 st col ( z 3 z 6 z 9 ) ( z 1 z z7 )
4
y
-1 0 1 -1 -1 -1
-1 0 1 0 0 0
-1 0 1 1 1 1
Saveetha Engineering College 113 x
EC8093- Dr.M.Moorthi,Professor-
Prewitt operator
EC8093- Dr.M.Moorthi,Professor-
Gy x3 2x 6 x9 x1 2x 4 x7
-1 -2 -1 -1 0 1
0 0 0 -2 0 2
1 2 1 -1 0 1
Figure Sobel masks for edge detection
The following shows the a 3 x3 region of image and various masks used to compute
the gradient at point z5
ii. SECOND DERIVATE can be approximated by the Laplacian mask given in Figure .
The drawbacks of Laplacian are its sensitivity to noise and incapability to detect the
direction of the edge. On the other hand, because it is the second derivate it produces
double peak (positive and negative impulse). By Laplacian, we can detect whether the pixel
lies in the dark or bright side of the edge. This property can be used in image segmentation.
0 -1 0
-1 4 -1
0 -1 0
Figure : Mask for Laplacian (second derivate).
a) Laplacian Operator:
It is a linear operator and is the simplest isotropic derivative operator, which is
defined as,
(1)
Where,
(2)-
(3)
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
0 1 0
1 -4 1
0 1 0
The digital laplacian equation including the diagonal terms is given as,
2 f (x 1, y) f (x 1, y) f (x, y 1) f (x, y 1) f (x 1, y 1) f (x 1, y 1)
f
f (x 1, y 1) f (x 1, y 1) 8 f (x, y)
1 1 1
1 -8 1
1 1 1
Similarly , two other implementations of the laplacian are,
-1 -1 -1
0 -1 0
-1 -8 -1
-1 4 -1
-1 -1 -
0 -1 0
Edge linking
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
Edge detection algorithm is followed by linking procedures to assemble edge pixels into
meaningful edges.
Basic approaches
1. Local Processing
2. Global Processing via the Hough Transform
3. Global Processing via Graph-Theoretic Techniques
1. Local processing
Analyze the characteristics of pixels in a small neighborhood (say, 3x3, 5x5) about every
edge pixels (x,y) in an image.All points that are similar according to a set of predefined
criteria are linked, forming an edge of pixels that share those criteria
Criteria
1. The strength of the response of the gradient operator used to produce the edge pixel
an edge pixel with coordinates (x0,y0) in a predefined neighborhood of (x,y)
is similar in magnitude to the pixel at (x,y) if
|f(x,y) - f (x0,y0) | E
2. The direction of the gradient vector
an edge pixel with coordinates (x0,y0) in a predefined neighborhood of (x,y)
is similar in angle to the pixel at (x,y) if
|(x,y) - (x0,y0) | < A
3. A Point in the predefined neighborhood of (x,y) is linked to the pixel at (x,y) if both
magnitude and direction criteria are satified.
4. The process is repeated at every location in the image
5. A record must be kept
6. Simply by assigning a different gray level to each set of linked edge pixels.
7. find rectangles whose sizes makes them suitable candidates for license plates
8. Use horizontal and vertical Sobel operators ,eliminate isolated short segments
• link conditions:
gradient value > 25
gradient direction differs < 15
2. Global Processing via the Hough Transform
Suppose that we have an image consisting of several samples of a straight line, Hough
[1962] proposed a method (commonly referred as Hough transform) for finding the line (or
lines) among these samples (or edge pixels). Consider a point (xi, yi). There is an infinite
number of lines passing through this point, however, they all can be described by
yi a xi b
This means that all the lines passing (xi, yi) are defined by two parameters, a and b. The
equation can be rewritten as
b xi a yi
Now, if we consider b as a function of a, where xi and yi are two constants,
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
the parameters can be represented by a single line in the ab-plane. The Hough transform is a
process where each pixel sample in the original xy-plane is transformed to a line in the ab-
plane. Consider two pixels (x1, y1) and (x2, y2). Suppose that we draw a line L across these
points. The transformed lines of (x1, y1) and (x2, y2) in the ab-plane intersect each other in
the point (a', b'), which is the description of the line L, see Figure . In general, the more
lines cross point (a, b) the stronger indication that is that there is a line y a x in the
b
image.
To implement the Hough transform a two-dimensional matrix is needed, see Figure. In each
cell of the matrix there is a counter of how many lines cross that point. Each line in the ab-
plane increases the counter of the cells that are along its way. A problem in this
implementation is that both the slope (a) and intercept (b) approach infinity as the line
approaches the vertical. One way around this difficulty is to use the normal representation
of a line:
x cos y sin d
Here d represents the shortest distance between origin and the line.
represents the angle of the shortest path in respect to the x-axis. Their corresponding ranges
are [0, 2D], and [- 90, 90], where D is the distance between corners in the image.Although
the focus has been on straight lines, the Hough transform is applicable to any other shape.
For example, the points lying on the circle
2 2
x c1 yc
2
c 2
3
can be detected by using the approach just discussed. The basic difference is the presence of
three parameters (c1, c2, c3), which results in a 3-dimensional parameter space.
y b
L b = -x a + y
x2 ,y2 2
x1 ,y1
b'
b = -x1 a + y1
x a
a'
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
bmax
...
0 ... ...
bmin
...
amin 0 a
amax
Figure : Quantization of the parameter plane for use in Hough transform.
y d
d max
...
0 ... ...
d
d
...
x min
min 0
max
Figure : Normal representation of a line (left); quantization of the d-plane into cells.
Thresholding
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
Suppose that the image consists of one or more objects and background, each
having distinct gray-level values. The purpose of thresholding is to separate these
areas from each other by using the information given by the histogram of the image.
If the object and the background have clearly uniform gray-level values, the object
can easily be detected by splitting the histogram using a suitable threshold value.
The threshold is considered as a function of:
T g i, j, xi, j , p i, j
where i and j are the coordinates of the pixel, xi,j is the intensity value of the pixel,
and p(i, j) is a function describing the neighborhood of the pixel (e.g. the average
value of the neighboring pixels).
A threshold image is defined as:
1 if xi, j T
f x, y
0 if x i, j T
Pixels labeled 1 corresponds to objects and Pixels labeled 0 corresponds to
background
T
T1 T2
Figure : Histogram partitioned with one threshold (single threshold),and with two
thresholds (multiple thresholds).
Saveetha Engineering 11
EC8093- Dr.M.Moorthi,Professor-
The thresholding can be classified according to the parameters it depends on:
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
Assume the larger PDF to correspond to the back ground level and the smaller PDF
to gray levels of objects in the image. The overall gray-level variation in the image,
p(z) P1 p1 (z) P2 p2 (z)
P1 P2 1
Where P1 is the pbb that the pixel is an object pixel, P2 is the pbb that the pixel is a
background pixel.
The probability of error classifying a background point as an object point is,as
T
E1 (T )
p
2 (z)dz
This is the area under the curve of P2(z) to the left of the threshold
The probability of error classifying a object point as an background point is,as
T
E2 (T ) p1 (z)dz
This is the area under the curve of P1(z) to the right of the threshold
Then the overall probability of error is E(T ) P E (T ) (T )
PE
2 1 1 2
where
1 and 12 are the mean and variance of the Gaussian density of one object
2 and 22 are the mean and variance of the Gaussian density of the other object
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
Substituting equation (2) in (1), the solution for the threshold T is given as ,
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
AT 2
BT C 0
where A
2 2
2 2
B 2 ( 1 ) 2
1 2 2 1
C 2 2 2 2 2 2 2 ln( P / P )
2
1 2 2 1 1 2 2 1 1 2
Since the quadratic equation has two possible solutions, two threshold values are
required to obtain the optimal solution.
If the variances are equal, a single threshold is sufficient given as,
2 P
T 1 2
ln 2
2 1 2 P1
Threshold selection Based on Boundary Characteristics:
The threshold value selection depends on the peaks in the histogram
The threshold selection is improved by considering the pixels lying on or near the
boundary between the objects and background
A three level image can be formed using the gradient, which is given as,
Light object of dark background
0
if f T
s(x, y)
if f T and f 0
2
if f T and 2 f 0
The transition from object to the back ground is represented by the occurrence of
‘+‘ followed by ‘-’
Thus a horizontal or vertical scan line containing a section of an object has the
following structure: (…)(-,+)(0 or +)(+,-)(…)
5. Write short notes on Region growing segmentation. Or Explain Region splitting and
Merging (May / June 2009)
Region growing is a procedure that groups pixels or sub regions in to layer regions
based on predefined criteria. The basic approach is to start with a set of seed points and
from there grow regions by appending to each seed these neighboring pixels that have
properties similar to the seed
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
Region growing (or pixel aggregation) starts from an initial region, which can be a
group of pixels (or a single pixel) called seed points. The region is then extended by
including new pixels into it. The included pixels must be neighboring pixels to the region
and they must meet a uniform criterion.
criteria: (i) The absolute gray-level difference between any pixel and the seed has
to be less than 65 and (ii) the pixel has to be 8-connected to at least one pixel in that region
(if more, the regions are merged) .This is selected from the following histogram
The growing continues until a stopping rule takes effect. The region growing has several
practical problems:
how to select the seed points
growing rule (or uniform criterion)
stopping rule.
All of these parameters depend on the application. In the infra-red images in
military applications the interesting parts are the areas which are hotter compared to their
surroundings, thus a natural choice for the seed points would be the brightest pixel(s) in the
image. In interactive applications the seed point(s) can be given by the user who gives (by
pointing with the mouse, for example) any pixels inside the object, and the computer then
finds the desired area by region growing algorithm.
The growing rule can be the differences of pixels. The value of the pixel in
consideration can be compared to the value of its nearest neighbor inside the region.
If the difference does not exceed a predefined threshold the pixel is attended to the
region. Another criterion is to examine the variance of the region, or gradient of the pixel.
Several growing criteria worth consideration are summarized in the following:
Global criteria:
Difference between xi,j and the average value of the
region. The variance 2 of the region if xi,j is attended to
Local criteria: it.
Difference between xi,j and the nearest pixel inside the region.
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
The quad tree segmentation is a simple but still powerful tool for image
decomposition.
In this technique , an image is divided into various sub images of disjoint
regions and then merge the connected regions together.
Steps involved in splitting and merging
Split into 4 disjoint quadrants any region Ri for which P(Ri)=FALSE.
Merge any adjacent regions Rj and Rk for which P(RjURk)=TRUE.
Stop when no further merging or splitting is possible.
The following shows the partitioned image,
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
Catchment basins
Local minima
Watershed lines
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
rise through the holes at a uniform rate. When the rising water in distinct catchment basins
is about to merge, a dam is built to prevent the merging. The flooding will eventually reach
a stage when only the tops of the dams arc visible above the water line. These dam
boundaries correspond to the divide lines of the watersheds. Therefore, they are the
(continuous) boundaries extracted by a watershed segmentation algorithm.
These ideas can be explained further with the aid of Fig.. Figure (a) shows a simple
gray-scale image and Fig. (b) is a topographic view, in which the height of the "'mountains ’’
is proportional to gray-level values in the input image. For ease of interpretation, the
backsides of structures are shaded. This is not to be confused with gray-level values; only
the general topography of the three-dimensional representation is of interest. In order to
prevent the rising water from spilling out through the edges of the structure, we imagine the
perimeter of the entire topography (image) being enclosed by dams of height greater than
the highest possible mountain, whose value is determined by the highest possible gray-level
value in the input image.
Suppose that a hole is punched in each regional minimum [shown as dark areas in
Fig. (b)] and that the entire topography is flooded from below by letting water rise through
the holes at a uniform rate. Figure (c) shows the first stage of flooding, where the “water”
shown in light gray, has covered only areas that correspond to the very dark background in
the image. In Figs. (d) and (e) we see that the water now has risen into the first and second
catchment basins, respectively. As the water continues to rise, it will eventually overflow
Saveetha Engineering 12
EC8093- Dr.M.Moorthi,Professor-
from one catchment basin into another. The first indication of this is shown in (f). Here,
water from the left basin actually overflowed into the basin on the right and a short "dam ”
(consisting of single pixels) was built to prevent water from merging at that level of
flooding (the details of dam building are discussed in the following section), The effect is
more pronounced as water continues to rise, as shown in Fig. (g).This figure shows a longer
dam between the two catchment basins and another dam in the top part of the right basin.
The latter dam was built to prevent merging of water from that basin with water from areas
corresponding lo the background. This process is continued until the maximum level of
flooding (corresponding to the highest gray-level value in the image) is reached. The final
dams correspond to the watershed lines, which are the desired segmentation result. The
result for this example is shown in Fig. (h) as a dark, one-pixel-thick path superimposed on
the original image. Note the important property that the watershed lines form a connected
path, thus giving continuous boundaries between regions.
One of the principal applications of watershed segmentation is in the extraction of
nearly uniform (blob like) objects from the background. Regions characterized by small
variations in gray levels have small gradient values. Thus, in practice, we often see
watershed segmentation applied to the gradient of an image, rather than to the image itself.
In this formulation, the regional minima of catchment basins correlate nicely with the small
value of the gradient corresponding to the objects of interest.
2. Dam Construction
Dam construction is based on binary images, which are members of 2-D integer
space Z2. The simplest way to construct dams separating sets of binary points is to use
morphological dilation.
The basics of how to construct dams using dilation are illustrated in Fig. Figure (a)
shows portions of two catchment basins at flooding step n - 1 and Fig. (b) shows the result
at the next flooding step, n. The water has spilled from one basin to the other and, therefore,
a dam must be built to keep this from happening. In order to be consistent with notation to
be introduced shortly,
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
let M1 and M2 denote the sets of coordinates of points in two regional minima.
Then let the set of coordinates of points in the catchment basin associated with these two
minima at stage n - 1 of flooding be denoted by and respectively.
These are the two black regions shown in Fig. (a). Suppose that each of the
connected components in Fig.(a) is dilated by the structuring element shown in Fig. (c),
subject to two conditions:
(1) The dilation has to be constrained to q (this means that the center of the structuring
element can be located only at points in q during dilation), and
(2) The dilation cannot be performed on points that would cause the sets being dilated to
merge (become a single connected component).
Figure (d) shows that a first dilation pass (in light gray) expanded the boundary of each
original connected component. Note that condition (1) was satisfied by every point during
dilation, and condition (2) did not apply to any point during the dilation process; thus the
boundary of each region was expanded uniformly.
Let the union of these two sets be denoted by C[n - 1], There are two connected
components in Fig. (a) and only one connected component in Fig. (b),This connected
component encompasses the earlier two components, shown dashed. The fact that two
connected components have become a single component indicates that water between the
two catchment basins has merged at flooding step n. Let this connected component be
denoted q. Note that the two components from step n - 1 can be extracted from q by
performing the simple AND operation .We note also that all points belonging to an
individual catchment basin form a single connected component.
In the second dilation (shown in medium gray), several points failed condition (1)
while meeting condition (2), resulting in the broken perimeter shown in the figure. It also is
evident that the only points in q that satisfy the two conditions under consideration describe
the one-pixel-thick connected path shown crossed-hatched in Fig. (d). This path constitutes
the desired separating dam at stage n of flooding. Construction of the dam at this level of
flooding is completed by setting all the points in the path just determined to a value greater
than the maximum gray-level value of the image. The height of all dams is generally set at
1 plus the maximum allowed value in the image. This will prevent water from crossing over
the part of the completed dam as the level of flooding is increased. It is important to note
that dams built by this procedure, which are the desired segmentation boundaries, are
connected components. In other words, this method eliminates the problems of broken
segmentation lines,
Although the procedure just described is based on a simple example, the method
used for more complex situations is exactly the same, including the use of the 3 X 3
symmetric structuring elements shown in Fig. (c).
1. At each step of the algorithm, the binary image in obtained in the following manner
1. Initially, the set of pixels with minimum gray level are 1, others 0.
2. In each subsequent step, we flood the 3D topography from below and the
pixels covered by the rising water are 1s and others 0s. M1, M2:
– Sets of coordinates of points in the two regional minima
2.Cn-1(M1), Cn-1(M2)
– Sets of coordinates of points in the catchment basins associated with M1 M2
at stage n-1 of flooding (catchment basins up to the flooding level)
3. C [n-1]
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
2.
q C[n-1] contains one connected component of C[n-1] (q2)
• q is incorporated into C[n-1] to form C[n]
3.
q C[n-1] contains more than one connected components of C[n-1] (q3)
• A ridge separating two or more catchment basins has been
encountered
• A dam has to be built within q to prevent overflow between the
catchment basins
4.
Repeat the procedure until n=max+1
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
5.
Algorithm keeps on increasing the level of flooding, and during the process Cn(Mi) and
T[n] either increase or remain constant.
Algorithm initializes C[min +1] = T[min+1], and then proceeds recursively assuming
that at step n C[n-1] has been constructed.
6.
Let Q be set of connected components in T[n].For each connected component q Є Q[n],
there are three possibilities:
7.
Condition (a) occurs when a new minima is encountered, in this case q is added to set
C[n-1] to form C[n].
8.
Condition (b) occurs when q lies within a catchment basin of some regional minima, in
that case
9.
Condition (c) occurs when ridge between two catchment basins is hit and further
flooding will cause the waters from two basins will merge, so a dam must be built within
q.
Consider the image and its gradient, shown in Figs. (a) and (b), respectively. Application of
the watershed algorithm just described yielded the watershed lines (white paths} of the
gradient image shown in Fig. (c).These segmentation boundaries are shown superimposed
on the original image in Fig. (d). As noted at the beginning of this section, (the
segmentation boundaries have the important property of being connected paths).
A morphological region growing approach.
Seed points:
– local minima points
Growing method:
– Dilation
Predicates
– Similar gradient values
Sub-region boundary
– Dam building
To avoid over-segmentation
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
– Use markers
Disadv: Concept is difficult to understand
8. Explain Morphological Image Processing in details
Morphological Image Processing is an important tool in the Digital Image processing, since
that science can rigorously quantify many aspects of the geometrical structure of the way
that agrees with the human intuition and perception.
Morphologic image processing technology is based on geometry. It emphasizes
on studying geometry structure of image. We can find relationship between each part of
image. When processing image with morphological theory. Accordingly we can
comprehend the structural character of image in the morphological approach an image is
analyzed in terms of some predetermined geometric shape known as structuring element.
Morphological processing is capable of removing noise and clutter as well as the
ability to edit an image based on the size and shape of the objects of interest. Morphological
Image Processing is used in the place of a Linear Image Processing, because it sometimes
distort the underlying geometric form of an image, but in Morphological image Processing,
the information of the image is not lost.
In the Morphological Image Processing the original image can be reconstructed
by using Dilation, Erosion, Opening and Closing operations for a finite no of times. The
major objective of this paper is to reconstruct the class of such finite length Morphological
Image Processing tool in a suitable mathematical structure using Java language.
Binary Image:-
The image data of Binary Image is Black and White. Each pixel is either ‘0’ or ‘1’.
A digital Image is called Binary Image if the grey levels range from 0 and 1.
Ex: A Binary Image shown below is,
1 1 1 1 0
1 1 1 1 0
0 1 1 1 0
0 1 1 1 0
(A 4* 5 Binary Image)
DILATION:-
Dilation
- grow image regions
Dilation causes objects to dilate or grow in size. The amount and the way that they
grow depends upon the choice of the structuring element [3]. Dilation makes an object
larger by adding pixels around its edges.
The Dilation of an Image ‘A’ by a structuring element ‘B’ is written as AB.
To compute the Dilation, we
position ‘B’ such that its origin is at pixel co-ordinates (x , y) and apply the rule.
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
Repeat for all pixel co-ordinates. Dilation creates new image showing all the
location of a structuring
element origin at which that structuring element HITS the Input Image. In this it adds a
layer of pixel to an object, there by enlarging it. Pixels are added to both the inner and
outer boundaries of regions, so Dilation will shrink the holes enclosed by a single region
and make the gaps between different regions smaller. Dilation will also tend to fill in any
small intrusions into a region’s boundaries.
The results of Dilation are influenced not just by the size of the structuring
element but by its shape also.
Dilation is a Morphological operation; it can be performed on both Binary and Grey Tone
Images. It helps in extracting the outer boundaries of the given images.
For Binary Image:-
Dilation operation is defined as follows,
D (A , B) = A B
Where,
A is the image
B is the structuring element of the order 3 * 3.
Many structuring elements are requested for Dilating the entire image.
EROSION:-
Erosion
- shrink image regions
Erosion causes objects to shrink. The amount of the way that they shrink depend
upon the choice of the structuring element. Erosion makes an object smaller by removing or
Eroding away the pixels on its edges [3].
The Erosion of an image ‘A’ by a structuring element ‘B’ is denoted as A Θ B.
To compute the Erosion, we position ‘B’ such that its origin is at image pixel co-ordinate
(x , y) and apply the rule.
1 if ‘B’ Fits ‘A’,
g(x , y) =
0 otherwise
Repeat for all x and y or pixel co-ordinates. Erosion creates new image that marks all the
locations of a Structuring elements origin at which that Structuring Element Fits the input
image. The Erosion operation seems to strip away a layer of pixels from an object,
shrinking it in the process. Pixels are eroded from both the inner and outer boundaries of
regions. So, Erosion will
enlarge the holes enclosed by a single region as well as making the gap between different
regions larger. Erosion will also tend to eliminate small extrusions on a regions boundaries.
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
The result of erosion depends on Structuring element size with larger Structuring
elements having a more pronounced effect & the result of Erosion with a large Structuring
element is similar to the result obtained by iterated Erosion using a smaller structuring
element of the same shape.
Erosion is the Morphological operation, it can be performed on Binary and
Grey images. It helps in extracting the inner boundaries of a given image.
For Binary Images:-
Erosion operation is defined as follows,
E (A, B) = A Θ B
Where, A is the image
B is the structuring element of the order 3 * 3.
Many structuring elements are required for eroding the entire image.
OPENING:-
Opening
- structured removal of image region boundary pixels
It is a powerful operator, obtained by combining Erosion and Dilation. “Opening
separates the Objects”. As we know, Dilation expands an image and Erosion shrinks it [3].
Opening generally smoothes the contour of an image, breaks narrow Isthmuses and
eliminates thin Protrusions [1].
The Opening of an image ‘A’ by a structuring element ‘B’ is denoted as A ○ B and is
defined as an Erosion followed by a Dilation, and is
written as [3],
A ○ B = (A Θ B) B
Opening operation is obtained by doing Dilation on Eroded Image. It is to
smoothen the curves of the image. Opening spaces objects that are too close together,
detaches objects that are touching and should not be, and enlarges holes inside objects.
Opening involves one or more Erosions followed by one Dilation.
CLOSING:-
Closing
- structured filling in of image region boundary pixels
It is a powerful operator, obtained by combining Erosion and Dilation. “Closing,
join the Objects” [3]. Closing also tends to smooth sections of contours but, as opposed to
Opening, it generally fuses narrow breaks and long thin Gulf’s, eliminates small holes
and fills gaps in the contour [1].
The Closing of an image ‘A’ by a structuring element ‘B’ is denoted as A● B
and defined as a Dilation followed by an Erosion; and is written as [3],
A● B = (A B) Θ B. Closing is obtained by doing Erosion on Dilated image.
Closing joins broken objects and fills in unwanted holes in objects.
Closing involves one or more Dilations followed by one Erosion.
FUTURE SCOPE:-
The Morphological Image Processing can be further applied to a wide spectrum
of problems including:
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
RD 1 1
CR
Where CR, commonly called the compression ratio, is
CR n
n1
2
7. What are the three basic data redundancies can be identified and exploited?
The three basic data redundancies that can be identified and exploited are,
i) Coding redundancy
ii) Interpixel redundancy
iii) Phychovisual redundancy
8. Define is coding redundancy?
If the gray level of an image is coded in a way that uses more code words than
necessary to represent each gray level, then the resulting image is said to contain coding
redundancy.
9. Define interpixel redundancy?
The value of any given pixel can be predicted from the values of its neighbors. The
information carried by is small. Therefore the visual contribution of a single pixel to an
image is redundant. Otherwise called as spatial redundant geometric redundant or interpixel
redundant.
Eg: Run length coding
10. Define psycho visual redundancy?
In normal visual processing certain information has less importance than other
information. So this information is said to be psycho visual redundant.
11. Define encoder
Source encoder is responsible for removing the coding and interpixel
redundancy and psycho visual redundancy.
Encoder has 2 components they are 1.source encoder2.channel encoder
12. Define source encoder
Source encoder performs three operations
1) Mapper -this transforms the input data into non-visual format. It reduces the
interpixel redundancy.
2) Quantizer - It reduces the psycho visual redundancy of the input images .This
step is omitted if the system is error free.
3) Symbol encoder- This reduces the coding redundancy .This is the final stage of
encoding process.
13. Define channel encoder
The channel encoder reduces the impact of the channel noise by inserting redundant
bits into the source encoded data.
Eg: Hamming code
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
14. What are the types of decoder?
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
Saveetha Engineering 13
EC8093- Dr.M.Moorthi,Professor-
Where C is the number of bits in the compressed file, and N (=XY) is the number of pixels
in the image. If the bit rate is very low, compression ratio might be a more practical
measure:
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
Input Construct
Forward Symbol Compressed
Image MXn sub
transform Quantiz encoder Image
image
Compressed
Symbol Inverse Merge Decompressed
Image decoder transform NXN image Image
Saveetha Engineering 14
EC8093- Dr.M.Moorthi,Professor-
PART B
1. Explain compression types
Now consider an encoder and a decoder as shown in Fig.. When the encoder receives the
original image file, the image file will be converted into a series of binary data, which is called
the bit-stream. The decoder then receives the encoded bit-stream and decodes it to form the
decoded image. If the total data quantity of the bit-stream is less than the total data quantity of
the original image, then this is called image compression.
1. Lossless compression can recover the exact original data after compression. It is used
mainly for compressing database records, spreadsheets or word processing files.
There is no information loss, and the image can be reconstructed exactly the same as the
original
Applications: Medical imagery
Input
Source encoder Symbol encoder
Image
Compressed
Image
Compressed
Symbol Source Decompressed Image
Image decoder decoder
1. Source encoder is responsible for removing the coding and interpixel redundancy
and psycho visual redundancy.
2. Symbol encoder- This reduces the coding redundancy .This is the final stage
of encoding process.
Source decoder- has two components
a) Symbol decoder- This performs inverse operation of symbol encoder.
b) Inverse mapping- This performs inverse operation of mapper.
Lossless Compression technique
Varaible length coding (Huffmann,arithmetic), LZW coding,Bit Plane coding,Lossless Predictive
coding
14
EC8093- Dr.M.Moorthi,Professor-
Input
Mapper Quantizer Symbol encoder Compressed
Image Image
Compressed
Symbol Inverse Decompressed
Image decoder Dequantizer transform Image
2.Explain in detail the Huffman coding procedure with an example. (May / June 2009),
(May / June 2006) (Variable length coding)
Huffman coding is based on the frequency of occurrence of a data item (pixel in images). The
principle is to use a lower number of bits to encode the data that occurs more frequently. Codes
are stored in a Code Book which may be constructed for each image or a set of images. In all
cases the code book plus encoded data must be transmitted to enable decoding.
14
EC8093- Dr.M.Moorthi,Professor-
A bottom-up approach
1. Initialization: Put all nodes in an OPEN list, keep it sorted at all times (e.g., ABCDE).
2. Repeat until the OPEN list has only one node left:
(a) From OPEN pick two nodes having the lowest frequencies/probabilities, create a parent node
of them.
(b) Assign the sum of the children's frequencies/probabilities to the parent node and insert it into
OPEN.
(c) Assign code 0, 1 to the two branches of the tree, and delete the children from OPEN.
A 15 1.38 0 15
B 7 2.48 100 21
C 6 2.70 101 18
D 6 2.70 110 18
E 5 2.96 111 15
TOTAL (# of bits): 87
a) Huffmann coding:
i) Order the given symbols in the decreasing probability
ii) Combine the bottom two or the least probability symbols into a single symbol that
replaces in the next source reduction
iii) Repeat the above two steps until the source reduction is left with two symbols per
probability
iv) Assign ‘0’ to the first symbol and ‘1’ to the second symbol
14
EC8093- Dr.M.Moorthi,Professor-
v) Code each reduced source starting with the smallest source and working back to
the original source.
1. Average Information per Symbol or Entropy:
n n
H (S ) pi I (si ) pi log2 Bits/symbol
pi
i1 i1
N
H (s)
3. Efficiency is given L
as
avg
Pi = probability of occurrence of the ith symbol,
Li = length of the ith code word,
N = number of symbols to be encoded
Problem:A DMS x has five equally likely symbols P ( x1) = p(x2) = P(X3) =p(x4)= P(x5) =0.2
construct the Huffman code cu calculate the efficiency.
14
EC8093- Dr.M.Moorthi,Professor-
X3 0.2m3 0.2m3 m1
0
X4 0.2m4
1
X5 0.2m5
0.6 0(m1+m4+m5)
0.4 1(m2+m3)
Xi code Length
x1 00 2
x2 10 2
x3 11 2
x4 000 3
x5 001 3
L = Σ(5,i=1) p(xi) ni
=(0.2) (2+2+2+3+3)
L = 2.4
Therefore %q = H(x)/ Llog2(M) = 2.32/2.4log2(2) = 96.7%
Properties of Huffman coding:
14
EC8093- Dr.M.Moorthi,Professor-
b) The Shannon-Fano
Algorithm Algorithm:
1. List the source symbols in order of decreasing probability.
2. partition the set in to two sets that are close to equiprobables as possible and assign O to the
upper set, 1 to the loser set.
3. continue this process, each time partitioning the sets with as nearly equal probabilities as
possible until further portioning is not possible.
Example A DMS x has four symbols x1,x2,x3 cu x4 with P(x1) = 1/2 , P(x2) = ¼ and P(x3) =
P(x4) = 1/8 construct the Shannon-fano code for x./ show that his code has the optimum
property that ni=I(xi) cu the code efficiency is 100 percent.
Xi p(xi) code
X1 0.5 0 1
X2 0.25 1 0 2
X3 0.125 11 0 3
X4 0.125 11 1 3
= p(x1) log(2) p(x1) + p(x2) log(2) p(x2) + p(x3) log(2) p(x3) + p(x4) log(2) p(x4)
14
EC8093- Dr.M.Moorthi,Professor-
L = Σ(4,i=1) p(xi) ni
L = 1.75
Therefore q = 100%
This is a basic information theoretic algorithm. A simple example will be used to illustrate the
algorithm:
Symbol A B C D E
Count 15 7 6 6 5
A top-down approach
2. Recursively divide into two parts, each with approx. same number of counts.
15
EC8093- Dr.M.Moorthi,Professor-
Encoding Procedure:
15
EC8093- Dr.M.Moorthi,Professor-
1. Construct code words using arithmetic coding for the sequence s0,s1,s2,s3 with probabilities
{0.1,0.3,0.4,0.2}
Answer:
15
EC8093- Dr.M.Moorthi,Professor-
• Form n binary matrices (called bitplanes), where the i-th matrix consists of the i-th bits of
the pixels of I.
Example: Let I be the following 2x2 image where the pixels are 3 bits long
101 110
111 011
Example:
15
EC8093- Dr.M.Moorthi,Professor-
0 000 000
1 001 001
2 011 010
3 010 011
4 110 100
5 111 101
6 101 110
7 100 111
Decoding a gray coded image:
The MSB is retained as such,i.e.,
ai g i ai 1 0im2
am1 gm1
15
EC8093- Dr.M.Moorthi,Professor-
Input Construct
Forward Symbol Compressed
Image MXn sub
transform Quantiz encoder Image
image
Compressed
Symbol Inverse Merge Decompressed Image
Image decoder transform NXN image
Explanation:
1. The NxN output image is subdivided into sub images of size nxn, which are then transformed
to generate (N/n)2 sub image transform arrays each of size n x m.
2. Transformation is done to decorrelate the pixels of each sub image or to pack as much
information as possible into the smallest number of transform coefficients.
The following transforms are used in the encoder block diagram
a. KLT
b. DCT
c. WHT
d. DFT
Mostly used is DCT or DFT
3. Quantization is done to eliminate the coefficients which carry least information. These omitted
coefficients will have small impacts on the quality of the reconstructed sub images.
4. Finally, the quantized coefficients are encoded.
5. The decoder is used to extract the code words, which are applied to the dequantizer, which is
used to reconstruct the set of quantized transform coefficients.
6. Then the set of transform coefficients that represents the image is produced from the quantized
coefficients.
7. These coefficients are inverse transformed and give the lossy version of the image. Here the
compression ratio is low.
STEPS:
i. Subimage size selection:
Normally 4x4, 8,x,8 and 16x16 image sizes are selected
ii. Transform Selection:
Consider an image f(x,y) of size N x N, then the transformed image is given as,
15
EC8093- Dr.M.Moorthi,Professor-
c. In case of K L transform:
T
v * u
Where, u = input image
Φk =Kth column of Eigen vectors
V = transformed image
d.In case of Discrete Cosine Transform (DCT)
15
EC8093- Dr.M.Moorthi,Professor-
Let Nt be the number of transmitted samples and the zonal mask is defined as an array which
takes the unity value in the zone of largest Nt variances of the transformed samples.
i.e.,
1, (k, l)I
m(k, l) t 0,
Here the pixel with maximum variance is replaced by 1 and a 0 in all other locations. Thus the
shape of the mask depends on where the maximum variances pixels are located. It can be of any
shape.
Threshold coding
The process in which the retained coefficients are selected on the basis of maximum magnitude
is called threshold coding.
1.
Here we encode the Nt coefficients of largest amplitudes.
2.
The amplitude of the pixels are taken and compared with the threshold value
3.If it is greater than the threshold η, replace it by 1 else by 0.
Threshold mask mη is given by,
m(k, l) 1, (k, l) I t
'
where, I '
0,
elsewhere
t (k, l); V (k, l)
4.Then the samples retained are quantized by a suitable uniform quantizer followed by an
entropy coder.
There are three basic ways to threshold a transformed sub image.
a.A single global threshold can be applied to all sub images – here the level of compression
differs from image to image depending on the number of coefficients.
b.A different threshold can be used for each sub image – The same number of coefficients is
discarded for each sub image.
The threshold can be varied as a function of the location of each coefficients within the sub
image.
Here,
T (u, v)
T (u, v) round
Z (u, v)
Z (0, 0) Z 0,1 ... Z 0, n 1
Z (1, 0) ... ... ...
Z
.... ... ... ...
Z (n 1, 0).........................Z (n 1, n 1)
Where, T (u, v) is a threshold and quantized approximation of T(u,v) and Z(u,v) is an
element of the transform normalization array.
The denormalized array is given as,
T ' (u, v) T (u, v)z(u, v)
The decompressed sub image is given as,
1
15
EC8093- Dr.M.Moorthi,Professor-
T ' (u, v)
15
EC8093- Dr.M.Moorthi,Professor-
7.Describe on the Wavelet coding of images. (May / June 2009) or Explain the lossy
compression wavelet coding.
ENCODER
Input
er Construct
Wavelet Symbol Compressed
MXn sub
Image transform Quantiz encoder Image
image
DECODER
Compressed
Symbol Merge Decompressed Image
Image decoder IWT NXN image
Explanation:
1. Input image is applied to wavelet transform, which is used to convert the original input image
into horizontal, vertical and diagonal decomposition coefficients with zero mean and laplacian
like distribution.
2. These wavelets pack most of the information into a small number of coefficients and the
remaining coefficients can be quantized or truncated to zero so that the coding redundancy is
minimized.
3. Runlength, huffmann, arithmetic, bit plane coding are used for symbol encoding
4.Decoder does the inverse operation of encoding
The difference between wavelet and transform coding is that the sub image processing is
not needed here.Wavelet transforms are computationally efficient and inherently local.
Wavelet selection:
The four Discrete waveler transforms are,
1.Haar wavelets – simplest
2.Daubechies wavelets – most popular
3.Symlet wavelet
4.Bio orthogonal wavelet
15
EC8093- Dr.M.Moorthi,Professor-
Wavelets
Wavelets are mathematical functions that cut up data into different frequency components,and
then study each component with a resolution matched to its scale.
These basis functions or baby wavelets are obtained from a single prototype wavelet called the
mother wavelet, by dilations or contractions (scaling) and translations (shifts).
wavelet Transform
Wavelet Transform is a type of signal representation that can give the frequency content of the
signal at a particular instant of time.
1.Wavelet transform decomposes a signal into a set of basis functions.
2.These basis functions are called wavelets
3. Wavelets are obtained from a single prototype wavelet y(t) called mother wavelet by dilations
and shifting:
(t)
1 tb
a,b
( )
a a
where a is the scaling parameter and b is the shifting parameter
The 1-D wavelet transform is given by :
Wavelet analysis has advantages over traditional Fourier methods in analyzing physical
situations where the signal contains discontinuities and sharp spikes
Types of wavelettransform
1. The discrete wavelet transform
2.The continuous wavelet transform
Input signal is filtered and separated into low and high frequency components
16
EC8093- Dr.M.Moorthi,Professor-
LL: Horizontal Lowpass & Vertical Lowpass , LH: Horizontal Lowpass & Vertical Highpass
HL: Horizontal Highpass & Vertical Lowpass , HH: Horizontal Highpass & Vertical Highpass
Quantizer Design:
The largest factor effecting wavelet coding compression and reconstruction error is coefficient
quantization.
The effectiveness of the quantization can be improved by ,(i) introducing an enlarged
quantization interval around zero, called a dead zone and (ii) adapting the size of the
quantization interval from scale to scale.
8. Write notes on JPEG standard with neat diagram. (May / June 2009), (May / June
2006) 1.JPEG:It defines three different coding systems:
1. A lossy baseline coding system, adequate for most compression applications
2. An extended coding system for greater compression, higher precision or progressive
reconstruction applications
3. A lossless independent coding system for reversible compression
Details of JPEG compression Algorithm
1) Level shift the original image
2) Divide the input image in to 8x8 blocks
3) Compute DCT for each block (matrix) (8x8)
4) Sort elements of the 8x8 matrix
5) Process one block at a time to get the output vector with trailing zeros
Truncated
16
EC8093- Dr.M.Moorthi,Professor-
16
EC8093- Dr.M.Moorthi,Professor-
2. JPEG 2000
TABLE SPECIFICATION
Lifting Parameter:
16
EC8093- Dr.M.Moorthi,Professor-
Scaling parameter:
K = 1.230174105
i. After the completion of scaling and lifting operations, the even indexed value Y(2n)
represents the FWT LP filtered output and odd indexed values of Y(2n+1) correspond to
the FWT HP filtered output.
j. The sub bands of coefficients are quantized and collected into rectangular arrays of “code
blocks..Thus after transformation, all coefficients are quantized.
k. Quantization is the process by which the coefficients are reduced in precision. Each of
the transform coefficients ab(u,v) of the sub band b is quantized to the value qb(u,v)
according to the formula,
Rb b [1 b ]
Where, Δb = quantization step 2size
211
Rb = nominal dynamic range of subband b
εb , μb = number of bits allocated to the exponent and mantissa of the subband’s
coefficients.
l. Encode the quantized coefficients using Huffmann or arithmetic coding.
TABLE SPECIFICATION
c. The dequantized coefficients are then inverse transformed by column and by row using
inverse forward wavelet transform filter bank or using the following lifting based
operations
16
EC8093- Dr.M.Moorthi,Professor-
16
EC8093- Dr.M.Moorthi,Professor-
Forward Forward
prediction prediction
I BBBPBBBPBBBI
Bidirectional
prediction
16
EC8093- Dr.M.Moorthi,Professor-
6. Also the encoder is designed to generate a bit stream that matches the capacity of the intended
video channel.This is done by using a rate controller which adjusts the quantization parameters
as a function of the occupancy of the output buffer.
7. As the buffer becomes fuller the quantization is made coarser, so that fewer bits stream into
buffer.
Rate
controller
Difference
Block Variable
+ DCT Quantizer Buffer
- Encoded
length coding
Block
Inverse
Quantizer
Inverse
DCT
+ Encoded
Variable Motion
length
coding Vector
Motion estimator
and
compensator W/
frame delay
I- frames provide the highest degree of random access, ease of editing, and greatest
resistance to the propagation of transmission error. As a result, all standards require their
periodic insertion into the compressed code stream.
16
EC8093- Dr.M.Moorthi,Professor-
3. Bidirectional frame (B- frame). A B- frame is the compressed difference between the
current frame and a prediction of it based on the previous I- or P- frame and next- P-
frame. Accordingly, the decoder must have access to both past and future reference
frames. The encoded frames are therefore reordered before transmission; the decoder
reconstructs and displays them in the proper sequence.
10. Explain the Basics of Vector Quantization in detail.
Definition:
Mapping of accurate data’s into inaccurate data’s is known as Quantization. If this procedure is
extended for vector data , then the procedure is called as Vector Quantization.
Vector Quantization is a process which can map n-dimensional vectors in the vector space R into
a finite set of vectors Y in the vector space R.
Y {y i , i 1,2,3,. ., n}
Set of all code words is called as code book
1. Here, original image is decomposed into n dimensional image vectors, vectors can be block
of pixel values or a 3-dimensional vector formed from the RGB color components.
16
EC8093- Dr.M.Moorthi,Professor-
VQ transforms the vectors of data into indexes that represents clusters of vectors
Voronoi region is the nearest neighbor region associated with each codeword Yi.
After the minimum distortion code vector has been found, the index k is transmitted using
log2Nc bits.
Vector Quantization Receiver:
1. VQ decoder uses a duplicate codebook and a table lookup to reproduce the image
2.Compression is obtained by using a codebook with relatively few code vectors compared to
the original image vectors
16
EC8093- Dr.M.Moorthi,Professor-
Linde-Buzo-Gray (LBG) Algorithm is one of the suboptimal code book design procedure
Start with:
1.a set of training vectors
2. initial codebook
3. distortion measure d
4. fractional distortion change threshold
Initialize:
1. iteration counter l to 1
2. average distortion over all training vectors, D(0) to a very large number
Based on the minimum euclidean distance, the input vector is attached to the code word and
average distortion is computed.Then revise the code words by computing the average of each
cluster.
Drawbacks:
The size of the codebook increases for higher compression ratio’s and
the search time required to find the minimal distortion code also increases.
Code book Design:
Consider an 8 bit image - Compressed to 1 bit per pixel,If n = 4 (2x2 blocks), codebook should
contain only 16 vectors,16 vectors will represent 2564 possible image vectors.
If block size is increased to 4x4, the codebook is 216 codevectors,The total number of vectors is
25616
Resulting mean-square quantization error becomes smaller as the block size increases.
Thus the Number of codebook vectors increase
In sub band coding, the spectrum of the input is decomposed into a set of bandlimitted
components, which is called sub bands. Ideally, the sub bands can be assembled back to
reconstruct the original spectrum without any error. Fig. shows the block diagram of two-band
filter bank and the decomposed spectrum. At first, the input signal will be filtered into low pass
and high pass components through analysis filters. After filtering, the data amount of the low
pass and high pass components will become twice that of the original signal; therefore, the low
pass and high pass components must be down sampled to reduce the data quantity. At the
receiver, the received data must be up sampled to approximate the original signal. Finally, the up
16
EC8093- Dr.M.Moorthi,Professor-
sampled signal passes the synthesis filters and is added to form the reconstructed approximation
signal.
After sub band coding, the amount of data does not reduce in reality. However, the human
perception system has different sensitivity to different frequency band. For example, the human
eyes are less sensitive to high frequency-band color components, while the human ears is less
sensitive to the low-frequency band less than 0.01 Hz and high-frequency band larger than 20
KHz. We can take advantage of such characteristics to reduce the amount of data. Once the less
sensitive components are reduced, we can achieve the objective of data compression.
Fig Two-band filter bank for one-dimension sub band coding and decoding
Now back to the discussion on the DWT. In two dimensional wavelet transform, a two-
dimensional scaling function, (x, y) , and three two-dimensional wavelet function H (x, y) ,
V (x, y) and D (x, y) , are required. Each is the product of a one-dimensional scaling function
(x) and corresponding wavelet function (x).
( x, y) ( x)( y) H (x, y) (x)( y)
V (x, y) ( y)(x) D (x, y) (x)( y)
where H measures variations along columns (like horizontal edges), V responds to variations
along rows (like vertical edges), and D corresponds to variations along diagonals.
Similar to the one-dimensional discrete wavelet transform, the two-dimensional DWT can be
implemented using digital filters and samplers. With separable two-dimensional scaling and
wavelet functions, we simply take the one-dimensional DWT of the rows of f (x, y), followed by
the one-dimensional DWT of the resulting columns. Fig. shows the block diagram of two-
dimensional DWT
.
16
EC8093- Dr.M.Moorthi,Professor-
As in the one-dimensional case, image f (x, y) is used as the first scale input, and output four
quarter-size sub-images W , WH , WV , and WD as shown in the middle of Fig. The
approximation output W H ( j, m, of the filter banks in Fig. can be tied to other input analysis
n)
filter bank to obtain more sub images, producing the two-scale decomposition as shown in the
left of Fig. Fig. shows the synthesis filter bank that reverses the process described above.
17
EC8093- Dr.M.Moorthi,Professor-
12. Explain the following terms 1.Bounary descriptors 2. Regional descriptors 3.Texture
descriptors
I. Simple Descriptors
The following are few measures used as simple descriptors in boundary descriptors,
Length of a boundary
Diameter of a boundary
Major and minor axis
Eccentricity and
Curvature
Length of a boundary is defined as the number of pixels along a
boundary.Eg.for a chain coded curve with unit spacing in both directions the
number of vertical and horizontal components plus √2 times the number of
diagonal components gives its exact length
The diameter of a boundary B is defined as
Diam(B)=max[D(pi,pj)]
D-distance measure
pi,pj-points on the boundary
The line segment connecting the two extreme points that comprise the diameter is
called the major axis of the boundary
The minor axis of a boundary is defined as the line perpendicular to the major
axis
Eccentricity of the boundary is defined as the ratio of the major to the minor
axis
Curvature is the rate of change of slope.
II. Shape number
Shape number is defined as the first difference of smallest magnitude.
The order n of a shape number is the number of digits in its representation.
17
EC8093- Dr.M.Moorthi,Professor-
K 1
a(u) 1
s(k )e j2uk / K ,u 0,1,2,....K 1
K
k 0
These complex coefficients a(u) are called the Fourier descriptors of the boundary.
The Inverse Fourier transform of these coefficients restores s(k) given as,
17
EC8093- Dr.M.Moorthi,Professor-
K 1
s(k )
a(u)e , k 0,1,2,. . .K 1
j2uk / K
u 0
Instead of using all the Fourier coefficients, only the first P coefficients are used.
(i.e.,) a(u) = 0, for u > P-1
Hence the resulting approximation is given as,
P 1
Thus, when P is too low, then the reconstructed boundary shape is deviated from the
original..
Advantage: It reduces a 2D to a 1D problem.
The following shows the reconstruction of the object using Fourier descriptors. As the
value of P increases, the boundary reconstructed is same as the original.
17
EC8093- Dr.M.Moorthi,Professor-
It is used to describe the shape of the boundary segments, such as the mean,
and higher order moments
Consider the following boundary segment and its representation as a 1D function
Let v be a discrete random variable denoting the amplitude of g and let p(v i)
be the corresponding histogram, Where i = 0,1,2,…,A-1 and A = the number
of discrete amplitude increments in which we divide the amplitude scale, then
the nth moment
A 1
of v about the mean is ,
n (v ) (v
i0
i
m ) n p (v ), i 0,1,2,.....A 1
i
m v p(v )
i i
i0
Another approach is to normalize g(r) to unit area and consider it as a histogram,
i.e., g(r) is the probability of occurrence of value ri and the moments are given as,
K 1
(v )
n
(r m) i
n
g ( r ), i 0,1,2,.... K 1
i
i0
Where,
K 1
m r g(r )
i i
i0
K is the number of points on the boundary.
The advantage of moments over other techniques is that the implementation of
moments is straight forward and they also carry a ‘physical interpretation of
boundary shape’
Regional and simple descriptors
The various Regional Descriptors are as follows,
a) Simple Descriptors
b) Topological Descriptors
c) Texture
Which are explained as follows,
a) Simple Descriptors
The following are few measures used as simple descriptors in region descriptors
Area
17
EC8093- Dr.M.Moorthi,Professor-
Perimeter
Compactness
Mean and median of gray levels
Minimum and maximum of gray levels
Number of pixels with values above and below mean
The Area of a region is defined as the number of pixels in the region.
The Perimeter of a region is defined as the length of its boundary
Compactness of a region is defined as (perimeter)2/area. It is a dimensionless
quantity and is insensitive to uniform scale changes.
b) Topological Descriptors
Topological properties are used for global descriptions of regions in the image
plane.
Topology is defined as the study of properties of a figure that are unaffected
by any deformation, as long as there is no tearing or joining of the figure.
The following are the topological properties
Number f holes
Number of connected components
Euler number
Euler number is defined as E = C – H,
Where, C is connected components
H is the number of holes
The Euler number can be applied to straight line segments such as polygonal
networks
Polygonal networks can be described by the number of vertices V, the number of
edges A and the number of faces F as,
E=V–A+F
Consider the following region with two holes
17
EC8093- Dr.M.Moorthi,Professor-
c) Texture
Texture refers to repetition of basic texture elements known as texels or
texture primitives or texture elements
Texture is one of the regional descriptors.
It provides measures of properties such as smoothness, coarseness and regularity.
There are 3 approaches used to describe the texture of a
region. They are:
•Statistical
•Structural
•Spectral
i. statistical approach
Statistical approaches describe smooth, coarse, grainy characteristics of
texture. This is the simplest one compared to others. It describes texture
using statistical moments of the gray-level histogram of an image or region.
The following are few measures used as statistical approach in Regional
descriptors
Let z be a random variable denoting gray levels and let p(zi) be the
corresponding histogram,
Where I = 0,1,2,…,L-1 and L = the number of distinct gray levels, then
the nth moment of z about the mean is ,
17
EC8093- Dr.M.Moorthi,Professor-
L1
z and it defines the average gray level of each region, is given as,
L1
m z p(z )
i i
i0
The second moment is a measure of smoothness given as,
R 1 1
, gray level contrast
1 2 (z)
R = 0 for areas of constant intensity, and R = 1 for large values of σ2(z)
The normalized relative smoothness is given as,
R 1 1
1 2 (z)
(L 1) 2
The third moment is a measure of skewness of the histogram given as,
L1
p (zi
U i0
An average entropy measure is given as,
L1
17
EC8093- Dr.M.Moorthi,Professor-
This technique uses the structure and inter relationship among components
in a pattern.
The block diagram of structural approach is shown below,
The input is divided into simple subpatterns and given to the primitive
extraction block. Good primitives are used as basic elements to provide
compact description of patterns
Grammar construction block is used to generate a useful pattern
description language
If a represents a circle, then aaa..means circles to the right, so the rule aS
allows the generation of picture as given below,
……………
17
EC8093- Dr.M.Moorthi,Professor-
By varying above co-ordinates we can generate two 1D functions, S(r) and S(θ) , that
gives the spectral – energy description of texture for an entire image or region .
18
EC8093- Dr.M.Moorthi,Professor-
18
EC8093- Dr.M.Moorthi,Professor-
18
EC8093- Dr.M.Moorthi,Professor-
18
EC8093- Dr.M.Moorthi,Professor-
QUESTION BANK
18
EC8093- Dr.M.Moorthi,Professor-
14. What are the properties of unitary transform? (May / June 2009)
15. State 2D sampling theory.
16. What is the need for quantization?
17. State and explain the convolution property of Fourier transform?
18. Explain the terms aliasing and fold over frequencies?
19. What is circulant matrix? Give an example.
20. Explain the decorrelation property of KL transform. (May / June 2006)
21. Compare SVD and KL transform. (May / June 2009)
22. Write N x N discrete cosine transform. (May / June 2009)
23. Define the SVD transform of images.
24. What is cosine transform?
25. What is the special feature of KL transform?
26. Define unitary transform
27. Name the working principle of digital camera
28. Define dither
29. Name the color models
30. Differentiate brightness and contrast
31. Differentiate hue and saturation
32. Give the HSI model
Part-B
1. Explain about the element of digital image processing (May / June 2009) (May /
June 2007)
2. Explain the different elements of visual perception in detail. (May / June 2009)
3. Illustrate the concept of Brightness adaptation. (May / June 2007)
4. Explain luminance and brightness and contrast with respect to human vision?
5. Write short notes on Non uniform sampling and Quantization. (May / June 2007)
6. Discuss in detail the process of uniform sampling and Quantization. (May / June
2007)
7. State the properties of optimum mean square quantizer. (Nov / Dec 2005)
8. Describe the properties of 2D Fourier transform and its application in image
processing. (May/June 2009)
9. Discuss the effects of non uniform sampling and quantization. (May/June 2009)
10. State and prove 2D sampling theorem. (May/June 2009)
11. Derive the expression for the signal to noise ratio of Lloyd max quantizer.
(May/June 2009)
12. State the condition for a matrix to be (1) unitary (2) orthogonal. Give an example of
a matrix which is both orthogonal and unitary. (May / June 2006)
13. Explain how a matrix which is represented using singular value decomposition.
(May / June 2006)
14. Explain the principle of vector quantization of images. (May / June 2006)
15. Explain Weber’s Law and Lateral inhibition with reference to human visual system.
(May / June 2006)
16. Describe Lloyd max quantizer for image given the probability density function?
(May / June 2006)
17. Explain any four properties of DFT. (Nov / Dec 2005)
18
EC8093- Dr.M.Moorthi,Professor-
Part-B
2. Explain how image subtraction and image averaging is used to enhance the image.
(May / June 2009)
3. Explain the various sharpening filters in spatial domain. (May / June 2009)
4. Discuss in detail the significance of Homomorphic filtering in image enhancement.
(May / June 2009) (May / June 2007)
5. Write short notes on Point operations, Transform operations and Pseudo coloring
with respect to image enhancement. (May / June 2009) (May / June 2006)
6. Explain any two spatial operations. (May / June 2006)
18
EC8093- Dr.M.Moorthi,Professor-
7. Write short notes on Contrast stretching and Grey level slicing. (May / June 2009)
8. Explain in detail about Sharpening filter, Edge Detection, Smoothing filter and
Noise removal. (May / June 2009)
9. What is histogram of an image? How does it modify an image in equalization and
specification techniques? (May / June 2009)
10. Explain Enhancement using point operations. (Nov / Dec 2005) (May / June 2006)
11. Write short notes on Histogram equalization and Histogram modification. (Nov /
Dec 2005) (May / June 2009) (May / June 2007)
12. Write short notes on Directional Smoothing and color image enhancement. (Nov /
Dec 2005) (May / June 2006)
13. Explain how a high pass filtered image can be generated using a low pass filter.
(May / June 2006)
14. How is the digital negative of an image generated? Where is this used? (May / June
2006)
15. Discuss the role of non linear filters in image enhancement. (May / June 2007)
16. Explain any one method of pseudo color image processing. (May / June 2007)
17. Write shorts notes on Geometric mean, Harmonic mean, Contra harmonic mean filters
1. Explain in detail the constrained least squares restoration. (May / June 2009)
18
EC8093- Dr.M.Moorthi,Professor-
2. Write notes on inverse filtering as applied to image restoration. (May / June 2009)
3. Give the degradation model for continuous function. (May / June 2009)
4. With the aid of block diagram , describe the digital image restoration system and
explain the image observation models
5. Explain in detail the function of Wiener filter. (May / June 2009)
6. Write short notes on Inverse filter and Pseudo inverse filter. (Nov / Dec 2005)
7. Write short notes on Wiener filter characteristics and Geometric mean filter. (Nov /
Dec 2005)
8. Derive the Weiner filter equation. Under what conditions does it become the pseudo
inverse filter? (May / June 2006)
9. Explain the image degradation model and various approaches of image restoration
technique. (May / June 2009)
10. Explain Constrained and Unconstrained Restoration in detail. (May / June 2009)
11. Describe the principle of wiener filtering in image Restoration.
12. Explain the method of removal of blur caused by uniform linear motion?
13. Discuss the differences between Geometric transformations and spatial transformations
techniques?
UNIT IV- IMAGE SEGMENTATION
Part A
Part-B
18
EC8093- Dr.M.Moorthi,Professor-
Part-B
1. Explain in detail the Huffman coding procedure with an example. (May / June 2009),
(May / June 2006)
2. Describe arithmetic coding of images with an example. (May / June 2006)
3. Explain the Pixel coding with example. (May / June 2009)
4. Describe on the Wavelet coding of images. (May / June 2009)
5. Explain in detail the method of Zonal and Threshold coding. (May / June 2009) ,
(May / June 2006)
6. Briefly discuss the MPEG compression standard. (May / June 2009)
7. Explain image compression and how it is achieved through wavelet transform. (May
/ June 2009)
18
EC8093- Dr.M.Moorthi,Professor-
8. Write notes on JPEG standard (jpeg and jpeg 2000)with neat diagram. (May / June
2009), (May / June 2006)
9. Apply Huffman coding procedure to the following message ensemble and determine
average length of encoded message and coding efficiency. The symbols
(x1,x2,x3,x4,x5,x6,x7,x8) are emitted with probabilities of
(0.22,0.2,0.18,0.15,0.08,0.05,0.02). (May / June 2009)
10. Explain the run length coding scheme. (May / June 2006)
11. Explain the image compression standards
12. What are the building blocks of an image encoder? State their function?
13. Explain Transform coding. (May / June 2009), (May / June 2009)
14. Explain Vector Quantization with neat diagram
15. Compute the Huffman’s code and IB code following gray level distribution.
levels(r) W W2 W3 W4 W5 W6
19