1.1 Image Fusion:: Implement of Hybrid Image Fusion Technique For Feature Enhancement in Medical Diagnosis
1.1 Image Fusion:: Implement of Hybrid Image Fusion Technique For Feature Enhancement in Medical Diagnosis
1.1 Image Fusion:: Implement of Hybrid Image Fusion Technique For Feature Enhancement in Medical Diagnosis
CHAPTER 1
INTRODUCTION
1.2 Motivation:
Fusion leads to more accurate data [1] and increased utility and it can also
improve the quality and increase the application of these data. Combine higher spatial
information in one band with higher spectral information in another dataset to create
1) Medical imaging
2) Microscopic imaging
3) Remote sensing
4) Robotics
5) Battle field surveillance
6) Automated target recognition
7) Guidance and control of autonomous vehicle.
Our project is related to multi modal medical image fusion, Generally for a
physician to analyze the condition of a patient, in most of cases he needs to study
different images like MRI, CT, PET, SPECT etc. This is a time taking process. So, our
idea is to fuse all these images into a single image to provide better diagnosis.
1.3 Objective:
In the recent years, Multimodal image fusion algorithms and devices, has
evolved as a powerful tool in the clinical applications, of medical imagining
techniques. It has shown significant achievements in improving clinical accuracy of
diagnosis based on medical images. The main motivation is to produce most relevant
information from different sources into a single output, which plays a crucial role in
medical diagnosis.
Medical imaging has gained significant attention due to its predominant role in
health care. Some of the different types of imaging modalities used now-a-days are X-
ray, computed tomography (CT), magnetic resonance imaging (MRI), magnetic
resonance angiography (MRA), etc., These imaging techniques are used for extracting
clinical information, which are although complementary in nature most of the times,
some are unique depending on the specific imaging modality used.
For example,
CT: Is used to provide more accurate information about calcium deposit, air and
dense structures like bones with less distortion, acute bleeds and tumours. But it
cannot detect physiological changes.
MRI: Under strong magnetic field and radio-wave energy, information about Nervous
system, structural abnormalities of soft tissue, muscles can be better visualized.
Hence, we can understand none of these modalities are able to carry all
relevant information in a single image. So, that anatomical and functional medical
images are needed to be combined for a concise view. For this purpose, the
multimodal medical image fusion has been identified as a source with better potential.
It aims to integrate information from multiple modalities to obtain a more complete
and accurate description of the same object which facilitate in more precise diagnosis
and better treatment. Fused image provides higher accuracy and reliability by
removing redundant information.
The applications of image fusion are found in radiology, molecular and brain
imaging, oncology, diagnosis of cardiac diseases, neuro-radiology and ultrasound.
Multimodal medical image fusion helps in diagnosing diseases, and also cost effective
by minimising storage to a single fused image instead of multiple-source images.
We have so many fusion techniques to perform medical image fusion but till
now no one will provide better results. Fusion techniques are broadly classified two
group’s spatial and spectral domain. Spatial domain transform directly deals with
pixels of an image. It leads to spatial distortion in the fused image. It does not give
directional information and also leads to spectral distortion, while the arithmetic
combination will lose original details as a result of low contrast of the fused image. It
becomes a negative factor while we go for further processing, such as classification
problem, of the fused image.
Till now we have wavelet, curvelet, and contourlet transforms, the resultant
fused image from these individual transformations doesn’t yield good fused image. In
wavelet transform it provide multi resolution fused image, but it fails to capture
curved edge information. This can be overcome by curvelet but it has limited
CHAPTER 2
LITERATURE SURVEY
Based on the input data of the fusion process and also based on the purpose of
fusion, fusion can be classified into the following types:
Fusion of images coming from different sensors (CT, MRI, visible, infrared,
ultra violet etc) is called Multi Modal Image fusion. It is used to decrease the amount
of data, to emphasize band-specific information. In our Project, we are focusing on
this type of image fusion.
(a)Near focused image (b) Far focused image (c) Fused image
It is defined as fusion of images from the same modality and taken at the same
time but from different viewpoints. A non-blind, shift-invariant image processing
technique that fuses multi-view three-dimensional image data sets into a single, high
quality three-dimensional image is presented in Figure 2.3.
Figure 2.3: Detection results of the motion-based tracker of the first run of the subject
“Alba”, for all camera views
It is effective for
1) Pixel level
2) Feature level
3) Decision level
A). The general block diagram of feature based fusion is shown in Figure 2.6.
The object segmentation routine is used only to bootstrap the feature selection
process, and hence any method that provides even rough, incomplete object
segmentation can be employed at this stage. Feature fusion techniques are used to
increase the accuracy of the feature measurement. Data fusion techniques at the
feature level rely on feature attribute combination techniques such as Kalman
filtering.
Image registration is one of the pre-processing techniques which align data sets in
an image using feature base algorithm. Before performing fusion we have to set images
to pre-processing stage.
The system level considerations that are required to implement image fusion is
shown in Figure 2.8. It contains the following stages.
Image registration
Image pre-processing
Image post-processing
For example if the sizes of images vary, so before fusion, the images are
needed to be resized so both the images are of the same size. This is done by
interpolating the smaller size image by rows and columns duplication.
Post Processing stage depends on type of display, fusion system is being used and the
preference of human operator.
The general requirement of an image fusing process is to preserve all valid and
useful information from the source images, while at the same time it should not
introduce any distortion in resultant fused image. There are various methods that have
been developed to perform image fusion. These methods can be divided into two
types, spatial domain method and frequency domain method.
Spatial domain methods work by combining the pixel values of the two or
more images to be fused in a linear or nonlinear way. The simplest form is a Weighted
Averaging method. Here, the resultant image is obtained by averaging every
corresponding pixel in the input to give the fused image.
In Frequency domain methods, the input images are decomposed into Multi-
scale coefficients initially. Various fusion rules are used in the selection or
manipulation of these coefficients and synthesized via inverse transforms to form the
fused image. The fusion techniques are classified as given in Figure 2.9.
Spatial Domain Fusion Method directly deals with image pixels by manipulating the
pixel values to achieve desired results.
1) Averaging
2) Select maximum.
3) Weighted average method.
4) Intensity-hue-saturation (IHS) transforms.
5) Brovery.
6) Principal component analysis (PCA).
2) Select maximum: The greater the pixel values the more in focus the image. Thus
this algorithm chooses the in-focus regions from each input image by choosing the
greatest value for each pixel, resulting in highly focused output. The value of the pixel
P (i, j) of each image is taken and compared to each other. The greatest pixel value is
assigned to the corresponding pixel.
p (i , j )= { A ( i. j ) + B (i , j ) } /2 (2.2)
1) Resulting in highly focused image output obtained from the input image as
compared to average method.
1) Pixel level method is affected by blurring effect which directly affect on the
contrast of the image.
3) Weighted Average Method: In this method the resultant fused image is obtained
by taking the weighted average intensity of corresponding pixels from both the input
images.
m n
p(i , j) ∑ ∑ W A ( i, j )+ ( 1−w ) B(i , j) (2.3)
i=0 j=0
Methods based on Intensity, Hue and Saturation (IHS) transform are probably
the most popular approaches used for enhancing the spatial resolution of multi-sensor
images. The IHS method is capable of quickly merging the massive volume of data. It
can transform the colour space form Red (R), Green (G), Blue (B) to Hue (H),
Saturation (S) and Intensity (I) space.
The IHS colour transformation effectively separates spatial (I) and spectral (H,
S) information from a standard RGB image. It relates to the human colour perception
parameters. The mathematical context is expressed by Eq. 2.4. I relates to the
intensity, while ‘v1’ and ‘v2’ represent intermediate variables which are needed in the
transformation. H and S stand for Hue and Saturation.
[] [ ] []
I 1/√ 3 1/√ 3 1/√ 3 R
v1 = 1 /√ 6 1/√ 6 −2/√ 6 G
v2 1/√ 2 −1 /2 0 B
H= tan
−1
[ ]
v2
v1
2
v1
S= (¿ + v 22 ) (2.4)
√¿
There are two ways of applying the IHS technique in image fusion: direct and
substitutional. The first refers to the transformation of three image channels assigned
to I, H and S. The second transforms three channels of the data set representing RGB
into the IHS colour space which separates the colour aspects in its average brightness
(intensity). The schematic diagram of IHS is shown in Figure 2.10.
[] [
R
G
B
=
1 /√ 3 1/ √ 6 1/√ 2
1 /√ 3 1/ √ 6 −1/√ 2
1 /√ 3 −2/√ 6 0 ] [] I
v1
v2
(2.5)
The use of IHS technique in image fusion is manifold, but based on one
principle: the replacement of one of the three components (I, H or S) of one data set
with another image. Most commonly the intensity channel is substituted. Replacing
the intensity (sum of the bands) by a higher spatial resolution value and reversing the
IHS transformation leads to composite bands. These are linear combinations of the
original (re-sampled) multispectral bands and the higher resolution panchromatic
band.
A variation of the IHS fusion method applies a stretch to the hue saturation
components before they are combined and transformed back to RGB. This is called
colour contrast stretching. The IHS transformation can be performed either in one or
in two steps. The two step approach includes the possibility of contrast stretching the
individual I, H and S channels. It has the advantage of resulting in colour enhanced
fused imagery. A closely related colour system to IHS is the HSV: hue, saturation and
value.
1) Perform image registration (IR) to PAN and MS(Multi spectral), and resample
MS
2) Convert MS from RGB space into IHS space.
3) Match the histogram of PAN to the histogram of the I component.
4) Replace I component with PAN.
5) Convert the fused MS back to RGB space.
5) Brovery Transform:
The spatial domain provide high spatial resolution and easy to perform, but it
has image blurring problem and outputs are less informative. Spatial distortion
becomes a negative factor.
1) Increases the contrast in the low and high ends of an image histogram.
2) It is a simple method to merge the data from different sensors.
3) This method is simple and fast.
4) It provide superior visual and high resolution multispectral image.
5) Very useful for visual Interpretation.
PCA based fusion is very suitable for merging the MS and PAN images.
Compared to the IHS fusion, the PCA fusion has the advantage that it does not have
the three band limitation and can be applied to any number of bands at a time. As
shown in Figure 2.11.
The PCA is useful for image encoding, image data compression, image
enhancement, digital change detection, multi-temporal dimensionality and image
fusion. It is a statistical technique that transforms a multivariate data set of inter-
correlated variables into a data set of new un-correlated linear combinations of the
original variables. It generates a new set of axes which are orthogonal. The approach
for the computation of the principal components (PCs) comprises the calculation of:
An inverse PCA transforms the combined data back to the original image
space. The use of the correlation matrix implies a scaling of the axes so that the
features receive a unit variance. It prevents certain features from dominating the
image because of their large digital numbers. The signal-to-noise ratio (SNR) is
significantly improved applying the standardized PCA. Better results are obtained if
the statistics are derived from the whole study area rather than from a subset area.
Two types of PCA can be performed: selective or standard. The latter uses all
available bands of the input image and the selective PCA uses only a selection of
bands which are chosen based on a priori knowledge or application purposes.
The first version follows the idea of increasing the spatial resolution of a
multichannel image by introducing an image with a higher resolution. The channel
which will replace PC1 is stretched to the variance and average of PC1. The higher
resolution image replaces PC1 since it contains the information which is common to
all bands while the spectral information is unique for each band.
1) This method is very simple to use and the images fused by this method have
high spatial quality.
2) It prevents certain features from dominating the image because of their large
digital numbers.
Here the high frequency details are injected into up-sampled version of MS
images. The disadvantage of spatial domain approaches is that they produce spatial
distortion in the fused image. It does not give directional information and also leads to
spectral distortion, while the arithmetic combination will lose original details as a
result of low contrast of the fused image. It becomes a negative factor while we go for
further processing, such as classification problem, of the fused image.
1) Multi-scale Decomposition
2) Multi-scale Geometric Analysis
Example: If the sizes of images vary, so before fusion, the images are needed to be
resized so both the images are of the same size. This is done by interpolating the
smaller size image by rows and columns duplication. In spectral domain the original
image is first transferred into frequency domain. It means that the Fourier transform
of the image is computed first.
Then the image is transferred in to frequency domain. It means that the Fourier
transform of the image is computed first. After that the image is transformed into
frequency domain using one of our proposed techniques. For the obtained image we
apply appropriate fusion rule. Now we perform inverse transformation to represent the
fused image in spatial domain.
format that is convenient to use, since the nodes in each level represent information
that is localized in both space and spatial frequency.
Decomposition
Formation of the initial image for decomposition.
Recomposition
Decomposition phase basically consists of the following steps. These steps are
performed l number of times, l being the number of levels to which the fusion will be
performed.
Low Pass filtering: The different pyramidal methods have a pre defined filter
with which are the input images convolved/filtered with.
Formation of the pyramid for the level from the filtered/convolved input
images using Burt’s method or Lis method.
The input images are decimated to half their size, which would act as the input
image matrices for the next level of decomposition.
Merging the input images is performed after the decomposition process. This
resultant image matrix would act as the initial input to the recomposition process.
operations in morphological filtering. Among many, the major operators used for
fusion are averaging, morphology towers, K-L transforms, and morphology pyramids.
The pyramid is produced by low-pass filtering the image and then sampling to
generate the next lower resolution level of the hierarchy. The basis for a
morphological pyramid requires a morphological sampling theorem. The overall
fusion strategy is shown in Figure 2.15. According to this strategy a morphological
pyramid is first produced for each of the input images. Then a morphological
difference pyramid is constructed, for each of the above pyramids. This is achieved by
taking the differences between the morphological images residing at successive levels
in the original pyramid. An intermediate pyramid is constructed. Combining
information from the two difference pyramids at each level. Finally, reconstruction of
the intermediate pyramid, using appropriate morphological operations, produces the
required fused image.
∞
y [ n ] =( x∗g ) [ n ] = ∑ x [ k ] g [n−k ] (2.6)
k=−∞
However, since half the frequencies of the signal have now been removed, half
the samples can be discarded according to Nyquist’s rule. The filter outputs are then
sub-sampled by 2 (It should be noted that Mallat's and the common notation is the
opposite, g- high pass and h- low pass):
∞
y [ n ] ∨low= ∑ x [ k ] g [2 n−k ] (2.7)
k=−∞
∞
y [ n ] ∨high= ∑ x [ k ] h[2 n−k ] (2.8)
k=−∞
This decomposition has halved the time resolution since only half of each filter
output characterizes the signal. However, each output has half the frequency band of
the input so the frequency resolution has been doubled. The 2D multi resolution
wavelet decomposition shown in Figure 2.16.
y│low = (x*g)2
At each level in the above diagram the signal is decomposed into low and high
frequencies. Due to the decomposition process the input signal must be a multiple of
2n where n is the number of levels. General frame work for DWT is shown in Figure
2.18.
p
DWT
(i, j) = ∑ ∑ (DWT (i , j))2
(2.11)
Step 1: Implement DWT on both the input image to create wavelet lower
decomposition.
Step 3: Carry inverse discrete wavelet transform on fused decomposed level, which
means to reconstruct the image, while the image reconstructed is the fused image
1) Curvelet transform.
2) Contourlet transform.
samples. The Fourier samples are wrapped around the origin. Finally the image is
reconstructed by performing the inverse FFT transform.
2) Contourlet-based Fusion:
CHAPTER 3
The existing method which is hybrid of two methods that is the wavelet based
image fusion and the curvelet based image fusion (hybrid of wavelet and curvelet
fusion rules). Curvelet based image fusion efficiently deals with the curved shapes,
therefore its application in medical fields would result in better fusion results than
obtained using wavelet transform alone.
flow diagram of existed method with is the combination of wavelet and curvelet
transformation is shown in Figure 3.1.
The flow diagram shows procedure of combining image 1 and image 2 into
single fused wavelet coefficients. These bands obtained are then passed through
curvelet transform which segments it into various additive components each of which
is sub-band of the image. These bands are then passed through tiling operation which
divides the band to overlapping tiles.
A hybrid of wavelet and curvelet integrates various pixel level rules in a single
fused image. Pixel based rules operates on individual pixels in the image but ignores
some important details such as edges, boundaries of the image. Wavelet based rule
alone may reduce the contrast in some images and cannot effectively remove the
ringing effects and noise appearing in the source images. Curvelet method can work
well with edges and boundaries and curve portions of the images using Ridgelet
transforms.
In the hybrid method first the decomposition of the input images is done up to
level N by passing the image through series of low and high pass filters. The low and
high pass bands are then subjected to curvelet transform by decomposing it further
into small tiles and then fused using wavelet transform and inverse wavelet transform
to get full size images. This will take into account the drawbacks of wavelet and
effectively remove it using curvelet transform and visual quality of the image is
improved. Wavelet transform of an image up to level N till different resolution is
obtained. This gives various frequency bands. In chapter 4 we discussed about the
operation of wavelet and curvelet transform in detail. There we understand clearly
how the medical images are fused using these transformations.
Step 2: These images are set to pre-processing which includes RGB to gray scale
conversation and also ensured image alignment.
Step 3: The images obtained from step-2 are first decomposed using Discrete Wavelet
Transform (WT).
Step 4: We get fused image in wavelet domain using following fusion rules.
Step 6: Apply contourlet transform on the fused image which is obtained in step-5.
Step 8: To get final hybrid fused image in spatial domain we apply inverse contourlet
transform.
CHAPTER 4
PROPOSED METHODOLOGIES
After pre-processing these source images are first decomposed using Wavelet
Transform (WT) in order to realize multi-scale sub band decompositions with no
redundancy. These sub-bands coefficients are predominantly low and the high
frequency sub-bands of the image. Now, the obtained approximation and detailed
coefficients after application of appropriate fusion rule are reconstructed using the
inverse DWT. The entire process carried out in this stage serves to provide significant
localization leading to a better preservation of features in the fused image.
discrete wavelet transforms (DWT). The coefficients of these filters are evaluated
using mathematical analysis. The wavelet transform is used to identify local features
in an image. It also used for decomposition of two dimensional (2D) signals such as
2D gray-scale image for multi-resolution analysis. The available filter banks
decompose the image into two different components i.e. high- and low- frequency.
When decomposition is carried out, the approximation and detail components can be
separated 2-D Discrete Wavelet Transformation (DWT) converts the image from the
spatial domain to transform domain. The image is divided by vertical and horizontal
lines and represents the first-order of DWT, and the image can be separated with four
parts those are LL1, LH1, HL1 and HH1.
The coefficient matrices IL (x, y) and IH(x, y) are both low pass and high pass
filtered in vertical direction and down sampled by a factor of two to create sub bands
(sub images) ILL(x, y) , ILH (x, y) , IHL(x, y), and IHH (x, y). Wavelet decomposition can
be implemented by two channel filter bank shown in Figure 4.3.
The Discrete Wavelet Transform has the property that the spatial resolution is
small in low-frequency bands but large in high frequency bands. This is because the
scaling function is treated as a low pass filter and the mother wavelet as high pass
filter in DWT implementation. The wavelet transform decomposition and
reconstruction take place column and row wise. Firstly row by row decomposition is
performed and then column by column. This can be shown in Figure 4.4.
The ILL(x, y) sub-band is the original image at the coarser resolution level,
which can be considered as a smoothed and sub-sampled version of the original
image. Most information of their source images is kept in the low frequency sub-
band. It represents the frequency usually contains slowly varying grey value
information in an image so called approximation.
The ILH (x, y), IHL(x, y) and IHH(x, y) are sub-bands contain the detail
coefficients of an image, which usually have large absolute values correspond to sharp
intensity changes and preserve salient information in the image.
1, 2, 3 - - - Decomposition Levels
There are different levels of decomposition which are shown in Figure 4.5.
After one level of decomposition, there will be four frequency bands, as listed above.
By recursively applying the same scheme to the LL sub-band a multi-resolution
decomposition with a desires level can then be achieved.
The schematic diagram of wavelet- based image fusion is shown in Figure 4.6.
In wavelet image fusion scheme, the source images I1(x, y) and I2(x, y), are
decomposed into approximation and detailed coefficients at required level using
DWT. Two images, I1(X1, X2) and I2(X1, X2) are registered. Wavelet transform is
applied on two images. It can be represented by the following equation,
Where W −1 and φ are the Inverse Discrete Wavelet Transform operator and fusion
operator. There is several wavelet fusion rules that can be used for the selection of the
wavelet coefficients from the wavelet transform of the images to be fused. The most
frequently used rule is the maximum frequency rule which selects the coefficients that
have the maximum absolute values. The Wavelet Transform concentrates on
representing the image in multi-scales and it’s appropriate to represent linear edges.
The multi-level image fusion using DWT is shown in Figure 4.6.
Fused Wavelet
Registered Wavelet
Source images coefficients
The fusion rule used in this paper is simply averages the approximation
coefficients and picks the detailed coefficient in each sub band with the largest
magnitude. Thus, N-level decomposition will finally have 3N+1 different frequency
bands, which include 3N high frequency bands and just one LL frequency band. This
decomposition carried out until desired resolution is reached. It depends upon the
ratio of spatial resolution of the image.
Step 1: The images to be fused must be registered to assure that the corresponding
pixels are aligned.
Step 2: These images are decomposed into wavelet transformed images, respectively,
based on wavelet transformation. The transformed images with K -level
decomposition will finally have 3K+1 different frequency bands, which
include one low-frequency portion (ILL) and 3K high-frequency portions
(low-high bands, high-low bands, and high-high bands).
Step 3: The transform coefficients of different portions or bands are performed with a
certain fusion rule.
A. Need for contourlet transform: Among these desiderata, the first three are
successfully provided by separable wavelets, while the last two require new
constructions. Moreover, a major challenge in capturing geometry and directionality
in images comes from the discrete nature of the data. For this reason we construct
multi-resolution and multi-direction image expansion using non-separable filter
banks.
Laplacian Pyramid
Directional Filter Bank
The overall result is an image expansion using basic elements like contour
segments, and thus is named contourlet. In particular, contourlet have elongated
supports at various scales, directions and aspect ratios. This allows contourlets to
efficiently approximate a smooth contour at multiple resolutions.
4.3.1 Laplacian Pyramid: In the image is first decomposed into four sub images and
also captures the point discontinuities in those images. The Figure 4.8 shows general
representation of Laplacian pyramid
This is the process that, Laplacian pyramids separates low frequency and high
frequency components. The obtained LF Sub-band (Scaled) image is further
decomposed to get the desired fused image.
4.3.2 Directional Filter Bank:
1) The High frequency components are given to the directional filter bank which links
point discontinuities into linear structure.
2) The High pass sub-band images are applied to Directional filter bank to further
decompose the frequency spectrum using an n-level iterated tree structured filter
banks as shown in Figure 4.11.
3) By doing this we capture smooth contours and edges at any orientation. Finally we
combine the scaled information with scaled multiplication. Since the directional
filter bank (DFB) was designed to capture the high frequency (representing
directionality) of the input image, the low frequency content is poorly handled.
4) In fact, with the frequency partition low frequency would “leak” into several
directional sub-bands, hence the DFB alone does not provide a sparse representation
for images. This fact provides another reason to combine the DFB with a multi-scale
decomposition as shown in Figure 4.12., where low frequencies of the input image
are removed before applying the DFB.
The fusion framework used in the experiments is shown in Figure 4.13. First,
source images are decomposed into multi-scale and multi-directional components
using contourlet transform, and these components are fused together based on a
certain fusion scheme. Next, inverse contourlet transform is performed in order to
obtain a final fused image.
The source images are fused according to the fusion scheme and fusion rule that are
described as follows in Figure 4.14.
Once the source images are decomposed, high frequency components are
selected from the PAN source image and then injected into detail sub-bands of the MS
source image via maximum frequency fusion rule which compares and selects the
frequency coefficient with the highest absolute value at each pixel.
Step 2: These images are set to pre-processing which includes RGB to gray scale
conversation and also ensured image alignment.
Step 3: The images obtained from step-2 are first decomposed using Discrete Wavelet
Transform (WT).
Step 4: We get fused image in wavelet domain using following fusion rules.
Step 6: Apply contourlet transform on the fused image which is obtained in step-5.
Step 8: To get final hybrid fused image in spatial domain we apply inverse contourlet
transform.
After pre-processing these source images are first decomposed using Curvelet
Transform (WT) in order to isolate different frequency components present of the
image into different planes without down sampling as in the traditional wavelet
transform. It efficiently deals with the curved shapes; therefore its application in
medical fields would result in better fusion results than obtained using wavelet
transform alone. Curvelet method can work well with edges and boundaries and curve
portions of the images using Ridgelet transforms. Now, the obtained approximation
and detailed coefficients after application of appropriate fusion rule are reconstructed
using Curvelet transform. The entire process carried out in this stage serves to provide
significant localization leading to a better preservation of features in the fused image.
Contourlet Transform (CT) works with two dimensional multi scale and directional
filter bank (DFB). In addition CT also uses iterated filter bank which makes it
computationally efficient. The perfect directional basis for discrete signals is created
with the help of DFB which was major drawback of the wavelet transformation.
Need for Curvelet: The wavelet transform concentrates on representing the image in
Multi-scales and it’s appropriate to represent linear edges. For curved edges, the
accuracy of edge localization in the wavelet transform is low. So, there is a need for
an alternative approach which has a high accuracy of curve localization such as the
curvelet transform.
4.8.1 Image Fusion by Discrete Curvelet Transform Method:
1. Sub band filtering: The purpose of this step is to decompose the image into additive
components, each of which is a Sub band of that image. This step isolates the different
frequency components of the image into different planes without down sampling as in
the traditional wavelet transform.
2. Tiling: Tiling is the process by which the image is divided into overlapping tiles. It
follows the sub-band decomposition, each of the sub-band filtered image is then
partitioned into blocks of NxN (N- blocks in horizontal direction & N-blocks in
vertical direction). These tiles are small in dimensions to transform curved lines into
small straight lines in the sub bands P1and P2. The tiling improves the ability of the
curvelet transform to handle curved edges. As observing in Figure 4.18.
3. Renormalization: Renormalization is nothing but centering each dyadic square to the
unit square [0, 1]x[0, 1].
4. Ridgelet Analysis: Before the Ridgelet Transform we need to perform the Ridgelet
tiling. The renormalized ridges have an aspect ratio of width= length2 . Now these
ridges can be encoded ridges efficiently using the Ridgelet Transform.
The curvelet transform was developed initially in the continuous domain via
multi-scale filtering and then applying a block Ridgelet transform on each band-pass
image. Later, the second generation curvelet transform was proposed that defined
directly via frequency partitioning without using the Ridgelet transform. Both curvelet
constructions require a rotation operation and correspond to a 2-D frequency partition
based on the polar coordinate. This makes the curvelet construction simple in the
continuous domain but causes difficulty in implementation for discrete images. So
that we go for contourlet transform.
Step 2: These images are set to pre-processing which includes RGB to gray scale
conversation and also ensured image alignment.
Step 3: The images obtained from step-2 are first decomposed using Curvelet
Transform to estimate the coefficients. Curvelet transformation has four stages
1. Sub-band decomposition
2. Smooth Partitioning
3. Renormalization
4. Ridgelet Analysis
Step 4: We get fused image in wavelet domain using following fusion rules.
Step 6: Apply contourlet transform on the fused image which is obtained in step-5.
Step 8: To get final hybrid fused image in spatial domain we apply inverse contourlet
transform.
The basic building block of MATLAB is MATRIX. The fundamental data type
is the array. Vectors, scalars, real matrices and complex matrix are handled as specific
class of this basic data type. The built in functions are optimized for vector operations.
No dimension statements are required for vectors or arrays.
Command Window: The command window is where the user types MATLAB
commands and expressions at the prompt (>>) and where the output of those
commands is displayed. It is opened when the application program is launched. All
commands including user-written programs are typed in this window at MATLAB
prompt for execution.
Work Space Window: MATLAB defines the workspace as the set of variables that
the user creates in a work session. The workspace browser shows these variables and
some information about them. Double clicking on a variable in the workspace
browser launches the Array Editor, which can be used to obtain information.
Current Directory Window: The current Directory tab shows the contents of the
current directory, whose path is shown in the current directory window. For example,
in the windows operating system the path might be as follows: C:\MATLAB\Work,
indicating that directory “work” is a subdirectory of the main directory “MATLAB”;
which is installed in drive C. Clicking on the arrow in the current directory window
shows a list of recently used paths.
MATLAB uses a search path to find M-files and other MATLAB related files.
Any file run in MATLAB must reside in the current directory or in a directory that is
on search path.
Editor Window: The MATLAB editor is both a text editor specialized for creating
M-files and a graphical MATLAB debugger. The editor can appear in a window by
itself, or it can be a sub window in the desktop. In this window one can write, edit,
create and save programs in files called M-files.
MATLAB editor window has numerous pull-down menus for tasks such as
saving, viewing, and debugging files. Because it performs some simple checks and
also uses color to differentiate between various elements of code, this text editor is
recommended as the tool of choice for writing and editing M-functions.
Graphics or Figure Window: The output of all graphic commands typed in the
command window is seen in this window.
Online Help Window: MATLAB provides online help for all it’s built in functions
and programming language constructs. The principal way to get help online is to use
the MATLAB help browser, opened as a separate window either by clicking on the
question mark symbol (?) on the desktop toolbar, or by typing help browser at the
prompt in the command window.
The help Browser is a web browser integrated into the MATLAB desktop that
displays a Hypertext Mark-up Language (HTML) documents. The Help Browser
consists of two panes, the help navigator pane, used to find information, and the
display pane, used to view the information. Self-explanatory tabs other than navigator
pane are used to perform a search.
Syntax: clc
Uigetfile: Displays a modal dialog box that lists files in the current directory and
enables the user to select or type the name of a file to be opened. If the filename is
valid and if the file exists, uigetfile returns the filename when the user clicks Open.
Syntax=uigetfile (FilterSpec,DialogTitle,DefaultName)
Imread: This command reads the image from the file specified by filename with the
standard file extension indicated by file type as given below:
Imresize: This command resizes the image of any type using this specified
interpolation method.
Syntax: figure(I)
Syntax: im2double(image)
Syntax: zeros(n)
Syntax: ones(n)
Size: This command returns the sizes of each dimension of array x in a vector d with
ndims(x).
Syntax: d=size(x)
Syntax: i=rgb2gray(i)
Syntax: imfuse(A,B)
CHAPTER 5
Example 1: Fusion of CT and MRI: Image Size [256 X 256] having tumour in
brain
In Figure 5.1 we considered CT and MRI images of brain with tumour. The
first one is the input source image which is obtained from CT scan of brain. CT
provides more accurate information about calcium deposit, air, bones, and any
blockages. The second input image which is considered as a reference which is
obtained from MRI scan of brain. MRI provides information about Nerve system, soft
tissues and muscles. We have applied Discrete Wavelet Transform (DWT) on the
images which then followed by contourlet transform.
We fuse these images using appropriate fusion rules which we have already
discussed in previous chapters. The result we obtained from contourlet transform
which will be in frequency domain. We need to reconstruct final hybrid fused image
in spatial domain, for that we apply inverse wavelet transform followed by inverse
contourlet transform.
Example 2: Fusion of MRI and PET: Image Size [256 X 256] for FTD (Neuron
Degeneration)
Example 3: Fusion of CT and MRI: Image Size [512 X512] of skull from top view
In Figure 5.3 we considered CT and MRI images of skull. The first one is the
input source image which is obtained from CT scan of skull. The second input image
which is considered as a reference which is obtained from MRI scan of skull. We have
applied Discrete Wavelet Transform (DWT) on the images which then followed by
contourlet transform.
We fuse these images using appropriate fusion rules. The result we obtained
from contourlet transform which will be in frequency domain. We need to reconstruct
final hybrid fused image in spatial domain, for that we apply inverse wavelet
transform followed by inverse contourlet transform.
Example 4: Fusion of MRI-T1 and MRI –T2: Image Size [256x 256] of brain
(cholesterol)
In Figure 5.4 we considered CT and MRI images of brain which shows the
cholesterol level. The first one is the input source image which is obtained from CT
scan of brain. The second input image which is considered as a reference which is
obtained from MRI scan of brain. We have applied Discrete Wavelet Transform
(DWT) on the images which then followed by contourlet transform. We fuse these
images using appropriate fusion rules
In Figure 5.5 we considered CT and MRI images of brain head. The first one is
the input source image which is obtained from CT scan of head. The second input
image which is considered as a reference which is obtained from MRI scan of brain.
We have applied Discrete Wavelet Transform (DWT) on the images which then
followed by contourlet transform.
We fuse these images using appropriate fusion rules. The result we obtained
from contourlet transform which will be in frequency domain. We need to reconstruct
final hybrid fused image in spatial domain, for that we apply inverse wavelet
transform followed by inverse contourlet transform.
5.2 Simulation Results of Curvelet – Contourlet Transform : In
this section we are discussing the simulation results of hybrid Curvelet
Transform and Contourlet Transform for various medical images obtained
from different modalities
Example 1: Fusion of CT and MRI: IMAGE SIZE [256 x256] having tumour in
brain
In Figure 5.6 we considered CT and MRI images of brain with tumour. The
first one is the input source image which is obtained from CT scan of brain. CT
provides more accurate information about calcium deposit, air, bones, and any
blockages. The second input image which is considered as a reference which is
obtained from MRI scan of brain. MRI provides information about Nerve system, soft
tissues and muscles.
Example 2: Fusion of MRI and PET: Image Size [256 x256] for FTD (Neuron De-
Generation)
image which is considered as a reference which is obtained from PET scan of brain.
PET can be used to provide better information on blood flow and flood activity with
low spatial resolution. As a result, the anatomical and functional medical images are
needed to be combined for a compendious view.
In Figure 5.8 we considered CT and MRI images of skull. The first one is the
input source image which is obtained from CT scan of skull. The second input image
which is considered as a reference which is obtained from MRI scan of skull. We have
applied curvelet transform on the images which then followed by contourlet
transform.
We fuse these images using appropriate fusion rules. The result we obtained
from contourlet transform which will be in frequency domain. We need to reconstruct
final hybrid fused image in spatial domain, for that we apply inverse contourlet
transform.
Example 4: Fusion of MRI-T1 and MRI –T2: IMAGE SIZE [256x 256] of brain
(cholesterol)
In Figure 5.9 we considered two images that are obtained from MRI Scans;
one is MRI_T1 and the second one MRI_T2. MRI_T1-weighted imaging is used to
differentiate anatomical structures mainly on the basis of T1 values.
Tissues with high fat content (e.g. white matter) appear bright and
compartments filled with water (e.g. CSF) appear dark. This is good for
demonstrating anatomy. MRI_T2 is vice versa. We have applied curvelet transform on
the images which then followed by contourlet transform. We fuse these images using
appropriate fusion rules. The result we obtained from contourlet transform which will
be in frequency domain. We need to reconstruct final hybrid fused image in spatial
domain, for that we apply inverse contourlet transform.
In Figure 5.10 we considered CT and MRI images of head. The first one is the
input source image which is obtained from CT scan of head. The second input image
which is considered as a reference which is obtained from MRI scan of brain.
In this section we will check the effectiveness of the proposed schemes that is
hybrid of the DWT-Contourlet, Curvelet- Contourlet transformation. Various
parameters such as Entropy, PSNR, and MSE are used to evaluate the effectiveness
and compared the performance metrics among the proposed methods and existed one.
We assume the source images to be in perfect registration. Here we consider different
source images like CT, MRI, and PET of brain tumour, Skull, MRI_T1, MRI_T2, and
Alzheimer’s disease (which is widely seen in aged people).
(a)CT-SKULL(b)MRI-SKULL(c)DWT-Curvelet(d)DWT-contourlet(e)Curvelet-Contourlet
(f)CT-TUMOR (g)MRI-TUMOR(h)DWT-Curvelet(i)DWT-contourlet(j)Curvelet-Contourlet
(p) MRI_T1 (q) MRI-T2 (r) DWT-Curvelet (s) DWT-Contourlet (t) Curvelet-Contourlet
The above images shown in Figure 5.11 gives the comparison of results obtained
from different techniques. The Figure (a) and (b) gives bone and tissues information
of skull, by fusing these images we obtained complete information using appropriate
hybrid technique. Here Figure (c), (d) and (e) represents existed and proposed
methods. Similarly the Figure (f) and (g) gives bone and tissues information of brain
having tumour, by fusing these images we obtained exact location of tumour using
appropriate hybrid technique. Here Figure (h), (i) and (j) represents existed and
proposed methods. Similarly the Figure (k) and (l) gives soft tissues and PET can be
used to provide better information on blood flow and flood activity with information
of brain, by fusing these images we obtained exact location of tumour using
appropriate hybrid technique. Here Figure (m), (n) and (o) represents existed and
proposed methods.
Similarly the Figure (p) and (q) gives MRI_T1-weighted imaging is used to
differentiate anatomical structures mainly on the basis of T1 values. Tissues with high
fat content (e.g. white matter) appear bright and compartments filled with water (e.g.
CSF) appear dark. This is good for demonstrating anatomy. MRI_T2 is vice versa. By
A. Entropy
L=1
E= ∑ Pilog (Pi) (5.1)
i=0
M N
2
∑ ∑ [ Xij−Yij ] (5.2)
MSE= i=1 j=1
MN
PSNR is the ratio between the maximum possible power of a signal and the power of
corrupting noise that affects the fidelity of its representation.
2 B−1
PSNR=20 log 10 ( ) (5.3)
MSE
Table 5.1: Performance metrics for Fusion of CT and MRI image size of skull
Figure 5.12: Comparison of PSNR, MSE and Entropy in terms of bar charts for CT
and MRI of skull
Table 5.2: Performance metrics for Fusion of CT and MRI image size having tumour
in brain
Chart 5.13: Comparison of PSNR, MSE and Entropy in terms of bar charts for CT and
MRI for brain having tumour
Table 5.3: Performance metrics for Fusion of MRI and PET image size for FTD
(NEURON DE-GENERATION)
Figure 5.14: Comparison of PSNR, MSE and Entropy in terms of bar charts for MRI
and PET for FTD (NEURON DE-GENERATION)
Table 5.4: Performance metrics for Fusion of CT and MRI image size for head
Figure 5.15: Comparison of PSNR, MSE and Entropy in terms of bar charts for CT
and MRI of Head
Table 5.5: Performance metrics for Fusion of MRI_T1 AND MRI_T2 image size [256
256] for brain (cholesterol)
Figure 5.16: Comparison of PSNR, MSE and Entropy in terms of bar charts for
MRI_T1 AND MRI_T2 for brain (cholesterol)
These bar charts and Tabular values give the experimental values performed on
different images we discussed previously. From these we can observe that curvelet-
contourlet transform has given better results for which PSNR and Entropy values are
of high values when compared to other transformation techniques and the MSE is less
when compared with other methods, which satisfied the conditions of better image
quality that will obtain after fusion.
CHAPTER 6
CONCLUSION
In this project, a hybrid technique for image fusion using the combinations of
wavelet, curvelet, contourlet is being simulated. The simulated results for different
hybrid combinations of above mentioned transforms are tested and compared for
various medical image combinations like MRI, and also for various input image sizes.
In all cases , curvelet-contourlet based hybrid technique is observed to be outsmarting,
which provide best quality fused image than other two combinations in terms PSNR ,
MSE and Entropy. Here Curvelet-Contourlet based hybrid technique suites best for
medical diagnosis
Future scope
In my project we have applied fusion on non noise images and this can be
further modified so that we can fuse and extract the features of noisy images. Also we
have performed fusion of only two images which can be further extended in fusing
multiple images.