Self-Organizing Maps For Texture Classification
Self-Organizing Maps For Texture Classification
Self-Organizing Maps For Texture Classification
1007/s00521-011-0797-x
ORIGINAL ARTICLE
Received: 4 March 2011 / Accepted: 26 December 2011 Springer-Verlag London Limited 2012
Abstract A further investigation of our intelligent machine vision system for pattern recognition and texture image classication is discussed in this paper. A data set of 335 texture images is to be classied into several classes, based on their texture similarities, while no a priori human vision expert knowledge about the classes is available. Hence, unsupervised learning and self-organizing maps (SOM) neural networks are used for solving the classication problem. Nevertheless, in some of the experiments, a supervised texture analysis method is also considered for comparison purposes. Four major experiments are conducted: in the rst one, classiers are trained using all the extracted features without any statistical preprocessing; in the second simulation, the available features are normalized before being fed to a classier; in the third experiment, the trained classiers use linear transformations of the original features, received after preprocessing with principal component analysis; and in the last one, transforms of the features obtained after applying linear discriminant analysis are used. During the simulation, each test is performed 50 times implementing the proposed algorithm. Results from the employed unsupervised learning, after training, testing, and validation of the SOMs, are analyzed and critically compared with results from other authors.
Keywords Self-organizing maps Texture classication Feature extraction Statistical analysis PCA LDA
1 Introduction Analysis, recognition, and classication of texture patterns and images are topics with current surge of research interest in the eld of digital image processing and pattern recognition, with wide areas of applications [15]. A number of different methods, algorithms, and paradigms have been or are being developed nowadays [69]. The investigated image classication and recognition systems may vary in their approach but most of them include data acquisition, data preprocessing, feature extraction, feature analysis, classication, and testing and evaluation stages [811]. The preprocessing of the raw data is difcult but important part of the whole process, whose aims are to extract useful and appropriate characteristics and features that are to be used in the later stages [8]. Often, the raw data are too large or complex to be used directly as input to a classier, leading to the curse of dimensionality and other problems related to the generalization abilities of the trained systems, especially when insufcient training samples are available. Even if this is not the case, reducing the number of variables representing the data can speed up and facilitate the learning process at later stages [11]. That is why principal component analysis (PCA), for example, is a widely accepted technique in such cases [1, 2, 12]. In [12], we investigated a texture images classication problem, using supervised neural network learning, for which a priori knowledge about the image classes was used. The aim of this research is to extend this previous work, considering the same classication problem, but assuming there is no expert knowledge available for the texture
N. Petrov (&) I. Jordanov School of Computing, University of Portsmouth, Portsmouth PO1 3HE, England, UK e-mail: Nedyalko.Petrov@port.ac.uk I. Jordanov e-mail: Ivan.Jordanov@port.ac.uk A. Georgieva NDOG, University of Oxford, Oxford OX3 9DU, England, UK e-mail: Antoniya.Georgieva@obs-gyn.ox.ac.uk
123
classes of the data set samples. This implies that no supervised learning can be used, and the knowledge about the texture patterns and their similarity and uniformity has to be extracted from the data set itself. Unsupervised classication of texture patterns and images is widely used approach with applications in a broad range of areas, for example: for determining water quality based on some chemical and physicochemical features [1], for classication of SAR images [2], for texture-based classication of atherosclerotic carotid plaque images for determining risk of stroke for individuals [13], for classifying volcanic ash using surface texture features [3], for automatically classifying texture structure of different fabric types using SOM [14], for classication of textures in scene images using biology inspired features [6], for classication of aerial images using SOMs [15]. In this investigation, a data set of 335 texture images, acquired via an intelligent visual recognition system, as reported in [12], is used. Each data sample of the set represents a grayscale image of an industrial cork tile that was classied in the previous paper into one of seven classes Beach, Corkstone, Desert, Lisbon, Pebble, Precision and Speckled. The distribution of the texture classes is nonuniform and is shown in Fig. 1.
The simulation of the investigated system is divided in ve main stages: data acquisition, feature extraction, feature analysis, classier training, and classier testing and evaluation. The rest of the paper is organized as follows: Sect. 2 presents information about the data acquisition, feature extraction, and feature analysis and reduction stages, while Sect. 3 covers the classication stage. The results from the conducted tests are given and discussed in Sect. 4. Finally, Sect. 5 concludes the paper and gives some ideas for future work.
2 Data acquisition and feature extraction The texture image data set used in this paper is acquired via an intelligent visual recognition system described in more detail in [12]. The system consists of a charge-coupled device camera, lightning devices, and scaffolding. Since the texture of the samples is of prime interest, the images are converted to a grayscale format. As mentioned above, a total of 335 grayscale images of size 230 9 340 pixels of cork tile samples of 7 predened by experts types were collected (see Fig. 2). The feature extraction phase in our investigation aims to identify characteristics and properties that make the classes of samples distinct from each other [16]. At this stage of the process, features that represent some valuable information about the texture of the images are obtained. This is preceded by image normalization. 2.1 Initial feature extraction In order to reduce the illumination effects on the analyzed images (e.g., due to a glare), a normalization technique is applied. In this process, a small window (15 9 15 pixels) is moved within each image and the local average is subtracted from the pixels values, in order to get images with average intensity of each neighborhood about a zero [9]. Afterward, 34 features are extracted using classical approaches.
Speckled 12%
Lisbon 12%
Desert 19%
Fig. 1 Distribution of the texture classes Fig. 2 Samples of the acquired texture dataimages of seven different types of wall cork tiles: Beach, Corkstone, Desert, Lisbon, Pebble, Precision and Speckled
123
2.1.1 Co-occurrence matrices Co-occurrence matrices, introduced by Haralick in [17], is a commonly applied statistical approach for texture features extraction that takes into account relative distances and orientation of pixels with co-occurring values [9, 15, 18]. The MATLABs Image Processing Toolbox is used for the computation of the co-occurrence matrices of the normalized images. As usually proposed by other authors [19], four relative orientations are usedhorizontal (0), right diagonal (45), vertical (90), and left diagonal (135). In this way, the energy, homogeneity, correlation, and contrast characteristics in each direction are computed, getting as a result the rotation invariant features [9, 11]. Also, two spatial relationships are consideredthe direct neighbors and the pixels with difference of ve. As a result, a total of eight co-occurrence matrices are obtainedfour for the direct neighbors and another four for the pixels with difference of ve. 2.1.2 Laws masks The Laws masks are used as a lter technique that is applied to identify points of high energy in an image [20]. Masks are derived from one-dimensional (1-D) vectors of ve pixels length, proposed by Laws, to pick up the average gray level, edges, ripples, spots, and waves [12, 13]: L5 (Level) = [1 4 6 4 1] ? Level detection; E5 (Edge) = [-1 -2 0 2 1] ? Edge detection; S5 (Spot) = [-1 0 2 0 -1] ? Spot detection; R5 (Ripple) = [1 -4 6 -4 1] ? Ripple detection; W5 (Wave) = [-1 2 0 -2 1] ? Wave detection. The vectors are multiplied each other (the second vector is transposed) and this way 25 different 5 9 5 masks are produced. The masks are then applied to the normalized set of samples and the obtained ltered images are converted to texture energy maps. The aim of this process (also called smoothing) is to deduce the local magnitudes of the quantities of interest (edges, spots, etc.). A smoothing window of size 15 9 15 [9] is applied to each ltered image Fk for the k-th mask and new energy images are obtained, where each pixel in the image is given by (1): Ek r; c
c7 r7 X X jc7 ir7
2.1.3 Entropy Entropy is a statistical measure of randomness that can be used to characterize the texture of an image [9, 14]. It takes low values for smooth images and vice versa. The entropy for each image sample is calculated using a MATLABs build-in function, according to (2): E
G X i1
where G is the number of gray levels in the images histogram, ranging between 0 and 255 for a typical 8-bit image, and d(i) is the normalized occurrence frequency of each gray level. 2.2 Statistical analysis and feature reduction Before applying any statistical analysis, a random subset of 25% of the available data is excluded for the purposes of further testing. This subset will be referred to as the testing set from now on and the remaining 75% of the available data will be the training set. During the feature extraction stage, a total of 34 features are obtained for each texture image (8 by the co-occurrence method, 25 by Laws masks and 1 entropy feature). The distribution of the seven classes of the training set, represented by two randomly selected from the 34 features is shown in Fig. 3. Figure 3b presents the classes distribution according to the 2nd and the 5th features of the original data set and Fig. 3a shows the classes means with 95% condence interval. As it can be seen from Fig. 3, the considerable overlap between the classes makes the classication process more challenging. In order to reduce the dimensionality of the classication problem (i.e., the number of inputs to the classier), to reduce the redundant information (i.e., the information contained in some highly correlated features), and to improve the class separability, two statistical analysis techniques [10] are used in some of the experiments. They are described in more details in the next two subsections. 2.2.1 Principal component analysis PCA is an eigenvalue-based multivariate technique that transforms a number of possibly correlated features into a number of uncorrelated features, called principal components (PC) [2, 9]. The number of the derived PCs is less than or equal to the number of the original features. It is an unsupervised technique and as such does not use any labeled information on the data. The rst PC accounts for as much of the variability (information) in the data, as possible, and each succeeding
jFk i; jj;
k 1; . . .; 25;
where (r, c) denotes the rows and columns indices. After obtaining 25 energy maps for each image, a power metric, representing the sum of the squared absolute values for each pixel in the map is used [9], to nally obtain 25 different values for each texture sample.
123
Neural Comput & Applic Fig. 3 Texture types distribution, according to two randomly selected features from the training set: a classes means with 95% condence intervals; b scatter plot of the samples
Fig. 4 Percentage of the information from the training set contained: a in the rst ve PCs for the PCA experiment; b in the rst ve eigenvalues for the LDA experiment
PC accounts for as much of the remaining variability as possible. Depending on the areas of application, PCA is also referred to as Hotelling transform, KarhunenLoeve transform, or proper orthogonal decomposition [9]. The PCA implementation of the MATLABs Statistics Toolbox is used for processing the extracted features of the training set. As a result, a new data set in which the rst 5 features contain about 97% of the total variation (information) is obtained (Fig. 4a). The PCA transformation matrix is saved for further use in the evaluation stage. Figure 5 shows the distribution of the seven texture classes, represented by the rst and second PCs. It can be seen that four out of the seven classes (Beach, Corkstone, Desert, and Pebble) are easily separable from the others. However, the rest of the classes are too close to each other
and partially overlap. This is because the PCA considers all the data samples independently, without taking into account which class they belong to. The overlapping in some of the classes however is expected to harden the classiers performance later on. 2.2.2 Linear discriminant analysis Linear discriminant analysis (LDA) is an eigenvaluesbased transformation technique that aims to nd a linear combination of features that characterize or separate two or more classes [9, 21]. LDA is not used in this work as a classication technique, but as a data preprocessing transform, before applying the classication technique, as recommended in [10]. The number of the newly generated
123
Neural Comput & Applic Fig. 5 Texture types distribution, according to the rst two PCs: a classes means with 95% condence intervals; b scatter plot of the samples
features is always one less than the number of the classes. An LDA implementation in MATLAB, following the algorithm presented in [21], is employed for this research. LDA is applied to the features extracted for each texture sample of the training set. As a result, the dimensionality of the feature space is reduced from 34 to 6 without loss of information about the class separability [11] and the LDA transformation matrix is saved for further use in the evaluation stage. Figure 4b shows the percentage contribution of each eigenvalue to the sum of the six eigenvalues. It can be seen that about 98.5% of the eigenvalues sum is contributed by the rst ve eigenvalues. The classes means with 95% condence intervals and the scatter plot of the processed with LDA data are shown in Fig. 6. It can be seen that the classes separability is considerably improved.
that they use a neighborhood function to preserve the topological properties of the input space [22]. Like most neural networks, SOMs operate in two modes: training and testing. The MATLABs implementation of SOM is employed for this research and the following algorithm is used for the classication: 1. 2. 3. Design of SOMs architecture (map topology, number of neurons, training parameters, etc.); Training of the SOM with data subset, representing the extracted texture features (75% of the available data set); As a result of step 2, a 2D map is obtained, in which each node and its closest neighbors represent similar data samples (Fig. 7); Based on the available expert knowledge for the training samples, the count of the samples belonging to a certain class is determined for each node of the map; Each node is then labeled to represent just one class the class with predominant number of associated samples. In case equal number of samples of different classes is mapped to a certain node, the node is labeled to the predominant class in its neighborhood (Fig. 7). A node gets no label if there are no data samples mapped to it (node [0,4] in Fig. 7b); The classiers testing is performed with the remaining 25% of the available data; Each testing sample label is compared to the label of the node that it is mapped to. A sample is counted as unclassied if it is mapped to an unlabeled node;
4.
5.
3 Classication For the classication of the texture samples data, selforganizing maps (SOM) are employed. As it is known, a SOM is an articial neural network (NN) that is trained using unsupervised learning to produce a low-dimensional (typically two-dimensional), discretized representation of the input space of the training samples, called map. A specic characteristic of SOMs (compared to other NNs) is
6. 7.
123
Neural Comput & Applic Fig. 6 Texture types distribution, according to the rst two eigenvalues: a classes means with 95% condence intervals; b scatter plot of the samples
Fig. 7 Sample SOM classier map. Image a presents the node hits for the samples from the training set and b from the testing set. The number in each node represents its hits. The nodes are colored according to the classes they are labeled to. Image c shows the relative distance between the map nodes. Darker color corresponds to larger distances
(a)
Speckled Lisbon
(b)
Precision Speckled
Precision
Lisbon
Corkstone
Desert
Pebble
Beach
Corkstone
Desert
Pebble
Beach
(c)
Precision Speckled Lisbon
Corkstone
Desert
Pebble
Beach
8.
4 Simulation and results MATLAB 2010B and its Neural Network, Image Processing and Statistics Toolboxes are used for the computations and simulations presented in this paper. Four major experiments are conducted: in the rst one, the classiers are trained using all the extracted features without any statistical preprocessing; in the second, the extracted features are normalized before being fed to a
where a is the accuracy of the classier, nc is the number of correctly classied samples, nw is the number of wrongly classied samples and nu is the number of unclassied samples.
123
classier; in the third experiment, the trained classiers use features obtained after preprocessing with PCA; and in the last one, features obtained after applying LDA are used. During the simulation, each test is performed 50 times using the algorithm given in Sect. 3. The minimum, maximum, and mean percentages of successfully classied texture images from the testing set are recorded, and the mean standard deviation over the 50 runs is also calculated. 4.1 Classication without statistical preprocessing In this experiment, SOMs are trained using all the 34 extracted features. No statistical preprocessing is performed, and random 75% (251 texture images) of the available data samples are used for training and the remaining 25% (84 texture images) for testing. Tables 1 and 3 show results from simulations with varying number of training epochs and varying number of neurons for different SOMs topologies. The sample confusion matrix given in Table 4 shows excellent performance of the classier for two of the classes (Lisbon and Speckled) and inferior results for the rest. 4.2 Classication with features normalization
run. It can be seen that the classiers performance is improved, and it is now able to better distinguish most of the classes. However, it still experiences some difculties with the Beach and the Corkstone samples. 4.3 Classication with PCA In this case, statistically preprocessed with PCA data is used for the training of SOMs. Again, random 75% (251 texture images) of the available data samples are used for training and the remaining 25% (84 texture images) for testing. Similarly to the previous case, the number of training epochs, the number of neurons in the SOM, the SOMs topology, and the number of principal components (PC) used for the training are varied. Each sub-experiment is performed 50 times, and the minimal, maximal, and the mean accuracy (%) for these runs are recorded. The results are presented in Tables 5, 7, and Fig. 8a. The sample confusion matrix given in Table 8 shows that this classier experience slight difculties recognizing some of the Corkstone samples, but performs very well on the rest of the classes. 4.4 Classication with LDA
In this experiment, all 34 features are used for the SOMs learning and the training set is normalized, so that the features have zero mean and unity standard deviation. Tables 2 and 3 show results from simulations with varying number of training epochs and varying number of neurons for different SOMs topologies. Table 4 gives a sample confusion matrix of the classiers performance for one
Table 1 Variation of the classiers accuracy (in %) for different number of training epochs for SOM with 120 neurons (15 9 8 map topology) and no statistical preprocessing Epochs Min Max Mean Std 50 48.2 63.0 55.1 3.6 100 58.0 75.3 66.7 3.9 250 70.3 81.5 77.0 2.6 500 70.4 80.3 77.0 1.9 1,000 75.3 81.5 78.4 1.4 2,500 75.3 81.5 78.3 1.6 5,000 74.1 82.7 78.0 1.9 7,500 75.3 82.7 78.1 1.8
In the last experiment, SOMs are trained using data statistically preprocessed with LDA, while the same training/ testing data ratio (75% training, 25% testing) is kept intact. The parameters for this experiment are varied through the number of eigenvalues used, the number of training epochs, the number of neurons, and the SOMs topology. Each simulation is performed 50 times, and the minimal, maximal, and the mean accuracy (in %) for these runs are given in Fig. 8b, Tables 6, and 7. Table 8 presents a sample confusion matrix of the classiers performance for one run. It can be seen that this classier is able to distinguish all the classes, and the classication error is mainly contributed by the unclassied samples (mapped to an unlabeled node). 4.5 Analysis of the results Figure 8a illustrates that no signicant improvement of the accuracy is obtained when more than 5 principal components are used (PCA case), and for the LDA case (Fig. 8b), the rst 3 eigenvalues bring the most signicant improvement. This could also be concluded from the graphics given in Fig. 4. Regarding the SOMs topology, no clear corelation between the accuracy and the number of used neurons was observed (Tables 3 and 7), but more experiments need to be done in order to investigate this in more detail.
SOMs with 120 neurons (15 9 8 map topology) are trained Table 2 Variation of the classiers accuracy (in %) for different number of training epochs for SOM with 120 neurons (15 9 8 map topology) after normalization Epochs Min Max Mean Std 50 71.6 86.4 77.8 3.6 100 79.0 90.1 84.9 3.1 250 84.0 93.8 88.7 2.4 500 84.0 93.8 89.8 2.0 1,000 85.2 93.8 89.9 2.1 2,500 85.2 93.8 89.8 1.8 5,000 87.7 95.1 90.8 1.8 7,500 87.7 93.8 90.9 1.6
123
Neural Comput & Applic Table 3 Variation of the classiers accuracy (in %) for different number of neurons and different SOM topology (trained for 500 epochs): with no statistical preprocessing on the left side of the cells and after normalization on the right
60 3 9 20 70.4/82.7 82.7/92.6 77.9/88.0 2.5/2.0 5 9 12 69.1/84.0 79.0/92.6 75.2/88.1 2.4/2.1 6 9 10 69.1/85.2 80.3/92.6 75.1/87.9 2.3/2.0
120 6 9 20 67.9/84.0 81.5/93.8 75.5/88.1 2.9/2.0 10 9 12 70.4/84.0 81.5/92.6 75.9/89.0 2.0/1.9 12 9 10 70.4/85.2 81.5/92.6 76.6/89.1 2.5/1.6
Table 4 Sample confusion matrix for SOM classier with 120 neurons (15 9 8 map topology) and 500 training epochs: with no statistical preprocessing on the left side of the cells and after normalization on the right
Actual
Predicted Beach Corkstone 1/1 8/7 0/0 0/0 1/0 0/0 0/0 Desert 0/0 0/1 10/15 0/0 0/1 1/0 0/0 Lisbon 0/0 0/0 0/0 11/11 2/0 1/1 1/0 Pebble 0/0 1/2 1/0 0/0 8/10 2/0 0/1 Precision 0/0 0/0 1/0 0/0 0/0 5/10 0/0 Speckled 0/0 0/0 1/0 0/0 0/0 1/0 9/9 Unclassied 0/1 1/1 0/0 0/0 0/0 0/0 0/0
Bold values represent the number of correctly classied samples of each class Table 5 Variation of the accuracy (in %) of the classier for different number of training epochs for SOM with 120 neurons, 15 9 8 map topology, and PCA preprocessing with 5 PCs
Precision Speckled
Fig. 8 Variation of the accuracy (in %) of the classier (SOM with 120 neurons, 15 9 8 map topology, 500 epochs). The border between the subbars shows the mean accuracy rate for the 50 runs. The top and
the bottom sections show the min and max rate, respectively, for: a different number of PCs used for the training (after PCA); b different number of eigenvalues used for training (after LDA)
Table 6 Variation of the accuracy (in %) of the classier for different number of training epochs for SOM with 120 neurons, 15 9 8 map topology, and LDA with 6 eigenvalues Epochs Min Max Mean Std 50 85.2 96.3 92.6 2.9 100 86.4 98.8 93.9 3.0 250 92.6 100.0 97.7 1.5 500 92.6 100.0 97.9 1.3 1,000 95.1 100.0 98.5 1.2 2,500 96.3 100.0 98.2 1.1 5,000 95.1 100.0 98.1 1.1 7,500 95.1 100.0 98.2 1.3
Figure 9 summerises and illustrates the obtained results for the four cases, presented in the previous section. It can be seen from the gure that, as expected, the worst accuracy is attained for the case with no statistical preprocessing. Although the accuracy of the normalized data looks better than the obtained one for the PCA case, it has to be noted that only ve principal components are considered during the training, whereas in the normalized case, all 34 extracted features are taken into account. The use of
123
Neural Comput & Applic Table 7 Variation of the classiers accuracy (in %) for different number of neurons, different SOM topology, 500 epochs after: PCA with 5 PCs on the left side of the cells and LDA with 6 eigenvalues on the right
60 3 9 20 81.5/96.3 91.4/100.0 86.7/98.7 2.1/1.1 5 9 12 81.5/96.3 92.6/100.0 87.8/99.2 2.2/0.9 6 9 10 82.7/96.3 91.4/100.0 87.4/99.1 1.8/1.0
120 6 9 20 81.5/95.1 93.8/100.0 87.1/98.6 2.2/1.4 10 9 12 82.7/93.8 92.6/100.0 88.7/97.9 2.0/1.2 12 9 10 84.0/93.8 91.4/100.0 88.4/97.6 1.7/1.4
Table 8 Sample confusion matrix for SOM classier with 120 neurons (15 9 8 map topology) and 500 training epochs: with PCA on the left side of the cells and with LDA on the right
Actual
Predicted Beach Corkstone 0/0 7/10 0/0 0/0 0/0 0/0 0/0 Desert 1/0 1/0 14/14 0/0 0/0 0/0 0/0 Lisbon 0/0 0/0 0/0 11/11 0/0 1/0 1/0 Pebble 0/0 2/0 0/0 0/0 11/11 0/0 0/1 Precision 0/0 0/0 0/0 0/0 0/0 10/11 0/0 Speckled 0/0 0/0 0/0 0/0 0/0 0/0 9/9 Unclassied 0/0 1/1 0/1 0/0 0/0 0/0 0/0
Bold values represent the number of correctly classied samples of each class
Precision Speckled
100
Accuracy
Training Epochs
Fig. 9 Bar graph showing the accuracy for the four case studies with increasing the number of training epochs
components. It can also be observed that the increase in the number of epochs for the runs does not lead to substantial increase in the accuracy, and above 250 epochs, an accuracy plateau is normally reached (Tables 1, 2, 5, and 6). The results for the PCA case, presented in Tables 5 and 7, are in good agreement with those given in [2], where the authors reported between 81 and 98% accuracy rate for a PCA-based unsupervised classication of SAR images. They are also very close to the [83, 95.5%] achieved in [15] and fall within the intervals with slightely larger accuracy variance, reported in [5, 6], where the results are within the [77, 100%] and [67, 92%] domains, respectively.
only ve PCs in the PCA case led to signicant reduction in the computational time, compared to the rst two experiments. Analyzing the sample confusion matrices for the four experiments (Tables 4 and 8), it can be said that the accuracy is improved (as expected) after applying LDA and PCA on the data sets, and this is especially valid for the Desert and Precision classes, while at the same time, the SOM kept excellent recognition rate for the Lisbon and Speckled classes. Overall, the achieved accuracy for the LDA case is superior for all runs, outperforming the others by 9% on average. The best results for the LDA are due to the nature of this approach, which uses the samples lables during the feature analysis. On the contrary, the PCA does not consider the classes when applying ortogonal linear transformation to convert the investigated features to principal
5 Conclusion The investigated texture image recognition of cork tiles is considered as unsupervised classication problem, and SOMs are employed for its solution. The proposed approach includes statistical feature preprocessing techniques (for the purposes of dimensionality reduction and dening optimal number of the features used for the classication) and employing SOM as a classier for unsupervised classication (NN architecture and topology design, investigating the complexity of the unsupervised learning and the performance of the SOM). For the purpose of comparison, the experiments and simulations of the system are also conducted using the raw data set without any statistical preprocessing. As expected, better results are obtained for the cases when statistical techniques such as PCA and LDA are used (on average about 92% accuracy
123
rate). When LDA is applied, the trained SOMs achieve very high accuracy rateabove 98%. This can be expected, as LDA is in fact supervised labeling technique, which makes the classication tasks for the subsequently used SOM much easier. The comparison of the sample confusion matrices for the four experiments (Tables 4 and 8) shows that the SOM classiers generally conrm the experts knowledge about the seven types of texture. However, the visual closeness of some of the misclassied samples to samples from other classes could assist experts to rene the classes boundaries or to introduce new classes. Although a straightforward comparison of the methods performance, based only on the accuracy, can be misleading due to the different complexity of the investigated problems (networks topology parameters, training convergence parameters, differences in the preprocessing techniques, and variations in the number of the investigated features and classes, size and quality of the datasets, etc.), it still can give some indication about the method quality. Nevertheless, as compared with results from other authors in the above paragraph, it can be concluded that while our results of 88% mean accuracy for the PCA case, and above 98% for the LDA case, are generally comparable and competitive for most of the cases, they are also superior in some of the comparisons. It is also interesting to note that in our previous paper [12], the achieved results (86% after PCA and 95% after LDA) are inferior to the ones presented here. This can be attributed to the added entropy feature and the feature normalization, applied before the analysis and classication stages, but would need further investigation in a future work.
4.
5.
6.
7.
13.
14.
15.
16.
17.
18.
References
1. Astel A, Tsakouski S, Barbieri S, Simeonov V (2007) Comparison of self-organizing maps classication approach with cluster and principal components analysis for large environmental data sets. Water Res 41:45664578 2. Chamundeeswari VV, Singh D, Singh K (2009) An analysis of texture measures in PCA-based unsupervised classication of SAR images. IEEE Geosci Remote Sens Lett 6:214218 3. Ersoy O, Aydar E, Gourgaud A, Artuner H, Bayhan H (2007) Clustering of volcanic ash arising from different fragmentation 19.
mechanisms using Kohonen self-organizing maps. Comput Geosci 33:821828 Guler I, Demirhan A, Karakis R (2009) Interpretation of MR images using self-organizing maps and knowledge-based expert systems. Digital Signal Process 19:668677 Lei Q, Zheng QF, Jiang SQ, Huang QG, Gao W (2008) Unsupervised texture classication: automatically discover and classify texture patterns. Image Vis Comput 26:647656 Martens G, Poppe C, Lambert P, Van de Walle R (2008) Unsupervised texture segmentation and labeling using biologically inspired features. In: IEEE 10th workshop on multimedia signal processing, vols 1 and 2, pp 163168 Paniagua B, Vega-Rodriguez MA, Gomez-Pulido JA, SanchezPerez JM (2010) Improving the industrial classication of cork stoppers by using image processing and neuro-fuzzy computing. J Intell Manuf 21:745760 Shih FY (2010) Image processing and pattern recognition: fundamentals and techniques. Wiley, Hoboken Umbaugh SE (2010) Digital image processing and analysis. CRC; Taylor & Francis, Boca Raton Bishop CM (2004) Neural networks for pattern recognition. Clarendon Press, Oxford Theodoridis S, Koutroumbas K (2009) Pattern recognition. Elsevier/Academic Press, Amsterdam Georgieva A, Jordanov I (2009) Intelligent visual recognition and classication of cork tiles with neural networks. IEEE Trans Neural Netw 20:675685 Christodoulou CI, Pattichis CS, Pantziaris M, Nicolaides A (2003) Texture-based classication of atherosclerotic carotid plaques. IEEE Trans Medical Imag 22:902912 Kuo CFJ, Kao CY (2007) Self-organizing map network for automatically recognizing color texture fabric nature. Fibers Polym 8:174180 Salah M, Trinder J, Shaker A (2009) Evaluation of the selforganizing map classier for building detection from lidar data and multispectral aerial images. J Spatial Sci 54:1534 Liu H, Yu L (2005) Toward integrating feature selection algorithms for classication and clustering. IEEE Trans Knowl Data Eng 17:491502 Haralick RM, Shanmuga K, Dinstein I (1973) Textural features for image classication. IEEE Trans Syst Man Cybern Smc 3:610621 Kohonen O, Hauta-Kasari M, Parkkinen J, Jaaskelainen T (2006) Co-occurrence matrix and self-organizing map based query from spectral image database. art. no. 603305, ICO20: Illumination, Radiation, and Color Technologies, vol 6033, pp 33053305 Randen T, Husoy JH (1999) Filtering for texture classication: a comparative study. IEEE Trans Pattern Anal Mach Intell 21:291310 Davies ER (2005) Machine vision: theory, algorithms, practicalities. Morgan Kaufmann, Amsterdam Dillon WR, Goldstein M (1984) Multivariate analysis: methods and applications. Wiley, New York Kohonen T (1990) The self-organizing map. Proc IEEE 78: 14641480
123