image processing and image mining using decision trees · image processing and image mining using...

15
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 25, 989-1003 (2009) 989 Image Processing and Image Mining using Decision Trees * KUN-CHE LU AND DON-LIN YANG Department of Information Engineering and Computer Science Feng Chia University Taichung, 407 Taiwan Valuable information can be hidden in images, however, few research discuss data mining on them. In this paper, we propose a general framework based on the decision tree for mining and processing image data. Pixel-wised image features were extracted and transformed into a database-like table which allows various data mining algorithms to make explorations on it. Each tuple of the transformed table has a feature descriptor formed by a set of features in conjunction with the target label of a particular pixel. With the label feature, we can adopt the decision tree induction to realize relationships be- tween attributes and the target label from image pixels, and to construct a model for pixel-wised image processing according to a given training image dataset. Both experi- mental and theoretical analyses were performed in this study. Their results show that the proposed model can be very efficient and effective for image processing and image min- ing. It is anticipated that by using the proposed model, various existing data mining and image processing methods could be worked on together in different ways. Our model can also be used to create new image processing methodologies, refine existing image proc- essing methods, or act as a powerful image filter. Keywords: data mining, pixel classification, low level image processing, image mining, decision tree 1. INTRODUCTION Mining large datasets of different types is a future trend [1]. As computer technolo- gies become more ubiquitous, besides numerical and categorical data, various digitalized images, sounds, voices, and videos have become part of daily life. Yet, although plenty of knowledge can be hidden in these data, very little literature has discussed data mining on them. Image mining concerns the extraction of implicit knowledge, image data relation- ship, or other patterns not explicitly stored in the images [2]. It is more than just an ex- tension of data mining to image domain. Image mining is an interdisciplinary endeavor which draws upon expertise in computer vision, image understanding, data mining, ma- chine learning, database, and artificial intelligence [3]. In [4, 5], the authors have classi- fied the issues of image mining into four classes: 1. association 2. classification 3. sequential pattern 4. time series pattern. Received March 26, 2008; revised June 16, 2008; accepted July 17, 2008. Communicated by H. Y. Mark Liao. * This paper was supported by the National Science Council of Taiwan, R.O.C. under grants No. NSC 96-2218- E-007-007 and NSC 95-2221-E-035-068-MY3.

Upload: others

Post on 26-Mar-2020

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 25, 989-1003 (2009)

989

Image Processing and Image Mining using Decision Trees*

KUN-CHE LU AND DON-LIN YANG

Department of Information Engineering and Computer Science Feng Chia University

Taichung, 407 Taiwan

Valuable information can be hidden in images, however, few research discuss data

mining on them. In this paper, we propose a general framework based on the decision tree for mining and processing image data. Pixel-wised image features were extracted and transformed into a database-like table which allows various data mining algorithms to make explorations on it. Each tuple of the transformed table has a feature descriptor formed by a set of features in conjunction with the target label of a particular pixel. With the label feature, we can adopt the decision tree induction to realize relationships be-tween attributes and the target label from image pixels, and to construct a model for pixel-wised image processing according to a given training image dataset. Both experi-mental and theoretical analyses were performed in this study. Their results show that the proposed model can be very efficient and effective for image processing and image min-ing. It is anticipated that by using the proposed model, various existing data mining and image processing methods could be worked on together in different ways. Our model can also be used to create new image processing methodologies, refine existing image proc-essing methods, or act as a powerful image filter. Keywords: data mining, pixel classification, low level image processing, image mining, decision tree

1. INTRODUCTION

Mining large datasets of different types is a future trend [1]. As computer technolo-gies become more ubiquitous, besides numerical and categorical data, various digitalized images, sounds, voices, and videos have become part of daily life. Yet, although plenty of knowledge can be hidden in these data, very little literature has discussed data mining on them.

Image mining concerns the extraction of implicit knowledge, image data relation-ship, or other patterns not explicitly stored in the images [2]. It is more than just an ex-tension of data mining to image domain. Image mining is an interdisciplinary endeavor which draws upon expertise in computer vision, image understanding, data mining, ma-chine learning, database, and artificial intelligence [3]. In [4, 5], the authors have classi-fied the issues of image mining into four classes: 1. association 2. classification 3. sequential pattern 4. time series pattern.

Received March 26, 2008; revised June 16, 2008; accepted July 17, 2008. Communicated by H. Y. Mark Liao. * This paper was supported by the National Science Council of Taiwan, R.O.C. under grants No. NSC 96-2218-

E-007-007 and NSC 95-2221-E-035-068-MY3.

Page 2: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

KUN-CHE LU AND DON-LIN YANG

990

Vailaya et al. [6] used low-level visual features to capture high-level concepts. They performed a hierarchical classification on vacation images to classify them as indoor or outdoor. Outdoor images were further classified as city or landscape; and a subset of landscape images was classified into sunset, forest, and mountain classes. Cromp et al. [7] examined various classifiers at pixel-level for automatic content extraction of remotely sensed images. No claims on superiority of one classifier over another were made be-cause results depended on the situation. In [8], a new type of texture feature based on association rules is proposed. It is shown that association rules can be adapted to capture frequently occurring low-level structures in images. Other research on developments in image mining such as techniques, particular frameworks, and systems has been sur-veyed in [2].

Image mining is just at its infancy [9]. Most existing data mining techniques have been designed for mining categorical or numerical data and are not well suited for image mining. On the other hand, various image processing techniques such as image segmen-tation, image enhancement, image restoration, or image compression focus on manipula-tion, not on analyzing image data. Thus, although many image processing techniques have been developed, few of them can be adopted to mine image data.

In image retrieval [10, 11], data mining techniques may be used on metadata and/or content of images for indexing an image. Yet, the primary concern of image retrieval is to effectively retrieve similar image objects under a user specified query. On the whole, there has been relatively little concern in extracting useful knowledge and/or processing the indexed images until recently.

Due to most existing technologies that cannot be directly applied for image mining, in this study, we tried to solve the problem by presenting a novel approach based on the decision tree for both image mining and image processing. The decision tree induction is a well-known methodology used widely in various domains, such as artificial intelli-gence, machine learning, data mining, and pattern recognition [12, 13]. It is a predictive model which usually operates as a classifier. The construction of a decision tree, also called decision tree learning process, uses a training dataset consisting of records with their corresponding features and a label attribute to learn the classifier. Once the tree is built, it can be used to predict the label attribute of unidentified records that are from the same domain. A decision tree has the hierarchical form of a tree [14]; an example is shown in Fig. 7.

There are many advantages to the use of decision trees for classification [15]. Deci-sion trees certainly are easy to use and efficient. Rules can be generated that are easy to interpret and understand. They scale well for large databases because the tree size is in-dependent of the database size. Each tuple in the database must be filtered through the tree. This classification takes time proportional to the height of the tree, which is fixed. Moreover, trees can be constructed for data with many attributes.

In our framework, the image used for input is formatted as a set of equally sized raw and label image pairs. These two kinds of images can be considered as the images “before” and “after” an ideal processing or the images “after” and “before” a corruption. With decision tree induction, the proposed model can be used to mine hidden relation-ships between attributes and the target label from image pixels; it can also select impor-tant features. Image processes or image effects can be modeled as understandable deci-sion rules. Moreover, these rules can be applied to perform pixel-wised image processing

Page 3: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

PROCESSING AND MINING IMAGES USING DECISION TREES

991

on new images. Based on different local pixel information, different image processing facilities can be applied easily at different image locations within a single scan of the test image. The built image processing model is efficient and effective, and can be more so-phisticated than other existing image processing methods.

2. PROPOSED MODEL

Fig. 1 depicts a general flow chart of the proposed image mining and image proc-essing framework. Each part will be described in the following subsections.

image transformation

mining results image model

new image

image mining

image processing

training image

Fig. 1. General processing flow of the proposed framework.

2.1 Image Transformation and Feature Extraction

As mentioned, the input data of the proposed model is formatted as a set of equal sized raw and label image pairs. The transformation of the input image dataset into a da-tabase-like table and subsuming of the related features is described in this subsection. For the sake of clarity, various terms used for this process are defined below. In addition, we propose three kinds of input data sources, which will be discussed further in section 4.1. Definition 1 The raw image is a d-dimensional light-intensity function, denoted by R(c1, c2, …, cd), where the amplitude (or value) of R at spatial coordinates (c1, c2, …, cd) gives the intensity of the raw image at that point (or pixel). Definition 2 The label image is a d-dimensional light-intensity function, denoted by L(c1, c2, …, cd), where the value of L at spatial coordinates (c1, c2, …, cd) gives the class identifier of the pixel at same spatial coordinates of its corresponding raw image. Definition 3 The database-like table X = {x1, x2, …, xt} is a set of records, where each record xr ∈ ℜk is a vector with elements <a1, a2, …, ak> being the value of attributes (or features) of X.

Page 4: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

KUN-CHE LU AND DON-LIN YANG

992

7 9 9 9 7

5 7 9 7 5

0 7 9 7 0

5 7 9 7 5

7 9 9 9 7

1 0 0 0 1

1 1 0 1 1

0 1 0 1 0

1 1 0 1 1

1 0 0 0 1

(a) Raw image. (b) Label image. Fig. 2. An example for an atom of the input image dataset.

In this work, only d = 2 is considered, i.e., images with dimensionality of 2. An

example for an atom of the input image dataset is shown in Fig. 2. Each pixel value of the raw image represents the gray level of a pixel. Each pixel value of the label image represents the class label of the pixel. Both pixel values are in the same position.

In this example, the raw image contains the capital English letter “I” with certain degree of blur. Thus, the inside pixels of the letter are darker and the outside pixels are brighter. If a pixel in the label image has the value “1”, the pixel in the same position of the raw image is a pixel of outside contour. It is assumed to be a pixel of interest (POI) in this case. In practice, the pixel value of the label image is not limited to the binary form but could take any kind of form. In addition, we can have as many raw and label image pairs at the same time as required for the input.

In order to mine useful information from a set of raw and label images, we propose a methodology to transform them into a database-like table and allow any data mining algorithms to work on top of the table. This process is simple and straightforward as shown in Fig. 3. Fig. 4 shows a part of the results of this transformation process accord-ing to the data in Fig. 2. Each row of such result table stands for a pixel. Hence its car-dinality (number of rows) equals the number of total pixels in the raw image. In addition, each column of such table represents a feature associated with the given pixels. In Fig. 4,

procedure img2tab(image: raw, label); begin set feature_generated_functions[1..n]; set label_generated_function; initiate table, pixel;

while pixel exists do {pixel scanning process} insert into table value :=

feature_generated1(raw, pixel), …, feature_generatedn(raw, pixel), label_generated(label, pixel);

continue to scan on the next pixel; end while return table;

end Fig. 3. Pseudocode of the image transformation algorithm.

Page 5: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

PROCESSING AND MINING IMAGES USING DECISION TREES

993

feature1 feature2 … featuren label pixel1 7 2 … value1,n 1 pixel2 9 1.25 … … 0 pixel3 9 0 … … 0

… … … … … … pixel4 7 2 … value25,n 1

Fig. 4. Result table of image transformation according to the input in Fig. 2.

feature1 represents the gray level and feature2 the local variation [16]. In order to sim-plify this demonstration, the local variation in this case is replaced with the average dif-ference of a pixel to its 4-neighbors. Other pixel-wised features [17, 18] such as entropy, contrast, mean, etc. can also be encoded into the table as long as they might have affec-tion on the collected dataset.

Various encoding strategies such as normalization (e.g., adjusting the value ranging from 0 to 1) or generalization (e.g., transforming the value to high, medium, or low) can be applied when generating the desired features. Moreover, the label image was included as a column in that table. With the presence of the label feature, hidden relationships between these two kinds of images can be mined.

2.2 Data Reduction

Because of the image characteristics, pixels from a neighboring area will generate similar feature vectors in the transformation process. Under some circumstances, it will cause remarkable redundant information in the result table; for example, an image with a large portion of background. Here we present some basic types of redundancy and show how they can be eliminated while converting the input image set.

Definition 4 The feature scope of a pixel M with spatial coordinates (c1, c2) is an n × n pixel area with center at M, from which all the desired features of M can be generated. Usually n is an odd number, and the sub-image within the feature scope, i.e., pixels within spatial coordinates (c1 ±

12

n − , c2 ± 12 ),n −

is called the root space of the pixel M, denoted as {RSM}.

Definition 5 Two root spaces {RSN}, {RSO} are rotation reachable if {RSN} = {RSO}R, where {.}R stands for a root space after rotating the angle once by 90°, 180°, or 270°.

Definition 6 Two root spaces {RSN}, {RSO} are mirror reachable if {RSN} = {RSO}F, where {.}F stands for a root space after flipping horizontally or vertically.

Given two pixels P and Q at different spatial coordinates of an image I, they are said to be:

1. equivalent redundant, if {RSP} is equal to {RSQ}, 2. rotation redundant, if {RSP} and {RSQ} are rotation reachable, 3. mirror redundant, if {RSP} and {RSQ} are mirror reachable, 4. conflict redundant, if {RSP} and {RSQ} satisfy any one of the first three conditions,

but the label information of pixels P and Q is not equal to each other.

Page 6: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

KUN-CHE LU AND DON-LIN YANG

994

function RR (image: raw, label; pixel: C); begin apply quantization on {RSC} if necessary; if {RSC} can be matched in Ξ do

{redundant pixel} discard {RSC} for further record generation; if the label information of the two matched entries are not equal do {conflict redundant pixel}

update the corresponding information in Ξ; retrieve or update previously generated record if necessary;

else {non-redundant pixel} record all characterized redundancies of {RSC} and the corresponding label information in Ξ;

end Fig. 5. Pseudocode of the redundancy reduction algorithm.

1. {RS1}= ∅  ∅  ∅ ∅  7  9 ∅  5  7 

add <{RS1},1> to Ξ  

2. {RS2}= ∅  ∅  ∅ 7  9  9 5  7  9 

add <{RS2},0> to Ξ  

3. {RS3}=∅ ∅ ∅

9 9 97 9 7

add <{RS3},0> to Ξ  

4. {RS4}=∅ ∅ ∅

9 9 79 7 5

discard <{RS4},0>   ∵{RS4}={RS2}

F & L4=L2

5. {RS5}= ∅  ∅  ∅ 9  7  ∅ 7  5  ∅ 

discard <{RS5},1>   ∵{RS5}={RS1}

F & L5=L1 

6. {RS6}= ∅  7  9 

∅  5  7 ∅  0  7 

add <{RS6},1> to Ξ  

7. {RS7}= 7  9  9 5  7  9 0  7  9 

add <{RS7},1> to Ξ  

8. {RS8}= 9 9 97 9 77 9 7

add <{RS8},0> to Ξ  

9. {RS9}= 9 9 79 7 59 7 0

discard <{RS9},1>   ∵{RS9}={RS7}

F & L9=L7

10. {RS10}= 9  7  ∅ 7  5  ∅ 7  0  ∅ 

discard <{RS10},1>   ∵{RS10}={RS6}

F & L10=L6 

11. {RS11}= ∅  5  7 ∅  0  7 ∅  5  7 

add <{RS11},0> to Ξ  

12. {RS12}= 5  7  9 0  7  9 5  7  9 

add <{RS12},1> to Ξ  

13. {RS13}= 7 9 77 9 77 9 7

add <{RS13},0> to Ξ  

14. {RS14}= 9 7 59 7 09 7 5

discard <{RS14},1>   ∵{RS14}={RS12}

F & L14=L12

15. {RS15}= 7  5  ∅ 7  0  ∅ 7  5  ∅ 

discard <{RS15},0>   ∵{RS15}={RS11}

F & L15=L11 

16. {RS16}= ∅  0  7 ∅  5  7 ∅  7  9 

discard <{RS16},1>   ∵{RS16}={RS6}

F & L16=L6 

17. {RS17}= 0  7  9 5  7  9 7  9  9 

discard <{RS17},1>   ∵{RS17}={RS7}

F & L17=L7

18. {RS18}= 7 9 77 9 79 9 9

discard <{RS18},0>   ∵{RS18}={RS8}

F & L18=L8

19. {RS19}= 9 7 09 7 59 9 7

discard <{RS19},1>   ∵{RS19}={RS7}

R & L19=L7

20. {RS20}= 7  0  ∅ 7  5  ∅ 9  7  ∅ 

discard <{RS20},1>   ∵{RS20}={RS6}

R & L20=L6 

21. {RS21}= ∅  5  7 ∅  7  9 ∅  ∅  ∅ 

discard <{RS21},1>   ∵{RS21}={RS1}

F & L21=L1 

22. {RS22}= 5  7  9 7  9  9 ∅  ∅  ∅ 

discard <{RS22},0>   ∵{RS22}={RS2}

F & L22=L2

23. {RS23}= 7 9 79 9 9∅ ∅ ∅

discard <{RS23},0>   ∵{RS23}={RS3}

F & L23=L3

24. {RS24}= 9 7 59 9 7∅ ∅ ∅

discard <{RS24},0>   ∵{RS24}={RS2}

R & L24=L2

25. {RS25}= 7  5  ∅ 9  7  ∅ ∅  ∅  ∅ 

discard <{RS25},1>   ∵{RS25}={RS1}

R & L25=L1  Fig. 6. Result of the redundancy reduction algorithm according to the images in Fig. 2.

Page 7: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

PROCESSING AND MINING IMAGES USING DECISION TREES

995

Users could characterize other types of redundancy according to the image problem they wish to solve. In order to pinch more redundancies, quantization techniques can be applied on the root space. The pseudocode regarding the function of redundancy reduc-tion is shown in Fig. 5. This function can be added to the pixel scanning process of the image transformation algorithm in Fig. 3. Fig. 6 shows the results of this reduction proc-ess according to the images in Fig. 2. The number of pixels for transformation after re-duction has reduced from 25 to 9. 2.3 Mining Results and their Applications

After having obtained such a database-like table in accordance to the desired input image dataset, mining algorithms can then be used on it. In this study, we have chosen the decision tree for this purpose. An advantage of the decision tree over other method-ologies, such as neural networks, is that it can provide understandable English-like rules or logic statements. For instance, if the gray level of a given pixel ranges between 180 and 240 and its entropy is greater than 0.5, then it is a pixel of interest, POI. This basic idea of simplicity and easy understandability is also the main principle of our approach.

The results of such a mining process may help us to better understand the image properties and relate to real world instances. The results can also be used to process new images of the same domain. Basically, the result of the proposed model is a decision-tree classifier. Fig. 7 depicts a classifier derived from the data shown in Fig. 4 by using CART [15]. A result classifier can be further straightforwardly translated into a set of human- readable if-then rules. For instance, from the three leaf nodes in Fig. 7, we can obtain the following three rules: − If the gray level of a given pixel is less than 8 and its local variation is less than 5,

then it is a pixel of outside contour. − If the gray level of a given pixel is less than 8 and its local variation is greater than or

equal to 5, then it is not a pixel of outside contour. − If the gray level of a given pixel is greater than or equal to 8, then it is not a pixel of

outside contour.

gray level?

local variation?

≥ 8 < 8

< 5 ≥ 5

no

noyes

Fig. 7. A decision tree for the concept is_outside_contour, derived from Fig. 2, indicating whether or not a pixel is a pixel of outside contour.

Page 8: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

KUN-CHE LU AND DON-LIN YANG

996

These rules can provide useful information about the training image. Besides, in order to obtain a higher level of appearance and meet the different information granular-ity requirements, the rules can be post-processed by rule induction algorithms [19]. More prominently, they can be used to process new images from the same domain. The practi-cal image processing capabilities include image restoration, image enhancement, image segmentation, etc. Both experimental and theoretical analyses were performed in this study to examine the proposed model.

The built classifier can also be used to select important features. Features used at higher tree levels for the splitting criteria show a higher significant influence on the pixel class. The selected features can reflect the characteristics of the label image and help design or refine other image processing algorithms.

3. EXPERIMENTS

For our mining process, we used the decision tree package of SAS Enterprise Miner (EM) [20]. First, we performed image restoration with enhancement by using two types of synthetic images, one by Arabic numerals and the other by English alphabet, with added noise artifacts. Afterwards, we performed image segmentation by using the same types of synthetic images with added blur artifacts. In all our experiments, the synthetic images were created with Photoshop by using 24-point Verdana fonts with a smooth boundary. The noise artifact was generated by using the filter, add_noise, with a pa-rameter of uniform distribution and the amount of 100%. Another blur artifact was gen-erated by using filter, Gaussian blur, with a radius of 2.5 pixels.

In the training dataset of Arabic numerals, the distorted and the original images syn-thesized by the numerals “3” to “9” and their two combinations (i.e., 33, 34, …, 98, 99) were used for the raw and label images, respectively. Fig. 8 shows the training dataset of the case with noises. The other numerals (i.e., 0, 1, and 2) were used to synthesize the testing image dataset.

(a) Raw image. (b) Label image.

Fig. 8. Syntactic training image dataset of the noised numerals.

Similarly, in the English alphabet training dataset, the distorted and the original im-

ages synthesized by the letters “F” to “Z” and their two combinations (i.e., FF, FG, …, YZ, ZZ) were used for the raw and label images, respectively. The other letters (i.e., A, B, …, E) were used to synthesize the testing image dataset.

Page 9: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

PROCESSING AND MINING IMAGES USING DECISION TREES

997

For image transformation, a feature scope of size 5 × 5 was used and the selected features included gray level, local variation, mean, local minimum, local maximum, and entropy. The label of a given pixel in the experiments of image restoration with en-hancement was set to its gray level in the label image. We did not apply any encoding strategies mentioned in section 2.1 on the features to simplify the demonstrations. How-ever, in practice, we can use any encoding strategy if required. In the image segmenta-tion experiments, the label feature was transformed to 0 or 1 according to the thresholded label image. In this way the segmentation nature was imitated to distinguish between “background” or “object”.

After we have settled the transformation details, a database-like table can be derived. By applying a classification algorithm on the database-like table, a classifier for label prediction can be obtained. Under the same way, testing images can be transformed into a database-like table to predict the label attributes. These predicted labels can moreover be visualized in a natural form of the input data, i.e., image. As we are proposing a gen-eral image mining and image processing framework and any existing decision tree algo-rithms can be used to do the job, we show only the testing result to simplify the demon-stration. For the other results regarding the constructed classifier or the corresponding rules, if interested, examples can be found in our previous work [21].

(a) Noise numerals case. (b) Noised alphabets case.

Fig. 9. Results of image restoration with enhancement; top row: original images (ground truth); middle row: distorted/testing images; third row: restored/resulting images.

(a) Blurred numerals case. (b) Blurred alphabets case.

Fig. 10. Results of image segmentation; top row: original images (ground truth); middle row: dis-torted/testing images; third row: restored/ resulting images.

Figs. 9 and 10 show the testing images along with the testing results and the ground

truth from the experiments for image restoration with enhancement and image segmenta-tion, respectively. The results in Fig. 9 show that the noised images were successfully restored and enhanced. The results in Fig. 10 show that the main portions of the image object were successfully segmented.

Fig. 11 compares the testing results of the proposed method and the thresholding technique. The first row in each of the examples were obtained by the proposed method and the others by the thresholding with different thresholds. The numerical testing results in Fig. 11 (a) were obtained by using the alphabet training dataset, which were better

Page 10: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

KUN-CHE LU AND DON-LIN YANG

998

t=64

t=32

t=1

t=170

t=190

t=210

t=160

t=170

t=190

(a) Numbers with noise. (b) Blurred numbers. (c) Blurred alphabets. Fig. 11. Comparison between the testing results of the proposed method and the thresholding tech-

nique. First row: result from the proposed method; other rows: results from thresholding using the thresholds annotated to the right.

(a) Blur phenomenon. (b) Noise phenomenon. Fig. 12. Real test image data acquired from a scanner.

than those obtained by using the numerical training dataset shown in Fig. 9. The reason is that there were only 56 training samples in the numerical training case, while the Eng-lish alphabet training case had 462 samples. The results in Fig. 11 show that our method can easily produce better results than the thresholding.

We also assessed the built model by using blurred and noised images acquired from a scanner. Fig. 12 shows parts of these two kinds of images that were used in our ex-periment. The blur phenomenon may be caused by the gutter and/or movement of the scanning material. The noise phenomenon may be caused by scanning crumpled and/or grained paper.

In this experiment, both the English alphabet and Arabic numerals were used to synthesize the training image, where synthetic blurred and noised phenomena were add-ed jointly to imitate real cases. The built image processing model was applied to both kinds of the real test images. OCR software was adopted to recognize the characters, and the numbers of characters erroneously recognized were used to evaluate the quality of the resulting images. We also compared the built model with the thresholding technique, where the threshold value was set automatically to 128. The results in Fig. 13 show that the built model is much more effective.

4. THEORETICAL ANALYSES AND DISCUSSIONS

4.1 The Input Data Source

Various image processing capabilities of the proposed model can be achieved by manipulating the input data source. As previously mentioned, there are three different sources for the label image:

Page 11: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

PROCESSING AND MINING IMAGES USING DECISION TREES

999

errors on

blurred part (361)

errors on normal part

(6234)

total errors on blurred image

(6595)

original 87 15 102

thresholding 55 12 67

built model 40 14 55 (a) Recognition results for the blurred image.

errors on

noised image (6710)

original 6710 (N/A)

thresholding 683

built model 48 (b) Recognition results for the

noised image. Fig. 13. Numbers of characters erroneously recognized for the original scanned image, the thresh-

olded image, and the image processed by the built model. Numbers in parenthesis show the amount of characters in the image.

4.1.1 User defined

With the advent of computer technology, more and more images will be presented in an electronic form. Hence, we will need more advanced methods to deal with these images as they are becoming an indispensable part of our daily life. Users would expect to obtain the effect according to their specification after processing an image. However, suitable processing methods may not always be available. In that case, the users or do-main experts can artificially manipulate the label image based on their requirements. By applying our approach, a desired image processing method can be acquired. Moreover, the mined information and selected features may help better understand the properties of the desired effects. This understanding also helps to design or refine other related image processing algorithms. 4.1.2 From other algorithms

If certain algorithms have been developed to solve particular image processing problems, users can use the outcomes they already obtained as the label image. As a result, the built model can process images similarly to or even better than the original method. In this manner, the proposed model is able to mine, learn, and explain the decision proc-ess of the original method. This explanation capability is particular helpful when the original method is more or less a black box for humans, such as a neural network.

On the other hand, users may include all the features that might have affection on the label image in the database-like table and leave the selecting process to the decision tree. The learning process of the decision tree is comprised of the following two proc-esses that are necessary when building image interpretation systems: feature selection and learning of the classifier [22]. This objective feature selection method can pick up more relevant features that are related to the label feature compared to the subjective feature selection methods. The mined information and selected features may help users better understand the studied properties and/or refine the original method. 4.1.3 Hybrid

The training data for the hybrid method are obtained through exploiting the out-come of other algorithms and additional modifications made by the user. In this way, the

Page 12: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

KUN-CHE LU AND DON-LIN YANG

1000

flaw of the original method may be overcome more easily. Furthermore, our expertise can be embedded in the new model, and thus make it more sophisticated.

4.2 Efficiency and Effectiveness

Once the image model has been trained, only one scan of the image dataset is re-quired to perform processing or testing on the new images. The processing time is pro-portional to the height of the result decision tree, which is fixed and independent of the original data size. The built decision tree model can also generate an efficient and com-pact code [23]. With these efficiency characteristics, our approach can be used to relieve the time-consuming problem of other image processing methods, such as neural net-works. Two ways to utilize the proposed approach are possible. In the first one, the original method is superseded with our method by using the strategies mentioned in sec-tion 4.1. The second one is to use the built model as an efficient filter to single out those images that need extra processing. For example, after singling out suspicious mammo-grams that might contain pixels of cancer [24], the original method can be applied for further diagnosis.

The first option is suitable for the case when the generated method, by applying our approach, is able to process the image better than the original method. Otherwise, a slight loss of accuracy does not make a significant difference to the users [22]. The second op-tion is suitable for the case when the processing result has a critical usage, and the gen-erated method is unable to fulfill this requirement. When the proposed approach is used as an efficient filter, to ensure the desired types of defects can be pinpointed, the mis-classification penalty for different classes can be adjusted in the decision tree training process. The capability to control a decision tree is a big advantage which can help ma-nipulate the classification process or even interactively train our model [20]. With these abilities, the well-known classification problem, over-fitting, can be handled more easily.

4.3 Extensibility and Flexibility

Many image processing methods possibly involve the use of spatial filters (or masks). Various kinds of masks can be integrated in the built model at the pixel scanning process of the image transformation procedure. After application of a mask on the train-ing images, the resulting gray level as well as the features of a given pixel can also be included as a column within the database-like table. Thus, the proposed model is able to mine across different layers of the image, and determine important features over dispar-ity of the masks as shown in Fig. 14. Apart from image masks, various image processing methods can also be incorporated into the model. Based on different local pixel informa-tion, suitable image processing capabilities can be easily derived via the different rules invoked from a built model. The proposed model can also be easily extended to handle 3D, 4D, or even higher dimensionality of image without the need of a big revolution.

5. CONCLUSION AND FUTURE WORK

In this paper, a novel, efficient, and effective model was proposed. With our ap-proach, data mining and image processing technologies can be fused. The results have

Page 13: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

PROCESSING AND MINING IMAGES USING DECISION TREES

1001

Raw Image

Label Image

Image A

Image B

Image C

mask M

mask N

mask M and O image convertingprocess

label fA.1 … fA.m fB.1 … fB.m fC.1 … fC.m

pixel1 v1,L v1,A.1 … v1,A.m v1,B.1 … v1,B.m v1,C.1 … v1,C.m

pixel2 v2,L v2,A.1 … v2,A.m v2,B.1 … v2,B.m v2,C.1 … v2,C.m

pixel3 v3,L v3,A.1 … v3,A.m v3,B.1 … v3,B.m v3,C.1 … v3,C.m

… … … … … … … … … … …

pixeln vn,L vn,A.1 … vn,A.m vn,B.1 … vn,B.m vn,C.1 … vn,C.m

Fig. 14. Concept of mining across different layers (A, B, C, …) of the image, and determining im-

portant features (1, …, m) over disparity of the masks (M, N, O, …).

shown clearly that the proposed model can mine and process image data well. The ex-periments have demonstrated the adaptability of the proposed model for various kinds of image processing problems, including image enhancement, image restoration, and image segmentation. In addition, the benefits resulting from application of the proposed model have been presented.

The proposed model requires label information of image pixels in advance; however, in some situations this label information may be unavailable or undetermined. Some-times, it might be necessary to find the actual hidden label properties, and our future work is to refine the proposed model to an unsupervised one, which can automatically analyze and determine the label information for further use. On the other hand, we are trying to tailor this general framework to a particular case. The specialization of the pro-posed model will involve more issues such as generation of raw image features, trans-formation of label image properties, integration of different masks, and cooperation of existing methods.

REFERENCES

1. U. Fayyad, D. Haussler, and P. Stolorz, “Mining scientific data,” Communications of the ACM, Vol. 39, 1996, pp. 51-57.

2. J. Zhang, W. Hsu, and M. L. Lee, “Image mining: Issues, frameworks and tech-niques,” in Proceedings of the 2nd International Workshop Multimedia Data Mining, 2001, pp. 13-20.

3. M. C. Burl, “Mining large image collections,” in R. Grossman, C. Kamath, V. Kumar, and R. Namburu, eds., Data Mining for Scientific and Engineering Applications, Kluwer Academic Publishers, New York, 2001, pp. 63-84.

4. C. Ordonez and E. Omiecinski, “Image mining: A new approach for data mining,” Technical Report GIT-CC-98-12, College of Computing, Georgia Institute of Tech-nology, 1998.

5. C. Ordonez and E. Omiecinski, “Discovering association rules based on image con-tent,” in Proceedings of IEEE Forum Research and Technology Advances in Digital Libraries, 1999, pp. 38-49.

6. A. Vailaya, A. T. Figueiredo, A. K. Jain, and H. J. Zhang, “Image classification for content-based indexing,” IEEE Transactions on Image Processing, Vol. 10, 2001, pp. 117-130.

7. R. F. Cromp and W. J. Cambell, “Data mining of multidimensional remotely sensed images,” in Proceedings of the 2nd International Conference on Information and

Page 14: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

KUN-CHE LU AND DON-LIN YANG

1002

Knowledge Management, 1993, pp. 471-480. 8. J. A. Rushing, H. S. Ranganath, T. H. Hinke, and S. J. Graves, “Using association

rules as texture features,” IEEE Transactions on Pattern Analysis and Machine In-telligence, Vol. 23, 2001, pp. 845-858.

9. K. C. Lu, D. L. Yang, and M. C. Hung, “Decision trees based image data mining and its application on image segmentation,” in Proceedings of International Conference on Chinese Language Computing, 2002, pp. 81-86.

10. K. M. Lee and W. N. Street, “Cluster-driven refinement for content-based digital image retrieval,” IEEE Transactions on Multimedia, Vol. 6, 2004, pp. 817-827.

11. A. M. W. Smeulders, M. Worring, S. Santini, A. Gupta, and R. Jain, “Content-based image retrieval at the end of the early years,” IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 22, 2000, pp. 1349-1380.

12. J. C. Shafer, R. Agrawal, and M. Mehta, “SPRINT: A scalable parallel classifier for data mining,” in Proceedings of the 22nd International Conference on Very Large Databases, 1996, pp. 544-555.

13. M. Mehta, R. Agrawal, and J. Rissanen, “SLIQ: A fast scalable classifier for data mining,” in Proceedings of International Conference on Extending Database Tech-nology, 1996, pp. 18-32.

14. E. A. Leonidaki, D. P. Georgiadis, and N. D. Hatziargyriou, “Decision trees for de-termination of optimal location and rate of series compensation to increase power system loading margin,” IEEE Transactions on Power Systems, Vol. 21, 2006, pp. 1303-1310.

15. M. H. Dunham, Data Mining Introductory and Advanced Topics, Prentice Hall, New Jersey, 2003.

16. R. C. Gonzalez and R. C. Woods, Digital Image Processing, Addison-Wesley, New York, 1992.

17. R. W. Conners and C. A. Harlow, “A theoretical comparison of texture algorithms,” IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 2, 1980, pp. 204-222.

18. R. M. Haralick, K. Shanmugam, and I. Dinstein, “Textural features for image classi-fication,” IEEE Transactions on Systems, Man, and Cybernetics, Vol. 3, 1973, pp. 610-621.

19. S. M. Weiss and N. Indurkhya, “Optimized rule induction,” Expert, IEEE [see also IEEE Intelligent Systems and Their Applications], Vol. 8, 1993, pp. 61-69.

20. SAS/STAT Version 8.1, SAS Enterprise Miner Version 4.0 System Help, SAS Insti-tute Inc., 2000.

21. K. C. Lu, Image Processing and Image Mining Using Decision Trees: Fusing Data Mining and Image Processing, a publication of master thesis by VDM Verlag, 2008.

22. P. Perner, U. Zscherpel, and C. Jacobsen, “A comparison between neural networks and decision trees based on data from industrial radiographic testing,” Pattern Rec-ognition Letters, Vol. 22, 2001, pp. 47-54.

23. L. Robert and G. Malandain, “Fast binary image processing using binary decision diagrams,” in Proceedings of IEEE Computer Society Conference on Computer Vi-sion and Pattern Recognition, 1997, pp. 97-102.

24. H. W. Chen, Y. C. Ouyang, S. K. Lee, P. C. Chung, and C. S. Lo, “Application of tele-consultation system in mammograms,” in Proceedings of the 18th Conference on Technology and Applications to Chinese and Western Medicine, 2000, pp. 39-40.

Page 15: Image Processing and Image Mining using Decision Trees · Image Processing and Image Mining using Decision Trees* KUN-CHE LU AND DON-LIN YANG Department of Information Engineering

PROCESSING AND MINING IMAGES USING DECISION TREES

1003

Kun-Che Lu (呂坤哲) received a B.E. degree and an M.S. degree in Computer Science from Feng Chia University, Taiwan, in 2000 and 2002, respectively. Mr. Lu is currently a Ph.D. stu-dent at the Department of Information Engineering and Com-puter Science, Feng Chia University. His research interests in-clude data mining, image processing, and databases.

Don-Lin Yang (楊東麟) received a B.E. degree in Com-puter Science from Feng Chia University, Taiwan, in 1973, an M.S. degree in Applied Science from the College of William and Mary in 1979, and a Ph.D. degree in Computer Science from the University of Virginia in 1985. He worked at IBM Santa Teresa Laboratory from 1985 to 1987 and at AT&T Bell Laboratories from 1987 to 1991. Since then, he joined the faculty of Feng Chia University and is currently a professor in the Department of In-formation Engineering and Computer Science. His research in-terests include image processing, data mining, and software en- gineering. Dr. Yang is a member of the IEEE Computer Society and the ACM.