rough set and tabu search based feature selection for credit scoring

8
Procedia Computer Science International Conference on Computational Science, ICCS 2010 Rough set and Tabu search based feature selection for credit scoring Jue Wang a,, Kun Guo b , Shouyang Wang a a Academy of Mathematics and Systems Science, Chinese Academy of Sciences, 100190, China b Research Center on Fictitious Economy and Data Science, Chinese Academy of Sciences, 100190, China Abstract As the credit industry has been growing rapidly, huge number of consumers’ credit data are collected by the credit department of the bank and credit scoring has become a very important issue. Usually, a large amount of redundant information and features are involved in the credit dataset, which leads to lower accuracy and higher complexity of the credit scoring model. So, eective feature selection methods are necessary for credit dataset with huge number of features. In this paper, a novel approach, called FSRT, to feature selection based on rough set and tabu search is pro- posed. In FSRT, conditional entropy is regarded as the heuristic to search the optimal solutions. The proposed method is introduced to credit scoring and Japan credit dataset in UCI database is selected to demonstrate the competitive performance of the proposed method. Moreover, FSRT shows a superior performance in saving the computational costs and improving classification accuracy. Keywords: Feature selection, Rough set, Tabu search, Credit scoring 1. Introduction The credit industry has experienced two decades of rapid growth with significant increases in auto-financing, credit card debt, and so on. Credit scoring models have been widely studied in the areas of statistics, machine learning, and artificial intelligence(AI) [1, 2, 3]. The advantages of credit scoring include reducing the cost of credit analysis, enabling faster credit decisions, closer monitoring of existing accounts, and prioritizing collections [4, 5]. With the growth of the credit industry and the large loan portfolios under management today, credit industry is actively developing more accurate credit scoring models. However, credit scoring datasets containing huge number of features are often involved. In such cases, feature selection will be necessary for significantly improving the accuracy of the credit scoring models [6, 7, 8]. This eort is leading to the investigation of eective approach to feature selection for credit scoring applications [9, 10, 11]. Due to the abundance of noisy, irrelevant or misleading features, the ability to handle imprecise and inconsistent information in real world problems has become one of the most important requirements for feature selection. Rough sets theory, proposed by Pawlak, is a novel mathematic tool handling uncertainty and vagueness, and inconsistent data [12, 13, 14]. The rough set approach to feature selection is to select a subset of features (or attributes), which Corresponding author. Email: [email protected] c 2010 Published by Elsevier Ltd. Procedia Computer Science 1 (2010) 2425–2432 www.elsevier.com/locate/procedia 1877-0509 c 2010 Published by Elsevier Ltd. doi:10.1016/j.procs.2010.04.273

Upload: jue-wang

Post on 26-Jun-2016

213 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Rough set and Tabu search based feature selection for credit scoring

Procedia Computer Science 00 (2010) 1–8

Procedia ComputerScience

International Conference on Computational Science, ICCS 2010

Rough set and Tabu search based feature selectionfor credit scoring

Jue Wanga,∗, Kun Guob, Shouyang Wanga

aAcademy of Mathematics and Systems Science, Chinese Academy of Sciences, 100190, ChinabResearch Center on Fictitious Economy and Data Science, Chinese Academy of Sciences, 100190, China

Abstract

As the credit industry has been growing rapidly, huge number of consumers’ credit data are collected by the creditdepartment of the bank and credit scoring has become a very important issue. Usually, a large amount of redundantinformation and features are involved in the credit dataset, which leads to lower accuracy and higher complexity ofthe credit scoring model. So, effective feature selection methods are necessary for credit dataset with huge number offeatures. In this paper, a novel approach, called FSRT, to feature selection based on rough set and tabu search is pro-posed. In FSRT, conditional entropy is regarded as the heuristic to search the optimal solutions. The proposed methodis introduced to credit scoring and Japan credit dataset in UCI database is selected to demonstrate the competitiveperformance of the proposed method. Moreover, FSRT shows a superior performance in saving the computationalcosts and improving classification accuracy.

Keywords: Feature selection, Rough set, Tabu search, Credit scoring

1. Introduction

The credit industry has experienced two decades of rapid growth with significant increases in auto-financing,credit card debt, and so on. Credit scoring models have been widely studied in the areas of statistics, machinelearning, and artificial intelligence(AI) [1, 2, 3]. The advantages of credit scoring include reducing the cost of creditanalysis, enabling faster credit decisions, closer monitoring of existing accounts, and prioritizing collections [4, 5].With the growth of the credit industry and the large loan portfolios under management today, credit industry is activelydeveloping more accurate credit scoring models. However, credit scoring datasets containing huge number of featuresare often involved. In such cases, feature selection will be necessary for significantly improving the accuracy of thecredit scoring models [6, 7, 8]. This effort is leading to the investigation of effective approach to feature selection forcredit scoring applications [9, 10, 11].

Due to the abundance of noisy, irrelevant or misleading features, the ability to handle imprecise and inconsistentinformation in real world problems has become one of the most important requirements for feature selection. Roughsets theory, proposed by Pawlak, is a novel mathematic tool handling uncertainty and vagueness, and inconsistentdata [12, 13, 14]. The rough set approach to feature selection is to select a subset of features (or attributes), which

∗Corresponding author. Email: [email protected]

c© 2010 Published by Elsevier Ltd.

Procedia Computer Science 1 (2010) 2425–2432

www.elsevier.com/locate/procedia

1877-0509 c© 2010 Published by Elsevier Ltd.doi:10.1016/j.procs.2010.04.273

Page 2: Rough set and Tabu search based feature selection for credit scoring

Jue Wang et al. / Procedia Computer Science 00 (2010) 1–8 2

can predict or classify the decision concepts as well as the original feature set [15]. Obviously, feature selection is anattribute subset selection process, where the selected attribute subset not only retains the representational power, butalso has minimal redundancy.

In this paper, a novel method of feature selection based on rough sets and tabu search(FSRT) is proposed forcredit scoring data. Tabu Search (TS), originally proposed by Glover in 1986, is a heuristic method which has shownsuccessful performance in solving many combinatorial search problems in some operations research literatures [16,17, 18]. FSRT uses a binary representation of solutions in the process of feature selection. The conditional entropyis invoked to measure the quality of solution and regarded as a heuristic. TS neighborhood search methodology hasbeen intensified in this method, i.e., avoiding be back to a recently visited solution, and adopting downhill movesto escape from local maximum information. Some search history information is reserved to help the search processto behave more intelligently [19, 20, 21]. The numerical results indicate that the proposed method shows a superiorperformance in saving the computational costs and improving the accuracy of several classification models.

The rest of the paper is organized as follows. In the next section, we briefly give the principles of rough set. InSection 3, we highlight the main components of FSRT and present the algorithm formally. In Section 4, numericalresults of FSRT are illustrated, and the classification results are presented comparing with the models without featureselection. Finally, the conclusion makes up Section 5.

2. Rough sets Preliminaries

An information system is a formal representation of a dataset to be analyzed, and it is defined as a pair S = (U,A),where U is a non-empty set of finite objects, called the universe of discourse, and A is a non-empty set of attributes.With every attribute a ∈ A, a set of its values Va is associated. In practice, we are mostly interested in dealing with aspecial case of information system called a decision system. It consists of a pair S = (U,C ∪ D), where C is called aconditional attributes set and D a decision attributes set.

The RS theory is based on the observation that objects may be indiscernible (indistinguishable) because of limitedavailable information. For a subset of attributes P ⊆ A , the indiscernibility relation is defined by IND(P):

IND(P) = {(ξ, η) ∈ U × U |∀a ∈ P, a(ξ) = a(η)}.It is easily shown that IND(P) is an equivalence relation on the set U. The relation IND(P), P ⊆ A, constitutes apartition of U, which is denoted U/IND(P) [13]. If (ξ, η) ∈ IND(P), then ξ and η are indiscernible by attributesfrom P. The equivalence classes of the P-indiscernibility relation are denoted [ξ]P. For a subset Ξ ⊆ U, the P-lowerapproximation of Ξ can be defined as follows,

PΞ = {ξ|[ξ]P ⊆ Ξ}.

For an information system S = (U,A) and a partition of U with classes Xi,1 ≤ i ≤ n, the entropy of attribute setB ⊆ A is defined by:

H(B) = −n∑

i=1

(p(Xi))log(p(Xi))

where IND(B) = {X1, X2, . . . , Xn}, p(Xi) = |Xi|/|U |, and |.| is the cardinality.The conditional entropy of an attribute set D with reference to another attribute set B is defined as follows:

H(D|B) = −n∑

i=1

(p(xi))m∑

j=1

(p(Yj|Xi))log(p(Yj|Xi))

where p(Yj|Xi) = |Yj ∩ Xi|/|Xi|,and U/IND(D) = {Y1,Y2, . . . ,Ym}, U/IND(B) = {X1, X2, . . . , Xn}, 1 ≤ i ≤ n, 1 ≤ j ≤ m.And some theorems have been proved in some literatures [22, 23].

Theorem 2.1. H(D|B) = H(D ∪ B)) − H(B)

2426 J. Wang et al. / Procedia Computer Science 1 (2010) 2425–2432

Page 3: Rough set and Tabu search based feature selection for credit scoring

Jue Wang et al. / Procedia Computer Science 00 (2010) 1–8 3

Theorem 2.2. If P and Q are the attribute sets of an information system S = (U, A) and IND(Q) = IND(P), thenH(Q) = H(P).

Theorem 2.3. If P and Q are the attribute sets of an information system S = (U, A), P ⊆ Q and H(Q) = H(P), thenIND(Q) = IND(P).

Theorem 2.4. An attribute r in an attribute set R ⊆ C is reducible if and only if H(r|R − {r}0 = 0.

Theorem 2.5. Given a relatively consistent decision table S = (U,C ∪D), an attribute set R is relatively independentif and only if H(D|R) = H(D|R − r) for every r ∈ R.

Theorem 2.6. Given a relatively consistent decision table S = (U,C ∪ D), an attribute set R ⊆ C ⊆ A of aninformation system S = (U, A) is a reduct of B if and only if

(1)H(R) = H(C)(2)The attribute set R is relatively independent.

One of the major applications of rough set theory is the attribute reduction, that is, the elimination of attributesconsidered to be redundant, while avoiding information loss [13, 14]. The reduction of attributes is achieved bycomparing equivalence relations generated by sets of attributes. Using the conditional entropy as a measure, attributesare removed so that the reduced set provides the same dependency degree as the original. In a decision system, areduct is formally defined as a subset R of the conditional attribute set C such that H(R) = H(C), where D is thedecision attributes set. A given dataset may have many reducts. Thus the set R of all reducts is defined as [13]:

R = {R : R ⊆ C; H(R) = H(C)}.The intersection of all the sets in R is called the core,

Core(R) =⋂

R∈RR.

The elements of the core are those attributes that cannot be eliminated without introducing more contradictions to thedataset. In the process of attribute reduction, a set Rmin ⊆ R of reducts with minimum cardinality is searched for:

Rmin = {R ∈ R : |R| ≤ |S |,∀S ∈ R}.which is called the minimal reduct set.

It is well known that finding a minimal reduct is NP-hard [13]. For most of the applications, only one minimalreduct is required, so all the calculations involved in discovering the rest are pointless. Therefore, an alternativestrategy is required for large datasets.

3. Credit scoring based on feature selection

In this section, the proposed feature selection method for credit scoring is presented, which is based on rough setand tabu search. Three credit scoring models and several performance measures are involved in the illustration of theproposed method.

3.1. Feature Selection based on Rough Set and Tabu SearchFSRT starts with an initial solution x0 and uses the tabu restriction rule in generating trial solutions in a neighbor-

hood of the current iterate. FSRT continues generating trial solutions until no improvement can be obtained throughconsecutive iterations. Then, FSRT proceeds to the search process starting from a diverse solution xdiv. If the numberof these consecutive iterations without improvement exceeds Ishak the number of non-improvement to apply shaking,FSRT invokes the search proceeds to improve the best reduct xbest obtained so far, if it exists. Then, the search iscontinued starting from xbest. The search may be terminated if the number of iterations exceeds the maximal num-ber of iterations Imax, or the number of consecutive iterations without improvement exceeds the maximal number ofnon-improvement iterations Iimp. Finally, Elite Reducts Inspiration is applied as a final diversification-intensificationmechanism. The FSRT formal algorithm is stated in the following and the detail of the algorithm can be found inliterature [24].

J. Wang et al. / Procedia Computer Science 1 (2010) 2425–2432 2427

Page 4: Rough set and Tabu search based feature selection for credit scoring

Jue Wang et al. / Procedia Computer Science 00 (2010) 1–8 4

Algorithm 3.1. FSRT

1. Let the Tabu List (TL) contain the two extreme solutions (0, . . . , 0) and (1, . . . , 1), set vF to be a zerovector, and set RedSet to be an empty set. Choose an initial solution x0, and set the counter k := 0.Select Imax, Iimp, Ishak and Idiv such that Imax > Iimp > Ishak > Idiv.

2. Generate neighborhood trials {y1, . . . , y�} around xk.

3. Set xk+1 equal to the best trial solution from y1, . . . , y�, and update TL, vF, xbest and RedSet. Setk := k + 1.

4. If the number of iterations exceeds Imax, or the number of iterations without improvement exceeds Iimp,go to Step 7.

5. If the number of iterations without improvement exceeds Ishak, apply shaking strategy to improve xbest,set xk := xbest and go to Step 2.

6. If the number of iterations without improvement exceeds Idiv, apply diversification strategy to obtain anew diverse solution xdiv, set xk := xdiv, and go to Step 2.

7. Apply Elite Reducts Inspiration to obtain xERI. Update xbest by xERI if the latter is better, andterminate.

3.2. credit scoring based on FSRT

This section investigates the credit scoring accuracy of three credit scoring models without feature selection andwith feature selection based on rough set and tabu search, including radial basis function(RBF), support vector ma-chine(SVM) and logistic regression. They use different algorithms to estimate the unknown credit scoring functionand employ different training methods to acquire information from the credit scoring examples available. Artificialneural networks has been widely used in credit scoring problems, and it has been reported that its accuracy is superiorto the traditional statistical methods [25]. RBF is the most frequently used neural network architecture in commercialapplications including credit scoring [26]. Support vector machines (SVM) have been applied successfully in manyclassification problems including credit scoring [27]. Logistic regression model is one of the most popular statisticaltools for classification problems. Logistic regression model, unlike other statistical tools (e.g. discriminant analysisor ordinary linear regression), can suit various kinds of distribution functions such as Gamble, Poisson, normal, etc.and is more suitable for the credit scoring problems [28, 29].

Depending on the credit features selected and the parameters setting of models, the estimation of the credit scor-ing function may result in different classification results. The credit scoring models are tested using 10-fold crossvalidation with a real world data set- Japanese credit data.

3.3. Performance measure

In addition to prediction accuracy, ROC curve is used to measure the performance of the models [30].Given a classifier and an instance, there are four possible outcomes.If the instance is positive and it is classified as

positive, it is counted as a true positive(TP); if the instance is negative and it is classified as positive, it is counted as afalse positive(FP). The tp rate(true positive rate), fp rate(false positive rate),Precision and F-measure are expressed as:

tp rate =T PP

; f p rate =FPN

; Precision =T P

T P + FP; Recall =

T PP

F − measure =2

1/precision + 1/recall

where P is the total positives and N is the total negatives.The ROC is a two-dimensional graph in which true positive rate is plotted on the Y-axis and false positive rate is

plotted on the X-axis, as illustrated in Figure 1. AUC is the area under the ROC curve. Generally, a model with alarger AUC will have a better average performance.

2428 J. Wang et al. / Procedia Computer Science 1 (2010) 2425–2432

Page 5: Rough set and Tabu search based feature selection for credit scoring

Jue Wang et al. / Procedia Computer Science 00 (2010) 1–8 5

Table 1: FSRT Parameter Setting

Parameter Definition Value|TL| size of tabu list 5� size of neighborhood round( |C|2 )nR num. of the best reducts xERI 3Imax max num. of iterations 100Iimp max num. of non-improvement iterations 40Idiv num. of non-improvement to apply Diverse 20Ishak num. of non-improvement to apply shaking 30

4. Numerical Experiments

This research investigates the potential for feature selection with Japan credit dataset, which is available from theUCI Repository of Machine Learning Databases.

4.1. Data set description

Japan credit dataset derived from various applications is used to evaluate the performance of the proposed algo-rithm. It is made public from the UCI Repository of Machine Learning Databases, and is mostly used to compare theperformance of various classification models. The dataset consists of 299 examples of creditworthy applicants and365 examples where credit should not be extended. For each applicant, 15 variables describe job status, applicant’sgender, marital status, house area, applicant’s age, saving, monthly loan repayment, No. of months to repay and soon.

4.2. Parameters setting

For Japan credit dataset, the FSRT MATLAB code was run 20 times with different initial solutions. Beforediscussing these results, we explain the setting of the FSRT parameters. In Table 1, we summarize all parametersused in the FSRT algorithm with their assigned values. These chosen values are based on the common setting in theliterature or based on our numerical experiments.

The performance of FSRT was tested using different values of these parameters. First, the tabu list size |TL|was set to 5 to save the three most recently visited solutions in addition to the all-zeros and all-ones solutions. Thenumber � of neighborhood trial solutions was set equal to round( |C|2 ). Since l depends on the problem size, it mayhelp avoid a deterioration of the method when the problem size increases. The numbers Idiv and Ishak of consecutivenon-improvement iterations before applying diversification or shaking are set equal to 20 and 30, respectively. Thenumber nR of best reducts is set equal to 3, which helps FSRT to improve the best reduct found so far if it was not aglobal one. Finally, the maximum numbers Imax and Iimp of iterations and of consecutive non-improvement iterations,respectively, are set as in Table 1. The numerical experiments showed that these setting were reasonable to avoidspending more non-improvement iterations, and sufficient to let the search process explore the current region beforesupporting it with diversification or intensification.

4.3. Feature selection and credit scoring

This section provides a discussion of the experimental procedure and the result of the experiment. To evaluatethe performance of feature selection and the classification accuracy of credit scoring model, the dataset was usuallyrandomly partitioned into training sets and independent test sets via a k-fold cross validation. Each of the k subsetsacted as an independent holdout test test for the model trained with the remaining k − 1 subsets. The advantage ofcross validation are that all of the test sets were independent and the reliability of the results could be improved. Atypical experiment uses k = 10, then we divide each credit data into 10 subsets for cross validation and carried out ourimplementation on 10 training data sets.

J. Wang et al. / Procedia Computer Science 1 (2010) 2425–2432 2429

Page 6: Rough set and Tabu search based feature selection for credit scoring

Jue Wang et al. / Procedia Computer Science 00 (2010) 1–8 6

Table 2: Feature selection results for Japan credit data sets

Japan credit dataset

Fold12345678910

Average

FRST result FSRT average7(8)8(6)9(6) 7.97(8)8(12) 7.67(10)8(8)9(2) 7.67(4)8(16) 7.87(2)8(16)9(2) 87(4)8(12)9(4) 87(18)8(2) 7.17(8)8(12) 7.67(14)8(2)9(4) 7.57(4)8(8)9(8) 8.2

7.73

Table 3: Performance comparison for different models in Japan dataset

Model Feature Selection TP rate FP rate Precision Recall F-measure AUC

RBFNO 0.798 0.226 0.807 0.798 0.794 0.867YES 0.816 0.19 0.816 0.816 0.816 0.888

SVMNO 0.861 0.126 0.872 0.861 0.862 0.868YES 0.861 0.126 0.872 0.861 0.862 0.868

LogisticNO 0.854 0.144 0.856 0.854 0.854 0.905YES 0.864 0.131 0.867 0.864 0.865 0.921

In Table 2, the result of feature selection for Japan credit dataset is illustrated, which will be contribute to decreasethe complexity of classification and improve the accuracy of credit scoring model. In the following experiment, afeature subset with minimum number of features is selected to the credit scoring models.

Three strategies were used in this study, namely RBF, SVM and Logistic. A comparison was made between usingall features with no pre-selection and using the features pre-selected by FSRT. The result for Japan credit dataset wasobtained and summarized in Tables 3.

From Table 3, we can see that there is no significant difference in performance between SVM classifiers withfeature selection and the baseline SVM classifier(all features). There is however a marginal difference in RBF and Lo-gistic classifiers with feature selection and the baseline classifiers. For Japan credit dataset, the average classificationaccuracy of both models without feature selection is 80.7% and 85.6%, respectively. An obvious improvement overbaseline classifiers has happened after feature selection. The precision of both models achieved 81.6% and 86.6%.There is also a higher true positive rate with a lower false positive rate after FSRT.

The performance of all models for Japan dataset are showed in the ROC space in Figure 1. Informally, one pointin ROC space is better than another if it is to the northwest where tp rate is higher while fp rate is lower. The points inFigure 1 show that the Logistic classifier with FSRT has a best performance as the point (FS-logistic) appearing in thenorthwest corner of the ROC space. Furthermore, the average performance of the classifiers with feature selection isbetter than the baseline classifiers. Moreover, the model achieved a higher AUC area after feature selection, take theRBF model for example, the ROC curve and AUC area of the RBF model is showed in Figure 2. The RBF model hasa better performance in average with FSRT as the ROC curve is almost always up the baseline model. In a word, themodels with fewer attributes after FSRT should not only save computational cost, but also improve the classificationaccuracy for Japan credit dataset.

2430 J. Wang et al. / Procedia Computer Science 1 (2010) 2425–2432

Page 7: Rough set and Tabu search based feature selection for credit scoring

Jue Wang et al. / Procedia Computer Science 00 (2010) 1–8 7

Figure 1: Performance comparison for different models Figure 2: ROC curve and AUC for RBF model

5. Conclusion

A novel approach to feature selection based on rough set and tabu search has been proposed for credit scoringproblem. New diversification and intensification elements have been embedded in FSRT to achieve better performanceand to fit the considered problem. Numerical experiment on Japan credit dataset has been presented to show theefficiency of FSRT. Comparisons with no pre-selection models have revealed that FSRT is promising and it is lessexpensive in computational cost. This research is leading to the effort for developing more refined and accurate creditscoring model.

Acknowledgements

This work is supported by Chinese Academy of Sciences(CAS) Foundation for Planning and Strategy Research(KACX1-YW-0906) and the National Natural Science Foundation of China (NSFC No. 70801058 and No. 70921061).

References

[1] Loretta J. Mester. What’s the point of credit scoring? Business Review, (9):3-16, 1997.[2] D. J. Hand, W.E. Henley. Statistical classification methods in consumer credit scoring: A review.Journal of the Royal Statistical Society.

Series A, 160(3): 523-541, 1997.[3] Yong Shi. Current Research Trend: Information Technology and Decision Making in 2008. International Journal of Information Technology

and Decision Making, 8(1):1-5, 2009.[4] David West. Neural network credit scoring models.Computers & Operations Research, 27(11):1131-1152, 2000.[5] Ligang Zhou, Kin Keung Lai, Jerome Yen. Credit Scoring Models with AUC Maximization Based on Weighted SVM. International Journal

of Information Technology and Decision Making, 8(4):677-696, 2009.[6] Guyon Isabelle, ElisseeffAndre. An Introduction to Variable and Feature Selection. Journal of Machine Learning Research, 3(7):1157C1182,

2003.[7] Huan Liu, Hiroshi Motoda. Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic Publishers, 1998.[8] Jianping Li, Zhenyu Chen, Liwei Wei, Weixuan Xu, Gang Kou. Feature Selection via Least Aquares Suport Feature Machine. International

Journal of Information Technology and Decision Making, 6(4):671-686, 2007.[9] R. Jensen and Q. Shen. Finding Rough Set Reducts with Ant Colony Optimization. Proceedings of the 2003 UK Workshop on Computational

Intelligence, 15-22, 2003.[10] R. Jensen and Q. Shen. Semantics-Preserving Dimensionality Reduction: Rough and Fuzzy-Rough-Based Approaches. IEEE Transactions

on Knowledge and Data Engineering, 16(12):1457-1471, 2004.[11] S. Tan. A Global Search Algorithm for Attributes Reduction. AI 2004: Advances in Artificial Intelligence, G.I. Webb and Xinghuo Yu (Eds.),

LNAI 3339:1004-1010, 2004.[12] Z. Pawlak. Rough sets. International Journal of Computer and Information Sciences, 11(5):341-356, 1982.[13] Z. Pawlak. Rough Sets: Theoretical Aspects of Reasoning About Data. Kluwer Academic Publishing, 1991.

J. Wang et al. / Procedia Computer Science 1 (2010) 2425–2432 2431

Page 8: Rough set and Tabu search based feature selection for credit scoring

Jue Wang et al. / Procedia Computer Science 00 (2010) 1–8 8

[14] Z. Pawlak and A. Skowron. Rough Set Methods and Applications: New Developments in Knowledge Discovery in Information Systems.Studies in Fuzziness and Soft Computing, L. Polkowski, T.Y. Lin, S. Tsumoto (Eds.), 56, Physica-Verlag, 2000.

[15] Roman W. Swiniarski, Andrzej Skowron. Rough Set Methods in Feature Selection and Recognition. Pattern Recognition Letters,24(6):833C849, 2003.

[16] C. Rego and B. Alidaee. Metaheursitic Optimization via Memory and Evolution. Springer-Verlag, Berlin, 2005.[17] F. Glover and M. Laguna. Tabu Search. Kluwer Academic Publishers, MA, USA, 1997.[18] F. Glover, G. Kochenberger. New Optimization Model for Data Mining. International Journal of Information Technology and Decision

Making, 5(4):605-609, 2006.[19] F. Glover. Tabu search–Part I, ORSA Journal on Computing, 1(3):190-206, 1989.[20] F. Glover, Tabu search–Part II, ORSA Journal on Computing, 2(1):4-32, 1990.[21] A. Hedar and M. Fukushima. Tabu Search Directed by Direct Search Methods for Nonlinear Global Optimization. European Journal of

Operational Research, 170(2):329-349, 2006.[22] Guoyin Wang£Algebra View and Information View of Rough Sets Theory£Proceedings of SPIE, 4384£200-207, 2001.[23] T.Y. Lin, Y.Y. Yao, L.A. Zadeh. Data Mining, Rough Sets and Granular Computing. Springer-Verlag, Berlin, 2002.[24] A. Hedar, Jue Wang and M. Fukushima. Tabu Search for Attribute Reduction in Rough Set Theory. Soft Computing, 12(9):909-918, 2008.[25] Vijay S. Desai, Jonathan N. Crook, George A. Berstreet Jr. A Comparison of Neural Network and Linear Scoring Models in the Credit Union

Enviroment. European Journal of Operational Research, 95(1):24-37, 1996.[26] Estefane Lacerda, Andre C.P.L.F. Carvalho, Antonio Padua Braga, Teresa Bernarda Ludermir. Evolutionary Radial Basis Functions for Credit

Assessment. Applied Intelligence, 22(3):167-181, 2005.[27] Cheng Lung Huang, Mu Chen Chen, Chieh Jen Wang. Credit Scoring with a Data Mining Approach Based on Support Vector Machines.

Expert Systems with Applications, 33(4):847-856, 2007.[28] Erkki K.Laitinen. Predicting a Corporate Credit Analyst’s Risk Estimate by Logistic and Linear Models. International Review of Financial

Analysis, 8(2):97-121, 1999.[29] Timothy H. Lee, Ming Zhang. Bias Correction and Statistical Test for Developing Credit Scoring Model through Logistic Regression Ap-

proach. International Journal of Information Technology and Decision Making, 2(3):299-311, 2003.[30] Fawcett, T., 2004. ROC Graphs: Notes and Practical Considerations for Researchers. Intelligent Enterprise Technologies Laboratory, HP

Laboratories Palo Alto, 2004.

2432 J. Wang et al. / Procedia Computer Science 1 (2010) 2425–2432