Hybrid Feature Selection Using Genetic Algorithm and Information Theory

  • cc icon
  • ABSTRACT

    In pattern classification, feature selection is an important factor in the performance of classifiers. In particular, when classifying a large number of features or variables, the accuracy and computational time of the classifier can be improved by using the relevant feature subset to remove the irrelevant, redundant, or noisy data. The proposed method consists of two parts: a wrapper part with an improved genetic algorithm(GA) using a new reproduction method and a filter part using mutual information. We also considered feature selection methods based on mutual information(MI) to improve computational complexity. Experimental results show that this method can achieve better performance in pattern recognition problems than other conventional solutions.


  • KEYWORD

    Pattern classification , Feature selection , Mutual information , Computational complexity

  • 1. Introduction

    Feature selection algorithms can be categorized based on subset generation and subset evaluation [1]. Subset generation is a search procedure that selects candidate feature subsets, based on certain search strategies, such as complete search, sequential search, and random search. Subset evaluation is a set of evaluation criteria used to evaluate a selected feature subset. The criteria can be categorized into two groups based on their dependency on inductive algorithms: independent criteria and dependent criteria. Some independent criteria include distance measures, information measures, dependency measures, and consistency measures [2-5]. A dependent criterion requires a predetermined inductive algorithm in feature selection. Based on the performance of the inductive algorithm applied on the selected subset, it determines which features are selected. Under evaluation criteria, algorithms are categorized into filter, wrapper, and hybrid. Filter methods are independent of the inductive algorithm and evaluate the performance of the feature subset by using the intrinsic characteristic of the data. In the filter methods, the optimal features subset is selected in one pass by evaluating some predefined criteria. Therefore, filter methods have the ability to quickly compute very high-dimensional datasets; however, they also have the worst classification performance, because they ignore the effect of the selected feature subset on the performance of the inductive algorithm. The wrapper methods utilize the error rate of the inductive algorithm as the evaluation function. They search for the best subset of features in all available feature subsets. Wrapper methods are generally known to perform better than filter methods.

    Information theory has been applied to feature selection problems in recent years. Battiti [6] proposed a feature selection method called mutual information feature selection (MIFS). Kwak and Choi [7] investigated the limitation of MIFS using a simple example and proposed an algorithm that can overcome the limitation and improve performance. The main advantage of mutual information methods is the robustness of the noise and data transformation. Despite these advantages, the drawback of feature selection methods based on mutual information is the slow computational speed due to the computation of a highdimensional covariance matrix. In pattern recognition, feature selection methods have been applied to various classifiers. Mao [8] proposed a feature selection method based on pruning and support vector machine (SVM), and Hsu et al. [9] proposed a method called artificial neural net input gain measurement approximation (ANNIGMA) based on weights of neural networks. Pal and Chintalapudi [10] proposed an advanced online feature selection method to select the relevant features during the learning time of neural networks.

    On the other hand, the techniques of evolutionary computation, such as genetic algorithm and genetic programming, have been applied to feature selection to find the optimal features subset. Siedlecki and Sklansky [11] used GA-based branchand- bound technique. Pal et al. [12] proposed a new genetic operator called self-crossover for feature selection. In the genetic algorithm (GA)-based feature selection techniques, each chromosomal gene represents a feature and each individual represents a feature subset. If the ith gene of the chromosome equals 1, then the ith feature is selected as one of the features used to evaluate a fitness function; if the chromosome is 0, then the corresponding feature is not selected. Kudo and Sklansky [13] compared a GA-based feature selection with many conventional feature selection methods, and they showed that GA-based feature selection performs better than others for highdimensional datasets.

    In this paper, we propose a feature selection method using both information theory and genetic algorithm. We also considered the performance of each mutual information (MI)-based feature selection method to choose the best MI-based method to combine with genetic algorithm. The proposed method consists of two parts: the filter part and the wrapper part. In the filter part, we evaluated the significance of each feature using mutual information and then removed features with low significance. In the wrapper part, we used genetic algorithm to select the optimal feature subsets with smaller sizes and higher classification

    performance, which is the goal of the proposed method. In order to estimate the performance of the proposed method, we applied our method on University of California-Irvine (UCI) machine-learning data sets [14]. Experimental results showed that our method is effective and efficient in finding small subsets of the significant features for reliable classification.

    2. Mutual Information-Based Feature Selection

       2.1 Entropy and Mutual Information

    Entropy and mutual information are introduced in Shannon’s information theory to measure the information of random variables [15]. Basically, mutual information is a special case of a more general quantity called relative entropy, which is a measure of the distance between two probability distributions. The entropy is a measure of uncertainty of random variables. More specifically, if a discrete random variable X has λ alphabets with its probability density function denoted as p(x) = Pr{X = x}, xλ, then the entropy of X can be defined as

    image

    The joint entropy of two discrete random variables X and Y is defined as follows:

    image

    where p(x, y)denotes the joint probability density function of X and Y. When some variables are known and the others are not, the remaining uncertainty can be described by the conditional entropy, which is defined as

    image

    The common information of two random variables X and Y is defined as the mutual information between them:

    image

    A large amount of mutual information between two random variables means that the two variables are closely related; otherwise, if the mutual information is zero, then the two variables are totally unrelated or independent of each other. The relation between the mutual information and the entropy can be described in (5), which is also illustrated in Figure 1.

    image

    In feature selection problems, the mutual information between two variables feature F and class C is defined in terms of their probabilistic density functions p(f), p(c), andp(f, c):

    image

    If the mutual information I(F;C) between feature F and class C is large, it means that feature F contains much information about class C. If I(F;C) is small, then feature F has little effect on output class C. Therefore, in feature selection problems, the optimal feature subset can be determined by selecting the features with higher mutual information.

       2.2 Mutual Information-Based Feature Selection

    2.21 Feature Selection Problem with Mutual Information

    Feature selection is a process that selects a subset from the complete set of original features. It selects the feature subset that can best improve the performance of a classifier or an inductive algorithm. In the process of selecting features, the number of features is reduced by excluding irrelevant or redundant features from the ones extracted from the raw data. This concept is formalized as selecting the most relevant k features from a set of n features. Battiti [6] named this concept a “feature reduction” problem.

    Let the FRn-k problem be defined as follows:

    Given an initial set of n features, find the subset with k<n features, such that the subset is “maximally informative” about the class.

    In information theory, the mutual information between two random variables measures the amount of information commonly found in these variables. The problem of selecting input features that contain the relevant information about the output can be solved by computing the mutual information between input features and output classes. If the mutual information between input features and output classes could be obtained accurately, the FRn-k problem could be reformulated as follows:

    Given an initial set F with n features and a C set of all output classes, find the subset SFwith k features such that the subset minimizesH(C|S) and maximizes the mutual information I(C; S).

    To solve this FRn-k problem, we can use three strategies: complete search, random search, and sequential search. Complete search guarantees to find the optimal feature subset according to an evaluation criterion. This strategy evaluates all possible subsets to guarantee completeness; however, it is almost impossible due to the large number of combinations. Random search starts with a randomly selected subset and proceeds in two different ways. One is to follow a sequential search, which injects randomness into the sequential approaches. The other is to generate the next subset in a completely random manner. The use of randomness helps to escape local optima in the search space. Sequential search gives up completeness and therefore risks losing optimal subsets. Many variations of the greedy algorithm to the sequential search are available, including sequential forward selection and sequential backward elimination. All these approaches add or remove features one at a time. Algorithms with sequential search are simple to implement and quickly produce results, because the order of the search space is usually O(N2) or less.

    Typically, the mutual information-based feature selection is performed by sequential forward selection. This method starts with an empty set of selected features, and then we add the best available input feature to the selected feature set one by one until the size of the set reaches k. This ideal sequential forward selection algorithm using mutual information is implemented as follows:

    1. (Initialization) Set F ← “initial set of n features,” S ← “empty set.”

    2. (Computation of the MI with the output class) If ∀fi∈ F, compute I(C; F).

    3. (Selection of the first feature) Select a feature that maximizes I(C; F), and set F ← F?{fi},S ← {fi}

    4. (Greedy sequential forward selection) Repeat until the desired number of selected features is reached.

    5. (Computation of the joint MI between variables) If ∀fi∈ F, computeI(C; fi,S).

    6. (Selection of the next feature) Choose the feature fi ∈ F that maximizes I(C; fi,S) and set F ← F ?{fi}, S ← {fi}. Output is the set S containing the selected features.

    where Cis a class, fi is the ith feature, and Sis a feature subset.

    2.22 Battiti’s Mutual Information-Based Feature Selection

    In the ideal sequential forward selection, we must estimate the joint mutual information between variables I(C; fi, S) and know the probabilistic density functions of variables to compute I(C; fi, S). However, it is difficult to compute probabilistic density functions of high-dimension data; therefore, we use a histogram of data.

    In selecting k features, if the output classes are composed of Kc classes and we divide the jth input feature space into Pj partitions to get the histogram, we must have

    image

    cells to compute I(C; fi, S).

    Because of this requirement, implementing the ideal sequential forward selection algorithm is practically impossible. To overcome this practical problem, Battiti [6] used only I(C; fi) and I(fi, fs), instead of calculatingI(C; fi, S). The mutual information I(C; fi)indicates the relative importance of the input feature fi, which was estimated based on the mutual information between the input feature fi and the output classC. I(fi, fs) indicates the redundancy between the input feature fi and the already-selected featuresfs. Battiti’s algorithm, also known as MIFS, is essentially the same as the ideal greedy sequential forward selection algorithm, except for Step 4, which was replaced in MIFS as follows [6]:

    4) (Greedy sequential forward selection) Repeat until the desired number of selected features is reached.

    a) (Computation of the MI between variables) For all couples of variables (fi, fs) with fiF, fsS, compute I(fi, fs), if it is not yet available.

    b) (Selection of the next feature) choose the feature fiF that maximizes I(C; fi) ?βΣfsS I(fi, fs), and set FF ?{fi}, S ← {fi}.

    image

    where β regulates the influence of the redundancy of input feature fi. If β= 0, the redundancy among features is not taken into consideration, and the algorithm selects features in the order of relative importance estimated by the mutual information between input features and output classes.

    3. Genetic Algorithm-Based Feature Selection

    Genetic algorithm is one of the best-known techniques for solving optimization problems. It is also a search method based on a random population. First, genetic algorithm randomly encodes the initial population, which is a set of created individuals. Each individual can be represented as bit strings that can be constructed using all possible permutations in a potential solution space. At each step, the new population is determined by processing the chromosome of the old population in order to obtain the best fitness in a given situation. This sequence continues until a termination criterion is reached. The chromosome manipulation is performed using one of three genetic operators: crossover, mutation, and reproduction. The selection step determines which individuals will participate in the reproduction phase. Reproduction itself allows the exchange of already existing genes, whereas mutation introduces new genetic material, where the substitution defines the individuals for the next population. This process efficiently provides optimal or near-optimal solutions.

    In the genetic algorithm-based feature selection, the size of chromosome n represents the total number of features N, and a gene represents a feature with values “1” and “0” meaning selected and removed, respectively. Therefore, we can define genetic algorithm-based feature selection as finding the optimal feature subset with the smallest number of genes set to “1” and with a higher classification performance.

    A generational procedure GA is shown below:

    image

    4. Proposed Method by Mutual Information and Genetic Algorithm

    The proposed method can be divided into two parts. The filter part is a preprocessing step to remove irrelevant, redundant, and noise features according to ranking by mutual information. In the wrapper part, the optimal feature subset is selected from the preprocessed features using genetic algorithm with the fitness function based on the number of features and on the accuracy of classification. Figure 2 shows the structure of the proposed method, described as follows.

       4.1 Filter Part by Mutual Information

    [Step 1] Evaluate mutual information. We determine the mutual information between each feature F and each class C by (6).

    [Step 2] Select the top-ranked features. Each feature is ranked using the evaluated mutual information. Then, we select the top-ranked features with higher mutual information to use as candidate individuals for the genetic algorithm. To determine the number of top-ranked features, we categorized the features into three types: full-top-ranked, half-top-ranked, and U-topranked. Full and half-top-ranked can be used on data with a small feature size, and U-top-ranked can be used on data with a

    high feature size, where U is the selection rate determined by the user.

       4.2 Wrapper Part by Genetic Algorithm

    [Step 3] Set initial parameters of GA and generate initial population. Set the initial parameters of GA, such as population size, probability of crossover and mutation, the number of generation, weight of fitness, and initial population rate. We also generate the initial population from the features selected in Step 2. In genetic algorithm-based feature selection methods, each chromosome is represented by an n-bit binary for an n?dimensional feature space {b1, b2, · · · , bn}. If the ith feature is present in the feature subset represented by the chromosome, then bi = 1; otherwise, bi = 0.

    An example of a simple procedure is as follows:

    image

    where |P| is the population size, random_number is a function that generates a random floating number within [0,1], and αis the expected number of selected features. Figure 3 shows the structure of an n-dimension binary chromosome.

    [Step 4] Evaluate fitness function of initial generation. Evaluate the fitness values of all individuals in the population using a classifier to evaluate each chromosome (the selected feature subset) based on the classification accuracy and the dimension of the feature subset. Tan et al. [14] proposed the following fitness function in order to optimize two objectives: maximize the classification accuracy of the feature subset and minimize the size of the feature subset.

    image

    where z is a feature vector representing a selected feature subset, and λ is a weight value between 0 and 1. The function is composed of two parts. The first part is the weighted classification accuracyacc(z)from a classifier and the second part is the weighted size feats(z)of the feature subset represented byz. In order to get a balance between accuracy and dimensionality reduction, the following fitness function is proposed:

    image

    where total_feat is the maximum number of features for the problem. In (8) and (9), in order to obtain the classifier with the best accuracy and the smallest size, λ is set to the rate 0.5.

    [Step 5] Perform selection, crossover, and mutation step. To produce the new feature subsets, these operators are carried out by the GA: selection operator, crossover operator, and mutation operator. The selection operator selects new feature subsets based on the fitness value of each feature, and then the crossover and mutation operators create the next generation feature subsets.

    [Step 6] Evaluate the fitness function of the next generation.

    [Step 7] Perform termination test. If a predefined generation is satisfied, then stop the algorithm. Otherwise, go to Step 5. In this study, the termination condition required only one generation.

    5. Experiments and Discussions

    Ten-fold cross-validation procedure is commonly used to evaluate the performance of k-nearest neighbor algorithm(k-NN) with 1-nearest neighbor. In the 10-fold cross-validation, the selected feature subsets are partitioned into ten. To test the MLP, one feature subset is retained as the validation data, and the remaining nine feature subsets are used as training data. The cross-validation process is repeated 10 times, and the 10 sets of results can be averaged to produce a single estimate. In

    the filter part, we used 0.7-top-ranked feature. Table 1 shows the parameters for the genetic algorithm and Table 2 shows the UCI datasets used in this experiment. The first two rows are artificial datasets and the rest are real-world datasets.

    In this experiment, we used the fitness values evaluated by using (9) and the three genetic operators: roulette-wheel selection, uniform crossover, and simple mutation.

    Figures 4 and 5 show the fitness values of GA in each generation and the number of selected features in each generation. We can see that the optimal feature subset was effectively found by the proposed method.

    In Table 3, all the results of different methods are obtained from [9]. The NN column lists the results with no feature subset selection. The second column shows the results of the standard wrapper-backward elimination method. The third column shows the results of the ANNIGMA-wrapper method proposed by Hsu et al. The fourth column represents the results of GA, and the final column shows the results of the proposed method. For each error and each number of selected features, we include the average and the standard deviation. As shown in Table 3, the proposed method shows better performance than the other methods for most datasets with small features. More specifically, the error rate is 4.2% when using the eight dominant features chosen bysed method, whereas the error rate is 11.4% for NN without feature selection. From this result, one can see thatsed method makes it possible to dramatically decrease the error.

    6. Conclusion

    The feature selection methods can be divided into two groups, filter method and wrapper method, based on their dependence and independence on the inductive algorithm. Filter methods

    have fast computational ability because the optimal feature subset is selected in one pass by evaluating some predefined criteria. However, they have the worst classification performance, because they ignore the effect of the selected feature subset on the performance of the inductive algorithm. The wrapper methods have higher performance than the filter methods, whereas they have high computational cost.

    In order to overcome the drawbacks of both filter methods and wrapper methods, we propose a feature selection method using both information theory and genetic algorithm. The proposed method was applied to UCI datasets and some gene expression datasets. For the various experimental datasets, the proposed method had better generalization performance than previous ones. More specifically, the error rate is 4.2% when using the eight dominant features chosen by the proposed method, whereas the error rate is 11.4% for NN without feature selection. From these results, one can see that the proposed method makes it possible to dramatically decrease the error without increasing the computational time.

      >  Conflict of Interest

    No potential conflict of interest relevant to this article was reported

  • 1. Dash M., Liu H. 1997 “Feature selection for classification” [Intelligent Data Analysis] Vol.1 P.131-156 google doi
  • 2. Rais M., Barrera J., Martins Jr D. C. 2010 “U-curve: a branch-and-bound optimization algorithm for U-shaped cost functions on Boolean lattices applied to the feature selection problem” [Pattern Recognition] Vol.43 P.557-568 google doi
  • 3. Foithong S., Pinngern O., Attachoo B. 2012 “Feature subset selection wrapper based on mutual information and rough sets” [Expert Systems with Applications] Vol.39 P.574-584 google doi
  • 4. Pal N. R., Malpani M. 2012 ”Redundancy-constrained feature selection with radial basis function networks” [in Proceedings of 2012 IEEE International Joint Conference on Neural Networks (IJCNN)] P.1-8 google doi
  • 5. Zhang T. 2009 ”On the consistency of feature selection using greedy least squares regression” [Journal of Machine Learning Research] Vol.10 P.555-568 google
  • 6. Battiti R. 1994 “Using mutual information for selecting features in supervised neural net learning” [IEEE Transactions on Neural Networks] Vol.5 P.537-550 google doi
  • 7. Kwak N., Choi C. H. 2002 “Input feature selection for classification problems” [IEEE Transactions on Neural Networks] Vol.13 P.143-159 google doi
  • 8. Mao K. Z. 2004 “Feature subset selection for support vector machines through discriminative function pruning analysis” [IEEE Transactions on Systems, Man, and Cybernetics, Part B] Vol.34 P.60-67 google doi
  • 9. Hsu C. N., Huang H. J., Dietrich S. 2002 “The ANNIGMAwrapper approach to fast feature selection for neural nets” [IEEE Transactions on Systems, Man, and Cybernetics, Part B] Vol.32 google doi
  • 10. Pal N. R., Chintalapudi K. 1997 “A connectionist system for feature selection” [Neural, Parallel and Scientific Computations] Vol.5 P.359-381 google
  • 11. Siedlecki W., Sklansky J. 1989 “A note on genetic algorithms for large-scale feature selection” [Pattern Recognition Letters] Vol.10 P.335-347 google doi
  • 12. Pal N. R., Nandi S., Kundu M. K. 1998 “Self-crossover: a new genetic operator and its application to feature selection” [International Journal of Systems Science] Vol.29 P.207-212 google doi
  • 13. Kudo M., Sklansky J. 2000 “Comparison of algorithms that select features for pattern classifiers” [Pattern Recognition] Vol.33 P.25-41 google doi
  • 14. Tan F., Fu X. Z., Zhang Y. Q., Bourgeois A. G. 2006 “Improving feature subset selection using a genetic algorithm for microarray gene expression data” [in Proceedings of 2006 IEEE Congress on Evolutionary Computation] P.2529-2534 google doi
  • 15. Shannon C. E., Weaver W. 1949 The Mathematical Theory of Communication google
  • [Figure 1.] Relation between entropy and mutual information.
    Relation between entropy and mutual information.
  • [Figure 2.] Flowchart of the proposed method. GA, genetic algorithm; k-NN, k-nearest neighbor algorithm.
    Flowchart of the proposed method. GA, genetic algorithm; k-NN, k-nearest neighbor algorithm.
  • [Figure 3.] n-dimension binary chromosome.
    n-dimension binary chromosome.
  • [Table 1.] Parameters for the genetic algorithm
    Parameters for the genetic algorithm
  • [Table 2.] UCI datasets and classifiers used in this experiment
    UCI datasets and classifiers used in this experiment
  • [Table 3.] Comparison results between the proposed and other methods
    Comparison results between the proposed and other methods
  • [Figure 4.] Fitness values of the genetic algorithm. (a) Breast cancerWisconsin data, (b) credit data, (c) monk3b data, (d) ionosphere data.
    Fitness values of the genetic algorithm. (a) Breast cancerWisconsin data, (b) credit data, (c) monk3b data, (d) ionosphere data.
  • [Figure 5.] The number of selected feature in each generation. (a) Breast cancerWisconsin data, (b) credit data, (c) monk3b data, (d) ionosphere data.
    The number of selected feature in each generation. (a) Breast cancerWisconsin data, (b) credit data, (c) monk3b data, (d) ionosphere data.