ACO Based Feature Subset Selection for Multiple k-Nearest Neighbor Classifiers
نویسندگان
چکیده
The k-nearest neighbor (k-NN) is one of the most popular algorithms used for classification in various fields of pattern recognition & data mining problems. In k-nearest neighbor classification, the result of a new instance query is classified based on the majority of k-nearest neighbors. Recently researchers have begun paying attention to combining a set of individual k-NN classifiers, each using a different subset of features, with the hope of improving the overall classification accuracy. In this paper we proposed Ant Colony Optimization (ACO) based feature subset selection for multiple k-nearest neighbor classifiers. The ACO is an iterative meta-heuristic search technique, which inspired by the foraging food behavior of real ant colonies. In ACO, real ants become artificial ants with the particular abilities such as distance determination & tour memory. The solution is constructed in a probabilistic way based on pheromone model in the form of numerical values. The concept of this approach is selecting the best possible subsets of feature from the original set with the help of ACO and combines the outputs from multiple k-NN classifiers. The experimental results show that this proposed method improves the average classification accuracy of k-NN classifier. KeywordsMachine Learning; k-Nearest Neighbor; Feature Subset Selection; Ant Colony Optimization.
منابع مشابه
A Classification Method for E-mail Spam Using a Hybrid Approach for Feature Selection Optimization
Spam is an unwanted email that is harmful to communications around the world. Spam leads to a growing problem in a personal email, so it would be essential to detect it. Machine learning is very useful to solve this problem as it shows good results in order to learn all the requisite patterns for classification due to its adaptive existence. Nonetheless, in spam detection, there are a large num...
متن کاملAn Improved K-Nearest Neighbor with Crow Search Algorithm for Feature Selection in Text Documents Classification
The Internet provides easy access to a kind of library resources. However, classification of documents from a large amount of data is still an issue and demands time and energy to find certain documents. Classification of similar documents in specific classes of data can reduce the time for searching the required data, particularly text documents. This is further facilitated by using Artificial...
متن کاملAn Improved K-Nearest Neighbor with Crow Search Algorithm for Feature Selection in Text Documents Classification
The Internet provides easy access to a kind of library resources. However, classification of documents from a large amount of data is still an issue and demands time and energy to find certain documents. Classification of similar documents in specific classes of data can reduce the time for searching the required data, particularly text documents. This is further facilitated by using Artificial...
متن کاملNearest Neighbor Ensembles Combines with Weighted Instance and Feature Sub Set Selection: A Survey
Ensemble learning deals with methods which employ multiple learners to solve a problem The generalization ability of an ensemble is usually significantly better than that of a single learner, so ensemble methods are very attractive, at the same time feature selection process of ensemble technique has important role of classifier. This paper, presents the analysis on classification technique of ...
متن کاملSimultaneous feature selection and feature weighting using Hybrid Tabu Search/K-nearest neighbor classifier
Feature selection and feature weighting are useful techniques for improving the classification accuracy of K-nearest-neighbor (K-NN) rule. The term feature selection refers to algorithms that select the best subset of the input feature set. In feature weighting, each feature is multiplied by a weight value proportional to the ability of the feature to distinguish pattern classes. In this paper,...
متن کامل