SlideShare uma empresa Scribd logo
1 de 8
A Fast Clustering-Based Feature Subset Selection Algorithm for
High-Dimensional Data
ABSTRACT:
Feature selection involves identifying a subset of the most useful features that produces
compatible results as the original entire set of features. A feature selection algorithm may be
evaluated from both the efficiency and effectiveness points of view. While the efficiency
concerns the time required to find a subset of features, the effectiveness is related to the quality
of the subset of features. Based on these criteria, a fast clustering-based feature selection
algorithm (FAST) is proposed and experimentally evaluated in this paper. The FAST algorithm
works in two steps. In the first step, features are divided into clusters by using graph-theoretic
clustering methods. In the second step, the most representative feature that is strongly related to
target classes is selected from each cluster to form a subset of features. Features in different
clusters are relatively independent, the clustering-based strategy of FAST has a high probability
of producing a subset of useful and independent features. To ensure the efficiency of FAST, we
adopt the efficient minimum-spanning tree (MST) clustering method. The efficiency and
effectiveness of the FAST algorithm are evaluated through an empirical study. Extensive
experiments are carried out to compare FAST and several representative feature selection
algorithms, namely, FCBF, ReliefF, CFS, Consist, and FOCUS-SF, with respect to four types of
GLOBALSOFT TECHNOLOGIES
IEEE PROJECTS & SOFTWARE DEVELOPMENTS
IEEE FINAL YEAR PROJECTS|IEEE ENGINEERING PROJECTS|IEEE STUDENTS PROJECTS|IEEE
BULK PROJECTS|BE/BTECH/ME/MTECH/MS/MCA PROJECTS|CSE/IT/ECE/EEE PROJECTS
CELL: +91 98495 39085, +91 99662 35788, +91 98495 57908, +91 97014 40401
Visit: www.finalyearprojects.org Mail to:ieeefinalsemprojects@gmail.com
well-known classifiers, namely, the probabilitybased Naive Bayes, the tree-based C4.5, the
instance-based IB1, and the rule-based RIPPER before and after feature selection. The results,
on 35 publicly available real-world high-dimensional image, microarray, and text data,
demonstrate that the FAST not only produces smaller subsets of features but also improves the
performances of the four types of classifiers.
EXISTING SYSTEM:
The embedded methods incorporate feature selection as a part of the training process and are
usually specific to given learning algorithms, and therefore may be more efficient than the other
three categories. Traditional machine learning algorithms like decision trees or artificial neural
networks are examples of embedded approaches. The wrapper methods use the predictive
accuracy of a predetermined learning algorithm to determine the goodness of the selected
subsets, the accuracy of the learning algorithms is usually high. However, the generality of the
selected features is limited and the computational complexity is large. The filter methods are
independent of learning algorithms, with good generality. Their computational complexity is
low, but the accuracy of the learning algorithms is not guaranteed. The hybrid methods are a
combination of filter and wrapper methods by using a filter method to reduce search space that
will be considered by the subsequent wrapper. They mainly focus on combining filter and
wrapper methods to achieve the best possible performance with a particular learning algorithm
with similar time complexity of the filter methods.
DISADVANTAGES OF EXISTING SYSTEM:
The generality of the selected features is limited and the computational complexity is
large.
Their computational complexity is low, but the accuracy of the learning algorithms is not
guaranteed.
The hybrid methods are a combination of filter and wrapper methods by using a filter
method to reduce search space that will be considered by the subsequent wrapper.
PROPOSED SYSTEM
Feature subset selection can be viewed as the process of identifying and removing as many
irrelevant and redundant features as possible. This is because irrelevant features do not
contribute to the predictive accuracy and redundant features do not redound to getting a better
predictor for that they provide mostly information which is already present in other feature(s).
Of the many feature subset selection algorithms, some can effectively eliminate irrelevant
features but fail to handle redundant features yet some of others can eliminate the irrelevant
while taking care of the redundant features. Our proposed FAST algorithm falls into the second
group. Traditionally, feature subset selection research has focused on searching for relevant
features. A well-known example is Relief which weighs each feature according to its ability to
discriminate instances under different targets based on distance-based criteria function.
However, Relief is ineffective at removing redundant features as two predictive but highly
correlated features are likely both to be highly weighted. Relief-F extends Relief, enabling this
method to work with noisy and incomplete data sets and to deal with multiclass problems, but
still cannot identify redundant features.
ADVANTAGES OF PROPOSED SYSTEM:
Good feature subsets contain features highly correlated with (predictive of) the class, yet
uncorrelated with (not predictive of) each other.
The efficiently and effectively deal with both irrelevant and redundant features, and
obtain a good feature subset.
Generally all the six algorithms achieve significant reduction of dimensionality by
selecting only a small portion of the original features.
The null hypothesis of the Friedman test is that all the feature selection algorithms are
equivalent in terms of runtime.
MODULES:
 Distributed clustering
 Subset Selection Algorithm
 Time complexity
 Microarray data
 Data Resource
 Irrelevant feature
MODULE DESCRIPTION
1. Distributed clustering
The Distributional clustering has been used to cluster words into groups based either on their
participation in particular grammatical relations with other words by Pereira et al. or on the
distribution of class labels associated with each word by Baker and McCallum . As
distributional clustering of words are agglomerative in nature, and result in suboptimal word
clusters and high computational cost, proposed a new information-theoretic divisive algorithm
for word clustering and applied it to text classification. proposed to cluster features using a
special metric of distance, and then makes use of the of the resulting cluster hierarchy to choose
the most relevant attributes. Unfortunately, the cluster evaluation measure based on distance
does not identify a feature subset that allows the classifiers to improve their original
performance accuracy. Furthermore, even compared with other feature selection methods, the
obtained accuracy is lower.
2. Subset Selection Algorithm
The Irrelevant features, along with redundant features, severely affect the accuracy of the
learning machines. Thus, feature subset selection should be able to identify and remove as much
of the irrelevant and redundant information as possible. Moreover, “good feature subsets
contain features highly correlated with (predictive of) the class, yet uncorrelated with (not
predictive of) each other. Keeping these in mind, we develop a novel algorithm which can
efficiently and effectively deal with both irrelevant and redundant features, and obtain a good
feature subset.
3. Time complexity
The major amount of work for Algorithm 1 involves the computation of SU values for TR
relevance and F-Correlation, which has linear complexity in terms of the number of instances in
a given data set. The first part of the algorithm has a linear time complexity in terms of the
number of features m. Assuming features are selected as relevant ones in the first part, when k
¼ only one feature is selected.
4. Microarray data
The proportion of selected features has been improved by each of the six algorithms compared
with that on the given data sets. This indicates that the six algorithms work well with
microarray data. FAST ranks 1 again with the proportion of selected features of 0.71 percent.
Of the six algorithms, only CFS cannot choose features for two data sets whose
dimensionalities are 19,994 and 49,152, respectively.
5. Data Resource
The purposes of evaluating the performance and effectiveness of our proposed FAST algorithm,
verifying whether or not the method is potentially useful in practice, and allowing other
researchers to confirm our results, 35 publicly available data sets1 were used. The numbers of
features of the 35 data sets vary from 37 to 49, 52 with a mean of 7,874. The dimensionalities of
the 54.3 percent data sets exceed 5,000, of which 28.6 percent data sets have more than 10,000
features. The 35 data sets cover a range of application domains such as text, image and bio
microarray data classification. The corresponding statistical information. Note that for the data
sets with continuous-valued features, the well-known off-the-shelf MDL method was used to
discredit the continuous values.
6. Irrelevant feature
The irrelevant feature removal is straightforward once the right relevance measure is defined or
selected, while the redundant feature elimination is a bit of sophisticated. In our proposed FAST
algorithm, it involves 1.the construction of the minimum spanning tree from a weighted
complete graph; 2. The partitioning of the MST into a forest with each tree representing a
cluster; and 3.the selection of representative features from the clusters.
SYSTEM FLOW:
SYSTEM CONFIGURATION:-
HARDWARE CONFIGURATION:-
 Processor - Pentium –IV
 Speed - 1.1 Ghz
 RAM - 256 MB(min)
Data set
Irrelevant feature removal
Selected Feature
Minimum Spinning tree
constriction
Tree partition & representation
feature selection
 Hard Disk - 20 GB
 Key Board - Standard Windows Keyboard
 Mouse - Two or Three Button Mouse
 Monitor - SVGA
SOFTWARE CONFIGURATION:-
 Operating System : Windows XP
 Programming Language : JAVA
 Java Version : JDK 1.6 & above.
REFERENCE:
Qinbao Song, Jingjie Ni, and Guangtao Wang, “A Fast Clustering-Based Feature
Subset Selection Algorithm for High-Dimensional Data”, IEEE TRANSACTIONS ON
KNOWLEDGE AND DATA ENGINEERING, VOL. 25, NO. 1, JANUARY 2013.

Mais conteúdo relacionado

Mais procurados

Iaetsd an efficient and large data base using subset selection algorithm
Iaetsd an efficient and large data base using subset selection algorithmIaetsd an efficient and large data base using subset selection algorithm
Iaetsd an efficient and large data base using subset selection algorithmIaetsd Iaetsd
 
A NEW TECHNIQUE INVOLVING DATA MINING IN PROTEIN SEQUENCE CLASSIFICATION
A NEW TECHNIQUE INVOLVING DATA MINING IN PROTEIN SEQUENCE CLASSIFICATIONA NEW TECHNIQUE INVOLVING DATA MINING IN PROTEIN SEQUENCE CLASSIFICATION
A NEW TECHNIQUE INVOLVING DATA MINING IN PROTEIN SEQUENCE CLASSIFICATIONcscpconf
 
Hybridization of Meta-heuristics for Optimizing Routing protocol in VANETs
Hybridization of Meta-heuristics for Optimizing Routing protocol in VANETsHybridization of Meta-heuristics for Optimizing Routing protocol in VANETs
Hybridization of Meta-heuristics for Optimizing Routing protocol in VANETsIJERA Editor
 
Extended pso algorithm for improvement problems k means clustering algorithm
Extended pso algorithm for improvement problems k means clustering algorithmExtended pso algorithm for improvement problems k means clustering algorithm
Extended pso algorithm for improvement problems k means clustering algorithmIJMIT JOURNAL
 
Network Based Intrusion Detection System using Filter Based Feature Selection...
Network Based Intrusion Detection System using Filter Based Feature Selection...Network Based Intrusion Detection System using Filter Based Feature Selection...
Network Based Intrusion Detection System using Filter Based Feature Selection...IRJET Journal
 
Application of three graph Laplacian based semisupervised learning methods to...
Application of three graph Laplacian based semisupervised learning methods to...Application of three graph Laplacian based semisupervised learning methods to...
Application of three graph Laplacian based semisupervised learning methods to...ijbbjournal
 
International Journal of Computer Science, Engineering and Information Techno...
International Journal of Computer Science, Engineering and Information Techno...International Journal of Computer Science, Engineering and Information Techno...
International Journal of Computer Science, Engineering and Information Techno...IJCSEIT Journal
 
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...IJERA Editor
 
C LUSTERING B ASED A TTRIBUTE S UBSET S ELECTION U SING F AST A LGORITHm
C LUSTERING  B ASED  A TTRIBUTE  S UBSET  S ELECTION  U SING  F AST  A LGORITHmC LUSTERING  B ASED  A TTRIBUTE  S UBSET  S ELECTION  U SING  F AST  A LGORITHm
C LUSTERING B ASED A TTRIBUTE S UBSET S ELECTION U SING F AST A LGORITHmIJCI JOURNAL
 
GPCODON ALIGNMENT: A GLOBAL PAIRWISE CODON BASED SEQUENCE ALIGNMENT APPROACH
GPCODON ALIGNMENT: A GLOBAL PAIRWISE CODON BASED SEQUENCE ALIGNMENT APPROACHGPCODON ALIGNMENT: A GLOBAL PAIRWISE CODON BASED SEQUENCE ALIGNMENT APPROACH
GPCODON ALIGNMENT: A GLOBAL PAIRWISE CODON BASED SEQUENCE ALIGNMENT APPROACHijdms
 
Differential Evolution Algorithm (DEA)
Differential Evolution Algorithm (DEA) Differential Evolution Algorithm (DEA)
Differential Evolution Algorithm (DEA) A. Bilal Özcan
 
Bioinformatics data mining
Bioinformatics data miningBioinformatics data mining
Bioinformatics data miningSangeeta Das
 
A ROBUST MISSING VALUE IMPUTATION METHOD MIFOIMPUTE FOR INCOMPLETE MOLECULAR ...
A ROBUST MISSING VALUE IMPUTATION METHOD MIFOIMPUTE FOR INCOMPLETE MOLECULAR ...A ROBUST MISSING VALUE IMPUTATION METHOD MIFOIMPUTE FOR INCOMPLETE MOLECULAR ...
A ROBUST MISSING VALUE IMPUTATION METHOD MIFOIMPUTE FOR INCOMPLETE MOLECULAR ...ijcsa
 
Parallel Evolutionary Algorithms for Feature Selection in High Dimensional Da...
Parallel Evolutionary Algorithms for Feature Selection in High Dimensional Da...Parallel Evolutionary Algorithms for Feature Selection in High Dimensional Da...
Parallel Evolutionary Algorithms for Feature Selection in High Dimensional Da...IJCSIS Research Publications
 
Semi-supervised learning approach using modified self-training algorithm to c...
Semi-supervised learning approach using modified self-training algorithm to c...Semi-supervised learning approach using modified self-training algorithm to c...
Semi-supervised learning approach using modified self-training algorithm to c...IJECEIAES
 
Optimal feature selection from v mware esxi 5.1 feature set
Optimal feature selection from v mware esxi 5.1 feature setOptimal feature selection from v mware esxi 5.1 feature set
Optimal feature selection from v mware esxi 5.1 feature setijccmsjournal
 
Minkowski Distance based Feature Selection Algorithm for Effective Intrusion ...
Minkowski Distance based Feature Selection Algorithm for Effective Intrusion ...Minkowski Distance based Feature Selection Algorithm for Effective Intrusion ...
Minkowski Distance based Feature Selection Algorithm for Effective Intrusion ...IJMER
 

Mais procurados (17)

Iaetsd an efficient and large data base using subset selection algorithm
Iaetsd an efficient and large data base using subset selection algorithmIaetsd an efficient and large data base using subset selection algorithm
Iaetsd an efficient and large data base using subset selection algorithm
 
A NEW TECHNIQUE INVOLVING DATA MINING IN PROTEIN SEQUENCE CLASSIFICATION
A NEW TECHNIQUE INVOLVING DATA MINING IN PROTEIN SEQUENCE CLASSIFICATIONA NEW TECHNIQUE INVOLVING DATA MINING IN PROTEIN SEQUENCE CLASSIFICATION
A NEW TECHNIQUE INVOLVING DATA MINING IN PROTEIN SEQUENCE CLASSIFICATION
 
Hybridization of Meta-heuristics for Optimizing Routing protocol in VANETs
Hybridization of Meta-heuristics for Optimizing Routing protocol in VANETsHybridization of Meta-heuristics for Optimizing Routing protocol in VANETs
Hybridization of Meta-heuristics for Optimizing Routing protocol in VANETs
 
Extended pso algorithm for improvement problems k means clustering algorithm
Extended pso algorithm for improvement problems k means clustering algorithmExtended pso algorithm for improvement problems k means clustering algorithm
Extended pso algorithm for improvement problems k means clustering algorithm
 
Network Based Intrusion Detection System using Filter Based Feature Selection...
Network Based Intrusion Detection System using Filter Based Feature Selection...Network Based Intrusion Detection System using Filter Based Feature Selection...
Network Based Intrusion Detection System using Filter Based Feature Selection...
 
Application of three graph Laplacian based semisupervised learning methods to...
Application of three graph Laplacian based semisupervised learning methods to...Application of three graph Laplacian based semisupervised learning methods to...
Application of three graph Laplacian based semisupervised learning methods to...
 
International Journal of Computer Science, Engineering and Information Techno...
International Journal of Computer Science, Engineering and Information Techno...International Journal of Computer Science, Engineering and Information Techno...
International Journal of Computer Science, Engineering and Information Techno...
 
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
 
C LUSTERING B ASED A TTRIBUTE S UBSET S ELECTION U SING F AST A LGORITHm
C LUSTERING  B ASED  A TTRIBUTE  S UBSET  S ELECTION  U SING  F AST  A LGORITHmC LUSTERING  B ASED  A TTRIBUTE  S UBSET  S ELECTION  U SING  F AST  A LGORITHm
C LUSTERING B ASED A TTRIBUTE S UBSET S ELECTION U SING F AST A LGORITHm
 
GPCODON ALIGNMENT: A GLOBAL PAIRWISE CODON BASED SEQUENCE ALIGNMENT APPROACH
GPCODON ALIGNMENT: A GLOBAL PAIRWISE CODON BASED SEQUENCE ALIGNMENT APPROACHGPCODON ALIGNMENT: A GLOBAL PAIRWISE CODON BASED SEQUENCE ALIGNMENT APPROACH
GPCODON ALIGNMENT: A GLOBAL PAIRWISE CODON BASED SEQUENCE ALIGNMENT APPROACH
 
Differential Evolution Algorithm (DEA)
Differential Evolution Algorithm (DEA) Differential Evolution Algorithm (DEA)
Differential Evolution Algorithm (DEA)
 
Bioinformatics data mining
Bioinformatics data miningBioinformatics data mining
Bioinformatics data mining
 
A ROBUST MISSING VALUE IMPUTATION METHOD MIFOIMPUTE FOR INCOMPLETE MOLECULAR ...
A ROBUST MISSING VALUE IMPUTATION METHOD MIFOIMPUTE FOR INCOMPLETE MOLECULAR ...A ROBUST MISSING VALUE IMPUTATION METHOD MIFOIMPUTE FOR INCOMPLETE MOLECULAR ...
A ROBUST MISSING VALUE IMPUTATION METHOD MIFOIMPUTE FOR INCOMPLETE MOLECULAR ...
 
Parallel Evolutionary Algorithms for Feature Selection in High Dimensional Da...
Parallel Evolutionary Algorithms for Feature Selection in High Dimensional Da...Parallel Evolutionary Algorithms for Feature Selection in High Dimensional Da...
Parallel Evolutionary Algorithms for Feature Selection in High Dimensional Da...
 
Semi-supervised learning approach using modified self-training algorithm to c...
Semi-supervised learning approach using modified self-training algorithm to c...Semi-supervised learning approach using modified self-training algorithm to c...
Semi-supervised learning approach using modified self-training algorithm to c...
 
Optimal feature selection from v mware esxi 5.1 feature set
Optimal feature selection from v mware esxi 5.1 feature setOptimal feature selection from v mware esxi 5.1 feature set
Optimal feature selection from v mware esxi 5.1 feature set
 
Minkowski Distance based Feature Selection Algorithm for Effective Intrusion ...
Minkowski Distance based Feature Selection Algorithm for Effective Intrusion ...Minkowski Distance based Feature Selection Algorithm for Effective Intrusion ...
Minkowski Distance based Feature Selection Algorithm for Effective Intrusion ...
 

Destaque

Top Secret: Large-Scale SPA
Top Secret: Large-Scale SPATop Secret: Large-Scale SPA
Top Secret: Large-Scale SPAAnderson Braz
 
Tugas Aksi Sosial - Fisip Unmer Malang
Tugas Aksi Sosial - Fisip Unmer MalangTugas Aksi Sosial - Fisip Unmer Malang
Tugas Aksi Sosial - Fisip Unmer MalangAulia Hamunta
 
Desenvolvimento de Games para Nokia Asha com software free/open source
Desenvolvimento de Games para Nokia Asha com software free/open sourceDesenvolvimento de Games para Nokia Asha com software free/open source
Desenvolvimento de Games para Nokia Asha com software free/open sourceMaurílio Silva
 

Destaque (7)

Top Secret: Large-Scale SPA
Top Secret: Large-Scale SPATop Secret: Large-Scale SPA
Top Secret: Large-Scale SPA
 
Java Web Start
Java Web StartJava Web Start
Java Web Start
 
Tugas Aksi Sosial - Fisip Unmer Malang
Tugas Aksi Sosial - Fisip Unmer MalangTugas Aksi Sosial - Fisip Unmer Malang
Tugas Aksi Sosial - Fisip Unmer Malang
 
JWS
JWSJWS
JWS
 
Effective unit testing
Effective unit testingEffective unit testing
Effective unit testing
 
Desenvolvimento de Games para Nokia Asha com software free/open source
Desenvolvimento de Games para Nokia Asha com software free/open sourceDesenvolvimento de Games para Nokia Asha com software free/open source
Desenvolvimento de Games para Nokia Asha com software free/open source
 
Gwt Presentation1
Gwt Presentation1Gwt Presentation1
Gwt Presentation1
 

Semelhante a JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset selection algorithm for high-dimensional data

2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...
2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...
2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...IEEEMEMTECHSTUDENTSPROJECTS
 
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...IEEEGLOBALSOFTTECHNOLOGIES
 
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...IEEEGLOBALSOFTTECHNOLOGIES
 
Cloudsim a fast clustering-based feature subset selection algorithm for high...
Cloudsim  a fast clustering-based feature subset selection algorithm for high...Cloudsim  a fast clustering-based feature subset selection algorithm for high...
Cloudsim a fast clustering-based feature subset selection algorithm for high...ecway
 
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...ecway
 
Android a fast clustering-based feature subset selection algorithm for high-...
Android  a fast clustering-based feature subset selection algorithm for high-...Android  a fast clustering-based feature subset selection algorithm for high-...
Android a fast clustering-based feature subset selection algorithm for high-...ecway
 
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATAEFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATAIJCI JOURNAL
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
Unsupervised Feature Selection Based on the Distribution of Features Attribut...
Unsupervised Feature Selection Based on the Distribution of Features Attribut...Unsupervised Feature Selection Based on the Distribution of Features Attribut...
Unsupervised Feature Selection Based on the Distribution of Features Attribut...Waqas Tariq
 
A Survey on Constellation Based Attribute Selection Method for High Dimension...
A Survey on Constellation Based Attribute Selection Method for High Dimension...A Survey on Constellation Based Attribute Selection Method for High Dimension...
A Survey on Constellation Based Attribute Selection Method for High Dimension...IJERA Editor
 
An integrated mechanism for feature selection
An integrated mechanism for feature selectionAn integrated mechanism for feature selection
An integrated mechanism for feature selectionsai kumar
 
A Threshold fuzzy entropy based feature selection method applied in various b...
A Threshold fuzzy entropy based feature selection method applied in various b...A Threshold fuzzy entropy based feature selection method applied in various b...
A Threshold fuzzy entropy based feature selection method applied in various b...IJMER
 
New Feature Selection Model Based Ensemble Rule Classifiers Method for Datase...
New Feature Selection Model Based Ensemble Rule Classifiers Method for Datase...New Feature Selection Model Based Ensemble Rule Classifiers Method for Datase...
New Feature Selection Model Based Ensemble Rule Classifiers Method for Datase...ijaia
 
Feature Selection : A Novel Approach for the Prediction of Learning Disabilit...
Feature Selection : A Novel Approach for the Prediction of Learning Disabilit...Feature Selection : A Novel Approach for the Prediction of Learning Disabilit...
Feature Selection : A Novel Approach for the Prediction of Learning Disabilit...csandit
 
Feature selection a novel
Feature selection a novelFeature selection a novel
Feature selection a novelcsandit
 
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...On Feature Selection Algorithms and Feature Selection Stability Measures : A ...
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...AIRCC Publishing Corporation
 
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...ijcsit
 
On Feature Selection Algorithms and Feature Selection Stability Measures : A...
 On Feature Selection Algorithms and Feature Selection Stability Measures : A... On Feature Selection Algorithms and Feature Selection Stability Measures : A...
On Feature Selection Algorithms and Feature Selection Stability Measures : A...AIRCC Publishing Corporation
 

Semelhante a JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset selection algorithm for high-dimensional data (20)

2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...
2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...
2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...
 
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...
 
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...
 
M43016571
M43016571M43016571
M43016571
 
Cloudsim a fast clustering-based feature subset selection algorithm for high...
Cloudsim  a fast clustering-based feature subset selection algorithm for high...Cloudsim  a fast clustering-based feature subset selection algorithm for high...
Cloudsim a fast clustering-based feature subset selection algorithm for high...
 
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...
 
Android a fast clustering-based feature subset selection algorithm for high-...
Android  a fast clustering-based feature subset selection algorithm for high-...Android  a fast clustering-based feature subset selection algorithm for high-...
Android a fast clustering-based feature subset selection algorithm for high-...
 
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATAEFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
SEO PROCESS
SEO PROCESSSEO PROCESS
SEO PROCESS
 
Unsupervised Feature Selection Based on the Distribution of Features Attribut...
Unsupervised Feature Selection Based on the Distribution of Features Attribut...Unsupervised Feature Selection Based on the Distribution of Features Attribut...
Unsupervised Feature Selection Based on the Distribution of Features Attribut...
 
A Survey on Constellation Based Attribute Selection Method for High Dimension...
A Survey on Constellation Based Attribute Selection Method for High Dimension...A Survey on Constellation Based Attribute Selection Method for High Dimension...
A Survey on Constellation Based Attribute Selection Method for High Dimension...
 
An integrated mechanism for feature selection
An integrated mechanism for feature selectionAn integrated mechanism for feature selection
An integrated mechanism for feature selection
 
A Threshold fuzzy entropy based feature selection method applied in various b...
A Threshold fuzzy entropy based feature selection method applied in various b...A Threshold fuzzy entropy based feature selection method applied in various b...
A Threshold fuzzy entropy based feature selection method applied in various b...
 
New Feature Selection Model Based Ensemble Rule Classifiers Method for Datase...
New Feature Selection Model Based Ensemble Rule Classifiers Method for Datase...New Feature Selection Model Based Ensemble Rule Classifiers Method for Datase...
New Feature Selection Model Based Ensemble Rule Classifiers Method for Datase...
 
Feature Selection : A Novel Approach for the Prediction of Learning Disabilit...
Feature Selection : A Novel Approach for the Prediction of Learning Disabilit...Feature Selection : A Novel Approach for the Prediction of Learning Disabilit...
Feature Selection : A Novel Approach for the Prediction of Learning Disabilit...
 
Feature selection a novel
Feature selection a novelFeature selection a novel
Feature selection a novel
 
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...On Feature Selection Algorithms and Feature Selection Stability Measures : A ...
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...
 
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...
 
On Feature Selection Algorithms and Feature Selection Stability Measures : A...
 On Feature Selection Algorithms and Feature Selection Stability Measures : A... On Feature Selection Algorithms and Feature Selection Stability Measures : A...
On Feature Selection Algorithms and Feature Selection Stability Measures : A...
 

Mais de IEEEGLOBALSOFTTECHNOLOGIES

DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Vampire attacks draining life from w...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Vampire attacks draining life from w...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Vampire attacks draining life from w...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Vampire attacks draining life from w...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT SSD a robust rf location fingerprint...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT SSD a robust rf location fingerprint...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT SSD a robust rf location fingerprint...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT SSD a robust rf location fingerprint...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Privacy preserving distributed profi...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Privacy preserving distributed profi...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Privacy preserving distributed profi...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Privacy preserving distributed profi...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Optimal multicast capacity and delay...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Optimal multicast capacity and delay...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Optimal multicast capacity and delay...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Optimal multicast capacity and delay...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT On the real time hardware implementa...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT On the real time hardware implementa...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT On the real time hardware implementa...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT On the real time hardware implementa...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Model based analysis of wireless sys...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Model based analysis of wireless sys...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Model based analysis of wireless sys...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Model based analysis of wireless sys...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Mobile relay configuration in data i...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Mobile relay configuration in data i...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Mobile relay configuration in data i...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Mobile relay configuration in data i...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Distributed cooperative caching in s...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Distributed cooperative caching in s...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Distributed cooperative caching in s...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Distributed cooperative caching in s...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Delay optimal broadcast for multihop...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Delay optimal broadcast for multihop...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Delay optimal broadcast for multihop...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Delay optimal broadcast for multihop...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Dcim distributed cache invalidation ...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Dcim distributed cache invalidation ...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Dcim distributed cache invalidation ...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Dcim distributed cache invalidation ...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Cooperative packet delivery in hybri...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Cooperative packet delivery in hybri...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Cooperative packet delivery in hybri...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Cooperative packet delivery in hybri...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Content sharing over smartphone base...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Content sharing over smartphone base...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Content sharing over smartphone base...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Content sharing over smartphone base...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Community aware opportunistic routin...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Community aware opportunistic routin...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Community aware opportunistic routin...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Community aware opportunistic routin...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Capacity of hybrid wireless mesh net...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Capacity of hybrid wireless mesh net...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Capacity of hybrid wireless mesh net...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Capacity of hybrid wireless mesh net...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Adaptive position update for geograp...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Adaptive position update for geograp...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Adaptive position update for geograp...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Adaptive position update for geograp...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT A scalable server architecture for m...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT A scalable server architecture for m...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT A scalable server architecture for m...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT A scalable server architecture for m...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Scalable and secure sharing of person...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Scalable and secure sharing of person...DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Scalable and secure sharing of person...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Scalable and secure sharing of person...IEEEGLOBALSOFTTECHNOLOGIES
 
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Qos ranking prediction for cloud serv...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Qos ranking prediction for cloud serv...DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Qos ranking prediction for cloud serv...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Qos ranking prediction for cloud serv...IEEEGLOBALSOFTTECHNOLOGIES
 

Mais de IEEEGLOBALSOFTTECHNOLOGIES (20)

DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Vampire attacks draining life from w...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Vampire attacks draining life from w...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Vampire attacks draining life from w...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Vampire attacks draining life from w...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT SSD a robust rf location fingerprint...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT SSD a robust rf location fingerprint...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT SSD a robust rf location fingerprint...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT SSD a robust rf location fingerprint...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Privacy preserving distributed profi...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Privacy preserving distributed profi...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Privacy preserving distributed profi...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Privacy preserving distributed profi...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Optimal multicast capacity and delay...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Optimal multicast capacity and delay...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Optimal multicast capacity and delay...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Optimal multicast capacity and delay...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT On the real time hardware implementa...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT On the real time hardware implementa...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT On the real time hardware implementa...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT On the real time hardware implementa...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Model based analysis of wireless sys...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Model based analysis of wireless sys...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Model based analysis of wireless sys...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Model based analysis of wireless sys...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Mobile relay configuration in data i...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Mobile relay configuration in data i...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Mobile relay configuration in data i...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Mobile relay configuration in data i...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Distributed cooperative caching in s...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Distributed cooperative caching in s...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Distributed cooperative caching in s...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Distributed cooperative caching in s...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Delay optimal broadcast for multihop...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Delay optimal broadcast for multihop...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Delay optimal broadcast for multihop...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Delay optimal broadcast for multihop...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Dcim distributed cache invalidation ...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Dcim distributed cache invalidation ...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Dcim distributed cache invalidation ...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Dcim distributed cache invalidation ...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Cooperative packet delivery in hybri...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Cooperative packet delivery in hybri...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Cooperative packet delivery in hybri...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Cooperative packet delivery in hybri...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Content sharing over smartphone base...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Content sharing over smartphone base...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Content sharing over smartphone base...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Content sharing over smartphone base...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Community aware opportunistic routin...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Community aware opportunistic routin...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Community aware opportunistic routin...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Community aware opportunistic routin...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Capacity of hybrid wireless mesh net...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Capacity of hybrid wireless mesh net...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Capacity of hybrid wireless mesh net...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Capacity of hybrid wireless mesh net...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Adaptive position update for geograp...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Adaptive position update for geograp...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Adaptive position update for geograp...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT Adaptive position update for geograp...
 
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT A scalable server architecture for m...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT A scalable server architecture for m...DOTNET 2013 IEEE MOBILECOMPUTING PROJECT A scalable server architecture for m...
DOTNET 2013 IEEE MOBILECOMPUTING PROJECT A scalable server architecture for m...
 
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...
 
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Attribute based access to scalable me...
 
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Scalable and secure sharing of person...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Scalable and secure sharing of person...DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Scalable and secure sharing of person...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Scalable and secure sharing of person...
 
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Qos ranking prediction for cloud serv...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Qos ranking prediction for cloud serv...DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Qos ranking prediction for cloud serv...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT Qos ranking prediction for cloud serv...
 

Último

(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...AliaaTarek5
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Mark Goldstein
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentPim van der Noll
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
Data governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationData governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationKnoldus Inc.
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxLoriGlavin3
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsNathaniel Shimoni
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesThousandEyes
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI AgeCprime
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxLoriGlavin3
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesKari Kakkonen
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPathCommunity
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfLoriGlavin3
 
[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality AssuranceInflectra
 
How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity PlanDatabarracks
 
Generative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfGenerative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfIngrid Airi González
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfNeo4j
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxLoriGlavin3
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersNicole Novielli
 
Modern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better StrongerModern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better Strongerpanagenda
 

Último (20)

(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
Data governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationData governance with Unity Catalog Presentation
Data governance with Unity Catalog Presentation
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directions
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI Age
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examples
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to Hero
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdf
 
[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance[Webinar] SpiraTest - Setting New Standards in Quality Assurance
[Webinar] SpiraTest - Setting New Standards in Quality Assurance
 
How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity Plan
 
Generative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfGenerative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdf
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdf
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptx
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software Developers
 
Modern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better StrongerModern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
Modern Roaming for Notes and Nomad – Cheaper Faster Better Stronger
 

JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset selection algorithm for high-dimensional data

  • 1. A Fast Clustering-Based Feature Subset Selection Algorithm for High-Dimensional Data ABSTRACT: Feature selection involves identifying a subset of the most useful features that produces compatible results as the original entire set of features. A feature selection algorithm may be evaluated from both the efficiency and effectiveness points of view. While the efficiency concerns the time required to find a subset of features, the effectiveness is related to the quality of the subset of features. Based on these criteria, a fast clustering-based feature selection algorithm (FAST) is proposed and experimentally evaluated in this paper. The FAST algorithm works in two steps. In the first step, features are divided into clusters by using graph-theoretic clustering methods. In the second step, the most representative feature that is strongly related to target classes is selected from each cluster to form a subset of features. Features in different clusters are relatively independent, the clustering-based strategy of FAST has a high probability of producing a subset of useful and independent features. To ensure the efficiency of FAST, we adopt the efficient minimum-spanning tree (MST) clustering method. The efficiency and effectiveness of the FAST algorithm are evaluated through an empirical study. Extensive experiments are carried out to compare FAST and several representative feature selection algorithms, namely, FCBF, ReliefF, CFS, Consist, and FOCUS-SF, with respect to four types of GLOBALSOFT TECHNOLOGIES IEEE PROJECTS & SOFTWARE DEVELOPMENTS IEEE FINAL YEAR PROJECTS|IEEE ENGINEERING PROJECTS|IEEE STUDENTS PROJECTS|IEEE BULK PROJECTS|BE/BTECH/ME/MTECH/MS/MCA PROJECTS|CSE/IT/ECE/EEE PROJECTS CELL: +91 98495 39085, +91 99662 35788, +91 98495 57908, +91 97014 40401 Visit: www.finalyearprojects.org Mail to:ieeefinalsemprojects@gmail.com
  • 2. well-known classifiers, namely, the probabilitybased Naive Bayes, the tree-based C4.5, the instance-based IB1, and the rule-based RIPPER before and after feature selection. The results, on 35 publicly available real-world high-dimensional image, microarray, and text data, demonstrate that the FAST not only produces smaller subsets of features but also improves the performances of the four types of classifiers. EXISTING SYSTEM: The embedded methods incorporate feature selection as a part of the training process and are usually specific to given learning algorithms, and therefore may be more efficient than the other three categories. Traditional machine learning algorithms like decision trees or artificial neural networks are examples of embedded approaches. The wrapper methods use the predictive accuracy of a predetermined learning algorithm to determine the goodness of the selected subsets, the accuracy of the learning algorithms is usually high. However, the generality of the selected features is limited and the computational complexity is large. The filter methods are independent of learning algorithms, with good generality. Their computational complexity is low, but the accuracy of the learning algorithms is not guaranteed. The hybrid methods are a combination of filter and wrapper methods by using a filter method to reduce search space that will be considered by the subsequent wrapper. They mainly focus on combining filter and wrapper methods to achieve the best possible performance with a particular learning algorithm with similar time complexity of the filter methods. DISADVANTAGES OF EXISTING SYSTEM: The generality of the selected features is limited and the computational complexity is large. Their computational complexity is low, but the accuracy of the learning algorithms is not guaranteed. The hybrid methods are a combination of filter and wrapper methods by using a filter method to reduce search space that will be considered by the subsequent wrapper.
  • 3. PROPOSED SYSTEM Feature subset selection can be viewed as the process of identifying and removing as many irrelevant and redundant features as possible. This is because irrelevant features do not contribute to the predictive accuracy and redundant features do not redound to getting a better predictor for that they provide mostly information which is already present in other feature(s). Of the many feature subset selection algorithms, some can effectively eliminate irrelevant features but fail to handle redundant features yet some of others can eliminate the irrelevant while taking care of the redundant features. Our proposed FAST algorithm falls into the second group. Traditionally, feature subset selection research has focused on searching for relevant features. A well-known example is Relief which weighs each feature according to its ability to discriminate instances under different targets based on distance-based criteria function. However, Relief is ineffective at removing redundant features as two predictive but highly correlated features are likely both to be highly weighted. Relief-F extends Relief, enabling this method to work with noisy and incomplete data sets and to deal with multiclass problems, but still cannot identify redundant features. ADVANTAGES OF PROPOSED SYSTEM: Good feature subsets contain features highly correlated with (predictive of) the class, yet uncorrelated with (not predictive of) each other. The efficiently and effectively deal with both irrelevant and redundant features, and obtain a good feature subset. Generally all the six algorithms achieve significant reduction of dimensionality by selecting only a small portion of the original features. The null hypothesis of the Friedman test is that all the feature selection algorithms are equivalent in terms of runtime.
  • 4. MODULES:  Distributed clustering  Subset Selection Algorithm  Time complexity  Microarray data  Data Resource  Irrelevant feature MODULE DESCRIPTION 1. Distributed clustering The Distributional clustering has been used to cluster words into groups based either on their participation in particular grammatical relations with other words by Pereira et al. or on the distribution of class labels associated with each word by Baker and McCallum . As distributional clustering of words are agglomerative in nature, and result in suboptimal word clusters and high computational cost, proposed a new information-theoretic divisive algorithm for word clustering and applied it to text classification. proposed to cluster features using a special metric of distance, and then makes use of the of the resulting cluster hierarchy to choose the most relevant attributes. Unfortunately, the cluster evaluation measure based on distance does not identify a feature subset that allows the classifiers to improve their original performance accuracy. Furthermore, even compared with other feature selection methods, the obtained accuracy is lower. 2. Subset Selection Algorithm The Irrelevant features, along with redundant features, severely affect the accuracy of the learning machines. Thus, feature subset selection should be able to identify and remove as much of the irrelevant and redundant information as possible. Moreover, “good feature subsets
  • 5. contain features highly correlated with (predictive of) the class, yet uncorrelated with (not predictive of) each other. Keeping these in mind, we develop a novel algorithm which can efficiently and effectively deal with both irrelevant and redundant features, and obtain a good feature subset. 3. Time complexity The major amount of work for Algorithm 1 involves the computation of SU values for TR relevance and F-Correlation, which has linear complexity in terms of the number of instances in a given data set. The first part of the algorithm has a linear time complexity in terms of the number of features m. Assuming features are selected as relevant ones in the first part, when k ¼ only one feature is selected. 4. Microarray data The proportion of selected features has been improved by each of the six algorithms compared with that on the given data sets. This indicates that the six algorithms work well with microarray data. FAST ranks 1 again with the proportion of selected features of 0.71 percent. Of the six algorithms, only CFS cannot choose features for two data sets whose dimensionalities are 19,994 and 49,152, respectively. 5. Data Resource The purposes of evaluating the performance and effectiveness of our proposed FAST algorithm, verifying whether or not the method is potentially useful in practice, and allowing other researchers to confirm our results, 35 publicly available data sets1 were used. The numbers of features of the 35 data sets vary from 37 to 49, 52 with a mean of 7,874. The dimensionalities of the 54.3 percent data sets exceed 5,000, of which 28.6 percent data sets have more than 10,000 features. The 35 data sets cover a range of application domains such as text, image and bio
  • 6. microarray data classification. The corresponding statistical information. Note that for the data sets with continuous-valued features, the well-known off-the-shelf MDL method was used to discredit the continuous values. 6. Irrelevant feature The irrelevant feature removal is straightforward once the right relevance measure is defined or selected, while the redundant feature elimination is a bit of sophisticated. In our proposed FAST algorithm, it involves 1.the construction of the minimum spanning tree from a weighted complete graph; 2. The partitioning of the MST into a forest with each tree representing a cluster; and 3.the selection of representative features from the clusters.
  • 7. SYSTEM FLOW: SYSTEM CONFIGURATION:- HARDWARE CONFIGURATION:-  Processor - Pentium –IV  Speed - 1.1 Ghz  RAM - 256 MB(min) Data set Irrelevant feature removal Selected Feature Minimum Spinning tree constriction Tree partition & representation feature selection
  • 8.  Hard Disk - 20 GB  Key Board - Standard Windows Keyboard  Mouse - Two or Three Button Mouse  Monitor - SVGA SOFTWARE CONFIGURATION:-  Operating System : Windows XP  Programming Language : JAVA  Java Version : JDK 1.6 & above. REFERENCE: Qinbao Song, Jingjie Ni, and Guangtao Wang, “A Fast Clustering-Based Feature Subset Selection Algorithm for High-Dimensional Data”, IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, VOL. 25, NO. 1, JANUARY 2013.