Show simple item record

dc.contributor.authorHadzic, Fedja
dc.contributor.authorDillon, Tharam S
dc.contributor.editorHuan Liu
dc.contributor.editorRobert Stine
dc.contributor.editorLeonardo Auslender
dc.date.accessioned2017-01-30T13:24:56Z
dc.date.available2017-01-30T13:24:56Z
dc.date.created2010-03-14T20:02:17Z
dc.date.issued2006
dc.identifier.citationHadzic, Fedja and Dillon, Tharam S. 2006. Using the symmetrical Tau criterion for feature selection decision tree and neural network learning, in Huan Liu, Robert Stine and Leonardo Auslender (ed), 2nd Workshop on Feature Selection for Data Mining: Interfacing Machine Learning and Statistics, in conjunction with the 2006 SIAM International Conference on Data Mining, Apr 20 2006, pp. 107-114. Bethesda, USA: ACM.
dc.identifier.urihttp://hdl.handle.net/20.500.11937/31352
dc.description.abstract

The data collected for various domain purposes usually contains some features irrelevant tothe concept being learned. The presence of these features interferes with the learning mechanism and as a result the predicted models tend to be more complex and less accurate. It is important to employ an effective feature selection strategy so that only the necessary and significant features will be used to learn the concept at hand. The Symmetrical Tau (t) [13] is a statistical-heuristic measure for the capability of an attribute in predicting the class of another attribute, and it has successfully been used as a feature selection criterion during decision tree construction. In this paper we aim to demonstrate some other ways of effectively using the t criterion to filter out the irrelevant features prior to learning (pre-pruning) and after the learning process (post-pruning). For the pre-pruning approach we perform two experiments, one where the irrelevant features are filtered out according to their t value, and one where we calculate the t criterion for Boolean combinations of features and use the highest t-valued combination. In the post-pruning approach we use the t criterion to prune a trained neural network and thereby obtain a more accurate and simple rule set. The experiments are performed on data characterized by continuous and categorical attributes and the effectiveness of the proposed techniques is demonstrated by comparing the derived knowledge models in terms of complexity and accuracy.

dc.publisherACM
dc.subjectfeature selection
dc.subjectnetwork pruning
dc.subjectrule simplification
dc.titleUsing the symmetrical Tau criterion for feature selection decision tree and neural network learning
dc.typeConference Paper
dcterms.source.startPage107
dcterms.source.endPage114
dcterms.source.titleProceedings of the 2nd workshop on feature selection for data mining: interfacing machine learning and statistics, in conjunction with the 2006 SIAM international conference on data mining
dcterms.source.seriesProceedings of the 2nd workshop on feature selection for data mining: interfacing machine learning and statistics, in conjunction with the 2006 SIAM international conference on data mining
dcterms.source.isbn9780898716115
dcterms.source.conference2nd Workshop on Feature Selection for Data Mining: Interfacing Machine Learning and Statistics, in conjunction with the 2006 SIAM International Conference on Data Mining
dcterms.source.conference-start-dateApr 20 2006
dcterms.source.conferencelocationBethesda, USA
dcterms.source.placeUSA
curtin.departmentDigital Ecosystems and Business Intelligence Institute (DEBII)
curtin.accessStatusOpen access
curtin.facultyCurtin Business School
curtin.facultyThe Digital Ecosystems and Business Intelligence Institute (DEBII)


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record