Show simple item record

dc.contributor.authorKursun, Olcay
dc.contributor.authorSakar, C. Okan
dc.date.accessioned2021-03-06T08:22:48Z
dc.date.available2021-03-06T08:22:48Z
dc.date.issued2012
dc.identifier.citationSakar C. O. , Kursun O., "A method for combining mutual information and canonical correlation analysis: Predictive Mutual Information and its use in feature selection", EXPERT SYSTEMS WITH APPLICATIONS, cilt.39, ss.3333-3344, 2012
dc.identifier.issn0957-4174
dc.identifier.othervv_1032021
dc.identifier.otherav_e09d932c-cf06-42a3-8550-9130969d8701
dc.identifier.urihttp://hdl.handle.net/20.500.12627/147920
dc.identifier.urihttps://doi.org/10.1016/j.eswa.2011.09.020
dc.description.abstractFeature selection is a critical step in many artificial intelligence and pattern recognition problems. Shannon's Mutual Information (MI) is a classical and widely used measure of dependence measure that serves as a good feature selection algorithm. However, as it is a measure of mutual information in average, under-sampled classes (rare events) can be overlooked by this measure, which can cause critical false negatives (missing a relevant feature very predictive of some rare but important classes). Shannon's mutual information requires a well sampled database, which is not typical of many fields of modern science (such as biomedical), in which there are limited number of samples to learn from, or at least, not all the classes of the target function (such as certain phenotypes in biomedical) are well-sampled. On the other hand, Kernel Canonical Correlation Analysis (KCCA) is a nonlinear correlation measure effectively used to detect independence but its use for feature selection or ranking is limited due to the fact that its formulation is not intended to measure the amount of information (entropy) of the dependence. In this paper, we propose a hybrid measure of relevance, Predictive Mutual Information (PMI) based on MI, which also accounts for predictability of signals from each other in its calculation as in KCCA. We show that PMI has more improved feature detection capability than MI, especially in catching suspicious coincidences that are rare but potentially important not only for experimental studies but also for building computational models. We demonstrate the usefulness of PM!, and superiority over MI, on both toy and real datasets. (C) 2011 Elsevier Ltd. All rights reserved.
dc.language.isoeng
dc.subjectMühendislik
dc.subjectOPERASYON ARAŞTIRMA VE YÖNETİM BİLİMİ
dc.subjectEkonomi ve İş
dc.subjectSosyal Bilimler (SOC)
dc.subjectSosyal ve Beşeri Bilimler
dc.subjectEkonometri
dc.subjectYöneylem
dc.subjectBilgi Sistemleri, Haberleşme ve Kontrol Mühendisliği
dc.subjectSinyal İşleme
dc.subjectBilgisayar Bilimleri
dc.subjectAlgoritmalar
dc.subjectMühendislik ve Teknoloji
dc.subjectBİLGİSAYAR BİLİMİ, YAPAY ZEKA
dc.subjectMühendislik, Bilişim ve Teknoloji (ENG)
dc.subjectBilgisayar Bilimi
dc.subjectMÜHENDİSLİK, ELEKTRİK VE ELEKTRONİK
dc.titleA method for combining mutual information and canonical correlation analysis: Predictive Mutual Information and its use in feature selection
dc.typeMakale
dc.relation.journalEXPERT SYSTEMS WITH APPLICATIONS
dc.contributor.departmentBahçeşehir Üniversitesi , ,
dc.identifier.volume39
dc.identifier.issue3
dc.identifier.startpage3333
dc.identifier.endpage3344
dc.contributor.firstauthorID74450


Files in this item

FilesSizeFormatView

There are no files associated with this item.

This item appears in the following Collection(s)

Show simple item record