Please use this identifier to cite or link to this item: https://ah.lib.nccu.edu.tw/handle/140.119/125517
DC FieldValueLanguage
dc.contributor.advisor周珮婷zh_TW
dc.contributor.author游上葦zh_TW
dc.contributor.authorYu, Shang-Weien_US
dc.creator游上葦zh_TW
dc.creatorYu, Shang-Weien_US
dc.date2019en_US
dc.date.accessioned2019-09-05T07:42:06Z-
dc.date.available2019-09-05T07:42:06Z-
dc.date.issued2019-09-05T07:42:06Z-
dc.identifierG0106354024en_US
dc.identifier.urihttp://nccur.lib.nccu.edu.tw/handle/140.119/125517-
dc.description碩士zh_TW
dc.description國立政治大學zh_TW
dc.description統計學系zh_TW
dc.description106354024zh_TW
dc.description.abstract特徵選取為機器學習領域中一重要部分,適當的選取特徵(變數),除了減少機器運算時間、人力與金錢外,也可以避免模型過度配適或是欠擬和的情況發生。雖然經過多年發展已有很多特徵選取的方法,但同一種模型,不一定適用所有資料情況,因此提出新方法希望在特徵選取上會有更多選擇。\n  本文提出一新方法衡量變數之間的關係,使用資訊理論中熵的概念,結合分類器支持向量機,獲取變數間關係,並將變數分群,給予每群適當的相對應分數,以此篩選變數。本文採用半監督式學習,計算敏感度、特異度與準確度之平均及所使用變數數量,並以高斯混合模型搭配EM演算法利用KS檢定之檢定統計量定義資料重要變數,評估方法能否選取重要變數,本文一共使用2筆模擬資料與5筆真實資料,並將結果與各大方法比較,結果顯示在各資料集中皆有穩定表現,即使在變數少的情況下也能有不錯表現。zh_TW
dc.description.abstractFeature selection technique plays a significant role in machine learning. Selecting features (variables) adequately can not only reduce the expenditure, operating time in machine and the cost of labor but also prevent under fitting or overfitting. Although lots of feature selection methods have been developed for decades, it is impossible to apply a unique method to all types of data sets. In this study, we propose a new method to calculate the correlation between variables based on the Shannon entropy from information theory and SVM classifier. Variables are grouped into several clusters and selected by the new correlation measurement. Besides, we define the importance of variable by the test statistic of KS test using Gaussian mixed model and E-M algorithm for the propose of result assessment. The performance of proposed method on two simulated data and five real data are demonstrated and compared with other feature selection methods. The predicted results are stable through the proposed method with a reduced dataset.en_US
dc.description.tableofcontents第一章 緒論 1\n第二章 文獻探討 3\n第三章 研究目的 5\n一、特徵選取(feature selection) 5\n第四章 研究方法 6\n第一節 ENTROPY特徵選取 6\n一、Shannon Entropy 6\n二、Mutual Entropy 7\n三、Support Vector Machine(SVM) 10\n四、Feature selection via Mutual Entropy based on classified model 11\n五、分群規則 13\n六、方法步驟 15\n第二節 其他特徵選取方法 17\n一、Random forest(RF) 17\n二、F-score 18\n三、Least Absolute Shrinkage and Selection Operation(LASSO) 19\n四、Fast Correlation-Based Filter(FCBF) 20\n第五章 資料介紹 22\n一、模擬資料一 22\n二、模擬資料二 24\n三、Wisconsin Diagnostic Breast Cancer 26\n四、Connectionist Bench(Sonar, Mines vs. Rocks) 27\n五、Credit Card Fraud Detection 28\n六、Oxford Parkinson’s Disease 29\n七、APS Failure at Scania Trucks 30\n第六章 研究結果 33\n第一節 資料結果 35\n一、模擬資料一及模擬資料二 35\n二、Wisconsin Diagnostic Breast Cancer 38\n三、Connectionist Bench(Sonar, Mines vs. Rocks) 41\n四、Credit Card Fraud Detection 44\n五、Oxford Parkinson’s Disease 47\n六、APS Failure at Scania Trucks 50\n第二節 結論 53\n第七章 未來展望 54\n第八章 參考資料 55zh_TW
dc.format.extent1809860 bytes-
dc.format.mimetypeapplication/pdf-
dc.source.urihttp://thesis.lib.nccu.edu.tw/record/#G0106354024en_US
dc.subject機器學習zh_TW
dc.subject特徵選取zh_TW
dc.subject維度縮減zh_TW
dc.subject支持向量機zh_TW
dc.subject支援向量機zh_TW
dc.subjectzh_TW
dc.subject相互熵zh_TW
dc.subjectMachine learningen_US
dc.subjectFeature selectionen_US
dc.subjectDimension reductionen_US
dc.subjectSupport Vector Machineen_US
dc.subjectSVMen_US
dc.subjectEntropyen_US
dc.subjectShannon Entropyen_US
dc.subjectMutual Entropyen_US
dc.title基於支持向量機計算的相互熵之特徵選取zh_TW
dc.titleA Feature Selection Study based on SVM and Mutual Entropyen_US
dc.typethesisen_US
dc.relation.referenceAkay, M.F., (2009). Support vector machines combined with feature selection for breast cancer diagnosis. Expert Systems with Applications, 36 (2), 3240–3247.\nArunasakthi, K., KamatchiPriya, & L., Askerunisa, A., (2014). Fisher Score Dimensionality Reduction for SVM Classification. 2014 International Conference on Innovations in Engineering and Technology, 3(3), 1900-1904.\nBlum, A., Langley, P., Selection of relevant features and examples in machine learning. Artificial Intelligence, 97(1-2), 245-271.\nBennasar, M., Hicks, Y., & Setchi, R., (2015). Feature selection using Joint Mutual Information Maximisation. Expert Systems with Applications, 42, 8520-8532.\nBoser, B.E., Guyon, I.M., & Vapnil, V.N., (1992). A Training Algorithm for Optimal Margin Classiers. Proceedings of the fifth annual workshop on Computational learning theory, 144-152.\nBreiman, L., (1996). Bagging Predictors. Machine Learning, 24, 123–140.\nBreiman, L., (2001). Random Forests. Machine Learning, 45(1), 5-32. doi:10.1023/a:1010933404324\nChen, Y.W., & Lin, C.J., (2005). Combining SVMs with Various Feature Selection Strategies. Department of Computer Science, National Taiwan University.\nhttps://www.csie.ntu.edu.tw/~cjlin/papers/features.pdf\nCortes, C., & Vapnik, V., (1995). Support-vector networks. Machine Learning, 20(3), 273-297. doi:10.1007/bf00994018\nDash, M., & Liu, H., (1997). Feature selection for classification. Intelligent Data Analysis, 1(1-4), 131-156.\nDíaz-Uriarte, R., & Alvarez de Andrés, S., (2006). Gene selection and classification of microarray data using random forest. BMC Bioinformatics, 7(1), 3.\ndoi:10.1186/1471-2105-7-3\nGuyon, I., & Elisseeff, A., (2003). An introduction to variable and feature selection. Journal of machine learning research, 3(March), 1157-1182.\nHaldurai, H., Madhubala, T., & Rajalakshmi, R., (2016). A Study on Genetic Algorithm and its Applications. International Journal of Computer Sciences and Engineering, 4(10), 139-143.\nHo, T.K., (1995). Random decision forests. Proceedings of the Third International Conference on Document Analysis and Recognition, 1, 278.\nHo, T.K., (1998). The Random Subspace Method for Constructing Decision Forests. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(8), 832-844.\nHolland, J.H., (1992). Genetic Algorithms. Scientific American, 267(1), 66-73.\nHsieh, F., Liu, S.Y., Hsieh, Y.C., McCowan, B., (2018). From patterned response\n  dependency to structured covariate dependency: Entropy based categorical-\n  pattern-matching. PLoS ONE, 13(6).\nHsu, C.W., Chang, C.C., & Lin, C.J., (2003). Machine Learning with Applications in Breast Cancer Diagnosis and Prognosis. Department of Computer Science, National Taiwan University. https://www.csie.ntu.edu.tw/~cjlin/papers/guide/guide.pdf\nHuang, C.L., & Wang, C.J., (2006). A GA-based feature selection and parameters optimization for support vector machines, Expert Systems with Applications, 31, 231-240.\nJolliffe, I.T., & Cadima, J., (2016). Principal component analysis: A review and recent developments. Philosophical Transactions of The Royal Society A Mathematical Physical and Engineering Sciences, 374(2065), 20150202.\nKumbhar, P., & Mali, M., (2016). A Survey on Feature Selection Techniques and Classification Algorithms for Efficient Text Classification. International Journal of Science and Research, 5(5), 1267-1275.\nPopovic, M., (2018). Researchers in an Entropy Wonderland: A Review of the Entropy Concept. Thermal Science, 22(2), 1163-1178.\nRaymer, M.L., Punch, W.F., Goodman, E.D., Kuhn, L.A., & Jain, A.K., (2000). Dimensionality reduction using genetic algorithms. IEEE Transactions on Evolutionary Computation, 4(2), 164-171.\nKohavi, R., John, G., (1997). Wrappers for feature selection. Artificial Intelligence, 97(1-2), 273-324.\nGenuer, R., Poggi, J. M., & Tuleau-Malot, C., (2010). Variable selection using Random Forests. Pattern Recognition Letters, 31(14), 2225-2236.\nSaeys, Y., Inza, I., & Larrañaga, P., (2007). A review of feature selection techniques in bioinformatics. Bioinformatics, 23(19), 2507-2517.\nScho ̈lkopf, B., Sung, K.K., Burges, Chris J.C., Girosi, F., Niyogi, P., Poggio, T., & Vapnik, V., (1997). Comparing support vector machines with Gaussian kernels to radial basis function classifiers. IEEE Transactions on Signal Processing, 45(11), 2758-2765.\nShannon, C. E., (1948). A Mathematical Theory of Communication. The Bell System Technical Journal. 27, 379-432, 623-656.\nSrivastava, S., Joshi, N., & Gaur, M., (2013). A Review Paper on Feature Selection Methodologies and Their Applications. International Journal of Engineering Research and Development, 7(6), 57-61.\nTibshirani, R., (1996). Regression Shrinkage and Selection via the Lasso. Journal of the Royal Statistical Society. Series B (Methodological), 58(1), 267-288.\nXu, X., (2006). Adaptive Intrusion Detection Based on Machine Learning: Feature Extraction, Classifier Construction and Sequential Pattern Prediction. International Journal of Web Services Practices, 2(1-2), 49-58.\nYu, L., & Liu, H., (2003). Feature selection for high-dimensional data: A fast correlation based filter solution. Department of Computer Science & Engineering, Arizona State University.zh_TW
dc.identifier.doi10.6814/NCCU201900996en_US
item.cerifentitytypePublications-
item.fulltextWith Fulltext-
item.openairecristypehttp://purl.org/coar/resource_type/c_46ec-
item.grantfulltextopen-
item.openairetypethesis-
Appears in Collections:學位論文
Files in This Item:
File SizeFormat
402401.pdf1.77 MBAdobe PDF2View/Open
Show simple item record

Google ScholarTM

Check

Altmetric

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.