學術產出-Theses

Article View/Open

Publication Export

Google ScholarTM

政大圖書館

Citation Infomation

  • No doi shows Citation Infomation
題名 無母數可加性迴歸模型下的變數選取方法之比較
A Comparative Study of Variable Selection Methods in Nonparametric Additive Regression Models
作者 楊侑
Yang, You
貢獻者 黃子銘<br>劉惠美
Huang, Tzee-Ming<br>Liu, Hui-Mei
楊侑
Yang, You
關鍵詞 無母數可加性迴歸模型
B樣條基底
樣條函數
變數選取
Nonparametric additive regression model
B-spline basis
Spline function
Variable selection
日期 2023
上傳時間 2-Jan-2024 15:17:32 (UTC+8)
摘要 本論文深入探討了無母數可加性迴歸模型下的變數選取方法。我 們採用了由 B 樣條基底組成的樣條函數建立模型,並比較了三種不同 的變數選取方法:Group Lasso、Adaptive Group Lasso 和 Group Lasso 結合向前選取法。透過四個不同的指標來評估這些變數選取方法的有 效性,同時透過圖表呈現的方式,深入研究變數選取對於樣條函數估 計的影響。 模擬實驗的結果顯示,在變數選擇的準確度方面,Group Lasso 方 法表現較低,但在保留非零函數方面表現優異。相較之下,Adaptive Group Lasso 方法在準確度上優於 Group Lasso 方法,但在保留非零函 數方面略遜。根據模擬實驗,我們認為 Adaptive Group Lasso 對懲罰參 數的敏感性是造成變數選取效果不理想的原因。未來的研究方向可以 致力於優化懲罰參數選擇演算法,以提升 Adaptive Group Lasso 方法的 性能。 最後,Group Lasso 結合向前選取法的方法在所有指標上表現出 色,因此,我們建議將其作為無母數可加性迴歸模型下的首選變數選 取方法。同時,模擬實驗結果顯示,當模型選擇剛好的變數時,樣條 函數估計效果最佳,進一步突顯了變數選取對於模型函數估計的重要 性。
This paper delves into variable selection methods under nonparametric additive regression models. We employed spline functions composed of B-spline basis to construct the model, comparing three different variable selection methods: Group Lasso, Adaptive Group Lasso, and Group Lasso combined with forward selection. The effectiveness of these variable selection methods was evaluated through four distinct metrics, and their impact on spline function estimation was thoroughly investigated through graphical representations. Simulation results indicate that, in terms of variable selection accuracy, Group Lasso performs relatively lower but excels in retaining non-zero functions. In comparison, Adaptive Group Lasso outperforms Group Lasso in accuracy but slightly lags behind in retaining non-zero functions. Based on the simulation experiments, we attribute the suboptimal variable selection performance of Adaptive Group Lasso to its sensitivity to penalty parameters. Future research directions could focus on optimizing penalty parameter selection algorithms to enhance the performance of Adaptive Group Lasso. Finally, the method of combining Group Lasso with forward selection demonstrates outstanding performance across all indicators. We recommend it as the preferred variable selection method under nonparametric additive regression models. Additionally, simulation results highlight that when the model selects precisely the right variables, the spline function estimation achieves optimal results, emphasizing the importance of variable selection in model function estimation.
參考文獻 De Boor, C. (1972). On calculating with B-splines. Journal of Approximation theory, 6(1):50–62. Efroymson, M. A. (1960). Multiple regression analysis. Mathematical methods for digital computers, 191–203. Fox, J. (2002). Nonparametric regression. Appendix to: An R and S-PLUS Companion to Applied Regression, 1–7. Huang, J., Horowitz, J. L., and Wei, F. (2010). Variable selection in nonparametric additive models. The Annals of Statistics, 38(4):2282–2313. Schoenberg, I. J. (1946). Contributions to the problem of approximation of equidistant data by analytic functions. Part B. On the problem of osculatory interpolation. A second class of analytic approximation formulae. Quarterly of Applied Mathematics, 4(2):112–141. Schwarz, G. (1978). Estimating the dimension of a model. The Annals of Statistics, 461–464. Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society: Series B (Methodological), 58(1):267–288. Yuan, M. and Lin, Y. (2006). Model selection and estimation in regression with grouped variables. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 68(1):49–67. Zou, H. (2006). The adaptive lasso and its oracle properties. Journal of the American statistical association, 101(476):1418–1429.
描述 碩士
國立政治大學
統計學系
110354021
資料來源 http://thesis.lib.nccu.edu.tw/record/#G0110354021
資料類型 thesis
dc.contributor.advisor 黃子銘<br>劉惠美zh_TW
dc.contributor.advisor Huang, Tzee-Ming<br>Liu, Hui-Meien_US
dc.contributor.author (Authors) 楊侑zh_TW
dc.contributor.author (Authors) Yang, Youen_US
dc.creator (作者) 楊侑zh_TW
dc.creator (作者) Yang, Youen_US
dc.date (日期) 2023en_US
dc.date.accessioned 2-Jan-2024 15:17:32 (UTC+8)-
dc.date.available 2-Jan-2024 15:17:32 (UTC+8)-
dc.date.issued (上傳時間) 2-Jan-2024 15:17:32 (UTC+8)-
dc.identifier (Other Identifiers) G0110354021en_US
dc.identifier.uri (URI) https://nccur.lib.nccu.edu.tw/handle/140.119/149018-
dc.description (描述) 碩士zh_TW
dc.description (描述) 國立政治大學zh_TW
dc.description (描述) 統計學系zh_TW
dc.description (描述) 110354021zh_TW
dc.description.abstract (摘要) 本論文深入探討了無母數可加性迴歸模型下的變數選取方法。我 們採用了由 B 樣條基底組成的樣條函數建立模型,並比較了三種不同 的變數選取方法:Group Lasso、Adaptive Group Lasso 和 Group Lasso 結合向前選取法。透過四個不同的指標來評估這些變數選取方法的有 效性,同時透過圖表呈現的方式,深入研究變數選取對於樣條函數估 計的影響。 模擬實驗的結果顯示,在變數選擇的準確度方面,Group Lasso 方 法表現較低,但在保留非零函數方面表現優異。相較之下,Adaptive Group Lasso 方法在準確度上優於 Group Lasso 方法,但在保留非零函 數方面略遜。根據模擬實驗,我們認為 Adaptive Group Lasso 對懲罰參 數的敏感性是造成變數選取效果不理想的原因。未來的研究方向可以 致力於優化懲罰參數選擇演算法,以提升 Adaptive Group Lasso 方法的 性能。 最後,Group Lasso 結合向前選取法的方法在所有指標上表現出 色,因此,我們建議將其作為無母數可加性迴歸模型下的首選變數選 取方法。同時,模擬實驗結果顯示,當模型選擇剛好的變數時,樣條 函數估計效果最佳,進一步突顯了變數選取對於模型函數估計的重要 性。zh_TW
dc.description.abstract (摘要) This paper delves into variable selection methods under nonparametric additive regression models. We employed spline functions composed of B-spline basis to construct the model, comparing three different variable selection methods: Group Lasso, Adaptive Group Lasso, and Group Lasso combined with forward selection. The effectiveness of these variable selection methods was evaluated through four distinct metrics, and their impact on spline function estimation was thoroughly investigated through graphical representations. Simulation results indicate that, in terms of variable selection accuracy, Group Lasso performs relatively lower but excels in retaining non-zero functions. In comparison, Adaptive Group Lasso outperforms Group Lasso in accuracy but slightly lags behind in retaining non-zero functions. Based on the simulation experiments, we attribute the suboptimal variable selection performance of Adaptive Group Lasso to its sensitivity to penalty parameters. Future research directions could focus on optimizing penalty parameter selection algorithms to enhance the performance of Adaptive Group Lasso. Finally, the method of combining Group Lasso with forward selection demonstrates outstanding performance across all indicators. We recommend it as the preferred variable selection method under nonparametric additive regression models. Additionally, simulation results highlight that when the model selects precisely the right variables, the spline function estimation achieves optimal results, emphasizing the importance of variable selection in model function estimation.en_US
dc.description.tableofcontents 第一章 緒論 1 第二章 文獻回顧 2 第一節 模型架構 2 第二節 樣條函數 3 第三節 變數選取方法 4 第三章 研究方法 7 第一節 近似無母數可加性迴歸模型 7 第二節 利用 Group Lasso 選取變數 8 第三節 利用 Adaptive Group Lasso 選取變數 10 第四節 Group Lasso 結合向前選取法 11 第四章 模擬實驗 12 第一節 資料生成與樣條函數基底的配置和處理 12 第二節 實驗變數選取結果 13 第三節 樣條函數估計效果 16 第五章 結論與建議 20 參考文獻 22zh_TW
dc.format.extent 930044 bytes-
dc.format.mimetype application/pdf-
dc.source.uri (資料來源) http://thesis.lib.nccu.edu.tw/record/#G0110354021en_US
dc.subject (關鍵詞) 無母數可加性迴歸模型zh_TW
dc.subject (關鍵詞) B樣條基底zh_TW
dc.subject (關鍵詞) 樣條函數zh_TW
dc.subject (關鍵詞) 變數選取zh_TW
dc.subject (關鍵詞) Nonparametric additive regression modelen_US
dc.subject (關鍵詞) B-spline basisen_US
dc.subject (關鍵詞) Spline functionen_US
dc.subject (關鍵詞) Variable selectionen_US
dc.title (題名) 無母數可加性迴歸模型下的變數選取方法之比較zh_TW
dc.title (題名) A Comparative Study of Variable Selection Methods in Nonparametric Additive Regression Modelsen_US
dc.type (資料類型) thesisen_US
dc.relation.reference (參考文獻) De Boor, C. (1972). On calculating with B-splines. Journal of Approximation theory, 6(1):50–62. Efroymson, M. A. (1960). Multiple regression analysis. Mathematical methods for digital computers, 191–203. Fox, J. (2002). Nonparametric regression. Appendix to: An R and S-PLUS Companion to Applied Regression, 1–7. Huang, J., Horowitz, J. L., and Wei, F. (2010). Variable selection in nonparametric additive models. The Annals of Statistics, 38(4):2282–2313. Schoenberg, I. J. (1946). Contributions to the problem of approximation of equidistant data by analytic functions. Part B. On the problem of osculatory interpolation. A second class of analytic approximation formulae. Quarterly of Applied Mathematics, 4(2):112–141. Schwarz, G. (1978). Estimating the dimension of a model. The Annals of Statistics, 461–464. Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society: Series B (Methodological), 58(1):267–288. Yuan, M. and Lin, Y. (2006). Model selection and estimation in regression with grouped variables. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 68(1):49–67. Zou, H. (2006). The adaptive lasso and its oracle properties. Journal of the American statistical association, 101(476):1418–1429.zh_TW