Publications-Theses

Article View/Open

Publication Export

Google ScholarTM

NCCU Library

Citation Infomation

Related Publications in TAIR

題名 集成式學習應用於電子零組件供應鏈需求預測
Ensemble learning for demand forecasting in electronics supply chains
作者 郭泰竹
Kuo, Tai-Chu
貢獻者 莊皓鈞<br>周彥君
Chuang, Hao-Chun<br>Chou, Yen-Chun
郭泰竹
Kuo, Tai-Chu
關鍵詞 電子供應鏈
需求預測
集成式學習
Electronics supply chains
Demand forecasting
Ensemble learning
日期 2019
上傳時間 7-Aug-2019 16:09:08 (UTC+8)
摘要 隨著半導體產業以及電子成品業快速發展,電子零組件之供應成為該產業中重要的一環。而通路商在整條供應鏈當中所扮演的角色尤其重要,需要及時反映下游需求並知會上游,達成良好的需求供給,也需要控制存貨水位,降低存貨成本。根據文獻探討,半導體通路商遇到以下三個困難,包括需求前置時間過長,需求變動量大,以及零組件生命週期短。本研究之研究問題為「如何在需求不穩、資料有限的狀況下,預測前置時間需求?」,希望能透過機器學習之技術,預測未來時間之前置需求量,協助通路商控制良好的存貨水位,提高其營業獲利。
本次研究與亞洲一代表性零組件通路商合作,其在擁有龐大營業額下,也需面對相當高的存貨成本,無論是訂貨超出或是不足,都需負擔存貨與缺貨成本。本研究運用其近三年之資料,結合下游製造廠商對於未來需求之預期數量矩陣,針對通路商零組件資料建立預測情境。目標是更精準預測在未來前置時間,下游廠商對於半導體通路商的零組件需求。
為達成以上目標,本研究使用機器學習演算法以及集成式模型架構,首先就不穩定的需求資料進行Temporal Aggregation,並利用跨產品的訓練樣本,以及不同演算法(包含XGBoost、LightGBM、Random Forest等),配合多種參數隨機組合,訓練多個需求預測模型。接著,利用集成式作法(ensemble)將模型混合,本研究使用的方法為平均法以及爬山法(hill-climbing)結合向前選取法(forward-selection),降低各個單一模型預測所產生的偏誤,經較小樣本的完整資料預測測試,和較大樣本的多缺失資料預測測試,本研究發現,利用所提出的資料建構和集成預測,能較各種單一模型顯著降低需求預測誤差,並較該通路商的既有預測方法,能優化70%-80%的零組件預測,使其更能優化存貨控制。
As the semi-conductor and electronics industries grow, the supply of components has become an important part of this industrial chain. Particularly, the distributor is the key of the supply chain because it is responsible for holding inventories to meet demand from downstream plants, in addition to sourcing from upstream vendors with long production lead time. According to the related literature, the distributor faces the problems including the long lead time, high demand volatility, and short life time cycle of the components. This study addresses the following question - How to predict demand over long lead time with limited data and unstable demand series?
This study collaborates with a representative electronics distributor in Asia. Despite its large revenue, the distributor operates under high inventory costs due to excessive holdings or shortages. This study first collects data on downstream demand and rolling forecast over three years. After data engineering, we build scenarios from data sets. The objective is to improve accuracy of demand forecasting, such that the distributor can better prepare for volatile demand from downstream production plants.
To achieve the goal, this study takes machine learning and ensemble learning techniques. First, we use temporal aggregation to smooth demand series and reduce noise. We construct cross-item training samples for different algorithms. After training models with random hyper-parameter search, we use model-averaging and hill-climbing methods to ensemble models` predictions. After testing a relatively small samples with complete data and a larger sample with incomplete data, we show that the proposed approaches can reduce prediction errors, compared with single models. Our method also outperforms the distributor`s internal forecast method for 70%-80% of components, such that the company can apply our method for better inventory control.
參考文獻 Breiman, L. (1996). Stacked regressions. Machine Learning, 24(1), pp. 49–64.
Breiman, L. (2001). Random forests. Machine learning, 45(1), pp. 5-32.
Carbonneau, R., Laframboise, K., & Vahidov, R. (2008). Application of machine learning techniques for supply chain demand forecasting. European Journal of Operational Research, 184(3), pp. 1140-1154.
Chen, T., & Guestrin, C. (2016, August). Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining (pp. 785-794). ACM.
Freund, Y., & Schapire, R. E. (1996, July). Experiments with a new boosting algorithm. In icml (Vol. 96, pp. 148-156).
Friedman, J. H. (2001). Greedy function approximation: a gradient boosting machine. Annals of statistics, pp. 1189-1232.
Graczyk, M., Lasota, T., Trawiński, B., & Trawiński, K. (2010, March). Comparison of bagging, boosting and stacking ensembles applied to real estate appraisal. In Asian conference on intelligent information and database systems (pp. 340-350). Springer, Berlin, Heidelberg.
Guo, H., Sun, F., Cheng, J., Li, Y., & Xu, M. (2016). A novel margin-based measure for directed hill climbing ensemble pruning. Mathematical Problems in Engineering, 2016.
Hill, T., O`Connor, M., & Remus, W. (1996). Neural network models for time series forecasts. Management science, 42(7), pp. 1082-1092.
Hyndman, R. J., & Koehler, A. B. (2006). Another look at measures of forecast accuracy. International journal of forecasting, 22(4), pp. 679-688.
Ke, G., Meng, Q., Finley, T., Wang, T., Chen, W., Ma, W., ... & Liu, T. Y. (2017). Lightgbm: A highly efficient gradient boosting decision tree. In Advances in Neural Information Processing Systems (pp. 3146-3154).
Meddahi, N., & Renault, E. (2004). Temporal aggregation of volatility models. Journal of Econometrics, 119(2), pp. 355-379.
Mendes-Moreira, J., Soares, C., Jorge, A. M., & Sousa, J. F. D. (2012). Ensemble approaches for regression: A survey. Acm computing surveys (csur), 45(1), 10.
Montero-Manso, P., Athanasopoulos, G., Hyndman, R. J., & Talagala, T. S. (2018). FFORMA: Feature-based forecast model averaging. Monash Econometrics and Business Statistics Working Papers, 19(18), 2018-19.
Natekin, A., & Knoll, A. (2013). Gradient boosting machines, a tutorial. Frontiers in neurorobotics, 7, 21.
Nikolopoulos, K., Syntetos, A. A., Boylan, J. E., Petropoulos, F., & Assimakopoulos, V. (2011). An aggregate–disaggregate intermittent demand approach (ADIDA) to forecasting: an empirical proposition and analysis. Journal of the Operational Research Society, 62(3), pp. 544-554.
Opitz, D., & Maclin, R. (1999). Popular ensemble methods: An empirical study. Journal of artificial intelligence research, 11, pp. 169-198.
Regattieri, A., Gamberi, M., Gamberini, R., & Manzini, R. (2005). Managing lumpy demand for aircraft spare parts. Journal of Air Transport Management, 11(6), pp. 426-431.
Ren, Y., Zhang, L., & Suganthan, P. N. (2016). Ensemble classification and regression-recent developments, applications and future directions. IEEE Computational intelligence magazine, 11(1), pp. 41-53.
Caruana, R., Niculescu-Mizil, A., Crew, G., & Ksikes, A. (2004, July). Ensemble selection from libraries of models. In Proceedings of the twenty-first international conference on Machine learning (pp. 18). ACM.
Rostami‐Tabar, B., Babai, M. Z., Syntetos, A., & Ducq, Y. (2013). Demand forecasting by temporal aggregation. Naval Research Logistics (NRL), 60(6), pp. 479-498.
Sill, J., Takács, G., Mackey, L., & Lin, D. (2009). Feature-weighted linear stacking. arXiv preprint arXiv:0911.0460.
Sun, Q., & Pfahringer, B. (2011, December). Bagging ensemble selection. In Australasian Joint Conference on Artificial Intelligence (pp. 251-260). Springer, Berlin, Heidelberg.
Syntetos, A. A., Babai, Z., Boylan, J. E., Kolassa, S., & Nikolopoulos, K. (2016). Supply chain forecasting: Theory, practice, their gap and the future. European Journal of Operational Research, 252(1), pp. 1-26.
Syntetos, A. (2001). Forecasting of intermittent demand (Doctoral dissertation, Brunel University).
Syntetos, A. A., & Boylan, J. E. (2005). The accuracy of intermittent demand estimates. International Journal of forecasting, 21(2), pp. 303-314.
Willemain, T. R., Smart, C. N., Shockor, J. H., & DeSautels, P. A. (1994). Forecasting intermittent demand in manufacturing: a comparative evaluation of Croston`s method. International Journal of forecasting, 10(4), pp. 529-538.
Zhou, Z. H., & Tang, W. (2003, May). Selective ensemble of decision trees. In International Workshop on Rough Sets, Fuzzy Sets, Data Mining, and Granular-Soft Computing (pp. 476-483). Springer, Berlin, Heidelberg.
Zhou, Z. H., Wu, J., & Tang, W. (2010). Corrigendum to “Ensembling neural networks: Many could be better than all”[Artificial Intelligence 137 (1–2)(2002) 239–263]. Artificial Intelligence, 174(18), 1570.
描述 碩士
國立政治大學
資訊管理學系
107356002
資料來源 http://thesis.lib.nccu.edu.tw/record/#G0107356002
資料類型 thesis
dc.contributor.advisor 莊皓鈞<br>周彥君zh_TW
dc.contributor.advisor Chuang, Hao-Chun<br>Chou, Yen-Chunen_US
dc.contributor.author (Authors) 郭泰竹zh_TW
dc.contributor.author (Authors) Kuo, Tai-Chuen_US
dc.creator (作者) 郭泰竹zh_TW
dc.creator (作者) Kuo, Tai-Chuen_US
dc.date (日期) 2019en_US
dc.date.accessioned 7-Aug-2019 16:09:08 (UTC+8)-
dc.date.available 7-Aug-2019 16:09:08 (UTC+8)-
dc.date.issued (上傳時間) 7-Aug-2019 16:09:08 (UTC+8)-
dc.identifier (Other Identifiers) G0107356002en_US
dc.identifier.uri (URI) http://nccur.lib.nccu.edu.tw/handle/140.119/124722-
dc.description (描述) 碩士zh_TW
dc.description (描述) 國立政治大學zh_TW
dc.description (描述) 資訊管理學系zh_TW
dc.description (描述) 107356002zh_TW
dc.description.abstract (摘要) 隨著半導體產業以及電子成品業快速發展,電子零組件之供應成為該產業中重要的一環。而通路商在整條供應鏈當中所扮演的角色尤其重要,需要及時反映下游需求並知會上游,達成良好的需求供給,也需要控制存貨水位,降低存貨成本。根據文獻探討,半導體通路商遇到以下三個困難,包括需求前置時間過長,需求變動量大,以及零組件生命週期短。本研究之研究問題為「如何在需求不穩、資料有限的狀況下,預測前置時間需求?」,希望能透過機器學習之技術,預測未來時間之前置需求量,協助通路商控制良好的存貨水位,提高其營業獲利。
本次研究與亞洲一代表性零組件通路商合作,其在擁有龐大營業額下,也需面對相當高的存貨成本,無論是訂貨超出或是不足,都需負擔存貨與缺貨成本。本研究運用其近三年之資料,結合下游製造廠商對於未來需求之預期數量矩陣,針對通路商零組件資料建立預測情境。目標是更精準預測在未來前置時間,下游廠商對於半導體通路商的零組件需求。
為達成以上目標,本研究使用機器學習演算法以及集成式模型架構,首先就不穩定的需求資料進行Temporal Aggregation,並利用跨產品的訓練樣本,以及不同演算法(包含XGBoost、LightGBM、Random Forest等),配合多種參數隨機組合,訓練多個需求預測模型。接著,利用集成式作法(ensemble)將模型混合,本研究使用的方法為平均法以及爬山法(hill-climbing)結合向前選取法(forward-selection),降低各個單一模型預測所產生的偏誤,經較小樣本的完整資料預測測試,和較大樣本的多缺失資料預測測試,本研究發現,利用所提出的資料建構和集成預測,能較各種單一模型顯著降低需求預測誤差,並較該通路商的既有預測方法,能優化70%-80%的零組件預測,使其更能優化存貨控制。
zh_TW
dc.description.abstract (摘要) As the semi-conductor and electronics industries grow, the supply of components has become an important part of this industrial chain. Particularly, the distributor is the key of the supply chain because it is responsible for holding inventories to meet demand from downstream plants, in addition to sourcing from upstream vendors with long production lead time. According to the related literature, the distributor faces the problems including the long lead time, high demand volatility, and short life time cycle of the components. This study addresses the following question - How to predict demand over long lead time with limited data and unstable demand series?
This study collaborates with a representative electronics distributor in Asia. Despite its large revenue, the distributor operates under high inventory costs due to excessive holdings or shortages. This study first collects data on downstream demand and rolling forecast over three years. After data engineering, we build scenarios from data sets. The objective is to improve accuracy of demand forecasting, such that the distributor can better prepare for volatile demand from downstream production plants.
To achieve the goal, this study takes machine learning and ensemble learning techniques. First, we use temporal aggregation to smooth demand series and reduce noise. We construct cross-item training samples for different algorithms. After training models with random hyper-parameter search, we use model-averaging and hill-climbing methods to ensemble models` predictions. After testing a relatively small samples with complete data and a larger sample with incomplete data, we show that the proposed approaches can reduce prediction errors, compared with single models. Our method also outperforms the distributor`s internal forecast method for 70%-80% of components, such that the company can apply our method for better inventory control.
en_US
dc.description.tableofcontents 第一章 緒論 1
第一節 研究背景與動機 1
第二節 研究問題與目的 2
第二章 文獻回顧 4
第一節 供應鏈預測問題 4
第二節 集成式學習模型 6
第三章 資料與方法 11
第一節 資料與情境敘述 11
第二節 模型與方法 14
第四章 預測結果 21
第五章 進階研究 24
第一節 資料與情境敘述 24
第二節 模型與方法 28
第三節 分析結果 29
第六章 結論 34
第一節 研究結果 34
第二節 研究限制與未來方向 35
參考文獻 37
zh_TW
dc.format.extent 1207484 bytes-
dc.format.mimetype application/pdf-
dc.source.uri (資料來源) http://thesis.lib.nccu.edu.tw/record/#G0107356002en_US
dc.subject (關鍵詞) 電子供應鏈zh_TW
dc.subject (關鍵詞) 需求預測zh_TW
dc.subject (關鍵詞) 集成式學習zh_TW
dc.subject (關鍵詞) Electronics supply chainsen_US
dc.subject (關鍵詞) Demand forecastingen_US
dc.subject (關鍵詞) Ensemble learningen_US
dc.title (題名) 集成式學習應用於電子零組件供應鏈需求預測zh_TW
dc.title (題名) Ensemble learning for demand forecasting in electronics supply chainsen_US
dc.type (資料類型) thesisen_US
dc.relation.reference (參考文獻) Breiman, L. (1996). Stacked regressions. Machine Learning, 24(1), pp. 49–64.
Breiman, L. (2001). Random forests. Machine learning, 45(1), pp. 5-32.
Carbonneau, R., Laframboise, K., & Vahidov, R. (2008). Application of machine learning techniques for supply chain demand forecasting. European Journal of Operational Research, 184(3), pp. 1140-1154.
Chen, T., & Guestrin, C. (2016, August). Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining (pp. 785-794). ACM.
Freund, Y., & Schapire, R. E. (1996, July). Experiments with a new boosting algorithm. In icml (Vol. 96, pp. 148-156).
Friedman, J. H. (2001). Greedy function approximation: a gradient boosting machine. Annals of statistics, pp. 1189-1232.
Graczyk, M., Lasota, T., Trawiński, B., & Trawiński, K. (2010, March). Comparison of bagging, boosting and stacking ensembles applied to real estate appraisal. In Asian conference on intelligent information and database systems (pp. 340-350). Springer, Berlin, Heidelberg.
Guo, H., Sun, F., Cheng, J., Li, Y., & Xu, M. (2016). A novel margin-based measure for directed hill climbing ensemble pruning. Mathematical Problems in Engineering, 2016.
Hill, T., O`Connor, M., & Remus, W. (1996). Neural network models for time series forecasts. Management science, 42(7), pp. 1082-1092.
Hyndman, R. J., & Koehler, A. B. (2006). Another look at measures of forecast accuracy. International journal of forecasting, 22(4), pp. 679-688.
Ke, G., Meng, Q., Finley, T., Wang, T., Chen, W., Ma, W., ... & Liu, T. Y. (2017). Lightgbm: A highly efficient gradient boosting decision tree. In Advances in Neural Information Processing Systems (pp. 3146-3154).
Meddahi, N., & Renault, E. (2004). Temporal aggregation of volatility models. Journal of Econometrics, 119(2), pp. 355-379.
Mendes-Moreira, J., Soares, C., Jorge, A. M., & Sousa, J. F. D. (2012). Ensemble approaches for regression: A survey. Acm computing surveys (csur), 45(1), 10.
Montero-Manso, P., Athanasopoulos, G., Hyndman, R. J., & Talagala, T. S. (2018). FFORMA: Feature-based forecast model averaging. Monash Econometrics and Business Statistics Working Papers, 19(18), 2018-19.
Natekin, A., & Knoll, A. (2013). Gradient boosting machines, a tutorial. Frontiers in neurorobotics, 7, 21.
Nikolopoulos, K., Syntetos, A. A., Boylan, J. E., Petropoulos, F., & Assimakopoulos, V. (2011). An aggregate–disaggregate intermittent demand approach (ADIDA) to forecasting: an empirical proposition and analysis. Journal of the Operational Research Society, 62(3), pp. 544-554.
Opitz, D., & Maclin, R. (1999). Popular ensemble methods: An empirical study. Journal of artificial intelligence research, 11, pp. 169-198.
Regattieri, A., Gamberi, M., Gamberini, R., & Manzini, R. (2005). Managing lumpy demand for aircraft spare parts. Journal of Air Transport Management, 11(6), pp. 426-431.
Ren, Y., Zhang, L., & Suganthan, P. N. (2016). Ensemble classification and regression-recent developments, applications and future directions. IEEE Computational intelligence magazine, 11(1), pp. 41-53.
Caruana, R., Niculescu-Mizil, A., Crew, G., & Ksikes, A. (2004, July). Ensemble selection from libraries of models. In Proceedings of the twenty-first international conference on Machine learning (pp. 18). ACM.
Rostami‐Tabar, B., Babai, M. Z., Syntetos, A., & Ducq, Y. (2013). Demand forecasting by temporal aggregation. Naval Research Logistics (NRL), 60(6), pp. 479-498.
Sill, J., Takács, G., Mackey, L., & Lin, D. (2009). Feature-weighted linear stacking. arXiv preprint arXiv:0911.0460.
Sun, Q., & Pfahringer, B. (2011, December). Bagging ensemble selection. In Australasian Joint Conference on Artificial Intelligence (pp. 251-260). Springer, Berlin, Heidelberg.
Syntetos, A. A., Babai, Z., Boylan, J. E., Kolassa, S., & Nikolopoulos, K. (2016). Supply chain forecasting: Theory, practice, their gap and the future. European Journal of Operational Research, 252(1), pp. 1-26.
Syntetos, A. (2001). Forecasting of intermittent demand (Doctoral dissertation, Brunel University).
Syntetos, A. A., & Boylan, J. E. (2005). The accuracy of intermittent demand estimates. International Journal of forecasting, 21(2), pp. 303-314.
Willemain, T. R., Smart, C. N., Shockor, J. H., & DeSautels, P. A. (1994). Forecasting intermittent demand in manufacturing: a comparative evaluation of Croston`s method. International Journal of forecasting, 10(4), pp. 529-538.
Zhou, Z. H., & Tang, W. (2003, May). Selective ensemble of decision trees. In International Workshop on Rough Sets, Fuzzy Sets, Data Mining, and Granular-Soft Computing (pp. 476-483). Springer, Berlin, Heidelberg.
Zhou, Z. H., Wu, J., & Tang, W. (2010). Corrigendum to “Ensembling neural networks: Many could be better than all”[Artificial Intelligence 137 (1–2)(2002) 239–263]. Artificial Intelligence, 174(18), 1570.
zh_TW
dc.identifier.doi (DOI) 10.6814/NCCU201900450en_US