學術產出-Theses

Article View/Open

Publication Export

Google ScholarTM

政大圖書館

Citation Infomation

  • No doi shows Citation Infomation
題名 波形時間序列經由小波轉換應用HuBERT於外匯市場預測
Time Series Data as Waveform Input with Wavelet Transform Using HuBERT on Foreign Exchange Market Prediction
作者 黃乾唐
Huang, Chien-Tang
貢獻者 蔡炎龍
Tsai, Yen-Lung
黃乾唐
Huang Chien-Tang
關鍵詞 Transformer
深度學習
語音辨識
HuBERT
貨幣對
小波轉換
外匯市場
Transformer
Deep Learning
Speech Recognition
HuBERT
Currency Pair
Foreign Exchange Market
Wavelet Transform
日期 2023
上傳時間 2-May-2023 15:04:55 (UTC+8)
摘要 近期在人工智慧研究中,Transformer,一種基於自注意力機制架構的模型在整個人工智慧研究裡面舉足輕重,最初是應用在自然語言處理上,有名的語言模型像是BERT或GPT等等都是使用此架;並且被廣泛的應用於其他研究方面,像是圖像分析、語音辨識等等。由於想解決人工資料標籤的問題,自監督式學習成為深度學習中一個很重要的研究方法;特徵學習是讓模型學習到資料的潛在向量,而無需像傳統方法去學習標籤。在這篇文章中,我們將使用Hidden Unit BERT (HuBERT) 一種基於自監督學習方法的語音辨識模型,應用於外匯市場的預測上。在資料處理方面,我們還採用小波轉換去比較,是否在價格資料上也擁有去除雜訊的效果。結果顯示HuBERT不僅在語音辨識上表現出色,而且在數值型時間序列資料上也有相當的表現。
Transformer, a self-attention-based architecture, has become the state-of-the-art model for most AI research. Initially used for natural language processing tasks, it has since been applied to other fields such as computer vision and speech, with models such as BERT and GPT based on it exhibiting exceptional performance. With the challenge of data labeling, self-supervised learning has emerged as an important approach in deep learning research, enabling models to learn latent vectors of data without the need for end-to-end labeling targets. In this paper, we explore the use of Hidden Units BERT (HuBERT), a speech representation model trained with a self-supervised approach, on forex data, which is numerical. We also incorporate Wavelet Transform to denoise the data, comparing its performance to that of numerical data. Our results show that HuBERT not only performs exceptionally well on speech tasks but also exhibits promising results on numerical time series data.
參考文獻 [1] Mohammad Zoynul Abedin, Mahmudul Hasan Moon, M Kabir Hassan, and Petr Hajek.
     Deep learning-based exchange rate prediction during the covid-19 pandemic. Annals of Operations Research, pages 1–52, 2021.
     [2] Mark Chen, Alec Radford, Rewon Child, Jeffrey Wu, Heewoo Jun, David Luan, and Ilya Sutskever. Generative pretraining from pixels. In International conference on machine
     learning, pages 1691–1703. PMLR, 2020.
     [3] Alexander Jakob Dautel, Wolfgang Karl Härdle, Stefan Lessmann, and Hsin-Vonn Seow.
     Forex exchange rate forecasting using deep recurrent neural networks. Digital Finance, 2:69–96, 2020.
     [4] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding, 2018.
     [5] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain
     Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale, 2020.
     [6] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition, 2015.
     [7] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
     [8] Wei-Ning Hsu, Benjamin Bolte, Yao-Hung Hubert Tsai, Kushal Lakhotia, Ruslan Salakhutdinov, and Abdelrahman Mohamed. Hubert: Self-supervised speech representation learning by masked prediction of hidden units, 2021.
     [9] Alec Radford, Karthik Narasimhan, Tim Salimans, Ilya Sutskever, et al. Improving language understanding by generative pre-training. 2018.
     [10] David E Rumelhart, Geoffrey E Hinton, and Ronald J Williams. Learning representations by back-propagating errors. nature, 323(6088):533–536, 1986.
     [11] Ilya Sutskever, Oriol Vinyals, and Quoc V. Le. Sequence to sequence learning with neural networks, 2014.
     [12] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need, 2017.
     [13] Shu wen Yang, Po-Han Chi, Yung-Sung Chuang, Cheng-I Jeff Lai, Kushal Lakhotia, Yist Y. Lin, Andy T. Liu, Jiatong Shi, Xuankai Chang, Guan-Ting Lin, Tzu-Hsien Huang,
     Wei-Cheng Tseng, Ko tik Lee, Da-Rong Liu, Zili Huang, Shuyan Dong, Shang-Wen Li, Shinji Watanabe, Abdelrahman Mohamed, and Hung yi Lee. Superb: Speech processing universal performance benchmark, 2021.
描述 碩士
國立政治大學
應用數學系
109751001
資料來源 http://thesis.lib.nccu.edu.tw/record/#G0109751001
資料類型 thesis
dc.contributor.advisor 蔡炎龍zh_TW
dc.contributor.advisor Tsai, Yen-Lungen_US
dc.contributor.author (Authors) 黃乾唐zh_TW
dc.contributor.author (Authors) Huang Chien-Tangen_US
dc.creator (作者) 黃乾唐zh_TW
dc.creator (作者) Huang, Chien-Tangen_US
dc.date (日期) 2023en_US
dc.date.accessioned 2-May-2023 15:04:55 (UTC+8)-
dc.date.available 2-May-2023 15:04:55 (UTC+8)-
dc.date.issued (上傳時間) 2-May-2023 15:04:55 (UTC+8)-
dc.identifier (Other Identifiers) G0109751001en_US
dc.identifier.uri (URI) http://nccur.lib.nccu.edu.tw/handle/140.119/144596-
dc.description (描述) 碩士zh_TW
dc.description (描述) 國立政治大學zh_TW
dc.description (描述) 應用數學系zh_TW
dc.description (描述) 109751001zh_TW
dc.description.abstract (摘要) 近期在人工智慧研究中,Transformer,一種基於自注意力機制架構的模型在整個人工智慧研究裡面舉足輕重,最初是應用在自然語言處理上,有名的語言模型像是BERT或GPT等等都是使用此架;並且被廣泛的應用於其他研究方面,像是圖像分析、語音辨識等等。由於想解決人工資料標籤的問題,自監督式學習成為深度學習中一個很重要的研究方法;特徵學習是讓模型學習到資料的潛在向量,而無需像傳統方法去學習標籤。在這篇文章中,我們將使用Hidden Unit BERT (HuBERT) 一種基於自監督學習方法的語音辨識模型,應用於外匯市場的預測上。在資料處理方面,我們還採用小波轉換去比較,是否在價格資料上也擁有去除雜訊的效果。結果顯示HuBERT不僅在語音辨識上表現出色,而且在數值型時間序列資料上也有相當的表現。zh_TW
dc.description.abstract (摘要) Transformer, a self-attention-based architecture, has become the state-of-the-art model for most AI research. Initially used for natural language processing tasks, it has since been applied to other fields such as computer vision and speech, with models such as BERT and GPT based on it exhibiting exceptional performance. With the challenge of data labeling, self-supervised learning has emerged as an important approach in deep learning research, enabling models to learn latent vectors of data without the need for end-to-end labeling targets. In this paper, we explore the use of Hidden Units BERT (HuBERT), a speech representation model trained with a self-supervised approach, on forex data, which is numerical. We also incorporate Wavelet Transform to denoise the data, comparing its performance to that of numerical data. Our results show that HuBERT not only performs exceptionally well on speech tasks but also exhibits promising results on numerical time series data.en_US
dc.description.tableofcontents 致謝 i
     中文摘要 ii
     Abstract iii
     Contents iv
     List of Tables vi
     List of Figures vii
     1 Introduction 1
     2 Related Work 3
     3 Deep Learning 4
     3.1 Neurons and Neural Networks 4
     3.2 Activation Functions 5
     3.3 Loss Function 8
     3.4 Optimization 9
     4 Transformer 11
     4.1 Recurrent Nueral Network 11
     4.2 Attention 12
     4.3 Transformer 14
     4.4 BERT 17
     4.5 HuBERT 18
     5 Experiments 20
     5.1 Wavelet Transform 20
     5.2 Data Processing 24
     5.3 Evaluation 27
     6 Conclusion 32
     Bibliography 34
     Appendix A Result Tables 36
zh_TW
dc.source.uri (資料來源) http://thesis.lib.nccu.edu.tw/record/#G0109751001en_US
dc.subject (關鍵詞) Transformerzh_TW
dc.subject (關鍵詞) 深度學習zh_TW
dc.subject (關鍵詞) 語音辨識zh_TW
dc.subject (關鍵詞) HuBERTzh_TW
dc.subject (關鍵詞) 貨幣對zh_TW
dc.subject (關鍵詞) 小波轉換zh_TW
dc.subject (關鍵詞) 外匯市場zh_TW
dc.subject (關鍵詞) Transformeren_US
dc.subject (關鍵詞) Deep Learningen_US
dc.subject (關鍵詞) Speech Recognitionen_US
dc.subject (關鍵詞) HuBERTen_US
dc.subject (關鍵詞) Currency Pairen_US
dc.subject (關鍵詞) Foreign Exchange Marketen_US
dc.subject (關鍵詞) Wavelet Transformen_US
dc.title (題名) 波形時間序列經由小波轉換應用HuBERT於外匯市場預測zh_TW
dc.title (題名) Time Series Data as Waveform Input with Wavelet Transform Using HuBERT on Foreign Exchange Market Predictionen_US
dc.type (資料類型) thesisen_US
dc.relation.reference (參考文獻) [1] Mohammad Zoynul Abedin, Mahmudul Hasan Moon, M Kabir Hassan, and Petr Hajek.
     Deep learning-based exchange rate prediction during the covid-19 pandemic. Annals of Operations Research, pages 1–52, 2021.
     [2] Mark Chen, Alec Radford, Rewon Child, Jeffrey Wu, Heewoo Jun, David Luan, and Ilya Sutskever. Generative pretraining from pixels. In International conference on machine
     learning, pages 1691–1703. PMLR, 2020.
     [3] Alexander Jakob Dautel, Wolfgang Karl Härdle, Stefan Lessmann, and Hsin-Vonn Seow.
     Forex exchange rate forecasting using deep recurrent neural networks. Digital Finance, 2:69–96, 2020.
     [4] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding, 2018.
     [5] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain
     Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale, 2020.
     [6] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition, 2015.
     [7] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
     [8] Wei-Ning Hsu, Benjamin Bolte, Yao-Hung Hubert Tsai, Kushal Lakhotia, Ruslan Salakhutdinov, and Abdelrahman Mohamed. Hubert: Self-supervised speech representation learning by masked prediction of hidden units, 2021.
     [9] Alec Radford, Karthik Narasimhan, Tim Salimans, Ilya Sutskever, et al. Improving language understanding by generative pre-training. 2018.
     [10] David E Rumelhart, Geoffrey E Hinton, and Ronald J Williams. Learning representations by back-propagating errors. nature, 323(6088):533–536, 1986.
     [11] Ilya Sutskever, Oriol Vinyals, and Quoc V. Le. Sequence to sequence learning with neural networks, 2014.
     [12] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need, 2017.
     [13] Shu wen Yang, Po-Han Chi, Yung-Sung Chuang, Cheng-I Jeff Lai, Kushal Lakhotia, Yist Y. Lin, Andy T. Liu, Jiatong Shi, Xuankai Chang, Guan-Ting Lin, Tzu-Hsien Huang,
     Wei-Cheng Tseng, Ko tik Lee, Da-Rong Liu, Zili Huang, Shuyan Dong, Shang-Wen Li, Shinji Watanabe, Abdelrahman Mohamed, and Hung yi Lee. Superb: Speech processing universal performance benchmark, 2021.
zh_TW