dc.contributor | 資科系 | - |
dc.creator (作者) | 劉吉軒 | - |
dc.creator (作者) | Liu , Jyi-Shane | - |
dc.creator (作者) | Lai, Huei-Ling | - |
dc.creator (作者) | Hsu, Hsiao-Ling | - |
dc.creator (作者) | Lin, Chia-Hung | - |
dc.creator (作者) | Chen, Yanhong | - |
dc.date (日期) | 2021-03 | - |
dc.date.accessioned | 27-Oct-2021 10:58:54 (UTC+8) | - |
dc.date.available | 27-Oct-2021 10:58:54 (UTC+8) | - |
dc.date.issued (上傳時間) | 27-Oct-2021 10:58:54 (UTC+8) | - |
dc.identifier.uri (URI) | http://nccur.lib.nccu.edu.tw/handle/140.119/137555 | - |
dc.description.abstract (摘要) | While word sense disambiguation (WSD) has been extensively studied in natural language processing, such a task in low-resource languages still receives little attention. Findings based on a few dominant languages may lead to narrow applications. A language-specific WSD system is in need to implement in low-resource languages, for instance, in Taiwan Hakka. This study examines the performance of DNN and Bi-LSTM in WSD tasks on polysemous BUNin Taiwan Hakka. Both models are trained and tested on a small amount of hand-crafted labeled data. Two experiments are designed with four kinds of input features and two window spans to explore what information is needed for the models to achieve their best performance. The results show that to achieve the best performance, DNN and Bi-LSTM models prefer different kinds of input features and window spans. | - |
dc.format.extent | 126 bytes | - |
dc.format.mimetype | text/html | - |
dc.relation (關聯) | International Journal of Asian Language Processing | - |
dc.title (題名) | Supervised Word Sense Disambiguation on Polysemy with Neural Network Models: A Case Study of BUN in Taiwan Hakka | - |
dc.type (資料類型) | article | - |
dc.identifier.doi (DOI) | 10.1142/s2717554520500113 | - |
dc.doi.uri (DOI) | https://doi.org/10.1142/s2717554520500113 | - |