學術產出-Theses

Article View/Open

Publication Export

Google ScholarTM

政大圖書館

Citation Infomation

  • No doi shows Citation Infomation
題名 以機器學習與規則方法辨識中文民事裁判書結構與爭點 : 以給付扶養費為例
Using Machine Learning and Pattern-Based Methods for Identifying Elements and Issues in Chinese Judgment Documents of Civil Cases : A Case Study with Alimony Cases
作者 林泓任
Lin, Hong-Ren
貢獻者 劉昭麟
Liu, Chao-Lin
林泓任
Lin, Hong-Ren
關鍵詞 自然語言處理
機器學習
深度學習
法學資訊
Natural language pocessing
Machine learning
Deep learning
Legal information
日期 2023
上傳時間 1-Sep-2023 15:23:33 (UTC+8)
摘要 因應近年來給付扶養費案件在家事事件案件中比例上升,為應用於後續法
學資訊、社會科學等領域研究及方便一般民眾閱讀並擷取裁判書中重點,在此我們選擇案由為給付扶養費案之裁判書作為語料並進行文字探勘與分析。

裁判書中除了案件基本資訊也就是給付扶養費案件中的聲請人所提出的主
張及相對人所提出的反駁或主張外,還包含有法院對此次案件的判決與法官如何做出判決的說明。因此本研究透過機器學習技術從裁判書的內文中找出為「聲請人主張」、「相對人主張」以及法院對該次案件的見解,也就是「案件涵攝」,還有該次案件所引用的法條及說明,也就是「法律見解」。

為了找出前述四種裁判書的主要內容分類,本實驗將其作為一種分類問
題,分別以段落及句子為單位進行分類。而在分類器上會分別嘗試使用傳統機器學習模型,此外也使用近年深度學習熱門的 BERT 模型或是將其作為一嵌入層(embedding layer)並接上其他深度學習模型的架構來對裁判書內容進行分類,而這四分類最終在使用深度學習模型後達到 0.816 的 F1 score。

除了「聲請人主張」、「相對人主張」、「法律見解」、「案件涵攝」四大類別的內容外,裁判書中還紀錄另一重要資訊,也就是案件中的「爭點」,其為該次案件中訴訟雙方所提出之主張或法律上的爭執點。初步建構成前述四大類別的分類器後,本研究會進一步找出爭點。此類問題在本研究中會分別轉化為以句為單位的分類問題,或視為一種文句生成問題。

在最後將前述四分類及爭點句的分類模型應用在裁判書判決准駁預測上,
透過比較及分析是否增加前述分類器提供的資訊對於裁判書判決預測的準確
度,而提供爭點及相比起沒有提供爭點的裁判預測模型,準確率自等同亂猜的0.623 提升到 0.877。
To response to the ratio of alimony in foundation’s civil cases is increasing in recent year. It need to apply to legal, social science and other fields and make people easily to read judgement documents. We choice the case of alimony in civil cases as our corpus and retrieve the information and analysis it.

In judgment documents, it contains not only foundation information of case but also contains the opinions of the courts, and uses of laws to reach the final decisions.
In this research, we use machine learning technique to find the pleadings of the applicants, the responses of the opposite parties, opinions of the courts and uses of laws to reach the final decisions.

In order to find four categories, it be seemed as a classification problem. We use counts and short sentences as instances, and try to classify it by traditional machine learning model and the popular deep learning model – BERT or use BERT as embedding layers and connect other deep learning model. At last, the classification model for four categories can reach 0.816 of F1 score.

Except the four categories claimed in previous count, It also has an important information in judgment document – issues. Issues means a point disputed by parties to a lawsuit. This research will also try to find issues, and takes it as five categories classification problem and sentences generation problem.

Finally, the classification model of four categories and issues will be used in judgment prediction. We compare the judgment prediction model with only information mentioned by parties and privies and model with information mentioned by parties and privies and issues. In previous one, it only can 0.623 accuracy and the later can reach 0.877 accuracy.
參考文獻 [1] Ronen Feldman and James Sanger, The Text Mining Handbook, Cambridge University Press.
[2] Mu Yang, Yin-Hsiang Liao, Wei-Yun Ma, “ckip-classic,” [線上]. Available: https://github.com/ckiplab/ckip-classic.
[3] Christopher D. Manning, Hinrich Schütze, Foundations of Statistical Natural Language Processing, MIT Press, 1999.
[4] Gerard Salton, Chris Buckley, Term Weighting Approaches in Automatic Text Retrieval, New York: Cornell University, 1987.
[5] Jenish Dhanani, Rupa Mehta & Dipti Rana, Effective and scalable legal judgment recommendation using pre-learned word embedding, Complex Intell. Syst. 8, 3199–3213, 2022.
[6] 藍家樑, 中文訴訟文書檢索系統雛形實作, 台北市: 國立政治大學 資訊科學學系, 2009.
[7] 蔡惠娟, 營造工程類法學判決書搜尋系統之研發, 台中市: 國立中興大學土木工程學系所, 2016.
[8] Zikun Hu, Xiang Li, Cunchao Tu, Zhiyuan Liu, Maosong Sun, Few-Shot Charge Prediction with Discriminative Legal Attributes, Proceedings of the 27th International Conference on Computational Linguistics, 2018, pp. 487--498.
[9] 何君豪, 階層式分群法在民事裁判要旨分群上之應用, 台北市: 國立政治大學 資訊科學學系, 2007.
[10] 鍾文傑, 陳哲文, 王駿發, 曾世邦, 王宗松, 基於多 BERT 模型之 NLLP 應用於建築工程訴訟之理解與預測, The Association for Computational Linguistics and Chinese Language Processing, 2020, pp. 72--84.
[11] 紀幸辰, 一個基於機器學習的醫療法判決書預測系統使用具文本相似性的法條分類決策樹, 台北市: 國立臺北大學 資訊工程學系, 2018.
[12] 林岡毅, 以資訊技術分析我國離婚贍養費相關裁判, 台北市: 國立臺灣大學科際整合法律學研究所, 2018.
[13] 陳聰富, 民法概要.
[14] “司法院資料開放平臺,” [線上]. Available: https://opendata.judicial.gov.tw/.
[15] Yi-Fan Liu, Chao-Lin Liu, Chieh Yang, 以民事訴訟之爭點分群為基礎的類似案件搜尋系統, The 34th Conference on Computational Linguistics and Speech Processing (ROCLING 2022).
[16] “electronjs,” OpenJS Foundation, 2021. [線上]. Available: https://www.electronjs.org/.
[17] J. Sun, “结巴中文分词,” [線上]. Available: https://github.com/fxsjy/jieba.
[18] Pedregosa et al., Scikit-learn: Machine Learning in Python,, JMLR 12, 2011.
[19] Subhajit Dey Sarkar, Saptarsi Goswami, Aman Agarwal, and Javed Aktar, A Novel Feature Selection Technique for Text Classification Using Naive Bayes, 2014.
[20] S. Fachrurrozi, Muljono, G. F. Shidik, A. Z. Fanani, Purwanto and F. A. Zami, Increasing Accuracy of Support Vector Machine (SVM) By Applying N-Gram and Chi-Square Feature Selection for Text Classification, International Seminar on Application for Technology of Information and Communication (iSemantic), 2021.
[21] Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova, BERT: Pretraining of Deep Bidirectional Transformers for Language Understanding, 2018.
[22] Ashish Vaswani, Attention Is All You Need, 2017.
[23] Guozheng Rao, Weihang Huang, Zhiyong Feng, Qiong Cong, LSTM with sentence representations for document-level sentiment classification, Neurocomputing, 2018, pp. 49 -- 57.
[24] Colin Raffel, Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer, The Journal of Machine Learning Research, 2020, p. pp 5485–5551.
[25] J. Su, T5 PEGASUS, 2021.
[26] Jingqing Zhang, PEGASUS: pre-training with extracted gapsentences for abstractive summarization, JMLR.org, 2020.
[27] Wei-Zhi Liu, Po-Hsien Wu, Hong-Ren Lin, Chao-Lin Liu, Predicting Judgments and Grants for Civil Cases of Alimony for the Elderly, Proceedings of the 34th Conference on Computational Linguistics and Speech Processing (ROCLING 2022), 2022, pp. 121--128.
[28] Zhongguo Wang, Bao Zhang, Chinese Text Similarity Calculation Model Based on Multi-Attention Siamese Bi-LSTM, CSSE 2021: 2021 4th International Conference on Computer Science and Software Engineering, 2021, pp. 93--98.
[29] Shang Li, Hongli Zhang, Lin Ye, Xiaoding Guo, Binxing Fang, MANN: A Multichannel Attentive Neural Network for Legal Judgment Prediction, IEEE Access, 2018, pp. 151144--151155.
[30] TensorFlow code and pre-trained models for BERT, “TensorFlow code and pretrained models for BERT,” [線上]. Available: https://github.com/googleresearch/bert.
Ilya Loshchilov, Frank Hutter, “Decoupled Weight Decay Regularization,” 14 Nov 2017.
[31] “TensorFlow,” 2015. [線上]. Available: https://github.com/tensorflow/tensorflow.
[32] Wei-Zhi Liu, Po-Hsien Wu, Hong-Ren Lin, Chao-Lin Liu, 老年扶養費請求案件之准駁及扶養金額預測, Taiwan: Rocling, 2022.
[33] Salton, G. and McGill, M. J, Introduction to modern information retrieval, McGraw-Hill, 1983.
描述 碩士
國立政治大學
資訊科學系
109753156
資料來源 http://thesis.lib.nccu.edu.tw/record/#G0109753156
資料類型 thesis
dc.contributor.advisor 劉昭麟zh_TW
dc.contributor.advisor Liu, Chao-Linen_US
dc.contributor.author (Authors) 林泓任zh_TW
dc.contributor.author (Authors) Lin, Hong-Renen_US
dc.creator (作者) 林泓任zh_TW
dc.creator (作者) Lin, Hong-Renen_US
dc.date (日期) 2023en_US
dc.date.accessioned 1-Sep-2023 15:23:33 (UTC+8)-
dc.date.available 1-Sep-2023 15:23:33 (UTC+8)-
dc.date.issued (上傳時間) 1-Sep-2023 15:23:33 (UTC+8)-
dc.identifier (Other Identifiers) G0109753156en_US
dc.identifier.uri (URI) http://nccur.lib.nccu.edu.tw/handle/140.119/147028-
dc.description (描述) 碩士zh_TW
dc.description (描述) 國立政治大學zh_TW
dc.description (描述) 資訊科學系zh_TW
dc.description (描述) 109753156zh_TW
dc.description.abstract (摘要) 因應近年來給付扶養費案件在家事事件案件中比例上升,為應用於後續法
學資訊、社會科學等領域研究及方便一般民眾閱讀並擷取裁判書中重點,在此我們選擇案由為給付扶養費案之裁判書作為語料並進行文字探勘與分析。

裁判書中除了案件基本資訊也就是給付扶養費案件中的聲請人所提出的主
張及相對人所提出的反駁或主張外,還包含有法院對此次案件的判決與法官如何做出判決的說明。因此本研究透過機器學習技術從裁判書的內文中找出為「聲請人主張」、「相對人主張」以及法院對該次案件的見解,也就是「案件涵攝」,還有該次案件所引用的法條及說明,也就是「法律見解」。

為了找出前述四種裁判書的主要內容分類,本實驗將其作為一種分類問
題,分別以段落及句子為單位進行分類。而在分類器上會分別嘗試使用傳統機器學習模型,此外也使用近年深度學習熱門的 BERT 模型或是將其作為一嵌入層(embedding layer)並接上其他深度學習模型的架構來對裁判書內容進行分類,而這四分類最終在使用深度學習模型後達到 0.816 的 F1 score。

除了「聲請人主張」、「相對人主張」、「法律見解」、「案件涵攝」四大類別的內容外,裁判書中還紀錄另一重要資訊,也就是案件中的「爭點」,其為該次案件中訴訟雙方所提出之主張或法律上的爭執點。初步建構成前述四大類別的分類器後,本研究會進一步找出爭點。此類問題在本研究中會分別轉化為以句為單位的分類問題,或視為一種文句生成問題。

在最後將前述四分類及爭點句的分類模型應用在裁判書判決准駁預測上,
透過比較及分析是否增加前述分類器提供的資訊對於裁判書判決預測的準確
度,而提供爭點及相比起沒有提供爭點的裁判預測模型,準確率自等同亂猜的0.623 提升到 0.877。
zh_TW
dc.description.abstract (摘要) To response to the ratio of alimony in foundation’s civil cases is increasing in recent year. It need to apply to legal, social science and other fields and make people easily to read judgement documents. We choice the case of alimony in civil cases as our corpus and retrieve the information and analysis it.

In judgment documents, it contains not only foundation information of case but also contains the opinions of the courts, and uses of laws to reach the final decisions.
In this research, we use machine learning technique to find the pleadings of the applicants, the responses of the opposite parties, opinions of the courts and uses of laws to reach the final decisions.

In order to find four categories, it be seemed as a classification problem. We use counts and short sentences as instances, and try to classify it by traditional machine learning model and the popular deep learning model – BERT or use BERT as embedding layers and connect other deep learning model. At last, the classification model for four categories can reach 0.816 of F1 score.

Except the four categories claimed in previous count, It also has an important information in judgment document – issues. Issues means a point disputed by parties to a lawsuit. This research will also try to find issues, and takes it as five categories classification problem and sentences generation problem.

Finally, the classification model of four categories and issues will be used in judgment prediction. We compare the judgment prediction model with only information mentioned by parties and privies and model with information mentioned by parties and privies and issues. In previous one, it only can 0.623 accuracy and the later can reach 0.877 accuracy.
en_US
dc.description.tableofcontents 第一章 緒論 1
1.1 研究動機 1
1.2 研究目的 1
1.3 論文架構 2
1.4 研究成果 2
第二章 文獻回顧 4
2.1 裁判書檢索及推薦系統 4
2.2 裁判書分類及分群 5
2.3 案件要素分析及結果預測 5
2.4 研究設計 5
第三章 研究方法 7
3.1 研究語料與研究架構 7
3.1.1 研究語料 7
3.1.2 研究架構 8
3.2 裁判書前處理與標記 9
3.2.1 原始資料分析與篩選 9
3.2.2 事實、理由段落分段 11
3.2.3 語料清理 13
3.2.4 斷句 14
3.2.5 規則與正規表示法標記段落類型 17
3.2.6 人工標記 19
3.3 向量化 24
3.3.1 n-gram 斷詞與詞頻 24
3.3.2 n-gram + tf-idf 句向量 24
3.3.3 降維 25
3.3.4 BERT 句向量 26
3.3.5 BERT 詞向量 27
3.4 文章結構分類模型 28
3.4.1 以段為單位的文章結構分類模型 28
3.4.2 句分類模型 30
3.4.3 裁判書爭點搜尋 33
3.5 架構分類模型應用於判決預測 35
3.5.1 BERT + multi-head attention 判決預測模型 35
3.5.2 BERT + Multichannel Attentive Neural Network 判決預測模型 37
第四章 實驗結果與分析 38
4.1 實驗語料 38
4.1.1 輸入特徵 38
4.1.2 語料標記 39
4.2 實驗評估方法 42
4.2.1 分類任務 42
4.2.2 生成任務 42
4.3 段落四分類模型 43
4.3.1 單純貝氏分類器 45
4.3.2 邏輯回歸 46
4.3.3 支持向量機 47
4.4 句子四分類模型 50
4.4.1 單純貝氏分類器 52
4.4.2 邏輯回歸 52
4.4.3 支持向量機 54
4.4.4 BERT + dense/LSTM 55
4.4.5 BERT + dense 增加前後文資訊 56
4.4.6 BERT + SR-LSTM 59
4.4.7 BERT + SR-LSTM 於大段落數為 6 的裁判書 61
4.5 爭點搜尋 62
4.5.1 分類模型 62
4.5.2 生成式摘要模型生成爭點句 63
4.6 判決預測 67
4.6.1 以聲請人與相對人段落預測 69
4.6.2 以聲請人、相對人段落加上爭點句預測 72
第五章 結論與未來展望 74
5.1 結論 74
5.2 未來展望 75
參考文獻 76
附錄A 裁判書原始資料 79
附錄B 論文口試相關討論 85
zh_TW
dc.format.extent 4367921 bytes-
dc.format.mimetype application/pdf-
dc.source.uri (資料來源) http://thesis.lib.nccu.edu.tw/record/#G0109753156en_US
dc.subject (關鍵詞) 自然語言處理zh_TW
dc.subject (關鍵詞) 機器學習zh_TW
dc.subject (關鍵詞) 深度學習zh_TW
dc.subject (關鍵詞) 法學資訊zh_TW
dc.subject (關鍵詞) Natural language pocessingen_US
dc.subject (關鍵詞) Machine learningen_US
dc.subject (關鍵詞) Deep learningen_US
dc.subject (關鍵詞) Legal informationen_US
dc.title (題名) 以機器學習與規則方法辨識中文民事裁判書結構與爭點 : 以給付扶養費為例zh_TW
dc.title (題名) Using Machine Learning and Pattern-Based Methods for Identifying Elements and Issues in Chinese Judgment Documents of Civil Cases : A Case Study with Alimony Casesen_US
dc.type (資料類型) thesisen_US
dc.relation.reference (參考文獻) [1] Ronen Feldman and James Sanger, The Text Mining Handbook, Cambridge University Press.
[2] Mu Yang, Yin-Hsiang Liao, Wei-Yun Ma, “ckip-classic,” [線上]. Available: https://github.com/ckiplab/ckip-classic.
[3] Christopher D. Manning, Hinrich Schütze, Foundations of Statistical Natural Language Processing, MIT Press, 1999.
[4] Gerard Salton, Chris Buckley, Term Weighting Approaches in Automatic Text Retrieval, New York: Cornell University, 1987.
[5] Jenish Dhanani, Rupa Mehta & Dipti Rana, Effective and scalable legal judgment recommendation using pre-learned word embedding, Complex Intell. Syst. 8, 3199–3213, 2022.
[6] 藍家樑, 中文訴訟文書檢索系統雛形實作, 台北市: 國立政治大學 資訊科學學系, 2009.
[7] 蔡惠娟, 營造工程類法學判決書搜尋系統之研發, 台中市: 國立中興大學土木工程學系所, 2016.
[8] Zikun Hu, Xiang Li, Cunchao Tu, Zhiyuan Liu, Maosong Sun, Few-Shot Charge Prediction with Discriminative Legal Attributes, Proceedings of the 27th International Conference on Computational Linguistics, 2018, pp. 487--498.
[9] 何君豪, 階層式分群法在民事裁判要旨分群上之應用, 台北市: 國立政治大學 資訊科學學系, 2007.
[10] 鍾文傑, 陳哲文, 王駿發, 曾世邦, 王宗松, 基於多 BERT 模型之 NLLP 應用於建築工程訴訟之理解與預測, The Association for Computational Linguistics and Chinese Language Processing, 2020, pp. 72--84.
[11] 紀幸辰, 一個基於機器學習的醫療法判決書預測系統使用具文本相似性的法條分類決策樹, 台北市: 國立臺北大學 資訊工程學系, 2018.
[12] 林岡毅, 以資訊技術分析我國離婚贍養費相關裁判, 台北市: 國立臺灣大學科際整合法律學研究所, 2018.
[13] 陳聰富, 民法概要.
[14] “司法院資料開放平臺,” [線上]. Available: https://opendata.judicial.gov.tw/.
[15] Yi-Fan Liu, Chao-Lin Liu, Chieh Yang, 以民事訴訟之爭點分群為基礎的類似案件搜尋系統, The 34th Conference on Computational Linguistics and Speech Processing (ROCLING 2022).
[16] “electronjs,” OpenJS Foundation, 2021. [線上]. Available: https://www.electronjs.org/.
[17] J. Sun, “结巴中文分词,” [線上]. Available: https://github.com/fxsjy/jieba.
[18] Pedregosa et al., Scikit-learn: Machine Learning in Python,, JMLR 12, 2011.
[19] Subhajit Dey Sarkar, Saptarsi Goswami, Aman Agarwal, and Javed Aktar, A Novel Feature Selection Technique for Text Classification Using Naive Bayes, 2014.
[20] S. Fachrurrozi, Muljono, G. F. Shidik, A. Z. Fanani, Purwanto and F. A. Zami, Increasing Accuracy of Support Vector Machine (SVM) By Applying N-Gram and Chi-Square Feature Selection for Text Classification, International Seminar on Application for Technology of Information and Communication (iSemantic), 2021.
[21] Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova, BERT: Pretraining of Deep Bidirectional Transformers for Language Understanding, 2018.
[22] Ashish Vaswani, Attention Is All You Need, 2017.
[23] Guozheng Rao, Weihang Huang, Zhiyong Feng, Qiong Cong, LSTM with sentence representations for document-level sentiment classification, Neurocomputing, 2018, pp. 49 -- 57.
[24] Colin Raffel, Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer, The Journal of Machine Learning Research, 2020, p. pp 5485–5551.
[25] J. Su, T5 PEGASUS, 2021.
[26] Jingqing Zhang, PEGASUS: pre-training with extracted gapsentences for abstractive summarization, JMLR.org, 2020.
[27] Wei-Zhi Liu, Po-Hsien Wu, Hong-Ren Lin, Chao-Lin Liu, Predicting Judgments and Grants for Civil Cases of Alimony for the Elderly, Proceedings of the 34th Conference on Computational Linguistics and Speech Processing (ROCLING 2022), 2022, pp. 121--128.
[28] Zhongguo Wang, Bao Zhang, Chinese Text Similarity Calculation Model Based on Multi-Attention Siamese Bi-LSTM, CSSE 2021: 2021 4th International Conference on Computer Science and Software Engineering, 2021, pp. 93--98.
[29] Shang Li, Hongli Zhang, Lin Ye, Xiaoding Guo, Binxing Fang, MANN: A Multichannel Attentive Neural Network for Legal Judgment Prediction, IEEE Access, 2018, pp. 151144--151155.
[30] TensorFlow code and pre-trained models for BERT, “TensorFlow code and pretrained models for BERT,” [線上]. Available: https://github.com/googleresearch/bert.
Ilya Loshchilov, Frank Hutter, “Decoupled Weight Decay Regularization,” 14 Nov 2017.
[31] “TensorFlow,” 2015. [線上]. Available: https://github.com/tensorflow/tensorflow.
[32] Wei-Zhi Liu, Po-Hsien Wu, Hong-Ren Lin, Chao-Lin Liu, 老年扶養費請求案件之准駁及扶養金額預測, Taiwan: Rocling, 2022.
[33] Salton, G. and McGill, M. J, Introduction to modern information retrieval, McGraw-Hill, 1983.
zh_TW