Please use this identifier to cite or link to this item: https://ah.lib.nccu.edu.tw/handle/140.119/136850
題名: 機器學習可解釋技術在商業智慧中對使用者信任之影響
The Effect of Explanation on User Trust in Business Intelligence
作者: 侯亮宇
Hou, Liang-Yu
貢獻者: 林怡伶
Lin, Yi-Ling
侯亮宇
Hou, Liang-Yu
關鍵詞: 人機互動
機器學習
資訊視覺化
可解釋性人工智慧
信任
Human computer interaction
machine learning
information visualization
trust
explainable artificial intelligence
XAI
日期: 2021
上傳時間: 2-九月-2021
摘要: 近年來機器學習引發了人工智慧 (Artificial Intelligence, AI) 應用的新趨勢。 AI 被應用於越來越複雜的任務和領域中。然而,大多數 AI 模型都在黑盒(Black box)中運行,導致人們難以理解或是分辨機器的運作以及決策過程。目前,可解 釋性人工智慧(Explainable Artificial Intelligence, XAI),大多著重於底層演算法的 解釋,並且集中於解釋圖形識別的結果。針對終端使用者的 XAI 應用則較多專 注於支援醫療保健領域的人類決策,少有研究調查商業領域的 AI 應用程序如何 與解釋性技術相結合。本研究以商業應用上終端使用者為中心為實際業務領域中 運用 AI 技術提出了一個通用的解釋框架。該框架基於商業智慧(Business Intelligence,BI) 所開發,為終端使用者提供在機器學習不同階段的完整解釋。為 了實踐我們的框架,我們在一個航空公司行李重量預測案例上應用了這個解釋性 架構。最後,為衡量該框架實踐後的有效性,我們在 Amazon Mechanical Turk 上 進行了實驗。我們的結果表明,使用解釋性框架的參與者對模型預測更有信心, 並且更信任系統,更願意採用系統提供的建議。我們的研究使企業能夠擴展他們 的商業智能,並結合這個解釋框架的不同階段,以提高機器學習技術在商業應用 中的透明度和可靠性。
Recently, machine learning has sparked a new trend in artificial intelligence (AI) applications. AI is applied to increasingly complex tasks and in many areas. Most AI models are running in a black box resulting in difficulty for understanding. From image recognition to sentiment analysis, XAI is used to support human decision-making in the healthcare domain, yet little research has been done to investigate how AI applications in the commercial domain can be integrated with explanatory techniques. This study proposes a generalized interpretative framework for end-user-centric applications in the business domain. The framework enables the provision of complete explanations to end users at different stages based on business intelligence. To validate our framework, we applied this explanatory framework in practice using an airline baggage weight prediction case. Finally, in order to measure the effectiveness of the framework in practice, we conducted an online experiment at Mturk. Our results show that participants who use the explanatory framework have more confidence in the model predictions, trust the system, and are more willing to adopt the recommendations provided by the system. Our research allows companies to extend their business intelligence and combine different stages of this explanatory framework to improve the transparency and reliability of machine learning technology in business applications.
參考文獻: Adadi, A., & Berrada, M. (2018). Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access.\nAlber, M., Lapuschkin, S., Seegerer, P., Hägele, M., Schütt, K. T., Montavon, G., Samek, W., Müller, K. R., Dähne, S., & Kindermans, P. J. (2019). INNvestigate neural networks! Journal of Machine Learning Research.\nAllen, W. L. (2018). Visual brokerage: Communicating data and research through visualisation. Public Understanding of Science, 27(8), 906–922.\nAmershi, S., Weld, D., Vorvoreanu, M., Fourney, A., Nushi, B., Collisson, P., Suh, J., Iqbal, S., Bennett, P. N., Inkpen, K., Teevan, J., Kikin-Gil, R., & Horvitz, E. (2019). Guidelines for human-AI interaction. Conference on Human Factors in Computing Systems - Proceedings.\nApley, D. W., & Zhu, J. (2020). Visualizing the effects of predictor variables in black box supervised learning models. Journal of the Royal Statistical Society. Series B: Statistical Methodology.\nBach, S., Binder, A., Montavon, G., Klauschen, F., Müller, K. R., & Samek, W. (2015). On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS ONE.\nBarredo Arrieta, A., Díaz-Rodríguez, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., Garcia, S., Gil-Lopez, S., Molina, D., Benjamins, R., Chatila, R., & Herrera, F. (2020a). Explainable Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion, 58, 82–115.\nBarredo Arrieta, A., Díaz-Rodríguez, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., Garcia, S., Gil-Lopez, S., Molina, D., Benjamins, R., Chatila, R., & Herrera, F. (2020b). Explainable Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information\nFusion.\nBien, J., & Tibshirani, R. (2011). Prototype selection for interpretable classification. Annals of Applied Statistics.\nBorkin, M. A., Vo, A. A., Bylinskii, Z., Isola, P., Sunkavalli, S., Oliva, A., & Pfister, H. (2013). What makes a visualization memorable. IEEE Transactions on\nVisualization and Computer Graphics.\nBruls, M., Huizing, K., & van Wijk, J. J. (2000). Squarified Treemaps.\nBurton, B., Geishecker, L., Schlegel, K., Hostmann, B., Austin, T., Herschel, G., Rayner, N., Sallam, R. L., Richardson, J., Hagerty, J., & Hostmann, B. (2006). Magic Quadrant for Business Intelligence Platforms WHAT YOU NEED TO KNOW. Gartner Research, January, 1–5. http://www.gartner.com/technology/about/ombudsman/omb_guide2.jsp\nCai, C. J., Jongejan, J., & Holbrook, J. (2019). The effects of example-based explanations in a machine learning interface. International Conference on Intelligent User Interfaces, Proceedings IUI, Part F1476, 258–262.\nCarvalho, D. V., Pereira, E. M., & Cardoso, J. S. (2019). Machine learning interpretability: A survey on methods and metrics. In Electronics (Switzerland).\nCawthon, N., & Moere, A. Vande. (2007). The effect of aesthetic on the usability of data visualization. Proceedings of the International Conference on Information Visualisation.\nChati, Y. S., & Balakrishnan, H. (2017). A Gaussian Process Regression approach to model aircraft engine fuel flow rate. Proceedings - 2017 ACM/IEEE 8th International Conference on Cyber-Physical Systems, ICCPS 2017 (Part of CPS Week).\nChaudhuri, S., Dayal, U., & Narasayya, V. (2011). An overview of business intelligence technology. In Communications of the ACM (Vol. 54, Issue 8). \nChen, H., Chiang, R. H. L., & Storey, V. C. (2012). Business intelligence and analytics: From big data to big impact. MIS Quarterly: Management Information Systems,\n36(4).\nClancey, W. J. (1983). The epistemology of a rule-based expert system -a framework for explanation. Artificial Intelligence.\nCollins, C. R., & Stephenson, K. (2003). A circle packing algorithm. Computational Geometry: Theory and Applications, 25(3).\nDavis, B., Glenski, M., Sealy, W., & Arendt, D. (2020). Measure Utility, Gain Trust: Practical Advice for XAI Researchers. Proceedings - 2020 IEEE Workshop on\nTRust and EXpertise in Visual Analytics, TREX 2020, 1–8.\nDawes, R. M. (1979). The robust beauty of improper linear models in decision making. American Psychologist, 34(7).\nDeng, X., & Chi, L. (2012). Understanding postadoptive behaviors in information systems use: A longitudinal analysis of system use problems in the business\nintelligence context. Journal of Management Information Systems, 29(3).\nDesai, M., Kaniarasu, P., Medvedev, M., Steinfeld, A., & Yanco, H. (2013). Impact of robot failures and feedback on real-time trust. ACM/IEEE International\nConference on Human-Robot Interaction.\nDietvorst, B. J., Simmons, J. P., & Massey, C. (2015). Algorithm aversion: People erroneously avoid algorithms after seeing them err. Journal of Experimental\nPsychology: General, 144(1).\nDoshi-Velez, F., & Kim, B. (2017). Towards A Rigorous Science of Interpretable Machine Learning. Ml, 1–13. http://arxiv.org/abs/1702.08608\nDresner, H. (2001). Business Intelligence in 2002: A Coming of Age - 103282.pdf. Gartner.\nFreedy, A., DeVisser, E., Weltman, G., & Coeyman, N. (2007). Measurement of trust in human-robot collaboration. Proceedings of the 2007 International Symposium on Collaborative Technologies and Systems, CTS.\nFriedman, J. H. (2001). Greedy function approximation: A gradient boosting machine. Annals of Statistics.\nFriedman, J. H., & Meulman, J. J. (2003). Multiple additive regression trees with application in epidemiology. Statistics in Medicine, 22(9).\nGillespie, PhD, MA, RN, T. W. (2012). Understanding Waterfall Plots. Journal of the Advanced Practitioner in Oncology, 3(2).\nGlass, A., McGuinness, D. L., & Wolverton, M. (2008). Toward establishing trust in adaptive agents. International Conference on Intelligent User Interfaces, Proceedings IUI.\nGoodman, B., & Flaxman, S. (2017). European union regulations on algorithmic decision making and a “right to explanation.” AI Magazine.\nGorchels, L. (2000). The Product Manager’s Handbook. In NTC Business Books. Groom, V., & Nass, C. (2007). Can robots be teammates? Benchmarks in human-robot\nteams. Interaction Studies.\nHenelius, A., Puolamäki, K., Boström, H., Asker, L., & Papapetrou, P. (2014). A peek into the black box: Exploring classifiers by randomization. Data Mining and\nKnowledge Discovery.\nHoffrage, U., & Gigerenzer, G. (1998). Using natural frequencies to improve diagnostic inferences. Academic Medicine.\nHong, S., & Zhang, A. (2010). An efficiency study of airlines and air cargo/passenger divisions: A DEA approach. World Review of Intermodal Transportation\nResearch.\nInselberg, A., & Dimsdale, B. (1990). Parallel coordinates: A tool for visualizing multi-dimensional geometry.\nItti, L., Koch, C., & Niebur, E. (1998). A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine\nIntelligence, 20(11).\nJiang, C., & Zheng, S. (2020). Airline baggage fees and airport congestion. Transportation Research Part C: Emerging Technologies.\nKay, M., Patel, S. N., & Kientz, J. A. (2015). How good is 85%? A survey tool to connect classifier evaluation to acceptability of accuracy. Conference on Human\nFactors in Computing Systems - Proceedings, 2015-April.\nKim, Y. S., Walls, L. A., Krafft, P., & Hullman, J. (2019). A Bayesian cognition approach to improve data visualization. Conference on Human Factors in Computing Systems - Proceedings, 1–14.\nKoh, P. W., & Liang, P. (2017). Understanding black-box predictions via influence functions. 34th International Conference on Machine Learning, ICML 2017, 4, 2976–2987.\nKosara, R. (2016). Presentation-Oriented Visualization Techniques. IEEE Computer Graphics and Applications, 36(1).\nKrause, J., Perer, A., & Ng, K. (2016). Interacting with predictions: Visual inspection of black-box machine learning models. Conference on Human Factors in\nComputing Systems - Proceedings, 5686–5697.\nLangley, P., & Simon, H. A. (1995). Applications of Machine Learning and Rule Induction. Communications of the ACM.\nLapuschkin, S., Binder, A., Montavon, G., Muller, K. R., & Samek, W. (2016). Analyzing Classifiers: Fisher Vectors and Deep Neural Networks. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition.\nLe Bras, P., Robb, D. A., Methven, T. S., Padilla, S., & Chantler, M. J. (2018). Improving user confidence in concept maps: Exploring data driven explanations. Conference on Human Factors in Computing Systems - Proceedings, 2018-April.\nLeBaron, B. (2001). Evolution and time horizons in an agent-based stock market. Macroeconomic Dynamics.\nLiao, Q. V., Gruen, D., & Miller, S. (2020). Questioning the AI: Informing Design Practices for Explainable AI User Experiences. Conference on Human Factors in Computing Systems - Proceedings.\nLipton, Z. C. (2018). The mythos of model interpretability. Communications of the ACM, 61(10).\nLou, Y., Caruana, R., Gehrke, J., & Hooker, G. (2013). Accurate intelligible models with pairwise interactions. Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.\nLundberg, S. M., & Lee, S. I. (2017). A unified approach to interpreting model predictions. Advances in Neural Information Processing Systems.\nMadsen, M., & Gregor, S. (2000). Measuring Human-Computer Trust. Proceedings of Eleventh Australasian Conference on Information Systems, 6–8.\nManikandan, S. (2011). Measures of central tendency: Median and mode. In Journal of Pharmacology and Pharmacotherapeutics (Vol. 2, Issue 3).\nMayer, R. C., Davis, J. H., & Schoorman, F. D. (1995). An Integrative Model of Organizational Trust. Academy of Management Review, 20(3), 709–734.\nMcAllister, D. J. (1995). Affect- and Cognition-Based Trust as Foundations for Interpersonal Cooperation in Organizations. Academy of Management Journal, 38(1).\nMcGovern, A., Lagerquist, R., Gagne, D. J., Jergensen, G. E., Elmore, K. L., Homeyer, C. R., & Smith, T. (2019). Making the black box more transparent: Understanding the physical implications of machine learning. Bulletin of the American Meteorological Society.\nMiller, T. (2019). Explanation in artificial intelligence: Insights from the social sciences.\nIn Artificial Intelligence.\nMolnar, C. (2019). Interpretable Machine Learning. A Guide for Making Black Box Models Explainable. Book, 247. https://christophm.github.io/interpretable-ml-book\nNguyen, A., Yosinski, J., & Clune, J. (2015). Deep neural networks are easily fooled: High confidence predictions for unrecognizable images. Proceedings of the IEEE\nComputer Society Conference on Computer Vision and Pattern Recognition. Nicolae, M., Arikan, M., Deshpande, V., & Ferguson, M. (2017). Do bags fly free? An empirical analysis of the operational implications of airline baggage fees. In Management Science.\nNunes, I., & Jannach, D. (2017). A systematic review and taxonomy of explanations in decision support and recommender systems. User Modeling and User-Adapted\nInteraction, 27(3–5), 393–444.\nPandey, A. V., Manivannan, A., Nov, O., Satterthwaite, M., & Bertini, E. (2014). The persuasive power of data visualization. IEEE Transactions on Visualization and\nComputer Graphics, 20(12), 2211–2220.\nPanniello, U., Gorgoglione, M., & Tuzhilin, A. (2016). In CARSs we trust: How context-aware recommendations affect customers’ trust and other business performance measures of recommender systems. Information Systems Research, 27(1).\nParasuraman, R., & Riley, V. (1997). Humans and automation: Use, misuse, disuse, abuse. Human Factors.\nPerrotta, F., Parry, T., & Neves, L. C. (2017). Application of machine learning for fuel consumption modelling of trucks. Proceedings - 2017 IEEE International Conference on Big Data, Big Data 2017.\nPieters, W. (2011). Explanation and trust: What to tell the user in security and AI? Ethics and Information Technology, 13(1).\nPizer, S. M., Amburn, E. P., Austin, J. D., Cromartie, R., Geselowitz, A., Greer, T., ter Haar Romeny, B., Zimmerman, J. B., & Zuiderveld, K. (1987). Adaptive histogram equalization and its variations. Computer Vision, Graphics, and Image Processing, 38(1), 99.\nPoursabzi-Sangdeh, F., Goldstein, D. G., & Hofman, J. M. (2021). Manipulating and measuring model interpretability. In Conference on Human Factors in Computing Systems - Proceedings.\nPower, D. J. (2002). Decision Support Systems: Concepts and Resources for Managers. In Information Systems Management (Vol. 20, Issue 4).\nPutnam, V., & Conati, C. (2019). Exploring the need for explainable artificial intelligence (XAI) in intelligent tutoring systems (ITS). CEUR Workshop Proceedings.\nQuinlan, J. R. (1987). Simplifying decision trees. International Journal of Man- Machine Studies.\nRibeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should i trust you?” Explaining the predictions of any classifier. Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 1135–1144.\nRobertson, G., Fernandez, R., Fisher, D., Lee, B., & Stasko, J. (2008). Effectiveness of animation in trend visualization. IEEE Transactions on Visualization and Computer Graphics, 14(6), 1325–1332.\nRose, J. M., Hensher, D. A., Greene, W. H., & Washington, S. P. (2012). Attribute exclusion strategies in airline choice: Accounting for exogenous information on decision maker processing strategies in models of discrete choice.\nTransportmetrica.\nRudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5), 206–215.\nSamek, W., Wiegand, T., & Müller, K.-R. (2017). Explainable Artificial Intelligence: Understanding, Visualizing and Interpreting Deep Learning Models.\nhttp://arxiv.org/abs/1708.08296\nSautto, J. M. (2014). Decision Support Systems for Business Intelligence, 2nd edition. In Investigación Operacional (Vol. 35, Issue 1).\nSaxena, R., & Srinivasan, A. (2013). Business intelligence. In International Series in\nOperations Research and Management Science.\nShafiei, F., & Sundaram, D. (2004). Multi-enterprise collaborative enterprise resource planning and decision support systems. Proceedings of the Hawaii International\nConference on System Sciences, 37.\nSimonyan, K., Vedaldi, A., & Zisserman, A. (2014). Deep inside convolutional networks: Visualising image classification models and saliency maps. 2nd International Conference on Learning Representations, ICLR 2014 - Workshop Track Proceedings.\nSwartout, W. R. (1983). XPLAIN: a system for creating and explaining expert consulting programs. Artificial Intelligence.\nTouchette, P. E., MacDonald, R. F., & Langer, S. N. (1985). A scatter plot for identifying stimulus control of problem behavior. Journal of Applied Behavior Analysis, 18(4).\nTrani, A. A., Wing-Ho, F. C., Schilling, G., Baik, H., & Seshadri, A. (2004). A neural network model to estimate aircraft fuel consumption. Collection of Technical Papers - AIAA 4th Aviation Technology, Integration, and Operations Forum, ATIO.\nvan Wijk, J. J., & van de Wetering, H. (1999). Cushion treemaps: visualization of hierarchical information. Proceedings of the IEEE Symposium on Information\nVisualization.\nVassiliades, A., Bassiliades, N., & Patkos, T. (2021). Argumentation and explainable artificial intelligence: A survey. In Knowledge Engineering Review.\nVeale, M., Van Kleek, M., & Binns, R. (2018). Fairness and accountability design needs for algorithmic support in high-stakes public sector decision-making. Conference on Human Factors in Computing Systems - Proceedings, 2018-April. Vilone, G., & Longo, L. (2020). Explainable Artificial Intelligence: a Systematic Review. Dl. http://arxiv.org/abs/2006.00093\nWang, D., Yang, Q., Abdul, A., Lim, B. Y., & States, U. (2019). Designing Theory-Driven User-Centric Explainable AI. 1–15.\nWang, J., Gou, L., Yang, H., & Shen, H. W. (2018). GANViz: A Visual Analytics Approach to Understand the Adversarial Game. IEEE Transactions on Visualization and Computer Graphics, 24(6).\nWang, N., Pynadath, D. V., & Hill, S. G. (2016). Trust calibration within a human-robot team: Comparing automatically generated explanations. ACM/IEEE\nInternational Conference on Human-Robot Interaction.\nWong, W. H., Zhang, A., Van Hui, Y., & Leung, L. C. (2009). Optimal baggage-limit policy: Airline passenger and cargo allocation. Transportation Science, 43(3),\n355–369.\nXia, M., Asano, Y., Williams, J. J., Qu, H., & Ma, X. (2020). Using Information Visualization to Promote Students’ Reflection on “gaming the System” in Online\nLearning. L@S 2020 - Proceedings of the 7th ACM Conference on Learning @ Scale, 37–49.\nYagoda, R. E., & Gillan, D. J. (2012). You Want Me to Trust a ROBOT? The Development of a Human-Robot Interaction Trust Scale. International Journal of\nSocial Robotics, 4(3).\nYang, F., Huang, Z., Scholtz, J., & Arendt, D. L. (2020). How do visual explanations foster end users’ appropriate trust in machine learning? International Conference\non Intelligent User Interfaces, Proceedings IUI.\nYu, K., Taib, R., Berkovsky, S., Zhou, J., Conway, D., & Chen, F. (2016). Trust and Reliance based on system accuracy. UMAP 2016 - Proceedings of the 2016\nConference on User Modeling Adaptation and Personalization.\nYur, E., & Vasil, V. (2013). Analytical Review of Data Visualization Methods in Application to Big Data. Journal of Electrical and Computer Engineering, 2013, Article ID 969458.\nZhang, Y., Vera Liao, Q., & Bellamy, R. K. E. (2020). Effect of confidence and explanation on accuracy and trust calibration in AI-assisted decision making.\nFAT* 2020 - Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency.
描述: 碩士
國立政治大學
資訊管理學系
108356028
資料來源: http://thesis.lib.nccu.edu.tw/record/#G0108356028
資料類型: thesis
Appears in Collections:學位論文

Files in This Item:
File SizeFormat
602801.pdf2.73 MBAdobe PDF2View/Open
Show full item record

Google ScholarTM

Check

Altmetric

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.