English  |  正體中文  |  简体中文  |  Post-Print筆數 : 27 |  Items with full text/Total items : 111206/142126 (78%)
Visitors : 48112574      Online Users : 68
RC Version 6.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
Scope Tips:
  • please add "double quotation mark" for query phrases to get precise results
  • please goto advance search for comprehansive author search
  • Adv. Search
    HomeLoginUploadHelpAboutAdminister Goto mobile version
    政大機構典藏 > 商學院 > 資訊管理學系 > 學位論文 >  Item 140.119/135329
    Please use this identifier to cite or link to this item: https://nccur.lib.nccu.edu.tw/handle/140.119/135329


    Title: 原像分析結合視覺化分析作為探究XAI之可行性研究
    The Study of Preimage Analysis With Visual Analytics For XAI
    Authors: 古聖釗
    Gu, Sheng-Zhao
    Contributors: 蔡瑞煌
    郁方

    Tsaih, Rua-Huan
    Yu, Fang

    古聖釗
    Gu, Sheng-Zhao
    Keywords: 可解釋人工智慧
    單隱層前饋神經網路
    整流線性單元
    原像分析
    視覺化分析
    Explainable Artificial Intelligence
    1-hidden layer feed-forward neural network
    ReLU
    Preimage analysis
    Visual analytics
    Date: 2021
    Issue Date: 2021-06-01 14:55:11 (UTC+8)
    Abstract: 本研究計畫探索可解釋人工智慧(Explainable Artificial Intelligence, XAI)議題。為了解決類神經網路(Artificial Neural Networks, ANN)的黑箱挑戰,本研究計畫將探討原像分析(Preimage analysis)的數學分析工具和視覺化分析(Visual analytics)是否可以用來打開黑箱。本研究將著眼於具有m個輸入節點,p個使用整流線性單元(Rectified Linear Unit, ReLU)激發函數的隱藏節點和一個使用線性激發函數的輸出節點的單隱層前饋神經網路(1-hidden Layer Feed-forward Neural network, 1HLNN)。近年來,ReLU在深度學習(Deep Learning)應用中被廣泛採用的原因是ReLU具有以下優點:(1)ReLU的計算成本低廉,因為它沒有復雜的數學運算,因此運算量較小,而訓練和執行的時間也較小;(2)線性是指輸入總合值變大時,該函數沒有“飽和”區域;(3)消失梯度問題可更容易地解決。還有,與深層神經網路(Deep Neural Networks, DNN)相比,1HLNN比較容易分析,比較容易打開其黑箱;而其推導得到的XAI結果可能可以擴展到DNN。因此,本研究計畫的重點之一乃在於探索具有ReLU激發函數的單隱層前饋神經網路的可解釋性。
    This research explores Explainable Artificial Intelligence (XAI) issues. In order to solve the black box of Artificial Neural Networks (ANN), this research project will explore whether the mathematical analysis tools of Preimage analysis and Visual analytics can be used. We will focus on 1- hidden Layer Feed-forward Neural Network(1HLNN) with m input nodes, p hidden nodes using Rectified Linear Unit (ReLU) excitation functions. In recent years, ReLU has been widely used in deep learning applications because ReLU has the following advantages: (1) ReLU has low computational cost; (2) Linearity means that when the total input value becomes larger, the function does not have a "saturated" area; (3) The vanishing gradient problem can be solved more easily. Also, compared with Deep Neural Networks (DNN), 1HLNN is easier to analyze and open its black box; and it may be extended to DNN. Therefore, one of the points of this research is to explore the interpretability of 1HLNN with ReLU.
    Reference: Arrieta, A. B., Díaz-Rodríguez, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., ... & Herrera, F. (2020). Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion, 58, 82-115.
    Belciug, S., & Gorunescu, F. (2018). Learning a single-hidden layer feedforward neural network using a rank correlation-based strategy with application to high dimensional gene expression and proteomic spectra datasets in cancer detection. Journal of biomedical informatics, 83, 159-166.
    Buchanan, B. G. (2005). A (very) brief history of artificial intelligence. Ai Magazine, 26(4), 53-53.
    Cao, J., & Lin, Z. (2015). Extreme learning machines on high dimensional and large data applications: a survey. Mathematical Problems in Engineering, 2015.
    Chen, H., Lundberg, S., & Lee, S. I. (2021). Explaining models by propagating Shapley values of local components. In Explainable AI in Healthcare and Medicine (pp. 261-270). Springer, Cham.
    Choo, J., & Liu, S. (2018). Visual analytics for explainable deep learning. IEEE computer graphics and applications, 38(4), 84-92.
    Deng, J., Li, K., & Irwin, G. W. (2011). Fast automatic two-stage nonlinear model identification based on the extreme learning machine. Neurocomputing, 74(16), 2422-2429.
    Glorot, X., Bordes, A., & Bengio, Y. (2011, June). Deep sparse rectifier neural networks. In Proceedings of the fourteenth international conference on artificial intelligence and statistics (pp. 315-323). JMLR Workshop and Conference Proceedings.
    Goodfellow, I., Bengio, Y., Courville, A., & Bengio, Y. (2016). Deep learning (Vol. 1, No. 2). Cambridge: MIT press.
    Gunning, D. (2017). Explainable artificial intelligence (xai). Defense Advanced Research Projects Agency (DARPA), nd Web, 2(2).
    He, K., Zhang, X., Ren, S., & Sun, J. (2015). Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE international conference on computer vision (pp. 1026-1034).
    High-Level Expert Group on AI (2019). Ethics guidelines for trustworthy AI (Report). European Commission.
    Huysmans, J., Dejaeger, K., Mues, C., Vanthienen, J., & Baesens, B. (2011). An empirical evaluation of the comprehensibility of decision table, tree and rule based predictive models. Decision Support Systems, 51(1), 141-154.
    Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems, 25, 1097-1105.
    Lifschitz, V. (2011). John McCarthy (1927–2011). Nature, 480(7375), 40-40.
    Lundberg, S., & Lee, S. I. (2017). A unified approach to interpreting model predictions. arXiv preprint arXiv:1705.07874.
    Maas, A. L., Hannun, A. Y., & Ng, A. Y. (2013, June). Rectifier nonlinearities improve neural network acoustic models. In Proc. icml (Vol. 30, No. 1, p. 3).
    Nielsen, M. A. (2015). Neural networks and deep learning (Vol. 25). San Francisco, CA: Determination press.
    Oke, S. A. (2008). A literature review on artificial intelligence. International journal of information and management sciences, 19(4), 535-570.
    Ribeiro, M. T., Singh, S., & Guestrin, C. (2016, August). " Why should i trust you?" Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining (pp. 1135-1144).
    Rudin, C. (2018). Please stop explaining black box models for high stakes decisions. arXiv preprint arXiv:1811.10154, 1.
    Shrikumar, A., Greenside, P., Shcherbina, A., & Kundaje, A. (2016). Not just a black box: Learning important features through propagating activation differences. arXiv preprint arXiv:1605.01713.
    Subudhi, B., & Jena, D. (2011). Nonlinear system identification using memetic differential evolution trained neural networks. Neurocomputing, 74(10), 1696-1709.
    Tsai, Y. H., Jheng, Y. J., & Tsaih, R. H. (2019, July). The Cramming, Softening and Integrating Learning Algorithm with Parametric ReLu Activation Function for Binary Input/Output Problems. In 2019 International Joint Conference on Neural Networks (IJCNN) (pp. 1-7). IEEE.
    Tsaih, R. H., Wan, Y. W., & Huang, S. Y. (2008, June). The rule-extraction through the preimage analysis. In 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence) (pp. 1488-1494). IEEE.
    Tsaih, R. R. (1993). The softening learning procedure. Mathematical and computer modelling, 18(8), 61-64.
    Tsaih, R. R. (1993, October). The softening learning procedure for the layered feedforward networks with multiple output nodes. In Proceedings of 1993 International Conference on Neural Networks (IJCNN-93-Nagoya, Japan) (Vol. 1, pp. 593-596). IEEE.
    Tsaih, R. H. R. (1997). Reasoning neural networks. In Mathematics of Neural Networks (pp. 366-371). Springer, Boston, MA.
    Tsaih, R. R. (1998). An explanation of reasoning neural networks. Mathematical and Computer Modelling, 28(2), 37-44.
    Copeland, B. J. (Ed.). (2004). The essential turing. Clarendon Press.
    Wang, F. Y., Zhang, J. J., Zheng, X., Wang, X., Yuan, Y., Dai, X., ... & Yang, L. (2016). Where does AlphaGo go: From church-turing thesis to AlphaGo thesis and beyond. IEEE/CAA Journal of Automatica Sinica, 3(2), 113-120.
    Watanabe, E., & Shimizu, H. (1993, October). Algorithm for pruning hidden units in multilayered neural network for binary pattern classification problem. In Proceedings of 1993 International Conference on Neural Networks (IJCNN-93-Nagoya, Japan) (Vol. 1, pp. 327-330). IEEE.
    Description: 碩士
    國立政治大學
    資訊管理學系
    107356043
    Source URI: http://thesis.lib.nccu.edu.tw/record/#G0107356043
    Data Type: thesis
    DOI: 10.6814/NCCU202100462
    Appears in Collections:[資訊管理學系] 學位論文

    Files in This Item:

    File Description SizeFormat
    604301.pdf11278KbAdobe PDF20View/Open


    All items in 政大典藏 are protected by copyright, with all rights reserved.


    社群 sharing

    著作權政策宣告 Copyright Announcement
    1.本網站之數位內容為國立政治大學所收錄之機構典藏,無償提供學術研究與公眾教育等公益性使用,惟仍請適度,合理使用本網站之內容,以尊重著作權人之權益。商業上之利用,則請先取得著作權人之授權。
    The digital content of this website is part of National Chengchi University Institutional Repository. It provides free access to academic research and public education for non-commercial use. Please utilize it in a proper and reasonable manner and respect the rights of copyright owners. For commercial use, please obtain authorization from the copyright owner in advance.

    2.本網站之製作,已盡力防止侵害著作權人之權益,如仍發現本網站之數位內容有侵害著作權人權益情事者,請權利人通知本網站維護人員(nccur@nccu.edu.tw),維護人員將立即採取移除該數位著作等補救措施。
    NCCU Institutional Repository is made to protect the interests of copyright owners. If you believe that any material on the website infringes copyright, please contact our staff(nccur@nccu.edu.tw). We will remove the work from the repository and investigate your claim.
    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - Feedback