English  |  正體中文  |  简体中文  |  Post-Print筆數 : 27 |  Items with full text/Total items : 113311/144292 (79%)
Visitors : 50912700      Online Users : 408
RC Version 6.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
Scope Tips:
  • please add "double quotation mark" for query phrases to get precise results
  • please goto advance search for comprehansive author search
  • Adv. Search
    HomeLoginUploadHelpAboutAdminister Goto mobile version
    政大機構典藏 > 資訊學院 > 資訊科學系 > 學位論文 >  Item 140.119/136970
    Please use this identifier to cite or link to this item: https://nccur.lib.nccu.edu.tw/handle/140.119/136970


    Title: 標註資料擴展應用於零樣本意向偵測的生成式模型
    Generative Model with Label-Augmentation for Zero-Shot Intent Detection
    Authors: 李鈺祥
    Li, Yu-Siang
    Contributors: 黃瀚萱
    Huang, Hen-Hsen
    李鈺祥
    Li, Yu-Siang
    Keywords: 意向偵測
    零樣本學習
    少樣本學習
    資料擴展
    生成式模型
    Intent Detection
    Zero-Shot Learning
    Few-Shot Learning
    Data Augmentation
    Generative Model
    Date: 2021
    Issue Date: 2021-09-02 16:57:08 (UTC+8)
    Abstract: 意向偵測是任務導向對話系統的子任務之一,針對使用者的輸入字句辨識其真正的意向為何,如此才能幫助後續的應用。現今多數已上線商用的意向偵測,皆是在特定的產業、領域。然而,更多的實際應用會面臨不同領域、資料的稀缺等問題。
    本篇論文針對跨領域、零樣本的情境進行探討,提出了具有新穎向生成式推理模型,並且進一步結合標註資料擴展,提升效能。在訓練階段,標註資料擴展包含了來回翻譯擴展、同義詞擴展,以及兩者的結合,使主任務模型進行意向偵測之結果可具多樣性。在跨領域零樣本推理的階段,集成模型匯聚基於BERT預訓練模型的句子向量餘弦相似度、基於T5預訓練模型的STSB任務字句相似度、基於詞彙的句子相似度等資訊,使來源領域的已預測意向類別,能被精準推理至目標領域的意向類別。
    本篇論文的實驗素材包含SNIPS、ATIS、MultiWOZ、SGD等常見於意向偵測的資料集。實驗設定則是利用每一個資料集當作來源領域,其他皆為目標領域,同時結合不同的推理與擴展方法組合進行實驗。實驗結果顯示,本研究所提出之方法能有效提升零樣本跨領域學習的效能。對於本研究的挑戰性因素,亦作了詳細的分析與討論。
    Intent detection is one of sub-tasks in task-oriented dialogue system, aiming at understanding how users think and want from their input queries, so that it can help them for downstream applications. Nowadays, more and more task-oriented dialogue systems using intent detection in in commercial and industrial fields. However, many practical problems remain to deal with.
    This work focuses on zero-shot cross-domain intent detection, where training instances in the target domain are unavailable, and the intent set of the target domain may greatly differ from that of the source domain. For this challenging scenario, we propose a novel approach that performs the inference as a generation task with label augmentation for performance enhancement.
    In the training stage, the main-task model learns from augmented label data, enabling the model to do prediction for diverse intent classes. We explore back-translation augmentation, synonym-set augmentation, and their combination in label augmentation. In the inference stage, our generative inference methodology facilitates zero-shot predictions for the new target domain. Our model is an ensemble of the cosine similarity based on BERT, the sentence similarity based on the built-in task, and the STSB measurement based on T5. The weighted combination is intensively explored.
    Our approach is evaluated on four famous datasets for intent detection, including SNIPS, ATIS, MultiWOZ, and SGD. We explore all combinations of the source and the target domains from these four datasets. Experimental results show the effectiveness of our approach. The challenging issues for zero-shot cross-domain intent detection are further analyzed and discussed.
    Reference: 1. Natural Language Processing, Retrieved Nov. 19 2020, from: https://zh.wikipedia.org/wiki/%E8%87%AA%E7%84%B6%E8%AF%AD%E8%A8%80%E5%A4%84%E7%90%86
    2. Ter Meulen, A. (2017). Logic and natural language. The Blackwell guide to philosophical logic, 461-483.
    3. Casanueva, I., Temčinas, T., Gerz, D., Henderson, M., & Vulić, I. (2020). Efficient intent detection with dual sentence encoders. arXiv preprint arXiv:2003.04807.
    4. Semantic Parsing, Retrieved May. 2 2021, from https://en.wikipedia.org/wiki/Semantic_parsing
    5. Jia, R., & Liang, P. (2016). Data recombination for neural semantic parsing. arXiv preprint arXiv:1606.03622.
    6. Andreas, J., Vlachos, A., & Clark, S. (2013, August). Semantic parsing as machine translation. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) (pp. 47-52).
    7. Berant, J., Chou, A., Frostig, R., & Liang, P. (2013, October). Semantic parsing on freebase from question-answer pairs. In Proceedings of the 2013 conference on empirical methods in natural language processing (pp. 1533-1544).
    8. Poon, H., & Domingos, P. (2010, July). Unsupervised ontology induction from text. In Proceedings of the 48th annual meeting of the Association for Computational Linguistics (pp. 296-305).
    9. Kaliszyk, C., Urban, J., & Vyskočil, J. (2017, September). Automating formalization by statistical and semantic parsing of mathematics. In International Conference on Interactive Theorem Proving (pp. 12-27). Springer, Cham.
    10. Rabinovich, M., Stern, M., & Klein, D. (2017). Abstract syntax networks for code generation and semantic parsing. arXiv preprint arXiv:1704.07535.
    11. Yin, P., & Neubig, G. (2017). A syntactic neural model for general-purpose code generation. arXiv preprint arXiv:1704.01696.
    12. Kumar, A., Gupta, A., Chan, J., Tucker, S., Hoffmeister, B., Dreyer, M., ... & Kumar, A. (2017). Just ASK: building an architecture for extensible self-service spoken language understanding. arXiv preprint arXiv:1711.00549.
    13. Bapna, A., Tur, G., Hakkani-Tur, D., & Heck, L. (2017). Towards zero-shot frame semantic parsing for domain scaling. arXiv preprint arXiv:1707.02363.
    14. Liu, B., & Lane, I. (2016). Attention-based recurrent neural network models for joint intent detection and slot filling. arXiv preprint arXiv:1609.01454.
    15. Liang, P., & Potts, C. (2015). Bringing machine learning and compositional semantics together. Annu. Rev. Linguist., 1(1), 355-376.
    16. Redko, I., Morvant, E., Habrard, A., Sebban, M., & Bennani, Y. (2019). Advances in domain adaptation theory. Elsevier.
    17. Bridle, J., & Cox, S. J. (1990). Recnorm: Simultaneous normalisation and classification applied to speech recognition. Advances in Neural Information Processing Systems, 3, 234-240.
    18. Ben-David, S., Blitzer, J., Crammer, K., Kulesza, A., Pereira, F., & Vaughan, J. W. (2010). A theory of learning from different domains. Machine learning, 79(1), 151-175.
    19. Sun, B., Feng, J., & Saenko, K. (2016, March). Return of frustratingly easy domain adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 30, No. 1).
    20. Amodei, D., Olah, C., Steinhardt, J., Christiano, P., Schulman, J., & Mané, D. (2016). Concrete problems in AI safety. arXiv preprint arXiv:1606.06565.
    21. Wang, Y., Yao, Q., Kwok, J. T., & Ni, L. M. (2020). Generalizing from a few examples: A survey on few-shot learning. ACM Computing Surveys (CSUR), 53(3), 1-34.
    22. Zero-Shot Learning, Retrieved Nov. 22 2020, from https://en.wikipedia.org/wiki/Zero-shot_learning
    23. Xian, Y., Schiele, B., & Akata, Z. (2017). Zero-shot learning-the good, the bad and the ugly. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 4582-4591).
    24. Lampert, C. H., Nickisch, H., & Harmeling, S. (2009, June). Learning to detect unseen object classes by between-class attribute transfer. In 2009 IEEE Conference on Computer Vision and Pattern Recognition (pp. 951-958). IEEE.
    25. Romera-Paredes, B., & Torr, P. (2015, June). An embarrassingly simple approach to zero-shot learning. In International conference on machine learning (pp. 2152-2161). PMLR.
    26. Atzmon, Y., & Chechik, G. (2018). Probabilistic and-or attribute grouping for zero-shot learning. arXiv preprint arXiv:1806.02664.
    27. Hu, R. L., Xiong, C., & Socher, R. Zero-Shot Image Classification Guided by Natural Language Descriptions of Classes: A Meta-Learning Approach.
    28. Srivastava, S., Labutov, I., & Mitchell, T. (2018, July). Zero-shot learning of classifiers from natural language quantification. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (pp. 306-316).
    29. Frome, A., Corrado, G., Shlens, J., Bengio, S., Dean, J., Ranzato, M. A., & Mikolov, T. (2013). Devise: A deep visual-semantic embedding model.
    30. Data Augmentation, Retrieved May. 2 2021, from https://en.wikipedia.org/wiki/Data_augmentation
    31. Shorten, C., & Khoshgoftaar, T. M. (2019). A survey on image data augmentation for deep learning. Journal of Big Data, 6(1), 1-48.
    32. Bird, J. J., Faria, D. R., Premebida, C., Ekárt, A., & Ayrosa, P. P. (2020, April). Overcoming data scarcity in speaker identification: Dataset augmentation with synthetic mfccs via character-level rnn. In 2020 IEEE International Conference on Autonomous Robot Systems and Competitions (ICARSC) (pp. 146-151). IEEE.
    33. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... & Polosukhin, I. (2017). Attention is all you need. arXiv preprint arXiv:1706.03762.
    34. Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. (2018). Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
    35. Open Sourcing BERT: State-of-the-Art Pre-training for Natural Language Processing. Retrieved Nov. 22 2020 from https://ai.googleblog.com/2018/11/open-sourcing-bert-state-of-art-pre.html
    36. Better Language Models and Their Implications. Retrieved Nov. 22 2020 from https://openai.com/blog/better-language-models/
    37. Mikolov, T., Sutskever, I., Chen, K., Corrado, G., & Dean, J. (2013). Distributed representations of words and phrases and their compositionality. arXiv preprint arXiv:1310.4546.
    38. Pennington, J., Socher, R., & Manning, C. D. (2014, October). Glove: Global vectors for word representation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) (pp. 1532-1543).
    39. Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., ... & Liu, P. J. (2019). Exploring the limits of transfer learning with a unified text-to-text transformer. arXiv preprint arXiv:1910.10683.
    40. Daumé III, H. (2009). Frustratingly easy domain adaptation. arXiv preprint arXiv:0907.1815.
    41. Kim, Y. B., Stratos, K., Sarikaya, R., & Jeong, M. (2015, July). New transfer learning techniques for disparate label sets. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers) (pp. 473-482).
    42. Obeidat, R., Fern, X. Z., & Tadepalli, P. (2016, August). Label Embedding Approach for Transfer Learning. In ICBO/BioCreative.
    43. Wang, Z., Qu, Y., Chen, L., Shen, J., Zhang, W., Zhang, S., ... & Yu, Y. (2018). Label-aware double transfer learning for cross-specialty medical named entity recognition. arXiv preprint arXiv:1804.09021.
    44. Lee, J. Y., Dernoncourt, F., & Szolovits, P. (2017). Transfer learning for named-entity recognition with neural networks. arXiv preprint arXiv:1705.06273.
    45. Sachan, D. S., Xie, P., Sachan, M., & Xing, E. P. (2018, November). Effective use of bidirectional language modeling for transfer learning in biomedical named entity recognition. In Machine learning for healthcare conference (pp. 383-402). PMLR.
    46. Yang, Z., Salakhutdinov, R., & Cohen, W. W. (2017). Transfer learning for sequence tagging with hierarchical recurrent networks. arXiv preprint arXiv:1703.06345.
    47. Lin, B. Y., & Lu, W. (2018). Neural adaptation layers for cross-domain named entity recognition. arXiv preprint arXiv:1810.06368.
    48. Jia, C., Liang, X., & Zhang, Y. (2019, July). Cross-domain NER using cross-domain language modeling. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (pp. 2464-2474).
    49. Liu, Z., Winata, G. I., & Fung, P. (2020). Zero-resource cross-domain named entity recognition. arXiv preprint arXiv:2002.05923.
    50. Shah, D. J., Gupta, R., Fayazi, A. A., & Hakkani-Tur, D. (2019). Robust zero-shot cross-domain slot filling with example values. arXiv preprint arXiv:1906.06870.
    51. Liu, Z., Winata, G. I., Xu, P., & Fung, P. (2020). Coach: A coarse-to-fine approach for cross-domain slot filling. arXiv preprint arXiv:2004.11727.
    52. Siddique, A. B., Jamour, F., & Hristidis, V. (2021). Linguistically-enriched and context-aware zero-shot slot filling. arXiv preprint arXiv:2101.06514.
    53. Yazdani, M., & Henderson, J. (2015, September). A model of zero-shot learning of spoken language understanding. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (pp. 244-249).
    54. Ferreira, E., Jabaian, B., & Lefevre, F. (2015, April). Online adaptative zero-shot learning spoken language understanding using word-embedding. In 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 5321-5325). IEEE.
    55. Dauphin, Y. N., Tur, G., Hakkani-Tur, D., & Heck, L. (2013). Zero-shot learning for semantic utterance classification. arXiv preprint arXiv:1401.0509.
    56. Kumar, A., Muddireddy, P. R., Dreyer, M., & Hoffmeister, B. (2017, August). Zero-Shot Learning Across Heterogeneous Overlapping Domains. In INTERSPEECH (pp. 2914-2918).
    57. Williams, K. (2019). Zero Shot Intent Classification Using Long-Short Term Memory Networks. In INTERSPEECH (pp. 844-848).
    58. Gangal, V., Arora, A., Einolghozati, A., & Gupta, S. (2020, April). Likelihood Ratios and Generative Classifiers for Unsupervised Out-of-Domain Detection in Task Oriented Dialog. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 34, No. 05, pp. 7764-7771).
    59. Breunig, M. M., Kriegel, H. P., Ng, R. T., & Sander, J. (2000, May). LOF: identifying density-based local outliers. In Proceedings of the 2000 ACM SIGMOD international conference on Management of data (pp. 93-104).
    60. Edunov, S., Ott, M., Auli, M., & Grangier, D. (2018). Understanding back-translation at scale. arXiv preprint arXiv:1808.09381.
    61. WordNet, Retrieved May. 5 2021, from https://wordnet.princeton.edu/
    62. Wei, J., & Zou, K. (2019). Eda: Easy data augmentation techniques for boosting performance on text classification tasks. arXiv preprint arXiv:1901.11196.
    63. Cosine Similarity, Retrieved May. 5 2021, from https://en.wikipedia.org/wiki/Cosine_similarity
    64. Gestalt Pattern Matching, Retrieved June. 5 2021, from https://en.wikipedia.org/wiki/Gestalt_Pattern_Matching
    65. Coucke, A., Saade, A., Ball, A., Bluche, T., Caulier, A., Leroy, D., ... & Dureau, J. (2018). Snips voice platform: an embedded spoken language understanding system for private-by-design voice interfaces. arXiv preprint arXiv:1805.10190.
    66. Price, P. (1990). Evaluation of spoken language systems: The ATIS domain. In Speech and Natural Language: Proceedings of a Workshop Held at Hidden Valley, Pennsylvania, June 24-27, 1990.
    67. Zang, X., Rastogi, A., Sunkara, S., Gupta, R., Zhang, J., & Chen, J. (2020). Multiwoz 2.2: A dialogue dataset with additional annotation corrections and state tracking baselines. arXiv preprint arXiv:2007.12720.
    68. Rastogi, A., Zang, X., Sunkara, S., Gupta, R., & Khaitan, P. (2020, April). Towards scalable multi-domain conversational agents: The schema-guided dialogue dataset. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 34, No. 05, pp. 8689-8696).
    69. Scikit-Learn Metrics F1 Score, Retrieved May. 5 2021, from: https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html
    70. McNemar`s test, Retrieved May. 5 2021, from: https://en.wikipedia.org/wiki/McNemar`s_test
    71. Hou, Y., Mao, J., Lai, Y., Chen, C., Che, W., Chen, Z., & Liu, T. (2020). Fewjoint: A few-shot learning benchmark for joint language understanding. arXiv preprint arXiv:2009.08138.
    72. Zhu, S., Zhao, Z., Zhao, T., Zong, C., & Yu, K. (2019, October). Catslu: The 1st chinese audio-textual spoken language understanding challenge. In 2019 International Conference on Multimodal Interaction (pp. 521-525).
    73. Chu-Ren Huang and Shu-Kai Hsieh. (2010). Infrastructure for Cross-lingual Knowledge Representation ─ Towards Multilingualism in Linguistic Studies. Taiwan NSC-granted Research Project (NSC 96-2411-H-003-061-MY3)
    Description: 碩士
    國立政治大學
    資訊科學系
    108753206
    Source URI: http://thesis.lib.nccu.edu.tw/record/#G0108753206
    Data Type: thesis
    DOI: 10.6814/NCCU202101275
    Appears in Collections:[資訊科學系] 學位論文

    Files in This Item:

    File Description SizeFormat
    320601.pdf8154KbAdobe PDF2151View/Open


    All items in 政大典藏 are protected by copyright, with all rights reserved.


    社群 sharing

    著作權政策宣告 Copyright Announcement
    1.本網站之數位內容為國立政治大學所收錄之機構典藏,無償提供學術研究與公眾教育等公益性使用,惟仍請適度,合理使用本網站之內容,以尊重著作權人之權益。商業上之利用,則請先取得著作權人之授權。
    The digital content of this website is part of National Chengchi University Institutional Repository. It provides free access to academic research and public education for non-commercial use. Please utilize it in a proper and reasonable manner and respect the rights of copyright owners. For commercial use, please obtain authorization from the copyright owner in advance.

    2.本網站之製作,已盡力防止侵害著作權人之權益,如仍發現本網站之數位內容有侵害著作權人權益情事者,請權利人通知本網站維護人員(nccur@nccu.edu.tw),維護人員將立即採取移除該數位著作等補救措施。
    NCCU Institutional Repository is made to protect the interests of copyright owners. If you believe that any material on the website infringes copyright, please contact our staff(nccur@nccu.edu.tw). We will remove the work from the repository and investigate your claim.
    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - Feedback