English  |  正體中文  |  简体中文  |  Post-Print筆數 : 27 |  Items with full text/Total items : 112704/143671 (78%)
Visitors : 49721617      Online Users : 604
RC Version 6.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
Scope Tips:
  • please add "double quotation mark" for query phrases to get precise results
  • please goto advance search for comprehansive author search
  • Adv. Search
    HomeLoginUploadHelpAboutAdminister Goto mobile version
    Please use this identifier to cite or link to this item: https://nccur.lib.nccu.edu.tw/handle/140.119/152959


    Title: 深度學習之中文歌詞段落情緒辨識
    Deep Learning-Based Paragraph-Level Emotion Recognition of Chinese Song Lyrics
    Authors: 標云
    Biao, Yun
    Contributors: 張瑜芸
    Chang, Yu-Yun
    標云
    Biao, Yun
    Keywords: 深度學習
    情感辨識
    中文歌詞
    效價
    喚醒
    BERT
    敘事理論
    Deep Learning
    Emotion Recognition
    Chinese Song Lyrics
    Valence
    Arousal
    BERT
    Narrative Theory
    Date: 2024
    Issue Date: 2024-08-05 15:06:03 (UTC+8)
    Abstract: 本研究探討結合深度學習技術與敘事理論在中文歌曲歌詞段落情感識別中的應用。本研究之動機源於音樂在人類生活中的重要性、個性化音樂串流服務的興起以及日益增長的自動情感識別之需求。本研究以BERT模型實現,訓練BERT模型來預測中文歌曲歌詞中的效價(正面或負面情感傾向)、喚醒程度(情感激動強度)及其二者之交織狀態(情感象限)。敘事理論中的主題和結構分析的整合提供了對歌詞情感表達更深入的理解。實驗結果證明了該模型在情感分類中的效率和準確性,表明其在提升音樂推薦系統品質方面的潛在實用性。即所有用於預測情感的 BERT 模型,包括正面或負面情感傾向(Accuracy = 0.91,F-score = 0.90)、情感激動強度(Accuracy = 0.86,F-score = 0.86)以及情感象限的 BERT 模型(Accuracy = 0.77,F-score = 0.76)都優於正面或負面情感傾向(Accuracy = 0.68,F-score = 0.65)、情感激動強度(Accuracy = 0.65,F-score = 0.64)和情感象限(Accuracy = 0.48,F-score = 0.45)的基線模型。此外,通過敘事理論進行的錯誤分析確定了導致誤分類的關鍵因素,這些因素包括詞彙歧義、句法複雜性和敘事之流動性,這些都在準確解釋歌詞中發揮著重要作用。整體而言,本研究強調了將敘事分析與深度學習技術相結合的價值,以實現更為複雜和準確的中文歌曲歌詞情感辨識系統。
    This study explores the implementation of deep learning techniques alongside narrative theory for paragraph-level emotion recognition in Chinese song lyrics. It is motivated by the integral role of music in human life and the growing demand for automatic emotion recognition systems driven by personalized music streaming services. We leverage the BERT model to implement and evaluate machine learning models trained to predict valence (positive or negative emotions), arousal (intensity of emotion), and their intertwined states (emotional quadrants) from Chinese song lyrics. The integration of thematic and structural analysis derived from narrative theory provides a deeper understanding of lyrics' emotional expression. Experimental results demonstrate the model's efficiency and accuracy in classifying emotions, indicating its potential utility in improving the quality of music recommendation systems. All BERT models for predicting valence (Accuracy = 0.91, F-score = 0.90), arousal (Accuracy = 0.86, F-score = 0.86) and quadrants (Accuracy = 0.77, F-score = 0.76) outperformed baseline models of valence (Accuracy = 0.68, F-score = 0.65), arousal (Accuracy = 0.65, F-score = 0.64), and quadrants (Accuracy = 0.48, F-score = 0.45). Furthermore, our error analysis, informed by narrative theory, identifies key factors contributing to misclassification. These factors include lexical ambiguity, syntactic complexity, and narrative flow, all of which play significant roles in the accurate interpretation of lyrics. Overall, this research underscores the value of blending narrative analysis with deep learning techniques to achieve a more sophisticated and accurate system for emotion recognition in Chinese song lyrics.
    Reference: Abdillah, J., Asror, I., Wibowo, Y. F. A., et al. (2020). Emotion classification of song
    lyrics using bidirectional lstm method with glove word representation weighting.
    Jurnal RESTI (Rekayasa Sistem Dan Teknologi Informasi), 4(4), 723–729.
    Agrawal, Y., Shanker, R. G. R., & Alluri, V. (2021). Transformer-based approach towards music emotion recognition from lyrics. European Conference on Information Retrieval, 167–175.
    Ahonen, H., & Desideri, A. M. (2007). Group analytic music therapy. voices, 14, 686.
    Alorainy, W., Burnap, P., Liu, H., Javed, A., & Williams, M. L. (2018). Suspended
    accounts: A source of tweets with disgust and anger emotions for augmenting
    hate speech data sample. 2018 International Conference on Machine Learning
    and Cybernetics (ICMLC), 2, 581–586.
    An, Y., Sun, S., & Wang, S. (2017). Naive bayes classifiers for music emotion classification based on lyrics. 2017 IEEE/ACIS 16th International Conference on
    Computer and Information Science (ICIS), 635–638.
    Arumugam, D., et al. (2011). Emotion classification using facial expression. International Journal of Advanced Computer Science and Applications, 2(7).
    Baker, F., Wigram, T., Stott, D., & McFerran, K. (2008). Therapeutic songwriting in
    music therapy: Part i: Who are the therapists, who are the clients, and why is
    songwriting used? Nordic Journal of Music Therapy, 17(2), 105–123.
    Barradas, G. T., & Sakka, L. S. (2022). When words matter: A cross-cultural perspective
    on lyrics and their relationship to musical emotions. Psychology of Music, 50(2),
    650–669.
    Besson, M., Faita, F., Peretz, I., Bonnel, A.-M., & Requin, J. (1998). Singing in the
    brain: Independence of lyrics and tunes. Psychological Science, 9(6), 494–498.
    Chaudhary, D., Singh, N. P., & Singh, S. (2021). Development of music emotion classification system using convolution neural network. International Journal of
    Speech Technology, 24, 571–580.
    Chiril, P., Pamungkas, E. W., Benamara, F., Moriceau, V., & Patti, V. (2022). Emotionally informed hate speech detection: A multi-target perspective. Cognitive
    Computation, 1–31.
    Desmet, B., & Hoste, V. (2013). Emotion detection in suicide notes. Expert Systems
    with Applications, 40(16), 6351–6358.
    Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2018). Bert: Pre-training of deep
    bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
    Edmonds, D., & Sedoc, J. (2021). Multi-emotion classification for song lyrics. Proceed-
    ings of the Eleventh Workshop on Computational Approaches to Subjectivity,
    Sentiment and Social Media Analysis, 221–235.
    Ekman, P. (1992). Facial expressions of emotion: New findings, new questions.
    Fludernik, M. (2009). An introduction to narratology. Routledge.
    Frijda, N. H. (1986). The emotions. Cambridge University Press.
    Genette, G. (1988). Narrative discourse revisited. Cornell University Press.
    Guillemette, L., & Lévesque, C. (2016). Narratology [In Louis Hébert (Ed.), Signo [online], Rimouski (Quebec)]. http://www.signosemio.com/genette/narratology.
    asp
    Habermas, T. (2018). Kinds of emotional effects of narratives. In Emotion and narrative: Perspectives in autobiographical storytelling (pp. 97–121). Cambridge
    University Press.
    Hallam, S., Cross, I., & Thaut, M. (2009). Oxford handbook of music psychology. Oxford
    University Press.
    He, H., Jin, J., Xiong, Y., Chen, B., Sun, W., & Zhao, L. (2008). Language feature
    mining for music emotion classification via supervised learning from lyrics. Advances in Computation and Intelligence: Third International Symposium, ISICA
    2008 Wuhan, China, December 19-21, 2008 Proceedings 3, 426–435.
    Herman, D., Phelan, J., Rabinowitz, P. J., Richardson, B., & Warhol, R. (2012). Narrative theory: Core concepts and critical debates. The Ohio State University
    Press.
    Hinchman, K. A., & Moore, D. W. (2013). Close reading: A cautionary interpretation.
    Journal of Adolescent & Adult Literacy, 56(6), 441–450.
    Houjeij, A., Hamieh, L., Mehdi, N., & Hajj, H. (2012). A novel approach for emotion
    classification based on fusion of text and speech. 2012 19th International Conference on Telecommunications (ICT), 1–6.
    Hu, X., & Downie, J. S. (2010). Improving mood classification in music digital libraries
    by combining lyrics and audio. Proceedings of the 10th annual joint conference
    on Digital libraries, 159–168.
    Hu, Y., Chen, X., & Yang, D. (2009). Lyric-based song emotion detection with affective
    lexicon and fuzzy clustering method. ISMIR, 123–128.
    Jain, S., & Wallace, B. C. (2019). Attention is not explanation. Proceedings of NAACLHLT, 3543–3556.
    Juslin, P. N., & Laukka, P. (2004). Expression, perception, and induction of musical
    emotions: A review and a questionnaire study of everyday listening. Journal of
    new music research, 33(3), 217–238.
    Kaan, S. (2021). Themes and narrative structures in the lyrics of hozier [B.S. thesis]. S.
    Kaan.
    Kim, M., & Kwon, H.-C. (2011). Lyrics-based emotion classification using feature selection by partial syntactic analysis. 2011 IEEE 23rd International Conference
    on Tools with Artificial Intelligence, 960–964.
    Ko, D. (2014). Lyric analysis of popular and original music with adolescents. Journal
    of Poetry Therapy, 27(4), 183–192.
    Kreuter, M. W., Green, M. C., Cappella, J. N., Slater, M. D., Wise, M. E., Storey, D.,
    Clark, E. M., O’Keefe, D. J., Erwin, D. O., Holmes, K., et al. (2007). Narrative
    communication in cancer prevention and control: A framework to guide research
    and application. Annals of behavioral medicine, 33, 221–235.
    Lee, L.-H., Li, J.-H., & Yu, L.-C. (2022). Chinese emobank: Building valence-arousal
    resources for dimensional sentiment analysis. Transactions on Asian and Low-
    Resource Language Information Processing, 21(4), 1–18.
    Li, C., Li, J. W., Pun, S. H., & Chen, F. (2021). An erp study on the influence of lyric
    to song’s emotional state. 2021 10th International IEEE/EMBS Conference on
    Neural Engineering (NER), 933–936.
    Liao, J.-Y., Lin, Y.-H., Lin, K.-C., & Chang, J.-W. (2021). 以遷移學習改善深度神經網
    路模型於中文歌詞情緒辨識 (using transfer learning to improve deep neural
    networks for lyrics emotion recognition in chinese). International Journal of
    Computational Linguistics & Chinese Language Processing, 26(2).
    Liu, T. Y. (2021). 台灣 2008 至 2020 年音樂治療相關碩士學位論文內容分析 (content analysis of music therapy-related master’s degree theses in taiwan from 2008 to 2020) [Doctoral dissertation].
    Liu, Y., Liu, Y., Zhao, Y., & Hua, K. A. (2015). What strikes the strings of your heart?— feature mining for music emotion analysis. IEEE TRANSACTIONS on Affective computing, 6(3), 247–260.
    Luck, G., Toiviainen, P., Erkkilä, J., Lartillot, O., Riikkilä, K., Mäkelä, A., Pyhäluoto, K., Raine, H., Varkila, L., & Värri, J. (2008). Modelling the relationships be- tween emotional responses to, and musical content of, music therapy improvisations. Psychology of music, 36(1), 25–45.
    Ma, W.-Y., & Chen, K.-J. (2003). Introduction to CKIP Chinese word segmentation system for the first international Chinese word segmentation bakeoff. Proceedings of the Second SIGHAN Workshop on Chinese Language Processing, 168–171. https://doi.org/10.3115/1119250.1119276
    Malheiro, R., Panda, R., Gomes, P., & Paiva, R. P. (2016). Emotionally-relevant features for classification and regression of music lyrics. IEEE Transactions on Affective Computing, 9(2), 240–254.
    McKinney, M., & Breebaart, J. (2003). Features for audio and music classification. Mohsin, M. A., & Beltiukov, A. (2019). Summarizing emotions from text using plutchik’
    s wheel of emotions. 7th Scientific Conference on Information Technologies for
    Intelligent Decision Making Support (ITIDS 2019), 291–294.
    Mokhsin, M. B., Rosli, N. B., Adnan, W. A. W., & Manaf, N. A. (2014). Automatic music emotion classification using artificial neural network based on vocal and
    instrumental sound timbres. SoMeT, 3–14.
    Negus, K. (2012). Narrative, interpretation and the popular song. Musical Quarterly,
    95(2-3), 368–395.
    Nicholls, D. (2007). Narrative theory as an analytical tool in the study of popular music
    texts. Music and Letters, 88(2), 297–315.
    Palmer, A. (2015). Narrative and minds in the traditional ballads of early country music.
    In Narrative theory, literature, and new media (pp. 205–220). Routledge. Plutchik, R. (1980). A general psychoevolutionary theory of emotion. In Theories of
    emotion (pp. 3–33). Elsevier.
    Rajesh, S., & Nalini, N. (2020). Musical instrument emotion recognition using deep recurrent neural network. Procedia Computer Science, 167, 16–25.
    Randle, Q. (2013). So what does” set fire to the rain” really mean? a typology for analyzing pop song lyrics using narrative theory and semiotics. MEIEA Journal, 13(1), 125–147.
    Revathy, V., Pillai, A. S., & Daneshfar, F. (2023). Lyemobert: Classification of lyrics’ emotion and recommendation using a pre-trained model. Procedia Computer Science, 218, 1196–1208.
    Riessman, C. (2005). Narrative analysis in narrative, memory, & everyday life. university of huddersfield, huddersfield.
    Rimé, B. (2009). Emotion elicits the social sharing of emotion: Theory and empirical review. Emotion review, 1(1), 60–85.
    Rolvsjord, R. (2001). Sophie learns to play her songs of tears: –a case study exploring the dialectics between didactic and psychotherapeutic music therapy practices. Nordic Journal of Music Therapy, 10(1), 77–85.
    Russell, J. A. (1980). A circumplex model of affect. Journal of personality and social psychology, 39(6), 1161.
    Russell, J. A. (2003). Core affect and the psychological construction of emotion. Psychological review, 110(1), 145.
    Ryan, M.-L. (2015). Texts, worlds, stories: Narrative worlds as cognitive and ontological concept. In Narrative theory, literature, and new media (pp. 11–28). Routledge. Salim, S., Iqbal, Z., & Iqbal, J. (2021). Emotion classification through product consumer
    reviews. Pakistan Journal of Engineering and Technology, 4(4), 35–40.
    Shi, W., & Feng, S. (2018). Research on music emotion classification based on lyrics and audio. 2018 IEEE 3rd Advanced Information Technology, Electronic and
    Automation Control Conference (IAEAC), 1154–1159.
    Shukla, S., Khanna, P., & Agrawal, K. K. (2017). Review on sentiment analysis on
    music. 2017 International Conference on Infocom Technologies and Unmanned
    Systems (Trends and Future Directions)(ICTUS), 777–780.
    Smith, B. H. (2016). What was “close reading”? a century of method in literary studies. The Minnesota Review, 2016(87), 57–75.
    Sundararajan, M., Taly, A., & Yan, Q. (2017). Axiomatic attribution for deep networks. International conference on machine learning, 3319–3328.
    Talebi, S., Tong, E., Li, A., Yamin, G., Zaharchuk, G., & Mofrad, M. R. (2024). Exploring the performance and explainability of fine-tuned bert models for neuroradiology protocol assignment. BMC Medical Informatics and Decision Making, 24(1), 40.
    Tan, E. S.-H. (1995). Film-induced affect as a witness emotion. Poetics, 23(1-2), 7–32. Thayer, R. E. (1990). The biopsychology of mood and arousal. Oxford University Press. Tzanetakis, G., & Cook, P. (2002). Musical genre classification of audio signals. IEEE
    Transactions on speech and audio processing, 10(5), 293–302.
    Ujlambkar, A. M., & Attar, V. Z. (2012). Automatic mood classification model for indian popular music. 2012 Sixth Asia Modelling Symposium, 7–12.
    Ullah, R., Amblee, N., Kim, W., & Lee, H. (2016). From valence to emotions: Exploring the distribution of emotions in online product reviews. Decision Support
    Systems, 81, 41–53.
    van Gulik, R., Vignoli, F., & van de Wetering, H. (2004). Mapping music in the palm
    of your hand, explore and discover your collection. Proceedings of the 5th In-
    ternational Conference on Music Information Retrieval.
    Wang, J., & Yang, Y. (2019). Deep learning based mood tagging for chinese song lyrics.
    arXiv preprint arXiv:1906.02135.
    Weninger, F., Eyben, F., Mortillaro, M., & Scherer, K. R. (2013). On the acoustics of
    emotion in audio: What speech, music, and sound have in common. Frontiers in
    psychology, 4, 51547.
    Wicentowski, R., & Sydes, M. R. (2012). Emotion detection in suicide notes using maximum entropy classification. Biomedical informatics insights, 5, BII–S8972. Wilson, T., Wiebe, J., & Hoffmann, P. (2005). Recognizing contextual polarity in phrase- level sentiment analysis. Proceedings of human language technology conference and conference on empirical methods in natural language processing, 347–354. Zad, S., & Finlayson, M. (2020). Systematic evaluation of a framework for unsupervised emotion recognition for narrative text. Proceedings of the First Joint Workshop
    on Narrative Understanding, Storylines, and Events, 26–37.
    Zhong, J., Cheng, Y., Yang, S., & Wen, L. (2012). Music sentiment classification integrating audio with lyrics. JOURNAL OF INFORMATION &COMPUTATIONAL SCIENCE, 9(1), 35–44.
    Description: 碩士
    國立政治大學
    語言學研究所
    110555005
    Source URI: http://thesis.lib.nccu.edu.tw/record/#G0110555005
    Data Type: thesis
    Appears in Collections:[語言學研究所] 學位論文

    Files in This Item:

    File Description SizeFormat
    500501.pdf9092KbAdobe PDF0View/Open


    All items in 政大典藏 are protected by copyright, with all rights reserved.


    社群 sharing

    著作權政策宣告 Copyright Announcement
    1.本網站之數位內容為國立政治大學所收錄之機構典藏,無償提供學術研究與公眾教育等公益性使用,惟仍請適度,合理使用本網站之內容,以尊重著作權人之權益。商業上之利用,則請先取得著作權人之授權。
    The digital content of this website is part of National Chengchi University Institutional Repository. It provides free access to academic research and public education for non-commercial use. Please utilize it in a proper and reasonable manner and respect the rights of copyright owners. For commercial use, please obtain authorization from the copyright owner in advance.

    2.本網站之製作,已盡力防止侵害著作權人之權益,如仍發現本網站之數位內容有侵害著作權人權益情事者,請權利人通知本網站維護人員(nccur@nccu.edu.tw),維護人員將立即採取移除該數位著作等補救措施。
    NCCU Institutional Repository is made to protect the interests of copyright owners. If you believe that any material on the website infringes copyright, please contact our staff(nccur@nccu.edu.tw). We will remove the work from the repository and investigate your claim.
    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - Feedback