政大機構典藏-National Chengchi University Institutional Repository(NCCUR):Item 140.119/49442
English  |  正體中文  |  简体中文  |  Post-Print筆數 : 27 |  全文笔数/总笔数 : 93779/124226 (75%)
造访人次 : 28876640      在线人数 : 531
RC Version 6.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
搜寻范围 查询小技巧:
  • 您可在西文检索词汇前后加上"双引号",以获取较精准的检索结果
  • 若欲以作者姓名搜寻,建议至进阶搜寻限定作者字段,可获得较完整数据
  • 进阶搜寻
    政大機構典藏 > 理學院 > 心理學系 > 學位論文 >  Item 140.119/49442

    请使用永久网址来引用或连结此文件: http://nccur.lib.nccu.edu.tw/handle/140.119/49442

    题名: 動畫角色肢體動作的情緒表達--探討動作特性及身體方向對情緒表達的影響
    Emotion expressiveness of animated character's body movement
    作者: 劉家揚
    贡献者: 黃淑麗
    关键词: 情緒表達
    emotion expression
    approach-avoidance motivation
    body movement
    movement quality
    body direction
    日期: 2010
    上传时间: 2010-12-08 11:39:48 (UTC+8)
    摘要: 肢體動作是情緒表達的方式之一,本研究旨在探討表達情緒時,表達者「動作特性」及「身體方向」對肢體動作情緒表達的影響,及其二者的關係。參考過去研究,本研究使用生氣、害怕、開心及悲傷四種情緒的肢體動作,並操弄平滑度、僵硬度、速度、力道及擴張度五種動作特性;0°、45°、90°三種身體方向,並且以正確率、反應時間及情緒強度作為指標,進行區辨作業及評分作業。本研究共分兩個實驗,每個實驗皆由兩個子實驗構成,兩個子實驗分別使用「一般動作」及「特殊動作」作為呈現刺激。實驗一目的在於探討「動作特性」對情緒表達的影響;實驗二目的在於探討「動作特性」和「身體方向」共同對情緒表達的影響。實驗一A使用一般動作並操弄五種動作特性,藉此得到各情緒的動作特性組合;實驗一B使用特殊動作並操弄動作特性相容性,結果顯示動作特性相容性確實有其效果,相容情況的情緒強度高於不相容情況。實驗二A、二B分別使用一般動作及特殊動作,並操弄五種動作特性及三種身體方向,結果顯示當身體方向符合情緒的趨避動機時,其正確率、反應時間及情緒強度的表現皆較好。综合而言,「動作特性」及「身體方向」在肢體動作表達情緒時確實有其效果,當動作特性相容時,可增進情緒表達力;當身體方向符合該情緒的趨避動機時,也可增進情緒表達力,但動作特性及身體方向二者間的關係為何,本研究仍無法下一定論。

    Body movement is one of the ways to express emotion. The purpose of this study is to explore the effects of movement quality and body direction on the emotion expressiveness of body movement. Referring to some previous studies, four kinds of emotion (anger, fear, happiness and sadness) were included in the present study. Five movement qualities (smoothness, stiffness, speed, strength and expansion) and three body directions (0°, 45° and 90°) were manipulated as the independent variables. Response accuracy and reaction times of an identification task and rating scores of the emotion rating task were measured as the dependent variables. There are two experiments in this study which includes two sub-experiments each. One of the sub-experiments adopted non-propositional body movements as stimuli (Experiment 1A & 2A), and the other adopted propositional body movements as stimuli (Experiment 1B & 2B). Experiment 1 aimed to explore the effect of movement qualities. In Experiment 1A, movement quality combinations which can express each of the four emotions were found. Based on this result, in Experiment 1B, movement quality compatibility was manipulated. The results of Experiment 1B confirmed the effect of movement quality compatibility on emotion strength in non-propositional body movement. Experiment 2 manipulated five movement qualities and three body directions. The results showed that when body direction was compatible with the approach-avoidance motivation of the expressed emotion, participant’s performances on response accuracy, reaction times and rating scores were better than that of incompatible ones. In conclusion, both movement quality and body direction can influence the emotion expressiveness of body movement. When movement qualities and body directions are compatible with the expressed emotion, the strength of emotion can be increased. But the relative contributions and the interaction effect of movement qualities and body directions are still unclear.

    Keywords: emotion expression, approach-avoidance motivation, body movement, movement quality, body direction
    林岳黌 (2009)。以參數化程序產生具情緒表達能力之3D肢體動作。未出版碩士論文,國立政治大學,台北市。
    Adams, R. B., Jr., & Kleck, R. E. (2003). Perceived gaze direction and the processing of facial displays of emotion. Psychological Science, 14, 644–647.
    Adams, R. B., Jr., & Kleck, R. E. (2005). The effects of direct and averted gaze on the perception of facially communicated emotion. Emotion, 5, 3-11.
    Alexander, S. (2006). Are interface agents scapegoats? Attributions of responsibility in human-agent interaction. Interacting with Computers, 19, 293-303.
    André, E., Muller, J., & Rist, T. (1996). The PPP Persona: A multipurpose animated presentation agent. In: Proceedings Advanced Visual Interfaces(p.245-247). New York, USA: ACM Press,.
    André, E., & Rist, T., Eds. (1997). WebPersona: A life-like presentation agent for the world-wide web. Proceedings of the IJCAI-97 Workshop on Animated Interface Agents: Making them Intelligent. Nagoya, Japan.
    Atkinson, A. P., Dittrich, W. H., Gemmell, A. J., & Young, A. W. (2004). Emotion perception from dynamic and static body expressions in point-light and full light displays. Perception, 33, 717–746.
    Ball, G., Ling, D., Kurlander, D., Miller, J., Pugh, D., Skelly, T., Stankosky, A., Thiel, D., Dantzich, M. V., & Wax, T. (1997). Lifelike computer characters: the Persona project at Microsoft Research. In: Bradshaw, J.M. (Ed.), Software Agents (p.191-222). The AAAI Press / The MIT Press, Menlo Park, CA.
    Bargh, J. A. (1997). The automaticity of everyday life. In R S. Wyer Jr. (Ed.), The automaticity of everyday life: Advances in social cognition (Vol. 10, pp. 1–61). Mahwah, NJ: Erlbaum.
    Berry, D. C., Butler, L. T., & Rosis, F. de. (2005). Evaluating realistic agent in an advice- giving task. Human–Computer Studies, 63(3), 304–327.
    Beskow, J. and McGlashan, S. (1997). Olga: A Conversational Agent with Gestures. In André, E., editor, Proceedings of the IJCAI 1997 Workshop on Animated Interface Agents: Making Them Intelligent, Nagoya, Japan. San Francisco: Morgan-Kaufmann Publishers.
    Brave, S., Nass, C., Hutchinson, K. (2005). Computers that care: investigating the effects of orientation of emotion exhibited by an embodied computer agent. International Journal of Human–Computer Studies, 62, 161–178.
    Cacioppo, J. T., Bernston, G. (1994). Relationship between attitudes and evaluative space: A critical review, with emphasis on the separability of positive and negative substrates. Psychological Bulletin, 115, 401-422.
    Camurri, A., Lagerlöf, I., & Volpe, G. (2003). Recognizing Emotion from Dance Movement: Comparison of Spectator Recognition and Automated Techniques. International Journal of Human‐Computer Studies, 59, 213‐225.
    Cassell, J., Pelachaud, C., Badler, N., Steedman, M., Achorn, B., Beckett, T., Douville, B., Prevost, S., & Stone, M. (1994). Animated conversation: rule-based generation of facial expression, gesture, and spoken intonation for multiple conversational agents. Proceedings of the SIGGRAPH Conference on Computer Graphics (p.413-420). Florida, USA: ACM press.
    Cohen, I., Sebe, N., Chen, L., Garg, A., & Huang, T. S., (2003). Facial expression recognition from video sequences:temporal and static modeling, computer vision and image understanding, 91,160–187.
    Coulson, M. (2004). Attributing emotion to static body postures: recognition accuracy, confusions, and viewpoint dependence. Journal of Nonverbal Behavior, 28(2),117-139.
    Corwin, G. (1921). Minor studies from the psychological laboratory of Cornell University . American Journal of Psychology, 32, 563-570.
    Curby, K. M., Glazek, K., & Gauthier, I. (2009). A visual short-term memory advantage for objects of expertise. Journal of Experimental Psychology: Human Perception and Performance, 35, 94–107.
    Davidson, R. J. (1993). Parsing affective space: Perspectives from neuropsychology
    and psychophysiology. Neuropsychology, 7, 464–475.
    Ekman, P. (2003): Emotions revealed. Weidenfeld & Nicolson.
    Fogg, B. J., & Tseng, H. (1999). The elements of computer credibility. CHI’99 conference proceedings (p.80-87). Pennsylvania, USA: ACM press.
    Gajewski, P. D., Schlegel, K., & Stoerig, P. (2008). Effects of human race and face inversion on the N170. Journal of Psychophysiology, 22(4), 157-165.
    Gunes, H., Piccardi, M., & Jan, T. (2007). Face and body gesture recognition for a vision-based multimodal analyzer. In Proc. 2003 Pan-Sydney Area Workshop on Visual Information Processing (p.19-28). Sydney, Australia: CRPIT, 36. Piccardi, M., Hintz, T., He, S., Huang, M. L., & Feng, D. D., Eds. ACS.
    Helmut, P., Chunling, M., & Mitsuru, I. (2007). Eye movements as indices for the utility of life-like interface agents: A pilot study. Interacting with Computers, 19 (2), 281-292.
    Huard, R. E., & Hayes-Roth, B. (1997). Character mastery with improvisational pupets. Proceedings of the IJCAI-97 Workshop on Animated Interface Agents: Making them Intelligent(p.85-89). Nagoya, Japan.
    James, W. T. (1932). A study of the expression of body posture. Journal of General Psychology, 7, 405–437.
    Klein, J., Moon,Y., & Picard, R.W. (2002). This computer responds to user frustration: theory, design, and results. Interacting with Computers, 14, 119–140.
    Lang, P. (1995). Studies of motivation and attention. American Psychologist, 50, 372-385
    LeDoux, J. (1995). Emotion:Clues from the brain. Annual Review of Psychology, 46, 209-235.
    Lester, J. C., & Stone, B. A. (1997). Increasing believability in animated pedagogical agents. Proceedings of the 1st International Conference on Autonomous Agents(p16-21). CA, USA.
    Lewin, K. (1935). A dynamic theory of personality. NY: McGraw-Hill.
    Maes, P. (1994). Agents that reduce work and information overload. Communication. ACM 37 (7), 31–40.
    Marsh, A. A., Ambady, N., & Kleck, R. E. (2005). The effects of fear and anger facial expressions on approach- and avoidance-related behaviors. Emotion, 5, 119–124.
    Montepare, J., Koff, E., Zaitchik, D., & Albert, M. S. (1999). The use of body movements and gestures as cues to emotion in younger and older adults. Journal of Nonverbal Behavior, 23, 133-152.
    Moundridou, M., Virvou, M. (2002). Evaluating the persona effect of an interface agent in a tutoring system. Journal of Computer Assisted Learning, 18, 253–261.
    Osgood, C., Suci, G., & Tannenbaum, P. (1957). The measurement of meaning. Urbana
    Petrushin, V. A. (2002). Creating emotion recognition agents for speech signal. In: Dautenhahn, K., Bond, A.H., Canamero, L., Edmonds, B. (Eds.), Socially Intelligent Agents: Creating Relationships with Computers and Robots. The Netherlands: Kluwer Academic Publishers, Dordrecht.
    Prendinger, H., Mori, J., Ishizuka, M., (2005). Using human physiology to evaluate subtle expressivity of a virtual quizmaster in a mathematical game. International Journal of Human-Computer Studies, 62, 231–245
    Reed, C. L., Stone, V., Bozova, S., & Tanaka, J. (2003). The body inversion effect. Psychological Science, 14, 302–308
    Rickel, J., & Johnson, W. L. (1999). Animated agents for procedural training in virtual reality: perception, cognition, and motor control. Applied Artificial Intelligence, 13, 343- 382.
    Sherington, C. A. (1906). The integrative actions of the nervous system. NY:Scribner’s.
    Shizgal, P. (1999). On the neural computation of utility: Implications from studies of brain stimulation and reward. In D. Kahneman, E. Diener, & N. Schwarz (Eds.), Well-being : The foundations of hedonic psychology(p.500-524). NY: Russell Sage Foundation.
    Sprengelmeyer, R., Young, A. W., Schroeder, U., Grossenbacher, P. G., Federlein, J., Buttner, T., & Przuntek, H. (1999). Knowing no fear. Proceedings of the Royal Society of London, Series B 266, 2451 – 2456.
    Stekelenburg, J. J., & Gelder, B. de (2004). The neural correlates of perceiving human bodies: An ERP study on the body-inversion effect. Neuroreport, 15(5), 777-780
    Stellar, J. R., & Stellar, E. (1985). The neurobiology of motivation and reward. NY: Springer-Verlag.
    Stock, J.V., Righart, R. & Gelder, B. (2007). Body expressions influence recognition of emotions in the face and voice. Emotion, 7, 487-494.
    Sullivan, J. & Cassell, J. (1998). Proceedings of the 1st Workshop on Embodied ConversationalCharacters (pp.12-15). Lake Tahoe, CA.
    Takeuchi, A., & Naito, T. (1995). Situated facial displays: towards social interaction. In I. Katz, R. Mack, L. Marks, M. B. Rosson, & J. Nielsen, Eds. Human Factors in Computing Systems: CHI’95 Conference Proceedings(p.450-455). New York: ACM Press.
    描述: 碩士
    資料來源: http://thesis.lib.nccu.edu.tw/record/#G0095752003
    数据类型: thesis
    显示于类别:[心理學系] 學位論文


    档案 描述 大小格式浏览次数
    200301.pdf951KbAdobe PDF2689检视/开启


    社群 sharing

    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - 回馈