Reference: | [1] B. Adams, C. Dorai, and S.Venkatesh, “Toward Automatic Extraction of Expressive Elements from Motion Pictures: Tempo,” IEEE Transactions on Multimedia, Vol. 4, No. 4, pp. 472-481, December 2002. [2] D. Arijon, Grammar of the Film Language. CA: Silman-James Press, 1976. [3] Christopher J. C. Burges, “A Tutorial on Support Vector Machines for Pattern Recognition,” Journal of Data Mining and Knowledge Discovery, Vol. 2, No. 2, pp. 121-167, 1998. [4] A. R. Damasio, The Feeling of What Happens: Body and Emotion in the Making of Consciousness. New York: Harcourt Brace, 1999. [5] R. Dietz and A. Lang, “Affective Agents: Effects of Agent Affect on Arousal, Attention, Liking and Learning,” Proceedings of Cognitive Technology Conference, San Francisco, CA, 1999. [6] N. Dimitrova, J. Martino, H. Elenbaas, and L. Agnihotri, “Color SuperHistograms for Video Representation,” IEEE International Conference on Image Processing (ICIP ‘99), Kobe, Japan, Vol. 3, pp. 314-318, October 1999. [7] P. Ekman, “Universals and Cultural Differences in the Judgments of Facial Expressions of Emotion,” Journal of Personality and Social Psychology, Vol. 54, No. 4, pp. 712-717, October 1987. [8] L. Giannetti, Understanding Movies, 10th ed. Englewood Cliffs, New Jersey: Prentice Hall, 2005. [9] A. Hanjalic and L. Q. Xu, “Extracting Moods from Pictures and Sounds: Towards truly personalized TV,” IEEE Signal Processing Magazine, Vol. 23, No. 2, pp. 90-100, March 2006. [10] A. Hanjalic and L. Q. Xu, “Affective Video Content Representation and Modeling,” IEEE Transaction on Multimedia, Vol. 7, No. 1, pp. 143-154, February 2005. [11] A. Hanjalic and L. Q. Xu, “User-oriented Affective Video Content Analysis,” Proceedings of IEEE CBAIBL, Kauai, Hawaii, pp. 50-57, December 2001. [12] H. B. Kang, “Affective Content Retrieval from Video with Relevance Feedback,” International Conference on Asian Digital Libraries, Kuala Lumpur, Malaysia, pp. 243-252, December 2003. [13] H. B. Kang, “Affective Content Detection using HMMs,” Proceedings of ACM International Conference on Multimedia, Berkeley, California, U.S.A, pp. 259-262, November 2003. [14] G. Kirouac, Les émotions: Monographies de psychologie. Sillery: Presses de l’Université du Québec, 1992. [15] F. F. Kuo, M. F. Chiang, M. K. Shan, and S. Y. Lee, “Emotion-based Music Recommendation by Association Discovery from Film Music,” Proceedings of ACM International Conference on Multimedia, Singapore, pp. 507-510, November 2005. [16] Y. Li, S. H. Lee, C. H. Yeh, and C. C. J. Kuo, ”Techniques for Movie Content Analysis and Skimming,” IEEE Signal Processing Magazine, Vol. 23, No. 2, pp. 79-89, March 2006. [17] L. Lu, H. Jiang, and H. J. Zhang, "A Robust Audio Classification and Segmentation Method," Proceedings of ACM International Conference on Multimedia, Ottawa, Ontario, Canada, pp. 203-211, September 2001. [18] L. Lu, H. J. Zhang, and H. Jiang, "Content Analysis for Audio Classification and Segmentation," IEEE Transaction on Speech and Audio Processing, Vol. 10, No. 7, pp. 504-516, October 2002. [19] F. H. Mahnke, Color, Environmental and Human Response. New York: Van Nostrand Reinhold, 1996. [20] S. Moncrieff, C. Dorai, and S. Venkatesh, “Affect Computing in Film through Sound Energy Dynamics,” Proceedings of ACM International Conference on Multimedia, Ottawa, Ontario, Canada, pp. 525-527, September 2001. [21] A. Ortony, G. Clore, and A. Collins, The Cognitive Structure of Emotions. New York: Oxford University Press, 1988. [22] C. E. Osgood, G. J. Suci, and P. H. Tannenbaum, The Measurement of Meaning. Urbana, IL: University of Illinois Press, 1957. [23] J. Y. Pan, H. J. Yang, P. Duygulu, and C. Faloutsos, “Automatic Image Captioning,” Proceedings of IEEE International Conference on Multimedia and Expo (ICME ’04), Taipei, Taiwan, pp. 1987-1990, June 2004. [24] J. Y. Pan, H. J. Yang, C. Faloutsos, and P. Duygulu, “Automatic Multimedia Cross-modal Correlation Discovery,” Proceedings of ACM International Conference on Knowledge Discovery on Database (SIGKDD ‘04), Seattle, Washington, pp. 653-658, August 2004. [25] D. S. Park, J. S. Park, and J. H. Han, “Image Indexing Using Color Histogram in the CIELUV Color Space,” Proceedings of 5th Japan-Korea Workshop on Computer Vision, Korea, pp. 126-132, 1999. [26] G. Peeters, “A Large Set of Audio Features for Sound Description (Similarity and Classification),” in the CUIDADO project. Technical report, Ircam, Paris, France, April 2004. [27] Z. Rasheed, Y. Sheikh, and M. Shah, “On the Use of Computable Features for Film Classification,” IEEE Transaction on Circuits and Systems for Video Technology (CSVT), Vol. 15, No. 1, pp. 52-64, January 2005. [28] J. A. Russell and A. Mehrabian, “Evidence for a Three-Factor Theory of Emotions,” Journal of Research in Personality, Vol. 11, pp. 273-294, 1977. [29] J. Saunders, “Real-Time Discrimination of Broadcast Speech/Music,” Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP ’96), Atlanta, Ga, Vol. 2, pp. 993-996, May 1996. [30] E. Scheirer and M. Slaney, “Construction and Evaluation of a Robust Multifeature Speech/Music Discriminator”, Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP ’97), Munich, Germany, Vol. 2, pp. 1331-1334, April 1997. [31] R. E. Thayer, The Biopsychology of Mood and Arousal. New York: Oxford University Press, 1989. [32] H. L. Wang and L. F. Cheong, “Affective Understanding in Film,” IEEE Transactions on Circuits and Systems for Video Technology (CSVT), Vol. 16, No. 6, pp. 689-704, June 2006. [33] C. Y. Wei, N. Dimitrova, and S.F. Chang, “Color-Mood Analysis of Films on Syntactic and Psychological Models,” Proceedings of IEEE International Conference on Multimedia and Expo. (ICME ’04), Taipei, Taiwan, pp. 831-834, June 2004. [34] H. Zettl, Sight Sound Motion: Applied Media Aesthetics, 3rd ed. Belmont, CA: Wadsworth Publishing Company, 1998. [35] http://www.intel.com/technology/computing/opencv/index.htm [36] http://eqi.org/fw.htm |