English  |  正體中文  |  简体中文  |  全文筆數/總筆數 : 84662/113307 (75%)
造訪人次 : 22350784      線上人數 : 531
RC Version 6.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
搜尋範圍 查詢小技巧:
  • 您可在西文檢索詞彙前後加上"雙引號",以獲取較精準的檢索結果
  • 若欲以作者姓名搜尋,建議至進階搜尋限定作者欄位,可獲得較完整資料
  • 進階搜尋
    請使用永久網址來引用或連結此文件: http://nccur.lib.nccu.edu.tw/handle/140.119/80531

    題名: Assessing the Attention Levels of Students by using a Novel Attention Aware System based on Brainwave Signals
    作者: Chen, Chih-Ming;Wang, Jung-Ying;Yu, Chih-Ming
    Chen, Chih-Ming
    貢獻者: 圖檔所
    日期: 2017-03
    上傳時間: 2016-01-13 11:14:01 (UTC+8)
    摘要: Rapid progress in information and communication technologies (ICTs) has fueled the popularity of e-learning. However, an e-learning environment is limited in that online instructors cannot monitor immediately whether students remain focus during online autonomous learning. Therefore, this study tries to develop a novel attention aware system (AAS) capable of recognizing students' attention levels accurately based on electroencephalography (EEG) signals, thus having high potential to be applied in providing timely alert for conveying low-attention level feedback to online instructors in an e-learning environment. To construct AAS, attention responses of students and their corresponding EEG signals are gathered based on a continuous performance test (CPT), ie, an attention assessment test. Next, the AAS is constructed by using training and testing data by the NeuroSky brainwave detector and the support vector machine (SVM), a well-known machine learning model. Additionally, based on the discrete wavelet transform (DWT), the collected EEG signals are decomposed into five primary bands (ie, alpha, beta, gamma, theta, and delta) as well as each primary band contains five statistical parameters (including approximate entropy, total variation, energy, skewness, and standard deviation), thus generating 25 potential brainwave features associated with students' attention level for constructing the AAS. An attempt based on genetic algorithm (GA) is also made to enhance the prediction performance of the proposed AAS in terms of identifying students' attention levels. According to GA, the seven most influential features are selected from 25 considered features; parameters of the proposed AAS are optimized as well. Analytical results indicate that the proposed AAS can accurately recognize individual student's attention state as either a high or low level, and the average accuracy rate reaches as high as 89.52%. Moreover, the proposed AAS is integrated with a video lecture tagging system to examine whether the proposed AAS can accurately detect students' low-attention periods while learning about electrical safety in the workplace via a video lecture. Four experiments are designed to assess the prediction performance of the proposed AAS in terms of identifying the periods of video lecture with high- or low-attention levels during learning processes. Analytical results indicate that the proposed AAS can accurately identify the low-attention periods of video lecture generated by students when engaging in a learning activity with video lecture. Meanwhile, the proposed AAS can also accurately identify the low-attention periods of video lecture generated by students to some degree even when students engage in a learning activity by a video lecture with random disturbances. Furthermore, strong negative correlations are found between the students' learning performance (ie, posttest score and progressive score) and the low-attention periods of video lecture identified by the proposed AAS. Results of this study demonstrate that the proposed AAS is effective, capable of assisting online instructors in evaluating students' attention levels to enhance their online learning performance.
    關聯: British Journal of Educational Technology, Volume48, Issue2, Pages 348-369
    資料類型: article
    DOI 連結: http://dx.doi.org/10.1111/bjet.12359
    DOI: 10.1111/bjet.12359
    顯示於類別:[圖書資訊與檔案學研究所] 期刊論文


    檔案 描述 大小格式瀏覽次數
    bjet12359.pdf804KbAdobe PDF726檢視/開啟


    社群 sharing

    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - 回饋