A Predictive Model To Evaluate Student Performance
A Predictive Model To Evaluate Student Performance
A Predictive Model To Evaluate Student Performance
net/publication/276863880
CITATIONS READS
23 1,931
4 authors, including:
Sachio Hirokawa
Kyushu University
287 PUBLICATIONS 860 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Tsunenori Mine on 28 August 2015.
Regular Paper
Abstract: In this paper we propose a new approach based on text mining techniques for predicting student per-
formance using LSA (latent semantic analysis) and K-means clustering methods. The present study uses free-style
comments written by students after each lesson. Since the potentials of these comments can reflect student learning at-
titudes, understanding of subjects and difficulties of the lessons, they enable teachers to grasp the tendencies of student
learning activities. To improve our basic approach using LSA and k-means, overlap and similarity measuring methods
are proposed. We conducted experiments to validate our proposed methods. The experimental results reported a model
of student academic performance predictors by analyzing their comments data as variables of predictors. Our proposed
methods achieved an average 66.4% prediction accuracy after applying the k-means clustering method and those were
73.6% and 78.5% by adding the overlap method and the similarity measuring method, respectively.
Keywords: comments data mining, latent semantic analysis (LSA), similarity measuring method, overlap method
c 2015 Information Processing Society of Japan 192
Journal of Information Processing Vol.23 No.2 192–201 (Mar. 2015)
contained in an unstructured collection of texts (student cluster similarity with LSA was much higher than that without
comment); we classify the results obtained after applying LSA. In addition, Hung [16] used clustering analysis as an ex-
LSA into 5 groups according to student grades by using the ploratory technique to examine e-learning literature and visual-
K-means clustering method. ized patterns by grouping sources that shared similar words and
• The similarity measuring method is proposed to calculate attribute values. In addition, Minami et al. [23] analyzed student
similarity between a new comment and comments in the attitudes towards learning, and investigated how the attitudes af-
nearest cluster, which is created in the training phase. fect final student evaluation; they pursued a case study of lecture
• The overlap method is introduced for a stable evaluation that data analysis in which the correlations exist between student at-
allows to accept the adjacent grade of its original grade cor- titudes to learning such as attendance and homework, as effort,
responding to 5-grade categories. To this end, we classify and the student examination scores, as achievement; they ana-
student marks into 9 grades. lyzed the student own evaluation and lectures based on a ques-
• The experiments were conducted to validate the proposed tionnaire. Through this study, Minami et al. showed that a lec-
methods: basic, overlap and similarity measuring methods turer could give feedback to students who tended to over-evaluate
by calculating the F-measure and the accuracy for each themselves, and let the students recognize their real positions in
method in estimating final student grades. The experimental the class.
results illustrate the validity of the proposed methods. Previous studies show that we need to understand individual
The rest of the paper is organized as follows. Section 2 gives students more deeply and recognize student learning status and
an overview of some related literature. Section 3 introduces the attitude to give feedback to them. We need to comprehend student
overview of our research and the procedures of the proposed characteristics by letting them describe themselves about their ed-
methods. Section 4 describes the methodology of our proposed ucational situations such as understanding of subjects, difficulties
methods. Section 5 discusses some of the highlighted experimen- to learn, learning activities in the classroom, and their attitude
tal results. Finally, Section 6 concludes the paper and describes toward the lesson. Researchers have used various classification
our future work. methods and various data in their studies to predict student aca-
demic performance.
2. Related Work Different from the above studies, Goda et al. [13] proposed the
The ability to predict student performance is very impor- PCN method to estimate student learning situations on the ba-
tant in educational environments. Increasing students’ success sis of their freestyle comments written just after the lesson. The
in their learning environment is a long-term goal in all aca- PCN method categorizes their comments into three items: P (Pre-
demic institutions. In recent years, there is a growing inter- vious), C (Current), and N (Next) so that it can analyze the com-
est in employing educational data mining techniques (EDM) to ments from the points of views of their time-oriented learning
conduct the automatic analysis and prediction of learner perfor- situations. Goda et al. [12] also conducted another study on us-
mance [2], [5], [8], [14], [24]. An emerging trend in EDM is ing PCN scores to determine the level of validity of assessment
the use of text mining which is an extension of data mining to based on student comments and showed strong correlations be-
text data [19], [20], [26]. Many researchers have successfully tween the PCN scores and the prediction accuracy of final student
used text mining techniques to analyze large amounts of tex- grades. They employed multiple regression analysis to calcu-
tual data in business, health science and educational domains late PCN scores. Their results indicated that students who wrote
[11], [21], [28], [29]. In our research we focus on using text min- comments with high PCN scores were considered as those who
ing in education fields to predict student grades. described their learning attitude appropriately. In addition, ap-
plying machine learning method support vector machine (SVM),
2.1 Educational Text Mining they illustrated that as student comments got higher PCN scores,
Text mining focuses on finding and extracting useful or in- prediction performance of their grades became higher. Goda et
teresting patterns, models, directions, trends, or rules from an al., however, did not discuss prediction performance of final stu-
unstructured text such as in text documents, HTML files, chat dent grades.
messages and emails. In addition, the major applications of text The current study is an extension of Goda et al. [12]; we fo-
mining include automatic classification (clustering), information cus on accuracy of prediction of final student grades. Using C-
extraction, text summarization, and link analysis [3]. As an au- comments from the PCN method, we try to predict their grade in
tomated technique, text mining can be used to efficiently and each lesson and discuss the changes in accuracy and F-measure
systematically identify, extract, manage, integrate, and exploit over a sequence of lessons.
knowledge for research and education [1].
Currently, there are only several studies about how to use text
3. Overview of the Prediction Method
mining techniques to analyze learning related data. For exam- 3.1 PCN Method and Student Grade
ple, Tane et al. [28] used text mining (text clustering techniques) To grasp student lesson attitudes and learning situations and to
to group e-learning resources and documents according to their give feedback to each student are educational foundations. Goda
topics and similarities. Antai et al. [4] classified a set of docu- et al. [13] proposed the PCN method to estimate learning situa-
ments according to document topic areas by using CLUTO pro- tions from comments freely written by students. Each student
gram with and without LSA. The results showed that the internal described his/her learning tendency, attitudes, and understanding
c 2015 Information Processing Society of Japan 193
Journal of Information Processing Vol.23 No.2 192–201 (Mar. 2015)
c 2015 Information Processing Society of Japan 194
Journal of Information Processing Vol.23 No.2 192–201 (Mar. 2015)
that most frequently appears in the cluster. We call the grade the last twenty years its capacity to simulate aspects of human
the dominant grade in the cluster. semantics has been widely demonstrated [18]. LSA is based on
( 4 ) Test Phase: The test phase evaluates the performance of three fundamental ideas: (1) to begin to simulate human seman-
prediction models by calculating Accuracy and F-measure. tics of language, we first obtain an occurrence matrix of terms
This phase first extracts words from a new comment, and contained in a comment, (2) the dimensionality of this matrix
transforms an extracted-word vector of the comment to a set is reduced using singular value decomposition, a mathematical
of K-dimensional vectors by using LSA. To evaluate the technique that effectively represents abstract concepts, and (3)
prediction performance, 10-fold cross validation was used. any word or text is represented by a vector in this new latent se-
90% of comments were classified as training data and con- mantic space [7], [18].
structed a model, then the model was applied to the rest 10% 4.2.1 Singular Value Decomposition
of comments as test data, and compared a predicted value LSA works through singular value decomposition (SVD), a
with corresponding the original data. The procedure was re- form of factor analysis. The singular value decomposition of A is
peated 10 times and the results were averaged. To improve defined as:
the prediction accuracy of student grade, we employed the
similarity measuring and overlap methods in addition to our A = US V T (2)
basic method. The details of the two methods are described where U and V are the matrices of the term vectors and document
in Sections 4.4 and 4.5, respectively. vectors. S = diag(r1 , . . . , rn ) is the diagonal matrix of singular
( 5 ) Noisy Data detection: Outlier analysis can be used to de- values. To reduce the dimensions, we can simply choose the k
tect data that adversely affect the results. In this paper, we largest singular values and the corresponding left and right sin-
detect outliers in two phases: training phase and test phase. gular vectors, the best approximation of A with rank-k matrix is
As a threashold to check the outliers in a cluster, we use the given by
standard deviation (S d) of each distance between each mem-
ber and the center in a cluster in the training phase, and the Ak = Uk S k VkT (3)
average of each distance between each member and the cen-
ter in a cluster, in the test phase. The detail is described in where Uk is comprised of the first k columns of the matrix U and
Section 4.3. VkT is the first k rows of matrix V T , S k = diag(r1 , . . . , rk ) is the
first k factors, the matrix Ak captures most of the important under-
4. Methodology lying structure in the association of terms and documents while
This section describes our methodology for predicting student ignoring noise due to word choice [30].
performance from free-style comments. When LSA is applied to a new comment, a query, a set of
words (like the new comment), is represented as a vector in a k-
4.1 Term Weighting of Comments dimensional space. The new comment query can be represented
In preparing for LSA, the C-comments are transformed into a by
standard word-by-comment matrix [6] by extracting words from
q = qT Uk S k−1 (4)
them. This word-by-comment matrix say A, is comprised of m
words w1 , w2 . . . , wm in n comments c1 , c2 , . . . , cn , where the where q and q are simply the vector of words in a new comment
value of each cell ai j indicates the total occurrence frequency of multiplied by the appropriate word weights and the k-dimensional
word wi in comment c j . vector transformed from q, respectively. The sum of these k di-
To balance the effect of word frequencies in all the comments, mensional word vectors is reflected in the term qT Uk in the above
a log entropy term weighting method is applied to the original equation. The right multiplication by S k−1 differentially weights
word-by-comment matrix, which is the basis for all subsequent the separate dimensions. Thus the query vector is located at the
analyses [17]. We apply a global weighting function to each weighted sum of its constituent term vectors [6].
nonzero element of ai j of A. The global weighting function trans- 4.2.2 Feature Selection and Semantic Feature Space
forms each cell ai j of A to a global term weight gi of wi for the Choosing the number of dimensions k for matrix A is an in-
entire collection of comments. teresting problem. While a reduction in k can remove much of
Here gi is calculated as follows: the noise, keeping to few dimensions or factors may lose impor-
n tant information [7]. In our study, we propose a method which is
gi = 1 + (pi j log(pi j ))/ log(n) (1) based on analyzing the first four dimensions of U, S and V from
j=1
comments data. We evaluated the first four columns of U results
where pi j = Lij /g fi , Lij = log(t fij + 1), t fij is the number of and confirmed they showed the relation between the meaning of
occurrences of wi in c j , g fi is the number of occurrences of word each column and the higher weight words. Therefore, we can
wi in all comments, and n is the number of all comments. predict student performance with more accuracy by employing
K-means clustering method [31]. Tables 3 and 4 show the mean-
4.2 Latent Semantic Analysis ing and the higher weight words of the first four columns after
Latent semantic analysis (LSA) is a computational technique analyzing U results by taking lesson 7 as an example. Words in
that contains a mathematical representation of language. During the first column include the subject of lesson 7 entitled by “An
c 2015 Information Processing Society of Japan 195
Journal of Information Processing Vol.23 No.2 192–201 (Mar. 2015)
c 2015 Information Processing Society of Japan 196
Journal of Information Processing Vol.23 No.2 192–201 (Mar. 2015)
Table 5 5-grades.
Grade S A B C D
Mark 90–100 80–89 70–79 60–69 0–59
#Student 21 41 23 17 21
Table 6 9-grades.
Grade S AS AB BA BC CB CD DC D
Mark 90–100 85–89 80–84 75–79 70–74 65–69 60–64 55–59 0–54
#Student 21 35 6 10 13 9 8 2 19
5. Experimental Results
5.1 Measures of Student Grade Prediction
In our experiment, we evaluated the prediction results by 10-
fold cross validation and run evaluation experiments according to
4 values: TP (True Positive), TN (True Negative), FP (False Pos-
itive) and FN (False Negative) and calculated Precision, Recall,
F-measure and Accuracy in each lesson as follows: Fig. 3 Results of classifying students.
Let G be 5-grade categories (S, A, B, C and D), and X be a
higher, we may more misdetect them. So our study showed the
subset of G; let obs (si , X) be a function that returns 1 if the grade
prediction results by calculating Precision, Recall, Accuracy, F-
of student si is included in X, 0 otherwise, where 1 ≤ i ≤ n, and n
measure and standard deviation to the 3 methods: the basic pre-
is the number of students; pred(si ) be a function that returns a set
diction method, the overlap method and the similarity measuring
of grade categories only including a predicted grade for student
method.
si ; !pred (si ) returns a complement of pred(si ).
c 2015 Information Processing Society of Japan 197
Journal of Information Processing Vol.23 No.2 192–201 (Mar. 2015)
Fig. 6 The results of accuracy and F-measure in each grade after detecting
noisy data.
lessons 7 and 12, and the lowest ones from the bottom in lessons
8 and 14.
In addition, Fig. 6 focuses on the prediction results in each
grade after detecting noisy data. We can see that grade A took
the highest results of prediction accuracy and F-measure, and the
grade D was the lowest.
c 2015 Information Processing Society of Japan 198
Journal of Information Processing Vol.23 No.2 192–201 (Mar. 2015)
Fig. 7 The prediction accuracy results with and without similarity measur- Fig. 9 The prediction accuracy by grade after applying the similarity mea-
ing method. suring method.
Table 8 S d of prediction accuracy for the proposed methods. and between the S d and the similarity measuring method. On the
Lesson 7 8 9 10 11 12 13 14 15 other hand, the correlation coefficient between the S d and that of
Basic Prediction Method 2.76 9.19 3.73 6.02 7.89 5.19 4.57 8.07 5.63 the overlap method only shows a weak correlation. We believe
Overlap Method 2.99 6.41 3.44 5.11 6.20 4.45 6.0 5.12 4.24 that this shows a fact that the overlap method gives stable evalua-
Similarity Measuring Method 2.16 5.95 4.05 4.76 7.22 4.03 4.56 6.35 4.03
tion of student grades.
Table 9 Correlation coefficient of S d and accuracy.
5.7 Prediction Accuracy Differences between 5 Grades
Basic Prediction Method −0.89485
Finally, Fig. 9 displays the relationship between C-comments
Overlap Method −0.5152
Similarity Measuring Method −0.91978 data from lessons 7 to 15 and the prediction accuracy in each
grade after applying the similarity measuring method and detect-
ing noisy data. As shown in Fig. 9, the average prediction accu-
racy results were between 0.65 and 0.93. We can clearly distin-
guish higher grade group: S, A, and B from lower one: C and
D. One of the reasons why prediction accuracy of grades C and
D became lower came from the smaller number of comments in
those grades.
c 2015 Information Processing Society of Japan 199
Journal of Information Processing Vol.23 No.2 192–201 (Mar. 2015)
diction method), the prediction accuracy became higher. Continuous Inquiry and Improvement, Southwest Educational Devel-
opment Laboratory (1997).
To sum up, there are still quite a few considerations that would
[16] Hung, J.: Trends of E-learning Research from 2000 to 2008: Use of
surely add even more value to the results obtained. It is nec- Text Mining and Bibliometrics, British Journal of Educational Tech-
essary to improve the prediction results of student performance nology, Vol.43, No.1, pp.5–15 (2012).
[17] Landauer, T.K., Foltz, P.W. and Laham, D.: An Introduction to Latent
from their comments; other machine learning techniques, such as, Semantic Analysis, Discourse Processes, Vol.25, pp.259–284 (1998).
neural network and support vector machine, will be candidate for [18] León, J., Olmos, R., Escudero, I., Jorge-Botana, G. and Perry, D.: Ex-
ploring the Assessment of Summaries: Using Latent Semantic Analy-
the improvement to compare with the present method (K-means sis to Grade Summaries Written by Spanish Students, Procedia - So-
Clustering). cial and Behavioral Sciences, Vol.83, pp.151–155 (2013).
[19] Li, S. and Ho, H.: Predicting Financial Activity with Evolutionary
Another interesting issue is expanding the problem to improve Fuzzy Case Bbased Reasoning, Expert Systems with Applications,
student performance by providing advice to them in different Vol.36, No.1, pp.411–422 (2009).
[20] Lin, F.-R., Hsieh, L.-S. and Chuang, F.-T.: Discovering genres of on-
classes according to the estimated performance of each student. line discussion threads via text mining, Computers and Education,
Measuring motivation after each lesson can help for giving feed- Vol.52, pp.481–495 (2009).
[21] Liu, B., Cao, S.G. and He, W.: Distributed Data Mining for E-
back to students and encourage them to improve writing skills; Business, Information Technology and Management, Vol.12, No.1,
they can describe attitudes and situations, understanding of sub- pp.1–13 (2011).
jects, difficulties to learn, and learning activities in the classroom. [22] Mansur, M.O. and Md. Sap, M.N.: Outlier Detection Technique in
Data Mining: A Research Perspective, Postgraduate Annual Research
This will help a teacher to give advice and improve their perfor- Seminar (2005).
mance. [23] Minami, T. and Ohura, Y.: Lecture Data Analysis towards to Know
How the Students’ Attitudes Affect to their Evaluations, International
Acknowledgments This work was supported in part by Conference on Information Technology and Applications (ICITA),
PEARL, enPiT of Project for Establishing a Nationwide Practi- pp.164–169 (2013).
[24] Parack, S., Zahid, Z. and Merchant, F.: Application of data mining
cal Education Network for IT Human Resources Development in educational databases for predicting academic trends and patterns,
under the MEXT, Japan, and JSPS KAKENHI Grant Number IEEE International Conference on Technology Enhanced Education
(ICTEE) (2012).
24500176, 25350311, 26350357 and 26540183. [25] Pellegrino, W., Chudowsky, N. and Glaser, R.: Knowing What Stu-
dents Know: The Science and Design of Educational Assessment,
Washington. DC: National Academy Press (2001).
References [26] Romero, C., Espejo, P.G., Zfra, A., Romero, J.R. and Ventura, S.: Web
[1] Abdousa, M.: A predictive Study of Learner Satisfaction and Out- Usage Mining for Predicting Final Marks of Students That Use Moo-
comes in Face-to-Face, Satellite Broadcast, and Live Video-Streaming dle Courses, Computer Applications in Engineering Education (2010).
Learning Environments, The Internet and Higher Education, Vol.13, [27] Shepard, L.A.: The Role of Classroom Assessment in Teaching and
No.4, pp.248–257 (2010). Learning, (CSE Technical Report 517), Los Angeles CA: Center for
[2] Adhatrao, K., Gaykar, A., Dhawan, A., Jha, R. and Honrao, V.: Pre- the Study of Evaluation (2000).
dicting Students’ Using ID3 and C4.5 Classification Algorithms, In- [28] Tane, J., Schmitz, C. and Stumme, G.: Semantic Resource Manage-
ternational Journal of Data Mining and Knowledge Management Pro- ment for the Web: An e-learning Application, WWW Conference, New
cess (IJDKP), Vol.3, No.5, pp.39–52 (2013). York, USA, pp.1–10 (2004).
[3] Ananiadou, S.: National Centre for Text Mining: Introduction to Tools [29] Ur-Rahman, N. and Harding, J.: Textual Data Mining for Industrial
for Researchers, (online), available from http://www.jisc.ac.uk/ Knowledge Management and Text Classification: A business Oriented
publications/publications/ (2008). Approach, Expert Systems with Applications, Vol.39, No.5, pp.4729–
[4] Antai, R., Fox, C. and Kruschwitz, U.: The Use of Latent Semantic 4739 (2012).
Indexing to Cluster Documents into Their Subject Areas., Fifth Lan- [30] Yu, B., ben Xu, Z. and hua Li, C.: Latent Semantic Analysis for
guage Technology Conference. Springer. ASME Design Engineering Text Categorization Using Neural Network, Knowledge-Based Sys-
Technical conferences, DETC2001/DTM (2011). tems, Vol.21, pp.900–904 (2008).
[5] Baker, R.S.J.D.: Data Mining for Education, International Encyclope- [31] Zaiane, O.R.: Introduction to Data Mining, chapter I, CMPUT690
dia of Education (3rd edition). Oxford, UK: Elsevier (2009). Principles of Knowledge Discovery in Databases (1999).
[6] Berry, M.W., Dumais, S.T. and O’Brien, G.W.: Using Linear Alge-
bra for Intelligent Information Retrieval, SIAM Review, Vol.37, No.4,
pp.573–595 (1995).
[7] Deerwester, S., Dumais, S.T., Furnas, G.W., Landauer, T.K. and
Harshman, R.: Indexing by Latent Semantic Analysis, Journal of the
American Society for Information Science, Vol.41, No.6, pp.391–407 Shaymaa E. Sorour received her B.S.
(1990). degree in Education Technology, at Fac-
[8] Delavari, N., Phon-Hon-Amnuaisuk, S. and Beikzadeh, M.R.: Data
Mining Application in Higher Learning Institutions, Informatics in
ulty of Specific Education, Kafr Elsheikh
Education, Vol.7, No.1, pp.31–54 (2008). University, Egypt and a M.S. degree in
[9] Dhillon, I.S. and Modha, D.S.: Concept Decompositions for Large Computer Education at Department of
Sparse Text Data using Clustering, Machine Learning, Springer,
Vol.42, No.1-2, pp.143–175 (2001). Computer Instructor Preparation, Faculty
[10] Earl, L.: Assessment of Learning, for Learning, and as Learning, of Specific Education, Mansoura Univer-
chapter 3, Thousand Oaks, CA, Corwin Press (2003).
[11] Fuller, C.M., Biros, D.P. and Delen, D.: An Investigation of Data sity, Egypt, in 2004 and 2010 respectively.
and Text Mining Methods for Real World Deception Detection, Ex- Since 2005 until now, she is working as an assistant lecture at
pert Systems with Applications, Vol.38, No.7, pp.8392–8398 (2011).
[12] Goda, K., Hirokawa, S. and Mine, T.: Correlation of Grade Predic- Department of Education Technology, Faculty of Specific Educa-
tion Performance and Validity of-Self-Evaluation Comments, SIGITE tion, Kafr Elsheikh University, Egypt. She is currently a Ph.D.
2013, The 14th Annual Conference in Information Technology Educa-
tion, pp.35–42 (2013). student in Graduate School of Information Science and Electrical
[13] Goda, K. and Mine, T.: Analysis of Students’ Learning Activities Engineering, Department of Advanced Information Technology,
through Quantifying Time-Series Comments, KES 2011, Part II, LNAI
6882, Springer-Verlag Berlin Heidelberg, pp.154–164 (2011). Kyushu University, Japan. She is a member of IEEE and IPSJ.
[14] Gorissen, P., van Bruggen, J. and Jochems, W.: Usage Reporting
on Recorded Lectures, International Journal of Learning Technology,
Vol.7, No.1, pp.23–40 (2012).
[15] Hord, S.M.: Professional Learning Communities: Communities of
c 2015 Information Processing Society of Japan 200
Journal of Information Processing Vol.23 No.2 192–201 (Mar. 2015)
c 2015 Information Processing Society of Japan 201