APJEE - Learning Engagement
APJEE - Learning Engagement
APJEE - Learning Engagement
Research Article:
Bricks to Clicks: Students’ Engagement in E-Learning during the
COVID-19 Pandemic
Donnie Adams1*, Kee Man Chuah2, Ahmed Mohamed3 and Bambang Sumintono1
1 Faculty of Education, Universiti Malaya, Jalan Universiti, 50603 Kuala Lumpur, Malaysia
2
Faculty of Language and Communication, Universiti Malaysia Sarawak, 94300 Kota Samarahan,
Sarawak, Malaysia
3
Faculty of Education, The Maldives National University, Male’, Maldives
Abstract
The COVID-19 pandemic has forced higher education institutions to execute mitigation efforts
such as an e-learning mode of instruction to reduce the impact on pedagogical activities. These
challenges have raised concerns on students’ engagement in e-learning as students are at risk
of falling behind in education. However, there is little empirical research focusing on students’
engagement in e-learning experiences. The purpose of this study is to investigate students’
engagement in e-learning as well as specifically assess their engagement based on gender,
age, ethnicity, level of education and field of study in a Malaysian public higher education
institution. This study used a non-experimental quantitative research design. Data were collected
from a sample of 281 students using the Blended Learning Engagement Questionnaire©.
WINSTEPS Rasch model measurement software was used to determine the reliability and
validity of the research instrument. Subsequently, Differential item functioning (DIF) was
conducted to assess students’ engagement in e-learning as well as specifically assessing student’s
demographic factors such as gender, age, ethnicity, level of education and field of study. Findings
indicate students have high levels of behavioural engagement as compared with cognitive and
emotional engagement in e-learning. Further analysis indicated there were differences in
students’ engagement based on demographic factors such as gender, age, and field of study.
This study provides insight into students’ engagement in e-learning that will help lecturers to
reflect on their own teaching practices. Implications and recommendations for future research
are presented.
To cite this article: Adams, D., Chuah, K. M., Mohamed, A., Sumintono, B., Moosa, V., & Shareefa, M. (2021). Bricks to
clicks: Students’ engagement in e-learning during the COVID-19 pandemic. Asia Pacific Journal of Educators and Education,
36(2), 99–117. https://doi.org/10.21315/apjee2021.36.2.6
© Penerbit Universiti Sains Malaysia, 2021. This work is licensed under the terms of the Creative Commons Attribution
(CC BY) (http://creativecommons.org/licenses/by/4.0/).
Donnie Adams et al.
INTRODUCTION
The COVID-19 pandemic has forced higher education institutions to pivot their instruction
online to allow teaching and learning to continue (Adams et al., 2021). Mitigation efforts
such as flexible online learning and e-assessment strategies were executed to reduce the
impact of the pandemic on pedagogical activities (Rapanta et al., 2020). Students are
required to study at home as an effort to control the pandemic (Adams & Dewitt, 2021).
Additionally, many lecturers need to teach online for the first time. Lecturers have to
carefully craft their instructional design to enable their students to fully digest and absorb
the learning.
LITERATURE REVIEW
The adoption of e-learning in higher education institutions across the globe has recorded
a remarkable growth over the last decade. This surge is contributed by the advancement
of learning technologies and the shift towards 21st century learning (Miller, 2018).
Prior to this, the use of e-learning is limited to institutions that offer distance learning
programmes (Osman et al., 2018). As technological devices and tools become more
100
Students’ Engagement in E-Learning
Over the last decade, researchers have shown great interest in examining e-learning
adoption from several perspectives. One area that was widely studies was related to faculty
and students’ readiness in accepting such mode of learning (Cheng et al., 2019; Graham,
2013; Matheos & Cleveland-Innes, 2018). While these studies were investigating
different higher education institutions, the results were consistent and indicative on the
inevitable need to adopt e-learning since more resources are now readily available online.
E-learning has been regarded as feasible option in extending learning beyond the class
hours (Yang, 2020). During the campus closures due to the COVID-19 pandemic,
studies on students’ readiness for e-learning or its variants were also widespread as more
institutions were putting emphasis on e-learning as a measure to mitigate the impacts of
the pandemic (Tang et al., 2021).
Another key focus on e-learning in higher education as identified in the previous studies
is on addressing the challenges to implement it effectively (Heilporn et al., 2021; Rasheed
et al., 2020; Roman & Plopeanu, 2021). The outcomes from e-learning adoption as
reported in Malaysian literature are mainly positive despite some commonly reported
challenges such as infrastructure and institutional support. Jie and Fernandez (2021) found
maintaining a stable internet connection in modest cities remains a continuous challenge.
Sukumaran et al. (2021) found there is lack of funding for technological innovations
and technical skills support. Almaiah et al. (2020), on the other hand, discovered several
factors not related to technology that could challenge e-learning implementation, namely
cultural, trust and self-efficacy factors. These factors essentially are linked to students’
engagement that affected perceived learning effectiveness as well as learners’ satisfaction.
101
Donnie Adams et al.
Lane et al. (2021) investigated students’ experiences of e-learning and their overall
satisfaction in the format. Their findings indicated how emotional engagement is an
applicable predictor of student satisfaction and success in e-learning courses. Vavasseur
et al. (2020) studied 353 medical students’ engagement and satisfaction on the online
learning contents. They discovered that online learning approach increased students’
performance and satisfaction. However, they measured engagement according to audience
retention of videos and attendance on a content management system, which may not be
sufficient to measure students’ cognitive and social engagement. Hence, there is a need to
measure students’ levels of engagement more holistically.
Al‐Qahtani & Higgins (2013) investigated the effect of three modes (traditional
classroom learning, e-learning and blended learning) on students’ achievement in higher
education. The study, which involved 148 students, indicated the positive effect of online
learning on students’ achievement as compared to the other two modes. The flexibility
of online learning in allowing students to be engaged with different activities is seen as a
contributing factor of students’ higher achievement. The same outcome was also reported
by Bazelais and Doleck (2018) in which students’ engagement in e-learning activities led
to higher performance and acquisition of more skills among students of a first-semester
physics course.
The study by Adams, Tan, Sumintono and Oh (2020) revealed high levels of cognitive,
emotional and behavioural engagements in e-learning activities among students of
higher education institutions in Malaysia. Students from engineering and medicine were
more highly engaged in online learning activities than their peers in social sciences and
natural sciences. Ożadowicz (2020) also found the same pattern among engineering
students. These two studies showed that despite the technicality and difficulty of certain
fields of study, online learning can still contribute to higher academic performances if
the design and implementation are done systematically. Tang et al. (2021) also studied
students’ readiness in adopting real-time or live online classes during the COVID-19
outbreak and found out that students were mainly complying to the requirements
to attend online classes and were not engaged when those classes were conducted.
They also found out that postgraduate students have higher readiness as compared to
undergraduates, prompting the need to devise measures in motivating them.
The review of previous studies shows that despite the reported benefits of e-learning
practices in increasing student satisfaction (Lane et al., 2021; Vavasseur et al., 2020)
and better academic performance (Al‐Qahtani & Higgins, 2013; Bazelais & Doleck,
2018), there is still a gap in understanding students’ levels of engagement in an
e-learning environment by scrutinizing its relationship with demographic factors such
as gender, ethnicity, locality and field of study.
102
Students’ Engagement in E-Learning
METHODOLOGY
Research Participants
A quantitative approach with survey design was employed in this study. The data were
collected using cross-sectional method utilising electronic format, namely google form
platform, from one public university in Malaysia. The respondents for this study were
students, where they invited by emails and social media to participate, it was stated in
first page that it was strictly anonymous and voluntary to address ethical concerns.
Thus, by completing the questionnaire, the students are given their consent.
The minimum sample size was 150 respondents, this is estimated based on Boone
et al. (2014) calculation of precise item calibration at 0.5 logit and 99% confidence
level. There were 400 students participated in the study. The next step is conducting
data cleaning and validation using WINSTEPS version 3.73, a Rasch measurement
model software, for detecting outlier (29 respondents who answered all minimum or
maximum value) and misfit response (90 respondents with Outfit MNSQ index bigger
than 2 or less than 0.4) (Andrich & Marais, 2019; Bond & Fox, 2015; Widhiarso &
Sumintono, 2016). As a result, 281 respondents were analysed further in this study, their
demographic profile is presented in Table 1.
103
Donnie Adams et al.
Table 1. (Continued)
Demographics Respondent Percentage (%)
Field of study
Social sciences 205 73
Natural Sciences 54 19
Engineering 18 7
Medicine 4 1
Instrumentation
Measurement Model
A Rasch rating scale model approach was used to assess to the data. This is because
students’ engagement in an online learning involves latent trait which refer to student
opinion, perception, and attitude in the activity which need precise and accurate
measurement model (Andrich & Marais, 2019; Boone et al., 2020). The study using
WINSTEPS software, where the raw data from the respondents (likert rating or ordinal
type) were mathematically transformed into logit (logarithm odd unit) via the logarithm
function which become an equal interval scale (Linacre, 2013). There were two type of
logits data produced from the software, which are item logit used to inform about
instrument quality and item calibration, and person logit which informs about respondents’
level of engagement (Boone et al., 2014; Bond & Fox, 2015).
To determine whether the instrument has a good quality, the validity and reliability aspect
in shown in Table 2. The table below shows that data gathered is fit to the model as
Outfit mean square value is close to one (an ideal value) both for person and item, it
also confirmed with significant value of Chi-square test. Reliability indices for person
(0.93), item (0.96) and alpha (0.94) indicates that the consistency of person and item
responses was ‘very good’ (Sumintono & Widhiarso, 2014). Separation indices, which
inform about grouping respondents (3.71) and how widespread item difficulty level (5.10)
showing the indices more than three (minimum acceptable value), supporting the fact
that very reliable instrument and data were collected.
104
Students’ Engagement in E-Learning
0.6
0.5
0.4
0.3
0.2
0.1
0.0
−7 −6 −5 −4 −3 −2 −1 0 1 2 3 4 5 6 7 8 9
Measure relative to item difficulty
Category Probability: 1 Category Probability: 2 Category Probability: 3 Category Probability: 4
105
Donnie Adams et al.
FINDINGS
The first result of Rasch model analysis in this study is estimation of item location
(calibrations) or logit value of item (LVI). The bigger LVI means the higher order of the
severity or difficulty level of the items on the scale, in other words the item tends to be
not easy to agree by respondents. The item logit mean is fixed at 0.0 logit, and the value
of item standard deviation (SD) in this study was 0.71 logit. To put into perspectives,
mean and SD were used to grouping the item based on difficulty level (Table 3).
There were 3 items (18%) in the category of very difficult to agree by respondents
(LVI > 0.71 logit); in the second category, which is difficult to agree (+0.71 > LVI > 0.00)
there were 4 items (25%); the next category which is easy to agree by respondents
(0.00 > LVI > −0.71) there were 6 items (39%); and lastly 3 items (18%) fall into category
very easy to agree by the respondents (LVI < −0.71 logit).
As shown in Table 3, three dimensions in the instrument were showing three different
response patterns. For the cognitive engagement, equally two items were categorized
both difficult and easy to agree respectively. Whereas for emotional engagement more
items (four out of five) were in the difficult to engage group by students. Interestingly,
behaviour engagement in e-learning tend to be easy to be conducted by the students,
where 6 out of 7 items fall into category easy and very easy to agree. These findings
indicate that student’s perception in e-learning behaviour, they do not have much
difficulty to do that compare to cognitive and emotional engagement process which
involve thinking and psychologically attached.
The complete item and person location in a logit measurement continuum is shown
in Figure 2. The Wright map or item hierarchy map demonstrates how the items and
participants fit together along on a logit continuum, the higher LVI meaning the item
difficult to endorse/agree by the respondents. At the right side of the map, item C3
(I prepare myself in reading materials online before attending classes with LVI +1.02 logit)
is the most difficult item to agree. This means that preparation before online lecture
seldom conducted by students. Meanwhile item B4 (I do my assignments and submit it
on time online with LVI −1.29 logit) located in the lower right of the map, informing
that regard to assignment student don’t have difficulty when change to online mode.
106
Students’ Engagement in E-Learning
This finding from the two different LVI items conclude student’s different situation in
before and after the online lecture.
107
Donnie Adams et al.
On the left side of Figure 2 is the spread of person level engagement. The higher logit
value person (LVP), which is located top left, indicating that the person answers to items
tend to state agree or strongly agree, showing a very high level of engagement in online
learning. The mean of LVP was +2.47 logit with standard deviation 2.34, showing that
the participants tended to be located higher than the items. This indicates that the level
of engagement of this sample was higher than the difficulty level as reflected in the items,
this also means that that the test-item targeting is satisfactory.
Like grouping of item difficulty level above, because Rasch model provide accurate and
precise measurement on the level of engagement in e-learning, categorising students’
responses is possible (Table 4). Using LVP of mean and SD, there are four levels of
e-learning engagement (from very high level to low level of engagement), which identify
number of students in each group. This analysis useful, because can go to individual
centred statistics rather than group centred statistics, which can provide many details
(Engelhard & Wind, 2018).
Table 4. Students online learning engagement level according to demographics (N = 281)
108
Students’ Engagement in E-Learning
Analysis on students’ gender found around 55 out of 89 male students (62%) and 86 out
of 192 female students (45%) were in very high- and high-level engagement of e-learning.
In the moderate level engagement, there were 21 male students (24%) and 78 female
students (40%); students who were low level of online learning engagement is the least
percentage (15% each respectively for male and female students). This show that proportion
of male student with higher level of engagement was higher from this study’s sample.
In terms of age, the proportion of students belong to very high- and high-level engagement,
nearly similar with those who consider themselves in moderate or low-level engagement
in every age group. Also, students who perceived themselves had low levels of e-learning
engagement are minority in all age groups, which were 22% for below 20 years old, 13% in
20–29 years, and 6% in more than 39 years age group, except for 30–39 years of age (29%).
In terms of students’ ethnicity, the similar result appears like in age group, except for
Chinese students, others belong to very high- and high-level online engagement. The
findings shows that the different also not big. Interestingly, except for Chinese students,
in other groups of student ethnicities, low level engagement also the minority, which were
range from 20% (International students) to 9% (Indian students). Analysis on student’s
level of education and their field of study found that less than in moderate and low-level
online learning engagement. Also, not so much different in term of students’ number in
each group were showing that nearly similar response in this sample.
Further analysis was conducted to examine the differences in test item responses based
on demographic profile which was Differential Item Functioning (DIF). The DIF
analysis is sensitive to number of respondents, where the group less than 30 person is
not suggested (Bond & Fox, 2015; Boone et al., 2020). So for this analysis there were
two demographic variables, namely gender and level of study, can be analysed with DIF.
An item was considered as having DIF if it had DIF contrast value of less than −0.5 or
more than 0.5, the t value of less than −2.0 or more than 2.0, and the p (Probability)
value of less than 0.05 or greater than −0.05 (Boone et al., 2014). Therefore, three items
for gender and four items for level of study respectively were suspected has DIF as
shown in Table 5. However, only one item, E1 (I participate actively in online activities)
which was fulfilling in all three criteria of DIF, the rest only one or two criteria.
In addition, looking at DIF plot for the whole items provide very interesting findings.
If the item location in the line close to the top, meaning that the item considers difficult
to agree; whereas if located below it is perceived easy to agree. Figure 3 which show
DIF plot based on gender, that each dimension has its unique pattern of response.
For cognitive dimension, item C1 (I’m able to do my best to complete online tasks) was no
different between male and female respondent; for item C2 and C3 it is considered difficult
to agree by female than male counterparts. However, for item C4 (I spend enough time and
effort to learn online), showing that male students agree lesser than female.
109
Donnie Adams et al.
For the emotional dimension, except for E1 (which is detected as item DIF), the other
items (E2 to E5) female respondents tend to response the item as difficult to agree
than male. For instance, in the item E3 (I feel my classmates respect my thoughts and views
during online discussion) and E5 (I feel my lecturers show interest to my views during online
discussion) male students found the item easy to endorse, showing that female student’s
has different thoughts towards online learning. For behavior dimension, two items
consider difficult for male students, B1 (I participate actively in online activities) and B2
(I behave fairly to all my classmates when performing task online), while the rest (B3 to
B7) male students found it was easy to agree than female students. In this behaviour
dimension shows that male students tend to be more active in online learning in nearly
all items.
110
Students’ Engagement in E-Learning
In the Figure 4 which shows a DIF plot based on level of study, which has two groups,
undergraduate (U) and postgraduate (P) students, the pattern of response a bit different.
Only certain items have distinctive response from the two groups, the rest tend to be similar
(such as C1, C2, C4, E3, B1, B2 and B5). For item C3 (I prepare myself in reading materials
online before attending classes) postgraduate student are well prepare than undergraduate,
showing the tendency and maturity of self-learning. The same response regard to item
E2 (I give importance to studying together with my classmates in a group online) showing
online learning more prefarafble for postgraduate students.
The real different can sense at the behaviour dimension, where identity as digital native
is salient for undergraduate. For instance, in the item B4 (I do my assignments and submit
it on time online), B6 (I try to do my best in online group work) and B7 (I share information
with my classmates during online discussion), younger generation like undergraduate students
don’t have any difficulty regard to technical challenges in online learning compare to
postgraduate students. The distance of behaviour engagement level actually showing
generation gap.
Figure 4. Person DIF plot based on level of study to all items
111
Donnie Adams et al.
DISCUSSION
Based on four levels of e-learning engagement identified in this study, the findings
disclosed male students were more engaged in e-learning activities than female students.
This study supports evidence from previous findings (e.g., Adams, Tan, Sumintono,
& Oh, 2020; Martin & Bolliger, 2018). DIF analysis further depicts male students
112
Students’ Engagement in E-Learning
participates more actively in online activities than female students. However, despite
male students outperforming their female counterparts in cognitive and emotional
components of engagement, female students achieved higher levels of engagement in
the behavioural components. This finding could be due to the higher level of excitement
and motivation that male students have in online activities than female students (Win &
Wynn, 2015). Previous findings also reported that male students preferred online learning
as compared to female students who preferred traditional face-to-face learning (Yu, 2021).
These findings may help instructors to carefully consider cognitive and emotional
engagement aspects when teaching female students in an e-learning setting.
In terms of students’ age, there was no significant difference across age groups in their
e-learning engagement. This finding is contrary to a previous study that suggests
students aged above 29 years old are more engaged in reading materials online and are
more prepared for online classes than younger students, whereas students aged below 20
prefer completing and submitting assignments on time, and study in groups (see Adams,
Tan, Sumintono, & Oh, 2020). In the literature, students’ age appears as a significant
factor that influences the level of both behavioural and cognitive engagement (Li & Lam,
2015). Even though no significant difference was identified within the age groups used
in this present study, postgraduate students remained well prepared and prefer e-learning
than undergraduate students. This result is in line with previous studies (Adams et al.,
2018; Li & Lam, 2015). Postgraduate students are found to be more cognitively engaged
(Li & Lam, 2015). A possible explanation for this result might be that postgraduate
students are more independent and matured learners who self-regulate their learning.
Most undergraduate students depend heavily on their lecturers; expecting lecturers to
send regular announcements and reminders to them as compared with postgraduate
students (Martin & Bolliger, 2018). According to DIF analysis, undergraduate students
found it difficult in preparing themselves before attending online classes. Despite this
result, undergraduate students preferred to complete assignments and submit online,
and try to engage themselves more in online group work as compared with postgraduate
students. These differences might indicate a generation gap in terms of digital nativity.
Undergraduate students in this study grew up in the cyber age, thus they are “native
speakers” of the digital language, and able to easily adapt to online classes (Wang et al.,
2019).
Overall findings of this study revealed students have high levels of behavioural engagement
as compared with cognitive and emotional engagement in e-learning. The findings also
disclosed male students were more engaged in e-learning activities than female students.
However, there was no significant difference across age groups in their e-learning
engagement. In terms of students’ level of education undergraduate students found it
difficult in preparing themselves before attending online classes. Thus, e-learning lessons
needs to be developed according to students’ competency, interest and skills (Adams, Tan,
& Sumintono, 2020). We might be turning students into passive learners and making
113
Donnie Adams et al.
them feel disconnected from their lessons if we fail to first identify their competence and
needs (Adams et al., 2018; Adams, Tan, & Sumintono, 2020; Vavasseur et al., 2020).
Lecturers must design their lessons that promotes student engagement in learning.
They must guide their students towards a more self-directed style of learning, away
from their usual preferred learning styles. Students should be guided in diagnosing their
learning needs, formulating their own learning goals, and sourcing their own resources or
materials for learning.
This study has several limitations. Firstly, the study was conducted only in a public in
Malaysia. Thus, generalising the finding is not possible as one group of students in a
public university cannot represent all students in other universities. Therefore, future
studies could be conducted to other universities, both public and private, while factoring
in the lecturers’ perspectives as well. Secondly, the addition of qualitative data such as
interviews and focus group discussions will further explore and explain the quantitative
findings. Furthermore, investigating what e-learning strategies that are most effective
in engaging students would add significantly to the current body of knowledge.
REFERENCES
Adams, D., & Dewitt, D. (2021). Innovative practices of technology-enhanced learning. Universiti
Pendidikan Sultan Idris.
Adams, D., Chuah, K. M., Sumintono, B., & Mohamed, A. (2021). Students’ readiness for
e-learning during the COVID-19 pandemic in a South-East Asian university: A Rasch
analysis. Asian Education and Development Studies. https://doi.org/10.1108/AEDS-05
-2020-0100
Adams, D., Sumintono, B., Mohamed, A., & Mohamad Noor, N. S. (2018). E-learning
readiness among students of diverse backgrounds in a leading Malaysian higher education
institution. Malaysian Journal of Learning and Instruction, 15(2), 227–256.
Adams, D., Tan, M. H. J., & Sumintono, B. (2020). Students’ readiness for blended learning in
a leading Malaysian private higher education institution. Interactive Technology and Smart
Education. https://doi.org/10.1108/ITSE-03-2020-0032
Adams, D., Tan, M. H. J., Sumintono, B., & Oh, S. P. (2020). Blended learning engagement
in higher education institutions: A differential item functioning analysis of students’
backgrounds. Malaysian Journal of Learning and Instruction, 17(1), 133–158. https://doi.org/
10.32890/mjli2020.17.1.6
114
Students’ Engagement in E-Learning
Al‐Qahtani, A. A., & Higgins, S. E. (2013). Effects of traditional, blended and e‐learning on
students’ achievement in higher education. Journal of Computer Assisted Learning, 29(3),
220–234. https://doi.org/10.1111/j.1365-2729.2012.00490.x
Almaiah, M. A., Al-Khasawneh, A., & Althunibat, A. (2020). Exploring the critical challenges and
factors influencing the E-learning system usage during COVID-19 pandemic. Education and
Information Technologies, 25, 5261–5280. https://doi.org/10.1007/s 10639-020-10219-y
Andrich, D., & Marais, I. (2019). A course in Rasch measurement theory, Measuring in the educational,
social and health sciences. Springer.
Bazelais, P., & Doleck, T. (2018). Investigating the impact of blended learning on academic
performance in a first semester college physics course. Journal of Computers in Education,
5(1), 67–94. https://doi.org/10.1007/s40692-018-0099-8
Beth, A. D., Jordan, M. E., Schallert, D. L., Reed, J. H., & Kim, M. (2015). Responsibility and
generativity in online learning communities. Interactive Learning Environments, 23(4), 471–
484. https://doi.org/10.1080/10494820.2013.788035
Bond, T. G., & Fox, C. M. (2015). Applying the Rasch model: Fundamental measurement in the
human sciences (3rd ed.). Routledge.
Boone, W. J., Staver, J. R., & Yale, M. S. (2014). Rasch analysis in the human sciences. Springer.
Boone, W. J., & Staver, J. R. (2020). Advances in Rasch analyses in the human sciences. Springer.
Cheng, L., Ritzhaupt, A. D., & Antonenko, P. (2019). Effects of the flipped classroom
instructional strategy on students’ learning outcomes: A meta-analysis. Educational
Technology Research and Development, 67, 793–824. https://doi.org/10.1007/s11423-018
-9633-7
Chuah, K. M., & Mohamad, F. (2020). Emergency remote teaching scenarios, struggles and
soundboxes: A case study on Malaysian teachers. Interaction Design and Architecture(s) Journal,
46(1), 13–28.
Coman, C., Țîru, L. G., Meseșan-Schmitz, L., Stanciu, C., & Bularca, M. C. (2020). Online
teaching and learning in higher education during the coronavirus pandemic: Students’
perspective. Sustainability, 12(24), 1–22. https://doi.org/10.3390/su122410367
Engelhard, G., & Wind, S. (2018). Invariant measurement with raters and rating scales. Routledge.
Fredricks, J. A., Blumenfeld, P. C., & Paris, A. H. (2004). School engagement: Potential
of the concept, state of the evidence. Review of Educational Research, 74(1), 59–109.
https://doi.org/10.3102/00346543074001059
Fisher, R., Perényi, Á., & Birdthistle, N. (2021). The positive relationship between flipped
and blended learning and student engagement, performance and satisfaction. Active
Learning in Higher Education, 22(2), 97–113. https://doi.org/10.1177/1469787418801702
Fisher, W. P. Jr. (2007). Rating scale instrument quality criteria. Rasch Measurement Transactions,
21(1), 1095.
Furlong, M. J., Whipple, A. D., St. Jean, G., Simental, J., Soliz, A., & Punthuna, S. (2003).
Multiple contexts of school engagement: Moving toward a unifying framework for
educational research and practice. The California School Psychologist, 8(1), 99–113.
https://doi.org/10.1007/BF03340899
Graham, C. R. (2013). Emerging practice and research in blended learning. In M. G. Moore (Ed.),
Handbook of distance education (pp. 333–350). Routledge.
Han, F., Pardo, A., & Ellis, R. A. (2020). Students’ self‐report and observed learning orientations
in blended university course design: How are they related to each other and to academic
performance? Journal of Computer Assisted Learning, 36(6), 969–980. https://doi.org/10.1111/
jcal.12453
115
Donnie Adams et al.
Heilporn, G., Lakhal, S., & Bélisle, M. (2021). An examination of teachers’ strategies to foster
student engagement in blended learning in higher education. International Journal of
Educational Technology in Higher Education, 18(1), 1–25. https://doi.org/10.1186/s41239
-021-00260-3
Hu, M., & Li, H. (2017). Student engagement in online learning: A review. In 2017 International
Symposium on Educational Technology (ISET) (pp. 39–43). IEEE.
Jie, S. S., & Fernandez, D. F. M. F. (2021). Factors influencing student’s behavioural intention
towards the use of e-learning during Covid-19 in Malaysia. Research in Management of
Technology and Business, 2(2), 952–964.
Lane, S., Hoang, J. G., Leighton, J. P., & Rissanen, A. (2021). Engagement and satisfaction:
Mixed-method analysis of blended learning in the sciences. Canadian Journal of Science,
Mathematics and Technology Education, 21(1), 100–122. https://doi.org/10.1007/s42330
-021-00139-5
Li, K. C., & Lam, H. K. (2015). Student engagement in a technology-mediated distance learning
course. International Journal of Services and Standards, 10(4), 172–191. https://doi.org/10
.1504/IJSS.2015.072445
Linacre, J. M. (1999). Investigating rating scale category utility. Journal of Outcome Measurement, 3,
103–122.
Linacre, J. M. (2013). A user’s guide to Winsteps Ministeps Rasch-model computer programs.
http://www.winsteps.com/index.htm
Lizcano, D., Lara, J. A., White, B., & Aljawarneh, S. (2020). Blockchain-based approach to
create a model of trust in open and ubiquitous higher education. Journal of Computing in
Higher Education, 32, 109–134. https://doi.org/10.1007/s12528-019-09209-y
Ma, J., Han, X., Yang, J., & Cheng, J. (2015). Examining the necessary condition for engagement
in an online learning environment based on learning analytics approach: The role of the
instructor. Internet and Higher Education, 24, 26–34. https://doi.org/10.1016/j.iheduc
.2014.09.005
Maatuk, A. M., Elberkawi, E. K., Aljawarneh, S., Rashaideh, H., & Alharbi, H. (2021). The
COVID-19 pandemic and E-learning: Challenges and opportunities from the perspective
of students and instructors. Journal of Computing in Higher Education. https://doi.org/10
.1007/2Fs12528-021-09274-2
Martin, F., & Bolliger, D. U. (2018). Engagement matters: Student perceptions on the importance
of engagement strategies in the online learning environment. Online Learning Journal, 22(1),
205–222. https://doi.org/10.24059/olj.v22i1.1092
Matheos, K., & Cleveland-Innes, M. (2018). Blended learning: Enabling higher education reform.
Revista Eletrônica de Educação, 12(1), 238–244.
Miller, R. (2018). Transforming the future: Anticipation in the 21st century. Taylor & Francis.
Mpungose, C. B. (2020). Emergent transition from face-to-face to online learning in a South
African University in the context of the Coronavirus pandemic. Humanities and Social Sciences
Communications, 7(1), 1–9.
Nguyen, Q., Rienties, B., & Richardson, J. T. E. (2020). Learning analytics to uncover inequality in
behavioural engagement and academic attainment in a distance learning setting. Assessment
and Evaluation in Higher Education, 45(4), 594–606. https://doi.org/10.1080/02602938
.2019.1679088
Osman, Z., Mohamad, W., Mohamad, R. K., Mohamad, L., & Sulaiman, T. F. T. (2018).
Enhancing students’ academic performance in Malaysian online distance learning
institutions. Asia Pacific Journal of Educators and Education, 33, 19–28. https://doi.org/
10.21315/apjee2018.33.2
116
Students’ Engagement in E-Learning
Ożadowicz, A. (2020). Modified blended learning in engineering higher education during the
COVID-19 lockdown—Building automation courses case study. Education Sciences,
10(10), 292.
Rapanta, C., Botturi, L., Goodyear, P., Guardia, L. and Koole, M. (2020), Online university
teaching during and after the Covid-19 crisis: Refocusing teacher presence and learning
activity. Postdigital Science and Education, 2(3), 923–945.
Rasheed, R. A., Kamsin, A., & Abdullah, N. A. (2020). Challenges in the online component
of blended learning: A systematic review. Computers & Education, 144, 103701.
https://doi.org/10.1016/j.compedu.2019.103701
Rasouli, A., Rahbania, Z., & Attaran, M. (2016). Students’ readiness for e-learning application in
higher education. Malaysian Online Journal of Educational Technology, 4(3), 51–64.
Roman, M., & Plopeanu, A. P. (2021). The effectiveness of the emergency eLearning during
COVID-19 pandemic: The case of higher education in economics in Romania.
International Review of Economics Education, 37, 100218. https://doi.org/10.1016/j.iree
.2021.100218
Schunk, D. H., & Mullen, C. A. (2012). Self-efficacy as an engaged learner. In S. L.
Christenson, A. L. Reschly, & C. Wylie (Eds.), Handbook of research on student engagement
(pp. 219–235). Springer. https://doi.org/10.1007/978-1-4614-2018-7
Sukumaran, S., Shahid, N. S. M., Abdullah, N., & Thiagarajah, S. (2021). E-learning of
STEM in Malaysian higher education institutions: Status and challenges. Asian Journal of
University Education, 17(4), 259–271.
Tao, Z., Zhang, B., & Lai, I. K. W. (2018) Perceived online learning environment and students’
learning performance in higher education: Mediating role of student engagement
(Vol. 843, pp. 56–64). Springer Verlag.
Tang, Y. M., Chen, P. C., Law, K. M., Wu, C. H., Lau, Y. Y., Guan, J., He, D., & Ho, G.
T. (2021). Comparative analysis of Student’s live online learning readiness during
the coronavirus (COVID-19) pandemic in the higher education sector. Computers &
Education, 168, 104211. https://doi.org/10.1016/j.compedu.2021.104211
Van Zile-Tamsen, C. (2017). Using Rasch analysis to inform rating scale development. Research
in Higher Education, 58(8), 922–933.
Vavasseur, A., Muscari, F., Meyrignac, O., Nodot, M., Dedouit, F., Revel-Mouroz, P.,
Dercle, L., Rozenblum, L., Wang, L., Maulat, C., Rousseau, H., Otal, P., Dercle,
L., & Mokrane, F. Z. (2020). Blended learning of radiology improves medical
students’ performance, satisfaction, and engagement. Insights into Imaging, 11, 1–12.
https://doi.org/10.1186/s13244-020-00865-8
Wang, H. Y., Sigerson, L., & Cheng, C. (2019). Digital nativity and information technology
addiction: Age cohort versus individual difference approaches. Computers in Human
Behavior, 90, 1–9.
Widhiarso, W., & Sumintono, B. (2016). Examining response aberrance as a cause of outliers in
statistical analysis. Personality and Individual Differences, 98, 11–15.
Win, N. L., & Wynn, S. D. (2015). Introducing blended learning practices in our classrooms.
Journal of Institutional Research Southeast Asia, 13(2), 17–27.
Yang, N. (2020). eLearning for quality teaching in higher education: Teachers’ perception, practice, and
interventions. Springer Nature.
Yu, Z. (2021). The effects of gender, educational level, and personality on online learning outcomes
during the COVID-19 pandemic. International Journal of Educational Technology in
Higher Education, 18(1), 14. https://doi.org/10.1186/s41239-021-00252-3
117