Upload
others
View
2
Download
0
Embed Size (px)
Citation preview
34 http://worldconferences.net/ejournals/glit/
GLIT E-Journal on Information Technology and Language Practice (e-ISSN: 2289-7445) Volume 2, Issue 2
CONDUCTING AND EXAMINING EFFECTIVENESS OF ONLINE
ASSESSMENTS USING THE MODULAR OBJECT-ORIENTED
DYNAMIC LEARNING ENVIRONMENT (MOODLE)
Rosyati Abdul Rashid & Wan Nurhafeza Wan Salam English Language Learning Centre (ELC)
Center for Foundation and Liberal Education
Universiti Malaysia Terengganu (UMT)
ABSTRACT
This paper aims to demonstrate the robustness of Moodle, a type of Learning Management System
(LMS) in assisting language instructors to administer and analyze the quality of test items, and the test
as a whole. Using Moodle Quiz Module, a public speaking test comprising 40 multiple-choice
questions (MCQ) was created, and administered online to 131 students. Descriptive statistics obtained
through the statistical analysis procedures performed by Moodle indicated that the test was reliable,
and was a difficult one for many students since many of the test items were of high levels of Bloom‟s
taxonomy of educational objectives. The test scores obtained by 58 students who also took the Cornell
Critical Thinking Test (CCTT), Level X, were found to be significantly correlated, at a moderate level,
with the scores that the students had in the public speaking test. Data generated by Moodle such as the
data from the analysis of students‟ responses to each test item, and the data provided by items statistics
like item facility (IF), item discrimination (ID), and distractor efficiency (DE) indices, were found to
be very useful in helping the test developers to examine the effectiveness of test items and to later
improve them so that only good items can be kept in the test bank for future use in assessing a
different group of students.
Keywords: Moodle, critical thinking, MCQ, item analysis, IF, DI, and DE indices
1. Introduction
To sustain the ability to compete in the present twenty-first century, Malaysia has to produce
knowledgeable and skilled workforce who are critical in thinking and be able to function well in a
variety of work conditions. Realizing that there is an urgent need to generate the desired Malaysian
citizens, the Malaysian government has recently launched the Malaysia Education Blueprint 2013-
2015, which lists eleven shifts required to transform the Malaysian education system. One of the shifts
is to "provide equal access to quality education of an international standard". Among the listed actions
to be taken to realize this change is the act of revamping examination questions, so as to allow a
greater percentage of questions at higher- order thinking levels to be included in school and all
national examinations by the year 2016. This newly revised education policy initiative is hoped to
assist the nation successfully in generating a high-quality workforce that will help to steer the country
towards greater prosperity. However, to the skeptics, it seems questionable that such reform in school
and national assessments can be implemented successfully within the target time considering the
rather short time frame given to actualize the change and the current somewhat less satisfactory
teachers' literacy assessment level.
35 http://worldconferences.net/ejournals/glit/
GLIT E-Journal on Information Technology and Language Practice (e-ISSN: 2289-7445) Volume 2, Issue 2
2. Literature review
2.1 Educational reforms in Malaysia and teachers’ assessment literacy
Many reform efforts have been taken by the Malaysian government to improve the national education
system (Rajendran, 2001). The first change made to the educational policy was based on the Razak
Report prepared before the country gaining its independence in 1957 and the most recent one is
written in the Malaysia Education Blueprint 2013-2025. One of the outcomes of the reform efforts was
the inclusion of critical thinking or higher-order thinking skills in the objective of the secondary
school curriculum (Curriculum Development Center, 1989, p.2), in the Integrated Curriculum for
Secondary Schools (ICSS), in the National Higher Education Action Plan which stressed that the focus
of the education system was “on generating world-class and holistic human capital who are
intellectually active, creative and innovative, ethically and morally upright, adaptable and capable of
critical thinking" (MOHE, 2007, p.8), and in the latest educational blueprint which states that “the
national curriculum aims to create Malaysian students that are balanced, resilient, inquisitive,
principled, informed, caring, patriotic, as well as effective thinker, communicator, and team player”
(Malaysia Education Blueprint 2013-2025, exhibit.4-2).
Although producing students with critical thinking ability has been considered as one of the
educational objectives for quite sometimes already, the question as to whether the goal is achievable
has frequently been raised. For example, Indramalar in 1997 (as cited in Rajendran, 2001, p.3)
reported the Minister of Education saying that "the education system will be revamped to encourage
rational and analytical thinking." The statement implies that the education system at that particular
time failed to realize the educational objective of producing critical thinkers. One of the measures
taken then to address the issue was to revamp public examination papers. A policy announced in 1994
targeted that by the year 2000, sixty percent of examination papers should comprise questions of the
higher-order thinking type.
The issue seems to have recurred recently when the analysis done by Pearson Education Group of the
UPSR and SPM English papers for the year of 2010 and 2011 revealed that about seventy percent of
the questions in the aforementioned national examination papers assessed "basic skills of knowledge
and comprehension"– the two low levels of cognitive domains in the Bloom's taxonomy (Malaysia
Education Blueprint, 2013-2025). To a certain extent, this finding had influenced the recent decision
of the Ministry of Education to include revamping examination questions as one of the main steps to
be taken in realizing the first type of transformation or shift to be made to the education system in this
21st century.
What went wrong? Perhaps Malaysian teachers are not yet fully prepared to assess students or they
have not adequately been given training on assessment. According to Lim, Wun, and Chew (2014),
Malaysian teachers do not have sufficient knowledge and skills to come up with assessments that can
measure the intended learning outcomes. Lim and his co-researchers made this conclusion based on
the findings of three local studies cited in their article. For example, they cited a recent survey
conducted by Suah (2012) involving 3,866 primary and secondary school teachers which presented
evidence showing that majority of the teachers had less satisfactory literacy assessment level. The
conclusion was further supported by two more studies that they cited. These studies, conducted by
Mohamad (2006) and Salbiah (1995), revealed that teachers' emphasis was more on the content of the
syllabus than the target learning outcomes whenever they were developing and carrying out
assessments.
It seems that much is said about teachers‟ low assessment literacy level and the poor quality of
examination papers, but little has been written about how to improve the standard so that the quality of
the exam papers can be enhanced. Experience and research have shown us that giving training alone
will not solve the problem that we are encountering now. In fact, to date, the Malaysian Ministry of
Education has supported teachers with various kinds of training and workshops including those on
assessments and the teaching of thinking skills (Rajendran, 2001; Lim, Wun, & Chew, 2014).
36 http://worldconferences.net/ejournals/glit/
GLIT E-Journal on Information Technology and Language Practice (e-ISSN: 2289-7445) Volume 2, Issue 2
Nevertheless, majority of teachers still perceived themselves not prepared for teaching the higher-
order thinking skills and claimed themselves not capable of coming up with examination questions to
test such skills (Rajendran, 2001). Perhaps what these teachers need is ample of opportunities to
practice creating and carrying out their assessment plus the chance to improve them.
2.2 Online assessment
A plausible solution to the issue being addressed so far is to train and motivate teachers to conduct the
online assessment. This type of assessment requires teachers to assess their students' performance
through an e-learning platform. This proposed solution seems to be a practical one because most
schools and universities are equipped with computer laboratories. Online assessment allows teachers
and any educators systematically to attest their teaching practice and test development. It has been
shown to improve teachers' teaching and learning method; teachers are not only able to continuously
evaluate their teachings but also concurrently, gauge their students' academic performance through
online assessment.
2.2.1 Learning Management System (LMS): Moodle
Learning Management System (LMS) is defined as “software that has been used in a learning content
presentation which has a significant role and complexity in e-learning environment” (Aydin and
Tirkes, 2010, p.2). LMS is an e-learning platform utilized to support the move to improve educational
quality and prepare a wider range of options. Concerning assessment purpose, LMS provides a
different range of testing and evaluation format, thus, offers flexibility in grading information.
Modular Object-Oriented Dynamic Learning Environment (Moodle) is one type of LMS popularly
used today. Moodle, with its modular design, can support a variety different kinds of assessment
which allows test developers to manipulate some test constraints such time, date and test duration
(Aydin & Tirkes, 2010). It provides an advanced online module in which questions can be created in
different formats, such as multiple-choice questions, matching questions, and short answer questions.
Also, educators can prepare their online assessments based on their specified learning objectives.
Moodle is one of the top choices of learning institutions regarding robustness. Among the best features
of Moodle are its notable outstanding ability to assist educators to improve pedagogical quality (Aydin
& Tirkes, 2010), ease of maintenance or usability (Wright & Wright, 2011), and „reliability and
functionality‟ of teaching and learning content especially in regard to testing and assessment (Whelan
& Bhartu, 2007, p.1055; Costa, Alvelos & Teixeira, 2012).
3. Main Objective of the Study
The main objective of the present study was to demonstrate that online assessment can be carried out
using Moodle and that this LMS could assist test developers in examining the effectiveness of their
test items, and also furnish them with data which would enable them to improve the quality of their
tests further.
3.1. Specific Objective
Specifically, this study aimed at assessing students‟ critical thinking ability through a standardized test
uploaded at the university e-learning website, and their performance in a public speaking test
developed using the quiz module, one of the tools available in Moodle.
4. Research Questions
Follows are the research questions addressed in the study:
a. What is the reliability of the two tests used in the study?
b. Is students‟ critical thinking related to their performance in the public speaking test?
c. How can item analysis statistics provided by Moodle help instructors improve the quality of the
public speaking test?
37 http://worldconferences.net/ejournals/glit/
GLIT E-Journal on Information Technology and Language Practice (e-ISSN: 2289-7445) Volume 2, Issue 2
i. How many items are considered easy and how many are categorized as difficult (i.e., what
is the IF index for each item)?
ii. How good are the test items at discriminating good students from weak ones (i.e., what is
the ID index for each item)?
iii. How good are the test item distractors (i.e., what is the DE index for test items)?
5. Methodology
5.1 Sample
The original sample of the present study consisted of 131 students from four public speaking classes
taught by the present teacher researchers. The final sample used to establish the relationship between
the two most important variables (i.e., performance in a public speaking test and critical thinking) in
this study was only fifty- eight (58) since not all of the students attempted the critical thinking test
posted on the university e-learning site.
5.2 Instruments
Two tests were used in the study: Cornell Critical Thinking Skills Test (CCTST), Level X, and a
public speaking test. The items for both tests were created using Moodle Quiz Module and were later
uploaded to the university e-learning site for students to access.
5.2.1 Cornell Critical Thinking Test Level X
Cornell Critical Thinking Skills Test (CCTST), Level X, was used to measure the critical thinking
ability of the students involved in the present study. This standardized test was developed by Ennis,
Millman, and Tomko (1985 & 2005) and later translated into Malay by Shaharom (2004). The test
contains seventy-one multiple-choice items, and a score of one (1) mark is given to a correctly
answered test item leading to a total mark of seventy-one (71) for performance on the test. A total high
score on the test is taken to reflect a high critical ability of the test taker.
5.2.2 Public Speaking Test
The public speaking test employed in this study had forty multiple-choice items. Each test item had
five options from which a test taker had to choose the correct answer. The test was used to gauge what
students had learned about the theories and guidelines provided in six chapters of The Art of Public
Speaking, a textbook (2010) that was adopted in the related course. The primary objective of the
public speaking course is to enable students to effectively develop and deliver their speeches. The
speech preparation process requires students to exercise their critical thinking skills so that their ability
to discern relationships among ideas are clearly indicated in their outlines and texts.
In the past, the public speaking test at University Malaysia Terengganu was administered by gathering
eight hundred to one thousand students in a few examination halls where test papers were distributed
to them. However, the establishment of an upgraded university e-learning site has motivated language
instructors to conduct the test online. The site adopted an LMS named Modular Object-Oriented
Dynamic Learning Environment (Moodle) as a teaching and learning platform. Moodle has made it
possible for the researchers and colleagues to construct and administer online public speaking tests
through its Quiz Module for the past six semesters. The earlier version of the online multiple-choice
question (MCQ) tests given to the students relied heavily on the test items provided in the test bank of
the adopted textbook. However, experience gained from administering online tests using such test
items had informed the present teacher researchers the need to discard some of the test bank items and
construct new items which suit local contexts to ascertain that the course educational goals could be
met and that cheating among students could be prevented. Also, the researchers felt the need to find a
mechanism to allow them to administer the test without having to call for all language instructors to
invigilate their respective groups of students taking the test in the computer laboratories. Observation
38 http://worldconferences.net/ejournals/glit/
GLIT E-Journal on Information Technology and Language Practice (e-ISSN: 2289-7445) Volume 2, Issue 2
of students' behaviour while taking the test and analyses of the test results enabled the present
researchers to gain insights into the problem. They discovered that students took a longer time to
answer test bank items of high levels of Bloom's taxonomy of educational objectives. The inclusion of
such items in tests given later to a different group of students, without the presence of instructors,
showed that cheating was very much reduced. Since then, it has been considered practical to include in
the high-level test items, which require students to apply, analyze, and evaluate information.
The frequent use of Moodle quiz module to create and administer online tests caused the researchers to
explore the potential of the LMS further. The exploration led to the discovery of a tool in the quiz
module which helped the researchers to determine the reliability of a test conducted and the quality of
the test items used. This tool is a statistical technique called item analysis procedure. Relevant
descriptive statistics computed from this procedure are made available by Moodle at the statistics
section of the quiz module. Following the discovery of such a useful tool, the researchers felt a strong
urge to conduct the present research as an attempt to improve further the quality of the test given to the
students. The researchers also wanted to determine the appropriateness of the decisions made by them
throughout the previous test construction process so that they would be better informed when
constructing items for future tests.
5.2.3 Data analysis procedures
This quantitative study employed item analysis procedures offered by Moodle Quiz Module to
determine the reliability of two tests and examine the effectiveness of items in the public speaking test.
The module allowed the researchers to export the collected data and convert them into different forms
such as Microsoft Excel and SPSS data for further statistical analysis. This mechanism enabled the
researchers to compute a correlation coefficient to determine the relationship between students'
performance on the two tests used in the study.
6. Results & Discussion
6.1. Cornell Critical Thinking Test (CCTT)
The analysis done on the data revealed that the CCTT was a reliable one. The alpha value computed
was .73, a value which is within the range of reliability estimates in the test manual (Ennis et al. 1985
& 2005). The mean score for the test indicated that on average the students obtained 33 out 71 test
items correct. This value is much lower than the mean score obtained by the sample norm (consisting
of American undergraduates) in the CCTT test manual. Although the computed mean did not reach the
expected level, it still provides evidence that Malaysian undergraduates can think critically. The lowest
score obtained in the test was 18, and the highest was 50. The mean score might probably be of a
higher value if the number of samples was larger.
6.2. Public Speaking test
The reliability coefficient for the public speaking test was obtained from the Moodle quiz module by
clicking the results and statistics blocks, respectively. As indicated in Figure 1, the computed
reliability value (i.e., the coefficient of internal consistency) was at a satisfactory level: α = .77. The
Figure also illustrates that the test was difficult for the students since the average test score (i.e., the
mean score) was only 54 percent. The skewness and kurtosis values were also below the value of 1.00
which indicates that the distribution of the data approaches the normal pattern.
39 http://worldconferences.net/ejournals/glit/
GLIT E-Journal on Information Technology and Language Practice (e-ISSN: 2289-7445) Volume 2, Issue 2
Figure 1: Descriptive statistics for Public Speaking Test
6.3. Relationship between Critical Thinking Public Speaking
Pearson Product-Moment Correlation analysis conducted on the collected data showed that critical
thinking was positively and significantly correlated to critical thinking. The computed value was .34
indicating that the relationship between the two variables was moderate. The correlation value also
suggests that critical thinking exerts an influence on students‟ performance in the public speaking test:
that is, at a moderate level, students who got a high mark for the public speaking test also received a
high score for critical thinking. This result implies that the effort taken by the researchers to include
higher-order test items is a worthwhile effort since the test items do not only manage to test students‟
comprehension of the content of their textbooks but also assess their ability to exercise their critical
thinking skills.
6.4. Quality of Public Speaking Test Items
One of the objectives of conducting this research was to observe and learn how statistical analysis
performed by Moodle enabled the teacher researchers to determine the efficacy of the test developed
through the Moodle Quiz Module. It was hoped that the information gathered from the module would
assist researchers in improving their test items, which would later be saved in the test bank for future
use.
6.4.1. Item analysis
Item analysis involves the application of statistical procedures on test scores to generate descriptive
and item statistics which can assist test developers in re-examining the effectiveness of their test items
and their test as a whole. These statistics can be computed manually using formulae provided in many
or most books on statistics. However, the procedures to be performed are rather tedious which may
cause many people not to consider using or applying them. With this issue, the Moodle Quiz Module
offers a practical solution. The module employs item analysis procedures which can generate the
statistical data that enables any test developers to examine the psychometric performance of each test
item and the effectiveness of a test as a whole.
The two most useful item statistics provided in the report produced by the Moodle Quiz Module are
item Facility Index (FI) and item Discrimination Index (DI). These indices are useful in assisting test
developers in making informed decisions regarding which items to retain, revise, and discard (Abdul
Rashid, 1996; ScorePak®, 2005; Kaplan & Saccuzzo, 2009; Cohen, Swerdlik & Sturman, 2013). The
two indices range from a value of 0 to 100. The FI is used to assess the difficulty level of a test item,
while the DI is used to examine the ability of an item to discriminate or distinguish between high and
40 http://worldconferences.net/ejournals/glit/
GLIT E-Journal on Information Technology and Language Practice (e-ISSN: 2289-7445) Volume 2, Issue 2
low scorers on the test given. A test item will be considered an easy item if its FI value approaches the
value of 100. An item that has a DI value approaching the zero value is deemed to have a poor
discriminating power.
There is no consensus among practitioners on the definite FI and DI values that can be referred to in
determining the difficulty of an item and its discrimination power. In designing a good MCQ test,
Kaplan and Saccuzzo (2009) recommend the use of test items with FI value ranging from .30 to .70
because such items enable test developers to maximize the information about the differences among
the test takers. The recommended difficulty value range is close to the value range observed in the
normal bell curve; that is, the value range of .25 to .75.
In this study, items with FI index lower than .30 were classified as difficult items while those with FI
above .75 were categorized as easy items. The ideal difficulty value for the five-option multiple-choice
test items used in this study was .60, a value suggested by Cohen et al. (2013). For the discrimination
index (DI), values proposed by ScorePak® (2005) were referred to. Items with DI above .30 are
considered as items with good discriminating power, those with DI between.10-.29 as items with fair
power, and those with DI below .10 as items with poor power.
6.4.2. Item Difficulty Index (FI)
The results of an item analysis performed on the test data were obtained from the Moodle Quiz
Module by clicking on the statistics block, which led the researchers to the page entitled Quiz
structure analysis. The sorting of relevant data can be done on this particular page, but the present
researchers preferred to download the data and converted it into an excel data form in case there was a
need to transform it later into the SPSS data. The data was then sorted to rank test items according to
their FI and DI values. Analysis of the data based on these two values revealed that out of forty test
items, twenty-five were in the desirable FI range of .30-.70, nine items in the easy range, and six in the
difficult range. Two of the items were found to be extremely easy since more than ninety percent of
the students answered them correctly. These items may need to be replaced with more challenging
items or with those having better FI values. One item (with FI value of .8) was found to be tough since
only three out of 131 students managed to answer it right.
6.4.3. Items Discrimination index (DI)
For the individual item power of distinguishing good students from poor ones, seventeen items were
listed as having good discriminating power,( DI ranging from .31- .54), fifteen with fair ability (DI
.12- .29) and eight with poor power (DI -.19-.9). Among the items with low DI, three were observed to
have a negative value which indicates that students who did not score on the test got these items right
while those who did well on the test did not get them right (Cohen et al., 2013; ScorePak®, 2005).
These items need to be attended to immediately and decisions must be made as to revise or discard
them.
6.4.4. Using FI, DI and DE values in evaluating the effectiveness of test items.
Most advocates of item analysis recommend test developers refer to both the FI and DI values of a test
item to determine its quality. In this study, the values of these two indices generated by the Moodle
Quiz Module were used to examine the strength and weakness of each of the 40 test items. For
example, based on its FI and Di values, test item number 1 is considered as an easy item but has a poor
discriminating power. Figure 2 shows the relevant statistics gathered from the item analysis procedure
performed on test item number 1. As shown in the Figure, item 1 has a high FI value but a low DI
value. These values indicate that the item was an easy one to answer for the majority of students who
took it; however, it had a poor ability to discriminate good students from poor ones. It is a good idea to
retain this item after improving its DI value because having a few easy items in a test will provide
motivation to, especially weak students to continue attempting the test.
41 http://worldconferences.net/ejournals/glit/
GLIT E-Journal on Information Technology and Language Practice (e-ISSN: 2289-7445) Volume 2, Issue 2
Figure 2: Item analysis of Question No. 1
To improve item 1, we need to analyse its distractor efficiency (DE) value and the students‟ responses
to the item. This data can be obtained from Moodle quiz structure analysis. Like its DI value, item 1
also had a low DE of 11.36%. This implies that many of its distractors were not functioning well.
According to Hingorjo and Jaleel (2012), a good test item, which has only one or without non-
functioning distractor at all, usually has DE within the range of 75-100%, while a poor item has DE of
less than 5%. Figure 3 presents the Moodle analysis of students' responses to test item 1. As indicated
in the analysis, four of the options were poor distractors. The first option only managed to attract nine
students while both the second and fourth options captured the eye of only one student. These options,
therefore, need to be further improved so that they can attract more students which in turn, can help to
increase the DI value of the item.
Figure 3: Percentage of students choosing the options given in Question No. 1
Another good example is the test item number 15 which is considered as a good item to keep in the
test bank for future use. Figure 4 shows the relevant statistics for the item. The FI value of this item is
within the desirable range (of .30-.70), and its DI value exceeds .31 indicating that the item has a good
discriminating power. However, the item only has DE of 53.61%, which is above 5% but still lower
than the desired DE of 75%.
Figure 4: Item analysis of Question No. 15
Detailed analysis of responses for test item 15, as presented in Figure 5, clearly shows that all options
for this item managed to attract students to them which accounts for its high DI value. This item is
challenging and is a high level of Bloom's taxonomy because it requires students to read each of the
options carefully and figure out their relationships. However, the item still has its weakness; the
second option only attracted one student and should, therefore, be improved to attract more students.
42 http://worldconferences.net/ejournals/glit/
GLIT E-Journal on Information Technology and Language Practice (e-ISSN: 2289-7445) Volume 2, Issue 2
Figure 5: Percentage of students choosing the options given in Question No. 15
Test developers should be wary of items that have negative DI values because those items indicate that
many weak students managed to select the correct option compared to good students (Kaplan &
Saccuzzo, 2009; ScorePak®, 2005). The negative DI value could either be due to a poorly constructed
item or wrongly keyed item. If the item is found to be poorly constructed on closer examination, it
should be eliminated from the test. However, if the value is due to test developer's carelessness in
keying the correct response option, it should then be revised, especially if its DI value is excellent.
Figure 6 highlights the quiz structure analysis for item 23, which is one of the items that has a negative
DI (item 23). The small FI value of about .34 reflects that the item is quite a difficult one.
Figure 6: Item analysis of Question No. 23
Nevertheless, as indicated by its negative DI, the item failed to discriminate students who took the test
according to their ability. Upon examining the analysis of responses for the item (given in Figure 7),
the researchers realized that one of the test distractors or options attracted only one student. Besides,
the item was not that clearly stated; preposition „of' should have been inserted after the word
„objective.' Thus, the item needs revision.
Figure 7: Percentage of students choosing the options given in Question No. 23
7. Conclusion and Future Recommendation
This paper has demonstrated that Moodle is indeed a robust LMS. It enables educators, particularly
language instructors, to develop and administer a paperless test to a big number of students without
having to score the test manually. It also generates statistical data that can assist instructors in
determining the effectiveness of their test items and guide them as to how to improve the items. It is
43 http://worldconferences.net/ejournals/glit/
GLIT E-Journal on Information Technology and Language Practice (e-ISSN: 2289-7445) Volume 2, Issue 2
clearly a useful tool for individuals who are keen to improve their test construction skills so that they
can enjoy their teaching profession better. However, educators should not rely solely on the item
analysis results generated by Moodle quiz module because the data are tentative and can be influenced
by many variables such as the type and size of test takers, and the instructional procedures used
(ScorePak®, 2005). Having had to develop expertise in test development procedures is literally a must
for all educators today. Thus, developing literacy in assessment should certainly be a constant and
deliberate effort of all educators to ensure that they have sound knowledge and skills in testing and
validation.
References
Abdul-Rashid, Rosyati. (1996). The development and validation of an institutional reading placement
test. Retrospective Theses and Dissertations. Paper 207. http://lib.dr.iastate.edu/rtd/207
Avison, D., Lau, F., Myers, M., & Nielsen, P.A.,(1999),Action Research, In Communications of the
ACM, 42:1, 94-97, Retrieved from:
http://www.researchgate.net/publication/220422055_Action_Research/file/60b7d51426bc9eb0d
Aydin, C.C., & Cagiltay, N.E., (2007), How Assessment System of an Open Source Learning
Management System can be Integrated to a remote Laboratory Application? Problems and
Solutions, In The 18th Annual IEEE International Symposium on Personal, Indoor and Mobile
Radio Communication (PIMRC’07), 1-3, Retrieved from:
http://ieeexplore.ieee.org/xpl/articleDetails.jsp?arnumber=4394850
Aydin, C.C., & Tirkes, G., (2010), Open source learning management systems in e-learning and
Moodle, Education Engineering (EDUCON), 2010 IEEE , pp.593,600, 14-16 April 2010.
Retrieved
from: http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=5492522&isnumber=5492336
Cohen, R. J., Swerdlik, M. E., & Sturman, E. D. (n.d.). Psychological Testing and Assessment: An
Introduction to Tests and Measurement (8th ed.). New York, United States of America: Mc
Graw-Hill.
Costa, C., Alvelos, H., & Teixeira, L. (2012). The Use of Moodle e-learning Platform: A Study in a
Portuguese University. In Procedia Technology, 5, 334–343. Retrieved from:
http://www.sciencedirect.com/science/article/pii/S2212017312004689
Ennis, R.H.,Millman,J. & Tomko, T.N.,(1985), Cornell Critical Thinking tests Level X and Level Z
manual, (1st ed.). CA:Midwest Publication.
Hingorjo, M.R., & Jaleel, F. (2012). Analysis of one-best MCQs: the difficulty index, discrimination
index and distractor efficiency. J Pak Med Association, 62(2):142-147.
Kaplan, R. M., & Saccuzzo, D. P. (2009). Psychological Testing: Principles, Applications and Issues
(7th ed.). CA, United States of America: Wadsworth, Cengage Learning.
Lim Hooi Lian, Wun Thiam Yew, & Chew Cheng Meng. (2014). Enhancing Malaysian teachers'
assessment literacy. International Education Studies, 7 (10): 74-81.
Malaysia Ministry of Higher Education (MOHE). (2007). Higher Education Action Plan 2007-2010.
Ministry of Education Malaysia (2013). Malaysia Education Blueprint 2013-2025: Preschool to
postsecondary education.
Rajendran Nagappan. (2001). The teaching of higher-order thinking skills in Malaysia. Journal of
Southeast Asia Education, 2(1).
ScorePak® (2005). Understanding Item Analysis Reports. Retrieved from:
http://www.washington.edu/oea/services/scanning_scoring/scoring/item_analysis.html
Whelan, R. & Bhartu, D., (2007). Factors in the deployment of a learning management system at the
University of the South Pacific. In ICT : providing choices for learners and learning. Retrieved
from: http://www.ascilite.org.au/conferences/singapore07/procs/whelan.pdf
Wright, P. & Wright, G., (2011), Using Moodle to Enhance Thai Language Learning: Instructor and
Learner Perspectives, In The Journal of Kanda University of International Studies, 23, 375-398,
Retrieved from: https://www.kandagaigo.ac.jp/kuis/about/bulletin/jp/023/pdf/019.pdf