Assessment Course
Assessment Course
Assessment Course
Evaluation of
Learning
(PGDT 423)
Unit 1: Assessment: Concept, Purpose, and Principles
• Introduction
• This part will familiarize you with some basic
concepts as: test, measurement, assessment and
evaluation;
• Brief explanation of the role of educational
objectives in assessment;
• Important principles that have to be adhered when
assessing students’ learning; and
• The importance of involving students in the
assessment process
Unit one Learning Outcomes
After the end of this unit, you will be able to:
• Define the meaning of test, measurement,
assessment and evaluation.
• Examine the purposes of assessment and evaluation
of learning
• Identify the principles of assessment and evaluation
of learning.
• Apply the principles of assessment and evaluation
of learning in the local context
1.1 Concepts
Other uses- The tests are also used for other purposes
such as information on grading of learners, giving feed-
back to learners, feed- back about effectiveness of
teaching, motivating learners to study and serving as a
scientific tool in research, Education and social sciences.
Examination
• covers large areas of contents,
• given at the end of the semester or the
course,
• the number of items included is large.
• Its main purpose in to assign grades.
Measurement:
• In our day to day life there are different things that
we measure.
• We measure our height and put it in terms of
meters and centimeters. We measure some of our
daily consumptions like sugar in kilograms and
liquids in liters.
• We measure temperature and express it in terms of
degree centigrade or degree Celsius.
• Hence to measure these things appropriate
instruments such as a meter, a weighing scale, or a
thermometer are mandatory in order to have
reliable measurements.
• Similarly, in education measurement is the process
by which the attributes of a person are described in
numbers.
• It is a quantitative description of the behavior or
performance of students.
• As educators we frequently measure human
attributes such as attitudes, academic achievement,
aptitudes, interests, personality and so forth.
• Measurement permits more objective description
concerning traits and facilitates comparisons.
• Hence, to measure we have to use certain
instruments so that we can conclude that a certain
student is better in a certain subject than others.
• Eg., a mathematics test is an instrument containing
questions and problems to be solved by students.
The number of right responses obtained is an
indication of performance of individual students in
mathematics
• Thus, the purpose of educational measurement is to
represent how much of ‘something’ is possessed by
a person using numbers.
• measurement is the assignment of quantitative
value to the results of a test or other assessment
techniques. Measurement can refer to both the
score obtained as well as the process itself
The process of measurement involves
• Specifying what is to be measured (learning objectives
and contents)
• Constructing devices to measure a certain attributes.
• Defining attributes to be measured (what behavior is to
be measured)
Results of measurement from class room tests may be used
to:
• Direct prescriptive study for the students who took the
test
• Indicate whether a students has mastered a well defined
body of subject matter or developed particular skill
• Indicate where a student stands in relation to others who
Characteristics of Psychological Measurement
• Psychological measurements have three
characteristics that put limitation on the
interpretation and use of psychological test. These
are:
1. It is descriptive:- a test score only describe
performance, It does not interpret or evaluate it
• e.g an IQ score of 102 describe a person’s
performance on particular test administration:
• interpretation of scores requires further
information, such as the performance of other
people to and knowledge of the test takers
background.
2. It is relative: scores usually are interpreted by
comparison with other people. Because:
• there are few absolute standards of performance
in areas that we test
• differences between people usually are more
interesting as useful than absolute
3. It is indirect:
• Characteristics are inferred from specific
behavior rather than being measured directly.
Assessment:
• In educational literature the concepts ‘assessment’ and
‘evaluation’ have been used with some confusion.
• Some educators have used them interchangeably to mean
the same thing. Others have used them as two different
concepts. Even when they are used differently there is too
much overlap in the interpretations of the two concepts.
• Cizek (in Phiye, 1997) provides a comprehensive definition of
assessment.
the planned process of gathering and synthesizing information
relevant to the purposes of:
(a) discovering and documenting students' strengths and
weaknesses,
(b) planning and enhancing instruction, or
(c) evaluating progress and making decisions about students.
• Generally, educational assessment is viewed as the
process of collecting information with the purpose of
making decisions about students.
• Instruments used to collect information includes tests,
observations, checklists, questionnaires and interviews.
• Rowntree (1974) views assessment as a human
encounter in which one person interacts with another
directly or indirectly with the purpose of obtaining and
interpreting information about the knowledge,
understanding, abilities and attitudes possessed by that
person.
• The key words in the definition of assessment is
collecting data to make decisions. Hence, to make
decisions one has to evaluate.
Types of Assessment
There are Informal and Formal types of Assessment
Informal Assessment –
• It is the process of gathering information about learning
on spur of movement or casually during class room
activity.
• It is not necessarily carefully planned but it meant provide
information that is critical to be known at that moment
• It includes a variety of techniques as:- questioning
learners,
• Observing learners work,
• Reviewing learners home work,
• Talking with learners and listening learners during reaction.
• Peer and self evaluation, and discussion.
• It doesn’t contribute to a students final grade
Formal assessment
• It is usually a written document, focused on assessing
specific competences of the learners.
• In a formal assessment score or grades is given on the
bases of student performance.
It includes variety of techniques as –
• Test, oral examination.
• Performance assessment tasks,
• Class work/Homework
• Examinations, project work,
• Portfolios and the like.
• Laboratory work, group work
• Research work, term paper
Evaluation:
• It is the process of making judgment about a given
situation,
• It refers to the process of judging the quality of
student learning on the basis of established
performance standards and assigning a value to
represent the worthiness or quality of that learning or
performance.
• It is concerned with determining how well they have
learned. When we evaluate, we are saying that
something is good, appropriate, valid, positive,
• Evaluation is based on assessment that provides
evidence of student achievement, the grade/course,
• Evaluation includes both quantitative and
qualitative descriptions of student behavior plus
value judgment concerning the desirability of that
behavior.
• Evaluation = Quantitative description of students’
behavior (measurement) + qualitative description of
students’ behavior (non-measurement) + value
judgment
• Thus, evaluation may or may not be based on
measurement (or tests) but when it is, it goes
beyond the simple quantitative description of
students’ behavior.
• Evaluation involves judgment. The quantitative
values that we obtain through measurement will
not have any meaning until they are evaluated
against some standards.
• Educators are constantly evaluating students and it
is usually done in comparison with some standard.
• Thus, evaluation is described as the comparison of
what is measured against some defined criteria and
to determine whether: it has been achieved,
appropriate, reasonable, valid and so forth.
• Evaluation summarizes and communicates what
students know and can do with respect to the
overall curriculum expectations to: parents,
teachers, institutions of further education, and
students themselves.
• In summary, test is one type of instrument used in
assessment, Measurement is assigning of numbers
to the results of a test or other forms of assessment
according to a specific rule
• Assessment is more comprehensive and inclusive
than testing and measurement. It includes the full
range of procedures (observations, rating of
performances, paper and pencil tests, etc)
• It may also include quantitative descriptions
(measurement) and qualitative descriptions (non-
measurement) of students’ behaviors.
• Evaluation, on the other hand, consists of making
judgments about the level of students’ achievement
for the purposes of grading and accountability and
for making decisions about promotion and
graduation.
Distinction between Assessment & Evaluation
?
• Angelo and Cross (1993) have listed seven basic
assumptions of classroom assessment which are
described as follows:
1. Although not exclusive, the quality of student
learning is directly, related to the quality of teaching.
• Therefore, one of the most promising ways to
improve learning is to improve teaching.
2. To improve their effectiveness, teachers need to
make their goals and objectives explicit. Then
they get specific, comprehendible feedback on
the extent to which they are achieving those
goals and objectives.
?
• According to current cognitive research, people are
motivated to learn by success and competence.
When students feel ownership and have choice in
their learning, they are more likely to invest time
and energy in it.
• Assessment can be a motivator, not through reward
and punishment, but by stimulating students’
intrinsic interest.
Assessment can enhance student motivation by:
• emphasizing progress and achievement rather than
failure
• providing feedback to move learning forward
• reinforcing the idea that students have control over,
and responsibility for, their own learning
• building confidence in students so they can and need
to take risks
• being relevant, and appealing to students’
imaginations
• providing the scaffolding that students need to
genuinely succeed
• There is strong evidence that involving students in
the assessment process can have very definite
educational benefits.
• Activity ?
• 1) As prospective teachers how do you think you
can involve your students in the assessment
process?
• 2) In what ways can students benefit if they are
involved in the assessment process?
• One way in which we can involve our students in
the assessment process is to establish the standards
or assessment criteria with them.
• This will help students understand what is to be
assessed.
• It helps students develop a clear picture of where
they are going, where they are now and how they
can close the gap.
• Another aspect is to involve students in trying to
apply the assessment criteria for themselves. The
evidence is that through trying to apply criteria, or
mark using a model answer, students gain much
greater insight into what is actually being required.
• Third by self-assessment and peer assessment. .
• self-assessment helps to aware one’s own
strengths and weaknesses in a particular subject
and identify one’s own next steps.
• Self-assessment allows students to think more
carefully about what they do and do not know, and
what they additionally need to know to accomplish
certain tasks.
• Peer assessment is making judgment about other
students’ work.
• Students learn how to make better sense of
assessment criteria if it gives feedback and/or
marks against them. Giving and receiving
feedback is an important aspect of student
learning and will be valuable skills for them in
professional contexts and for future learning.
Assessment and Teacher Professional Competence in Ethiopia
advantage
• It does not require much time to answer
• It can cover a wide range of content
• It can be scored quickly, reliably, and objectively
disadvantage
• It promotes memorization of factual information:
names, dates, definitions, and so on.
• encourage students for guessing. because there is
a 50 percent probability of getting the right answer.
• The diagnostic value of such test is nil
suggestions help to construct true/false test items.
• Avoid negative particularly double negative, if used
under line or bold them to minimize confusion.
• Restrict single-item statements to single concepts.
two concepts in a single item statement, may have two
responses
• Use approximately equal number of items
Make proportional number of true and false statement
• Make statements representing both categories equal in
length.
CONSTRUCTING MATCHING TEST ITEM
Disadvantage
• Stress on rote recall and encourage student to spend
their time on rote memorizing of trivial details rather
than seeking important understanding
• In other words it is not suitable to measure higher
order of learning outcomes.
Guideline to construct supply test item
• Omit key words or phrases and substitute by blank space so that the
required answer will be definite and clear
• Avoid indefinite statement that may be logically answered by several
item
• Avoid excessive blanks in a single item when too many blanks are left
an incomplete statement has no meaning or it becomes ambiguous
• Do not take statements directly from the text as basis for short
answer item because it promotes memorization.
• Be sure that the question or statement poses problem to the
examine. A direct question is often more desirable than an
incomplete statement
• The length of the blank space should be equal in length unless it
gives clue to the students. However, be certain to include sufficient
space for the longest response.
CONSTRUCTING MULTIPLE- CHIOCE TEST ITEMS
Table of Specification
• The validity, reliability and usability of test depend
on the care with which the test are planned and
prepared.
• planning classroom test involves identifying the
instructional objectives earlier stated and the
subject matter (content) covered during the
teaching/learning process
• This planning leads to the preparation of table of
specification (the test blue print) for the test
Guides in planning a classroom test.
i. Determine the purpose of the test;
ii. Describe the instructional objectives and content to be
measured;
iii. Determine the relative emphasis to be given to each learning
outcome;
iv. Select the most appropriate item formats (essay or objective);
v. Develop the test blue print to guide the test construction;
vi. Prepare test items that is relevant to the learning outcomes
specified in the test plan;
vii. Decide on the pattern of scoring and the interpretation of result;
viii. Decide on the length and duration of the test, and
ix. Assemble the items into a test, prepare direction and administer
the test.
• A table of specification is a two-way table that
matches the objectives and content
• A table of specification guides the selection of test
items that measures a representative sample of
contents.
• Developing a table of specification involves:
-Preparing a list of learning outcomes, i.e. the type
of performance students are expected to
demonstrate
-Outlining the contents of instruction, i.e. the area
in which each type of performance is to be shown,
and
-Preparing the two way chart that relates the
learning outcomes to the instructional content.
Instructional Objectives
Assessment:
Concept, Purpose,
2 2 1 1 - - 6 24%
and Principles
assessment
strategies 1 1 1 1 - - 4 16%
interpretation of
scores 1 2 1 - 1 - 5 20%
3.1 Introduction
• In the previous unit you learned about various
assessment strategies, planning, construction and
administration of classroom tests.
• In this unit, you are going to learn the techniques
of analyzing responses to test items to determine
their validity and reliability.
• You will also learn about the advantages and
techniques of test item banking.
Learning Outcomes
• 70 X
Groups A B C D* Total
High Scorers 0 1 1 8 10
Low Scorers 1 1 5 3 10
Total 1 2 6 11 20
• Item difficulty interpretation
Alternatives
Group
A* B C D
Upper 8 12 6 4
Lower 5 10 7 8
Unit 4: Interpretation of Scores
• Introduction
• In unit three you learned about how to analyze test
items in order to determine each test item and the
overall test in general. In this unit you are going to
be familiarized with the idea of test score
interpretation and the major statistical techniques
that can be used to interpret test scores.
Particularly, you will learn about the methods of
interpreting test scores, measures of central
tendency, measures of dispersion or variability,
measures of relative position, and measures of
relationship or association.
• Test interpretation is a process of assigning meaning.
• It is necessary because the raw score obtained from
a test standing on itself rarely has meaning.
• For instance, a score of 60% in one test cannot be
said to be better than a score of 50% obtained by the
same test taker in another test of the same subject.
• The test scores on their own lack a true zero point
and equal units.
• Moreover, they are not based on the same standard
of measurement and as such meaning cannot be
read into the scores on the basis of which academic
and psychological decisions may be taken.
• 4.1 Kinds of scores
• Data differ in terms of what properties of the real number
series (order, distance, or origin) we can attribute to the
scores.
The most common kinds of scores include nominal, ordinal,
interval, and ratio scales.
• A nominal scale involves the assignment of different
numerals to categorize that are qualitatively different.
• For example, we may assign the numeral 1 for males and
2 for females.
• These symbols do not have any of the three
characteristics (order, distance, or origin) we attribute to
the real number series.
• The 1 does not indicate more of something than the 0.
• An ordinal scale has the order property of a real
number series and gives an indication of rank
order.
• For example, ranking students based on their
performance would involve an ordinal scale.
• We know who is best, second best, third best,
etc.
• But the ranked do not tell us anything about the
difference between the scores
• With interval data we can interpret the
distances between scores.
• If, on a test with interval data, a Almaz has a
score of 60, Abebe a score of 50, and Beshatu
a score of 30, we could say that the distance
between Abebe’s and Beshadu’s scores (50 to
30) is twice the distance between Almaz”s and
Abebe’s scores (60 t0 50).
• If one measures with a ratio scale, the ratio of the
scores has meaning. Thus, a person whose height
is 2 meters is twice as a tall as a person whose
height is 1 meter.
• We can make this statement because a
measurement of 0 actually indicates no height.
That is, there is a meaningful zero point. However,
if a student scored 0 on a spelling test, we would
not interpret the score to mean that the student
had no spelling ability
4.2 Methods of Interpreting test scores
10 2 4
7 -1 1
9 1 1
6 -2 4
8 0 0
=40x-)2=10
40/5=8
• Raw score method
Score (x) X2
10 100
7 49
9 81
6 36
8 64
• S2 = - = 330- == =2
N 5
nxy xy
rxy =
nx 2
x ny 2 y
2 2
102670 56437
=
10388 56 1020427 437
2 2
2228
= = 0.71
74413301
• Activity: Calculate the Pearson product
moment correlation coefficient
X 44 20 40 10 11 12
Y 3 4 3 40 12 20
UNIT 5: Ethical Standards of assessment
Introduction
• This unit introduce you with ethics as a mechanism
of maintaining quality in assessment practice.
• You will be familiarized with some basic standards
expected of professional teachers to be ethical in
their assessment practices.
• You will also be familiarized with some general
considerations in addressing diversity in the
classroom to make the assessment procedures
accessible and free of bias.
Learning Outcomes
Upon completion of this unit, you should be able to:
• List down ethical and professional standards of
assessment
• Propose contextualized ethical and professional
standards in using assessment
• Sensitize the consequences of unethical use of
assessments
• Adhere to the ethical standards of tests and test
uses
5.2. Sections and sub-sections