Part I: The Instrument: Researcher-Completed Instruments Subject-Completed Instruments
Part I: The Instrument: Researcher-Completed Instruments Subject-Completed Instruments
Instrument is the generic term that researchers use for a measurement device (survey, test, questionnaire, etc.). To help distinguish
between instrument and instrumentation, consider that the instrument is the device and instrumentation is the course of action (the
process of developing, testing, and using the device).
Instruments fall into two broad categories, researcher-completed and subject-completed, distinguished by those instruments that
researchers administer versus those that are completed by participants. Researchers chose which type of instrument, or instruments,
to use based on the research question. Examples are listed below:
Researcher-completed Instruments
Subject-completed Instruments
Rating scales
Questionnaires
Interview schedules/guides
Self-checklists
Tally sheets
Attitude scales
Flowcharts
Personality inventories
Performance checklists
Achievement/aptitude tests
Time-and-motion logs
Projective devices
Observation forms
Sociometric devices
Usability
Usability refers to the ease with which an instrument can be administered, interpreted by the participant, and scored/interpreted by
the researcher. Example usability problems include:
1.
Students are asked to rate a lesson immediately after class, but there are only a few minutes before the next class begins
(problem with administration).
2.
Students are asked to keep self-checklists of their after school activities, but the directions are complicated and the item
descriptions confusing (problem with interpretation).
3.
Teachers are asked about their attitudes regarding school policy, but some questions are worded poorly which results in
low completion rates (problem with scoring/interpretation).
Validity and reliability concerns (discussed below) will help alleviate usability issues. For now, we can identify five usability
considerations:
1.
2.
3.
4.
5.
Have any problems been reported by others who used it?
It is best to use an existing instrument, one that has been developed and tested numerous times, such as can be found in the Mental
Measurements Yearbook. We will turn to why next.
Inter-Rater/Observer Reliability: The degree to which different raters/observers give consistent answers or estimates.
2.
3.
Parallel-Forms Reliability: The reliability of two tests constructed the same way, from the same content.
4.
Internal Consistency Reliability: The consistency of results across items, often measured with Cronbachs Alpha.
Relating Reliability and Validity
Reliability is directly related to the validity of the measure. There are several important principles. First, a test can be considered
reliable, but not valid. Consider the SAT, used as a predictor of success in college. It is a reliable test (high scores relate to high GPA),
though only a moderately valid indicator of success (due to the lack of structured environment class attendance, parent-regulated
study, and sleeping habits each holistically related to success).
Second, validity is more important than reliability. Using the above example, college admissions may consider the SAT a reliable
test, but not necessarily a valid measure of other quantities colleges seek, such as leadership capability, altruism, and civic
involvement. The combination of these aspects, alongside the SAT, is a more valid measure of the applicants potential for
graduation, later social involvement, and generosity (alumni giving) toward the alma mater.
Finally, the most useful instrument is both valid and reliable. Proponents of the SAT argue that it is both. It is a moderately reliable
predictor of future success and a moderately valid measure of a students knowledge in Mathematics, Critical Reading, and Writing.