DIBELS Scoring Guide
DIBELS Scoring Guide
DIBELS Scoring Guide
DIBELS 8 Edition
August 2018
University of Oregon (2018). 8th Edition of Dynamic Indicators of Basic Early Literacy Skills (DIBELS®).
Eugene, OR: University of Oregon. Available: https://dibels.uoregon.edu
Table of Contents
Chapter 1: Rationale and Overview of DIBELSâ 8th Edition ....................................................... 1
References ................................................................................................................................ 49
Over the past 50 years, empirical work in the field of reading science has evolved sufficiently to
establish consensus on how to assess early literacy skills, especially in reading. This consensus is
built on foundational research demonstrating that our most important challenge related to
preventing reading problems is to ensure that effective literacy instruction and early intervention
upon the first signs of trouble are accessible for all students (Snow, Burns, and Griffins, 1998).
Oregon, 2018) has gained widespread use in the United States as a measure of reading skills.
DIBELS includes subtests designed primarily to measure the foundational reading skills
emphasized in the National Reading Panel Report (National Institute of Child Health and Human
Development, 2000), including phonological awareness, phonics (or decoding), fluency, and
DIBELS 8th Edition is conceptualized, including the theory of action upon which the test is built.
We also provide an overview of the DIBELS suite of subtests and the reasons for testing each of
Overview of DIBELSâ
DIBELS 8th Edition takes a curriculum-based measurement (CBM) approach to assessing
reading. It is intended for assessing reading skills from the beginning of kindergarten through the
end of eighth grade. DIBELS was originally developed for use in identifying students
intervene early and prevent the occurrence of later compounded reading difficulties.
DIBELS was also designed to evaluate the effectiveness of the interventions delivered to
those students receiving support, to inform changes in intervention when indicated as a means of
maximizing student learning and growth. The use of DIBELS is consistent with the 2004
called multi-tiered systems of support (MTSS) -- to identify students with learning disabilities. In
indicates risk for the development of learning difficulties. Data are gathered to determine which
students need more support, whether they are responsive to the intervention provided, and who
needs more intensive support based on a lack of response (e.g., Al Otaiba, Connor, Foorman,
Greulich, & Folsom, 2009; Fuchs & Vaughn, 2012; Gersten, Compton, Connor, Dimino,
DIBELS 8th Edition builds on this rich history by implementing innovative design
features to address some of the most common critiques of CBM reading assessments. For
instance, oral reading fluency (ORF) has been critiqued for having “form” effects in that some
passages are simply harder to read than others, even when readability is held constant (e.g.,
Baker, Biancarosa, Park, Bousselot, Smith, Baker, Kame’enui et al., 2015; Christ & Ardoin,
2009; Cummings, Park, Schaper, 2013; Francis, Santi, Barr, Fletcher, Varisco, & Foorman,
2008). Another critique centers on “floor” and “ceiling” effects, which occur when many
students receive scores of zero or the maximum possible score respectively. These effects can
create problems when monitoring for growth (e.g., Catts, Petscher, Schatschneider, Bridges, &
DIBELS, including DIBELS Next. But DIBELS 8th Edition has a lot to offer that no other
edition of DIBELS and no other curriculum-based measurement (CBM) system ever has before.
In fact, DIBELS 8th Edition is more useful for more students in more grades than ever before.
• New grade levels. DIBELS has been extended through the end of eighth grade.
• Consistent subtests within grade. Subtests relevant to a given grade are provided at
every benchmark period.
• New and revised subtests. DIBELS now includes a new subtest and existing subtests
have undergone extensive improvement efforts to maximize their usefulness.
o Expanding the safety net. DIBELS now offers a word reading fluency measure
that can help identify students with poor sight word and irregular word reading
skills that other subtests miss.
o Expanding the utility. DIBELS forms now have items that progress in difficulty
beyond risk cut-points designed to provide data teachers can use in planning
instruction for all students.
• Dyslexia screening. DIBELS subtests offer efficient and cost-effective measures of
processing speed, phonological awareness, and the alphabetic principle for dyslexia
screening purposes.
• New composite scores. DIBELS composite scores are also getting an overhaul to make
them better than ever at predicting risk (arriving in 2019).
• Equated scores. DIBELS will offer equated scores for subtests within a grade level (also
arriving in 2019).
reading comprehension. DIBELS subtests are timed measures so that efficiency in reading skills
is considered as well as accuracy. In many ways the DIBELS subtests represent not only the
constructs in the National Reading Panel Report (NICHD, 2000), but also a developmental
continuum. As a result, although DIBELS 8th Edition provides consistent subtests within a grade
level, the subtests included change across grades in a manner that parallels student development
and instructional foci (Adams, 1990; Chall, 1996; Ehri, 2005; Paris & Hamilton, 2009).
Letter Naming Fluency (LNF) is a standardized, individually administered test that provides a
measure of risk for reading achievement. LNF is based on research by Marston and Magnusson
(1988) and is administered to students in the fall of kindergarten through the spring of first grade.
In LNF, students are presented with a page of 100 upper- and lowercase letters arranged
in a random order and are asked to name as many letters as they can. Students are given 1 minute
to provide letter names. If a student does not know a letter name, the examiner provides the letter
For DIBELS 8th Edition, LNF now accounts for how frequently letters appear in both
upper- and lower-case forms. To better control differences in difficulty between forms,
consistent rules are used in both kindergarten and first grade regarding when less frequent letters
can appear on the forms. Each form in both grades begins with a sampling of the 20 most
frequently seen letters (Jones & Mewhort, 2004), thereby preventing students from getting
frustrated by forms that begin with rarer letters, such as X or q. The kindergarten version of LNF
LNF excludes three letters on all forms: upper- and lower-case W and lower-case L.
Although these are obviously important letters for students to know, they introduce real problems
in a fluency assessment. W is the only letter with a multi-syllabic name: three syllables to be
exact. As a result, any time W appears, it takes three times as long to name as other letters, which
negatively affects a student’s LNF score. The lower-case L (l) was eliminated because it is easily
confused with both the upper-case I and the number 1. Not only does this visual similarity pose
problems for students, but it has also historically created scoring problems for the adult
administering the assessment. By avoiding these letters, each included item (or letter) is equally
The LNF score is the number of letters named correctly within 1 minute. LNF takes about
2 minutes to administer. For students identified as at risk on LNF, the primary instructional goals
should be in phonological awareness, alphabetic principle, and accuracy and fluency with
connected text. LNF remains the dominant predictor of student success for the fall of
PSF assesses a student’s ability to fluently segment two- to six-phoneme words into their
individual phonemes. In PSF, the examiner orally presents a series of words and asks the student
“sat,” and the student said “/s/ /a/ /t/", the student would receive three points for the word. After
each response, the examiner presents the next word. Students are given 1 minute to segment the
In DIBELS 8th Edition, PSF accounts for both word frequency and the number of
phonemes in a word. All forms draw only from the 2,500 most frequent words in English (Balota
et al., 2007) to minimize vocabulary familiarity from interfering with student performance. In
addition, to better control differences in difficulty between forms, consistent rules are used in
both grades regarding where less frequent words can appear on the forms. Moreover, spelling
patterns are ordered in terms of the number of phonemes, proceeding from two phoneme words
In kindergarten, the first 20% of items have two phonemes, while the remaining 80%
have three phonemes. In this way, PSF now avoids the distinct floor effects (i.e., many students
scoring zero) in kindergarten that have plagued previous versions and, thus, eliminates the need
for a separate measure of initial sound fluency. In first grade, the progression in difficulty is a bit
more rapid, with the first 13% of items having two phonemes and then increasing in phonemes
The PSF score is the number of correct phonemes produced within 1 minute. The PSF
measure takes about 2 minutes to administer and has 20 alternate forms per grade for monitoring
progress.
alphabetic principle. NWF is seen as a “pure” measure of the alphabetic principle, because
vocabulary and sight word knowledge cannot play a role in recognizing nonsense words. NWF is
administered to students in the fall of kindergarten through the spring of third grade.
NWF assesses a student’s ability to decode words based on the alphabetic principle.
Traditionally in NWF, the student is presented with an 8.5 inch x 11 inch sheet of paper with
randomly order VC and CVC nonsense words (e.g., sig, rav) and asked to verbally produce (a)
the whole nonsense word or (b) individual letter sounds. For example, if the stimulus word is
“hap”, the student could say the nonsense word as a whole or “/h/ /a/ /p/” to receive three letter
sounds correct (and one whole word correct, if they read the whole word without hesitations,
segmenting, or repetition). Students are given 1 minute to read or sound out as many nonsense
In DIBELS 8th Edition, NWF now accounts for the frequency of spelling patterns (Jones
& Mewhort, 2004; Norvig, 2012). As a result, all forms utilize only phonetically regular letter
combinations that actually appear in English. Thus, students will no longer be asked to decode
nonsense words like “fev” or “kaj”, and nonsense words like “kex” will appear less often than
DIBELS 8th Edition also expands the spelling patterns assessed beyond simply
CVC patterns, the first grade forms also include vowel-consonant (VC) spelling patterns. In
addition, the latter half of first grade forms include additional spelling patterns typically taught in
Edition also now offers NWF in second and third grade by including more complex phonics
patterns in these grades. As a result, DIBELS NWF forms provide instructionally relevant
information even for students who are at minimal risk in kindergarten through third grade. New
spelling patterns included in first through third grade appear in the table below.
An additional improvement to NWF is that we have gone back to scoring words recoded
correctly (WRC) rather than whole words read (WWR). Whereas with WWR students only
received credit if they correctly read a nonsense word at first sight (i.e., without sounding out),
with WRC they also receive credit if they blend a nonsense word after sounding out the
component sounds. Because both methods of scoring predict student risk, in DIBELS 8th Edition,
students receive credit for blending nonsense words whether they sound them out first or not.
The final scores for NWF are the number of letter sounds read correctly and the number
of nonsense words read or recoded correctly as a whole word within 1 minute. The NWF
measure takes about 2 minutes to administer and has 20 alternate forms per grade for monitoring
progress.
The new Word Reading Fluency (WRF) subtest represents a part of reading that was not
previously assessed in DIBELS. WRF involves reading words out of context. Word Reading
reading “sight” words. Sight words include words with irregular pronunciations (non-decodable
words like “the” and “was” and “of”) as well as common words with regular pronunciations
(decodable words like “in” and “we” and “no”). WRF is administered to students from the fall of
While most CBM systems assess either nonsense word reading or real word reading,
DIBELS 8th Edition offers tests of both NWF and WRF in kindergarten through third grade.
WRF targets real words based on age of acquisition in students’ vocabulary (Brysbaert &
Biemiller, 2017) and their frequency in written text (Balota et al., 2007).
In DIBELS 8th Edition, WRF assesses only words that are typically acquired orally in or
before a given grade. This reduces the likelihood that students will encounter words on the
assessment that they have never heard before and are not yet expected to know. In addition, each
form starts with a sample of the most frequent words seen in text and then moves on to less
frequent words in the latter half of the form. In this way, WRF yields instructionally relevant
information both for students at risk and students at minimal risk. Finally, DIBELS WRF
accounts for word complexity, as measured by the number of syllables in a word. All forms
include one-syllable words. Grades 1-3 include two-syllable words, and Grades 2-3 include
again only those that are typically acquired by Grade 3 and are frequently seen in print.
These features ensure the instructional relevance of DIBELS WRF results for all students.
Importantly, our research has shown that the inclusion of WRF helps to identify students at risk
In WRF, the examiner presents the student with a page of words arranged in rows and
asks the student to read as many words as possible within 1 minute. Words must be blended for
the student to receive credit. The final score is the number of words read correctly within 1
minute. The WRF measure takes about 2 minutes to administer and has 20 alternate forms per
and fluency with connected text. ORF is administered to students in the fall of first grade through
ORF assesses a student’s ability to read words in connected text. In ORF, the examiner
presents the student with a passage and asks the student to read the passage aloud for 1 minute.
Words omitted, substituted, and hesitations of more than three seconds are scored as errors.
DIBELS 8th Edition marks the first time that DIBELS ORF requires the administration of
only one passage per benchmark period. Research has shown that administering more than one
passage does little to improve the reliability and validity of ORF, meaning that the supposed
benefits of administering three passages just does not pan out (Baker et al., 2015; Petscher &
An additional unique and exciting feature of DIBELS 8th Edition ORF passages is that
they were written by published authors and elementary and middle school teachers, most of
whom had previous experience writing for students. The authors had diverse backgrounds, came
from across the US, and had experience writing in a range of genres. In addition, all ORF
passages were reviewed by a panel of parents and former teachers for developmental
appropriateness and text complexity in the grades for which they were intended. Passages
deemed inappropriate, too complex, or not complex enough for a given grade were either revised
(and reviewed again) or discarded. As a result, ORF passages are not only more engaging for
both students and assessors, but also feel more authentic and appropriate for the grades in which
they appear.
The final scores for ORF are (a) the number of words read correctly within 1 minute and
(b) the accuracy percentage. The ORF measure takes about 2 minutes to administer and has 20
Maze
Maze subtests (formerly known as DAZE) will be included as part of DIBELS 8th Edition
reading comprehension. Maze is administered to students in the fall of second grade through the
The new version of maze will be based on research that has shown consistently that maze
measures tend to get at only very low-level comprehension. To make DIBELS maze measures
written by published authors and experienced teachers. Second, more work has gone into the
selection of distractors. Third, we have revised formatting to make reading the passages easier on
the eye. Finally, maze measures will be available in second through eighth grade instead of only
In Maze, the examiner presents students with a passage that has every seventh word
removed and replaced with three options. In third through eighth grade, the first and last sentence
are left intact, and in second grade, the first two sentences and last sentence are left intact. The
final score is the number of maze words selected correctly within 3 minutes. The Maze measure
takes about 5 minutes to administer. Forms for progress monitoring are not provided.
of special intervention programs through progress monitoring, and to provide minimum levels of
performance for all students to reach to be considered on track for becoming a reader through
Screening and benchmark goals for DIBELS 8th Edition subtests were based on research
that examined the longitudinal predictive validity of a score on the Reading subtests Iowa
Assessments (University of Iowa, 2015) toward the end of the school year for every grade except
kindergarten. In kindergarten, DIBELS Next (Good & Kaminsky, 2009) composite scores were
used as the criterion. In both cases, risk was defined as performing at or below the 20th percentile
rank and benchmark (or minimal risk) was defined as performing at or above the 40th percentile
The first step in RTI and MTSS models is to identify students early who may need additional
instructional support to meet a benchmark goal. Consistent with legislation in many states, early
screening also helps to identify risk for dyslexia. To identify risk and the need for support,
DIBELS benchmark assessments are administered to all students in the school three times per
year (at the beginning, middle, and end of the school year). Periodic screening of all students
DIBELS 8th Edition is also undergoing study as a dyslexia screener. DIBELS cut-scores
detect risk for reading problems, including dyslexia, making DIBELS measures an efficient and
cost-effective way to screen for dyslexia. In addition, to improve DIBELS 8th Edition’s validity
for these purposes, the LNF, PSF, and NWF measures are being validated against other measures
of rapid automatized naming (RAN), phonological awareness, and the alphabetic principle that
are typically used in dyslexia identification. As a result, DIBELS users can be confident that
their screening measures are just as trustworthy for dyslexia screening as they are for detecting
Progress Monitoring
The purpose of the ongoing progress monitoring data provided by DIBELS 8th Edition is to
inform teachers of the extent to which the instructional support provided to each student
identified as at risk is working. Teachers can use data about students’ progress to adjust the
amount and type of support, including group membership and/or grouping strategies to meet
© 2018 University of Oregon. All rights reserved. DIBELS® 8th Edition
August 2018 13
students’ changing needs (Al Otaiba et al., 2009; Fuchs & Vaughn, 2012; Gersten et al., 2009).
Appropriate ongoing progress monitoring data places an upper limit on the amount of time an
intervention is allowed to continue without either modifying instructional variables (e.g., group
for a different intervention (Al Otaiba et al., 2009; Fuchs & Vaughn, 2012; Gersten et al., 2009).
Benchmark Goals
The purpose of the DIBELS 8th Edition benchmark goals is to provide educators with standards
for gauging the progress of all students. The benchmark goals represent levels of performance
for all students to reach to be at minimal risk of developing reading difficulties, which is defined
as performing at or above the 40th percentile rank on the criterion reading assessment.
Benchmark assessment multiple times a year can help to assure that students who are on track for
DIBELS 8th Edition is intended for use with students enrolled in kindergarten through eighth
grade. Subtests can be administered to students with or without reading difficulties and
disabilities, with frequency of assessment adjusted based on the assessment purpose (e.g.,
Examiners who give and interpret DIBELS 8th Edition must receive training in
standardized administration and scoring procedures. Standardization ensures reliable scores and
The next section presents general instructions that apply to all DIBELS 8th Edition
subtests. That section is followed by specific instructions for administering and scoring the five
DIBELS 8th Edition subtests: (a) letter naming fluency, (b) phonemic segmentation fluency, (c)
nonsense word fluency, (d) word reading fluency, and (e) oral reading fluency. Specific materials
required for each subtest are listed in the subtest descriptions. Throughout this chapter, bold
Arial font is used to indicate scripted directions or prompts provided to the student.
administered subtests. With the exception of Maze, all DIBELS 8th Edition subtests are 60-
second timed measures. When 60 seconds have elapsed, the examiner always places a bracket
(i.e., ]) after the last item completed in 60 seconds and says Stop. Also, if a student makes an
error, put a slash (i.e., /) through the incorrect item. If a student makes an error but self-corrects
the error within 3 seconds, mark SC over the item. Each subtest has a specific discontinue rule.
assessment is also discontinued altogether. The benchmark discontinue rules are explained where
Rule
Kindergarten, Fall If PSF is discontinued, do not Enter 0 for PSF. Do not enter scores
administer NWF and WRF. for the remaining subtests: NWF and
WRF
Kindergarten, Winter If NWF is discontinued, do Enter 0 for NWF. Do not enter scores
First grade, Fall If WRF is discontinued, do Enter 0 for WRF. Do not enter scores
These rules are intended to save time and to spare the student unnecessary frustration.
The rules were derived from a national field trial that indicated students who scored 0 for the
indicated assessments in the periods specified above were extremely unlikely to get any items
correct on the remaining subtests. Nonetheless, examiners have the option of administering the
DIBELS 8th Edition measures early literacy skills in English. Therefore, students should use the
English pronunciation of words. However, it is important to mention that students are not
penalized for varied pronunciation due to dialect or articulation. For example, if the student
consistently says /th/ for /s/ and pronounces “thee” for “see” when naming the letter “C”, credit
is given for naming the letter correctly. This is a professional judgment and should be based on
the student’s responses and any prior knowledge of the student's speech patterns.
Different regions of the country use different dialects of American English. In the next section is
a pronunciation guide that is particularly helpful with the Phonemic Segmentation and Nonsense
Word Fluency subtests. These pronunciation examples may be modified consistent with regional
vowels (e.g., word, far), which are sometimes also called r-colored vowels. Considerable
disagreement exists about how many phonemes exist in words with r-controlled vowels and thus
in American English (e.g., Bizzocchi, 2017; Fry, 2004; Lockenvitz, Kuecker, & Ball, 2015).
Whereas earlier editions treated some as single phonemes and others as two or more phonemes,
DIBELS 8th Edition simplifies the treatment of r-controlled vowels by treating them as single
phonemes. Nonetheless, test administrators should take local dialects and articulation issues into
account when scoring nonsense words or phonemic segmentations involving r-controlled vowels.
In some regions in the US, r-controlled vowels are more clearly separated into multiple sounds or
phonemes (e.g., “lair” might be pronounced as “layer”). Students using such a pronunciation
Note. Both voiced and unvoiced forms of ‘th’ and ‘sh’ are acceptable for nonsense words
Materials
• Scoring book
• Student form
• Pen or pencil
• Clipboard
• Timer
Administration
1. Position the clipboard and scoring book so that the student cannot see what you record.
2. Place the student copy of the LNF subtest in front of the student.
3. Say these specific directions:
(point).
Begin.
Acceptable prompts
There are two acceptable prompts for LNF: a prompt for when students hesitate and for when
Hesitation Prompt. If the student hesitates for 3 seconds on a letter, score the letter as
incorrect, provide the correct letter, point to the next letter, and say:
Keep going.
This prompt may be repeated. For example, if the letters are “p T n” and the student
says, “p” then does not say anything for 3 seconds, prompt by saying T, then point to "n"
and say:
Keep going.
Repeat this as many times as needed throughout administration. The maximum time for
Letter Sound Prompt. If the student provides the letter sound rather than the letter
name, say:
Discontinue rules
Discontinue LNF Rule. If the student reads 0 correct letter names within the first line,
discontinue LNF, put a bracket after the last letter attempted and record a score of 0.
Scoring rules
LNF provides one score: the number of letters named correctly. Mark student responses
according to the rules in the first table below. The second table provides several examples of
Incorrect responses Make a slash (/) through each letter named incorrectly.
Self-corrections If a student makes an error but corrects it within 3 seconds, write "SC"
Letter reversals A letter is incorrect if the student substitutes a different letter for the
(Note that lower-case L does not appear on LNF forms, and the font
Procedure Letters
b T n E “d…T…n…E” b T n E 3 /4
p S n L “q…S…m…L” p S n L 2 /4
M I k L “M…L…k…L” M I k L 3 /4
Letter sounds A letter is incorrect if the student provides the letter-sound for the
stimulus letter (e.g., /d/ for “D”). A prompt for providing letter-sounds is
Procedure Letters
b T n E “/b/…T…n…E” b T n E 3 /4
p S n L “p…/s/…m…L” p S m L 3 /4
M I k L “M…I…/k/…L” M I k L 3 /4
Omissions A letter is incorrect if the student skips the letter. If the student skips an
entire line, cross out the line and record a score of 0 for that line
The observer should judge the full test administration. That includes observing setup and
directions, timing and scoring the test in parallel with the examiner, checking the examiner’s
accuracy in procedures using the fidelity checklist in the Appendix, and deciding if the examiner
Materials
• Scoring book
• Pen or pencil
• Clipboard
• Timer
Administration
1. Position the clipboard and timer so that the student cannot see what you record.
2. Say these specific directions:
After I say it, you tell me all the sounds in the word.
(1 second pause)
3. Give the student the first word and start the timer.
4. Follow along in the Scoring Booklet. As the student says the sounds, underline each
different, correct, sound segment produced. Put a slash (/) through sounds produced
incorrectly. See Acceptable Prompts and Scoring Rules for more details.
5. As soon as the student is finished saying the sounds in the current word, present the next
word promptly and clearly.
6. At the end of 60 seconds, stop presenting words and stop the timer. Place a bracket
(]) after the last sound provided by the student.
Acceptable prompts
There is only one acceptable prompt for PSF: a prompt for when students hesitate.
Hesitation Prompt. If the student hesitates for 3 seconds, give the next word, and score
the word (or remaining sounds in the word if word has been partially segmented) as
incorrect by leaving it unmarked (no slashes or underlines). Repeat this prompt as many
Discontinue rules
Discontinue PSF Rule. If a student has not given any sound segments correctly in the
first 5 words, discontinue PSF, put a bracket after the last word attempted and record a
score of 0.
student does not get any sounds correct in the first 5 words, discontinue PSF and any
further benchmark assessments (i.e., NWF and WRF) for that time of year. At all other
Scoring rules
PSF provides one score: the sum of sound segments produced. Students receive 1 point for each
different, correct, part of the word. Mark student responses according to the rules in the first
table below. The second table provides several examples of common situations and how to score
in them.
Correct responses Underline the sound segments in the word the student produces that are
correctly pronounced.
Incorrect responses Make a slash (/) through sounds pronounced incorrectly. Circle the
segmentation.
Self-corrections If a student makes an error but corrects it within 3 seconds, write "SC"
Schwa sounds Schwa sounds (/u/) added to consonants are not counted as errors. Some
and some early learning of sounds includes the schwa. For example, if the
word is “track,” and the student says “tu...ru...a...ku” they would receive 4
of 4 points.
Additions Additions are not counted as errors if they are separated from the other
sounds in the word. For example, if the word is “track,” and the student
Sound The student may elongate the individual sounds and run them together as
long enough to make it clear they know the sounds in the word, they would
Partial The student is given credit for each correct sound segment, even if they
segmentation have not segmented to the phoneme level. Use the underline to indicate the
size of the sound segment. For example, if the word is “track,” and the
Overlapping The student receives credit for each different, correct, sound segment of
segmentation the word. Thus, if the word is “track,” and the student says “tra...ack,” the
student would receive 2 of 4 points because /tra/ and /ack/ are both
Mispronounced The student does not receive credit for sound segments that are
segment mispronounced. For example, if the word is “track,” and the student says
“t...r...a...gs” they would receive no credit for /gs/ because there is no /g/ or
No If the student repeats the entire word, no credit is given for any sounds. For
segmentation example, if the word is “track,” and the student says “track,” circle the
Procedure Segments
Spelling If the student spells the word, no credit is given. For example, if the word
is “track,” and the student says “t …r…a...c…k”, cross out each sound.
Procedure Segments
Omissions A sound is incorrect if the student omits the sound, but the sound is left
unmarked.
Procedure Segments
The observer should judge the full test administration. That includes observing setup and
directions, timing and scoring the test in parallel with the examiner, checking examiner’s
accuracy in procedures using the fidelity checklist in the Appendix, and deciding if examiner
Materials
• Scoring book
• Student form
• Pen or pencil
• Clipboard
• Timer
Administration
1. Position the clipboard and timer so that the student cannot see what you record.
2. Place the student copy of the NWF practice items in front of the student.
3. Say these specific directions:
Or “lum.”
word)
Start here
When I say “Begin,” read the words the best you can.
Point to each letter and tell me the sound or read the whole word.
Ready?
Begin.
Acceptable prompts
There is only one acceptable prompt for NWF: a prompt for when students hesitate. Execution of
the prompt depends on whether a students is initially blending nonsense words or sounding them
out. If the student is reading words, the rule applies to words; if the student is sounding words
Keep going.
Repeat this as many times as needed throughout administration. The maximum time for
Discontinue rules
Discontinue NWF Rule. If a student does not get any sounds correct in the first 5 words,
discontinue NWF, put a bracket after the last nonsense word attempted and record a score
does not get any sounds correct in the first 5 words, discontinue NWF and any further
benchmark assessments for that time of year (i.e., WRF). At all other times of year,
Scoring rules
NWF provides two scores: the sum of correct letter sounds (CLS) and the sum of words read or
recoded correctly (WRC). Every correct letter sound receives 1 point for CLS, regardless of
whether a student blends. Words read correctly, whether sounded out initially or not, receive 1
point each for WRC. Mark student responses according to the rules in the first table below. The
second table provides several examples of common situations and how to score in them.
Self-corrections If a student makes an error but corrects it within 3 seconds, write "SC"
Sounds When a student sounds out a nonsense word and then blends it. Underline the
followed by individual letters and then the nonsense word as a whole and score a 3 for CLS
Procedure CLS
WRC
Repeated Letter sounds given twice receive credit once. For example, if stimulus word is
sounds “rab” and the student says /r/ /a/ /ab/, the student receives only 1 point for the
letter sound “a” even though the correct sound was provided twice, and a total
CLS WRC
Partially If a word is partially correct, underline the corresponding letters for the sounds
correct produced correctly and word parts for any sounds blended. Put a slash (/)
responses through incorrectly produced letter sounds (to distinguish from omissions; see
Omissions scoring rule). For example, if the word is “rab” and the student says
“rayb” (with a long /a/), the letters “r” and “b” would be underlined, and the
letter “a” would be slashed with a score of 2 for CLS and 0 for WRC.
CLS WRC
nar “n…er” /n /a r/ 1 /2 0 /1
Sounds out Letter sounds produced out of order are scored as incorrect. For example, if
of order stimulus word is “mot” and the student says /t/ /o/ /m/, only /o/, a letter sound
read correctly, would be underlined with a score of 1 for CLS and 0 for WRC.
This is true even if the student using partial or full blending. Blended letter
sounds must be correct and in the correct position (beginning, middle, end) to
receive credit (see last two examples). If a student reads a nonsense word
using blending, letter sounds produced out of order are scored as incorrect.
CLS WRC
Omissions If a student skips a word or row, skip marking any slash and move to the next
The observer should judge the full test administration. That includes observing setup and
directions, timing and scoring the test in parallel with the examiner, checking examiner’s
accuracy in procedures using the fidelity checklist in the Appendix, and deciding if examiner
Materials
• Scoring book
• Student form
• Pen or pencil
• Clipboard
• Timer
Administration
1. Position the clipboard and timer so that the student cannot see what you record.
2. Place the student copy of the WRF form in front of the student.
3. Say these specific directions:
Start here
When I say “Begin,” point to each word and read it the best you
can.
If you get stuck, I will tell you the word, so you can keep reading.
Begin.
Acceptable prompts
There is only one acceptable prompt for WRF: a prompt for when students hesitate.
Hesitation Prompt. If student hesitates for 3 seconds on a word, give the correct word,
mark the word as incorrect, point to the next word, and say:
Keep going.
Repeat this as many times as needed throughout administration. The maximum time for
Discontinue rules
Discontinue WRF Rule. If a student does not get any words correct in the first line (5
words), discontinue WRF, put a bracket after the last word attempted and record a score
of 0.
student does not get any sounds correct in the first 5 words: discontinue WRF and any
further benchmark assessments for that time of year (i.e., ORF). At all other times of
WRF provides one score: the sum of words read correctly. Mark student responses according to
the rules in the first table below. The second table provides several examples of common
Self-correct If a student makes an error and corrects it within 3 seconds, write “SC”
joy draw cloud “/j/ /oy/ draw cloud” joy draw cloud 2 /3
joy draw cloud “/j/ /oy/ joy draw joy draw cloud 3 /3
cloud”
Word Words read correctly but in the wrong order are scored as incorrect.
Omissions A word is incorrect if the student skips the word. If the student skips an entire
line, cross out the line and record a score of 0 for that line.
The observer should judge the full test administration. That includes observing setup and
directions, timing and scoring the test in parallel with the examiner, checking examiner’s
accuracy in procedures using the fidelity checklist in the Appendix, and deciding if examiner
Materials
• Scoring book
• Student form
• Pen or pencil
• Clipboard
• Timer
Administration
1. Position the clipboard and timer so that the student cannot see what you record.
2. Place the student copy of the ORF form in front of the student.
3. Say these specific directions:
out loud.
If you get stuck, I will tell you the word, so you can keep reading.
When I say “Stop” I may ask you to tell me about what you read, so
do your best reading.
Start here
Ready?
Begin.
Acceptable prompts
There is only one acceptable prompt for ORF: a prompt for when students hesitate.
Hesitation Prompt. If student hesitates for 3 seconds on a word, give the correct word,
and mark the word as incorrect. Repeat this as many times as needed throughout
Discontinue rules
Discontinue ORF Rule. If the student does not read any words correctly in the first line
of the passage, discontinue ORF, put a bracket after the last word attempted and record a
score of 0.
Scoring rules
ORF provides two scores: the sum of words read correctly and an accuracy percentage. The
accuracy percentage is calculated by dividing the sum of words read correctly by the number of
Mark student responses according to the rules in the first table below. The second table provides
Self-correct If a student makes an error and corrects it within 3 seconds, write “SC”
Insertions Inserted words are ignored and not counted as errors. The student does not
get points for inserted words. If the student frequently inserts extra words,
it may be worth noting the pattern at the bottom of the scoring page.
The walk was “The walk was The walk was fun. 4 /4
Repetitions Words that are repeated are not scored as incorrect so long as they are read
dog.”
Sounded out A word is scored as incorrect if it is sounded out correctly but not blended.
eeee … d read.”
eeee … d.”
Abbreviations Abbreviations should be read in the way you would normally pronounce
pea.”
teeth.”
teeth.”
Word order All words that are read correctly but in the wrong order are scored as
incorrect.
The green park “The park green has The green park has 3 /5
Omissions Omitted words are scored as incorrect. If a student skips an entire row,
cross out the row and mark the skipped words incorrect.
The observer should judge the full test administration. That includes observing setup and
directions, timing and scoring the test in parallel with the examiner, checking examiner’s
accuracy in procedures using the fidelity checklist in the Appendix, and deciding if examiner
Adams, M. J. (1990). Beginning to read: Thinking and learning about print. Cambridge, MA:
MIT Press.
Al Otaiba, S., McDonald Connor, C., Foorman, B., Schatschneider, C., Greulich, L., & Sidler, J.
F. (2009). Identifying and intervening with beginning readers who are at-risk for
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4296731/
Baker, D. L., Biancarosa, G., Park, B. J., Bousselot, T., Smith, J. L. M., Baker, S. K., . . . Tindal,
G. (2015). Validity of CBM measures of oral reading fluency and reading comprehension
on high-stakes reading assessments in grades 7 and 8. Reading and Writing, 28, 57–104.
doi: 10.1007/s11145-014-9505-4
Balota, D. A., Yap, M. J., Hutchison, K. A., Cortese, M. J., Kessler, B., Loftis, B., . . . Treiman,
R. (2007). The English lexicon project. Behavior Research Methods, 39(3), 445–459. doi:
10.3758/bf03193014
Bizzocchi, A. (2017). How many phonemes does the English language have? International
3134.0510006
Brysbaert, M., & Biemiller, A. (2017). Test-based age-of-acquisition norms for 44 thousand
10.3758/s13428-016-0811-4
effects associated with universal screening and their impact on the early identification of
10.1177/0022219408326219
Chall, J. S. (1996). Stages of reading development (2nd ed.). Fort Worth, TX: Harcourt Brace.
Christ, T. J., & Ardoin, S. P. (2009). Curriculum-based measurement of oral reading: Passage
equivalence and probe-set development. Journal of School Psychology, 47, 55–75. doi:
10.1016/j.jsp.2008.09.004
Cummings, K. D., Park, Y., & Schaper, H. A. B. (2013). Form effects on DIBELS Next oral
38(2), 91-104.
Ehri, L. C. (2005). Learning to read words: Theory, findings, and issues. Scientific Studies of
Francis, D. J., Santi, K. L., Barr, C., Fletcher, J. M., Varisco, A., & Foorman, B. R. (2008). Form
effects on the estimation of students' oral reading fluency using DIBELS. Journal of
Fry, E. (2004). Phonics: A large phoneme - grapheme frequency count revised. Journal of
Gersten, R. M., Compton, D., Connor, C. M., Dimino, J., Santoro, L., Linan-Thompson, S., &
No. NCE 2009-4045). Washington, DC: National Center for Education Evaluation and
Retrieved from
https://ies.ed.gov/ncee/wwc/Docs/PracticeGuide/rti_reading_pg_021809.pdf
Good, R. H., & Kaminsky, R. A. (2009). DIBELS Next. Dallas, TX: Voyager Sopris Learning.
Jones, M. N., & Mewhort, D. J. K. (2004). Case-sensitive letter and bigram frequency counts
Lockenvitz, S., Kuecker, K., & Ball, M. J. (2015). Evidence for the distinction between
delivery systems: Enhancing instructional options for all students (pp. 137–172).
National Institute of Child Health and Human Development. (2000). Teaching children to read:
from http://www.nichd.nih.gov/publications/pubs/nrp/documents/report.pdf
mixture modeling for emergent and developing skills with distributional changes over
Norvig, P. (2012). English letter frequency counts Mayzner Revisited or ETAOIN SRHLDCU.
Paris, S., & Hamilton, E. E. (2009). The development of children's reading comprehension. In S.
Israel & G. Duffy (Eds.), Handbook of research on reading comprehension (pp. 32–53).
Petscher, Y., & Kim, Y.-S. (2011). The utility and accuracy of oral reading fluency score types
10.1016/j.jsp.2010.09.004
Riedel, B. W. (2007). The relation between DIBELS, reading comprehension, and vocabulary in
10.1598/RRQ.42.4.5
Snow, C. E., Burns, M. S., & Griffin, P. (1998). Preventing reading difficulties in young
University of Iowa. (2015). Iowa assessments. Orlando, FL: Houghton Mifflin Harcourt.
University of Oregon. (2018). Dynamic Indicators of Basic Early Literacy Skills (DIBELS®)
Needs
Pass practice
1. Holds clipboard and timer so student cannot see what is recorded.
5. Follows along and marks the scoring book as the student responds.
9. At the end of 60 seconds, puts a bracket (]) after the last letter named and
says Stop.
10. Accurately determines and records the total number of correct letter
names in 60 seconds. Score is within 2 points of the expert examiner.
Needs
Pass practice
1. Holds clipboard and timer so student cannot see what is recorded.
4. Follows along and marks the scoring book as the student responds.
5. As soon as the student is finished saying the sounds in the current word,
presents the next word promptly and clearly.
6. Administers acceptable prompts correctly and when appropriate.
9. Stops at the end of 60 seconds and puts a bracket (]) after the last response.
10. Accurately determines and records the total number of correctly produced
phonemes in 60 seconds. Score is within 2 points of the expert examiner.
Pass Needs
Practice
1. Holds clipboard and timer so student cannot see what is recorded.
5. Follows along and marks the scoring book as the student responds.
9. At the end of 60 seconds, puts a bracket (]) after the last sound provided
and says Stop.
10. Accurately determines and records the correct letter sounds produced
and words read correctly within 60 seconds. Score is within 2 points of
the expert examiner.
Needs
Pass Practice
1. Holds clipboard and timer so student cannot see what is recorded.
5. Follows along and marks the scoring book as the student responds.
9. At the end of 60 seconds, puts a bracket (]) after the last sound provided
and says Stop.
10. Accurately determines and records the number of words read correctly.
Score is within 2 points of the expert examiner.
Needs
Pass Practice
1. Holds clipboard and timer so student cannot see what is recorded.
5. Follows along and marks the scoring book as the student responds.
9. At the end of 60 seconds, puts a bracket (]) after the last sound provided
and says Stop.
10. Accurately determines and records the number of words read correctly.
Score is within 2 points of the expert examiner.