Should We Test Reading or DIBELS?

  • 21 January, 2018
  • 15 Comments

Teacher question:

I teach first-grade and this year I switched schools. In my previous school, we tested our students with DIBELS three times a year. The idea was to figure if the students were having trouble with decoding so that we could help them. That isn’t how my new principal does it. He has us giving kids a reading comprehension test with leveled books. I asked him about it and he said that the district didn’t care about DIBELS and he didn’t care about DIBELS (he only cares about how kids do on the ____ test). I’m confused. I thought the idea of teaching phonics and fluency was to enable comprehension, but the emphasis on this test seems to suggest—at least in my new school—that is no longer necessary. What should I do?

Shanahan response:

Educators often complain about the intrusive accountability testing imposed by politicians and bureaucrats who wouldn’t know the difference between a textbook and a whiteboard. But many of the dumbest decisions made by teachers and principals are in pursuit of the tests we ourselves impose.

The accountability tests—PARRC, SBAC, and all the other state tests—are there to check up on how well we are doing our jobs. Not surprising that we don’t like those, and that, consequently, we might bend over backward to try to look good on such tests. That’s why many schools sacrifice real reading instruction—that is, the instruction that could actually be expected to help kids read better—in favor of so much test prep and test practice.

But what of the formative assessments that are supposed to help us do our jobs? You know, the alphabet soup of inventories, diagnostic tests, screeners, monitors, and dipsticks that pervade early reading instruction like DIBELS, ERDA, PALS, CTOPP, TPRI, ISIP, CAP, TRC, NWEA, AIMSweb, and TOWRE.

None of these instruments are problematic in and of themselves. Most set out to measure young children’s ability to…well, to… do something relevant to early literacy learning. For instance, they might evaluate how many letters the children can name, or how well they can hear the sounds within words. Sometimes, as in your school, they ask kids to read graded passages or little books and to answer questions about them, or as in your previous school, they might gauge student ability to perceive correctly the sounds within words.

The basic idea of these testing schemes is to find lacks and limitations. If Johnny doesn’t know his letters, then his kindergarten teacher should provide extra tuition in that. If Mary can’t understand the first-grade text, perhaps she should get her teaching from a somewhat easier book. And so on.

That is all well and good… but how we do twist those schemes out of shape! My goodness.

As a result, educators increasingly have grown restive with the “instructional validity” of assessment. Instructional validity refers to the appropriateness of the impact these tests have upon instruction.

DIBELS itself has often been the target of these complaints. These tests shine a light on parts of the reading process—and teachers and principals tend to focus their attention on these tested parts—neglecting anything about literacy development that may not have this kind of flashlight. Thus, one sees first-grade teachers spending inordinate amounts of time on word attack trying to raise NWF (nonsense word fluency) scores, but with little teaching of untested skills like vocabulary or comprehension or writing.

Even worse, we sometimes find instruction aimed at mastery of the nonsense words themselves too with the idea that this will result in higher scores.

Of course, this is foolishness. The idea of these formative testing regimes is to figure out how the children are doing with some skill that supports their reading progress, not to see who can obtain the best formative test scores.

The reason why DIBELS evaluates how well kids can read (decode or sound out) nonsense words is that research is clear that decoding ability is essential to learning to read and instruction that leads students to decode better eventually improves reading ability itself (including reading comprehension). Nonsense words can provide a good avenue for the assessment of this skill because they would not favor any particular curriculum (as real words would), they correlate with reading as well as real words do, and no one in their right mind would have children memorizing nonsense words. Oops… apparently, the last consideration is not correct. Teachers, not understanding or caring the purpose of the test, are sometimes willing to raise scores artificially by just this kind of memorization.

And, to what end? Remember, the tests are aimed at identifying learning needs that can be addressed with extra teaching. If I artificially make it appear that Hector can decode well when he can’t (memorizing the test words is one way to do this), then I get out of having to provide him the instruction that he needs. In other words, I’ve made it look like I’m a good teacher, but what I’ve really done is disguised the fact that Hector isn't succeeding, and I'm delaying any help that may be provided until it is too late.

Another example of this kind of educational shortsightedness has to do with the idea of using the tests to determine who gets extra help, like from a Title I reading teacher, perhaps. In most schools, the idea is to catch kids literacy learning gaps early so we can keep them on the right track from the beginning. But what if you are in a school with high mobility (your kids move a lot)?

I know of principals who deploy these resources later—grades 2 or 3—to try to make certain that these bucks improve reading achievement at their schools. Research might find it best to use these tests early to target appropriate interventions in Kindergarten and Grade 1, but these schmos don’t want to "waste" resources in that way since so many students don’t stick around all the way to the accountability testing. Instead of targeting the testing and intervention at the points where these will help kids the most, these principals aim them at what might make the principals themselves look better (kind of like the teachers teaching kids the nonsense words).

Back to your question… your school is only going to test an amalgam of fluency (oral reading of the graded passages) and reading comprehension. If all that you want to know is how well your students can read, that is probably adequate. If all the first-grade teachers tested her charges with that kind of test, the principal will end up with a pretty good idea of how well the first-graders in his school are reading so far. Your principal is doing nothing wrong in imposing that kind of test if that is what he wants to know. I assume those results will be used to identify which kids will need extra teaching.

I get your discomfort with this, however. You are a teacher. You are wondering… if little Mary needs extra teaching what should that extra teaching focus on?

Because of the nature of reading, that kind of assessment simply can’t identify which reading skills are causing the problem. Mary might not read well—the test is clear about that, but we can’t tell whether this poor reading is due to gaps in phonological awareness (PA), phonics, oral reading fluency, vocabulary, or reading comprehension itself.

The default response for too many teachers, with this test or any other, is to teach something that looks like the test. In first grade that would mean neglecting those very skills that improve reading ability. The official panels that have carefully examined the research and concluded that decoding instruction was essential did so because such teaching resulted in better overall reading achievement (not just improvements in the skill that was taught). The same can be said about PA, fluency, and vocabulary instruction.

I’d love to tell you I have a great solution to your problem… for instance, perhaps all the children could be tested in the way that your principal requires and then anyone who failed to reach a particular reading level could then be tested further using DIBELS or something like DIBELS to identify the underlying skills that are likely holding those kids back. That sounds pretty sensible since it would keep teachers from just focusing on the underlying skills (and then ignoring reading comprehension), and yet, I quake at those teachers who will now teach reading with the test passages or who will coach the kids on the answering the test questions so that no one needs to be tested further--in other words, hiding the fact that their kids are struggling.

The key to making all of this work for kids is: 

  1. All teachers and principals need to know what skills are essential to reading success. There are skills and abilities inherent in reading comprehension itself (so testing comprehension is not unreasonable) but there are also enabling skills that make comprehension possible for young readers (and testing those skills makes sense too). Knowledge of letters, ability to perceive sounds, decoding facility, knowledge of high-frequency words, oral reading fluency, awareness of word meanings, and ability to make sense of text are all part of the reading process—and all of these should be taught and tested from the start.
  2. It is also critical for educators to know that this list of essential skills is not a sequence of teaching… in which one starts with letters and sounds and ends up with ideas. In fact, good early reading instruction provides a combination of instruction in decoding, fluency, comprehension, and writing—from the very beginning.
  3. Formative assessment can help us to monitor student progress in all of these areas, one is no more important than another. Because a student lags in one area, is no reason to neglect instruction in the other areas. If you find that a youngster does not decode well, I would provide added or improved decoding instruction—but I would also maintain a daily teaching regimen with all of the other literacy components, too.
  4. It is essential that educators know what tests can be used to measure the various components of literacy and how these assessments work. A nonsense word test, for instance, isn’t trying to find out if kids can read nonsense words (but how well they can decode any words). A fluency test is not about speed reading, but about being able to read the text so that it sounds like language. A comprehension test score might reveal comprehension problems, but you can’t determine that without additional tests of the underlying skills—since low comprehension may be, and usually is at this point, the result of poor decoding or lack of fluency.
  5. No educator should ever teach the test, nor should lessons look like the test. These kinds of tests are not competitive. They are there to help us identify who needs help and what they may need help with. Every screwy thing we do to make the scores look better than they are is harmful to the education of children.

So, a pox on both your houses… That your principal doesn’t care why kids are having reading trouble is a serious impediment for the boys and girls in that school. That you don’t recognize the value of a test of your students’ actual reading ability concerns me as it might indicate a willingness to go off the deep end, teaching some aspects of reading to the neglect of others. Teach it all, monitor it all... and help these children to succeed.

Comments

See what others have to say about this topic.

Harriett Janetos Jan 21, 2018 08:18 PM

Let me come to this teacher's defense--possibly. Tim, you say: "That you don’t recognize the value of a test of your students’ actual reading ability concerns me as it might indicate a willingness to go off the deep end, teaching some aspects of reading to the neglect of others." The teacher didn't mention which test with "leveled books" she gives, but if her district is like mine and many others, it's the Fountas & Pinnell Benchmark Assessments, which are very counterproductive in K-1. Why? Because the first few levels use predictable text rather than decodable text. What they assess is whether students have mastered around 30 high-frequency words and then are able to look at the picture and read words like "elephant" or "swing", words most kindergarteners could not independently decode. As you've previously poined out (In your discussion of Reading at the Speed of Sight), relying on context support is characteristic of poor readers. Both Louisa Moats and Marilyn Adams--cowriters of the Foundational Skills portion of the Common Core Standards--have spoken out eloquently against using predictable text for beginning readers. After the National Reading Panel's report, California only approved two textbooks for adoption--and both used decodable text as part of their reading assessments. Many districts have taken a step backward.

Nancy Steinbock Jan 21, 2018 09:10 PM

This is an excellent compilation of the points that need to be addressed in early literacy education and beyond, i.e., that 'reading' is the amalgam of a number of skills. We have been in touch over the summer. I am an SLP and specialist in reading and writing acquisition as part of my larger experience with identifying and treating language disorders, including 'written oral language disorders' commonly identified as 'dyslexia.' I understand that there is a vicious entrapment of accountability, measured by tests resulting in the wholesale adoption of balanced literacy programs which are meant to address PA, phonics and 'comprehension.' You are absolutely right. . .we don't teach kids to read but push them through these hoops and then, I hear stories of 3rd graders having troubling reading (and writing).
The default solution often is to put them on some kind of computer program as 'teacher', not as practice. Coupled with this, is the sad neglect of handwriting even though it has been shown through viable, valid research, to be an important pathway into integrating sound/symbols, fluent spelling and importantly, memory. As children move through to secondary education, often they are given even more apps. . essentially z-apping them into under-performance, whether secondary to underlying linguistic deficits or frankly, misteaching. SLPs in schools, are often hobbled by reliance on tests to pinpoint these deficits and then, by lack of time or knowledge to help address them.

This becomes essential when teaching ESL students and in fact, I am working to develop now, I hope, with the local library system, ESL courses for parents who not only want to improve their language uptake by working through phonology, vocabulary, reading comprehension and writing but also, to expose their preschoolers and early elementary kids to these very concepts to assist them in developing areas that might not be fully valued and taught in their own classrooms.

Nancy Rose Steinbock, M.A., CCC-SLP
Edgartown, Massachusetts

Kay Stahl Jan 21, 2018 11:56 PM

I am always concerned when schools exclusively use leveled text kits to assess primary level readers.
1. Typically, that indicates that there is less instructional emphasis on explicit, systematic instruction of the code. Many of these schools are not using a consistent program of instruction. Without assessing foundational skills, teachers are unlikely to know where students are along the word recognition developmental continuum. All students require systematic instruction in word recognition.
2. Although some test using a finely defined text-reading gradient is necessary K-1, there is a great deal of variabillity in test administration, scoring, and interpretation by teachers using kits of leveled readers. It is troubling if you only have one test and the data is not collected in reliable, standardized ways. This is particularly true if the school’s test kit does not have standardized questions for each book and consistent quantitative ways to code the retelling. Schools should be having annual (or more frequent) booster sessions to ensure fidelity.
3. The teacher requesting Tim’s feedback mentioned that her school valued the leveled reader as a gauge of first grade comprehension. This is troubling to me. Texts at the easiest levels (A-G) do not adhere to comprehensive narrative text structures. Therefore, they may be harder for young children to remember than a more higher level book that has a more complex story line. Additionally, without a script to prompt children to “think about what is happening in the book, because you will have to retell what you read and answer questions when you finish,” young readers get so focused on performing and getting words right that they may be startled by those activities at the end, especially if they are only doing collaborative retellings during instruction. At that reading stage (A-G/H) the M-S-V analysis is an important tool. In many schools that I visit, they are just using the percentage to inform instructional level, omitting the time-consuming MSV analysis. Finally, tests of listening comprehension can distinquish between comprehension issues and decoding problems.

Harriett Janetos Jan 22, 2018 12:07 AM

MSV analysis is discredited in Mark Seidenberg's Language at the Speed of Sight and David Kilpatrick's Essentials of Assessing, Preventing, and Overcoming Reading Difficulties along with works by other researchers.

Tara Mikell Jan 22, 2018 12:26 AM

I do not think this teachers principal is uncaring. She is new to the school and may just not know the thoughts behind the principal's reasoning. DIBELS has it place, but apparently this principal does not feel that it is at his school. DIBELS, to me, does not provide instruction for reading skills and strategies. It provides the educator with the necessary information to plan instruction for the students that will give them the reading skills and strategies to do well on the state test. I teach 3rd grade and my students take the state test for the first time in their school career during 3rd grade. My school does not use DIBELS in 3rd grade, but our 1st and 2nd grades do use it. By 3rd grade, we expect that our students can decode words so that they can read fluently through a text. Of course, often that is not the case. This does not mean that I just teach the rest of the class and leave them behind just because I do not use DIBELS. I still have that state test to look forward to in April and I know that, ultimately, I will be held accountable for my students scores. My goal is to provide instruction through any means necessary to increase the reading comprehension of my students. If that means that I need to incorporate some aspects of DIBELS, then I will do that. If it means that I need to focus on other areas of reading instruction, then I will do that. I believe that the principal in this situation only wants what is best for his students and he does not think it is DIBELS.

Nancy Steinbock Jan 22, 2018 12:39 AM

This is an excellent compilation of the points that need to be addressed in early literacy education and beyond, i.e., that 'reading' is the amalgam of a number of skills. We have been in touch over the summer. I am an SLP and specialist in reading and writing acquisition as part of my larger experience with identifying and treating language disorders, including 'written oral language disorders' commonly identified as 'dyslexia.' I understand that there is a vicious entrapment of accountability, measured by tests resulting in the wholesale adoption of balanced literacy programs which are meant to address PA, phonics and 'comprehension.' You are absolutely right. . .we don't teach kids to read but push them through these hoops and then, I hear stories of 3rd graders having troubling reading (and writing).
The default solution often is to put them on some kind of computer program as 'teacher', not as practice. Coupled with this, is the sad neglect of handwriting even though it has been shown through viable, valid research, to be an important pathway into integrating sound/symbols, fluent spelling and importantly, memory. As children move through to secondary education, often they are given even more apps. . essentially z-apping them into under-performance, whether secondary to underlying linguistic deficits or frankly, misteaching. SLPs in schools, are often hobbled by reliance on tests to pinpoint these deficits and then, by lack of time or knowledge to help address them.

This becomes essential when teaching ESL students and in fact, I am working to develop now, I hope, with the local library system, ESL courses for parents who not only want to improve their language uptake by working through phonology, vocabulary, reading comprehension and writing but also, to expose their preschoolers and early elementary kids to these very concepts to assist them in developing areas that might not be fully valued and taught in their own classrooms.

Nancy Rose Steinbock, M.A., CCC-SLP
Edgartown, Massachusetts

Tim Shanahan Jan 22, 2018 02:27 AM

Harriet— the teacher did include the test name (I removed it). However, only testing decoding and fluency—even in grade 1 is problem as it leads teachers to narrow their teaching inappropriately.

Kay— I agree with all of your points. There are very real problems with any assessments, including levels testing. Teachers can inflate scores inappropriately making them useless,

Tara—it doesn’t matter that the principal’s heart is in the right place only that his plan ignores the nature of reading and what it takes to learn reading. His wishful thinking will prevent some kids from learning as well as they could.

Nancy- Thanks.

Harriett Janetos Jan 22, 2018 03:01 AM

You're right, Tim. But only testing beginning reading with text emphasizing high-frequency words and "picture reading" is equally problematic. This principal is wrong to use just one type of assessment, but if I could only choose one I would choose decodable rather than predictable text.

Troy Fredde Jan 22, 2018 01:42 PM

Tim is saying these assessments are not evil. You have to know what they are intended for, what data you gain from them, and how to use the information. You can’t over rely on 1 assessment. The F&P Benchmark is not evil. Know how to use it. Decodable texts have their issues also. They dumb down language and natural comprehension, but can be used for certain purposes. We need to be assessing our Ss in multiple ways using multiple texts continually. And formative standardized assessments have their place used correctly. We all have opinions, but we have to acknowledge facts and facts and proper use of assessments. You have think and be intentional with your teaching and testing. Know the limitations of your assessments and the proper way to use them! Know that certain assessments are supposed to be used to flag Ss as possible you needing more support and teaching, but and that after they are flagged more in depth assessments are needed to pinpoint what kind of intervention is needed.

Susan Hall Jan 26, 2018 04:15 AM

Tim, I really enjoyed reading your blog about appropriate use of assessment data, and the importance of selecting the right type of assessment for various purposes. Universal early literacy screening data with assessments such as DIBELS provide teachers and administrators with critical information about a student's acquisition of skills that are highly predictive of later reading milestones. As you point out, teaching to the test shows a fundamental misunderstanding of WHY these assessments work. They measure indicators that happen to be highly predictive. Yet when a student doesn't perform well on an indicator, the next step is to use a diagnostic assessment that measures more skills in that area (typically phonological awareness or phonics) to ponpoint the lowest missing skill along a continuum. This is where instruction should begin. Susan Hall, author of I've DIBEL'd, Now What?

Jessica Foster Mar 05, 2018 12:53 AM

This article had several excellent points and was relevant to some of the issues I have had during my years teaching. I think one of the biggest concerns for teachers, unfortunately, is the standardized tests that our students have to take at the end of the year. These tests are compared to other classrooms, other schools, and other districts, and just about everyone can get their hands on the results. Since there is such a demand for your students to do well, a lot of times teachers tend to focus on “teaching to the test”. It really is such a disservice to the most important part of the whole puzzle, our students. They are the ones that are not getting quality teaching and quality learning. Instead they are being taught how to answer questions on a test. I believe this is why we see such a problem with retention of skills. Our students are memorizing, not learning.

With that being said, I understand exactly where the teacher that wrote in the question is coming from. Many times, the non-classroom teachers in our school think that a single test such as DIBELS or Reading Inventory is a true picture of the student’s reading ability, and that is not always the case. When I taught 5th grade several years ago, our entire school would DIBEL three times a year. There was a push to make sure the students were reading the “words per minute” norm for their grade level. I remember several students who would read well above the norm but would continue to fail assignments in the classroom. When I would bring these students up at our RTI meetings, their DIBELS scores was always the focus. Since their WPM was where it should be, several people in these meetings would say that they should be passing work in the classroom and would move on. How frustrating! Asking for help for these students and it being pushed to the side. So as a teacher, I figured out what I needed to do in my classroom to provide the best reading instruction to these students. That is why one of the last sentences from this article really hits home for me. “Teach it all, monitor it all... and help these children to succeed.” That is what we must do as educators. It is all about the children that are in our classrooms each year.

Kara West Mar 12, 2023 09:46 PM

Check out Emily Hanford’s Sold a Story podcast or the forthcoming documentary The Right to Read. You are right to be concerned if your district is not assessing phonics.

Sarah R Nov 22, 2024 01:34 AM

What about with testing that is truly high stakes? Our state (Indiana) recently passed legislation that will prevent the promotion of any third graders who don’t pass IREAD.

We were redistricted this year and started with 75% new students. Most of whom are “well below” on all DIBELS measures. In years past we’ve seen that the majority of students who pass IREAD in third grade pass because they seem to be able to compensate for a lack of foundational skills. The foundational portion of the test is only 25-35% of the total score—so the students with an overall passing score typically fail the foundational part of the test, and pass the fiction/non-fiction portions, which are between 60-80% of the test.

All students get 60 minutes of EL education and 30 minutes of UFLI daily. Our admin is in favor of directing interventions to focus heavily on test prep: passages that are in a similar testing format to IREAD, teaching clue words, test taking strategies, etc., with the goal of teaching students how to pass the fiction/non-fiction parts of the test. Last year, I wasn’t a fan of this, because it took intervention resources away from tier 2. But this year, with so many students facing retention, I don’t know which is worse: being retained and dealing with the negative impacts of that (but having the opportunity to catch up on foundational skill gaps), or being taught to compensate to pass a test and being promoted (but without ever receiving foundational skills).

Please advise, this is so heavy on my heart.

Timothy Shanahan Nov 22, 2024 02:22 PM

Sarah--
I think principals (and, perhaps, you now) have great faith that the kind of practice you describe will enable students to do better on those tests. That is not necessarily the case. There are all kinds of things that your school could do to improve student performance on those tests including increasing the amount of reading instruction that you provide, making sure that instruction addresses decoding, fluency, comprehension, and writing, using your Tier 2 resources to address weak areas, getting parents involved in helping increase the amount of instruction and practice, etc. Don't try to raise test scores, try to teach students to read better.

tim

Sarah R. Nov 24, 2024 12:18 AM

Hi Tim,

Thanks for the response. I failed to add that we are also doing 45 min./day of Corrective Reading in addition to the 45 min of UFLI and the 45 min of EL Education each day. We also have students from a local college coming in to tutor 15 students, 4 days a week for 45 min/day. Only 15% of 3rd graders met benchmark on NWF-CLS at BOY, and the goals set by DIBELS show that even with above average growth, many students would still be “well below” benchmark.

All 3 of the third grade teachers are new to teaching foundational reading. I’m an instructional coach, and am coaching 1 of the 3, but the other 2 have declined support. They all have named that they’re too overwhelmed to accept much coaching support, even when I offer to take other things off of their plate. Other than involving parents (which we definitely could do a better job with) I just don’t know what more we could do to do to get this cohort caught up in time to pass IREAD by the end of year.

I don’t care about passing the test, other than the fact that it has some pretty significant consequences for those students who don’t pass it (retention). I’m not sold on the idea of offering that type of practice (test prep), and before retention was at stake, I was opposed to it. I just don’t know where to go from here. I just want to do right by our students; I’m open to any/all evidence-based ideas.

What Are your thoughts?

Leave me a comment and I would like to have a discussion with you!

Comment *
Name*
Email*
Website
Comments

Should We Test Reading or DIBELS?

15 comments

One of the world’s premier literacy educators.

He studies reading and writing across all ages and abilities. Feel free to contact him.