Discussion about this post

User's avatar
James Cantonwine's avatar

I think this misses a few key points with state tests. I'll caveat this with a note that state assessment systems vary; however, many states are using the same vendors for similar assessments under different names. My own experience is as a district assessment coordinator in Washington State.

1. Preliminary results are available to school and district staff FAR earlier than they are made public. Rarely are the official results any different than preliminary. For that reason, my district now sends preliminary score reports home at the end of the school year. Reasons for public delays at the state level include: staff shortages within data departments, discrepancies with determining accountability site for high mobility students, and disputes around individual student scores. (I'm sure there are others I don't see from a district point-of-view.)

2. State tests are not always measuring the same constructs as other assessments like MAP and NAEP, which feeds into the varied testing times. If the state is assessing English Language Arts, some amount of writing will need to be produced by the student. When assessments only measure reading, they can be shorter and are easier for machine scoring. Something similar happens with math tests that are designed to elicit evidence for the thinking and reasoning on the student's end. Whether we should measure these is an open question. MAP is highly correlated to state test results and maybe we don't need to assess writing, too.

3. State tests aren't necessarily any longer than tests like MAP or NAEP: it just looks like they are. First, assessment developers report the "typical" time taken for a test, but teachers are affected by the time taken by slowest student. That's generally much longer than vendors will share. Second, weeks-long testing windows show when the assessment could happen, not when it did happen. When testing does drag on too long, which it often does, it's frequently a result of the perceived stakes of the assessment. Staff may feel an incentive to really drag out the big accountability measure to eke out any last score improvements.

Expand full comment
3 more comments...

No posts