« Are U.S. Teachers Teaching Too Much? | Main | D.C. Update: Allegedly False Test Scores Used for Value-Added Calculations »

When Value-Added Evaluations Meet Cheating Allegations

By now you've read that more than 200 teachers were fired in Washington based on their results under an evaluation system known as IMPACT. (We've been inundated with news on this here in the D.C. area.)

You'll also remember that a number of D.C. public schools are being investigated for allegations of cheating on standardized tests between 2008 and 2010.

So, I'm wondering how the intersection of these two events has (or has not) affected individual teachers. Value-added scores determine teacher effectiveness by looking at student growth on test scores from year to year. Were student test scores from schools where the cheating investigations are ongoing used to evaluate teachers? Let's say a student came from a school that had a high number of erasures in previous years, and that student's scores dropped dramatically in 2011. Is his or her 2011 teacher held accountable for the decline? Has DCPS found a way to take possible cheating into account?

I've contacted DCPS about the issue and am waiting to hear back. I'd like to hear from teachers on this as well (especially those from D.C.). And are teachers in Atlanta, where the cheating investigation has been more conclusive, dealing with similar implications?

Notice: We recently upgraded our comments. (Learn more here.) If you are logged in as a subscriber or registered user and already have a Display Name on edweek.org, you can post comments. If you do not already have a Display Name, please create one here.
Ground Rules for Posting
We encourage lively debate, but please be respectful of others. Profanity and personal attacks are prohibited. By commenting, you are agreeing to abide by our user agreement.
All comments are public.

Follow This Blog


Most Viewed On Teacher



Recent Comments