Poverty and Performance in Virginia Schools

A recent study out of Stanford looked at eleven years of district level NAEP data by race and economic disadvantage from all public school districts and concluded that

racial segregation is strongly associated with the magnitude of achievement gaps in third grade and the rate at which gaps grow from third to eighth grade. The association of racial segregation with achievement gap growth is completely accounted for by racial differences in school poverty (termed “racial economic segregation”). Thus, racial segregation is harmful because it concentrates minority students in high-poverty schools, which are, on average, less effective than lower-poverty schools.

On the subject of economic disadvantage, the VDOE database offers Virginia SOL data on a granular level, so let’s look at some of those numbers. In particular, the graphs below present 2022 SOL pass rates by subject, per school, for students who are “economically disadvantaged” (“ED”) (those who qualify for the free/reduced lunch program, TANF, or Medicaid) and for their more affluent peers (“Not ED”).

Notes: The percentage of ED students below is calculated from the number of ED students taking a particular test and the number of Not ED students taking that test. Data are school averages for the subject area for all grades tested, 3-8 and End of Course. The VDOE suppression rules omit data for just over 3.5% of the entries in the 2022 SOL database, primarily for cases where the number of students in a group is <10.

Aggrieved note: As of 12/20/22 and again on 12/29 (well after I complained), the links on the VDOE Web site to the data dictionary (definition of “economically disadvantaged”) and to the suppression rules, among others, do not work.

To start, here are the school average ED and Not ED data for the reading tests. .

image

The slopes of both fitted lines are negative, consistent with school performance declining with increasing economic disadvantage. The Not ED rate decreases more steeply so that the Not ED/ED difference shrinks from 19.9% at 0% ED to 13.8% at 100%. The R-squared values speak to a modest correlation for the Not ED data and a less robust one for the ED.

Notes: Please remember that correlation does not imply causation. These data speak of correlations between the relative number of ED students and pass rates. The R-squared values measure the strength of the correlations. However, the correlations, standing alone, do not prove that the increasing ED percentage causes the decreases in pass rates. And a different matter: In addition to the 3.5% suppressed data in the 2022 database, the data from some schools are suppressed for only one of the groups (ED or Not ED), leading to a further 1.9% of unusable data for the reading tests, which is 0.15% of the database total.

The database also gives the school average pass rate (here the “All” rate). Adding that All rate to the ED/Not ED graph produces an unreadable mess. But we can look at the All data separately:

image

Or, more usefully, we can add the All fitted line to the ED/Not ED graph.

image

The intercept of the All data fitted line,  95.7, approaches the 96.9 value of the Not ED data. The steeper slope of the All line leads to a 100% ED value of 45.4, which is less than the 46.7 value of the ED line at 100% ED. And the R-squared value for All data winds up greater than either the ED or the Not ED value.

The explanation is simple: If the All data were the simple average of the ED and Not ED pass rates, the All line would lie exactly between the ED and Not ED lines. The different slope and intercept are reflections of the different sizes of the ED and Not ED groups: At low %ED, the All averages are raised by the larger proportion Not ED students, while at high ED, the All values are lowered by the larger numbers of ED results in the average.

Let’s say that again: Because of the roughly fifteen point difference between Not ED and ED pass rates, schools with small populations of ED students see their averages boosted by the larger population of Not ED students, while those with large ED populations see their averages dragged down.

Here are numbers for the two extreme cases in the reading data. “Avg. Pass” is the average of the ED and Not ED rates; “Diff.” is the difference between the All Pass (school average) and Avg. Pass.

image

You read that correctly: The school average at Greenville is 17.6 points higher than the average of the ED and Not ED rates; At Newsome Park, the school average is 17.9 points lower.

Yet the Board of Education measures school quality (and accreditation) by the All data, among other factors. Thus, the evaluation process of our Board penalizes schools with larger populations of ED students and rewards those with fewer ED students, even if both groups were taught with equal effectiveness.

The math data tell the same story.

image

The ED and Not ED data from Virginia imply substantial correlations between ED population and those pass rates. But these data do not tell us the cause(s) of decreasing school performance with increasing ED population. We might expect the following factors to affect that performance:

  • Behavior problems with large ED populations;
  • Schools with large Not Ed populations can cherry pick their hires; schools with large ED populations often must take what they can get;
  • Social pressure from ED students on their Not ED peers;
  • Not ED families in high percentage ED neighborhoods may represent the lower end of the Not ED income spectrum; and
  • Probably others.

In any case, isn’t it long past time for our Board of Education to abandon their school-average poverty penalty and evaluate schools separately for ED and Not ED pass rates, or by the average of the two?