Short-Term Impact of $3 Billion School Improvement Grants: Zilch; Long-Term Impact: TBD

A hefty study commissioned by the Institute of Education Sciences (IES) and written by researchers at Mathematica has examined the impact of School Improvement Grants (SIGs) and found that implementing a SIG-funded model "had no impact on math or reading test scores, high school graduation or college enrollment."

As "School Improvement Grants: Implementation and Effectiveness" stated, SIGs were a signature $3 billion program delivered during the Obama administration as part of its American Recovery and Reinvestment Act of 2009. The idea was to award grants to states that agreed to implement one of four intervention models — transformation, turnaround, restart or closure — in their lowest-performing schools. Each model "prescribed specific practices" that were designed to improve student learning, including outcomes for high-need students.

With the implementation of Every Student Succeeds Act, that best-practices approach (with its guard rails) no longer shows up in the SIG program. Now states and districts have a great deal more flexibility in how they attempt to turn around their lowest-achieving schools. But before institutional memory fades regarding that initial SIG investment, the institute wanted to find out the answers to four basic questions:

  • Did schools implementing a SIG-funded model use the improvement practices pushed by the SIGs?
  • Did those practices include a focus on English language learners (ELLs), and did any focus differ between those schools implementing a SIG-funded model and those that didn't?
  • Did the funding have any impact on outcomes for low-performing schools?
  • Did the specific intervention model implemented have any impact on outcomes?

This isn't the first time SIGs have undergone evaluations. Three previous reports examined SIG implementation practices and states' capabilities in supporting school turnaround. But this fourth report builds on those earlier research efforts and adds an additional year of data. It also adds the overall question regarding efficacy of SIGs on student outcomes.

The improvement practices that made up the four intervention models focused on four areas:

  • Adopting comprehensive instructional reform strategies, such as the use of data in implementing instructional programs;
  • Developing and increasing teacher and principal effectiveness, such as using "rigorous, transparent and equitable evaluation systems";
  • Increasing learning time and creating "community-oriented" schools that engage families and meet "students' social, emotional and health needs"; and
  • Having operational flexibility in hiring, discipline and budget; and receiving technical assistance and support.

The research team looked at the SIG-promoted practices in two ways: It conducted a descriptive analysis that compared the use of the practices for 290 schools under SIG-funded models in 2012-2013 and 190 schools that weren't under those models; and it analyzed data from 460 schools to figure out whether implementation of a SIG-funded model in 2010-2011 had any impact on the use of the practices.

Short-Term Impact of $3 Billion School Improvement Grants: Zilch; Long-Term Impact: TBD 

To sort out whether the SIG program had any impact on student outcomes, the researchers analyzed data to measure impact on test scores, graduation rates and college enrollment in a sample of 190 schools eligible for SIG and 270 schools that weren't.

They also undertook an examination of the various intervention models themselves to figure out whether student achievement varied by model. In this, the researchers tapped data from a sampling of 270 schools implementing a SIG model.

The project whittled down its conclusions covered in the 419-page report to five key findings:

  • Although schools implementing SIG-funded models reported using more SIG-promoted practices than other schools, there was no statistically significant evidence that SIG caused those schools to implement more practices.
  • Across all study schools, use of SIG-promoted practices was highest in comprehensive instructional reform strategies (89 percent or 7.1 of eight SIG-promoted practices) and lowest in operational flexibility and support (43 percent or 0.87 of the two promoted practices).
  • There were no significant differences in the use of ELL-focused practices between schools running a SIG-funded model and those that didn't.
  • Overall, across all grades, the implementation of any of the SIG-funded models had "no significant impacts on math or reading test scores, high school graduation or college enrollment."
  • In grades 2 through 5, there was no evidence of one model being more effective at generating higher learning gains than any other model. In higher grades (6 through 12), the turnaround model saw larger student gains in math than the transformation model; however, the report noted, other factors may explain the variances as well, such as "baseline differences between schools implementing different models."

In her coverage of the report for the Thomas Fordham Institute Research Intern Lauren Gonnelli pointed to two disparate reactions to the findings.

First, there was Andy Smarick's response. Smarick, a resident fellow at the conservative think-tank American Enterprise Institute, called the results, "devastating to Arne Duncan's and the Obama administration's education legacy." As he declared, "The program failed spectacularly."

Morgan Polikoff, an associate professor of education at the University of Southern California's Rossier School of Education, was more circumspect. His take: When it comes to policy evaluation, judgment based on early evaluation can turn out to be "misleading." As he noted in an article on C-SAIL, a website supported by IES, although the impact evaluation of the SIG initiative was "neutral, several studies have found positive effects and many have found impacts that grow as the years progress (suggesting that longer-term evaluations may yet show effects)." In other words, sustained improvements such as the ones prompted by the school improvement grants could take a long time to prove their impact. His suggestion: Don't declare failure too soon.

The full report is freely available on the Mathematica Policy Research site.

Featured

  • educators seated at a table with a laptop and tablet, against a backdrop of muted geometric shapes

    HMH Forms Educator Council to Inform AI Tool Development

    Adaptive learning company HMH has established an AI Educator Council that brings together teachers, instructional coaches and leaders from school district across the country to help shape its AI solutions.

  • illustration of stacked coins, bar graphs, downward arrows, and two school buildings

    Survey: Top Education and Budget Challenges for Schools

    A recent survey of more than 2,500 educators, school leaders, and district administrators across the country identified the top challenges schools are facing this year. The 2025 National Educator Survey, conducted by PowerSchool, found that teacher shortages and mounting financial uncertainty are persistent pain points across K-12 education.

  • laptop displaying a phishing email icon inside a browser window on the screen

    ED Grant Portal Target of Phishing Campaign

    Threat researchers at BforeAI have identified a phishing campaign spoofing the U.S. Department of Education's G5 grant management portal.

  • laptop with AI symbol on screen

    Google Launches Lightweight Gemma 3n, Expanding Emphasis on Edge AI

    Google DeepMind has officially launched Gemma 3n, the latest version of its lightweight generative AI model designed specifically for mobile and edge devices — a move that reinforces the company's focus on on-device computing.