Research Asks If Machine Learning Can Make Better Enrollment Predictions
- By Dian Schaffhauser
- 12/07/21
Could
artificial intelligence help school districts better predict their
fall enrollments? That was the question a team
of researchers asked on behalf of the School
District of Philadelphia.
Philadelphia's
school system, like many others, offers residents school choice.
Families may choose to send their students to the neighborhood school
or to another one elsewhere in the city, including public charters or
private schools. That flexibility poses a planning problem: Each year
schools have no idea about the size of their incoming cohorts. While
spring planning may tell district and site leaders that students
intend to go to their local school, in the intervening months
families may choose to send
their kids elsewhere.
The
district allocates resources in March for the upcoming school year
and then reallocates resources as needed after October 1, once it has
taken measure of discrepancies between the prospective spring
enrollment and actual attendance in the fall, a process the district
calls "leveling." The October leveling process can result
in the elimination of classrooms, reassignment of students to
different teachers and shifting of teachers to different schools or
grade levels. This kind of disruption can have a big impact on
classroom engagement for students
To
see if there was a better way to plan for fall enrollment, the
district turned to the Regional
Educational Laboratory Mid-Atlantic, part of the
federal Institute
of Education Sciences network. In a project
administered by Mathematica,
researchers developed three machine learning algorithms for doing
student enrollment prediction, to see if the results were any better
than a "simple regression model" —
figuring out which variables matter and which ones don't, and then
applying the ones with impact to the calculations.
The
AI didn't do any better. According to a recently published report
and infographic,
all four methods had similar accuracy, differing from actual fall
cohort sizes by about six students, on average, including across
schools "with larger proportions of black
students, economically disadvantaged students and English learners."
Each method resulted in the need to reallocate between 20% and 30% of
students to different teachers in the October recount.
While
machine learning didn't generate any better results, the research
project did come up with some guidance:
-
First, for
districts with high rates of mobility, predictions would probably be
refined if the schools gathered additional data later in the spring
and early summer; and
-
Second, while
any of the methods would generate similar predictions, regression
models are easier and less expensive for districts to implement.
Also,
while the research used 259 predictors from multiple schools
(2016-2017 through 2018-2019), just a handful of variables really
makes a difference:
-
The current
enrollment for each grade, which had a predictor importance of 0.97;
-
The number of
students with more than five in-school suspensions (0.91);
-
The number of
students with more than five out-of-school suspensions (0.83); and
-
The number of
students with fewer than six absences (0.65).
Finally,
the researchers offered a caveat: The results may not be applicable
in years to come since COVID-19 has wreaked havoc. The pandemic, the
report stated, "might have fundamentally altered patterns of
attrition and new entrants in a way that models based on historic
data are unable to capture."
Details
are openly available on
the Regional Educational Laboratory website.
About the Author
Dian Schaffhauser is a former senior contributing editor for 1105 Media's education publications THE Journal, Campus Technology and Spaces4Learning.