IE 11 Not Supported

For optimal browsing, we recommend Chrome, Firefox or Safari browsers.

Preparing K-12 and higher education IT leaders for the exponential era

Using Machine Learning to Understand College Closures

Researchers worked with the Federal Reserve to create a predictive model that assesses hundreds of institutional characteristics to estimate the likelihood that a college might close.

A yellow sign reads "private property" on the grounds of Wells College in Aurora, N.Y., which closed in 2024.
Wells College in Aurora, N.Y., closed abruptly in 2024, after 156 years in operation.
David Kidd
As financial pressures mount across higher education, researchers are turning to machine learning to better predict which colleges are at risk of closing.

In recent work, higher education researchers collaborated with the Federal Reserve to develop a predictive model that combines hundreds of institutional characteristics to estimate the likelihood a college might close. The model outperforms financial monitoring systems currently used by the federal government, offering a more nuanced understanding of financial distress in higher ed.


WHY COLLEGES CLOSE


College closures remain relatively uncommon, but financial distress is more widespread. Small institutions, particularly those that rely heavily on tuition revenue, have little margin for error when enrollment drops. Zack Mabel, director of research at the Georgetown University Center on Education and the Workforce (CEW) studying higher education finance, said colleges can experience financial distress for years without closure, and most institutions under stress ultimately do not close. This makes predicting which schools will shutter especially challenging.

“[I] think there are many different factors that contribute to why colleges close,” he said. “There’s no silver bullet.”

Phillip Levine, a professor of economics at Wellesley College, said closures should be understood as the most extreme outcome of broader financial pressures across higher education. For example, higher education institutions rely heavily on human infrastructure, which makes rising labor costs that outpace inflation put sustained stress on budgets. Over the past decade, inflation-adjusted tuition has been flat or declining, leaving institutions with persistent operating stress. When colleges can no longer cut costs without undermining their ability to function, closure becomes the remaining option.

“You can fire people, that reduces costs, but then someone’s got to stand in front of the room and teach the kids,” he said. “Then, the kids need food and support and other services. And at some point, it just becomes not sustainable.”

When closures do occur, they can be disruptive for students, employees and the surrounding community that relied on the institution to bolster the local economy.


CURRENT METHODS


Today, the federal government’s primary tool for assessing college financial health is the U.S. Department of Education’s Financial Responsibility Composite Scores (FRCS), a metric built from three accounting ratios that measure liquidity, equity and profitability. Institutions that fail the composite score or fall into a warning range may face heightened cash monitoring, which requires colleges to disburse financial aid to students before being reimbursed.

These measures function as compliance and risk control mechanisms rather than predictive tools. They are based on historical financial data and use a limited number of pre-specified indicators. They are also based on information from institutions that is not standardized, as not all schools use the same fiscal year.

The usefulness of these measures is constrained by a lack of timely data, Robert Kelchen, head of the Department of Educational Leadership and Policy Studies at the University of Tennessee, Knoxville, and contributor to the model, said. Federal data is based on audited financial reports and submissions to the Integrated Postsecondary Education Data System (IPEDS). This means financial stress becomes visible in federal systems only after it has been building for years. For example, he said, IPEDS data from the fiscal year 2024 only just became available.


MACHINE LEARNING


Machine learning models can address this problem by incorporating many variables simultaneously and allowing the data to determine how those variables interact. Rather than assuming which financial indicators matter most, these models test combinations of enrollment trends, revenue sources, institutional size, sector and other characteristics to estimate closure risk.

In the researchers’ model, enrollment declined and reliance on tuition revenue emerged as particularly strong predictors of closure. Other factors, such as institutional size, sector (public, private, nonprofit or for-profit) and even missing data also contributed to risk assessments. The model was tested against historical closure events and demonstrated higher predictive accuracy than FRCS-based monitoring alone — with 83 percent average accuracy compared to the federal methods’ 77 percent.

According to Mabel, the advantage of machine learning is not just that it processes more data, but that it does not require researchers to specify in advance how different factors should relate to one another.

“The model itself is going to figure out and test all the different ways in which multiple types of information can work together to provide a more comprehensive and more holistic portrait of the likelihood of closing,” he said.
Abby Sourwine is a staff writer for the Center for Digital Education. She has a bachelor's degree in journalism from the University of Oregon and worked in local news before joining the e.Republic team. She is currently located in San Diego, California.