A comprehensive review of the current math placement model and exploration of model improvement utilizing historical student data
Review the accuracy of Montana State University's current math placement model and explore model improvement using historical student data and supervised machine learning models to more accurately place incoming students into math courses.
Driven by the pandemic, higher education institutions have stopped requiring placement tests when students apply to college. This has led to fewer incoming students providing ACT or SAT scores (as low as 20% at MSU). Given that these scores have been historically used to place students, this has caused a need for other student success proxies to be used to place students into first-year math and writing courses. This project reviews the accuracy of the current math placement model at Montana State University and then looks to use other models to place students more accurately in courses. The overall goal is to boost student success, retention, and graduation rates, recognizing the crucial role of accurate course placement in a student's educational journey and future success.
The flow chart to the left lays out the current math placement model at MSU. The boxes to the left indicate ways incoming students can be placed into seven possible course levels. Placement into different levels is driven by a student's ACT, SAT, ACT & HS GPA, SAT & HS GPA, or Ed Ready scores (MSU's online testing platform). Once students are placed into a course level, they register for the correct course based on their degree requirements (courses noted in the boxes to the left).
The accuracy of the placement model was assessed by categorizing student grade outcomes as successful if a student received a B- or higher and unsuccessful if they received a C+ or lower (percentages noted near arrows). This data, a tangible measure of the model's accuracy, was then used to calculate percentages. The chart to the left displays the outcomes at each course level, revealing higher success rates in higher-level courses. The lowest level of success was seen in 250-level courses, with 44%, and 300-level courses, with 54%.
To review the accuracy of the current model, f-scores were calculated by using the following logic: students who are compliant and unsuccessful are considered a false negative, students who are uncompliant and successful are a false positive, students who are compliant and successful are a true positive, and lastly students that are compliant and unsuccessful are a true negative. Based on this logic, the table below shows the corresponding f-scores at each course level. The lowest f-scores are seen in 250 and 290-level courses. An overall f-score for the current model of .90 indicates that the current model is performing well in both precision and recall. It is worth noting that there are many more instances where students were complaint vs. uncompliant, so this measure of f-score may be skewed to a higher score, as false positives and false negatives are factored into the denominator in calculating precision and recall.
Note that for all machine learning models, the tables below depict the f-scores calculated at the course level for each input variable and an overall f-score for each variable and variable combination model explored. These scores show the accuracy of the models holistically and at the course level.
Since the outcome variable for this project is whether a student was successfully placed in a course, the first model explored to improve student placement was a logistic regression. Logistic regressions predict binary outcomes, identifying the relationship between a dependent variable and multiple independent variables.
Using logistic regression to forecast student success based on variables like placement scores and high school GPA resulted in models that, while statistically significant in some cases, displayed low pseudo-R-squared values across various course levels. This suggests that, despite a certain level of statistical significance indicating a better fit than the null model, the actual explanatory power of the models still needs to be improved. Essentially, the models confirm a relationship between the chosen variables and student success, yet they capture only a tiny fraction of the variance in outcomes. The low pseudo-R-squared values indicate that other unconsidered factors may significantly influence student success or that the relationships between the variables and success still need to be fully captured by logistic regression in its current form.
Intending to place students at the correct course level accurately, classification models were explored since they efficiently categorize data. The first classification model reviewed was a decision tree. Decision trees work by repeatedly splitting the data into subsets based on the values of the input variables and creating a tree. At each node, a decision is made based on the value of a particular input variable, and the data is split into two or more subsets. F-scores were calculated for each course level and then an overall f-score. The f-scores at the individual level indicate the model's accuracy at the course level and then the overall f-scores indicate the accuracy of the entire model.
Interestingly, the decision tree models revealed that Ed Ready scores are more accurate in course placement than SAT and ACT scores. The model incorporating both Ed Ready and high school GPA achieved the highest single variable f-score of .47. At the same time, SAT scores proved less effective, with an f-score of .30. Further analysis of the decision tree model suggests that while Ed Ready is the most accurate variable in placing students, high school GPA offers additional accuracy with an f-score of .47. This was seen particularly for 100, 250 and 300 level courses. In reviewing the visualization of this model, the course placements shown have slightly different breakpoints from the current placement model. This suggests that there may be benefits in adjusting the score thresholds for 250 and 300-level courses to improve student success rates, which currently linger around .44 to .54 in effectiveness.
After reviewing the outcomes of the decision tree, a random forest model was explored. Random forests aggregate many decision trees, which could lead to improved placement. We saw improvement in the model's overall accuracy except when using Ed Ready scores and high school GPA as the input variables. A f-score of .47 was seen in the decision tree model and only .45 in the random forest model.
Ed Ready scores alone appeared to have the highest accuracy at .47. However, in reviewing, accuracy in student placement is 0 in 150 and 290. Depending on the course, different variables appear to do better at placing students. For example, at 100 course levels, Ed Ready scores seem more accurate in placing students. In contrast, Ed Ready Scores at the 500 level appear less precise in placing students than other variables. Ed Ready Scores are the most successful in predicting student course placement. Even with improvements to the accuracy of student placement seen in the random forest model, the overall accuracy of student placement is lower than the current success levels seen using the current math placement model.
Lastly, to see if the accuracy of student placement could be further improved, another classification model was used, in this case, XG boost. Like the random forest model, it builds multiple decision trees; however, each tree corrects the errors of the previous tree, and then an algorithm is used to minimize loss when adding new trees. Here, slight improvements to accuracy were seen in all input variables except for Ed Ready scores, where the accuracy stayed the same with an f score of .47. This model appears to have higher levels of accuracy where we see seven f scores above .6. In contrast only 5 of scores above .6 were seen in the random forest output and only 4 in the classification tree output. Even though this is the best statistical model for placing students, the overall placement accuracy is again lower than students' success with the current placement model.
F-scores were calculated using the flow chart model to compare the current and machine learning models. This calculation was done by determining if a student complied with the model’s constraints and whether the student succeeded in the course. It's important to note that the F-scores, derived from the flow chart model, serve as a benchmark rather than a direct outcome from a statistical model. Consequently, while these F-scores provide valuable points of reference, they should not be interpreted as precise measures of model accuracy like F-scores calculated from predictive modeling techniques. It is also important to note that these scores are likely skewed to the high side as there are many more instances where students were complaint vs. uncompliant, so this measure of f-score may be skewed to a higher score, as false positives and false negatives are factored into the denominator in calculating precision and recall.
In analyzing the success rate of the current model and the outcomes of the logistic regression and classification tree models, it is worth noting the limitations of the data. The currently available data is based on historical student placement, determined by the current model's thresholds. The math department allows students to improve their original scores to be placed into a higher course level. Since students can enhance their Ed Ready score by doing additional work in the system, the historical student placement does not always follow the current model’s placement criteria. Since we have chosen only to include students who adhere to the model placement, this analysis does not include students placed into a higher-level math course.
The other major limitation of the data is the volume of available input data at any given course level (i.e., ACT, SAT, high school GPA, and Ed Ready scores). Certain variables are used to predict student placement at different course levels, which leads to the data being separated into different course levels and input variables. This can lead to the historical data in each course level placement becoming sparse, creating small sample sizes in specific courses and variable types, which can cause the outcome of the analysis to be misleading. This is due to more minor data sets being swayed by outliers, which can ultimately lead to predictions and outcomes that are not accurate in the models. Also, since the historical data only includes historical information on MSU students, the modeling is complex and would likely be irrelevant to other universities. If these models would like to be explored in different institutions, they would likely need to use their datasets against their current placement methodology.
Adjusting the current model's thresholds based on the classification tree model's insights is advised to enhance the accuracy of student course-level placement and subsequently increase student success rates. Specifically, the Ed Ready score thresholds for placing students into 250-level courses should be raised from 25 to 27 and 300-level courses from 30 to 35. Furthermore, incorporating high school GPA alongside Ed Ready scores, placing students into 300-level courses with an Ed Ready score of 30 and a GPA of 3.5 or higher and into 400-level courses with an Ed Ready score of 35 and a GPA of 3.5 or higher could refine the placement process. If recommendations are implemented, then continued evaluation of the model's performance is suggested to ensure improvement of math placement accuracy and allow for adjustments to the model as needed.
Ensuring first-year students are accurately placed into courses that reflect their competency levels is crucial for fostering academic success, reducing dropout rates, and facilitating timely graduation. The proposed adjustments to the placement model, including revising Ed Ready score thresholds and incorporating high school GPA, are grounded in a detailed analysis of student performance data. By implementing these recommendations and continuously reassessing the model with new student data each semester, it's possible to create a dynamic and responsive placement process. This approach aims to enhance student achievement in the short term and supports broader goals of increasing retention and graduation rates, thereby improving overall satisfaction with the educational experience.
Contact brittany.thompson5.bt@gmail.com to get more information on the project