Potential benefit: Understanding factors related to scores can improve model clarity and later inform real world, focused outreach.
Ethical & equity considerations: As an Autistic researcher and Early Intervention/Early Childhood Special Education (EI/ECSE) specialist, I want to understand how predictive models are created, what their limitations are, and the ethical considerations.
Categorical (gender f/m, jaundice yes/no, family history of Autism yes/no, used app before yes/no, results from screener results YES (6+) NO (~5), ethnicity, country of residence)
AQ-10 screener
Allison et al. (2012) evaluated shortened 10-item versions of the Autism Spectrum Quotient (AQ-50) as quick screening tools for Autism. Using data from over 1,000 Autistic individuals and 3,000 controls, the short forms showed high sensitivity (accurately identifying most people with Autism) and high specificity (correctly excluding those without).
Before building the models, it’s important to review the Ordinary Least Squares (OLS) assumptions for linear regression (Ch. 4 & 6, Boehmke, & Greenwell, 2019):
Linearity: There should be a roughly linear relationship between predictors and outcome variables.
Sample size: The number of observations (n) should be larger than that of the predictors (p).
Multicollinearity: The independent variables cannot be highly correlated to each other (p. 269, Vogt & Johnson, 2016).
Modeling assumptions
To reduce overfitting, the models used 10-fold cross-validation, providing more reliable measure of generalizability.
Model 1: Ridge Regularization model
plot(ridge)
ridge$bestTune
alpha lambda
49 0 0.49
Model 1: Ridge Regularization model
Held-out test metrics
Model
R2
MAE
RMSE
Baseline (mean-only)
NA
2.263
2.601
Ridge (10-CV, keep_screen=FALSE)
0.309
1.886
2.233
Model 2: Lasso Regularization model
alpha lambda
6 1 0.005
Model 2: Lasso Regularization model
Held-out test metrics: Baseline vs Ridge vs Lasso
Model
R2
MAE
RMSE
Baseline (mean-only)
NA
2.263
2.601
Ridge (10-CV, keep_screen=FALSE)
0.309
1.886
2.233
Lasso (10-CV, keep_screen=FALSE)
0.281
1.880
2.241
Model 3: Elastic Net Regression model
plot(elastic)
elastic$bestTune
alpha lambda
26 0 0.09249147
Evaluating model performance
The Elastic Net model chose alpha = 0, meaning it behaved like Ridge. With many correlated demographic features, this gentle ridge-style shrinkage worked better than LASSO’s stronger variable-dropping penalty. Ridge and Elastic Net performed the best (R² = 0.309), and while LASSO was slightly lower, all three outperformed the mean-only baseline.
Held-out test metrics: Baseline vs Ridge vs Lasso vs Elastic Net (sorted by R²)
Model
R2
MAE
RMSE
Ridge (10-fold CV)
0.309
1.886
2.233
Elastic Net (10-fold CV)
0.309
1.886
2.234
Lasso (10-fold CV)
0.281
1.880
2.241
Baseline (mean-only)
-0.004
2.263
2.601
Model fit - Final model
Model efficiency & effectiveness
ggplot(comp_long, aes(x = Metric, y = Value, fill = Model)) +geom_col(position ="dodge") +labs(title ="Model Performance by Regularization",x ="Metric", y ="Value") +theme_minimal() +theme(plot.title =element_text(hjust =0.5))
Model fit - Final model
Winner (Ridge) 🥇
Model fit - Cut off point & other considerations
The AQ-10 uses a cut-off score of 6, but I chose not to create a binary outcome. Modeling the continuous score offered more nuance and reduced the risk of oversimplifying an already limited dataset.
The dataset is highly skewed. Because I do not know who was able to or motivated to access this online screener, the sample is not representative of a broader population at all.
Given these limitations, linear regression was more informative for exploring patterns.
What I learned: Machine Learning
As with any method, the quality of data determines the quality of the outcome. In ML, this is more critical because it is easy to run models with limited understanding of the datasets or the variables.
As Gould et al. (2023) noted, “To… effectively eliminate health disparities requires recognition of the subjectivity of data and of the power of data to dictate and reinforce narratives, accompanied by intentional reform of data practices” (p. 12).
What I learned: Variables & Findings
Participants from the U.S., Canada, and Brazil showed higher positive correlations, while those from the UAE, India, South Asia, and New Zealand showed lower ones. Because reported Autism diagnosis rates are much higher in the U.S. and Canada than in Brazil (World Population Review, 2025), Brazil’s similarity to these countries was unexpected.
This likely reflects who chose to participate rather than true population patterns, since the sample is based on self-selection. Without knowing participants’ motivations or access factors, these results are difficult to interpret.
Family history of Autism was a meaningful predictor, but findings like these must never be misread as suggesting that certain racial/ethnic or national groups are “more likely” to be Autistic.
Other considerations
The AQ was developed using adults in the UK who spoke English, of unknown race, with men overrepresented (Baron-Cohen et al., 2001). This makes the norming very skewed.
I struggled with the ethical implications of applying ML. It is unclear if participants consented to secondary use of data. Conducting these models requires not only methodological care but also ethical reflection about data ownership, economical/environmental impact, and the risks of misinterpretation.
Conclusion
“…all model-building efforts are constrained by the existing data” (Kuhn & Johnson, 2016, p. 61), and equally shaped by the assumptions and interpretive choices researchers make throughout the analytic process.
References
Allison, C., Auyeung, B., & Baron-Cohen, S. (2012). Toward brief “Red Flags” for Autism screening: The short Autism Spectrum Quotient and the short Quantitative Checklist in 1,000 cases and 3,000 controls. Journal of the American Academy of Child & Adolescent Psychiatry, 51(2), 202-212.e7.
Baron-Cohen, S., Wheelwright, S., Skinner, R., Martin, J., & Clubley, E. (2001). The Autism-Spectrum Quotient (AQ): Evidence from Asperger Syndrome/High-functioning autism, males and females, Scientists and Mathematicians, 31(1), 5-18.
Boehmke, B. & Greenwell, B. (2019). Hands on Machine Learning with R. Taylor & Francis.
Gould, L. H., Farquhar, S. E., Greer, S., Travers, M., Ramadhar, L., Tantay, L., Gurr, D., Baquero, M., & Vasquez, A. (2023). Data for equity: Creating an antiracist, intersectional approach to data in a local health department. Journal of Public Health Management and Practice, 29(1), 11–20. https://doi.org/10.1097/PHH.0000000000001579