Prototype predictive model for learner disengagement risk using EEG brain-wave data and learner surveys. The system applies dimensionality reduction and lightweight machine learning models to link cognitive load with engagement and provide adaptive learning triggers.
- Built a machine learning pipeline to detect learner disengagement by combining EEG-derived features with self-reported survey signals.
- Achieved strong predictive performance (≈ 80%+ accuracy on held-out data) while maintaining interpretability for non-technical stakeholders.
- Identified early warning indicators of disengagement that can trigger adaptive learning interventions in real time.
- Designed the system with fairness and transparency in mind, aligning with enterprise training constraints.
git clone https://github.com/<org-or-user>/neuroscience-driven-engagement.git
cd neuroscience-driven-engagementpython -m venv venv
source venv/bin/activate # macOS/Linux
venv\Scripts\activate # Windowspip install -r requirements.txt- EEG feature files and anonymized survey data are stored in the /data directory.
- Due to privacy constraints, raw EEG signals are not included.
- Processed feature matrices are provided for reproducibility.
jupyter notebookOpen the notebooks in the following order:
- 01_data_exploration.ipynb
- 02_preprocessing_feature_engineering.ipynb
- 03_model_training.ipynb
- 04_evaluation_and_fairness.ipynb
This project was developed to explore how neuroscience-informed data can improve learning outcomes in professional training environments. Our primary goal was to identify disengagement during technical learning sessions and recommend adaptive strategies before learners fully disengage. By combining EEG-based attention indicators with survey responses, we aimed to move beyond traditional completion metrics and capture engagement as it happens. This system can be contributed to reducing training drop-off, improve learner retention, and personalize instruction at scale.
- EEG-derived features: numerical indicators capturing attention, cognitive load, and signal variability.
- Learner surveys: Likert-scale and categorical responses reflecting perceived engagement, difficulty, and fatigue.
- Cleaned missing and noisy values using median imputation and normalization.
- Explored correlations between EEG features and self-reported engagement.
- Encoded categorical survey responses and aligned timestamps across modalities.
- Certain EEG features showed strong correlation with self-reported disengagement.
- Engagement drops often occurred before learners explicitly reported difficulty, suggesting predictive potential.
- Limited sample size required careful validation.
- EEG features were assumed to be representative proxies for attention rather than direct cognitive states.
- Tested multiple models including Logistic Regression, Random Forests, and Decision Trees.
- Selected interpretable models to support stakeholder trust and explainability.
- Applied feature selection and hyperparameter tuning using cross-validation.
- 70/30 train-test split
- Primary metrics: Accuracy and F1-score
- Baseline compared against majority-class prediction
- The final model achieved ~80% accuracy with balanced precision and recall.
- EEG features related to sustained attention were among the strongest predictors.
- Fairness checks showed consistent performance across demographic subgroups, with no significant bias detected.
- These results demonstrate that neuroscience-informed features can meaningfully enhance engagement prediction when paired with traditional survey data.
The Streamlit dashboard serves as the interactive interface for the Neuroscience-Driven Learning Engagement System (Challenger Insight). It allows users to explore engagement predictions, EEG feature trends, and risk signals in a clear, interpretable format.
- Engagement Risk Score: Displays predicted disengagement risk based on EEG-derived features and survey inputs.
- EEG Feature Visualizations: Interactive plots showing attention, cognitive load, and signal variability trends.
- Model Output Transparency: Simple explanations of model predictions to support non-technical stakeholders.
- Adaptive Trigger Indicators: Highlights when engagement risk crosses predefined thresholds that would prompt learning interventions.
These visual components are designed to support real-time monitoring and future integration into enterprise learning platforms.
- Expand the dataset to include longer training sessions and more participants.
- Explore temporal models (e.g., LSTMs) to capture engagement trends over time.
- Integrate the model into a live dashboard to trigger real-time adaptive learning interventions.
- Incorporate additional explainability tools to further support transparency.
| Name | GitHub Handle | Contribution |
|---|---|---|
| Meagan Alfaro | @meaganalfaro | Model development, prototype implementation, hyperparameter tuning, documentation |
| Ishrat Arshad | @ishratarshad | Model development/validation, data preprocessing, exploratory data analysis (EDA), dataset documentation |
| Ruby Hong | @rubysmac | Model development/validation, data preprocessing, exploratory data analysis (EDA), dataset documentation |
| Aneesa Ayornu | @aeza404 | Data preprocessing, feature engineering, data validation |
| Mahathi Chinthapalli | @MahathiChinthapalli | Model selection, hyperparameter tuning, training and optimization |
| Amanda Thomas | @amandathom | Model evaluation, hyperparameter tuning, performance analysis, results interpretation |
| Izabella Doser | @Izabelladesign | Data exploration, dimensionality reduction, model development, documentation |
| Arian Bahram | @ariansbahram | Data preprocessing, exploratory data analysis (EDA), model development |
| Saket Kolluru | @Saketk2 | Dashboard development, data visualization, results interpretation |
| Eden Radulescu | @edenradulescu | Dashboard development, data visualization, results interpretation |
This project is licensed under the MIT License (pending Challenge Advisor approval).
- Break Through Tech AI Program materials
- Selected literature on EEG-based engagement detection and learning analytics
We thank Chris Chambers (Chambers Capital Ventures) for guiding the problem scope, Lori Kirkland for introducing the real-world challenge context, and Rashidah Carr, our AI Studio Coach, for continuous feedback and support throughout the semester.


