Shayan Doroudi
Carnegie Mellon University
Prior Work:
My Approach:
Combine aspects of both for greater impact.
I focus on scalable automated content selection, using methods that combine machine learning, human computation, and principles from the learning sciences
I focus on scalable automated content selection, using methods that combine machine learning, human computation, and principles from the learning sciences
Cognitive Mastery Learning
Overconstrained
Reinforcement Learning (RL)
Underconstrained
Corbett and Anderson, 1994
Cognitive Mastery Learning
Heuristic Mastery Learning
Corbett and Anderson, 1994
Kelly, Wang, Thompson, and Heffernan, 2015
Use student response data to find the optimal adaptive content selection policy.
Use student response data to find the optimal adaptive content selection policy.
Use student response data to find the optimal adaptive content selection policy.
Use student response data to find the optimal adaptive content selection policy.
Use student response data to find the optimal adaptive content selection policy.
Corbett and Anderson, 1994
Perceived Mastery State
Actual Mastery State
We may give fewer practice opportunities to students than they actually need to reach mastery.
Doroudi and Brunskill, EDM 2017, Best Paper Nominee
| Baseline | Adaptive Policy | |
|---|---|---|
| Simulated Results | 5.9 ± 0.9 | 9.1 ± 0.8 |
Doroudi, Aleven, and Brunskill, L@S 2017
| Baseline | Adaptive Policy | |
|---|---|---|
| Simulated Results | 5.9 ± 0.9 | 9.1 ± 0.8 |
| Experimental Results | 5.5 ± 2.6 | 4.9 ± 1.8 |
Doroudi, Aleven, and Brunskill, L@S 2017
Doroudi, Thomas, and Brunskill, UAI 2017, Best Paper
| Policy 1 | Policy 2 | Policy 3 | |
|---|---|---|---|
| Student Model 1 | |||
| Student Model 2 | |||
| Student Model 3 |
\(V_{SM_1,P_1}\)
\(V_{SM_2,P_1}\)
\(V_{SM_3,P_1}\)
\(V_{SM_1,P_2}\)
\(V_{SM_2,P_2}\)
\(V_{SM_3,P_2}\)
\(V_{SM_1,P_3}\)
\(V_{SM_2,P_3}\)
\(V_{SM_3,P_3}\)
Baseline |
Adaptive Policy |
|
|---|---|---|
G-SCOPE Model |
5.9 ± 0.9 |
9.1 ± 0.8 |
Doroudi, Aleven, and Brunskill, L@S 2017
|
Baseline |
Adaptive Policy |
|
|---|---|---|
|
G-SCOPE Model |
5.9 ± 0.9 |
9.1 ± 0.8 |
|
Bayesian Knowledge Tracing |
6.5 ± 0.8 |
7.0 ± 1.0 |
Doroudi, Aleven, and Brunskill, L@S 2017
|
Baseline |
Adaptive Policy |
|
|---|---|---|
|
G-SCOPE Model |
5.9 ± 0.9 |
9.1 ± 0.8 |
|
Bayesian Knowledge Tracing |
6.5 ± 0.8 |
7.0 ± 1.0 |
|
Deep Knowledge Tracing |
9.9 ± 1.5 |
8.6 ± 2.1 |
Doroudi, Aleven, and Brunskill, L@S 2017
|
Baseline |
Adaptive Policy |
Awesome Policy |
|
|---|---|---|---|
|
G-SCOPE Model |
5.9 ± 0.9 |
9.1 ± 0.8 |
16 |
|
Bayesian Knowledge Tracing |
6.5 ± 0.8 |
7.0 ± 1.0 |
16 |
|
Deep Knowledge Tracing |
9.9 ± 1.5 |
8.6 ± 2.1 |
16 |
Doroudi, Aleven, and Brunskill, L@S 2017
Let students construct their own understandings.
Share those constructions with future students to help them co-construct their understandings.
Building on recent literature on learnersourcing (Kim et al., 2015).
Advantages:
Can create content in a more cost-effective way.
Both content creators and content consumers can learn from this content.
Students can engage with content in new, authentic ways.
Complex Crowdsourcing Tasks (Web Search)
Introduction to Mathematical Thinking MOOC
Doroudi, Kamar, Brunskill, and Horvitz, CHI 2016
Doroudi, Kamar, Brunskill, and Horvitz, CHI 2016
Doroudi, Kamar, Brunskill, and Horvitz, CHI 2016
Doroudi, Kamar, Brunskill, and Horvitz, CHI 2016
# Workers |
Accuracy |
|
|---|---|---|
Control |
150 |
0.50 ± 0.27 |
Reviewing Expert Examples |
140 |
0.61 ± 0.26 |
Validating Peer Solutions |
107 |
0.56 ± 0.26 |
Doroudi, Kamar, Brunskill, and Horvitz, CHI 2016
Experiment I
Doroudi, Kamar, Brunskill, and Horvitz, CHI 2016
* Workers see one short (<800 char) and one long (>800 char) solution
Experiment II
# Workers |
Accuracy |
|
|---|---|---|
Reviewing Expert Examples |
102 |
0.59 ± 0.26 |
Validating Peer Solutions |
95 |
0.58 ± 0.23 |
Validating Filtered Solutions* |
88 |
0.60 ± 0.25 |
Experiment II
# Workers |
Accuracy |
|
|---|---|---|
Reviewing Expert Examples |
102 |
0.59 ± 0.26 |
Validating Peer Solutions |
95 |
0.58 ± 0.23 |
Validating Filtered Solutions* |
88 |
0.60 ± 0.25 |
|
Extra Filtered Solutions** |
34 |
0.74 ± 0.17 |
* Workers see one short (<800 char) and one long (>800 char) solution
** Post Hoc analysis of workers who saw one medium (500-800 char)
and one extra long (>1000 char) solution
Doroudi, Kamar, Brunskill, and Horvitz, CHI 2016
Keith Devlin, Introduction to Mathematical Thinking
Evaluate the following proofs
Read the following proofs and then construct your own
Logically correct
Logically correct
Good reasons
Clear
Good opening
Good conclusion
Williams, Kim, Rafferty, Maldonado, Gajos, Lasecki, and Heffernan, 2016
Williams, Kim, Rafferty, Maldonado, Gajos, Lasecki, and Heffernan, 2016
\(p_j\) is the probability of answering a question correctly after seeing peer solution \(j\).
\(\mathcal{F}\) is the set of features. Could include:
Solution length
Number of symbols/urls
Bag of words representation
Score on rubric items
Shown to be a good sequence according to cognitive load theory (Kalyuga, 2003; Kalyuga and Sweller, 2005)
Review the following example
Fill in the missing steps
Answer the following question
Adapt instruction based on cognitive efficiency
(Kalyuga and Sweller, 2005; Najar, Mitrovic, and McLaren, 2016; Salden, Aleven, Schwonke, and Renkl 2010)
If CE > T1
If CE > T2
Review the following example
Answer the following question
Fill in the missing steps
Based on cognitive load theory, hypothesize that this would be a good sequence
Review the following example
Answer the following question
Evaluate the following proof
Use reinforcement learning to learn the appropriate transition points
Review the following example
Answer the following question
Evaluate the following proof
If CE > T1
If CE > T2
If CE < T3
If CE < T4
By January 2018:
Complete content curation experiments for web search tasks and submit to IJCAI.
By March 2018:
Complete review of RL approaches to adaptive content selection and submit paper to JEDM.
Complete experiment testing different learnersourcing activities on MOOC and submit as Work-in-Progress to L@S.
By May 2018:
Complete initial activity sequencing experiment on MOOC.
By August 2018:
Complete adaptive sequencing experiment on MOOC.
Start writing dissertation.
By December 2018:
Replicate any interesting findings.
Complete dissertation.
I have proposed a number of methods for scalable automated adaptive content selection that combine machine learning, human computation, and principles from the learning sciences.
My work demonstrates how both:
Insights from computer science and statistics can inform the learning sciences.
Insights from the learning sciences can guide computational approaches.
Ed Tech
Machine Learning
My approach
Rule-Based AI
Statistical Machine Learning
Combine aspects of both
Adaptive Content Selection Approaches
AI
Approaches
The research reported here was supported, in whole or in part, by the Institute of Education Sciences, U.S. Department of Education, through Grants R305A130215 and R305B150008 to Carnegie Mellon University. The opinions expressed are those of the authors and do not represent views of the Institute or the U.S. Dept. of Education. Some of the research was also funded with the generous support of Microsoft Research
I am fortunate to have worked on (and continue to work on) much of the research presented here with a number of collaborators including Emma Brunskill, Vincent Aleven, Ken Holstein, Phil Thomas, Ece Kamar, Eric Horvitz, Minsuk Chang, Juho Kim, and Keith Devlin.
Special thanks to my thesis committee members,
Emma Brunskill, Vincent Aleven, Ken Koedinger, Chinmay Kulkarni, and Eric Horvitz
as well as Sharon Carver and David Klahr