150 likes | 251 Vues
Detecting the Learning Value of Items In a Randomized Problem Set. Zachary A. Pardos , Neil T. Heffernan. Worcester Polytechnic Institute Department of Computer Science. The Problem. What is the learning value of content in ITS? - does it promote learning?. Ways to find out: Run an RCE
E N D
Detecting the Learning Value of Items In a Randomized Problem Set Zachary A. Pardos, Neil T. Heffernan Worcester Polytechnic Institute Department of Computer Science
The Problem What is the learning value of content in ITS? - does it promote learning? • Ways to find out: • Run an RCE • Data mine responses (using this method) • 100s of Items of learning content • 1000s of students’ responses
Dataset • Student main problem responses (correct/incorrect) to 25 problem sets of 2,3 and 4 questions • Questions within a problem set relate to the same skill • 160-800 students completed each problem set in 2006-2007 school year data • 2,400 students total with 54,000 responses (14-16 year olds) • Questions in the problem sets were presented in a randomized order (required for this analysis) Main problem hint
Confound • Since only main question responses are being analyzed, the learning from the main question is confounded with the learning from the scaffolding and hints of the problem. • learning could be attributed to • The immediate feedback to the main problem of question 1 • The scaffolding of question 1 • Applying concepts from question 1 the next main problem Main problem hint
Model Parameters can be learned with the EM algorithm! .. ? • Modeling or measuring learning requires modeling knowledge • Knowledge Tracing used to model learning Parameters (probability of learning) (guess/slip) P(Skill: 0 → 1) P(Skill: 0 → 1) S S S Latent (skill knowledge) (dichotomous) P(correct| Skill = 0) P(incorrect| Skill = 1) Observables (question answers) incorrect correct correct
Model • Knowledge tracing assumes that learning rate is the same between each opportunity • Our model associates the learning rate with the particular problem that was encountered • Learning rate between opportunities are the same regardless of which problem the student saw Knowledge Tracing 0.12 0.12 Our Item Effect Model • Learning rates are an attribute of specific problems • Learning rates must be associated with problem for all permutations. 0.11 0.15
Model Three question sequence permutations modeled with shared Bayesian parameters Also known as Equivalence classes of CPTs (conditional probability tables )
Reliability measure • Data for a problem set randomly split into 20 equal size bins by student • Each bin was evaluated separately by the model • Binomial test used to estimate the probability of the null hypothesis, that each item is equally likely to have the highest learning rate • ie: binopdf(best_choice_mode,20,0.25) Item learning rates
Method Application Compute the learning rates of the three questions in the problem set Which one is BEST? Definition of BEST in this analysis: The question in a problem set that has the highest probability of learning.
Results • Problem sets with four questions were analyzed and the parameters of prior, guess/slip and learning rates were learned using the described method • The question with the highest probability of learning was identified • Method needs validation. Another method may report different results and reliability. • Ground truth of the parameters is necessary to validate the method and results
Simulation Validation • Since ground truth of learning rates in the real world are impossible to know, a simulation study was run • The simulation set a variety of values for the parameters of prior, guess/slip and learning rates and then simulated user responses • These responses could then be analyzed by the method using the same technique as was used on real data • An error analysis could done since the underlying simulation parameters of the data were known (did the method pick the right best question?) • Opportunity to learn what the method can & can’t do
Simulation Results • More students increases chance of a result • Larger learning difference between questions also increases the change of a result • Of the 160 experiments evaluated, 89 were reported as reliable (56%) • Of the 89 reported reliable results (using p < 0.05), seven were incorrect (7.8% FP)
Limitations • Only problem sets of five questions or less can be reasonably evaluated • Larger problem sets become intractable to compute due to the exponential increase in nodes and permutations as question count increases • for a four question set (4+4)*24 = 192 nodes • for a five question set (5+5) *120 = 1,200 nodes • Possible optimization is to only model the sequences for which there is data • Randomization of question order must be present to control for factors including problem difficulty and allow for detecting learning rates of all item pairs in the problem set
Contribution • No methods previous to this • Estimate learning rates per problem • Allows for the best (and worse) content to be identified without RCEs • Extends knowledge tracing to support randomization of problem order • Uses permutations of sequences to estimate stable Bayesian parameters with EM
Conclusions & Future Work • We think that this method, and ones built off of it, will facilitate better tutoring systems • Randomization gives many of the properties of a RCE. This method can perform a similar function but in the form of data mining to find what content works best • Method could be applied to aid in improving accuracy of question skill tagging