Affiliations: Human-Sciences and Technologies Advanced Research Institute, Stanford University, CA, USA. minchi@stanford.edu | School of Computing, Informatics and Decision Science Engineering, Arizona State University, AZ, USA. Kurt.Vanlehn@asu.edu | Department of Computer Science and Intelligent Systems Program and Learning Research and Development Center, University of Pittsburgh, Pittsburgh, PA, USA. litman@cs.pitt.edu | Learning Research and Development Center, University of Pittsburgh, Pittsburgh, PA, USA pjordan@pitt.edu
Abstract: Pedagogical strategies are policies for a tutor to decide the next action when there are multiple actions available. When the content is controlled to be the same across experimental conditions, there has been little evidence that tutorial decisions have an impact on students' learning. In this paper, we applied Reinforcement Learning (RL) to induce two sets of pedagogical policies from pre-existing human interaction data. The NormGain set was derived with the goal of enhancing tutorial decisions that contribute to learning while the InvNormGain set was derived with the goal of enhancing those decisions that contribute less or even nothing to learning. The two sets were then tested with human students. Our results show that when the content was controlled to be the same, different pedagogical policies did make a difference in learning and more specifically, the NormGain students outperformed their peers. Overall our results suggest that content exposure and practice opportunities can help students to learn even when tutors have poor pedagogical tutorial tactics. However, with effective tutorial tactics, students can learn even more.
Keywords: Reinforcement learning, human learning, intelligent tutoring systems, pedagogical strategy