Chi - Induction of Adaptive Pedagogical Tutorial Tactics
Contents
Project Overview
The goal of this project is to investigate on application of Reinforcement Learning (RL) to derive adaptive pedagogical strategies directly from pre-existing interaction data. Pedagogical strategies are policies to decide the next system's action when there are multiple ones available. More specifically, this project is designed to: 1) help computer tutors employ effective, adaptive pedagogical policies; 2) test the viability of using RL, especially POMDP, to induce pedagogical policies, 3) show that pedagogical policies is a potential source of learning power for computer tutors to improve students' learning; and 4) explore the underlining causes of the effectiveness of the induced policies.
For any forms of learning environment including ITSs, the system's behaviors can be viewed as a sequential decision process wherein, at each discrete step, the system is responsible for selecting the next action to take. Each of these system decisions affects successive user's actions and performances. It is unclear how to make each decision effectively because its impact on learning cannot often be observed immediately and the effectiveness of one decision also depends on the effectiveness of subsequence decisions. Ideally, an effective learning environment should craft and adapt its actions to users' needs. However, there is no existing well-established theory on how to make these system decisions effectively. Most of existing ITSs, for example, either employ fixed pedagogical policies providing with little adaptability or employ hand-coded pedagogical rules that seek to implement existing cognitive or instructional theories. These theories may or may not have been well-evaluated.
In this project, we apply RL to improve the effectiveness of an ITS by induce pedagogical policies direct from a pre-existing student-computer interactivity data. More specifically, we focused on the two types of tutorial decisions: Elicit vs.Tell (ET) and Justify vs. Skip-Justify (JS). When making ET decisions the tutor decides whether to elicit the next step from the student or to tell them the step directly. The JS decisions address points where the tutor may optionally ask students to justify an answer they have taken or entry they have made. Neither type of decisions is well-understood in that there are many theories but no widespread consensus on how or when an action should be taken. Thus, we investigate on applying and evaluating RL to induce pedagogical tutorial tactics from pre-existing interactivity data.
Planned accomplishments for PSLC Year 6
Previously, we applied MDPs model to induce pedagogical policies from the data. However, a framework more suitable for this task is the Partially Observable Markov Decision Process (POMDP). POMDPs allow for realistic modeling of the students’ knowledge levels, the students’ intentions, and other hidden state components by incorporating them into the state space. POMDPs explicitly represent two sources of uncertainty: non-determinism in the control process and partial observability of the students’ knowledge levels. In the former case, outcomes of the tutorial actions or students’ knowledge level are not deterministic; in the latter, the underlying students’ knowledge levels are observed indirectly via incomplete or imperfect observations. The goal of year 6 is to explore the use of POMDP models. Additionally, we have been worked on investigating the causes of the effectiveness of the induced pedagogical tactics in the previous study.
Integrated Research Results and High Profile Publication
We focused on the two types of tutorial decisions: Elicit vs.Tell (ET) and Justify vs. Skip-Justify (JS). When making ET decisions the tutor decides whether to elicit the next step from the student or to tell them the step directly. The JS decisions address points where the tutor may optionally ask students to justify an answer they have taken or entry they have made. Neither type of decisions is well-understood in that there are many theories but no widespread consensus on how or when an action should be taken. Thus, we investigate on applying and evaluating RL to induce pedagogical tutorial tactics from pre-existing interactivity data.
Previously, a particular RL model, a Markov Decision Process (MDP), was applied to automatically derive adaptive pedagogical strategies directly from pre-existing student-computer interactivity data. The effectiveness of RL-induced tutorial tactics was then tested on real human subjects with random assignments. Results showed that after solving the same problems in the same amount of time, the induced pedagogical policies significantly improved students' learning gains up to about 60% compared with less effective pedagogical policies: t(55) = 3.058, p = 0.003, d = 0.81 (M = 0.41, SD = 0.19 for the Experimental Group and M = 0.25, SD = 0.21 for the Control group) (Chi et al., 2010a). Overall, our results showed that these fine-grain tutorial decisions indeed do matter to learning.
Moreover, the pedagogical policies employed by the Experimental group were derived from the log files from two pre-existing training corpora. Since the Experimental Group experienced the identical procedure and training materials as the students collected in the pre-existing training corpora, a post-hoc comparison was done among the three groups. Results showed that while no significant differences were found on pre-test score and time on training among the three groups, there were significant differences among the three groups on both post-test scores and NLG scores: F(2, 127) = 5.16, p = 0.007 and F(2, 127) = 7.57, p = 0.001 respectively (Chi et al., 2010b). More specifically, the Experimental Group significantly out-performed the two previous groups in terms of posttest scores and NLG. This result suggested that RL can be fruitfully applied to induce more effective, adaptive pedagogical strategies from less effective pre-existing data.
Compared with previous studies on applying RL to induce pedagogical policies on ITSs, this project so far has at least two major contributions. First, we showed that using a relatively small exploratory corpus as training corpus for inducing pedagogical policies is a feasible approach. Second, we empirically showed that the RL induced policies indeed made students learn deeper or better. Moreover, while much of previous research on applying RL to ITSs and non-tutoring NL Dialogue systems used pre-defined state representation, our approach in this project is to begin with a large set of features to which a series of feature-selection methods were applied to reduce them to a tractable subset. In this paper, we will shed some lights on the relative effectiveness of different feature selection methods and which features among the ones defined were most involved in the final induced policies.
Year 6 Project Deliverables
- Develop code and human-computer interfaces for applying, comparing and interpreting cognitive model discovery algorithms across multiple data sets in DataShop.
- Demonstrate the use of the model discovery infrastructure for at least two discovery algorithms applied to at least 4 DataShop data sets.
- For at least one of these data sets, work with associated researchers to perform a “close the loop” experiment whereby we test that a better cognitive model leads to better or more efficient student learning.
6th Month Milestone
By March, 2010 we will 1) be able to run the LFA algorithm on PSLC data sets from the DataShop web services, 2) have run model discovery with using at least one algorithm on at least two data sets, and 3) we will have designed and ideally run the close-the-loop experiment.