1 Learning from Learning Curves: Item Response Theory & Learning Factors Analysis Ken Koedinger Human-Computer Interaction Institute Carnegie Mellon University.

Slides:



Advertisements
Similar presentations
Modelling with expert systems. Expert systems Modelling with expert systems Coaching modelling with expert systems Advantages and limitations of modelling.
Advertisements

Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 January 23, 2012.
Learning Factors Analysis – A General Method for Cognitive Model Evaluation and Improvement Hao Cen, Kenneth Koedinger, Brian Junker Human-Computer Interaction.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
The Network of Dynamic Learning Communities C 107 F N Increasing Rigor February 5, 2011.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Educational data mining overview & Introduction to Exploratory Data Analysis Ken Koedinger CMU Director of PSLC Professor of Human-Computer Interaction.
1 LearnLab: Bridging the Gap Between Learning Science and Educational Practice Ken Koedinger Human-Computer Interaction & Psychology, CMU PI & CMU Director.
Improving learning by improving the cognitive model: A data- driven approach Cen, H., Koedinger, K., Junker, B. Learning Factors Analysis - A General Method.
Machine Learning in Practice Lecture 7 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Model Assessment, Selection and Averaging
An Individualized Web-Based Algebra Tutor D.Sklavakis & I. Refanidis 1 An Individualized Web-Based Algebra Tutor Based on Dynamic Deep Model Tracing Dimitrios.
Supporting (aspects of) self- directed learning with Cognitive Tutors Ken Koedinger CMU Director of Pittsburgh Science of Learning Center Human-Computer.
Learning from Learning Curves: Item Response Theory & Learning Factors Analysis Ken Koedinger Human-Computer Interaction Institute Carnegie Mellon University.
x – independent variable (input)
Data mining with DataShop Ken Koedinger CMU Director of PSLC Professor of Human-Computer Interaction & Psychology Carnegie Mellon University Ryan S.J.d.
Project  Now it is time to think about the project  It is a team work Each team will consist of 2 people  It is better to consider a project of your.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
+ Doing More with Less : Student Modeling and Performance Prediction with Reduced Content Models Yun Huang, University of Pittsburgh Yanbo Xu, Carnegie.
Educational Data Mining Overview John Stamper PSLC Summer School /25/2011 1PSLC Summer School 2011.
Part I: Classification and Bayesian Learning
A Value-Based Approach for Quantifying Scientific Problem Solving Effectiveness Within and Across Educational Systems Ron Stevens, Ph.D. IMMEX Project.
Educational data mining overview & Introduction to Exploratory Data Analysis with DataShop Ken Koedinger CMU Director of PSLC Professor of Human-Computer.
Science and Engineering Practices
Chapter 5 Data mining : A Closer Look.
Classification and Prediction: Regression Analysis
DataShop: An Educational Data Mining Platform for the Learning Science Community John Stamper Pittsburgh Science of Learning Center Human-Computer Interaction.
Educational Data Mining and DataShop John Stamper Carnegie Mellon University 1 9/12/2012 PSLC Corporate Partner Meeting 2012.
1 Prediction of Software Reliability Using Neural Network and Fuzzy Logic Professor David Rine Seminar Notes.
PSLC DataShop Introduction Slides current to DataShop version John Stamper DataShop Technical Director.
Unit 2: Engineering Design Process
Applying the Multimedia Principle: Use Words and Graphics Rather than Words Alone Chapter 4 Ken Koedinger 1.
Tuteurs cognitifs: La théorie ACT-R et les systèmes de production Roger Nkambou.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 April 2, 2012.
Learning Progressions: Some Thoughts About What we do With and About Them Jim Pellegrino University of Illinois at Chicago.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
Data Shop Introduction Ken Koedinger & Alida Skogsholm Human-Computer Interaction Institute Carnegie Mellon University.
Evidence-based Practice Chapter 3 Ken Koedinger Based on slides from Ruth Clark 1.
Noboru Matsuda Human-Computer Interaction Institute
Educational Data Mining: Discovery with Models Ryan S.J.d. Baker PSLC/HCII Carnegie Mellon University Ken Koedinger CMU Director of PSLC Professor of Human-Computer.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Jennifer Lewis Priestley Presentation of “Assessment of Evaluation Methods for Prediction and Classification of Consumer Risk in the Credit Industry” co-authored.
CROSS-VALIDATION AND MODEL SELECTION Many Slides are from: Dr. Thomas Jensen -Expedia.com and Prof. Olga Veksler - CS Learning and Computer Vision.
CONFIDENTIAL1 Hidden Decision Trees to Design Predictive Scores – Application to Fraud Detection Vincent Granville, Ph.D. AnalyticBridge October 27, 2009.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
711: Intelligent Tutoring Systems Week 1 – Introduction.
Applying the Redundancy Principle ( Chapter 7) And using e-learning data for CTA Ken Koedinger 1.
Data mining with DataShop Ken Koedinger CMU Director of PSLC Professor of Human-Computer Interaction & Psychology Carnegie Mellon University.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
RULES Patty Nordstrom Hien Nguyen. "Cognitive Skills are Realized by Production Rules"
1 The Software Development Process ► Systems analysis ► Systems design ► Implementation ► Testing ► Documentation ► Evaluation ► Maintenance.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
A M ACHINE L EARNING A PPROACH FOR A UTOMATIC S TUDENT M ODEL D ISCOVERY Nan Li, Noboru Matsuda, William Cohen, and Kenneth Koedinger Computer Science.
Core Methods in Educational Data Mining HUDK4050 Fall 2015.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 January 25, 2012.
Using DataShop Tools to Model Students Learning Statistics Marsha C. Lovett Eberly Center & Psychology Acknowledgements to: Judy Brooks, Ken Koedinger,
Core Methods in Educational Data Mining HUDK4050 Fall 2014.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 February 6, 2012.
2011 Data Mining Industrial & Information Systems Engineering Pilsung Kang Industrial & Information Systems Engineering Seoul National University of Science.
Core Methods in Educational Data Mining
Big Data, Education, and Society
Addressing the Assessing Challenge with the ASSISTment System
Overview of Machine Learning
The Behavior of Tutoring Systems
Neil T. Heffernan, Joseph E. Beck & Kenneth R. Koedinger
Core Methods in Educational Data Mining
Core Methods in Educational Data Mining
Presentation transcript:

1 Learning from Learning Curves: Item Response Theory & Learning Factors Analysis Ken Koedinger Human-Computer Interaction Institute Carnegie Mellon University Cen, H., Koedinger, K., Junker, B. Learning Factors Analysis - A General Method for Cognitive Model Evaluation and Improvement. 8th International Conference on Intelligent Tutoring Systems Stamper, J. & Koedinger, K.R. Human-machine student model discovery and improvement using data. Proceedings of the 15 th International Conference on Artificial Intelligence in Education

3(2x - 5) = 9 6x - 15 = 92x - 5 = 36x - 5 = 9 Cognitive Tutor Technology Use cognitive model to individualize instruction Cognitive Model: A system that can solve problems in the various ways students can If goal is solve a(bx+c) = d Then rewrite as abx + ac = d If goal is solve a(bx+c) = d Then rewrite as abx + c = d If goal is solve a(bx+c) = d Then rewrite as bx+c = d/a Model Tracing: Follows student through their individual approach to a problem -> context-sensitive instruction

3(2x - 5) = 9 6x - 15 = 92x - 5 = 36x - 5 = 9 Cognitive Tutor Technology Use cognitive model to individualize instruction Cognitive Model: A system that can solve problems in the various ways students can If goal is solve a(bx+c) = d Then rewrite as abx + ac = d If goal is solve a(bx+c) = d Then rewrite as abx + c = d Model Tracing: Follows student through their individual approach to a problem -> context-sensitive instruction Hint message: “Distribute a across the parentheses.” Bug message: “You need to multiply c by a also.” Knowledge Tracing: Assesses student's knowledge growth -> individualized activity selection and pacing Known? = 85% chanceKnown? = 45%

4 Cognitive Model Discovery Traditional Cognitive Task Analysis  Interview experts, think alouds, DFA Result: cognitive model of student knowledge  Cognitive model drives ITS behaviors & instructional design decisions Key goal for Educational Data Mining  Improve Cognitive Task Analysis  Use student data from initial tutor  Employ machine learning & statistics to discover better cognitive models

5 Overview Using learning curves to evaluate cognitive models Statistical models of student performance & learning  Example of improving tutor  Comparison to other Psychometric models Using Learning Factors Analysis to discover better cognitive models Educational Data Mining research challenges

Student Performance As They Practice with the LISP Tutor Mean Error Rate

Evidence for Production Rule as an appropriate unit of knowledge acquisition Production Rule Analysis

8 Using learning curves to evaluate a cognitive model Lisp Tutor Model  Learning curves used to validate cognitive model  Fit better when organized by knowledge components (productions) rather than surface forms (programming language terms) But, curves not smooth for some production rules  “Blips” in leaning curves indicate the knowledge representation may not be right Corbett, Anderson, O’Brien (1995)  Let me illustrate …

9 Curve for “Declare Parameter” production rule How are steps with blips different from others? What’s the unique feature or factor explaining these blips? What’s happening on the 6th & 10th opportunities?

10 Can modify cognitive model using unique factor present at “blips” Blips occur when to-be-written program has 2 parameters Split Declare-Parameter by parameter-number factor:  Declare-first-parameter  Declare-second-parameter (defun second (lst) (first (rest lst))) (defun add-to (el lst) (append lst (list lst)))

11 Can learning curve analysis be automated? Manual learning curve analysis  Identify “blips” in learning curve visualization  Manually create a new model  Qualitative judgment of fit Toward automatic learning curve analysis  Blips as deviations from statistical model  Propose alternative cognitive models  Evaluate cognitive model using prediction accuracy statistics

12 Overview Using learning curves to evaluate cognitive models Statistical models of student performance & learning  Example of improving tutor  Comparison to other Psychometric models Using Learning Factors Analysis to discover better cognitive models Educational Data Mining research challenges

13 Representing Knowledge Components as factors of items Problem: How to represent KC model? Solution: Q-Matrix (Tatsuoka, 1983) Items X Knowledge Components (KCs)  Single KC item = when a row has one 1  Multi-KC item = when a row has many 1’s Item | KCs:AddSubMulDiv 2* * Q matrix is a bridge between a symbolic cognitive model & a statistical model

14 Additive Factors Model Assumptions Logistic regression to fit learning curves (Draney, Wilson, Pirolli, 1995) Assumptions about knowledge components (KCs) & students  Different students may initially know more or less  Students generally learn at the same rate  Some KCs are initially easier than others  Some KCs are easier to learn than others These assumptions are reflected in a statistical model  Intercept parameters for each student  Intercept & slope parameters for each KC Slope = for every practice opportunity there is an increase in predicted performance

15 Simple Statistical Model of Performance & Learning Problem: How to predict student responses from model? Solution: Additive Factor Model  i students, j problems/items, k knowledge components (KCs) Student intercept KC intercept KC slope Model parameters:

16 Area Unit of Geometry Cognitive Tutor 15 skills: Circle-area Circle-circumference Circle-diameter Circle-radius Compose-by-addition Compose-by-multiplication Original cognitive model in tutor: Parallelogram-area Parallelogram-side Pentagon-area Pentagon-side Trapezoid-area Trapezoid-base Trapezoid-height Triangle-area Triangle-side

17 Log Data Input to AFM StudentStep (Item)KCOpportunitySuccess Ap1s1Circle-area00 Ap2s1Circle-area11 Ap2s2Rectangle-area01 Ap2s3 Compose-by- addition00 Ap3s1Circle-area20 Items = steps in tutors with step- based feedback Q-matrix in single column: works for single KC items Opportunities Student has had to learn KC

18 AFM Results for original KC model SkillInterceptSlope Avg OpportuntiesInitial ProbabilityAvg ProbabilityFinal Probability Parallelogram-area Pentagon-area StudentIntercept student01.18 student10.82 student20.21 Model Statistics AIC3,950 BIC4,285 MAD0.083 Higher intercept of skill -> easier skill Higher slope of skill -> faster students learn it Higher intercept of student -> student initially knew more The AIC, BIC & MAD statistics provide alternative ways to evaluate models MAD = Mean Absolute Deviation

19 Overview Using learning curves to evaluate cognitive models Statistical models of student performance & learning  Example of improving tutor  Comparison to other Psychometric models Using Learning Factors Analysis to discover better cognitive models Educational Data Mining research challenges

20 Application: Use Statistical Model to improve tutor Some KCs over-practiced, others under (Cen, Koedinger, Junker, 2007) 20 initial error rate 76% reduced to 40% after 6 times of practice initial error rate 12% reduced to 8% after 18 times of practice

“Close the loop” experiment In vivo experiment: New version of tutor with updated knowledge tracing parameters vs. prior version Reduced learning time by 20%, same robust learning gains Knowledge transfer: Carnegie Learning using approach for other tutor units 21

22 Additive Factor Model (AFM) generalizes Item Response Theory (IRT) Instance of logistic regression  Example: In R use generalized linear regression with family=binomial  glm(prob-correct ~ student + KC + KC:opportunity, family=binomial,…) Generalization of item response theory (IRT)  IRT simply has i student & j item parameters  glm(prob-correct ~ student + item, family=binomial,…) AFM is different from IRT because:  It clusters items by knowledge components  It has an opportunity slope for each KC

Comparing to other psychometric models AFM adds a growth component to “LLTM” (Wilson & De Boeck)  LTTM is an “item explanatory” generalization of IRT or “Rasch”  “Person explanatory” models are related to factor analysis and other matrix factorization techniques

24 Model Evaluation How to compare cognitive models?  A good model minimizes prediction risk by balancing fit with data & complexity (Wasserman 2005) Model-data fit metrics  Log likelihood, root mean squared error (RMSE), mean average deviation (MAD), area under curve (AUC), … Prediction metrics  BIC, AIC: Faster metrics add a penalty for # parameters BIC = -2*log-likelihood + numPar * log(numOb)  Cross validation: Slower but better Split data in training & test sets, optimize parameters with training set, apply fit metrics on test set

A good cognitive model produces a learning curve Recall LISP tutor example above Without decomposition, using just a single “Geometry” skill, Is this the correct or “best” cognitive model? no smooth learning curve. a smooth learning curve. But with decomposition, 12 skills for area, Rise in error rate because poorer students get assigned more problems

DataShop visualizations to aid “blip” detection Some do not => Opportunity to improve model! Many curves show a reasonable decline

27 Learning Factors Analysis

28 Overview Using learning curves to evaluate cognitive models Statistical models of student performance & learning  Example of improving tutor  Comparison to other Psychometric models Using Learning Factors Analysis to discover better cognitive models Educational Data Mining research challenges

29 Learning Factors Analysis (LFA): A Tool for Cognitive Model Discovery LFA is a method for discovering & evaluating alternative cognitive models  Finds knowledge components that best predict student performance & learning transfer Inputs  Data: Student success on tasks in domain over time  Codes: Factors hypothesized to drive task difficulty & transfer Outputs  A rank ordering of most predictive cognitive models  Parameter estimates for each model

30 Learning Factors Analysis (LFA) draws from multiple disciplines Cognitive Psychology  Learning curve analysis (Corbett, et al 1995) Psychometrics & Statistics  Q Matrix & Rule Space (Tatsuoka 1983, Barnes 2005)  Item response learning model (Draney, et al., 1995)  Item response assessment models (DiBello, et al., 1995; Embretson, 1997; von Davier, 2005) Machine Learning & AI  Combinatorial search (Russell & Norvig, 2003)

31 Item Labeling & the “P Matrix”: Adding Alternative Factors How to improve existing cognitive model? Have experts look for difficulty factors that are candidates for new KCs. Put these in “P matrix” Item | SkillAddSubMul 2*8001 2*8 – * *8101 Q MatrixP Matrix Item | SkillDeal with negative Order of Ops … 2*800 2*8 – 300 2* *801

32 Using P matrix to update Q matrix Create a new Q’ by using elements of P as arguments to operators  Add operator: Q’ = Q + P[,1]  Split operator: Q’ = Q[, 2] * P[,1] Item | SkillAddSubMulDivneg 2* *8 – * Q- Matrix after add P[, 1] Item | SkillAddSubMulDivSub- neg 2* *8 – * Q- Matrix after splitting P[, 1], Q[,2]

33 LFA: KC Model Search How to find best model given Q and P matrices? Use best-first search algorithm (Russell & Norvig 2002)  Guided by a heuristic, such as BIC or AIC Do model selection within space of Q matrices Steps: 1. Start from an initial “node” in search graph using given Q 2. Iteratively create new child nodes (Q’) by applying operators with arguments from P matrix 3. Employ heuristic (BIC of Q’) to rank each node 4. Select best node not yet expanded & go back to step 2

34 Example in Geometry of split based on factor in P matrix StudentStepSkill Opportunity A p1s1Circle-area-alone0 A p2s1Circle-area-embed0 Ap2s2Rectangle-area0 Ap2s3Compose-by-add0 Ap3s1Circle-area-alone1 StudentStepSkillOpportunityEmbed Ap1s1Circle-area0alone Ap2s1Circle-area1embed Ap2s2Rectangle-area0 Ap2s3Compose-by-add0 Ap3s1Circle-area2alone After Splitting Circle-area by Embed Factor in P matrix Original Q matrix New Q matrix Revised Opportunity

LFA –Model Search Process Automates the process of hypothesizing alternative cognitive models & testing them against data Search algorithm guided by a heuristic: BIC Start from an existing cog model (Q matrix) Cen, H., Koedinger, K., Junker, B. (2006). Learning Factors Analysis: A general method for cognitive model evaluation and improvement. 8th International Conference on Intelligent Tutoring Systems.

36 Example LFA Results: Applying splits to original model Common results:  Compose-by-multiplication split based on whether it was an area or a segment being multiplied  Circle-radius is split based on whether it is being done for the first time in a problem or is being repeated Model 1Model 2Model 3 Number of Splits:3 Number of Splits:2 1.Binary split compose- by-multiplication by figurepart segment 2.Binary split circle- radius by repeat repeat 3.Binary split compose- by-addition by backward backward 1.Binary split compose-by- multiplication by figurepart segment 2.Binary split circle-radius by repeat repeat 3.Binary split compose-by- addition by figurepart area- difference 1.Binary split compose-by- multiplication by figurepart segment 2.Binary split circle-radius by repeat repeat Number of Skills: 18 Number of Skills: 17 BIC: 4, BIC: 4,251.07

37 Compose-by-multiplication KC examples Composing SegmentsComposing Areas

38 Tutor Design Implications 1 LFA search suggests distinctions to address in instruction & assessment With these new distinctions, tutor can  Generate hints better directed to specific student difficulties  Improve knowledge tracing & problem selection for better cognitive mastery Example: Consider Compose-by-multiplication before LFA InterceptslopeAvg Practice OpportuntiesInitial ProbabilityAvg ProbabilityFinal Probability CM With final probability.92, many students are short of.95 mastery threshold

39 Tutor Design Implications 2 However, after split: CM-area and CM-segment look quite different  CM-area is now above.95 mastery threshold (at.96)  But CM-segment is only at.60  Original model penalizes students who have key idea about composite areas (CM-area) -- some students solve more problems than needed Instructional redesign implications:  Change skillometer so CM-area & CM-segment are separately addressed Set parameters appropriately -- CM-segment with have a lower initial known value  Add more problems to allow for mastery of CM-segment  Add new hints specific to the CM-segment situation InterceptslopeAvg Practice Opportunties Initial Probability Avg Probability Final Probability CM CMarea CMsegment

40 Summary of Learning Factors Analysis (LFA) LFA combines statistics, human expertise, & combinatorial search to discover cognitive models Evaluates a single model in seconds, searches 100s of models in hours  Model statistics are meaningful  Improved models suggest tutor improvements Can currently be applied, by request, to any dataset in DataShop with at least two KC models

41 Mixed initiative human- machine discovery 1. Human  Hypothesize possible “learning factors” and code steps 2. Machine  Search over factors, report best models discovered 3. Human  Inspect results  If needed, propose new factors. Go to 2.  If good, modify tutor and test.

42 Human-machine discovery of new cognitive models Better models discovered in Geometry, Statistics, English, Physics

43 Some Open EDM Research Problems

44 Open Research Questions: Technical What factors to consider? P matrix is hard to create  Enhancing human role: Data visualization strategies  Other techniques: Matrix factorization, LiFT  Other data: Do clustering on problem text Interpreting LFA output can be difficult  How to make interpretation easier? => Researcher can’t just “go by the numbers” 1) Understand the domain, the tasks 2) Get close to the data

45 Model search using DataShop: Human & machine improvements DataShop datasets w/ improved KC models:  Geometry Area ( ), Geometry Area Hampton Unit 34, … New KCs (learning factors) found using DataShop visualization tools  Learning curve, point tool, performance profiler  Example of human “feature engineering” New KC models also discovered by LFA  Research goal: Iterate between LFA & visualization to find increasingly better KC models

46 Most curves “curve”, but if flat, then KC may be bad

47 Detecting planning skills: Scaffolded vs. unscaffolded problems Scaffolded  Prompts are given for subgoals Unscaffolded –Prompts are not given for subgoals (initially)

48 1. Not smooth 2. No learning 3. Uneven error rate Easy tasks do not require subgoals, hard tasks do! Discovering a new knowledge component Each KC should have: 1. smooth learning curve 2. statistical evidence of learning 3. even error rates across tasks Create new KCs by finding a feature common to hard tasks but missing in easy ones

New model discovery: Split “compose” into 3 skills Hidden planning knowledge: If you need to find the area of an irregular shape, then try to find the areas of regular shapes that make it up Redesign instruction in tutor  Design tasks that isolate the hidden planning skill  Given square & circle area, find leftover 1 2 When prompts are initially present for component areas 3

Before unpacking compose-by-addition After -- unpacked into subtract, decompose, remaining compose-by-addition

3-way split in new model (green) better fits variability in error rates than original (blue)

52 Automate human-machine strategies for “blip” detection Research goal: Automate low slope, non-low intercept, & high residual detection Uses:  speed up LFA search  point human coders to bad KCs cluster harder vs. easier tasks

53 Developing & evaluating different learning curve models Many papers in Educational Data Mining (EDM) conference  Also in Knowledge Discovery & Data mining (KDD) Papers comparing knowledge tracing, AFM, PFA, CPFA, IFA  See papers by Pavlik, Beck, Chi …

54 Open Research Questions: Psychology of Learning Change AFM model assumptions  Is student learning rate really constant? Does a Student x Opportunity interaction term improve fit? What instructional conditions or student factors change rate?  Is knowledge space “uni-dimensional”? Does a Student x KC interaction term improve fit?  Need different KC models for different students/conditions? Is learning curve an exponential or power law?  Long-standing debate, which has focused on “reaction time” not on error rate!  Compare use of Opportunity vs.Log(Opportunity) Other outcome variables: reaction time, assistance score Other predictors: Opportunities => Time per instructional event; Kinds of opportunities: Successes, failures, hints, gamed steps, …

55 Open Research Questions: Instructional Improvement Do LFA results generalize across data sets?  Is AIC or BIC a good estimate for cross-validation results?  Does a model discovered with one year’s tutor data generalize to a next year?  Does model discovery work for ed games, other domains? Use learning curves to compare instructional conditions in experiments Need more “close the loop” experiments  EDM => better model => better tutor => better student learning

56 END

57 To do Shorten – by how much?  Which slides to delete?  Remove details on geometry model application Put “other” alternatives at end  Cottage industry in EDM & KDD  Papers comparing knowledge tracing, AFM, PFA, CPFA, IFA … see Pavlik, Beck, Chi … Table with LFA search results Demo parts of DataShop? Add some interactive questions  Use Learning Objectives to aid that

58 If time: DataShop Demo and/or Video See video on “about” page “Using DataShop to discover a better knowledge component model of student learning”

Before unpacking compose-by-addition After -- unpacked into subtract, decompose, remaining compose-by-addition

60 Detecting planning skills: Scaffolded vs. unscaffolded problems Scaffolded  Columns given for area subgoals Unscaffolded  Columns not given for area subgoals

61 Knowledge Decomposibility Hypothesis Human acquisition of academic competencies can be decomposed into units, called knowledge components (KCs), that predict student task performance & transfer Performance predictions  If item I1 only requires KC1 & item I2 requires both KC1 and KC2, then item I2 will be harder than I1  If student can do I2, then they can do I1 Transfer predictions  If item I1 requires KC1, & item I3 also requires KC1, then practice on I3 will improve I1  If item I1 requires KC1, & item I4 requires only KC3, then practice on I4 will not improve I1 Fundamental EDM idea:  We can discover KCs (cog models) by working these predictions backwards! KC1 add KC2 carry KC3 subt I1: I2: I3: I4: Example of Items & KCs

62 Using Student Data to Make Discoveries Design Cognitive Tutor courses: Tech, Text, Training Deploy Address social context Data Qual, quant; process, product Discover Cognition, learning, instruction, context Research base Cognitive Psychology Artificial Intelligence Practice base Educators Standards

63 Cognitive Task Analysis is being automated Use ed tech to collect student data Develop data visualizations & model discovery algorithms Machine learning systems & cognitive scientists working together Cen, Koedinger, Junker (2006). Learning Factors Analysis: A general method for cognitive model evaluation and improvement. Intelligent Tutoring Systems.

64 Can this data-driven CTA be brought to scale? Combine Cognitive Science, Psychometrics, Machine Learning … Collect a rich body of data Develop new model discovery techniques PSLC & DataShop are facilitating

65 Cognitive modeling from symbolic to statistical Abstract from a computational symbolic cognitive model to a statistical cognitive model For each task label the knowledge components or skills that are required: AddSubMul 2*8001 2*8 – * *8101 Q Matrix

Geometry Tutor Scaffolding problem decomposition Problem decomposition support

67 Good Cognitive Model => Good Learning Curve An empirical basis for determining when a cognitive model is good Accurate predictions of student task performance & learning transfer  Repeated practice on tasks involving the same skill should reduce the error rate on those tasks => A declining learning curve should emerge

68 Statistical Model of Student Performance & Learning “Additive Factor Model” (AFM) (cf., Draney, Pirolli, Wilson, 1995) Evaluate with BIC, AIC, cross validation to reduce over-fit

Comparing to other psychometric models Adds a growth component to “LLTM” (Wilson & De Boeck)  LTTM is an “item explanatory” generalization of Rasch/IRT  AFM is “item learning explanatory”

70 Automating the Cognitive Model Discovery Process Learning Factors Analysis Input: Factors that may differentiate tasks Output: Best cognitive model Cen, H., Koedinger, K., Junker, B. (2006). Learning Factors Analysis: A general method for cognitive model evaluation and improvement. 8th International Conference on Intelligent Tutoring Systems.