Presentation is loading. Please wait.

Presentation is loading. Please wait.

March 20, 2004Symposium on Reasoning and Learning in Cognitive Systems Mixing Automatic and Deliberative Learning During Problem Solving Randolph M. Jones.

Similar presentations


Presentation on theme: "March 20, 2004Symposium on Reasoning and Learning in Cognitive Systems Mixing Automatic and Deliberative Learning During Problem Solving Randolph M. Jones."— Presentation transcript:

1 March 20, 2004Symposium on Reasoning and Learning in Cognitive Systems Mixing Automatic and Deliberative Learning During Problem Solving Randolph M. Jones Soar Technology & Colby College

2 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 2 Background There are alternative ways we might incorporate multi-step learning into a model –One approach would be to automate explicit instruction of desired task behavior Even this is difficult –This talk focuses on models that can discover new problem-solving knowledge and strategies on their own

3 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 3 Knowledge Tuning and Acquisition There are two primary ways a model can learn new strategies –Acquiring new task knowledge that allows more complete or efficient coverage of a problem space –Tuning existing task knowledge so it is retrieved more oportunistically Knowledge acquisition in its own right is also important –But this work suggests that knowledge acquisition depends on knowledge tuning

4 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 4 Knowledge Tuning Basic representational structure of knowledge chunk remains unchanged Retrieval/selection patterns associated with the knowledge do change

5 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 5 Knowledge Acquisition Entirely new structured representations of long-term knowledge are added to the models knowledge base Or existing chunks of knowledge undergo structural changes

6 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 6 Task Example: Solving Physics Problems Learning to solve physics problems involves learning new equations relevant to the problems, and learning the situations in which those equations should be used Students who self-explain study examples show greater improved performance than those who dont (Chi et al., 1989) –Are they tuning knowledge or acquiring knowledge? Cascade (VanLehn, Jones, & Chi, 1992) models the self- explanation effect observed in humans learning to solve physics problems

7 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 7 Task Example: Simple Addition There are a variety of strategies that can be used to perform elementary addition, some more efficient than others Children are usually instructed using a basic strategy, but invent a particular set of more efficient strategies on their own (Siegler & Jenkins, 1989) –Are they tuning knowledge or acquiring knowledge? GIPS (Jones & VanLehn, 1994) models the series of strategy shifts exhibited by children

8 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 8 Cascade: Typical Problem 10 kg What is the tension in the string?

9 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 9 Cascade: Typical Problem A B C What is the magnitude of each force?

10 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 10 Cascade: Typical Example Let the knot be the body FA, FB, FC are all the forces acting on the body The body is at rest, so FA+FB+FC=0 By projection, FA X +FB X =0 By projection, FA Y +FB Y +FC Y =0 FA X =–FA cos 30 = –0.8666FA etc. FA FB FC

11 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 11 Cascade: Modeling Goal Explain the learning process and other factors that cause students who carefully study examples to learn more effectively than students who do not

12 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 12 Cascade: Knowledge Representation Long-term task knowledge is a set of physics equations, geometric equations, and rules for representing free-body diagrams –Implemented in Prolog Default problem-solving strategy is exhaustive depth-first search with backtracking –Straightforward application of Prolog Problem-solving goals are quantities (variables) for which the problem solver must compute a value Selection knowledge allows heuristic search by using past solution paths as analogies to the current problem

13 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 13 Cascade: Learning Processes Knowledge tuning –Analogical Search Control –When Cascade succeeds in computing a value for a sought quantity, it records a triple including the name of the problem, the sought quantity, and the equation that was used to compute the value –The caching process occurs automatically and frequently, every time a subgoal is achieved –On subsequent problems, Cascade Attempts to map the current problem quantities and relations to the analog problem Searches for cached triples that mention problem analogs to the current problem, together with an analogous sought quantity Attempts the retrieved equation before falling back on the default ordering of knowledge (if backtracking occurs)

14 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 14 Cascade: Learning Processes Knowledge acquisition –Explanation-based Learning of Correctness –If Cascade cannot solve a problem (after exhaustive search), it begins the search again, this time attempting a repair at the first point that backtracking is encountered Repairs occur by attempting to apply relevant overly general rules to the problem On success, Cascade stores a specialization of the overly general rule with the rest of the task knowledge –The rule learning process occurs deliberatively and infrequently, only after the model has recognized an impasse in problem solving

15 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 15 Cascade: Learning Interactions Knowledge acquisition only works if the model is repairing the right gap in a potential solution space The model can be guided toward the right gap: –By the directions in a worked example –By the quality of knowledge tuning

16 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 16 Cascade: Learning Interactions Initial problem state Solution Dead ends False paths Knowledge gap

17 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 17 Cascade: Experimental Results No Analogical Search Control –Learns 3 correct rules –Solves 9 problems correctly No EBLC on examples –Learns 13 correct rules –Learns 4 incorrect rules –Solves 21 problems correctly (many using a backup transformational analogy strategy) ASC & EBLC –Learns 22 correct rules –Solves 23 problems correctly

18 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 18 GIPS: Typical Problem Sum Strategy

19 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 19 GIPS: Modeling Goal Model how children independently invent the Min strategy with experience –Min is a more efficient strategy, suggesting that it may be produced primarily by knowledge tuning –However, there appear to be structural changes to the steps the children are taking to solve problems

20 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 20 GIPS: Knowledge Representation Task knowledge is represented as STRIPS-like operators with preconditions, constraints, add conditions, and delete conditions –Problem-solving algorithm is flexible means-ends analysis TRANSFORM goal: Use features describing current state and goal to retrieve a candidate operator to APPLY for the next step in the transformation APPLY goal: Execute the operator if possible, else set up a new TRANSFORM to the preconditions of the operator Retrieval/selection knowledge is encoded as probability estimates (for logical sufficiency and logical necessity) attached to each potential triggering feature for each operator –State and Goal relations

21 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 21 Example Bayesian Concept Liftable FEATURELSLN size is small3.00.0 weight is light2.00.3 has handle2.00.3 attached to floor0.03.0 color is red1.01.0 Note this example has propositional features, but features in GIPS are relational –GIPS uses a graph-based maximal partial match procedure to map combinations of relations to propositions

22 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 22 GIPS: Learning Processes Knowledge tuning –Every time an APPLY goal leads to success or failure, GIPS updates the appropriate probability estimates for each state and goal feature present when the APPLY goal was created A = Action A is the right thing to do next F = Feature F is true in the problem situation –A similar process occurs every time an operator executes (or not)

23 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 23 GIPS: Learning Processes Knowledge acquisition –When feature values for an operators execution concept receive particularly strong logical necessity values, a deliberative process explicitly adds the new feature as a condition of the operator Another process removes features from the operator conditions

24 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 24 The SUM-to-MIN Strategy Shift I just counted X and X is the addend. (I have to count.) I see X fingers and X is the addend. (I dont have to count! I can use the counter for something else!) X is the addend. (I dont have to raise any fingers.)

25 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 25 GIPS Learning Interactions Bayesian updates happen continuously and automatically, leading to performance shifts based on retrieval of operators Based on accumulating evidence, the model periodically tries more drastic structural changes to operator preconditions, which have larger effects on subsequent retrieval patterns (because operator preconditions are used as subgoal retrieval cues and determine satisfaction of APPLY goals)

26 March 20, 2004Mixing Automatic and Deliberative Learning During Problem Solving 26 Lessons It is difficult to acquire new knowledge without first tuning old knowledge Tuning old knowledge implies that you have some old knowledge to tune For complex learning, we need to focus on learning in the context of significant prior knowledge Tuning can help guide the search for building new operators (Cascade) as well as for adjusting the structural representations of existing operators (GIPS) –You only want to acquire new knowledge after you have accumulated some evidence (from tuning) that the knew knowledge is appropriate and useful


Download ppt "March 20, 2004Symposium on Reasoning and Learning in Cognitive Systems Mixing Automatic and Deliberative Learning During Problem Solving Randolph M. Jones."

Similar presentations


Ads by Google