Difference between revisions of "Personalized Problems"

From Open Pattern Repository for Online Learning Systems
Jump to navigation Jump to search
(Initial commit for the Match problems to skill levels pattern)
 
m (Edited format)
 
(25 intermediate revisions by 2 users not shown)
Line 1: Line 1:
<!-- Design pattern template
Fill in all succeeding fields that are either surrounded by a comment tag (<!--  - ->) or angle brackets (< >) when applicable -->
{{Infobox_designpattern
{{Infobox_designpattern
|author=[[User:Pinventado|Paul Inventado]]<br/>Peter Scupelli
|image=Personalized_problems.png
|contributor=
|contributor=[[Paul Salvador Inventado]], [[Peter Scupelli]]
|source= Inventado and Scupelli (in press 2015)<ref name="Inventadoip">Inventado, P.S. & Scupelli, P. (in press 2015). [https://cmu.box.com/shared/static/m6qfs01z71gt38a7tf85gcgl8t84iw50.pdf A Data-driven Methodology for Producing Online Learning System Design Patterns]. In ''Proceedings of the 22nd Conference on Pattern Languages of Programs (PLoP 2015)''. New York:ACM.</ref>
|dataanalysis=[[Analysis:Student_affect_and_interaction_behavior_in_ASSISTments#skillleve|Student affect and interaction behavior in ASSISTments]]
|domain= General
|domain= General
|stakeholders= Teachers<br/>Students
|stakeholders= Teachers, Students, System developers
|evaluation = PLoP 2015 writing workshop <br/>[[Talk:ASSISTments]]
|evaluation = [http://www.hillside.net/plop/2015/ PLoP 2015] shepherding and writing workshop<br/> [[Talk:ASSISTments]]
|application = [[ASSISTments]]
|application = [[ASSISTments]]
|appliedevaluation = [[ASSISTments]]
|appliedevaluation = [[ASSISTments]]
}}
}}


If students have difficulty answering a problem, then give students problems that fit their level of understanding.
Assign appropriate problem-solving activities to a student’s skill level<ref name="Inventadoip"/>


[[File:Match_problems_to_skill_levels.png|center|800px]]
==Context==
==Context==
Teachers use ASSISTments to select the problems in an assignment, to specify the sequence and conditions for presenting questions to students, and to assign the homework or activity to their students.  
Content creators design problems for students to solve to better understand the concepts taught.


==Evidence==
==Problem==
 
Students become bored or disengage from an activity if they are asked to solve problems that are either too easy or too difficult.
===Literature===
Learners experience cognitive overload when they are given a task that is too difficult for them to accomplish. Experiments showed higher learning outcomes, more task involvement, and less effort when tasks were adapted to learner's skill levels <ref>Corbalan, G., Kester, L., & van Merriënboer, J. J. (2008). [http://www.sciencedirect.com/science/article/pii/S0361476X08000118 Selecting learning tasks: Effects of adaptation and shared control on learning efficiency and task involvement]. Contemporary Educational Psychology, 33(4), 733-756.</ref>.
 
===Data===
Data showed that students got bored or disengaged when a problem was too easy or too difficult. [http://http://www.contrib.andrew.cmu.edu/~lelab/patternwiki/ASSISTments_experiments].


==Forces==
==Forces==
* Teachers assign the same homework to all students.
#'''Prior knowledge.''' Students cannot solve a problem if they lack the necessary skills<ref name="Sweller2004">Sweller, J. (2004). [http://link.springer.com/article/10.1023%2FB%3ATRUC.0000021808.72598.4d Instructional design consequences of an analogy between evolution by natural selection and human cognitive architecture]. ''Instructional science, 32''(1-2), 9-31.</ref>.
* The ASSISTments interface provides teachers with simple control mechanisms (e.g., sequencing of questions in a homework, selecting questions based on the correctness) to control how questions are presented to students.
#'''Desirable difficulty.''' Activities that are too easy or do not challenge learners’ understanding of a concept require less mental processing and often, result in less learning<ref>Bjork, R.A. (1994). [http://psycnet.apa.org/psycinfo/1994-97967-009 Memory and metamemory considerations in the training of human beings]. In J. Metcalfe and A. Shimamura (Eds.), ''Metacognition: Knowing about knowing''. (pp.185-205). Cambridge, MA: MIT Press.</ref><ref>Piaget, J. (1952).The origins of intelligence. New York: International University Press.</ref>.
* Problem definitions (e.g., difficulty, presentation, wording, sequence) affect students’ learning experiences.
#'''Learning rate.''' Student learning rates vary because of differences in prior knowledge, learning experiences, and quality of instruction received<ref name="Bloom1974">Bloom, B. S. (1974). [http://psycnet.apa.org/journals/amp/29/9/682/ Time and learning]. ''American psychologist, 29''(9), 682.</ref>.
#'''Persistence.''' Students may disengage from a learning activity if they get stuck too long while trying to solve it<ref>Arnold, A., Scheines, R., Beck, J. E., and Jerome, B. (2005). [https://oli.cmu.edu/wp-oli/wp-content/uploads/2012/05/Arnold_2005_Time_and_Attention.pdf Time and attention: Students, sessions, and tasks]. In ''Proceedings of the AAAI 2005 Workshop Educational Data Mining'' (pp. 62-66).</ref><ref>D’Mello, S., and Graesser, A.  (2012). [http://www.sciencedirect.com/science/article/pii/S0959475211000806 Dynamics of affective states during complex learning]. ''Learning and Instruction, 22''(2), 145-157.</ref>.


==Problem==
To ensure that students are challenged but not overwhelmed by the questions they are asked to answer.


==Solution==
==Solution==
Assign to students math problems that they have enough skills to solve.
Therefore, assign problems that are appropriate for a student’s skill level. A student’s capability to solve a problem can be identified using assessments of their knowledge on pre-requisite skills, or model-based predictors<ref name="Koedinger2007">Koedinger, K. R., and Aleven, V.  (2007). [http://link.springer.com/article/10.1007/s10648-007-9049-0 Exploring the assistance dilemma in experiments with cognitive tutors]. ''Educational Psychology Review, 19''(3), 239-264.</ref>.


==Consequences==
==Consequences==


===Benefits===
===Benefits===
* Students do not get bored or too frustrated by the level of question difficulty
#Students are asked to answer problems that they are capable of solving themselves, or with some assistance<ref>Vygotsky, L. S.  (1962). ''Language and thought''. Massachusetts Institute of Technology Press, Ontario, Canada.</ref>.
* Students are more likely to complete their homework
#The problem challenges students because it requires skills that students may not have mastered yet.
* Students master the skills being taught in the assignment
#Each student is assigned to a different problem that is appropriate for his/her skill level.
#Students may continue to solve a challenging problem if they have enough prerequisite knowledge.


===Liabilities===
===Liabilities===
* Teachers and content experts will need to create content and define pedagogies for handling differences in student skills.
#Content writers will need to provide different content for each skill level.
* ASSISTments will need to be modified to adapt content based on a students’ skill level.
#The system needs to keep track of pre-requisite and post-requisite skills, as well as problems associated with those skills so they can be assigned appropriately.
#The system needs to be capable of measuring students’ skill level and selecting problems dynamically.
#If students’ skill level is incorrectly identified, the system can still give students problems that are too easy or too difficult.
 
==Evidence==
 
===Literature===
Research showed that personalizing content according to students’ skill level resulted in similar learning gains as non-personalized content, but took a shorter amount of time. This was observed in various domains such as simulated air traffic control,<ref>Salden, R. J., Paas, F., Broers, N. J., and Van Merriënboer, J. J. (2004)
. [http://link.springer.com/article/10.1023/B:TRUC.0000021814.03996.ff Mental effort and performance as determinants for the dynamic selection of learning tasks in air traffic control training]. ''Instructional science, 32''(1-2), 153-172.</ref>, Algebra<ref>Cen, H., Koedinger, K. R., and Junker, B. (2007). [http://dl.acm.org/citation.cfm?id=1563681 Is Over Practice Necessary?-Improving Learning Efficiency with the Cognitive Tutor through Educational Data Mining]. ''Frontiers in Artificial Intelligence and Applications, 158'', 511.</ref>, Geometry<ref>Salden, R.J.C.M., Aleven, V., Schwonke, R. and Renkl, A. (2010). [http://link.springer.com/article/10.1007/s11251-009-9107-8 The expertise reversal effect and worked examples in tutored problem solving]. ''Instructional Sicience, 38'', 289--307.</ref>, and health sciences<ref>Corbalan, G., Kester, L. and van Merrieonboer, J.J.G. (2008). [http://www.sciencedirect.com/science/article/pii/S0361476X08000118 Selecting learning tasks: Effects of adaptation and shared control on learning efficiency and task involvement]. ''Contemporary Educational Psycholoy, 33'', 733--756.</ref>.
 
===Discussion===
Shepherds, writing workshop participants, and learning system stakeholders (i.e., data mining experts, learning scientists, and educators) agreed that the design pattern’s solution could address the identified problem.
 
===Data===
According to an [[Analysis:Student_affect_and_interaction_behavior_in_ASSISTments#skilllevel | ASSISTments math online learning system data]], boredom and gaming behavior correlated with problem difficulty (i.e., evidenced by answer correctness and number of hint requests).
<!--===Applied evaluation===
Results from randomized controlled trials (RCTs) or similar tests that measures the pattern's effectiveness in an actual application. For example, compare student learning gains in an online learning system with and without applying the pattern. -->


==Example==
==Example==
A teacher would create problems for a homework with different difficulty levels appropriate for students who show low, medium or high performance. As students answer questions in their homework, ASSISTments will keep track of their progress to classify them as low performance (i.e., student makes mistakes ≥ 60% of the time), medium performance (i.e., student makes mistakes < 60% and ≥ 40% of the time) or high performance (i.e., student makes mistakes < 40% of the time). Depending on students’ performance level, ASSISTments will provide them with the corresponding question type so it is more likely for students to receive questions that are fit for their skill level.  
Many online learning systems were designed to adapt to students’ skill level. For example, SQL-Tutor provides students with problems on SQL programming that are appropriate to their level of knowledge<ref>Mitrovic, A., and Martin, B.  (2004). [http://link.springer.com/chapter/10.1007/978-3-540-27780-4_22 Evaluating adaptive problem selection]. In ''Adaptive hypermedia and adaptive web-based systems'' (pp. 185-194). Springer Berlin Heidelberg.</ref>. [[Cognitive_tutor_algebra | Cognitive Tutor Algebra]] is another learning system that tracks student mastery on a particular knowledge component and provides them with algebra problems that are appropriate to their skill level<ref name="Koedinger2007"/>. The [[ASSISTments]] online learning system provides an IF-THEN-ELSE functionality that allows teachers to control the problem sets assigned to students based on their performance<ref>Donnelly, C.J.  (2015). [https://www.wpi.edu/Pubs/ETD/Available/etd-042315-135723/unrestricted/cdonnelly_thesis.pdf Enhancing Personalization Within ASSISTments (Doctoral dissertation)].</ref>. This can be used to identify students’ skill level and to assign the appropriate problem set.
 
 
A concrete example for applying this pattern is a teacher that encodes multiple math problem sets with varying levels of difficulty into an online learning system (e.g., single-digit subtraction, multiple-digit subtraction, subtraction by regrouping). As students answer questions in their homework, the online learning system would keep track of students’ progress to identify their skill level such as low (i.e., student makes mistakes ≥ 60% of the time), medium (i.e., student makes mistakes < 60% and ≥ 40% of the time) or high (i.e., student makes mistakes < 40% of the time). Based on students’ performance, the online learning system would provide the appropriate problem set so that it is more likely for students to receive questions that are fit for their skill level.


==Related patterns==
==Related patterns==
Problems can be personalized much like Content Personalization (Danculovic et al. 2001), but will heavily depend on a model of student knowledge.
The {{Patternlink|Personalized Problems}} design pattern is an implementation of the {{Patternlink|Different Exercise Levels}} design pattern in online learning systems. The system should use the {{Patternlink|Just Enough Practice}} design pattern so that students reach mastery before switching to a more challenging problem set. The {{Patternlink|Differentiated Feedback}} or {{Patternlink|Worked examples}} design patterns may be used to facilitate learning.  
 


==References==
==References==
<references/>
<references/>


[[Category:Design_patterns]] [[Category:ASSISTments]]
==External Links==
*[http://assistments.org ASSISTments]
* [http://www.carnegielearning.com/learning-solutions/software/cognitive-tutor/ Cognitive Tutor Software]
 
 
[[Category:Design_patterns]] [[Category:ASSISTments]] [[Category:Full_Pattern]]  [[Category:Pattern Language for Math problems and Learning Support in Online Learning Systems]] [[Category:Online Learning System]] [[Category:Intelligent Tutoring System]]

Latest revision as of 09:18, 6 June 2017


Personalized Problems
Personalized problems.png
Contributors Paul Salvador Inventado, Peter Scupelli
Last modification June 6, 2017
Source Inventado and Scupelli (in press 2015)[1]
Pattern formats OPR Alexandrian
Usability
Learning domain General
Stakeholders Teachers, Students, System developers
Production
Data analysis Student affect and interaction behavior in ASSISTments
Confidence
Evaluation PLoP 2015 shepherding and writing workshop
Talk:ASSISTments
Application ASSISTments
Applied evaluation ASSISTments

Assign appropriate problem-solving activities to a student’s skill level[1]

Context

Content creators design problems for students to solve to better understand the concepts taught.

Problem

Students become bored or disengage from an activity if they are asked to solve problems that are either too easy or too difficult.

Forces

  1. Prior knowledge. Students cannot solve a problem if they lack the necessary skills[2].
  2. Desirable difficulty. Activities that are too easy or do not challenge learners’ understanding of a concept require less mental processing and often, result in less learning[3][4].
  3. Learning rate. Student learning rates vary because of differences in prior knowledge, learning experiences, and quality of instruction received[5].
  4. Persistence. Students may disengage from a learning activity if they get stuck too long while trying to solve it[6][7].


Solution

Therefore, assign problems that are appropriate for a student’s skill level. A student’s capability to solve a problem can be identified using assessments of their knowledge on pre-requisite skills, or model-based predictors[8].

Consequences

Benefits

  1. Students are asked to answer problems that they are capable of solving themselves, or with some assistance[9].
  2. The problem challenges students because it requires skills that students may not have mastered yet.
  3. Each student is assigned to a different problem that is appropriate for his/her skill level.
  4. Students may continue to solve a challenging problem if they have enough prerequisite knowledge.

Liabilities

  1. Content writers will need to provide different content for each skill level.
  2. The system needs to keep track of pre-requisite and post-requisite skills, as well as problems associated with those skills so they can be assigned appropriately.
  3. The system needs to be capable of measuring students’ skill level and selecting problems dynamically.
  4. If students’ skill level is incorrectly identified, the system can still give students problems that are too easy or too difficult.

Evidence

Literature

Research showed that personalizing content according to students’ skill level resulted in similar learning gains as non-personalized content, but took a shorter amount of time. This was observed in various domains such as simulated air traffic control,[10], Algebra[11], Geometry[12], and health sciences[13].

Discussion

Shepherds, writing workshop participants, and learning system stakeholders (i.e., data mining experts, learning scientists, and educators) agreed that the design pattern’s solution could address the identified problem.

Data

According to an ASSISTments math online learning system data, boredom and gaming behavior correlated with problem difficulty (i.e., evidenced by answer correctness and number of hint requests).

Example

Many online learning systems were designed to adapt to students’ skill level. For example, SQL-Tutor provides students with problems on SQL programming that are appropriate to their level of knowledge[14]. Cognitive Tutor Algebra is another learning system that tracks student mastery on a particular knowledge component and provides them with algebra problems that are appropriate to their skill level[8]. The ASSISTments online learning system provides an IF-THEN-ELSE functionality that allows teachers to control the problem sets assigned to students based on their performance[15]. This can be used to identify students’ skill level and to assign the appropriate problem set.


A concrete example for applying this pattern is a teacher that encodes multiple math problem sets with varying levels of difficulty into an online learning system (e.g., single-digit subtraction, multiple-digit subtraction, subtraction by regrouping). As students answer questions in their homework, the online learning system would keep track of students’ progress to identify their skill level such as low (i.e., student makes mistakes ≥ 60% of the time), medium (i.e., student makes mistakes < 60% and ≥ 40% of the time) or high (i.e., student makes mistakes < 40% of the time). Based on students’ performance, the online learning system would provide the appropriate problem set so that it is more likely for students to receive questions that are fit for their skill level.

Related patterns

The Personalized Problems design pattern is an implementation of the Different Exercise Levels design pattern in online learning systems. The system should use the Just Enough Practice design pattern so that students reach mastery before switching to a more challenging problem set. The Differentiated Feedback or Worked examples design patterns may be used to facilitate learning.


References

  1. 1.0 1.1 Inventado, P.S. & Scupelli, P. (in press 2015). A Data-driven Methodology for Producing Online Learning System Design Patterns. In Proceedings of the 22nd Conference on Pattern Languages of Programs (PLoP 2015). New York:ACM.
  2. Sweller, J. (2004). Instructional design consequences of an analogy between evolution by natural selection and human cognitive architecture. Instructional science, 32(1-2), 9-31.
  3. Bjork, R.A. (1994). Memory and metamemory considerations in the training of human beings. In J. Metcalfe and A. Shimamura (Eds.), Metacognition: Knowing about knowing. (pp.185-205). Cambridge, MA: MIT Press.
  4. Piaget, J. (1952).The origins of intelligence. New York: International University Press.
  5. Bloom, B. S. (1974). Time and learning. American psychologist, 29(9), 682.
  6. Arnold, A., Scheines, R., Beck, J. E., and Jerome, B. (2005). Time and attention: Students, sessions, and tasks. In Proceedings of the AAAI 2005 Workshop Educational Data Mining (pp. 62-66).
  7. D’Mello, S., and Graesser, A. (2012). Dynamics of affective states during complex learning. Learning and Instruction, 22(2), 145-157.
  8. 8.0 8.1 Koedinger, K. R., and Aleven, V. (2007). Exploring the assistance dilemma in experiments with cognitive tutors. Educational Psychology Review, 19(3), 239-264.
  9. Vygotsky, L. S. (1962). Language and thought. Massachusetts Institute of Technology Press, Ontario, Canada.
  10. Salden, R. J., Paas, F., Broers, N. J., and Van Merriënboer, J. J. (2004) . Mental effort and performance as determinants for the dynamic selection of learning tasks in air traffic control training. Instructional science, 32(1-2), 153-172.
  11. Cen, H., Koedinger, K. R., and Junker, B. (2007). Is Over Practice Necessary?-Improving Learning Efficiency with the Cognitive Tutor through Educational Data Mining. Frontiers in Artificial Intelligence and Applications, 158, 511.
  12. Salden, R.J.C.M., Aleven, V., Schwonke, R. and Renkl, A. (2010). The expertise reversal effect and worked examples in tutored problem solving. Instructional Sicience, 38, 289--307.
  13. Corbalan, G., Kester, L. and van Merrieonboer, J.J.G. (2008). Selecting learning tasks: Effects of adaptation and shared control on learning efficiency and task involvement. Contemporary Educational Psycholoy, 33, 733--756.
  14. Mitrovic, A., and Martin, B. (2004). Evaluating adaptive problem selection. In Adaptive hypermedia and adaptive web-based systems (pp. 185-194). Springer Berlin Heidelberg.
  15. Donnelly, C.J. (2015). Enhancing Personalization Within ASSISTments (Doctoral dissertation).

External Links