Chen Et Al.
Chen Et Al.
www.wileyonlinelibrary.com
Background. The worked example effect in cognitive load theory suggests that
providing worked examples first followed by solving similar problems would facilitate
students’ learning. Using problem solving–worked example sequence is another way of
implementing example-based instruction. Although research has demonstrated the
superiority of worked example–problem solving sequence on learning materials that
presumably are high in element interactivity for novices, none of the previous studies have
compared the two sequences with levels of element interactivity experimentally
manipulated in a strictly controlled manner.
Aim. The reported study aimed to investigate the effects of levels of element
interactivity of the learning tasks and levels of learner prior knowledge on the
effectiveness of two alternative example-based sequences, worked example–problem
solving versus problem solving–worked example.
Sample. Fifty-two Year five students, around 10 to 11 years old, from a primary school
in Indonesia participated in Experiment 1, and 96 Year eight students, around 13 to
14 years old, from a secondary school in Indonesia participated in Experiment 2.
Methods. 2 (sequences: worked example–problem solving vs. problem solving–worked
example) 9 2 (levels of element interactivity: low vs. high) experimental design, with the
second factor repeatedly measured, was used in the two experiments conducted with
learners at different levels of prior knowledge.
Result. The results showed the advantage of using worked example–problem solving
sequence for learning materials high in element interactivity, especially for novice
learners, whereas there were no differences between the worked example–problem
solving and problem solving–worked example sequences for learning materials low in
element interactivity for more knowledgeable learners.
Conclusion. This study not only replicated the results of previous studies, but also
extended their findings by experimentally manipulating levels of element interactivity of
learning materials.
*Correspondence should be addressed to Endah Retnowati, Faculty of Mathematics and Sciences, Yogyakarta State University,
Colombo St. No. 1, Karang Malang, Sleman, Yogyakarta 55281, Indonesia (email: [email protected]).
DOI:10.1111/bjep.12317
Worked example–problem solving and problem solving–worked example 211
worked example conditions. The results suggested that the completion of partial worked
example and worked example conditions were both superior to conventional problem-
solving condition with less time and effort. The similar findings were also demonstrated in
the areas of physics (Van Gog et al., 2011) and geometry (Paas & Van Merri€enboer, 1994;
Schwonke et al., 2009). Kirschner, Sweller, and Clark (2006) also strongly argued for the
effectiveness of using worked examples rather than problem-based approaches in
teaching, especially for novice learners.
However, the effectiveness of using worked example–problem solving sequence may
be moderated by the levels of element interactivity. The next section will discuss this issue
in details.
Element interactivity
Element interactivity is the core concept in cognitive load theory. Interactive elements are
the interconnected elements of information that must be processed simultaneously in
working memory in order to achieve understanding (Sweller et al., 2011).
mathematics. In their experiments, two instructional effects, the worked example and
generation effects, were tested. When testing for the generation effect, a list of simple geometry
formulas (low in element interactivity) was used: students either generated or read formulas.
The results showed preference for the generation group in memorizing the list of formulas.
When testing for the worked example effect, students were required to apply some of
the geometry formulas in the list to calculate the area of compound shapes (high element
interactivity tasks). Some of the participants were presented with a worked example
followed by a similar problem to solve, while others were required to solve the equivalent
number of problems without any worked examples. The results showed advantages of the
worked example–problem solving sequence, indicating a worked example effect.
Therefore, for materials low in element interactivity, generation was superior to worked
examples, whereas for materials high in element interactivity, the result was reversed.
Since levels of learner expertise affect the levels of element interactivity, when testing
the same sets of materials with more knowledgeable learners, all materials became low in
element interactivity. Accordingly, the results indicated the generation effect for all sets of
materials with expertise reversal effects found across multiple experiments. It is very
likely that a similar kind of interaction exists between the effectiveness of worked
example–problem solving, problem solving–worked example sequences and levels of
element interactivity (which are affected by levels of learner expertise).
Role of expertise
According to the expertise reversal effect, the worked example effect is influenced by the
level of learner expertise. With increased levels of learner knowledge in the task domain,
the initial advantage of worked examples disappears, and with sufficiently high levels of
learner expertise, problem solving may become even more effective than studying
214 Ouhao Chen et al.
worked examples (Kalyuga, 2007). For example, Reisslein et al. (2006) investigated how
learner expertise affected the effectiveness of different instructional sequences (worked
example–problem solving vs. problem solving–worked example). The results supported
an expertise reversal effect: novices benefitted more from the worked example–problem
solving sequence, compared to more knowledgeable learners who benefitted more from
the problem solving–worked example sequence.
Present study
The experiments reported in this paper conceptually replicated the studies of Van Gog
et al. (2011) and Reisslein et al. (2006) on comparing the alternative example-based
instructional sequences: worked example–problem solving versus problem solving–
worked example. However, the current study added a new experimental factor, element
interactivity, into consideration. None of the previously conducted studies compared the
two sequences from this perspective, with levels of element interactivity experimentally
manipulated in a strictly controlled manner.
As mentioned above, levels of element interactivity are closely related to levels of
learner expertise; therefore, it could be assumed that the worked example–problem
solving sequence would be superior to the problem solving–worked example sequence
for materials high in element interactivity (when novice learners are recruited; Hypothesis
1), but this advantage would disappear or reverse for materials low in element interactivity
(when expert learners are recruited; Hypothesis 2).
Experiment 1
Participants
Fifty-two Year five students, around 10 to 11 years old, from one primary school in
Indonesia participated in this experiment. In the first phase of this study (low element
interactivity tasks), 28 students were randomly assigned to the worked example–problem
solving sequence, and the rest 24 students were assigned to the problem solving–worked
example sequence (See Figure 1). In the second phase of this study (high element
interactivity tasks), half of the students in the Phase 1 worked example–problem solving
sequence and half of the students in the problem solving–worked example sequence
were randomly assigned to a new worked example–problem solving sequence in Phase 2,
while the rest of students were assigned to a new problem solving–worked example
sequence in Phase 2 (See Figure 2). Four students from the problem solving–worked
example group in Phase 2 were excluded from the analysis, as they did not complete the
whole procedure. All students were novices in geometry tasks used in this study.
Materials
Learning materials
There were two sets of materials designed for each phase of this study separately. For the
first phase, materials were 11 geometry formulas for calculating area of simple geometric
shapes, such as the Area of a Circle. In the worked example–problem solving group, the
students studied the list of formulas without pen/pencil, and then, they generated the
formulas in writing, whereas in the problem solving–worked example group, students
generated the formulas in writing first, followed by studying the same list of formulas
Worked example–problem solving and problem solving–worked example 215
Studied 11 formulas
Cued-recall test
without pen/pencil. In Phase 2, the materials included tasks about calculating the area of
composite shapes by applying formulas studied in Phase 1. Two pairs of worked example–
problem solving and problem solving–worked example tasks were designed. For each
pair of worked example–problem solving tasks, an example showing the full procedure
for calculating the area of a composite shape (see Figure 3 for an example) was followed
by a similar problem to solve; for each pair of problem solving–worked example tasks, the
only difference was the order of presenting the worked example and the similar problem.
Reviewed 11 formulas
Overview of the compound
shapes
Post-content test
trapezium and addition, multiplication, and division by 2, that together added nine
elements. Finally, adding together those two separate area values involved another
element. Therefore, in total, around 18 interactive elements were involved in this task,
making it high in element interactivity, compared to the formulas in Phase 1.
Worked example–problem solving and problem solving–worked example 217
Post-content tests
A cued-recall test (for materials low in element interactivity) was designed and
administered at the end of Phase 1. The order of the 11 geometry formulas in the cued-
recall test was randomized, based on the order used in the learning phase. A post-content
test (for materials high in element interactivity) with five questions on calculating the area
of composite shapes (similar to the problems in Phase 2 learning) was designed and
administered at the end of Phase 2.
Procedure
This joint study was approved by the University ethics committee of the Institute of
Research and Community Service. The approval numbers are 378/UN34.21/TU/2018 for
Experiment 1 and 817/UN34.21/TU/2018 for Experiment 2.
In Phase 1, students were randomly allocated to the alternative experimental
sequences (worked example–problem solving or problem solving–worked example; See
Figure 1) and sat in different rooms (10 min for moving class). All students studied 11
formulas first (10 min). Then, the students in the worked example–problem solving
group were presented the same 11 formulas again (10 min), followed by generating these
formulas (10 min), whereas students in the problem solving–worked example group
were required to generate the 11 formulas first (10 min), followed by the presentation of
these formulas (10 min). A cued-recall test of the 11 formulas was conducted at the end of
this phase (8 min), the Cronbach a = .80.
In Phase 2, all participants had been firstly randomly re-allocated to new experimental
groups, as described in the Participants section (See Figure 2). All the students were
again presented with the same 11 formulas as in Phase 1 as a review, and the researcher
provided an introduction to the compound shapes as combinations of common shapes.
The students were shown two examples of compound shapes (10 min). After that,
students in different groups sat in different rooms (10 min for moving class). The students
in the worked example–problem solving group were presented the first pair of worked
example–problem solving tasks, followed by the second pair, whereas the students in the
problem solving–worked example group were presented the first pair of problem
solving–worked example tasks, followed by the second pair (20 min). The only difference
between the two groups was the order of presenting the worked example and the
problem. Finally, a post-content test, including questions that were similar to those used
during the Phase 2 learning, on calculating the area of compound shapes was conducted
(10 min), the Cronbach a = .84.
Scoring
When scoring the cued-recall test, 1 mark was given to each correctly recalled formula;
therefore, the total of 11 marks could be awarded for the cued-recall test. For the second
phase test, three marks were given to each question: 1 mark for correctly calculating the
area of one separated shape; 1 mark for correctly calculating the area of another separated
shape; and 1 mark for adding the two separate areas together. Therefore, the total of 15
marks for five questions could be awarded for the test. The carry-over errors were
considered when marking the test. For example, if the area of the first shape was
calculated incorrectly (0 mark), but the area of the second shape was correct (1 mark) and
the addition of the two values was performed correctly (1 mark), even if the final answer
was incorrect, two marks were allocated for the task.
218 Ouhao Chen et al.
Results
A 2 (sequences: worked example–problem solving vs. problem solving–worked
example) 9 2 (levels of element interactivity: low vs. high) repeated measures ANOVA
was conducted with the second factor repeatedly measured. The raw test scores for both
post-tests were converted to the percentage correct scores (Table 1).
The main effect of sequence was not significant, F(1, 46) = .005, MSE = 729.56,
p = .944, g2p = 0. The main effect of element interactivity was also not significant, F(1,
46) = .144, MSE = 488.72, p = .706, g2p = .003. However, the interaction between
sequence and element interactivity was significant, F(1, 46) = 8.291, MSE = 488.72,
p = .006, g2p = .153. Following the significant interaction, simple effect analyses were
conducted. For materials that were low in element interactivity (the cued-recall test),
there was no significant difference between the sequences, t(47) = 1.45, SEdiff = 7.486,
p = .154, d = .41, whereas, for materials that were high in element interactivity (the post-
content test), the effect of sequence was significant, t(47) = 2.15, SEdiff = 6.913,
p = .037, d = .59, indicating that the worked example–problem solving sequence was
superior to the problem solving–worked example sequence.
Experiment 1 demonstrated the superiority of the worked example–problem solving
sequence with novice learners, but only for materials high in element interactivity (area of
composite shapes). For materials low in element interactivity (formulas for simple
common geometric shapes), the superiority of the problem solving–worked example
sequence was not found (although the effect size of 0.41 indicated a possibility of the
effect). Thus, for novice learners, the superiority of the worked example–problem solving
sequence for materials high in element interactivity disappeared with materials low in
element interactivity.
Since the level of element interactivity depended on levels of learner expertise,
materials high in element interactivity for novices could become low in element
interactivity for relatively more experienced learners. Therefore, in Experiment 2, more
knowledgeable learners were used, with the same sets of materials, to further compare the
effectiveness of worked example–problem solving and problem solving–worked exam-
ple sequences.
Experiment 2
Participants
Ninety-six Year eight students, around 13 to 14 years old, from one secondary school in
Indonesia participated in this experiment. In the first phase of this study, 48 students were
randomly assigned to the worked example–problem solving sequence, and the rest 48
students were assigned to the problem solving–worked example sequence (Figure 1). In
the second phase of this study, exactly the same procedure as in Experiment 1 was used
for randomly re-allocating participants to new experimental groups (Figure 2). Four
students from the problem solving–worked example group in Phase 2 were excluded
from the analysis, as they did not complete the whole procedure. All students had been
previously taught the type of geometry tasks that were used in this study and therefore
were considered to be relatively knowledgeable in the task domain.
Materials
All materials were identical to those used in Experiment 1. However, as students in this
experiment had been previously instructed on how to calculate the area of composite
shapes, they could retrieve the relevant schemas as single entities. Therefore, the
materials used in both phases of this experiment were low in element interactivity for the
participants.
Results
A 2 (sequences: worked example–problem solving vs. problem solving–worked
example) 9 2 (levels of element interactivity: low vs. high) repeated measures ANOVA
was conducted with the second factor repeatedly measured. The raw post-tests scores
were converted to the percentage correct scores (Table 2).
The effect of sequence was not significant, F(1, 90) = .233, MSE = 351.11, p = .631,
g2p = .003. The effect of element interactivity was not significant, F(1, 90) = 1.814,
MSE = 368.80, p = .181, g2p = .020. The interaction between sequence and element
interactivity was also not significant, F(1, 90) = 2.152, MSE = 368.80, p = .146,
g2p = .023. Therefore, the results confirmed that for materials low in element interactivity
(geometric formulas and area of composite shapes) in both phases of Experiment 2, there
were no significant differences between the worked example–problem solving and
problem solving–worked example sequences. The superiority of the worked example–
problem solving sequence for high element interactivity tasks found in Experiment 1 with
novices disappeared in Experiment 2 with more knowledgeable learners, which might
indicate an expertise reversal effect.
General discussion
This randomized, controlled experimental study was designed to investigate relations
between the effectiveness of worked example–problem solving and problem solving–
conditions under which this approach becomes preferable to the worked example–
problem solving instructional sequence. In relation to this, one major limitation of this
study was that the materials were mostly associated with learning procedural knowledge.
There is a strong indication though, originating from research in productive failure and
invention learning approaches (DeCaro & Rittle-Johnson, 2012; Kapur, 2014; Loibl &
Rummel, 2014a,b), that problem solving–worked example sequences could be beneficial
for learning conceptual knowledge by novice learners. It is possible to speculate that in
case of conceptual knowledge, the initial problem solving attempts, even if unsuccessful,
could familiarize learners with some essential elements of the task, at least in isolated
rather than interactive form and without much meaning assigned at this stage. Such initial
rudimentary learning could allow reducing working memory load during the following
comprehensive explicit instruction phase. In cognitive load theory, this method has been
associated with the so-called isolated-interactive elements effect (Sweller et al., 2011).
Future empirical studies within a cognitive load framework may need to take a closer look
at this matter when investigating the effectiveness of alternative sequences of explicit
instruction and problem solving for learning conceptual knowledge.
Other limitations of this study are the absence of delayed post-tests and differential
retention and transfer tasks in the post-tests. Judging by the results of studies within
productive failure and invention learning frameworks, problem-first sequences might be
most effective when evaluated by delayed transfer tests. Such tests need to be included in
future studies, together with thorough analyses and comparisons of levels of element
interactivity of learning tasks.
In summary, the major theoretical implication of this study is that by considering the
factor of element interactivity, the results of the two reported experiments provide a
potential way of explaining the contradictory findings within different theoretical
frameworks in relation to the effectiveness of alternative sequences of worked examples
and problem solving tasks. For learning materials high in element interactivity, the worked
example–problem solving sequence is likely to be more effective, whereas, for materials
low in element interactivity, the problem solving–worked example sequence could be
superior.
Educational implications
The results of this study may inform educational practitioners that novice learners dealing
with complex information that is difficult to process in limited-capacity working memory
should initially be presented with sufficient instructional guidance, such as that provided
by worked examples. With increases in learner levels of expertise, a gradual shift to a
heavier emphasis on problem solving may be beneficial. Also, the teachers need to
carefully consider both the nature of learning materials and the levels of learner prior
knowledge when using worked example–problem solving and problem solving–worked
example instructional sequences.
Acknowledgement
Authors would like to thank the teachers and students, from SMP Negeri 5 Depok, Yogyakarta,
Indonesia, and SD Negeri Percobaan 2, Yogyakarta, Indonesia, participated in our experiments.
Also, authors would like to thank Marissa and Ibrohim, from Yogyakarta State University,
Yogyakarta, Indonesia, assisted in marking and data entry.
222 Ouhao Chen et al.
Ethics statement
This joint study is approved by the ethics committee at the institute of Research and
Community Service, Yogyakarta State University, Yogyakarta, Indonesia. The approval
numbers are 378/UN34.21/TU/2018 for Experiment 1 and 817/UN34.21/TU/2018 for
Experiment 2.
References
Chen, O., Kalyuga, S., & Sweller, J. (2015). The worked example effect, the generation effect, and
element interactivity. Journal of Educational Psychology, 107, 689–704. https://doi.org/10.
1037/edu0000018
Chen, O., Kalyuga, S., & Sweller, J. (2016a). Relations between the worked example and generation
effects on immediate and delayed tests. Learning and Instruction, 45, 20–30. https://doi.org/
10.1016/j.learninstruc.2016.06.007
Chen, O., Kalyuga, S., & Sweller, J. (2016b). When instructional guidance is needed. The
Educational and Developmental Psychologist, 33, 149–162. https://doi.org/10.1017/edp.
2016.16
Chen, O., Kalyuga, S., & Sweller, J. (2017). The expertise reversal effect is a variant of the more
general element interactivity effect. Educational Psychology Review, 29, 393–405. https://doi.
org/10.1007/s10648-016-9359-1
Chen, O., Retnowati, E., & Kalyuga, S. (2019). Effects of worked examples on step performance in
solving complex problems. Educational Psychology, 39, 188–202. https://doi.org/10.1080/
01443410.2018.1515891
Cooper, G., & Sweller, J. (1987). Effects of schema acquisition and rule automation on mathematical
problem-solving transfer. Journal of Educational Psychology, 79, 347–362. https://doi.org/10.
1037/0022-0663.79.4.347
Cowan, N. (2001). Metatheory of storage capacity limits. Behavioral and Brain Sciences, 24, 154–
176.
DeCaro, M. S., & Rittle-Johnson, B. (2012). Exploring mathematics problems prepares children to
learn from instruction. Journal of Experimental Child Psychology, 113, 552–568. https://doi.
org/10.1016/j.jecp.2012.06.009
Ericsson, K. A., & Kintsch, W. (1995). Long-term working memory. Psychological Review, 102,
211–245. https://doi.org/10.1037/0033-295X.102.2.211
Glogger-Frey, I., Fleischer, C., Gr€ uny, L., Kappich, J., & Renkl, A. (2015). Inventing a solution and
studying a worked solution prepare differently for learning from direct instruction. Learning
and Instruction, 39, 72–87. https://doi.org/10.1016/j.learninstruc.2015.05.001
Glogger-Frey, I., Gaus, K., & Renkl, A. (2017). Learning from direct instruction: Best prepared by
several self-regulated or guided invention activities? Learning and Instruction, 51, 26–35.
https://doi.org/10.1016/j.learninstruc.2016.11.002
Hsu, C.-Y., Kalyuga, S., & Sweller, J. (2015). When should guidance be presented in physics
instruction? Archives of Scientific Psychology, 3, 37–53. https://doi.org/10.1037/arc0000012
Kalyuga, S. (2007). Expertise reversal effect and its implications for learner-tailored instruction.
Educational Psychology Review, 19, 509–539. https://doi.org/10.1007/s10648-007-9054-3
Kapur, M. (2014). Productive failure in learning math. Cognitive Science, 38, 1008–1022. https://d
oi.org/10.1111/cogs.12107
Kirschner, P. A., Sweller, J., & Clark, R. E. (2006). Why minimal guidance during instruction does not
work: An analysis of the failure of constructivist, discovery, problem-based, experiential, and
inquiry-based teaching. Educational psychologist, 41, 75–86. https://doi.org/10.1207/
s15326985ep4102_1
Leppink, J., Paas, F., Van Gog, T., van Der Vleuten, C. P., & Van Merrienboer, J. J. (2014). Effects of
pairs of problems and examples on task performance and different types of cognitive load.
Learning and Instruction, 30, 32–42. https://doi.org/10.1016/j.learninstruc.2013.12.001
Worked example–problem solving and problem solving–worked example 223
Loibl, K., & Rummel, N. (2014a). The impact of guidance during problem-solving prior to instruction
on students’ inventions and learning outcomes. Instructional Science, 42, 305–326. https://doi.
org/10.1007/s11251-013-9282-5
Loibl, K., & Rummel, N. (2014b). Knowing what you don’t know makes failure productive.
Learning and Instruction, 34, 74–85. https://doi.org/10.1016/j.learninstruc.2014.08.004
Miller, G. A. (1956). The magical number seven, plus or minus two: Some limits on our capacity for
processing information. Psychological Review, 63, 81–97. https://doi.org/10.1037/h0043158
Paas, F. (1992). Training strategies for attaining transfer of problem-solving skill in statistics: A
cognitive-load approach. Journal of Educational Psychology, 84, 429–434. https://doi.org/10.
1037/0022-0663.84.4.429
Paas, F., & Van Merri€enboer, J. J. (1994). Variability of worked examples and transfer of geometrical
problem-solving skills: A cognitive load approach. Journal of Educational Psychology, 86, 122–
133. https://doi.org/10.1037/0022-0663.86.1.122
Peterson, L., & Peterson, M. J. (1959). Short-term retention of individual verbal items. Journal of
Experimental Psychology, 58, 193–198. https://doi.org/10.1037/h0049234
Reisslein, J., Atkinson, R. K., Seeling, P., & Reisslein, M. (2006). Encountering the expertise reversal
effect with a computer-based environment on electrical circuit analysis. Learning and
Instruction, 16, 92–103. https://doi.org/10.1016/j.learninstruc.2006.02.008
Schwonke, R., Renkl, A., Krieg, C., Wittwer, J., Aleven, V., & Salden, R. (2009). The worked-example
effect: Not an artefact of lousy control conditions. Computers in Human Behavior, 25, 258–
266. https://doi.org/10.1016/j.chb.2008.12.011
Sweller, J., Ayres, P., & Kalyuga, S. (2011). Cognitive load theory. New York, NY: Springer.
Sweller, J., & Chandler, P. (1994). Why some material is difficult to learn. Cognition and Instruction,
12, 185–233. https://doi.org/10.1027/s1532690xci1203_1
Sweller, J., & Cooper, G. (1985). The use of worked examples as a substitute for problem solving in
learning algebra. Cognition and Instruction, 2, 59–89.
Sweller, J., & Paas, F. (2017). Should self-regulated learning be integrated with cognitive load theory?
A commentary. Learning and Instruction, 51, 85–89. https://doi.org/10.1016/j.learninstruc.
2017.05.005
Sweller, J., & Sweller, S. (2006). Natural information processing systems. Evolutionary Psychology,
4, 434–458. https://doi.org/10.1177/147470490600400135
Tindall-Ford, S., Chandler, P., & Sweller, J. (1997). When two sensory modes are better than one.
Journal of Experimental Psychology: Applied, 3, 257–287. https://doi.org/10.1037/1076-
898X.3.4.257
Van Gog, T., Kester, L., & Paas, F. (2011). Effects of worked examples, example-problem, and
problem-example pairs on novices’ learning. Contemporary Educational Psychology, 36, 212–
218. https://doi.org/10.1016/j.cedpsych.2010.10.004