Catapult Experiment: Design of Experiments (Doe) and Response Surface Methods (RSM)
Catapult Experiment: Design of Experiments (Doe) and Response Surface Methods (RSM)
(RSM)
Module Coursework
Catapult Experiment
The Catapult Experiment consist of a small wooden toy catapult that can be fired in the
classroom that has a variety of setting factors – which makes it an excellent example for an
academic exercise to apply the DoE and RSM tools.
The methodology outlines the sequential strategy that will be followed – that starts with the
screening experiment (planning, execution and analysis), and the follow-up experiment
(planning, execution and analysis) to inform which factors have the greatest influence on the
ball firing response and should be used on the prediction model. Once the model is
constructed, it can be used to make predictions whiting the confidence interval.
As such, Elastic Position (most statistically significant effect), Pull Back Angle and Stop
Position were identified as the factors with the greatest effect on the response. The model
accuracy was +/- 159 mm.
Conclusions and recommendations for future work and discussions conclude the report,
including personal reflections.
Figure 1. The operator and the measurer for the experiments – including the Catapult
Experiment setup and factors: Pull-back angle (A), Hold time in armed position (B), Elastic
Twists (C), Elastic Position - Fixing Arm (D), Stop position (E) ............................................... 9
Figure 2. Available Factorial Designs – Minitab...................................................................... 10
Figure 3. A face cantered central composite design [8] ............................................................12
Figure 4. The Fractional Factorial Design and results (coded) ................................................13
Figure 5. The Fractional Factorial Design Factors and Levels ..................................................14
Figure 6. Main effects plot - screening experiment (coded) .....................................................14
Figure 7. Interaction Plot - screening experiment .................................................................... 15
Figure 8. Surface Plot - screening experiment ..........................................................................16
Figure 9. Pareto Chart of the Effects (α = 0.05) - screening experiment ................................. 17
Figure 10. Pareto Chart of the Effects (α = 0.15) - screening experiment ................................ 17
Figure 11. Half Normal Plot (α = 0.05) - screening experiment .............................................. 18
Figure 12. Face Centred Composite Experiment Model ...........................................................19
Figure 13. Normal Probability Plot – Follow-up Experiment ...................................................19
Figure 14. Deleted Residual vs Fit Plot – Follow-up Experiment ............................................ 20
Figure 15. Pareto Chart – Follow-up Experiment .................................................................... 20
Figure 16. Normal Probability Plot, Deleted Residuals, Pareto Chart and Half Normal –
Follow-up Experiment Re-run ................................................................................................. 23
Figure 17. Surface Plots – Follow-up Experiment ................................................................... 24
Figure 18. Interaction Plots – Follow-up Experiment ............................................................. 24
Figure 19. Contour Plots – Follow-up Experiment .................................................................. 25
Figure 20. The Fractional Factorial Design and results (not coded) ....................................... 32
Figure 21. The Fractional Factorial Design Factors and Levels (not coded) ........................... 32
Figure 22. Main effects plot - screening experiment (un-coded) ............................................ 32
Figure 23. Interaction Plot - screening experiment ................................................................. 33
Figure 24. Surface Plot - screening experiment ....................................................................... 33
Figure 25. Half Normal Plot (α = 0.15) - screening experiment ............................................. 34
Figure 26. Surface Plots – Follow-up Experiment ................................................................... 34
Figure 27. Interaction Plots – Follow-up Experiment ............................................................. 34
Figure 28. Contour Plots – Follow-up Experiment ................................................................. 35
Figure 29. Normal Probability Plot – Follow-up Experiment Re-run ..................................... 35
Figure 30. Deleted Residual vs Fit Plot – Follow-up Experiment Re-run ............................... 35
Figure 31. Pareto Chart – Follow-up Experiment – Re-run .................................................... 36
Table of Tables
Table of Equations
Glossary of Notations
Acronym Meaning
1.1 Background
The catapult has the advantage that is has multiple factors that can be adjusted to control the
firing distance response – which makes it very suitable to use as a case study for these
learning outcomes.
The Objectives are to conduct analysis on the catapult experiment by using advanced DoE
and RSM statistical analysis that will serve future BB project – as this is an academic exercise
for developing the candidate’s practical skills in applying the statistical tools.
The Methodology used is looking to select and make use of the most appropriate statistical
tools and deploy a sequential methodologies to plan, run and analyse experiments, with the
appropriate justification to support the engineering problem.
Following the scope, in terms of Limitations and Constraints, the major limitation
comes from the small number of experiments that was carried out in the class. Due to time
constrains, not understanding the importance or difficulty – not enough repeats where
carried out.
There are many benefits of knowing the response surface for a product or process. It is
advisable to follow a sequential strategy for experimentation, which is what we will do in this
report: identify trends with a two-level experiment and then run a multi-level experiment
aiming to model the response surface accurately with a smaller number of factors.
The response in our case is the total horizontal distance travelled by the ball in flight -
the projectile distance of the catapult. Initially, we will include 5 quantitative factors in
the experiment and we will conduct a 2-level experiment, followed by the selection of 3 of
them that have the strongest interaction and effect, which we will investigate in detail in a
multi-level experiment.
Varying the factor levels allows the identification of the largest effects on the response. The
factors levels where selected from the possible setting options and by using engineering
knowledge. The tensioning peg was not selected as one of the factors due to the drastic effect
on the catapult function.
The 5 selected factors and the 2 initial levels or settings for the experiment are presented in
Table 2, with the corresponding coding level.
Several noise factors where identified, which are presented in the table below (Table 3).
Figure 1. The operator and the measurer for the experiments – including the Catapult Experiment
setup and factors: Pull-back angle (A), Hold time in armed position (B), Elastic Twists (C), Elastic
Position - Fixing Arm (D), Stop position (E)
Based on the above described factors and the potential noise factors, the following
operational definition was applied in the experiment, which was carried out by the team
using the roles and responsibilities highlighted in Table 4.
1. Coordinator reads the factor levels from the experimental design in Minitab
2. The operator follows the set up instructions for the defined levels:
a. Sets the firing parameter levels (Coordinator verifies the set up levels)
b. Re-positions the catapult in the marked spot
c. Securely holds it down
d. Arms the catapult arm and load the ball
e. Waits for fire instructions from timekeeper
f. Fires catapult at the timekeeper instructions in the same way was for each run
g. Releases elastic to allow stretch and resets catapult in neutral position
3. The measurer measures the distance the ball travelled using the tape meter and the
instructions from the roles and responsibilities
4. The coordinator records the measurement in Minitab
5. Repeat for each of the factors levels
Coding / standardising
Coding or standardising the analysis factors means to use -1 to represent the lowest observed
value and to use +1 to represent the highest, whilst coding the intermediate values by linear
interpolation. This is beneficial in several ways, such as giving the ability to compare and
interpret the parameter estimates, the independency of the units and scale of the parameter.
[6] Also, the intercept is now a meaningful number that is the predicted response of the
travelled distance when all x’s are at coded 0 – which is at the ‘centre’ of the data. [7]
Next, we need to select a factorial design that includes the number of levels and factors (5
factors with 2 levels). When selecting a design, we aim to minimise the number of runs that
we have to carry out, while still maintaining the factors list.
This experiment is designed to fit a 2nd order surface, which was not done in the screening
experiment, which will include the quadratic terms for a more accurate curved response
surface.
The aim of this experiment is to fit the 2nd order response surface that can be used to predict,
and optimise the projectile distance of the ball from the catapult.
Table 5 presents the factors and the levels – with the codding for all 3 levels, including the
middle setting which is codded at 0.
The other 2 discarded factors are keep constant at the +1 positions - Hold time in armed
position (B) = 15 sec and Elastic Twists (C) = 10.
Central Composite (CC) design was chosen, which gave us 17 runs in total.
The below extract from Minitab shows the CC design set up. The design contains 3 factors, on
the basis of one replicate, with 3 centre points in cube.
Factors: 3 Replicates: 1
Base runs: 17 Total runs: 17
Base blocks: 1 Total blocks: 1
Cube points: 8
Center points in cube: 3
Axial points: 6
Center points in axial: 0
α: 1
PRESS
Equation 2. Press RSME = √
Number of Runs
Using this prediction interval, we will obtain a range in which our response will fall within
given confidence level. To obtain the prediction, we will use Minitab.
For each experiment run, we repeated the measurement 3 times and calculated the Average –
as presented in Figure 4 (uncoded results Figure 20 in Annexes). The Average was used as
the response variable. This was done in order to ensure that the confidence intervals obtained
are greater, whilst still having an efficient design.
Alias Structure
I + ABCDE
A + BCDE
B + ACDE
C + ABDE
D + ABCE
E + ABCD
AB + CDE
AC + BDE
AD + BCE
AE + BCD
BC + ADE
BD + ACE
It is easy to observe from the plot that factors Hold Time and the Twists have very little
effect on the distance travelled by the ball, and therefore it can be concluded that these
factors can be removed from the next phase of the experiment.
The remaining 3 factors – Pull Back Angle, Elastic Position and Stop Position appear
to have large magnitude of effect on the response as per the slope steepness.
1100 1500
1000
ver age ver age
1000
900
1 1
800
0 500 0
-1 St op Pos it ion -1 St op Pos it ion
0 -1 0 -1
1 1
Twis t s Elast ic Pos it ion
We can observe the magnitude and importance of effects – with D (Elastic Position) having
the largest effect by far, followed by A and E as main factors and interactions (DE, AE, AD,
BC, CD) up to the reference line. For α = 0.05, the reference line is 58.3 and for α = 0.15, the
reference line is 38.6.
The limitations of the Pareto chart are due to the fact that the absolute value of the effect is
given, but not the ability to determine which effects increase or decrease the response.
Above we used two values for α – the conventional value of 0.05 (strict) and the proposed
value for the analysis of 0.15 (less strict – industry experience) – in order to help to
determine which effects are statistically significant – by using P-values. If P-value is greater
than α, then the effect is not significant and can be removed, and if the opposite is true, the
effect is significant.
In this analysis, moving forward, α = 0.05 will be used, with the results for α = 0.15 being
presented in the annexes.
Similar to Pareto Plots, from this plots we cannot determine the effect of the factor in the
response – increase or decrease.
In the analysis above we uses 2 different SE alpha α level of 5% and 15% (P-value 0.15) using
Length’s pseudo-standard error (PSE), the second largest level allows for the capturing all
statistically significant occurrences (Length’s PSE =22.6875).
Two different type of errors are possible with Half Normal Plots - false positives (type 1 error)
and false negatives (type 2 error) – i.e. points that are off the line by chance and small real
effects that cannot be isolated from actual random effects.
In conclusion, based on all the plots analysed above, we can confirm that the Number of
Twists and Hold time have negligible effect on the result and can be removed from the
analysis.
Elastic Position is confirmed as the factor which most statistically significant effect, and Pull
Back Angle and Stop Position are similar magnitude, all 3 will be used in the Follow-up
Experiment.
The Average of three measurements was used as the response variable. This was done in
order to ensure that the confidence intervals obtained are greater, whilst still having an
efficient design.
There are 2 more interactions – AC and BB – that are close to the limit, but a decision was
made for those to be left in and retest.
Pull-Back Angle (A) 147.2 18.6 (103.3, 191.1) 7.93 Pull-Back Angle (A) 0.000 1.00
Stop Position (E) -95.2 18.6 (-139.1, -51.3) -5.13 Stop Position (E) 0.001 1.00
Elastic Position (D) 269.8 18.6 (225.9, 313.7) 14.54 Elastic Position (D) 0.000 1.00
Pull-Back Angle (A)*Pull-Back Angle (A) -29.9 35.9 (-114.7, 54.9) -0.83 Pull-Back Angle (A)*Pull-Back Angle (A) 0.432 1.54
Stop Position (E)*Stop Position (E) -128.2 35.9 (-213.0, -43.4) -3.58 Stop Position (E)*Stop Position (E) 0.009 1.54
Elastic Position (D)*Elastic Position (D) 0.1 35.9 (-84.7, 84.9) 0.00 Elastic Position (D)*Elastic Position (D) 0.998 1.54
Pull-Back Angle (A)*Stop Position (E) -107.3 20.8 (-156.4, -58.2) -5.17 Pull-Back Angle (A)*Stop Position (E) 0.001 1.00
Pull-Back Angle (A)*Elastic Position (D) 66.5 20.8 (17.4, 115.5) 3.20 Pull-Back Angle (A)*Elastic Position (D) 0.015 1.00
Stop Position (E)*Elastic Position (D) -114.8 20.8 (-163.9, -65.7) -5.53 Stop Position (E)*Elastic Position (D) 0.001 1.00
Regression Equation in Uncoded Units
Average = -18489 + 149 Pull-Back Angle (A) + 2927 Stop Position (E)- 201 Elastic Position (D) - 0.299 Pull-Back Angle (A)*Pull-
Distance Back Angle (A)- 128.2 Stop Position (E)*Stop Position (E) + 0.03 Elastic Position (D)*Elastic Position (D) - 10.73 Pull-
Back Angle (A)*Stop Position (E)+ 3.32 Pull-Back Angle (A)*Elastic Position (D) - 57.4 Stop Position (E)*Elastic Position (D)
Fits and Diagnostics for Unusual Observations
Average
Obs Distance Fit SE Fit 95% CI Resid Std Resid Del Resid HI Obs Cook’s D DFITS
4 646.7 582.8 52.3 (459.0, 706.5) 63.9 2.40 5.32 0.794718 4 2.24 10.4736 R
5 951.7 1018.4 52.3 (894.7, 1142.1) -66.8 -2.51 -7.35 0.794718 5 2.44 -14.4645 R
R Large residual
As presented in the methodology, PRESS RMSE will be calculated using Equation 2. We can
observe that PRESS is quite high and the number of runs quite low – which means we need
to be cautious when using Press RSME, however, it will be recalculated after the re-run and it
is expected to decrease.
PRESS 512426
Press RSME = √Number of Runs = √ 17
= 173.6165 mm (Model Confidence Interval)
From the model summary, we can see that the R-sq value dropped (from 98.25% to 98.06%),
the PRESS reduced from 512426 to 433896.
Pull-Back Angle (A) 147.2 17.3 (108.1, 186.2) 8.53 0.000 Pull-Back Angle (A) 1.00
Stop Position (E) -95.2 17.3 (-134.2, -56.1) -5.51 0.000 Stop Position (E) 1.00
Elastic Position (D) 269.8 17.3 (230.8, 308.9) 15.63 0.000 Elastic Position (D) 1.00
Stop Position (E)*Stop Position (E) -143.5 26.9 (-204.4, -82.7) -5.34 0.000 Stop Position (E)*Stop 1.00
Position (E)
Pull-Back Angle (A)*Stop Position (E) -107.3 19.3 (-150.9, -63.6) -5.56 0.000 Pull-Back Angle 1.00
(A)*Stop Position (E)
Pull-Back Angle (A)*Elastic Position (D) 66.5 19.3 (22.8, 110.1) 3.44 0.007 Pull-Back Angle 1.00
(A)*Elastic Position (D)
Stop Position (E)*Elastic Position (D) -114.8 19.3 (-158.4, -71.1) -5.95 0.000 Stop Position 1.00
(E)*Elastic Position (D)
Regression Equation in Uncoded Units
Average = -10104 + 47.66 Pull-Back Angle (A) + 3049 Stop Position (E)- 200 Elastic Position (D)
Distance - 143.5 Stop Position (E)*Stop Position (E)- 10.73 Pull-Back Angle (A)*Stop Position (E) + 3.323 Pull-
Back Angle (A)*Elastic Position (D) - 57.40 Stop Position (E)*Elastic Position (D)
Fits and Diagnostics for Unusual Observations
Average
Obs Distance Fit SE Fit 95% CI Resid Std Resid Del Resid HI Cook’s D Obs DFITS
2 711.7 764.0 48.1 (655.3, 872.7) -52.4 -2.02 -2.58 0.775 1.76 2 -4.7931 R
3 590.0 641.9 48.1 (533.2, 750.6) -51.9 -2.00 -2.54 0.775 1.73 3 -4.7102 R
4 646.7 588.7 48.1 (480.0, 697.4) 58.0 2.24 3.17 0.775 2.16 4 5.8834 R
5 951.7 1024.4 48.1 (915.7, 1133.1) -72.7 -2.81 -7.53 0.775 3.40 5 -13.9661 R
8 976.7 1031.7 48.1 (923.0, 1140.4) -55.0 -2.13 -2.84 0.775 1.95 8 -5.2719 R
R Large residual
a ge Distance1 0 00 a ge Distance1200
10 0 0
1 1
800 800
0 0
-1 St op Posit ion ( E ) -1 E la st ic Posit ion ( D)
0 -1 0 -1
1 1
Pull-Back Angle ( A) Pull-Ba ck Angle ( A)
1250
a ge Distance1 0 00
750 1
500 0
-1 E la st ic Posit ion ( D
0 -1
1
Stop Posit ion ( E )
We can observe that we need to have a pull-back angle as high as possible to maximise the
response, as well as a high elastic position setting. As a particular example, if we were to set
up the A=180º and D=5 we would have a firing distance larger than 1400 mm (for E=4), and
in contrast, the response would be less than 600 mm for all coded levels 0.
Hold Values
-1.0 -1.0
Pull-Back Angle (A) 0
-1.0 -0.5 0.0 0.5 1.0 -1.0 -0.5 0.0 0.5 1.0
Stop Position (E) 0
1.0 Elastic Position (D)*Stop Position (E) Elastic Position (D) 0
0.5
0.0
-0.5
-1.0
-1.0 -0.5 0.0 0.5 1.0
Settings Prediction
Variable Setting Setting Setting Fit SE Fit 95% CI 95% PI
Pull-Back Angle (A) 175 165 179 1082.63 22.3633 (1032.04, 1133.22) (949.192, 1216.07)
Predicted Actual
1082.63 1132.12
1184.83 1255.19
1634.12 1596.65
From the table above, we can observe that the predicted and the actual distance are close,
within the PRESS RMSE model confidence expectation of 95% confidence intervals. It can
also be observed that the longer the firing distance, the higher the variation, therefore the
accuracy is decreasing with distance.
As seen in the Results section, it is important to follow the outlined methodology from the
DoE to enable the construction of a proper understanding of the system and the factors that
influence the response.
The methodology was appropriate because it enabled us to use the right statistical tools
appropriate for the problem, which included the selection of the most important factors from
the screening experiment, followed by in-depth model development and optimisation to
enable prediction of the response. This methodology is very powerful in a variety of scenarios
that could be under investigation or optimisation.
None of the steps outlined in the methodology should be missed, as under the false
impression of saving time, jumping over the screening experiment into the follow-up 3 factor
experiment can be very costly.
The report presents graphs and descriptions that where a duplication of other plots that
presented a similar message (such as Pareto) that where not essential. This was done for my
personal benefit – to explore all the angles and the options to look at the data and analyse the
problem, and less for the benefit of the marker. In turn, this lead to a large report and a large
number of graphs in the annexes.
Whilst running the experiments, the team (some members presented in Figure 1) was aligned
on the importance of the operational definition and planning stage, which led to us lagging
behind other teams and having to rush to perform the follow up experiment. It would have
been ideal to have more time to repeat and replicate in randomised runs the follow up
experiment and to get a larger accuracy. Even with our extensive work on the operational
definition, it was still the case that some of the observations had large residuals.
For me, in particular, the most important learning from this module and PMA is the capacity
to use the Minitab DoE and RSM tools and to plan and run an experiment following a robust
methodology. Although the catapult case was an academic exercise, I feel that I have learned
a lot from this example and the PMA and I would be confident to apply it into a future
project.
I feel that I have gained a lot from attending this module, and even if I will not be using this
tools in the DoE and RSM in the foreseeable future, there is always the possibility of
returning into engineering to support with BB work. Even beyond this, a detailed
understanding of advanced statistical tools is beneficial in my future work – as I am now
starting to observe opportunities for me to apply the regression analysis and RSM to
leadership data and the Pulse survey data.
There is scope for improvement to look to utilise more designs and the model fit selection, as
a theoretical exercise.
On a cautionary note, it is important to always use engineering judgement when making the
initial factor selection – as without this, some very important factors could be missed, or
others that have a drastic effect like the tensioning peg in our case could be selected that
would skew the model.
Noise factors are equally important to note and obverse and to capture proper mitigation
actions against them in the Operational definition – which will make it more robust.
We can draw the following main conclusions from the results – the screening experiment is
very powerful to identify from the initial factors studies which have the largest and smallest
effect on the response (projectile distance). In this study, it was concluded from that section
that Hold Time and the Twists have very little effect and where removed from the follow-up
experiment. The three remaining factors have a large effect – which is easily seen in the main
effects plot.
The conclusion from the follow-up experiment is to not accept a surface response and a
model until detailed analysis is not carried out – looking at the regression analysis – as it
might be the case that terms might need to be removed (outliers) and the model re-fitted.
It is recommended that in order to bring the Press RSME down, we should have done more
experiments (runs) – which was not possible due to time constraints. In the future, it is
recommended for repeats to be planned in.
Recommendations for future work include to expand the application of these tools in either
engineering projects or in HR projects. I have considered to use the RSM tools for the
prediction of different leadership behavioural tests with Pulse scores – however, for the
purpose of this report, would have not enabled me to apply the DoE tools, therefore the
Catapult was used instead.
The use of the tools was relatively easy and the data interpretation very visual and
straightforward.
I particularly liked the prediction tool – which enabled us to predict the firing distance
without needing to actually carry the experiment. Given that we would have a good model, it
is a very efficient way to predict responses and to do tests.
[2] “Interpret the key results for Interaction Plot,” Minitab Express Support, [Online].
Available: [Link]
to/modeling-statistics/anova/how-to/interaction-plot/interpret-the-results/.
[Accessed May 2018].
[3] M. 1. Support, “Interpret the key results for Surface Plot,” [Online]. Available:
[Link]
statistics/using-fitted-models/how-to/surface-plot/interpret-the-results/key-results/.
[Accessed May 2018].
[4] M. 1. Support, “Effects plots for Analyze Factorial Design,” Minitab, [Online].
Available: [Link]
to/modeling-statistics/doe/how-to/factorial/analyze-factorial-design/interpret-the-
results/all-statistics-and-graphs/effects-plots/. [Accessed May 2018].
[5] M. 1. Support, “Interpret the key results for Analyze Response Surface Design,”
Minitab , [Online]. Available: [Link]
and-how-to/modeling-statistics/doe/how-to/response-surface/analyze-response-
surface-design/interpret-the-results/interpret-the-results/?SID=129050. [Accessed
May 2018].
[6] M. 18, “What is the difference between coded units and uncoded units?,” Minitab,
[Online]. Available: [Link]
to/modeling-statistics/doe/supporting-topics/basics/coded-units-and-uncoded-
units/. [Accessed May 2018].
[7] U. o. Bradford, “BB3 Design of Experiments and Response Surface Methods Course
Notes,” 2015.
Figure 20. The Fractional Factorial Design and results (not coded)
Figure 21. The Fractional Factorial Design Factors and Levels (not coded)
Hold Tim e
800
5
10
Twis t s
750
500
0 3
5
St op Pos it ion
Elastic Position 3
1
160 160 1
160 160
Stop Position 4
170 5 170 0 170 1 170 3
180 180 180 180
Pull B ack A ngle Pull B ack A ngle Pull B ack A ngle Pull B ack A ngle
1100
0 1500
1000
A ver age A ver age
900 1000
5 5
800
0
4 500 4
0 St op Pos it ion 1 St op Pos it ion
5 3 3 3
10 5
Twis t s Elast ic Pos it ion
1200
ge Distance1000 ge Distance
5 900 5
800
4 600 3
160 St op Posit ion ( E ) 160
16 E la st ic Posit ion ( D)
170 3 170 1
180 180
Pull-Back Angle ( A) Pull-Ba ck Angle ( A)
1 2 50
ge Distance1 0 0 0
75 0 5
500 3
3 E la st ic Posit ion ( D)
4 1
5
St op Posit ion ( E )
Hold Values
3.0 1
Pull-Back Angle (A) 170
160 165 170 175 180 160 165 170 175 180
Stop Position (E) 4
5 Elastic Position (D)*Stop Position (E) Elastic Position (D) 3
1
3.0 3.5 4.0 4.5 5.0
At α = 0.15