Quantitative Problem Solving
Quantitative Problem Solving
2018
Recommended Citation
Moore, Peter L., "Quantitative Problem Solving in Natural Resources" (2018). Open Textbooks. 1.
https://lib.dr.iastate.edu/opentextbooks/1
This Book is brought to you for free and open access by Iowa State University Digital Repository. It has been accepted for inclusion in Open Textbooks
by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected].
P. L. Moore
Quantitative
Problem Solving in
Natural Resources
O C TO B E R 5, 2018
D I S C O V E R Y I N T H E S O L U T I O N O F A N Y P R O B L E M . Y O U R P R O B L E M M AY B E
M O D E S T ; B U T I F I T C H A L L E N G E S Y O U R C U R I O S I T Y A N D B R I N G S I N T O P L AY
Y O U R I N V E N T I V E F A C U LT I E S , A N D I F Y O U S O LV E I T B Y Y O U R O W N M E A N S ,
Y O U M AY E X P E R I E N C E T H E T E N S I O N A N D E N J O Y T H E T R I U M P H O F D I S C O V -
E R Y. S U C H E X P E R I E N C E S AT A S U S C E P T I B L E A G E M AY C R E AT E A TA S T E
F O R M E N TA L W O R K A N D L E A V E T H E I R I M P R I N T O N M I N D A N D C H A R A C T E R
FOR A LIFETIME.
G E O R G E P Ó LYA , H O W T O S O LV E I T
Copyright © 2018 P. L. Moore
1 Introduction 11
9 Triangles 93
Part V MODELING
12 Modeling 125
Index 141
Preface
Note to instructors
I have written this book with my students and their needs in mind,
but hope that others may find it useful. It may be worth taking a
moment to clarify how I use this book. While I expect my students
to read this text and work on the exercises in it, I build the course
around a group of “focus problems” not contained here, but with an
open-ended form and practical flavor like the sample problems in
Chapter 3. As we work on the first focus problem, I guide students
through the problem-solving process described in Chapter 2, making
frequent reference to examples contained in the book. Therefore, to
appreciate the processes involved, I think it is wise to have students
read the first three chapters early. The remainder of the book is or-
ganized more by problem type, and can be assigned or referred to as
appropriate to support the focus problems, rather than proceeding
linearly through each chapter.
I have deliberately avoided discussing particular software tools or
internet resources, partly in the interest of ensuring that this text does
not rapidly become obsolete, but also to allow for flexibility. In online
video tutorials and in-class exercises, I ask my students to work with
10 iowa state university
data and create graphs using spreadsheets, but those seeking richer
computing environments can just as easily use this text alongside
instruction in R, Matlab, Mathematica, and others.
1
Introduction
4x2 − 13x + 6 = 0.
often see equations exactly like the example above, because the roots
turn out to be icky decimal numbers rather than nice, clean integers.
Furthermore, things get complex (literally!) if the numerical coeffi-
cients on the left-hand side of the equation are such that the term un-
der the square-root in the quadratic formula turn out to be negative.
Such a poorly-behaved case belongs to a completely different subject
in the mathematics curriculum (complex analysis), and so cannot
be imposed upon an unsuspecting algebra student. However, in the
“real world”, there is no more reason to suspect a real result than a
complex one, in those rare practical instances when one needs to find
the roots of a second-order polynomial. Thus, in this approach we
learn a very strict set of rules applicable only to an idealized subset
of problems that may or may not have any significance outside of
abstract trivia.
The approach we use in this course is to encounter math and
statistics in the process of finding solutions to real—or at least plausible—
problems in the natural sciences. Sometimes these real problems are
messier than those out of a textbook. Often they will be open-ended
and will require multiple steps and a variety of techniques. We’ll
need to decide for ourselves what tools and techniques to use, ac-
cording to the needs of the problem. Being creative in math classes
isn’t what we’ve been trained to do, and at times it may be uncom-
fortable. That’s OK, we’ll take our time. But no matter the problem,
we’ll always have a reason to be doing math or statistics – every quan-
tity in an equation will have real meaning or role, and we can apply
our non-quantitative knowledge and experience with these entities to
help us solve our problems.
The rebound in the roadside counts from 2013 to 2014 was char-
acterized in the DNR report as a 151% increase, and the change from
2014 to 2015 was described as a 37% rise. The last few numerical
value pairs plotted in the graph are also shown in Table 1.1 to the
right. year count
This example may seem straight forward at first glance, and in 2012 7.8
2013 6.5
some ways it is. The trend appears to indicate an overall decline in 2014 16.3
pheasant numbers across the state, and perhaps we should be pre- 2015 23.2
pared to take a more hands-on approach to managing pheasant num- 2016 20.4
2017 14.4
bers if we wish to sustain a viable game resource in the coming years. 2018 20.6
On the other hand, the year-to-year changes seem to be erratic, ris-
ing and falling in a way that seems to lack a pattern. Addressing the Table 1.1: The most recent five years of
data from the roadside pheasant count.
guiding question with any confidence, however, could be a bit chal-
lenging. If, for example, we were looking at the dataset at the end
of 2011 following 6 straight years of steady decline, would we have
been able to anticipate a rebound in 2014 or 2015? Probably not with-
out a robust and reliable model4 of the factors that cause population 4
a model in this context means an ap-
proximate mathematical representation
changes and how those factors could change in subsequent years.
of the real system from which predic-
These are advanced topics, but ones that wildlife managers have to tions about the behavior of the real
incorporate into their management strategies in some situations. system may be made and tested.
the larger population, at least one step removed from the quantity
we seek. How does such a sample relate to the larger quantity we are
after? That’s a pretty simple question in theory (i.e., in stats class),
but when we account for the sampling methodology, timing, and
observer variability, and we consider that pheasant visibility may not
always be directly linked to population, it isn’t so straight-forward
after all.
The observations we’ve made from this dataset are just some of
the many complexities that we might uncover as we endeavor to
solve problems in pheasant population or habitat management in
Iowa. This example hints at the concepts of time series analysis,
forecasting, and measurement uncertainty, as well as functional rela-
tionships between multiple variables and between samples and pop-
ulations. Each of these concepts represents a quantitative tool that
can be applied toward the larger problem-solving task. We’ll visit
most of these concepts and many more en route to addressing prac-
tical problems and methods. But as we will see in the next chapter,
the quantitative procedures that we employ in the problem-solving
process are just a part of the arsenal necessary to solve practical prob-
lems. Furthermore, it should go without saying that some of the
quantitative tools we do have at our disposal are not appropriate for
some problems, and a key job of the problem-solver is to determine
which tools those are. We’ll delve deeper into this issue a few pages
ahead.
“In a lake the bud of a water-lily was observed, one span above the
water, and when moved by the gentle breeze, it sunk in the water at
two cubits’ distance. Required the depth of the water.”
-Henry Wadsworth Longfellow
The lines above are from a poem in which a Mr. Churchill play-
fully challenges his wife with mathematical puzzlers. The problem
he describes actually dates back many centuries, to a 12th-century In-
dian mathematician named Bhascaracharya who posed the problem
5
One translation that can be readily in verse in his book Lilavati5 . As we look forward to learning the pro-
found online is an 1817 translation by cess of problem-solving, let’s imagine a dialogue between a student
H.T. Colebrooke: Algebra, with Arith-
metic and Mensuration, Brahmegupta (S) confronted with this problem and a patient instructor (I). This
and Bhascara, London, John Murray. isn’t quite the sort of problem we’re interested in really diving into,
but the dialogue serves to illustrate a few points that we’ll address in
the next chapter.
quantitative problem solving in natural resources 15
mathematics, and all your arithmetic is sound, you’ll get the correct
answer and you’ll know it.
S: Oh. But... how do I know if I’ve got the right answer? You’re not
going to tell me?
I: Welcome to the real world.
Exercises
1. Make a list of the variables or factors that one might need to take
into account when projecting future pheasant populations from
historic data such as the Iowa Roadside Pheasant Survey.
3. See if you can arrive at a solution for the Lilavati water lily prob-
lem. While you’re working on this, occasionally step back and
think about what sorts of things you are doing to make progress.
Are you trying to understand what the problem is about? Trying
to appreciate the geometry? Trying to recall algorithms that relate
different quantities in a geometric shape? If you feel stuck at any
point, what do you think is preventing you from making progress?
Record and annotate all of your thoughts and solution attempts,
and try not to erase or scribble anything out.
Part I
PROBLEM SOLVING
2
Problem Solving as a Process
4. Check the result. Does the result satisfy the conditions stated
in the problem? Is it consistent with expectations or within
reasonable bounds? Can you arrive at the same result using a
different approach?
were to point the student toward a solution method every time she
was confronted with a challenging problem, she would learn only
two things: 1) how to implement algorithms and compute numerical
results as instructed; and 2) to relinquish all control of choosing how
to approach and solve a problem to somebody else. Sadly, this is of-
ten the best outcome of the standard school mathematics curriculum.
The worst outcome is that students dismiss math as boring, difficult,
or irrelevant. In some cases, a diligent student develops some facil-
ity2 with basic manipulations of mathematical symbols and numbers, 2
Though the memory of how to use
but little or no ability to create the frame of mind and methodological mathematical algorithms certainly
degrades with time unless used or
structure needed to begin and confidently proceed trying solutions. reviewed frequently
For the challenge of getting started, Pólya’s framework offers Un-
derstand. What is the problem really asking, and what exactly do
you want to end up with as a result? Take, for instance, the pheas-
ant count problem that we introduced in the first chapter: what is
22 iowa state university
• Try a few values. Where some values are unknown but are not
the desired quantity, try to solve the problem with a few supposed
numerical values. Sometimes the outcome of algebraic relation-
ships is relatively insensitive to the precise value of the quantities
included in the relationship. When the relationship is strongly sen-
sitive to unknown or poorly-constrained quantities, identify those
quantities as important intermediate goals or sub-problems.
• Visualize the data. If data or input values are given and a rela-
tionship or summary-statistic is desired, make a graph or diagram
from the data. In some cases, visual representation of the data can
suggest or substantiate approximate values for the unknowns or
can illustrate the form of functional relationships.
• Resources
These are the things you know or understand about the prob-
lem domain, constraints on quantities and their representation
in the problem domain, and the skills you possess in perform-
ing algorithmic procedures.
• Heuristics
The decision-making tools used to make sense of challenging
problems that allow you to make progress or develop insight.
Most of the chapters of this booklet are devoted to developing
strategies useful in natural science or resource management
domains.
• Control
Control is the management and self-awareness of the
problem-solving process. It includes planning, execution
and evaluation decisions and and the selection of resources
and heuristics for the problem.
• Belief
Belief includes the set of notions one has about the problem
domain as well as one’s own abilities or challenges in ap-
plying math and statistics to the problem domain; this also
includes preconceptions and (mis)understandings that could
lead to the use of (in)correct resources and heuristics in a
given problem.
Exercises
costs?
Figure 3.2: Schematic illustration of
one conceptual model for the optimal
management of fire fuels. Total cost
3.6 Maximized effluent C of management increases with
suppression and pre-suppression effort.
How much of phosphorus (P) can be discharged from point sources With little suppression effort however,
the risk of high losses or net change
into an urban stream without exceeding total maximum daily loads in value, NVC, is high, declining
(TMDLs)?1 with suppression effort. This view of
The U.S. Clean Water Act of 1972 establishes criteria for identify- management incentives indicates that
overall cost is C + NVC.
ing polluted water bodies and designing plans to reduce pollutant 1
Inspired by Litwack et al., 2006, Journal
loads and improve water quality. One of the measures available to of Environmental Engineering, 132(4):
water resource managers is the establishment of Total Maximum 538-546.
Exercises
NUMERICAL
REASONING
4
Quantities in the Real World
Properties of quantities
not meaningful.
Paying attention to units and dimen- Quantities with practical meaning to us will often have units. Con-
sions is not mere formality. In the sider a few random examples:
scaled quantities we use in science,
units and dimensions constrain what
mathematical operations are permissi-
• 71 foot tall tree
ble.
• 16.4 grams of soil
• 42 snow geese
could spend lots of time (not wisely, I would argue) setting up frac-
tions on paper to figure this out (how many ounces in a pound, how
many pounds in a kilogram, how many kilograms in a ton. . . ), but
as long as we have access to a computer we might as well use it.
Plenty of apps and websites will do basic unit conversions for you,
including google itself. Employing one of these, we see that there are
35274 ounces in a metric ton, and now we can express our measured
concentration as 15.84 ounces / 35274 ounces, or about 0.00045. To
convert from this decimal value to parts per million, we just need to
multiply by one million or 106 (much like you would multiply by 100
to get a percent), giving us roughly 450 ppm lead. So at 450 ppm, the
lead concentration in our soil exceeds the threshold that the EPA has
deemed safe, but is not cause for great alarm.
• Graphs, images, figures: any visual display of quantities or re- numbers equations
lationships between them. Single-variable statistical charts (e.g.,
Figure 4.1: Ways of communicating and
histograms) and two-dimensional graphs (i.e., x-y scatter-plots)
understanding quantitative information.
will be the most frequently encountered, but maps are another
example. Often the most efficient way to demonstrate patterns or
large volumes of data.
44 iowa state university
40
4.6.1 Example: brook trout recruitment (teaser Problem 3.7)
20
kinds of variables, data tables like Table 4.1 are perhaps the best op-
tion.
Given only the information presented already about the brook
trout electrofishing catch, communicating through equations is prob-
ably unwarranted. However, a textual narrative is essential for com-
municating the nature and meaning of the data presented in this
example. Suppose you were unfamiliar with electrofishing and de-
pletion methods for fish population assessment, would the bars in
Figure 4.2 and the numbers in Table 4.1 tell a clear story? The figure
captions and the paragraphs above are essential for drawing meaning
from the figure and table. This is why words are at the center of the
triangle in Figure 4.1. And this is also why it quantitative problem
solving is a writing-intensive endeavor. If we are unable to communi-
cate clearly about our ideas, strategies, results, and conclusions, most
of the effort is for naught.
Exercises
In high school chemistry, we learn that there are more than 602 sex-
1
6.022×1023
602 sextillion, or is Avo- tillion molecules in a mole of a chemical substance1 . But we don’t
gadro’s constant, the number of normally see Avogadro’s constant written as some number of sextil-
molecules in one mole of a chemical
substance lions, nor do we see it elaborated with all of the 24 digits necessary
to write it in integer form: it is difficult to keep track of all those dig-
its when writing them, and even more difficult to keep track when
you’re reading or comparing different numbers. Instead of writing
the entire number out, we use the shorthand of scientific notation,
where Avogadro’s constant looks more like 6.022×1023 . In general,
scientific notation has the form:
a = 10b
where a and b are sometimes called the mantissa and power, respec-
tively. So Avogadro’s constant has a mantissa of about 6.022 and a
power of 23, which is equivalent to saying that the complete quantity
2
A related issue is that of significant has 23 digits after the mantissa2 . Obviously this is a very large num-
digits. Scientific notation allows us ber. We can just as easily express very small numbers with scientific
to clearly specify how precise we are
claiming to be through the number of notation. An e. coli bacterium is roughly 2 µm (micrometers) long,
digits included in the mantissa: in this which is 2×10−6 m. Here, the power of −6 indicates not that it’s a
case, 4.
negative number (it would be absurd to say something has a negative
length, because length is a ratio scale!), but that it is smaller than 1
and that there should be 6 digits to the right of a decimal point if we
wished to express it as a decimal number. So we could express this
equivalently in a few ways:
Note that these equalities both amount to unit conversions, but the
second equality is specifically a conversion to scientific notation.
Negative exponents indicate numbers smaller than 1, and there
are occasions where it can be helpful at times to can write these as
fractions. When we have a quantity expressed in scientific notation
with a negative exponent 10−b , that is equivalent to the same quan-
tity divided by 10b . Therefore, another way to express the length of e.
coli is:
1
2 × 10−6 m = 2 × 6 m.
10
So dividing by 106 is the same as multiplying by 10−6 . Notice here
In the standard order of operations, that the order of operations is important. By convention, exponents
parentheses take precedence, then
exponents, then multiplication or
division, and finally addition and
subtraction.
quantitative problem solving in natural resources 49
83
× 100% = 23.0556% (5.1)
360
En route to computing this, we created the ratio 83 to 360, which is
around 0.23 if you simplify it with your calculator. As with many
such ratios, we can choose from a variety of different but equivalent
ways of expressing this quantity. We could just express it as the ratio
4
this is the way we usually express a of two whole numbers like 83:3604 , or as the fraction:
map scale, like 1:24,000. See Part III of
this book for more on that issue. 83
. (5.2)
360
Or as we’ve already seen, it is simple to write it as a decimal number
(0.230556). But since we encounter percentages frequently, we may
more readily appreciate it expressed as a percentage. For the present
purposes, we could describe a percentage as “parts per hundred”,
since it is just the same ratio scaled to an arbitrary reference value of
100. In other words, for every hundred deer in the sample, about 23
have CWD. Expressing a quantity in “per mil” is closely analogous,
except instead of multiplying by the factor 100% we’d multiply by
5
Although not common in many 1000h(that’s the per mil symbol)5 . In this case, we’d end up saying
disciplines, isotope concentrations that about 83/360 ×1000h= 231h(or 231 per thousand deer) are in-
are often expressed in h, where the
reference value is the isotopic ratio of a fected. To make this even more absurd, we could express the same
standard substance. information just as easily as parts per million (ppm) or parts per bil-
lion (ppb) following a similar tactic. Each of these ways of expressing
a normalized quantity is arithmetically equivalent, but implies a dif-
ferent realm of precision about the quantity of interest and the scope
quantitative problem solving in natural resources 51
of its possible values. We’d likely never talk about deer in parts per
million, but we might talk about lead concentrations that way!
Other types of normalized quantities in science include frequen-
cies, concentrations, and probabilities, to name a few. The quantities
may be expressed somewhat differently, but in most cases there is
a comparison being made between values of the same dimensions
(and often the same units!). Indeed this is sometimes a simplifying
strategy: when you normalize a quantity to a standard of the same
units, details about the specific units by which the quantities were
measured can be discarded. Often this is a good thing. For exam-
ple, when we use a map that is scaled at, for example, 1:24,000, we
are not told what units that ratio was constructed with, because it
doesn’t matter! If you use a ruler to find that the map distance be-
tween two features on the map is 2 inches, that distance in the real
world is equal to 2 × 24, 000 = 48, 000 in. It doesn’t matter whether
your ruler is ruled in inches, centimeters, furlongs or rods, the quan-
tity you measure on the map only needs to be multiplied by the scale
factor (24,000) to find the true distance! As we’ve seen, however, ne-
glecting the specific units used to derive a normalized quantity can
also be the cause of some confusion (is the concentration of one sub-
stance mixed with another computed on the basis of their masses,
volumes, or something else?). It becomes a good thing if the proce-
dural statement for the quantity is either made clear or is known by
convention.
How do we use a normalized quantity to our advantage? Suppose
I extrapolate from our sample of CWD in deer carcasses to predict
that 23% of the deer in the entire county are infected with CWD. If
we take for granted that my science is good, all we need to know to
find out the number of CWD-infected deer in the county is the total
number of deer in the county, Ndeer . Once we recognize that the ratio
of infected deer to total deer is 0.23 (23% of the total population of
100%), we need only perform a simple multiplication:
NCWD 23%
= = 0.23 (5.3)
Ndeer 100%
Using one of the above rules for exponents to modify the right-hand
side of this relationship, we can find that:
and we have our result. There are 106 cm3 , and therefore 106 ml
in a cubic meter! It would be just as easy to look up the conver-
sion online, but the same basic approach can be readily applied to
more complex problems with murkier solutions. In the next section
we’ll consider a more challenging and engaging example that can be
worked out with a similar strategy.
The exponent therefore becomes −6, which you recall is the base for
a “parts per million” ratio. We can simplify the fraction 6/5 either
directly on a calculator, in our heads10 , or by multiplying both nu- 10
One great benefit of using scientific
merator and denominator by two (= 12/10) and dividing by 10 to notation is that computations can be
approximated easily by hand!
get 1.2:
6/5 × 10−6 = 1.2 ppm (5.11)
So the result for Wapello is 1.2 ppm, which exceeds safe limits for
consumption. For the Maquoketa River, the Hg concentraion is only
0.2, so it is safe to eat and no advisory need be issued.
tissas together, put the powers together, and put the units together.
Multiplying the mantissas through we get 6, and using the rules for
manipulating exponents (see the next section!) the exponents are
added together (3 + 4 + −1) = 6. Canceling units, we see that USD is
the only remaining unit. So our ballpark solution is that the value of
the standing timber in this 1000 ha parcel is 6 × 106 USD, or about $6
million.
Exercises
14
Based on the article ”Prevalence of 2. 14 Leucism is partial albinism, manifested in penguins as a lack
leucism in Pygocelid penguins of the of (or substantial reduction in) pigment in plumage. A study of
Antarctic Peninsula” by Forrest and
Naveen, Waterbirds 23(2): 283-285, 2000. three species of penguin (Adélie, Gentoo and Chinstrap) in the
Antarcic peninsula sought to identify the prevalence of leucism
in these different species. The paper cited in the margin provides
species prevalence count the following information derived from detailed counts of penguin
Adélie 1:114,000 1,144,000
Chinstrap 1:146,000 293,800 breeding colonies made during the years 1994-1997:
Gentoo 1:20,000 41,550
Perform the following manipulations of the prevalence data for
each species:
The random nature of this set of data allows us to use some of the
familiar ways of describing our data, while boosting our confidence
that we are also properly characterizing the larger system that we are
sampling.
assessing the health of all the humans in a small town! Instead of try-
ing to track down every individual though, we can do a decent job by
simply taking a random sample from the population and performing
the desired analysis on that random sample. As we have seen, if we
are sufficiently careful about avoiding bias in our sampling, we can
be reasonably confident that our sample will tell us something use-
ful (and not misleading) about the larger population that the sample
came from.
If our concern is mainly with the population of a target species in
a certain area, we can use a method called mark-recapture, or capture-
recapture. The basic premise is simple: we capture some number of
individuals in a population at one time, band, tag or mark them
in such a way that they can be recognized later as individuals that
were previously captured, then release them. Some time later, after
these individuals have dispersed into the population as a whole,
we capture another set. The proportion of the individuals in the
second capture who are marked should, in theory, be the same as the
proportion of the whole population that we marked to begin with. If
the number of individuals we marked the first time around is N1 , the
number we captured the second time around is N2 , and the number
in the second group that bore marks from the first capture is M, the
population P may be estimated most simply as:
N1 N2
P= (6.1)
M
This comes from the assumption that our sample each time is ran-
dom, and that the marked individuals have exactly the same like-
lihood of being in the second capture as they did in the first: 1/P.
Therefore, if we sampled and marked a fraction N1 /P the first time
around and sample N2 the second time around, then we should ex-
pect a fraction M/N2 of them to be marked.
Of course this whole plan can be foiled if some key assumptions
are not met. For example, we need the population to be “closed” –
that is, individuals do not enter and leave the population such that
our sample is not coming from the same set of individuals each time.
Problems could also ensue if our “random” sample isn’t random, if
somehow the process of marking individuals either harmed them
or made their likelihood of re-capture more or less likely, or if the
time we allowed for them to re-mix with their population was not
appropriate. On the last point, you can imagine that if we recapture
tortoises 10 minutes after releasing them from their first capture, our
second sample will not be very random. On the other hand, if we
recapture marked fish 20 years after they were first marked, many of
them may have died and been replaced by their offspring, and thus
our assumption of a “closed” population is violated. So in planning
62 iowa state university
The first point suggests the use of our measures of central ten-
dency: mean, median and mode. The second goal relates to measures
of spread or dispersion in the data. For example, how close are most
values in the data set to the mean?
that i begins with a value of 1 and increases with each added term
until i = n, which is the last term. So in the end, you can interpret
this to have a meaning identical to the equivalent expressions above,
but in some cases this notation can be more compact and explicit. It
also looks fancier and more intimidating, so people will sometimes
use this notation to scare you off, even though it gives you the same
result as the second equation above.
6.3.2 Mode
The mode is the value or range of values that occurs most frequently
in a data set. Since you caught 5 half-pound fish and fewer of every
other weight value in the dataset, the mode of this distribution is
0.5 pounds. Now if the weights we’ve reported above are actually
rounded from true measured weights that differ slightly, this defini-
tion becomes less satisfactory. For example, suppose the half-pound
crappies actually weighed 0.46, 0.49, 0.5, 0.55 and 0.61 pounds. None
of these are actually the same value, so can we say that this is still a
mode? Indeed we can if we choose to discretize or bin these data. We
might say that our fish weights fall into bins that range from 0.375
to 0.625, 0.625 to 0.875, 0.875 to 1.125, and so on. In this case, since
all of our crappies fall in the range 0.375 to 0.625 (which is 5 ± 1/8
lbs), this size range remains the mode of the data set. We can see this
quantitative problem solving in natural resources 65
6.4 Spread
xi − x̄, (6.5)
The gives us a good sense for how far from the mean a typical mea-
surement lies. We can now characterize a sample as having a mean
value of x̄ and standard deviation of σ, or saying that typical values
are x̄ ± σ. But in reality, if we computed x̄ and σ, the bounds set by
x̄ − σ and x̄ + σ only contain about 68% of the data points. If we want
to include more of the data, we could use two standard deviations
above and below the mean, in which case we’ve bounded more than
95% of the data.
One piece of information we have thus far omitted from our list of
properties that fully define a quantity’s value is uncertainty. This is
particularly important when we are quantifying something that has
been measured directly or derived from measurements. Thus, to even
more completely define the value of a measured quantity, we should
include some estimate of the uncertainty associated with the number
assigned to it. This will often look like:
where x is the thing we are trying to quantify, xbest is our best guess
of its value, and δx is our estimate of the uncertainty. Though it will
depend on the quantity in question, our best estimate will often be
the result of a single measurement or – better yet – the mean of a
The preferred value xbest for a quantity number of repeated measurements.
of interest will often be the mean of
repeated measurements of that quantity.
6.5.1 Uncertainty in measured quantities
All measurements are subject to some degree of uncertainty, arising
from the limited resolution of the instrument or scale used to make
the measurement, or from random or systematic errors resulting
from the method or circumstances of measurement. Let’s consider an
example:
Suppose two fisheries biologists each measured the lengths of
ten of the brook trout captured during the electrofishing traverse
from Problem 3.7. Both used boards with identical scales printed
on them, graduated to half of a centimeter. They then plan to put
their measurements together to get a data set of 20 fish. One of them
was trained to pinch together the tail fins to make this measurement,
while the other was not. In addition, because they wished not to
quantitative problem solving in natural resources 67
harm the fish, they made their measurements quickly, even if the fish
flopped and wiggled during the measurement. What are the poten-
tial sources of error and how big are they relative to one another?
For starters, implicit in the graduations on this board is that the
user cannot confidently read any better than half-centimeters off the
scale. He or she can, however, visually interpolate between two adja-
cent graduations to improve precision (see below). However, this step
is inherently subjective and limits the certainty of the measurement.
We might call this instrumental error because its magnitude is set by
the instrument or device use to make the measurement. One way to
reduce this source of error is to use a more finely-graduated scale. Instrumental error is fixed by the
A second source of error arises from the hasty measurements and resolution of the device used to make
a measurement, and can usually only
the fact that the fish were not necessarily cooperative. Perhaps the be reduced by using a more precise
mouth was sometimes not pressed up all the way against the stop, or instrument.
the fish wasn’t well aligned with the scale. Some lengths may have
been too large or small as a result, yielding a source of error that was
essentially random. Indeed, we can call this random error since its
sign and magnitude are largely unrelated from one measurement
to the next. Reducing this source of error in this case would require Random measurement errors may be
either more careful and deliberate effort at aligning and immobilizing mitigated by repeating measurements.
6.6 Distributions
The kind of data we’ve been talking about thus far is univariate: a
single quantity with variable values like the diameter of a stream-
bed particle, or the length of a fish. As we know, not all age-0 brook
trout are the same size. In a first-pass capture of 50 fish, for example,
we should expect some variability in length that might reflect age,
genetics, social structure, or any other factor that might influence de-
velopment. The variation may be visualized graphically in a number
of ways. We’ll start with a histogram.
A histogram shows the distribution of a set of discrete measure-
ments – that is the range of values and the number of data points
falling into each of a number of bins, which are just ranges of val-
ues (112.5 to 117.5 is one bin, 117.5 to 122.5 another. . . ). This can be
Figure 6.3: A histogram showing the
called a frequency distribution, and a histogram is one of the best distribution of simulated (random)
ways to visualize a frequency distribution (Figure 6.3). measurements of the length of 100
snakes.
But what if we had uniformly distributed data? A uniform dis-
tribution means that it is equally likely that we’ll find an individual
with a length on the low end (97.5-102.5 mm) of the range as any
other. That would look quite different – there would be no hump
in the middle of the histogram, but rather a similar number of mea-
surements of each possible length. The uniform distribution is great:
in fact, we count on uniformity sometimes. If you are at the casino
and rolling the dice, you probably assume (unless you’re dishonest)
that there is an equal probability that you’ll roll a 6 as there is that
you’ll roll a 1 on any given die. We can call that a uniform probabil-
ity distribution for a single roll of a die. But What if the game you
are playing counts the sum of the numbers on 5 dice? Is there still a
uniform probability of getting any total value from 5 to 30?
We could actually simulate that pretty easily by randomly choos-
ing (with a computer program like R6 or Excel) five integers between
1 and 6 and adding them together. Figure 6.4 shows the plot that
comes out. Looks sorta like a bell curve, right? Well, how likely is
it that you’ll get five 1’s or five 6’s? Not very, right? You’re no more Figure 6.4: Sum of the values of five
likely to get one each of 1,2,3,4 and 5 either, right? However, there dice, rolled 100 times each.
6
R is a top choice software for general-
purpose data analysis and modeling.
It is free software, works on most com-
puter platforms, and has nearly infinite
capabilities due to the user-contributed
package repository. Learn more about R
at https://cran.r-project.org/
70 iowa state university
are multiple ways to get a 1,2,3,4 and 5 with different dice showing
each of the possible numbers, whereas there is only one way to get
all sixes and one way to get all ones. So there are better chances that
you’ll get a random assortment of numbers, some higher and some
lower, and their sum will tend toward a central value, the mean of
the possible values. So, since your collection of rolls of the dice rep-
resent a random sample from a uniform distribution, the sum of
several rolls will be normally distributed.
What’s it got to do with fish? If we sample brook trout randomly
from one stream reach and measure their lengths, we might expect
them to be normally distributed. Describing such a normal distribu-
tion with quantities like the mean and standard deviation gives us
the power to compare different populations, or to decide whether
some individuals are outliers. The nuts and bolts of those compar-
isons depend on how the type of distribution represented by the
population. An ideal normal distribution is defined by this equation:
( x − µ )2
1
f (x) = √ exp (6.12)
2πσ 2σ2
2. The graph and data table below and right show measurements of
brook trout lengths from pass #1 of the electrofishing campaign
described in Problem 3.7. Use these resources to answer the fol-
lowing questions:
(a) Judging from the histogram in Figure 2, does the dataset con-
tain just one mode or more than one? What might be the reason
for this?
(b) What is the mean and standard-deviation for the (presumed)
age-0 portion of this sample?
Data is information. Data is the result of somebody’s efforts to record There are some great resources for
and store information, often to provide an opportunity for insight. It data management out there in various
forms, including some geared toward
can be used to discover patterns, test hypotheses, and support argu- biologists. A few great ones are:
ments, among other things. But the numbers themselves often cannot Data Carpentry
www.datacarpentry.org.
convey much meaning – it is through manipulation and interpreta- Wickham, H., 2014, Tidy Data. Journal
tion of the data that those uses can be realized. We therefore need of Statistical Software 59(10).
to be conscious about how data are structured and managed, so that Saltz, J.H. and J.M. Stanton, 2017,
Introduction to Data Science, Sage publ.
they can be manipulated and interpreted to reveal insights. If data
are badly structured and managed, the risks can be great. Optimisti-
cally, we risk wasting lots of time trying to restructure data to allow
the kind of analysis we wish to perform; worse yet, we risk compro-
mising the integrity of or, heaven-forbid, the complete loss of data
through poor structuring and mismanagement.
First, lets be clear about what is meant by data structure and data
management:
that the humans who need to make sense of the data will be using
tools like computers to facilitate this approach. Therefore the data
structure needs to also accommodate the demands of the computer
hardware and software that it is used on as well as the humans.
Thus, data should be organized in a logical, self-consistent way and
it should be accompanied by documentation that helps to explain
the content and and context of the data. Similarly, accessible data
archiving, in principle, allows colleagues and competitors to test, ver-
ify, reproduce, and/or compare results with their own, ensuring that
scientific advances that you make with the help of your data can also
lead to advancement of science and management more broadly.
Table 7.1: A portion of an untidy number bluegill (2008) bluegill (2007) Iowa Darter (2007)
dataset.
1 L:1.5 W:0.7 L:1.9 W:1.3 L:2.1 W:0.9
2 L:1.0 W:0.7 L:1.6 W:1.3 L:2.0 W:1.3
3 L:2.6 W:1.5 L:2.4 W:1.7 L:1.7 W:0.7
4 L:1.1 W:0.6
fish in each column. There is also an index value in the first column
that facilitates counting the number of fish of each species captured
in each year. This is reasonably straight-forward for a human to
interpret, particularly if we are told that L corresponds to a length
in inches and W corresponds to weight in grams. However a larger
dataset organized like this table would be miserable to analyze for a
variety of reasons, including:
• two variables (length and width) are listed together within each
column, with numbers and letters mixed
One instructive question to ask is how many variables there are here.
We notice that data span multiple years, so year could be a variable.
There are also multiple species here, so species could be viewed as a
variable. Then length and width should each be variables. Finally, if
we wish to have an index or ID number for each fish, that might be 3
If there are multiple datasets derived
a fifth variable3 . In general, tidy data is organized in a rectangular from the same group of fish, assigning a
fish ID number would be a simple way
array in which each column represents a variable and each row an to connect these datasets using database
observation. In most cases, the first row contains descriptive but sim- methods.
ple column headers. This simple prescription seems unthreatening,
but it is often surprising how pervasive untidy data is. Tidy data has:
So with five variables, how many observations do we have? Each • one column for each variable
fish represents an observation, with a unique ID number, year of cap- • one row for each observation
ture, species, weight and length. From Table 7.2, there appear to be • a header row
ten fish listed among the three columns. so since each fish is an ob-
servation. According to the principles of tidy data, then, there should
be ten rows of data with values in each of the five columns. Below is
a tidy representation of this dataset. This table is now organized in a
way that can easily be sorted, filtered, and summarized in common
statistics and computational software packages.
fishID year species weight (g) length (in) Table 7.2: The data from Table 7.2
transformed to a tidy dataset.
1 2008 bluegill 1.5 0.7
2 2007 bluegill 1.9 1.3
3 2007 Iowa darter 2.1 0.9
4 2008 bluegill 1.0 0.7
5 2007 bluegill 1.6 1.3
6 2007 Iowa darter 2.0 1.3
7 2008 bluegill 2.6 1.5
8 2007 bluegill 2.4 1.7
9 2007 Iowa darter 1.7 0.7
10 2007 bluegill 1.1 0.6
SPATIAL REASONING
8
Geometry and Geography
Consider the wetland shown in the map below (Figure 8.1). How
might we characterize its size? Perhaps the answer depends on the
context of the question. Are we interested in how far it is to cross
it in a canoe? How long the shoreline habitat is for waterfowl? The
number of acres it occupies? What about the amount of water it
holds? In turn these questions point to linear distance, curvilinear
distance, area and volume, respectively. Each of these types of quan-
tities can be expressed in a variety of ways1 according to the setting, 1
For example, we can express all the
the application, or the purpose of communication. quantities in terms of SI units of m, m2 ,
and m3 , or we can use more traditional
The distinctions between length, area, and volume are more than U.S. agricultural units like feet, acres,
just trivia. They clearly reflect not only different ways of estimat- and acre-feet.
and the instructor nods approval but says, “and now we need the
area in square kilometers”. Our downtrodden student proceeds to
82 iowa state university
look up the conversion factor between miles and kilometers: 1mi '
1.609 km. Great. The calculator buttons click away until the student,
exasperated, inquires “so the area in square km is 13.02 × 1.609,
Common Mistake #2: using length which equals about 20.95 right?” The ever-patient instructor shakes
conversion factor for area conversion. her head: “there are 1.609 km in a mile, but how many square km
in a square mile?”. Our student stares at the map and feigns interest
in the question. On a whim, he asks “do I need to square the 1.609
too?” The instructor pats him on the shoulder and remarks “yep,
write it all out, and don’t forget the units” as she walks away. Our
student, relieved at having guessed correctly, writes:
The map scalebar in Figure 8.1.1 shows not only a graphical scale
that can be used to directly measure real-world distances from the
(scaled) map representation, but also indicates a ratio: 1:31,680. What
does this scale mean? Does it have units?
Map scales are like most other dimensionless proportions, as we
discussed in Chapter 4. The beauty of many dimensionless propor-
tions is that we can use any units we want in them provided that both
parts of the scale ratio or proportion have the same units. So for the
map scale, we can say that 1 inch on the map is equal to 31,680 inches
in the world that the map represents6 . If we measure out a path on 6
For better or worse, the US still per-
the map that is x inch long, the distance we would cover walking sists with using inches, feet, and miles
as conventional measures of distance
along that path in the real world is x × 31680 inches. That’s not a in official maps and documents. Even
very easy distance to envision because there aren’t any very familiar though most scientists adopted the met-
ric system long ago, we retain imperial
benchmarks near that quantity of inches, but we could convert the units here to recognize the persistence
latter to feet or miles to make it simpler, and then we can re-express of legacy units in our maps.
the scale in as a dimensional ratio: 31680 in. × 1 ft./12 in. = 2640 ft.. Recall that benchmarking is a process
of conceptualizing the size of a quantity
We can go one step further still: 2640 ft. × 1 mi./5280 ft. = 0.5 mi.. by comparing it with a known reference
That works out pretty nicely, and it often does so by design! So we quantity.
can restate the map scale as 1 inch = 0.5 miles, or equivalently 2 1 mile = 5280 feet
84 iowa state university
inches per mile. This means the same thing as the scale ratio 1:31,680
but is more specific because we have already chosen the units that we
wish to measure with. Note that we cannot say that the map scale is
2:1 or 1:0.5 because in converting the second number from inches to
miles we’ve made the scale statement unit-specific.
Maps aren’t the only thing that we encounter that are scaled repre-
sentations of reality. When we are learning about microscopic prop-
erties of molecules or cells, we often look at diagrams or physical
models of things that are too small to see. When looking through
a microscope, we perceive a much enlarged version of the object of
study. In each case, we are seeing representations of reality scaled
7
This number comes from: Trans- to size that is easier for us to grasp. Importantly, we are also (usu-
portation Research Board and Na- ally) seeing things scaled isometrically, meaning that all dimensions
tional Research Council, 2005. As- are enlarged or shrunk by the same constant factor. We’ll see in later
sessing and Managing the Ecological
Impacts of Paved Roads. Washington, chapters some interesting problems associated with scaling that is not
DC: The National Academies Press. isometric.
https://doi.org/10.17226/11535.
total road length (in miles) yields about 4.35 square miles. That’s
about 4.35/574 × 100% = 0.76% of the county area!
V = πr2 h (8.2)
described in an earlier chapter: write out the problem with just di-
mensions. We’re looking for how much it weighs, but really what we
want is a mass. If we list the dimensions of the variables we have and
want, they look like this:
we see that the mass we are looking to solve for appears in the den-
sity term. The volume term appears as a negative 3rd power in the
density term and a positive 3rd power in the volume itself, so when
multiplied together, volume (V ) and density (ρ) should yield dimen-
sions of mass: m = ρV. If we remind ourselves that the definition
of density is indeed ρ = m/V, this makes sense as a simple alge-
braic modification of that definition. Thus, our estimated mass for the
86 iowa state university
Is this close enough? Perhaps, but that depends on the nature of the
problem: why do we wish to know how much the snake weighs, and
what will we do with that information?
Not all spatial bodies of interest to us are easily measured using the
simple idealized shapes reviewed above. Alternatively, geometric
idealizations introduce too much error for certain applications. Non-
ideal shapes can, however, be approximated by irregular polygons in
some settings. Here, we describe a method for computing the area
of an arbitray two-dimensional polygon using a clever trick that is
frequently built into CAD, GIS, and other geospatial software pack-
ages. The primary requirement we must meet to use this method is
to have coordinate pairs for each vertex in the polygon in a Cartesian
(i.e., a plane with two orthogonal, linear axes; a.k.a. an x-y plane) co-
ordinate system. This algorithm may be implemented easily enough
by hand, but finding the area of a more complex shape is better left
up to a computer. Let’s have a look at how this algorithm works.
quantitative problem solving in natural resources 87
circle
circumference: 2πr
r
area: πr2
rectangle h
perimeter: 2b + 2h
area: bh b
a
triangle h b
perimeter: a + b + c
c
area: 12 ch
sphere r
surface area: 4πr2
volume: 43 πr3
h
rectangular prism
surface area: 2bw + 2bh + 2hw
volume: bwh w
b
cylinder h
surface area: 2πrh
volume: πr2 h
r
88 iowa state university
1
A= [( x2 − x1 )(y2 + y1 ) + ( x3 − x2 )(y3 + y2 ) + ...
2
... + ( x1 − xn )(y1 + yn )] (8.7)
Exercises
easting, m northing, m
1 477698 4651450
2 477720 4651488
3 477738 4651524
4 477746 4651549
5 477768 4651554
6 477776 4651574
7 477784 4651607
8 477792 4651630
9 477789 4651652
10 477794 4651668
11 477825 4651670
12 477855 4651675
13 477882 4651671
14 477906 4651678
15 477923 4651701
16 477946 4651714
17 477971 4651721
18 477994 4651723
19 478002 4651737
20 478009 4651761
21 478015 4651779
22 478035 4651794
23 478053 4651792
24 478066 4651792
25 478083 4651794
26 478099 4651794
27 478116 4651789
28 478131 4651780
29 478136 4651757
30 478132 4651736
31 478114 4651714
32 478099 4651694
33 478081 4651678
34 478066 4651661
35 478070 4651633
36 478068 4651604
37 478065 4651587
38 478065 4651562
39 478065 4651539
40 478061 4651516
41 478053 4651500
42 478040 4651491
43 478033 4651472
44 478023 4651449
45 478010 4651425
46 477986 4651406
47 477953 4651394
48 477923 4651379
49 477905 4651371
50 477882 4651364
51 477857 4651356
52 477837 4651359
53 477814 4651368
54 477797 4651363
55 477781 4651368
56 477761 4651368
57 477740 4651368
58 477716 4651376
59 477705 4651391
60 477698 4651412
61 477690 4651432
62 477698 4651450
9
Triangles
the secant of an angle? Not likely. But it isn’t too uncommon to en-
counter the likes of sine and cosine, which are often written sin and
1
At this point you may just need to take cos, respectively. That’s because they are really, really useful1 . And
my word for it, but I hope that you’ll it turns out, nearly all the other trig functions you may have learned
appreciate this fact by the time you
finish this chapter. about are readily defined using sines and cosines! For example, the tan-
gent of an angle can be defined as the ratio of the angle’s sine to its
cosine, but it is so useful that we should recognize it as well.
As far as I’m concerned, on the off chance that you ever need to
manipulate an equation containing the hyperbolic cosine (cosh) of
something, you can look it up or type it into an internet tool like
Wolfram|Alpha is a web-based tool Wolfram|Alpha. If this is your first experience with trigonometry,
developed by mathematician and en- have no fear, we’ll take it slow!
trepreneur Stephen Wolfram. It is based
on the same underlying computational
engine as the math software Mathe-
matica, but can be used (for free with
slightly-limited functionality) from
any web browser. In addition to per-
forming computation and algebraic
simplification, it can attempt to com-
prehend simple written questions and
can retrieve data from a few established
databases, concerning for example Similarity
weather, fincance, and sports.
Triangle Similarity
Two triangles are similar if any one of the following can be es-
tablished:
• AAA. The angles of one triangle are equal to the angles of the
second.
• SAS. Two side-lengths and one angle of one triangle are equal
to those of the second. Side lengths may be scaled by a con-
stant C if that constant is the same for each side.
When you’ve reached the end of this chapter, you should be able to
show how each of these criteria for similarity could be derived from
one of the others. That is left as an exercise for you to work on, one
that can build your intuition for using triangle properties for problem
solving.
a
and doesn’t touch the side that is opposite it. For reasons that might
be apparent in a moment, we choose names for the sides and vertices
that imply a relationship between a vertex and the side opposite A C
b
it. So for example, side a is opposite vertex A (meaning the vertex
A is not one of the endpoints of side a). It probably satisfies your Figure 9.2: A right triangle.
intuition that the size of the angle at a vertex might have some simple
relationships to the length of its opposite side – at least there is a
more intuitive relationship there than between vertex A and one of
the other two sides. Imagine keeping vertices A and C stationary,
but allowing the angle at A to grow. It is plain to see that if the angle
∠ A increases, the vertex B must move up and the length of side a
increases accordingly. This thought experiment produces similar
96 iowa state university
results for the other opposing side/angle pairs as well, and we’ll use
it to our advantage shortly in dealing with triangulation.
One of the most fundamental properties of all triangles is that the
three vertex angles sum to 180◦ (∠ A + ∠ B + ∠C = 180◦ ). For the
special case of a right triangle, the right angle by definition is 90◦ , so
B
the other two angles must be smaller than 90◦ . This seems obvious,
but it has an important consequence: the longest side of a triangle is
se the one that is opposite the largest angle. Therefore, since the right
opposite
te nu
po angle is the largest angle in a right triangle, the longest side (which
hy
we call the hypotenuse) is across from the right angle (Figure 9.3).
In addition to the rule that angles must sum to 180◦ , one of the
A C most powerful properties of right triangles is their adherence to the
adjacent
Pythagorean theorem:
Figure 9.3: A right triangle. a2 + b2 = c2 . (9.1)
This is always true provided that c is the hypotenuse of a right trian-
gle. It turns out that there is a simple modification that can be made
to this if we are dealing with any arbitrary triangle. But before we go
much farther, consider an example setting where a right triangle can
be a useful aid to measurement.
q r
Knowing that the second nest is about 513 m away from the first is
great. But if you were to give instructions to a field assistant to walk
513 m from the first nest to find the second nest, that alone is insuffi-
cient information to get to the correct place. Which direction does she
have to go? You could, of course, simply have her walk north 366 m
and then east 359 m, but that wouldn’t be terribly efficient. What is
missing is obviously the direction. If she’s carrying a compass, you
could give her a bearing or compass direction to follow, but what is
that bearing, and do we have enough information to determine it? p
Figure 9.4: Dickcissel nest distances.
North is toward the top of the page.
9.2.3 Angles and azimuths
At this point, we need to draw more important distinctions regard-
ing coordinate systems and conventions. When we need to be more
accurate than simply saying “northeast”, compass bearings are of-
ten given as angles in degrees. Some people prefer to use quadrant 0
30
bearings, where directions are given with respect to deviations from 330
north or south. For example, due northeast might be expressed as
300
60
“north 45 east”, or equivalently N45◦ E. That can be interpreted
as 45◦ east of due north. Similarly, southeast could be S45◦ E and
270
southwest is S45◦ W. This can sometimes be easier to understand
90
in conversation, but bearings expressed in azimuth are less prone
to mis-interpretation. Azimuth is the compass direction in degrees
120
240
clockwise from north, increasing continuously from 0◦ to 360◦ . In this
system, north is both 0◦ and 360◦ , east is 90◦ , south is 180◦ and west
150
210
is 270◦ (Figure 9.5).
180
Trigonometric Ratios
The sine of an angle is the ratio of the length of the opposite
side (O) to the length of the hypotenuse side (H).
The cosine of an angle is the ratio of the length of the adjacent
side (A) to the length of the hypotenuse side (H).
The tangent of an angle is the ratio of the length of the opposite
side (O) to the length of the adjacent side (A).
• The opposite and adjacent sides are defined relative to the angle
that is the argument of the trig function.
• If we know one angle (other than the right angle) and one side
length, we can find the two remaining side lengths using the
trigonometric functions.
quantitative problem solving in natural resources 99
O
sin θ = (9.5)
H
A
cos θ = (9.6)
H
Since we know θ and H on the unit circle, we can rearrange these
equations to solve for the unknowns. Multiplying both sides of each
equation by H, we get:
H sin θ = O (9.7)
H cos θ = A (9.8)
Now the reason we have done this in the unit circle is that H = 1,
so we essentially end with the definitions sin θ = O and cos θ = A.
An important thing to realize, then, is that we can scale up to any
arbitrary side lengths. Suppose we were interested in not a triangle
with hypotenuse 1 unit, but one with hypotenuse 55 meters. Defining
f = 55 and isometrically scaling all the triangle sides by that factor,
we can see for example that:
fO m
sin θ = . (9.9)
fH m
up
merchantable timber in our forest parcel in Problem 3.5 could include
a set of representative height measurements with a clinometer.
θu D
E In measuring the height of a tree, two readings are often taken
θ Hl
lower sig l
ht line with the clinometer: one to the crown θu and one to the base or
stump θl . The observation point E is a pre-determined distance D
Figure 9.7: Clinometry measurement of from the tree itself. From this information, what is the height of the
tree height.
tree?
We identify H as the quantity of interest, and observe that H =
Hu + Hl . As a first step, we must therefore determine Hu and Hl .
We assume the geometry of the problem allows us to construct two
imaginary right triangles as illustrated and that our clinometer gives
us angles in degrees from the horizontal. Since we know the hori-
zontal distance to the tree D and have measured the angles to the top
(θu ) and bottom (θl ) of the tree, we know the adjacent side length (D)
and an angle for both triangles. The target unknowns are the oppo-
site sides of each triangle, and from sohcahtoa we know that we can
use tan to find the opposite sides when we know the adjacent sides.
quantitative problem solving in natural resources 101
Thus:
Hu
tan θu = (9.10)
D
D tan θu = Hu (9.11)
and
Hl
tan θl = (9.12)
D
D tan θl = Hl (9.13)
and since H = Hu + Hl ,
Law of Sines
Note that this equation is strange in that there are two equals
signs. Don’t get worried, this is just a shorthand that allows us to
say on one line that each ratio between the sine of an angle and
the length of its opposite side is equal to the other corresponding
sine/side ratios. If we wrote each equation with a single ‘=’, there
would be three of them and it would simply take more space. But in
actual implementation you can use any of the implied equalities in
Equation 9.16, such as:
sin A sin C
= . (9.17)
a c
The key concept here is that there is a simple and consistent relation-
ship between the sine of each angle and the length of its opposite
side, and that this applies to all triangles, no matter what size or shape.
That makes plenty of sense right? If you imagine taking a triangle
formed by three knotted rubber bands and lengthening one side
(without changing the length of the other sides), what happens to the
angle opposite that stretched side? It grows right? But to accommo-
date the growth of that angle, the other two angles must get smaller.
The law of sines is especially helpful if you know two sides and one
angle (SSA) or two angles and one side (AAS).
Exercises
1. Use Figure 9.4 and the right triangles formed by dropping the
perpendicular h, to derive the law of sines from the trig functions
you already know.
3. What happens if you try to apply the law of sines to a right trian-
gle?
ALGEBRAIC REASONING
10
Generalizing Relationships
l = ra (10.1)
where we are calling tusk length l and age a. Notice that this way of
writing the relationship implies that we are treating a as an indepen-
dent variable (that is, we can think of it as sort of a cause) and l is the
dependent variable (an effect that depends on the cause). Depend-
ing on the circumstances, these roles could be switched. Indeed, it
is easier to measure tusk length than age for a given elephant, so we
might wish to use measurements of tusks to help determine the age
distribution in a wild elephant population.
Also important to remember is that when we are doing science
instead of just math, the variables usually have units and dimen-
sions, which we discussed previously. If we expect our equation to
be meaningful, the dimensions on the left- and right-hand sides of
the equation need to be consistent (i.e. equal). So in the elephant tusk
example r = l/a is a growth rate, must have dimensions of length
per time, or [ L T −1 ] (see how we get that? would that be the same if
we swapped our independent and dependent variables?). If we have
been measuring length in inches and age in years, our value for r
should be in inches per year.
Great! But as we said above, we might wish instead to know the
age as a function of tusk length. So we need to rearrange things a
bit. Let’s now define m as the number of years of age per inch of tusk
length, which is just the reciprocal of r. In other words, m = 1/r,
which also means r = 1/m. Since we’ve just taken a reciprocal here,
the dimensions of m are just the reciprocal of the dimensions of r,
[ T L −1 ] .
Now we can re-state our new linear relationship as l = a/m, or
a = ml. In this form, we have the dependent variable (age) on the
left-hand side of the equation and the independent variable (tusk
length) on the right hand side, as is convention. But at this point,
110 iowa state university
a = a0 + ml (10.2)
a − a0
m= , (10.3)
l
a1 ( l1 , a 1 )
which we might recognize as the “rise” of the function, ( a − a0 )
divided by the “run” l.
In some cases it is unnecessary, but in others we may need to spec-
a0 length, l
ify something about the domain (a set of upper and lower constraints
l1
on the values of the independent variable) over which a proposed
Figure 10.2: Plot of tusk length l as a
relationship is valid. For example, it doesn’t make sense for an ele-
function of age a. phant to have a negative age any more than it does to have a negative
tusk length. There is probably an upper limit to tusk length as well,
though it is hard to be confident what that might be. To be complete
but conservative, we may specify that the domain of the function as
0 < l < 160 inches. The range for our linear function is the spread
of minimum to maximum values of a corresponding to the minimum
and maximum values in the domain. Note that this last comment
applies to linear functions (though sometimes the signs are reversed),
but for some non-linear functions of interest, maximum and min-
imum values in the range may not correspond to maximum and
minimum values bounding the domain. We’ll see examples of this
later.
Functional relationships that are approximately linear are very
common in the sciences. Indeed, a routine procedure in the analysis
quantitative problem solving in natural resources 111
the independent variable? One area where these polynomials are use-
ful in natural science is spatial measurement. you probably remem-
ber that the areas of squares and circles each depend on the second
power of a characteristic length (side or radius). Likewise, volumes of
spheres and cubes depend on the 3rd power (coincidence?). While we
may never encounter perfect spheres and cubes in the natural world,
we may find occasion to idealize the size and shape of something
(like a sand grain, egg or raindrop as a sphere, a tree root or snake as
a cylinder, etc.) in a simple model so that we can better understand
something about it.
Likewise, some physical phenomena can be described with equa-
tions that depend on a whole number power (often 2) of time or
position. In more complicated problems in the real world, it can also
be advantageous to approximate an unruly function using a so-called
series expansion of the function, which often amounts to a polynomial.
These examples notwithstanding, true polynomial functions do
not arise as commonly in the natural sciences as linear and some
other non-linear functions do2 . An important exception, which we’ll 2
But as we’ll see below, there are
grapple with quite a bit later this term, is the so-called logistic or certainly relationships in the natu-
ral sciences where the relationships
density-dependent growth function. In ecology, this function de- between variables are best described
scribes the theoretical growth of populations constrained by limited with functions that have non-integer
exponents.
space or resources. We can write the basic relationship as:
r 2
G = rN − N , (10.6)
K
where the dependent variable G is the population growth rate y = x3/2
[1 T −1 ], the independent variable N is the number of individuals,
r is a growth constant and K is the carrying capacity. This function is y
y = ax b (10.7)
S = cAiz (10.8)
y = e− x y = abx . (10.9)
x
Exercises
11.1 Correlation
two variables indicates that when one goes up, so does the other. A
high negative correlation indicates that as one goes up, the other goes
down. Low correlation coefficients indicate that a consistent linear
relationship cannot be established. If correlation is established, how-
ever, this analysis doesn’t yet provide details about the functional
relationships present.
11.2 Regression
Exercises
that relationship.
Part V
MODELING
12
Modeling
A = RKLSCP (12.1)
differential equations:
dS
= − βSI (12.2)
dt
dI
= βSI − γI (12.3)
dt
dR
= γI (12.4)
dt
These differential equations are not easily solved in most cases,
but we can use them as a basis for a numerical simulation of disease
dynamics if we are able to estimate the parameters β and γ. A nu-
merical representation of the first equation might look something like
this, for example:
OK, great, but how would we know if this is true more generally?
And what exactly do we mean by larger? Does that mean taller?
Larger volume? This brings up a few issues that become important
when we’re talking about real quantities rather than abstract vari-
ables. Unambiguously defining quantities can be an important first
130 iowa state university
M ∝ V? (12.8)
M
=c (12.10)
V
Now recall that the definition of density is mass per unit volume.
Volume
That’s exactly what we have on the left-hand side of the equation! So
Figure 12.1: Plot of some hypothetical our equation now says that c, the parameter we used to transform
measurements of animal mass and volume into mass, is the same as density! So for an individual ani-
volume.
mal, the parameter that relates mass to volume is density. As we have
done previously, let’s assume that most animals have a density close
to that of water so this proportionality parameter c doesn’t vary sig-
nificantly among species. So to the extent that it is correct to say that
most animal’s body density is close to that of water, we can argue
that larger animals do indeed weigh more, in general.
1
An entertaining and well-composed This is probably not a very profound revelation to you1 . But with
article on some not-so-obvious conse-
only a few more small leaps in logic, we can get somewhere consider-
quences of size differences in animals
is On Being the Right Size, byt J.B.S. Hal- ably more interesting. For more than a century, biologists have been
dane, published in Harper’s Magazine, intrigued by a remarkable relationship between the basal metabolic
March 1926.
rate and body mass for animals of a wide range of sizes and shapes.
Amazingly, if one assembles a large set of data and plots it on a
graph with a logarithmic scale, mice, humans and elephants and
most of the rest fall along a straight line! An equation that describes
quantitative problem solving in natural resources 131
this relationship and the line on the graph looks like this:
B = B0 Mb (12.11)
where B is the basal metabolic rate, M is body mass as before, and
B0 and b are constants (we’ll see what they mean later!). This equa-
tion is yet another power law, and equations with this form pop up
surprisingly often in ecology once you start looking. We’ll get more
into functions and power laws later on. But for now, some important
points should be made:
birth and death rates may scale with the population, such that they
can be represented like this:
B = b × N, D = d×N (13.2)
where b and d are birth and death rates per individual. So, for exam-
ple, if the birth rate is approximately 0.15 individuals per murrelet
2
Note that this birth rate is given per per year2 , and death rate is 0.05 individuals per murrelet per year, we
individual. Obviously males cannot can write our equation for population as:
give birth to offspring, so a better way
to express fertility or fecundity is in
terms of birth rates per female; however N = N0 + 0.15N0 − 0.05N0 (13.3)
the per individual or per capita birth rate
is easier to work with. If we simplify the right-hand side of this, we have N after one year as
a simple function of N0 :
N = 1.1N0 (13.5)
If you plug in 100 for N0 , this gives us an unsurprising result that
population is 110 murrelets. This makes sense, since we get 0.15 ×
100 = 15 births and 0.05 × 100 = 5 deaths during that year.
Now if we project into future years (where t is the number of
years after our initial measurement of population N0 ) with the same
relationship, we’ll see that after another year of births and deaths,
we’ll get:
Nt=2 = 1.1(1.1N0 ) (13.6)
where the quantity in parentheses is the population after one year,
now incremented by another series of births and deaths. We can
rearrange that equation slightly to yield:
And by now you probably see the pattern. If t is the number of years
after an initial population census N0 , our projection of population is:
Nt = N0 × 1.1t (13.9)
B = B0 Mb (13.12)
R = e−1/T (13.14)
N1 = N0 (1 + r ) (13.16)
N = N0 (1 + r )t (13.17)
While we arrived at this result with just some reason and algebra,
a more general solution can be found using calculus. We won’t
worry too much with how this solution is obtained, nor will you
be expected to reproduce it, but it is always nice to see how more
advanced topics can help us with the problem at hand. So here is a
quick summary of how the calculus version works:
If we re-write our first incremental population change equation
above
N1 = N0 + rN0 (13.18)
N1 − N0 = rN0 (13.19)
Notice that the left-hand side is now just the population change over
one year. One of the strategies of calculus that allows elegant solu-
tion of complex problems is to imagine “smooth” changes, where
the increment over which those changes are measured in vanish-
ingly small. While this is obviously an oversimplification of popu-
lation dynamics (i.e., many animals have discrete breeding seasons
so that births are clustered during a relatively small period of time,
and no births occur during the remainder of the year), but in many
cases we don’t need to worry too much about this. We express these
vanishingly-small change increments with derivatives, where the
138 iowa state university
dN
= rN (13.20)
dt
Applying some second semester calculus, we’d come up with the
following solution, which works at all t:
N = N0 ert (13.21)
(1 + r ) N02
N1 = (1 + r ) N0 − (13.22)
K
This looks a bit clunky, but we can clean it up with a little bit of alge-
bra and by making the same kinds of calculus-oriented modifications
that we made above:
dN N
= rN 1 − (13.23)
dt K
As above, the derivative term on the left hand side is the rate of pop-
ulation change as a function of time, or the population growth rate. If
we write the equation with G for growth rate on the left-hand side, it
looks a bit more manageable:
N
G = rN 1 − (13.24)
K
r 2
G = rN − N (13.25)
K
quantitative problem solving in natural resources 139
N0 K
N= (13.26)
N0 + (K − N0 )e−rt
where the first term on the right-hand side is the cost of suppres-
sion activities, while the second term is the net value change in case
of fire. The lowest-cost state is clearly the bottom of the dip in Fig-
ure 3.2, but can we identify that point algebraically? If we use a little
calculus, we can indeed.
In first-semester calculus, we learn that the maxima and minima of
functions can be found by setting the derivative equal to zero. In this
case:
d
(C + Vnc ) = w − kV0 e−kE = 0. (13.28)
dE
For our purposes here, I won’t explain how we arrive at this, but
suffice it to say that when we solve the right-hand equality for E, we
retrieve the effort corresponding to the minimum total C + Vnc . We’ll
140 iowa state university
Exercises
1. Review Section 13.2.1. In the equation for cost plus net value
change, there is a constant k. What are it’s units?
3. Review Section 12.1.2 and ensure that you are comfortable with
the analysis presented there – or that you have developed and
justified your own approach to achieving an analogous solution.
Propose and execute a strategy for incorporating car trips through
the county road network in order to estimate the probable number
of collisions in a given span of time.