Intercomparison of Photogrammetry Software For Three-Dimensional Vegetation Modelling
Intercomparison of Photogrammetry Software For Three-Dimensional Vegetation Modelling
Intercomparison of
rsos.royalsocietypublishing.org
photogrammetry software
for three-dimensional
Research
vegetation modelling
Cite this article: Probst A, Gatziolis D, Strigul Alexandra Probst1 , Demetrios Gatziolis2 and
N. 2018 Intercomparison of photogrammetry
software for three-dimensional vegetation Nikolay Strigul3
modelling. R. Soc. open sci. 5: 172192. 1 Department of Biology, University of Washington, Seattle, WA, USA
http://dx.doi.org/10.1098/rsos.172192 2 USDA Forest Service, Pacific Northwest Research Station, Portland, OR, USA
3 Department of Mathematics and Statistics, Washington State University Vancouver,
2018 The Authors. Published by the Royal Society under the terms of the Creative Commons
Attribution License http://creativecommons.org/licenses/by/4.0/, which permits unrestricted
use, provided the original author and source are credited.
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
2.2. Imagery
2.2.1. Unmanned aerial system-based aerial images
The set of aerial images used is detailed in [2]. A small UAS equipped with a GoPro 3+ Black camera
was programmed to follow a circular trajectory (20 m radius) around a 16 m tall deciduous tree at a
constant 12 m above-ground elevation with the camera oriented towards the vertical middle of the tree.
The UAS was moving at a constant speed and acquired 200 5 MB images during a windless day. The
camera features an f/2.8 wide-angle lens placed in front of a 12-megapixel sensor. No permissions were
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
required prior to conducting our fieldwork. Using unmanned aerial vehicle (UAV)-based, nadir looking
4
imagery featuring sparse and low vegetation on flat land, Wu [14], the author of the VisualSfM software,
documented that scene reconstructions obtained by using the generic image calibration model embedded
quantitative metric suitable for our purposes. All calculations were performed using the R software
5
(www.r-project.org).
3.1.3. Ghosts
By this term, we refer to a single objects or object parts that appear in more than one instance in
a point cloud. They are probably produced because of errors in the derivation of certain camera
positions. CMPMVS replicated parts of the synthetic tree’s main stem but not branches or foliage.
The duplicated stem instance was accompanied by a separate, distinct shadow cast on the grass
background. VisualSFM also generated ghosts albeit smaller in size compared with those from CMPMVS.
Duplicates of large, solid objects such as the main stems of trees are easily discernible. Duplicates of
foliage dispersed among tree crowns, however, are very difficult to identify, and can have ramifications
on desired vegetation measurements, such as volume and area values. We were able to detect
these in the VisualSFM-derived point clouds because they had distinct spectral features compared to
their surroundings.
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
4. Discussion
The pioneering work by Snavely et al. [20] was designed to accommodate any collection of digital images
irrespective of origin, resolution and effective camera lens focal length or combinations thereof. Since
then there has been a proliferation of proposed improvements, either novel or adaptations of pre-existing,
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
Figure 2. UAS-acquired scene image (a), and software-generated dense three-dimensional reconstructions (b–f ).
analogue photogrammetry ideas. The set of software packages evaluated in this study are only a subset of
a range of solutions available today. Considering the impetus provided by technological advancements
and public interest in relevant applications, we expect further advancements to photogrammetric
software.
A characteristic shared by the software packages examined is the absence of detailed technical
documentation. For a few of them this issue is partially mitigated by online forums where users share
experiences and ask for and provide feedback to processing challenges. While the information exchanged
in these discussions can be valuable, it can also be speculative, subjective or applicable to a narrow set
of conditions. In this study, we aimed at providing a detailed quantitative evaluation of performance at
natural scenes.
In addition to the dearth of technical documentation, comprehensive sensitivity analysis with intent
to optimize parameter values for a given set of UAS images is inhibited by the fact that photogrammetric
processing, and dense cloud derivation in particular, is a very computationally intensive process. Based
on our prior experience and the work performed in this study, we believe it is indeed possible, with a lot
of effort and time investment, to occasionally improve on a structural attribute (completeness, positional
accuracy, etc.) of a dense point cloud by trying combinations of values for the three primary controls
mentioned in §2.1 instead of using the default values. However, the improvement is rarely substantial,
regardless of whether the evaluation is visual or quantitative. Further, we have observed numerous cases
where the parameter value combination proven to improve the dense point cloud of one scene has little
effect on another similar scene.
This apparent absence of consistency is probably rooted to the fact that the concept of obtaining
three-dimensional scene information using structure-from-motion techniques and the algorithms that
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
Figure 3. Lateral and nadir views of real scene reconstructions at varying scales.
support it have been designed for opaque objects with Lambertian (diffuse) or approximately Lambertian
surface reflectance. Indeed, scenes comprising exclusively opaque objects tend to contain few artefacts.
Similar behaviour is observed with point clouds generated from UAV-based imagery with nadir-oriented
cameras over forested landscapes, a popular application [3–5]. In this configuration, the crowns of
trees always have a terminal background, the forest floor and usually exhibit minute changes in solar
illumination direction from one image to the next. In the viewing configuration of the real-world scene of
this study, the tree crown background can be at infinite distance. The implication is that two neighbouring
pixels positioned along an epipolar ray can be at markedly different distances from the camera. Besides,
in two successive camera positions, one of them can be subject to direct solar illumination while the other
is not, thanks, for example, to an intervening piece of foliage, leading to two images with very different
overall brightness and contrast. Algorithms that anticipate only gradual changes in object parallax and
illumination geometry, typical of opaque objects, fail to perform consistently for non-solid ones. Leaves
behaving as nearly specular surface reflectors and of profiles that vary dramatically with changes in
viewing geometry further compound the frequency and magnitude of artefacts.
Table 1. Summary of artefacts in three-dimensional reconstructions.
floating
artefacts virtual attached partial background
software versus real scene artefacts reconstructions issues ghosts
CMVS/ PMVS 150 / 411 few, thin layer of grass on synthetic tree incomplete real tree, missing most tree and synthetic scene targets small pieces of real tree foliage
branches of its upper half partially reconstructed reconstructed elsewhere
.............................................................................................................................................................................................................................................................................................................................................................................
CMPMVS 39 / 58 few in synthetic scene. Large number of a few missing virtual tree branches ground discontinuities large sections of the synthetic tree
artefacts attached to the top of real trees reconstructed elsewhere in
scene
.............................................................................................................................................................................................................................................................................................................................................................................
MVE 1/7 sky attached on trees large missing section in the upper object shape in real scene no ghosts
middle of the real tree background deteriorates with
distance
.............................................................................................................................................................................................................................................................................................................................................................................
SURE 127 / 18 sky artefacts on upper parts of crowns, more complete tree reconstructions practically no background in real no ghosts
pronounced in the real scene scene
.............................................................................................................................................................................................................................................................................................................................................................................
PhotoScan 93 / 54 sky artefacts at the top of tree crown, larger in complete tree reconstruction but ground discontinuities, distorted no ghosts
(lowest quality) the real scene hazy shape with hollow background trees
appearance
.............................................................................................................................................................................................................................................................................................................................................................................
PhotoScan 35 / 70 small grass and sky artefacts on synthetic tree. complete reconstruction of trees ground discontinuities, distorted no ghosts
(low quality) Large sky artefacts in real scene but somewhat hazy shape background trees
.............................................................................................................................................................................................................................................................................................................................................................................
PhotoScan 6 / 24 thin layers of grass mixed in synthetic tree complete reconstruction of trees ground discontinuities no ghosts
(medium quality) crown. Large upper crown artefact in real
scene
.............................................................................................................................................................................................................................................................................................................................................................................
PhotoScan 3 / 27 misplaced thin layers of grass and sky in complete reconstruction of trees discontinuities in real scene no ghosts
(high quality) synthetic scene. Small sky artefacts except for selected branches ground. Missing parts of
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
PhotoScan 0 / 29 no synthetic scene artefacts, small layer of sky almost half of the synthetic tree is discontinuities in real scene no ghosts
(highest quality) to the real scene tree missing ground. Partially reconstructed
ground.
.............................................................................................................................................................................................................................................................................................................................................................................
................................................
9
1.0
10
0.6
0.4
PhotoScan (high)
PhotoScan (medium)
CMVS/PMVS
CMPMVS
0.2 SURE
PhotoScan (low)
MVE
PhotoScan (lowest)
PhotoScan (highest)
0
0 0.2 0.4 0.6 0.8 1.0
false positive rate (fraction of points that are not in reference cloud)
Variability in the distance of tree crown components depicted in overlapping image regions from
corresponding camera locations induces variability in representation scale. Owing to occlusion from
crown components at the near end of the crown, components at the middle or far end may be visible only
partially, even where they are positioned within the overlapping field of view of successively acquired
images. Scale and occlusion rate variability paired with a high-quality setting specified by the user filter
out scene components with representation frequency below the internal image number threshold. They
thus lead to sizeable discontinuities or gaps and explain the high omission rates observed when using the
‘highest’ PhotoScan quality setting. SURE avoids this issue by excluding background scene components
from the point clouds, while MVE follows the exactly opposite strategy. It prefers to deliver scene
representations with larger spatial extent while accepting higher frequencies of artefacts and reduced
point densities for background objects.
The apparent commission errors observed in the actual (figure 2) and synthetic scene (figure 5) for
MVE and CMVS/PMVS can probably be reduced, if not completely removed, by masking on each image
the regions representing the sky background. The masking operation can be accomplished by applying
a combination of spectral and textural filters, given that a clear or cloudy sky has distinct digital pixel
number ranges and texture from those of vegetation, ground or man-made objects. Even with this image
preprocessing step, however, the upper portions of tree crowns will still inherit some of the sky’s spectral
signature, as foliage and vegetation material occupy only a portion of each pixel. Apparently, point cloud
derivatives that capitalize solely on geometric attributes would not be affected by such colour-related
artefacts.
For the rest of the tree crowns, simultaneously reducing the omission and commission artefacts
in a systematic manner is probably infeasible, at least in the present state of software development.
This is because in all software tested, the parametric configuration and application of pertinent
algorithms appears to be static, in the sense that it does not adapt to local conditions. Enabling dynamic
parametrization could be programmatically complex and further reduce processing efficiencies given
that a second pass over the entire image set would be required, after the initial dense reconstruction is
complete and the approximate structure of the scene is known.
It is suggested that for scenes dominated by crowns with complete and rigorous foliage along their
entire vertical profile, the user specifies settings that require crown components to be present in a larger
number of images, four or more, with processing of every other pixel along epipolar rays. Conversely,
for more open, see-through crowns the minimum number of images required for scene component
inclusion in the dense point cloud can be lower to avoid discontinuities in the dense point clouds
generated.
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
Figure 5. Software-derived point clouds aligned to reference synthetic tree in lateral and nadir views (first two columns) and coloured
by classes of local distance discrepancy between reference and models (third and fourth columns). The class colouring scheme is blue for
0.0, green for 0.0075, yellow for 0.015 distance, red for 0.0225 and purple for larger distances (outliers). Distance values are relative to unit
scene width.
software AUC
PhotoScan (high quality) 0.948
.........................................................................................................................................................................................................................
CMVS/PMVS 0.937
.........................................................................................................................................................................................................................
CMPMVS 0.935
.........................................................................................................................................................................................................................
SURE 0.930
.........................................................................................................................................................................................................................
MVE 0.898
.........................................................................................................................................................................................................................
The AUC metric computed from the ROC curves for the synthetic scene shows that the ‘high’ and
‘medium’ settings of PhotoScan, the commercial product, is performing better than all other software
tested, probably thanks to embedded heuristics, but still contains notable artefacts. Whether the observed
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
performance is significantly superior to that of CMVS/PMVS, CMPMVS and SURE, the freeware options,
12
is not particularly clear. MVE and the other PhotoScan settings clearly have inferior performance. The
same software ranking persists for the actual scene, although the evaluation in this case is purely
5. Conclusion
Photogrammetry-based analysis of vegetation structure is an emerging area of research. This work
introduces an original and flexible approach for intercomparison of workflows and software, potentially
useful for alternative scene compositions and application areas. It evaluates their ability to generate
dense point cloud reconstructions of trees and shrubs. By including a synthetic, yet highly realistic scene
with precisely known object dimensionality, it delivers a detailed, quantitative assessment of software
performance. Study findings confirm that the same set of UAV-based images, or synthetic alternatives,
processed with different software implementations of the structure-from-motion concept yield point
clouds with different spatial characteristics. Findings suggest that the commercial software evaluated has
slightly superior performance compared to freeware alternatives but scene representation completeness
and positional accuracy does not improve monotonically with increases in processing complexity and
execution time. Our findings pertain to vegetation structure and scene illumination conditions similar
to those used in this study. Additional investigations would be needed prior to claiming applicability
to other conditions. The methodology presented can serve as a guide to forest inventory specialists and
analysts interested in obtaining detailed, three-dimensional representations of trees present in field plots
economically, following an established road map.
Data accessibility. Original aerial UAS-based and synthetics imagery data used for the comparison of photogrammetric
algorithms are available on the Dryad Digital Repository: http://dx.doi.org/10.5061/dryad.2459s12 [21].
Authors’ contributions. A.P., D.G. and N.S. equally participated in the design of the study and wrote the manuscript. A.P.
has conducted three-dimensional reconstructions. All authors gave final approval for publication.
Competing interests. We declare we have no competing interests.
Funding. This work was partially supported by Simons Foundation (no. 283770 to N.S.) and a grant by the US Forest
Service titled ‘Evaluation of Visual Structure from Motion Technology for Forest Inventory Field Operations’.
Acknowledgements. We are grateful to Jean Lienard for technical help with software and to Mathias Rothermel for
providing the SURE software package for evaluation.
CMPMVS depth map, dense point cloud, mesh command line 0.6.0 M. Jancosek, T. Pajdla
mesh recon.
.........................................................................................................................................................................................................................
MVE depth map, dense point cloud, image orientation, dense command line, GUI 05/2016 S. Fuhrmann,
floating surface recon., point cloud, mesh F. Langguth,
mesh cleaning M. Goessele
.........................................................................................................................................................................................................................
OpenMVS dense point cloud, mesh mesh command line 0.7 Git-hub user
recon., mesh refining, cdcseacave
mesh texturing
.........................................................................................................................................................................................................................
SURE depth map, dense point cloud, mesh command line, GUI 0.0 M. Rothermel,
mesh K. Wenzel
.........................................................................................................................................................................................................................
PhotoScan image orientation, dense command line, GUI 1.3.1 Agisoft LLC
point cloud, mesh
.........................................................................................................................................................................................................................
CMPMVS. CMPMVS was developed by Michal Jancosek and Tomas Pajdla. It is a multi-view
reconstruction software specifically designed to reconstruct weakly supported surfaces, such as
transparent glasses or obscured ground planes [27]. It requires a priori known camera positioning
and orientation information, in our case supplied by VisualSFM. Using a plane sweeping algorithm,
CMPMVS creates a depth map for each image, which is then used to create a point cloud and finally a
three-dimensional mesh. We implemented CMPMVS using the default parameters set in the batch file
provided.
MVE. Researchers Simon Fuhrmann, Fabian Langguth and Michael Goessele created the
reconstruction software pipeline known as Multi-View Environment (MVE) [28]. Like VisualSFM, MVE
contains software for the complete reconstruction pipeline. However, we chose to use VisualSFM’s sparse
reconstruction in order to maintain consistency across comparisons.
OpenMVS. OpenMVS is a recently released open-source library aiming to provide a complete set of
dense reconstruction algorithms (http://cdcseacave.github.io/openMVS/). OpenMVS creates a dense
reconstruction and a mesh and furnishes the mesh surfaces with texture. At the time of this study, despite
our best efforts, we could not manage to obtain reconstructions of the virtual reality environment with a
quality consistent to the other workflows, and decided to omit this program from the comparisons.
SURE. SURE is a three-dimensional reconstruction software developed by Mathias Rothermel &
Konrad Wenzel [29]. It is not an open-source program but provides licences for academic use. At SURE’s
core is the LibTSgm library, which contains modules that perform image triangulation from camera
parameters. SURE requires a sparse reconstruction input, and accepts many forms including VisualSFM’s
nvm file.
Agisoft PhotoScan. Agisoft PhotoScan is a commercial three-dimensional reconstruction software
produced by Agisoft LLC [30]. It can be used under commercial and educational licensing. PhotoScan is
an all-in-one three-dimensional photogrammetry software which handles the entire modelling process
from feature matching to dense reconstruction.
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
Figure 7. Images of the synthetic scene acquired at various viewing points around the targeted tree.
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
2a 2b
Figure 8. CMPMVS ghosts. 1b and 2b regions depict duplications of regions 1a and 2a, respectively.
Figure 9. MVE attached artefacts. Region 3 shows a mixture of grass and sky enveloping the upper portion of the tree crown. Region 4
shows sky texture attached to the top of the tree.
Figure 10. SURE artefacts. Region 5 shows upper crown leaves layered by points coloured as sky background. Region 6 shows the same
phenomenon but this time of leaves and branches layered by points coloured as grass.
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
Figure 11. Lateral and nadir views of real scene reconstructions obtained with different quality setting of PhotoScan software.
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
17
Figure 12. PhotoScan-derived point clouds with different quality settings, aligned to reference synthetic tree in lateral and nadir views
(first two columns) and coloured by classes of local distance discrepancy between reference and models (third and fourth columns). The
class colouring scheme is blue for 0.0, green for 0.0075, yellow for 0.015 distance, red for 0.0225 and purple for larger distances (outliers).
Distance values are relative to unit scene width.
References
1. Snavely N, Seitz S, Szeliski R. 2008 Modeling the 2. Gatziolis D, Lienard JF, Vogs A, Strigul NS. (doi:10.1371/journal.pone.
world from internet photo collections. Int. J. 2015 3D tree dimensionality assessment 0137765)
Comput. Vis. 80, 189–210. (doi:10.1007/s11263-007- using photogrammetry and small 3. Puliti S, Ørka HO, Gobakken T, Næsset E. 2015
0107-3) unmanned aerial vehicles. PLoS ONE 10, e0137765. Inventory of small forest areas using an unmanned
Downloaded from http://rsos.royalsocietypublishing.org/ on November 18, 2018
aerial system. Remote. Sens. (Basel) 7, 9632–9654. with multi-view stereo quality assurance by 22. Wu C. 2013 Towards linear-time incremental
18
(doi:10.3390/rs70809632) confidence prediction. In The IEEE Conf. on Computer structure from motion. In 2013 Int. Conf. on 3D
4. Dandois JP, Olano M, Ellis EC. 2015 Optimal altitude, Vision and Pattern Recognition (CVPR) Workshops, Vision-3DV, pp. 127–134. New York, NY: IEEE. See
& 2018 The Authors. Published by the Royal Society under the terms of the Creative
Commons Attribution License http://creativecommons.org/licenses/by/4.0/, which permits
unrestricted use, provided the original author and source are credited.