Functions of gis
Functions of gis
The GIS then allows the individual layers, or themes as they are called to be
linked. Analysis of the information can then be undertaken using the statistical
and analytical tools that are provided as part of the GIS. By providing spatial
analysis of suitably coded data it is possible to provide striking, visual
representations of data. These representations can often reveal patterns and
trends that might otherwise have gone unnoticed without the use of GIS
techniques.
Under this heading we will consider a large number of functions which a GIS
may be required to perform in order to get the digital mapped data into the
desired format so as to obtain requisite map output or to confidently allow for
any subsequent data analysis. Essentially, this means that the original digital
data may need to be changed in some way, i.e. either by correcting it, updating
it, refining it or by altering it in some desired way. It is possible that many of
the functions can be performed using other types of software, e.g. image
processing packages. The capacity of a GIS to perform pre-processing means
that the user has a huge opportunity to “interactively experiment” with the
available data, thereby allowing for the appropriate data to be derived according
to the task in hand. The efficiency in which individual GIS's perform
manipulations will depend upon the particular algorithms which they use and
the way in which the data is structured.
In essence this function represents the checking and revising of any data which
has previously been captured, with the obvious aim being to minimise errors. In
the case of digitised data, it is often possible and desirable to perform editing
immediately following data capture, i.e. as a final stage in the digitising process,
but it is important to note that many GIS software programmes allow for the
detection and correction of digitising errors as a pre-processing function.
GIS software also contains programmes for verifying the correctness of all
geometric, topological and attribute data, e.g. making certain that all graphical
data is suitably defined, that attribute data does not exceed expected ranges and
that impossible combinations of attributes do not occur. Data may be copied,
deleted, moved, joined, altered, etc. Any of these data editing functions should
be capable of being performed on both the graphical and the textual data. If data
is not carefully verified, and errors remain, then manipulations of the data at a
later processing stage will cause error propagation and multiplication, thereby
invalidating, or at least making less useful, any final GIS output.
For many manipulations it may also be preferable to convert data from a raster
to vector structure, or vice versa. This is necessary since there are still no truly
integrated GIS's which are able to handle both raster and vector data with equal
ease. It is important to note that in the vector to raster conversion (rasterising)
there will be an inevitable loss of accuracy, a factor which would be
exacerbated both with increasing sinuosity of the lines and with increasing
raster cell size. In the raster to vector conversion (vectorising) the GIS software
programme performs a vectorising process which “threads” a line through
groups of pixels using a special “thinning” algorithm. There will be a
consequent need for topological information to be constructed and for
individual features to be identified. These latter requirements can call for
considerable operator intervention, but there are GIS functions which
automatically compute new nodes and links and compile topology tables.
Scale changes are easily accommodated via a simple multiplier function and
maps can easily be rotated to particular orientations. A more complex function
which most GIS's can achieve is the correction for distortions (rectification).
These distortions may occur in the original source data for a number of reasons,
e.g.
(b) The angle of view or relief differences also cause variations in scale.
(d) Using data reduction algorithms to generalise or smooth linear data, e.g. to
thin out coordinates in digitised lines (in order to greatly reduce the amount
of data storage).
(g) Annotations can be added to maps using labels, text, legends or cartographic
symbols.
v. Integrations
These GIS functions simply consist of a series of operations which allow for the
cartographic refinement of the finished map, i.e. at the manipulation stage
factors concerned with map presentation can be improved. This may include
adding a suitable border to the map, varying the width of mapped lines, altering
chosen colours, varying the fonts or font size, altering the layout of the map or
the position of textual features such as the key or title. If a 3-D image has been
created this might be the stage at which a land use categorisation could be
draped over the image.
vii. Interpolation
This is the procedure for estimating the values for any continuous (rather than
discrete) “properties” at unsampled sites along a line or within an area. This
must be based on existing point observation data within the area (or along the
line), which themselves should have been derived using valid measuring and
sampling techniques. The problem in interpolation is choosing the model which
is best able to produce correct interpolations, i.e. a model which suits the data
array and the way in which actual variability occurs. Many simple GIS
interpolation procedures rely on the use of various weighting functions, i.e. this
allows for the logical fact that near points used in an interpolation should count
for more than distant points. There are more complex interpolation models
which cover 2-D arrays of data points e.g. Thiessen polygons, the use of kriging
or Fourier series, or models which can be applied to interpolating linear
pathways, e.g. the fitting of spline functions. A special case of interpolation is
centroid determination. In this function the GIS is able to calculate the co-
ordinate location of the centre of a polygon. Extrapolation is simply the using of
interpolation techniques to extend calculated trends beyond the area of specific
study or interest, or beyond the range of the data held.
The analytical functions which most GIS software provides operate on both the
spatial or the attribute data (or a combination of these). Most of the following
analyses can be performed on both vector or raster structured data, though
inevitably one or other of these is more efficient depending on the actual
analysis being performed. Some authors have reviewed analytical capacities
under headings which correspond to the types of data, i.e. point analysis,
polygon analysis and linear analysis or vector analysis and raster analysis,
though we have chosen to discuss the techniques under the headings of spatial,
statistical and measurement analyses.
i. Spatial Analysis
(a) Connectivity (or network) analysis is useful for determining how well
connected any particular site is via any method of communications. Thus a
connectivity index can be worked out which shows, for example, for all towns
(nodes) in any selected area, the relative number of road or other
communications or pipeline connections (links) which exist between each town
and all the other towns in that area. Connectivity can also be conceived in terms
of distance, cost or time, and it is useful in optimising route allocations.
(e) Location optimisation is now being widely used as a GIS based method
which allows for the selection of optimum locations for the siting of any
activity. This analysis is usually used by larger commercial companies when
seeking, for instance, sites for new retailing outlets or for centralised
distribution points. In these cases various spatially variable economic and social
indicators, such as the social class structure of an area and the population
density, would need to held in a digital geo-referenced form. Similar analyses
are also used by the forestry and agricultural sectors in seeking to optimise their
operations, though here physical rather than economic criteria might be more
important.
As with the spatial analyses, there are a huge range of statistical functions which
any individual GIS might be able to perform. Since many of these functions are
not particular to GIS, i.e. they are commonly performed by statistical packages,
or spreadsheet or database packages, then we need not elaborate on them here,
except to mention that it is possible to link many existing statistical packages
with GIS software as a means of executing statistical analyses. These include
simple descriptive statistics showing measures of centrality, frequency analyses
or measures of dispersion, plus more complex correlations and multi-variate
analyses. Many GIS's are now capable of performing some complex spatial
statistical analyses such as spatial autocorrelation and nearest neighbour
analysis. Spatial autocorrelation is used to provide a measure of contiguity
between areas. If a marine species were shown to be distributed in any of these
ways, then we might need to seek explanations. Nearest neighbour analysis
provides a relative measure of the dispersion of points in a given area, i.e. they
may tend towards clustering, randomness or a uniform spread.
iii. Measurement
3. Data Display
A major anticipated use for any GIS will be to display the data, i.e. the display
capacity will represent the output from the system as presented initially on the
VDU. A fundamental usefulness of the concept of GIS is that it can display
output at any stage in the processing of the data. So the GIS provides the facility
for maps to be incrementally built up, with desired modifications being possible
at any stage. Modifications might be in terms of changes to the data inputs to
the map, or in terms of the visual representation of the map. So the GIS user can
control, review or experiment at any stage in order to achieve a meaningful final
output. All good GIS software will have a range of graphic display features to
control factors such as label size, fonts, colour or shading ranges, line widths,
symbolism, map feature positions, etc. And the format of the display is not
confined to maps - it may be in graphical, tabular or textual forms. The early
1990's has witnessed a huge emphasis in the GIS field on the perceptual science
of “visualization” - how we look at maps, what information is being conveyed,
how people may each view the same mapped scene differently, different ways
of communicating information, etc.
In the near future it is likely that GIS output, and indeed its full range of
functionality, will be capable of display and/or use aboard suitably equipped
fisheries vessels. Many vessels now have quite sophisticated navigation systems
which utilize electronic charts in conjunction with radar and plotting facilities.
These can display a variety of static information covering bathymetry,
navigational features, land masses, restrictive areas, etc, plus the tracks of
moving vessels. It will be a simple progression to extend this functionality so
that other desirable layers will be capable of being integrated and displayed in
an interactive mode, i.e. such that the vessel has on-board ability to perform a
range of required GIS functions.
4. Database Management
A database may be defined as a large collection of related data which has been
structured in an ordered way but which is independent of any particular
application. For GIS purposes this collection of data may be stored externally in
digital form in a purpose created computer software database package, e.g.
Oracle or dBASE, or internally in a database which forms part of the GIS
software. Attribute data is frequently stored externally whilst the geographic
data are more commonly stored within the GIS software.
Distributed databases are also commonly used in GIS and becoming more so. In
this case the data may be held in disparate sources within or outside an
organisation. Obviously, if a GIS has access to such databases then a vast range
of extra data becomes available to the system. In the fisheries context, it is easy
to envisage that a GIS could offer greater functionality if it could have direct
access to oceanographic, meteorological and perhaps environmental databases,
all of which were likely to be situated separately from the fisheries GIS. Anon
(1993b) provides detailed information on how a major database, the Regional
Maritime Database (BDRM), is being built to cover the marine areas along the
West African coastline. This database will be capable of being accessed by any
or all of the 10 participating countries, and all information is being specifically
geo-referenced so that GIS functionality can be ensured.
The reasons for having a DBMS are fairly basic. Thus firstly, it is obvious that
the data needs to be maintained in the sense of being kept up-to-date, correct,
properly ordered, properly structured, etc. It also has to be managed so that it is
properly understood, and that it is available how, when, where and to whom it is
required. Storage structures must be constantly monitored so as to minimise
storage space and to maximise searching efficiency. And since most databases
are constantly growing, new fields might need adding or indeed old fields might
need deleting. A database manager is usually assigned who can not only
manage all of the above, but who can also regulate access to the database, cope
with systems failures when they occurred and who is able to link databases with
external databases if required. He or she may also be required to cope with the
legal sides of data management, i.e. making certain that inaccurate decisions are
unlikely to be made as a consequence of using his data, and making sure that
copyright infringements are not made.
(a) Hierarchical. In the hierarchical model each record can have a number of
links to lower “levels”, but only one link to a higher level. The highest link is
the “root”, lower levels are called “children” and levels above are called
“parents”. Hierarchical data structures are easy to understand and to update or
expand and they are useful where up and down searching is required, but they
are not very good in circumstances where horizontal searching is carried out,
i.e. where it might be necessary to locate all records which are at one level,
since there are no connections at the same level.
(b) Networked. This is similar to the hierarchical data model but here it is
possible to have more than one parent, and thus many-to-many relationships can
be found. A networked database structure can be analogous to a
communications network where there may be many linkages between any
combinations of centres. This type of data structure makes good use of the
available data, with rapid connections being possible, but it is difficult to create
and maintain. Both hierarchical and networked structures are now seldom used
in GIS's.
Relational databases are very flexible and are easy to establish, use and
maintain. Almost any relationship is possible to work out by the use of Boolean
logic and mathematical operations, and additional data can easily be added to
the database. The main drawback lies in the fact that searches can be very time
consuming, i.e. given the calculations and other specifications involved and the
huge numbers of tables which might be needed. The Standard Query Language
(SQL) has been developed as the standard language for use with the relational
DBMS, and this kind of database is the one most frequently used by GIS's.
(d) Object Oriented. This group intends to make object oriented DBMS the
norm for the future by ensuring that they are extremely efficient and that there is
a world-wide common framework for their development so that they can work
across all environments, on all hardware platforms and under any operating
system. These databases take a sophisticated view of geographical entities and
as such object oriented DBMS are conceptually difficult to explain and to
understand. Basically all entities are considered as having the three fundamental
concepts of: object, class and inheritance. For instance, geographic objects
might be “road”, “port”, “sea”, etc. Each object can then in turn have a class
such as “Secondary (road)”, “Fishing (port)”or “Shallow (sea)”, and they may
also have sub-classes. Each object can also be defined as having certain
properties, e.g. a fishing vessel may have an owner, a value, a size, etc., and
each object may have certain functions (operations or methods) which can be
performed on them, e.g. select, measure, locate, modify or draw. Inheritance
means that when a class is added to an object, then all aspects of “class” can
potentially be included in the new description, e.g. all aspects pertaining to
“Fishing” can potentially be linked to the class of “Fishing port”. For those
seeking a more detailed explanation of object-oriented DBMS, then or Laurini
and Thompson (1992) or Cooper (1993) give reasonably easy definitions.