Syllabus: Total: 45 Periods
Syllabus: Total: 45 Periods
UNIT – V: APPLICATIONS 9
GIS Applicant - Natural Resource Management - Engineering - Navigation - Vehicle
tracking and fleet management - Marketing and Business applications - Case studies.
TOTAL: 45 PERIODS
TABLE OF CONTENTS
UNIT - I
UNIT I FUNDAMENTALS OF GIS
1.1. GIS Overview 1.1
1.1.1. Introduction to GIS 1.1
1.2. Basic Spatial Concepts 1.2
1.2.1. Location 1.2
1.2.2. Distance 1.3
1.2.3. Network 1.3
1.2.4. Neighborhoods and regions 1.4
1.2.5. Scale 1.5
1.2.6. Spatial Heterogeneity 1.5
1.2.7. Spatial Dependence 1.6
1.2.8. Objects and Fields 1.7
1.3. Geographic Coordinate Systems 1.8
1.3.1. Sphere and Ellipsoid 1.9
1.3.2. Geoids 1.9
1.3.3. Datum 1.10
1.4. GIS as an Information System 1.11
1.5. Defining GIS 1.11
1.6. History of GIS 1.11
1.7. Components of a GIS 1.13
1.7.1. Software 1.13
1.7.2. Hardware 1.13
1.7.3. Data 1.14
1.7.4. Users 1.14
1.7.5. People 1.14
1.7.6. Methods 1.14
1.8. Proprietary and Open Source GIS Software 1.15
1.8.1. AGIS-Map 1.15
1.8.2. Autodesk 1.15
1.8.3. Bentley Systems, Inc. 1.15
1.8.4. Cartographica 1.15
1.8.5. DeLorme 1.15
1.8.6. ESRI 1.15
1.8.7. Intergraph 1.16
1.8.8. Manifold 1.16
1.8.9. Ortelius 1.17
1.8.10. MapInfo – Pitney Bowes Business Insight (PBBI) 1.17
1.8.11. Maptitude 1.17
1.8.12. My World 1.17
1.8.13. Super-geo Technologies 1.18
1.8.14. Tatuk-GIS Editor 1.18
1.8.15. Terrain Tools 1.18
1.8.16. TerrSet Geospatial Monitoring and Modeling System 1.18
1.8.17. TNT Products 1.19
1.9. Open Source GIS Software 1.19
1.9.1. Quantum GIS (QGIS) 1.19
1.9.2. DIVA-GIS 1.19
1.9.3. Free QGIS Courses from Geo-Academy 1.19
1.9.4. Geo-Da 1.19
1.9.5. GVSIG 1.20
1.9.6. Map-Window GIS 1.20
1.9.7. SAGA (System for Automated Geo-scientific Analyses) 1.20
1.9.8. UDIG 1.20
1.9.9. Environmental Benefits Mapping & Analysis Program 1.20
1.10. Data Types 1.21
1.10.1. Spatial Data 1.21
1.10.2. Non-Spatial Data 1.22
1.10.3. Spatial Data Models 1.22
1.10.4. Attribute Data Models 1.23
1.11. Types of Attribute Data 1.24
1.11.1. Character Data 1.24
1.11.2. Numeric Data 1.25
1.11.3. Date/Time Data 1.25
1.11.4. BLOB Data 1.25
1.12. Scale of measurement/Level of measurements 1.25
1.12.1. Nominal scaling 1.27
1.12.2. Ordinal scaling 1.28
1.12.3. Interval scaling 1.28
1.12.4. Ratio scaling 1.28
Two Marks Questions with Answers 1.29
Review Questions 1.33
UNIT-II
SPATIAL DATA MODELS
2.1. Introduction 2.1
2.2. Database Structures 2.1
2.3. Data Structure Models 2.2
2.3.1. Hierarchical Data Structure Model 2.2
2.3.2. Network Data Structure Model 2.3
2.3.3. Relational Data Structure Model 2.4
2.3.4. Object Oriented Database Structure 2.5
2.4. Definition - What does Entity-Relationship Diagram (ERD) 2.6
2.4.1. Entity-Relationship Diagram (ERD) 2.6
2.4.2. E-R Diagram 2.7
2.5. Spatial Data Models 2.8
2.5.1. Vector Data Formats 2.9
2.5.2. Raster Data Formats 2.10
2.5.3. Image Data 2.12
2.5.4. Vector and Raster – Advantages and Disadvantages 2.13
2.6. Raster Data Structure 2.14
2.6.1. Run length encoding 2.15
2.6.2. Block encoding 2.15
2.6.3. Chain encoding 2.16
2.6.4. Quadtree 2.16
2.6.5. A Single Layer Raster Data can be represented using 2.17
2.6.6. Comparison between Vector and Raster Data Models 2.17
2.7 Data Compression 2.18
2.7.1. Raster Data Compression 2.18
2.7.2. Run length Encoding 2.19
2.7.3. Quadtree 2.20
2.8. Vector data structure 2.20
2.8.1. Connectivity 2.21
2.8.2. Contiguity 2.22
2.8.3. Containment 2.22
2.8.4. Simple Features 2.23
2.8.5. Topologic Features 2.23
2.8.6. Fully topological polygon network structure 2.24
2.8.7. Triangular Irregular Network (TIN) 2.24
2.9. Raster vs Vector Models 2.26
2.9.1. Vector Data 2.27
2.9.2. Raster Types: Discrete vs Continuous 2.29
2.9.3. Discrete Raster’s have Distinct Values 2.30
2.9.4. Continuous Rasters have Gradual Change 2.30
2.10. Vector vs Raster: Spatial Data Types 2.30
2.10.1. Vector and Raster – Advantages and Disadvantages 2.31
2.10.2. Vector Data 2.31
2.10.4. Raster Data 2.31
2.11. TIN and GRID data models 2.32
2.11.1. Different Types of TIN Methods and Processes 2.33
2.11.2. Components of TIN 2.33
2.11.3. TIN Data Model 2.34
2.12. Grid/Lunr/Magi 2.35
2.12.1. IMGRID GIS 2.35
2.13. Open Geospatial Consortium (OGC) 2.36
2.13.1. Standards 2.36
2.14. Spatial Data Quality 2.38
2.14.1. Spatial Data Quality 2.38
2.14.2. Usage Components 2.40
2.14.3. Causes of Error 2.40
2.14.4. Map Generalization 2.41
2.14.5. Sources of Possible Errors 2.43
2.14.6. Controlling Errors 2.44
Two Marks Questions with Answers 2.46
Review Questions 2.51
UNIT-III
UNIT - V
APPLICATIONS
FUNDAMENTALS OF GIS
GIS therefore refers to a set of three aspects of our modern world, and offers new ways to
deal with them. The concept of information is indeed the heart of the rapidly growing field of
Geographic Information Systems or GIS. As the world moves into the Information Age,
meaningful data and information are becoming the major „currency‟. With the continuous
advances in computer technology, it is easy to concentrate on dazzling systems and software, but
the real value of any product is the data and the information such data provides. At the heart of
any GIS is information.
These concepts have been a foundation for researchers for centuries (see Classics in
Spatial Thinking). They have been augmented in recent decades with computational and
visualization tools and with vast and easily accessible information resources. These concepts and
tools must be as central to general education as reading, writing, and arithmetic. In conjunction
with the appropriate spatial tools, they provide a basic scaffold for designing research, solving
problems, and structuring education programs.
Spatial concepts invites contributions about other concepts for spatial thinking (e.g., in
design fields and in the humanities and arts). Examples may include the link between form and
function in architecture, the search for pattern in speech and text, the use of spatial notation in
music, the use of spatial metaphor in the sciences and humanities, the importance of place in
cultural and social studies, and the spatial elements of aesthetics in the visual arts. If you wish to
add concepts to the listing.
1.2.1. Location:
Every type of spatial data has two components: a location and some attribute(s). In this
way, location can be seen as a fundamental trait that both defines spatial data and separates it from
other types of information. Broadly speaking, there are two types of locations: absolute and
relative. Absolute location refers to an exact position on the Earth‟s surface defined by some
coordinate system. Street addresses and latitude/longitude coordinates are good examples of
absolute locations. Relative locations, on the other hand, are defined in reference to other objects.
For example, one could define UCSB‟s relative location as 10 miles west of downtown Santa
Barbara.
Fundamentals of GIS 1.3
The method of determining or measuring location is called georeferencing. There are
many different ways of defining a georeference, but they all must meet a few requirements. First,
the georeference must be unique so that only one location is described. Second, georeferences
must have an accepted meaning that is shared so that most users understand its implication.
Finally, georeferences must be viable throughout time so that their meaning is not lost. Metric
georeferences are those which define location by measurement and are of particular importance in
analyzing spatial data. In order to properly define a metric georeference, consideration must be
given to the shape of the Earth, map projections and coordinate systems, and positional accuracy.
Representing location is also an important concept in the use of spatial data. Spot locations
are represented as zero-dimension points, lines as one-dimensional poly-lines, areas as two-
dimensional polygons, and volumes as three-dimensional polyhedra.
1.2.2. Distance
Distance describes the measurement or separation of two objects or places. At its most
basic, measured distance provides a mechanism for describing spatial extent. In spatial reasoning,
the classical case is that a positive relationship exists between distance and locational attribute
similarity. The relationship, as worded by the Swiss-American cartographer, Waldo Tobler, is,
“Everything is related to everything else, but near things are more related to each other.”
Examining the strength, limitations, and exceptions of the distance-similarity correlation serves as
the foundation of much quantitative and qualitative spatial research. Measured distance also
provides a necessary dimension for describing position. All real points can be plotted with respect
to another by coupling measures of distance with heading and/or time. Common application of
distance include spatial analysis and modeling, physics and gravitation, distance decay, buffers,
geodesics, route description and optimization, and qualitative comparison of place.
1.2.3. Network
A network is a physical or conceptual system of linkages among entities. Networks offer
an infrastructure for representing the anisotropic relationships of various constituents and
constituent attributes. Typically, a network connection denotes increased accessibility or
relatedness along a link, and linkages may override the default notion that closer features are more
accessible and related. For instance, driving from one side of a river to the other often requires a
1.4 Geographic Information System
circuitous path many orders of magnitude greater than the direct distance from origin to
destination.
Utilizing neighborhoods and regions can allow one to make inferences about data from its
spatial context. Many fields use neighborhoods and regions in just this way. In remote sensing,
neighborhood statistics can be calculated using the values of adjacent pixels. In business, service
regions are defined to maximize the number and locations of stores or restaurant franchises. In
landscape ecology, metrics have been developed to measure the fragmentation of environmental
patches within an ecosystem.
There are problems with making inferences between data at different scales, however. The
modifiable area unit problem (MAUP), or ecological fallacy, involves two issues that underscore
the use of data aggregated to the neighborhood or region. First, there is a problem of scale. In this
way, results of analysis at one scale are not comparable to results at higher or lower levels of
aggregation. For example, it does not make sense to use population density calculated at the state
level to describe density at the county level. Second, there is a problem of aggregation. Changing
region boundaries can have immense effects on the neighborhood statistics. The gerrymandering
of voting districts is one example of an aggregation problem.
Fundamentals of GIS 1.5
1.2.5. Scale
In spatial reasoning, scale describes the dimensional relationship between a representation
and reality. Due to the large variation of all space, scale is used to project reality to more useful
and meaningful sizes. For large expanses, scale is reduced (e.g., fitting the entirety of earth‟s
surface on to a paper map), and for miniscule distances, scale is increased (e.g., enlarging and
schematizing chemical reactions). Often, scale is denoted as a fraction where a unit of measure in
reality is compared to the same unit on the projection. For instance, a paper map of a landscape
showing a scale of 1:10,000 would mean that a drawn unit on the map represents 10,000 of the
same unit in reality.
Offering spatial data at scales different from their original data collection granularity may
imply changes of meaning; thus, data product limitations should be carefully considered by both
creators and consumers of spatial information. As a fundamental spatial concept, applications of
scale permeate many human activities including data collection, cartography, art, architecture,
engineering, and nanotechnology.
There are many research implications caused by the fundamental heterogeneity of spatial
data. In most cases, spatial data can be said to have a non-constant mean and variance throughout
a study area. That is, local statistical parameters change with location and are thus not uniformly
distributed. This characteristic of spatial data is termed non-stationarity and directly impacts
research in areas such as sampling design. For example, it is very difficult to obtain a
representative sample of a region because of the heterogeneous nature of many spaces.
1.6 Geographic Information System
Another characteristic of spatial data that impacts research is the general rule that spatial
data tend to become more heterogeneous as the study area gets larger. This means the variance
observed in a small region is less than that in a larger region. That is, expected variation around
the mean in small regions underestimates and, thus, is not applicable in subsequently larger
regions. This uncontrolled variance in spatial data has impacts in the study design of projects
involving large areas and times, such as global warming.
Researchers can quantify spatial correlation by using indices such as Moran‟s I and
Geary‟s C. Researchers can also model spatial dependence by using the methods developed in the
field of geostatistics. Techniques such as kriging allow researchers to quantify the changes in
attribute variance versus distance. This model of spatial dependence is called a semi-variogram, or
just variogram. Variograms offer a mechanism to predict attribute values in locations where data
are not present. These powerful interpolation tools have many real-world applications in fields
Fundamentals of GIS 1.7
such as mining and petroleum discovery, epidemiology, atmospheric science, oceanography, and
soil science.
The object-field dichotomy serves as the underpinning for all methods of spatial
representation and analysis, and each perspective offers a characteristic set of abilities and
limitations. For instance, objects offer logical concepts like “inside” and “outside,” and
comparisons such as intersections and buffers. Fields are particularly suited for spatial phenomena
considered less bounded, like air, water, temperature, and elevation. Not constrained by distinct
edges, fields lend themselves to analyses related to interpolation and global attribute comparison.
Analysis and representation of fields often yield areas of peaks, valleys, aspects, and slopes.
Figure 1.7: Examples of latitudinal lines are shown on the left and examples of longitudinal
lines are shown on the right. The 0° degree reference lines for each are shown in red (equator
for latitudinal measurements and prime meridian for longitudinal measurements).
A latitude measures the angle from the equatorial plane to the location on the earth‟s
surface. A longitude measures the angle between the prime meridian plane and the north-south
plane that intersects the location of interest. For example Colby College is located at around
45.56° North and 69.66° West. In a GIS system, the North-South and East-West directions are
encoded as signs. North and East are assigned a positive (+) sign and South and West are assigned
a negative (-) sign. Colby College‟s location is therefore encoded as +45.56° and -69.66°.
The reason the earth has a slightly ellipsoidal shape has to do with its rotation which
induces a centripetal force along the equator. This results in an equatorial axis that is roughly 21
km longer than the polar axis.
Our estimate of these radii is quite precise thanks to satellite and computational
capabilities. The semi-major axis is 6,378,137 meters and the semi-minor axis is 6,356,752
meters.
Differences in distance measurements along the surfaces of an ellipsoid vs. a sphere are
small but measurable (the difference can be as high as 20 km) as illustrated in the following lattice
plots.
1.3.2. Geoid
Representing the earth‟s true shape, the geoid, as a mathematical model is crucial for a
GIS environment. However, the earth‟s shape is not a perfectly smooth surface. It has undulations
resulting from changes in gravitational pull across its surface. These undulations may not be
visible with the naked eye, but they are measurable and can influence locational measurements.
Figure 1.10. Earth’s geoid with gravitational field shown in rainbow colors. The
ondulations depicted in the graphics are exaggerated for visual effects. (source: NASA)
1.10 Geographic Information System
Note that we are not including mountains and ocean bottoms in our discussion, instead we
are focusing solely on the earth‟s gravitational potential which can be best visualized by
imagining the earth‟s surface completely immersed in water and measuring the sea surface level
over the entire earth surface.
The earth‟s gravitational field is dynamic and is tied to the flow of the earth‟s hot and fluid
core. Hence its geoid is constantly changing, albeit at a large temporal scale.The measurement and
representation of the earth‟s shape is at the heart of geodesy–a branch of applied mathematics.
1.3.3. Datum
So how are we to reconcile our need to work with a (simple) mathematical model of the
earth‟s shape with the on dulating nature of the earth‟s surface (i.e. its geoid). The solution is to
align the geoid with the ellipsoid (or sphere) representation of the earth and to map the earth‟s
surface features onto this ellipsoid/sphere. The alignment can be local where the ellipsoid surface
is closely fit to the geoid at a particular location on the earth‟s surface (such as the state of
Kansas) or geocentric where the ellipsoid is aligned with the center of the earth. How one chooses
to align the ellipsoid to the geoid defines a datum.
Local Datum
There are many local datums to choose from, some are old while others are more recently
defined. The choice of datum is largely driven by the location of interest. For example, when
working in the US, a popular local datum to choose from is the North American Datum of 1927
(or NAD27 for short). NAD27 works well for the US but it‟s not well suited for other parts of the
world. For example, a far better local datum for Europe is the European Datum of 1950 (ED50 for
short). Examples of common local datums are shown in the following table:
Geocentric Datum
Many of the modern datums use a geocentric alignment. These include the popular World
Geodetic Survey for 1984 (WGS84) and the North American Datums of 1983 (NAD83). Most of
the popular geocentric datums use the WGS84 ellipsoid or the GRS80 ellipsoid. These ellipsoids‟
semi-major and semi-minor axes are nearly identical: 6,378,137 meters and 6,356,752 meters
respectively. Examples of popular geocentric datums are shown in the following table:
Possibly the earliest use of the geographic method, in 1854 John Snow depicted a cholera
outbreak in London using points to represent the locations of some individual cases. His study of
the distribution of cholera led to the source of the disease, a contaminated water pump within the
heart of the cholera outbreak. While the basic elements of topology and theme existed previously
in cartography, the John Snow map was unique, using cartographic methods, not only to depict
but also to analyze, clusters of geographically dependent phenomena for the first time.
1.12 Geographic Information System
The early 20th century saw the development of "photo lithography" where maps were
separated into layers. Computer hardware development spurred by nuclear weapon research led to
general-purpose computer "mapping" applications by the early 1960s. In the year 1962, the
world's first true operational GIS was developed by the federal Department of Forestry and Rural
Development in Ottawa, Canada by Dr. Roger Tomlinson. It was called the "Canada Geographic
Information System" (CGIS) and was used to store, analyze, and manipulate data collected for the
Canada Land Inventory (CLI). It is an initiative to determine the land capability for rural Canada
by mapping information about soils, agriculture, recreation, wildlife, forestry, and land use at a
scale of 1:50,000.
CGIS was the world's first "system" and was an improvement over "mapping" applications
as it provided capabilities for overlay, measurement, and digitizing or scanning. It supported a
national coordinate system that spanned the continent, coded lines as "arcs" having a true
embedded topology, and it stored the attribute and location specific information in a separate files.
Dr. Tomlinson is known as the "father of GIS," for his use of overlays in promoting the spatial
analysis of convergent geographic data.
In 1964, Howard T Fisher formed the Laboratory for Computer Graphics and Spatial
Analysis at the Harvard Graduate School of Design, where a number of important theoretical
concepts in spatial data handling were developed. This lab had major influence on the
development of GIS until early 1980s. Many pioneers of newer GIS "grew up" at the Harvard lab
and had distributed seminal software code and systems, such as 'SYMAP', 'GRID', and
'ODYSSEY'.
By the early 1980s, M&S Computing (later Intergraph), Environmental Systems Research
Institute (ESRI) and CARIS emerged as commercial vendors of GIS software, successfully
incorporating many of the CGIS features, combining the first generation approach to separation of
spatial and attribute information with a second generation approach to organizing attribute data
into database structures. More functions for user interaction were developed mainly in a graphical
way by a user friendly interface (Graphical User Interface), which gave to the user the ability to
sort, select, extract, reclassify, reproject and display data on the basis of complex geographical,
topological and statistical criteria. During the same time, the development of a public domain GIS
begun by the U.S. Army Corp of Engineering Research Laboratory (USA-CERL) in Champaign,
Illinois, a branch of the U.S. Army Corps of Engineers to meet the need of the United States
military for software for land management and environmental planning.
In the years 1980s and 1990s industry growth were spurred on by the growing use of
GIS on Unix workstations and the personal computers. By the end of the 20th century, the
rapid growth in various systems had been consolidated and standardized on relatively few
platforms and users were beginning to export the concept of viewing GIS4 data over the
Internet, requiring uniform data format and transfer standards. More recently, there is a
growing number of free, open source GIS packages, which run on a range of operating
systems and can be customized to perform specific tasks. As computing power increased and
hardware prices slashed down, the GIS became a viable technology for state development
planning. It has become a real Management Information System (MIS), and thus able to
support decision making processes.
Fundamentals of GIS 1.13
1.7. COMPONENTS OF A GIS
A working GIS integrates five key components: -
i) Hardware ii) Software iii) Data iv) People v) Methods
GIS enables the user to input, manage, manipulate, analyze, and display geographically
referenced data using a computerized system. To perform various operations with GIS, the
components of GIS such as software, hardware, data, people and methods are essential.
1.7.1. Software
GIS software provides the functions and tools needed to store, analyze, and display
geographic information. Key software components are (a) a database management system
(DBMS) (b) tools for the input and manipulation of geographic information (c) tools that support
geographic query, analysis, and visualization (d) a graphical user interface (GUI) for easy access
to tools. GIS software are either commercial software or software developed on Open Source
domain, which are available for free. However, the commercial software is copyright protected,
can be expensive and is available in terms number of licensees.
1.7.2. Hardware
Hardware is the computer on which a GIS operates. Today, GIS runs on a wide range of
hardware types, from centralized computer servers to desktop computers used in stand-alone or
networked configurations. Minimum configuration required to Arc/Info Desktop 9.0 GIS
application is as follows:
Hardware Requirements
CPU Speed: 800 MHz minimum, 1.0 GHz recommended or higher
Processor: Pentium or higher
Memory/RAM: 256 MB minumum, 512 MB recommended or higher
Display Properties: Greater than 256 color depth
Swap Space: 300 MB minimum
Disk Space: Typical 605 MB NTFS, Complete 695 MB FAT32 + 50 MB forinstallation
1.14 Geographic Information System
Browser: Internet Explorer 6.0 Requirement:
(Some features of ArcInfo Desktop 9.0 require a minimum installation of Microsoft
Internet Explorer Version 6.0.)
1.7.3. Data
The most important component of a GIS is the data. Geographic data or Spatial data and
related tabular data can be collected in-house or bought from a commercial data provider. Spatial
data can be in the form of a map/remotely-sensed data such as satellite imagery and aerial
photography. These data forms must be properly geo-referenced (latitude/longitude). Tabular data
can be in the form attribute data that is in some way related to spatial data. Most GIS software
comes with inbuilt Database Management Systems (DBMS) to create and maintain a database to
help organize and manage data.
1.7.4. Users
GIS technology is of limited value without the users who manage the system and to
develop plans for applying it. GIS users range from technical specialists who design and maintain
the system to those who use it to help them do their everyday work.
These users are largely interested in the results of the analyses and may have no interest or
knowledge of the methods of analysis. The user-friendly interface of the GIS software allows the
nontechnical users to have easy access to GIS analytical capabilities without needing to know
detailed software commands. A simple User Interface (UI) can consist of menus and pull-down
graphic windows so that the user can perform required analysis with a few key presses without
needing to learn specific commands in detail.
1.7.5. People
GIS technology has limited value without the people who manage and develop plans for
applying it to real world problems. GIS user range from technical specialists who design and
maintain the system to those who use it to help them perform their everyday work. The
identification of GIS specialist's vs. end users is often critical to the proper implementation of GIS
technology. This is what called 'brain ware' which is equally important as the Hardware and
software. Brain ware refers to the purpose and objectives, and provides the reason and
justification, for using GIS.
This component of GIS includes all those individuals (such as the programmer, database
manager, GIS researcher etc.) who are making the GIS work, and also the individuals who are at
the user end using the GIS services, applications and tools.
1.7.6. Methods
A successful GIS operates according to a well-designed plan and business rules, which are
the models and operating practices unique to each organization.
Fundamentals of GIS 1.15
1.8. PROPRIETARY AND OPEN SOURCE GIS SOFTWARE
1.8.1. AGISMap
AGIS for Windows is a mapping and simple GIS package specifically designed to be easy
to use, and distributed as shareware via the world wide web.
Platforms: Windows
1.8.2. Autodesk
Autodesk has a series of software applications designed to meet GIS needs in a variety of
areas that interfaces with their CAD software.
Platforms: Windows
Platforms: Windows
1.8.4. Cartographica
Commercial software package for Mac OS featuring: Support for a huge number of import
formats, including popular Raster formats, manual editing and georeferencing, automatic
geocoding, integration with online mapping, output to large-format printers. Thirty-day demo
available for download.
Platforms: Macintosh
1.8.5. DeLorme
DeLorme is the producers of XMap, a GIS application “with 80% of the functionality
found in a traditional GIS at 15% of the cost”. Performs such functions as geocoding, image
rectification, 3D visualization and coordinate transformation.
Platforms: Windows
1.8.6. Esri
Environmental Systems Research Institue has been creating GIS software for over 30
years. Recognized as the leader in GIS software, it‟s been estimated that about seventy percent of
1.16 Geographic Information System
GIS users use Esri products. Esri overhauled their software packages into an interoperable model
called ArcGIS (the desktop GIS is referred to as ArcMap). In addition, Esri has developed plug-
ins called extensions which add to the functionality of ArcGIS. Demo and light versions of Esri
software are available for downloading. You can also find free data to use with Esri products.
Platforms: Windows
Further Resources: ArcGIS, ArcView 3.x (no longer in production)
1.8.7. Intergraph
Intergraph makes several GIS applications. Most of the GIS packages are designed with an
Open GIS in mind and therefore can work with a variety of other GIS software formats.
Intergraph has developed products that help merge GIS with information technology (IT) and
business process improvement tools. Intergraph offers the Geo-Media family of solutions and
Modular GIS Environment MGE Suite of mapping and GIS applications.
Using an open architecture, the Geo-Media product suite integrates geospatial information
throughout the enterprise and provides the tools needed to develop business-to-business and
custom client applications using industry standard development tools. Geo-Media offers
uninhibited access to all geospatial data formats without the need for data translations. Currently
in Version 4.0 the Geo-Media family is made up of Geo-Media, Geo-Media Professional, Geo
Media Web Map, and Geo Media Web Enterprise.
• GeoMedia is the universal information integrator, serving as a visualization and
analysis tool and as an open platform for custom GIS solution development.
• GeoMedia Professional is a product specifically designed to collect and manage
spatial data using standard databases.
• GeoMedia WebMap is a Web-based map visualization tool with real-time links to one
or more GIS data warehouses.
• GeoMedia WebEnterprise creates dynamic, custom web-mapping applications that can
analyze and manipulate geographic data.
• In addition to these products, Intergraph offers MFworks for GeoMedia which
provides users of grid-based software the power of visualization, mapping, and
analysis. Intergraph also offers SMMS for GeoMedia which is a desktop tool for
geographic metadata creation and geographic data management.
The Modular GIS Environment (MGE) product suite provides production-ready
capabilities for automating, managing, analyzing, and presenting GIS data, and is completely
interoperable with GeoMedia.
1.8.8. Manifold
Manifold System provides comprehensive, professional grade GIS software for $245 that
includes a very wide array of features. Manifold imports data from over 80 different GIS formats,
including all formats used by Federal government sites for free Internet downloads, and Manifold
allows seamless, simultaneous work with vector drawings, raster images, terrain elevations and
raster data sets either as 2D displays or 3D terrain visualizations. Manifold includes exceptional
DBMS capabilities, full development facilities and includes a built-in Internet Map Server for fast
Fundamentals of GIS 1.17
and easy publication of GIS projects to the web without programming. Options include US Street
address geocoding and the Enterprise Edition, for centralized geospatial data storage on enterprise
servers that can be used by many GIS operators at once. Manifold was the first GIS to attain
“designed for XP” status with Microsoft and the Manifold Internet Map Server works perfectly
within ASP.NET servers. For info, see http://www.manifold.net/professional.
Platforms: Windows
1.8.9. Ortelius
Ortelius is a “map illustration” software package that adds one more selection to the very
limited mapping software options out there for Macintosh users. A free trial download is
available.
Platforms: Macintosh
1.8.11 Maptitude
The Maptitude Mapping Software is a full-featured mapping package for Windows.
Designed for ease-of-use, data visualization and geographic analysis, Maptitude comes with
comprehensive nationwide and worldwide maps, including complete US street maps, and Census
tract and ZIP Code boundaries and demographics. Caliper also produces TransCAD for
transportation and logistics. TransCAD is used for solving key analytical problems in
transportation planning, management, and operations. TransCAD is used extensively for
transportation database development and maintenance, demand forecasting, operations
management, and vehicle routing and scheduling.
Platforms: Windows OS
1.8.12. MyWorld
My World GIS is a full-featured GIS designed for educational use. My World provides a
carefully selected subset of the features of a professional GIS environment. These features include
multiple geographic projections, table and map views of data, distance-measurement tools,
buffering and query operations, customizable map display. They have been selected to provide the
greatest value to students without overwhelming them with complexity The features are accessed
through a supportive interface designed with the needs of students and teachers in mind. My
1.18 Geographic Information System
World can import data from the industry-standard shapefile format, as well as from tab and
comma-delimited text files.
Platforms: Windows OS
Platforms: Windows OS
Platforms: Windows
Platforms: Windows
Fundamentals of GIS 1.19
1.8.17. TNT Products
Created by Microimages, The TNT Products is a suite of GIS applications for fully
integrated GIS, image processing, CAD, TIN, desktop cartography, and geospatial database
management.
(ii) QGIS and DIVA-GIS are more widely used than the others. Both have MAC
versions available for downloading.
1.9.2. DIVA-GIS
DIVA-GIS are a free computer program for mapping and geographic data analysis (a
geographic information system (GIS).
1.9.4. GeoDa
GeoDa is a free GIS software program primarily used to introduce new users into spatial
data analysis. It‟s main functionality is data exploration in statistics.
One of the nicest things about it is how it comes with sample data for you to give a test-
drive. From simple box-plots all the way to regression statistics, GeoDa has complete arsenal of
statistics to do nearly anything spatially.
It‟s user base is strong. For example, Harvard, MIT and Cornell universities have
embraced this free GIS software to serve as a gentle introduction to spatial analysis for non-GIS
users. From economic development to health and real estate, it‟s been used as an exciting
analytical in labs as well.
1.20 Geographic Information System
From the Center for Spatial Data Science at the University of Chicago, software for
geospatial analysis, geovisualization and other techniques.
1.9.5. gvSIG
In 2004, the gvSIG project emerged as a free, open source GIS software option in Spain.
We illustrate in this gvSIG guide and review why we like it SO much:
gvSIG really outperforms QGIS 2 for 3D. It really is the best 3D visualization available in
open source GIS.
The NavTable is agile in that it allows you to see records one-by-one vertically.
The CAD tools are impressive on gvSIG. Thanks to the OpenCAD Tools, you can trace
geometries, edit vertices, snap and split lines and polygons. If you need GIS on your mobile
phone, gvSIG Mobile is perfect for field work because of its interface and GPS tools.
1.9.8. uDig
uDIG is an acronym to help get a better understanding what this Free GIS software is all
about.
• u stands for user-friendly interface
• D stands for desktop (Windows, Mac or Linux). You can run uDIG on a Mac.
• I stand for internet oriented consuming standard (WMS, WFS or WPS)
• G stands for GIS-ready for complex analytical capabilities.
When you start digging into uDig, it‟s a nice open source GIS software option for basic
mapping. uDig‟s Mapnik lets you import basemaps with the same tune as ArcGIS
Specifically, its easy-to-use, the catalog, symbology and Mac OS functionality are some of
the strong points. But it has limited tools and the bugs bog it down to really utilize it as a truly
complete free GIS software package.
Spatial data: describes the absolute and relative location of geographic features.
Attribute data: describes characteristics of the spatial features. These characteristics can
be quantitative and/or qualitative in nature. Attribute data is often referred to as tabular data.
Attribute Data:
The attributes refer to the properties of spatial entities. They are often referred to as non-
spatial data since they do not in themselves represent location information. This type of data
describes characteristics of the spatial features. These characteristics can be quantitative and/or
qualitative in nature. Attribute data is often referred to as tabular data.
Spatial Data:
Geographic position refers to the fact that each feature has a location that must be
specified in a unique way. To specify the position in an absolute way a coordinate system is used.
For small areas, the simplest coordinate system is the regular square grid. For larger areas, certain
approved cartographic projections are commonly used. Internationally there are many different
coordinate systems in use. This locational information is provided in maps by using Points, Lines
and Polygons. These geometric descriptions are the basic data elements of a map. Thus spatial
data describes the absolute and relative location of geographic features.
The coordinate location of a forest would be spatial data, while the characteristics of that
forest, e.g. cover group, dominant species, crown closure, height, etc., would be attribute data.
Other data types, in particular image and multimedia data, have become more prevalent with
changing technology. Depending on the specific content of the data, image data may be
considered either spatial, e.g. photographs, animation, movies, etc., or attribute, e.g. sound,
descriptions, narration‟s, etc.
Spatial relationship: The connections between spatial objects are described as spatial
relationships (e.g. A contains B; A is adjacent to B, A is North of B, etc.).
1.22 Geographic Information System
Attributes: Attributes capture the thematic mode by defining different characteristics of
objects. Spatial features in the real world are reduced in the form of point, line, area and surface.
GIS will store the data either in tabular form, geographical map, digital map or remotely sensed
map.
The following diagram reflects the two primary spatial data encoding techniques. These
are vector and raster. Image data utilizes techniques very similar to raster data, however typically
lacks the internal formats required for analysis and modeling of the data. Images reflect pictures or
photographs of the landscape.
Fundamentals of GIS 1.23
1.10.4. Attribute Data Models
A separate data model is used to store and maintain attribute data for GIS software. These
data models may exist internally within the GIS software, or may be reflected in external
commercial Database Management Software (DBMS). A variety of different data models exist for
the storage and management of attribute data. The most common are:
• Tabular
• Hierarchical
• Network
• Relational
• Object oriented
Tabular Model
The simple tabular model stores attribute data as sequential data files with fixed formats
(or comma delimited for ASCII data), for the location of attribute values in a predefined record
structure. This type of data model is outdated in the GIS arena. It lacks any method of checking
data integrity, as well as being inefficient with respect to data storage, e.g. limited indexing
capability for attributes or records, etc.
Hierarchical Model
The hierarchical database organizes data in a tree structure. Data is structured downward
in a hierarchy of tables. Any level in the hierarchy can have unlimited children, but any child can
have only one parent. Hierarchical DBMS have not gained any noticeable acceptance for use
within GIS.
They are oriented for data sets that are very stable, where primary relationships among the
data change infrequently or never at all. Also, the limitation on the number of parents that an
element may have is not always conducive to actual geographic phenomenon.
Network Model
The network database organizes data in a network or plex structure. Any column in a plex
structure can be linked to any other. Like a tree structure, a plex structure can be described in
terms of parents and children.
This model allows for children to have more than one parent. Network DBMS have not
found much more acceptance in GIS than the hierarchical DBMS. They have the same flexibility
limitations as hierarchical databases; however, the more powerful structure for representing data
relationships allows a more realistic modeling of geographic phenomenon. However, network
databases tend to become overly complex too easily. In this regard it is easy to lose control and
understanding of the relationships between elements.
Relational Model
The relational database organizes data in tables. Each table, is identified by a unique table
name, and is organized by rows and columns. Each column within a table also has a unique name.
1.24 Geographic Information System
Columns store the values for a specific attribute, e.g. cover group, tree height. Rows represent one
record in the table.
In a GIS each row is usually linked to a separate spatial feature, e.g. a forestry stand.
Accordingly, each row would be comprised of several columns, each column containing a specific
value for that geographic feature.
Object-Oriented Model
The object-oriented database model manages data through objects. An object is a
collection of data elements and operations that together are considered a single entity. The object-
oriented database is a relatively new model.
This approach has the attraction that querying is very natural, as features can be bundled
together with attributes at the database administrator's discretion. To date, only a few GIS
packages are promoting the use of this attribute data model.
However, initial impressions indicate that this approach may hold many operational
benefits with respect to geographic data processing. Fulfillment of this promise with a commercial
GIS product remains to be seen.
Aside from descriptors, character fields can contain other attribute values such as
categories and ranks. For example, a character field may contain the categories for a street:
avenue, boulevard, lane, or highway. A character field could also contain the rank, which is a
relative ordering of features. For example, a ranking of the traffic load of the street with “1” being
the street with the highest traffic.
Because character data is not numeric, calculations (sum, average, median, etc.) can‟t be
performed on this type of field, even if the value stored in the field are numbers (to do that, the
field type would need to be converted to a numeric field). Character fields can be summarized to
produced counts (e.g. the number of features that have been categorized as “avenue”).
Fundamentals of GIS 1.25
1.11.2. Numeric Data
Integer and floating are numerical values (see: the difference between floating and integer
values). Within the integer type, the is a further division between short and long integer values.
As would be expected, short integers store numeric values without fractional values for a shorter
range than long integers. Floating point attribute values store numeric values with fractional
values. Therefore, floating point values are for numeric values with decimal points (i.e numbers to
the right of the decimal point as opposed to whole values).
Nominal Data:
Nominal data describe different kinds of different categories of data such as land use types
or soil types.
Ordinal Data:
Ordinal data differentiate data by ranking relationship. For example-cities may be grouped
into large, medium and small cities by population size.
Interval Data:
Interval data have known intervals between values such as temperature reading. For
example- a temperature reading of 700 F is warmer than 600 F by 100 F.
Ratio Data:
Ratio data are the same as interval data except that ratio data are based on a meaningful or
absolute zero value. Population densities are an example of ratio data, because a density of 0 is an
absolute zero.
Measurement scale (or level) of attribute can be broadly divided into two categories. Some
attribute are measured in a numerical scale (such as job accessibility) whereas others are not (such
as world language)
Level of measurement prescribe the information required for an attribute reference system
Level of measurement Information required
Nominal Definitions of categories
Ordinal Definitions of categories plus ordering
Interval Unit of measure plus zero point
Ratio Unit of measure
Fundamentals of GIS 1.29
TWO MARK QUESTIONS WITH ANWERS
1) What is GIS?
It‟s a computer based information system primarily aims in collecting, classifying,
crosschecking, manipulating, interpreting, retrieving and displaying data which are spatially
referred to the earth in an appealing way.
These x and y coordinates are based on different projection system and there are various
types of projection system. Most of the time GIS is used to create maps and to print. To
perform the basic task in GIS, layers are combined, edited and designed.
QGIS supports shape files, coverage‟s, personal geo-databases, dxf, MapInfo, Post-GIS, and
other formats. Web services, including Web Map Service and Web Feature Service, are also
supported to allow use of data from external sources.
QGIS integrates with other open-source GIS packages, including Post-GIS, GRASS GIS, and
Map-Server. Plug-ins written in Python or C++ extends QGIS's capabilities. Plug-ins can
geo-code using the Google Geo-coding API, perform geo-processing functions similar to
those of the standard tools found in Arc-GIS, and interface with PostgreSQL / Post-GIS,
SpatiaLite and MySQL databases.
1) Describe in detail, the Basic spatial concepts. Add a note on introduction to GIS.
2) Write an essay on Geographic Coordinate Systems.
3) Discuss the software and hardware components of GIS (May/June - 2012)
4) Write detailed notes on proprietary and open source GIS Software
5) Explain in detail about Open Source GIS Software.
6) What is data model? Add notes on different types of data.
7) Write an essay on attribute data model. (May/June - 2014)
8) Write an essay on scale of measurement/Level of measurements
CHAPTER- 2
2.1. INTRODUCTION
Spatial data are what drive a GIS. Every functionality that makes a GIS separate from
another analytical environment is rooted in the spatially explicit nature of the data.
Spatial data are often referred to as layers, coverage’s, or layers. We will use the term
layers from this point on, since this is the recognized term used in Arc-GIS. Layers represent, in a
special digital storage format, features on, above, or below the surface of the earth. Depending on
the type of features they represent, and the purpose to which the data will be applied, layers will
be one of two major types.
a) Vector data represent features as discrete points, lines, and polygons.
b) Raster data represent the landscape as a rectangular matrix of square cells.
Depending on the type of problem that needs to be solved, the type of maps that need to be
made, and the data source, either raster or vector, or a combination of the two can be used. Each
data model has strengths and weaknesses in terms of functionality and representation. As you get
more experience with GIS, you will be able to determine which data type to use for a particular
application.
Vector, e.g,
• ArcInfo Coverages
• ArcGIS Shape Files
• CAD (AutoCAD DXF & DWG, or Micro Station DGN files)
• ASCII coordinate data
Raster, e.g,
• ArcInfo Grids
• Images
2.2 Geographic Information System
• Digital Elevation Models (DEMs)
• generic raster datasets
Each pointer establishes a parent-child relationship where a parent can have more than one
child but a child can only have one parent. There is no connection between the elements at the
same level. To locate a particular record, you have to start at the top of the tree with a parent
record and trace down the tree to the child.
Advantages
• Easy to understand: The organization of database parallels a family tree understanding
which is quite easy.
Spatial Data Models 2.3
• Accessing records or updating records are very fast since the relationships have been
predefined.
Disadvantages
• Large index files are to be maintained and certain attribute values are repeated many
times which lead to data redundancy and increased storage.
• The rigid structure of this model doesn’t allow alteration of tables, therefore to add a
new relationship entire database is to be redefined.
Advantages
• The many too many relationships are easily implemented in a network data model.
• Data access and flexibility in network model is better than that in hierarchical model.
An application can access an owner record and the member records within a set.
• It enforces data integrity as a user must first define owner record and then the member
records.
• The model eliminated redundancy but at the expense of more complicated
relationships.
2.4 Geographic Information System
2.3.3. Relational Data Structure Model
• The relational data model was introduced by Codd in 1970. The relational database
relates or connects data in different files through the use of a common field.
• A flat file structure is used with a relational database model. In this arrangement, data
is stored in different tables made up of rows and columns as shown in figure.2.3.
• The columns of a table are named by attributes. Each row in the table is called a tuple
and represents a basic fact.
• No two rows of the same table may have identical values in all columns.
Advantages
• The manager or administrator does not have to be aware of any data structure or data
pointer. One can easily add, update, delete or create records using simple logic.
Disadvantages
• A few search commands in a relational database require more time to process
compared with other database models.
• For example, let us try to represent a thought: “Hawaii is an island that is a state of
USA” in GIS. In this case, we don’t mind the geographic location with latitude and
longitude in the conventional GIS model. This is not appropriate to use the layers. In
an object-oriented model, we are more careful with spatial relationships for example,
“is a” (the island is a land) and “part of” (the state is a part of the country).
• In addition, Hawaii (state) has Honolulu City and also is in Pacific Region. Figure 2.4
(a) shows “is an” inheritance for the super class of land, while Figure 2.4 (b) shows
the spatial relationships for the object of the state.
An entity is a real-world item or concept that exists on its own. Entities are equivalent to
database tables in a relational database, with each row of the table representing an instance of that
entity.
Spatial Data Models 2.7
An attribute of an entity is a particular property that describes the entity. A relationship is
the association that describes the interaction between entities. Cardinality, in the context of ERD,
is the number of instances of one entity that can, or must, be associated with each instance of
another entity. In general, there may be one-to-one, one-to-many, or many-to-many relationships.
For example, let us consider two real-world entities, an employee and his department. An
employee has attributes such as an employee number, name, department number, etc. Similarly,
department number and name can be defined as attributes of a department. A department can
interact with many employees, but an employee can belong to only one department, hence there
can be a one-to-many relationship, defined between department and employee.
In the actual database, the employee table will have department number as a foreign key,
referencing from department table, to enforce the relationship.
You could use this database to query how many of a particular species of game fish were
examined at a specific park during a data range of interest. This would be non-spatial query
because we are just counting an occurrence at one particular location. We are not using the
coordinates to perform some type of buffer analysis or other spatial analysis to query the data.
The following diagram reflects the two primary spatial data encoding techniques. These
are vector and raster. Image data utilizes techniques very similar to raster data, however typically
lacks the internal formats required for analysis and modeling of the data. Images reflect pictures or
photographs of the landscape.
Vector lines are often referred to as arcs and consist of a string of vertices terminated by a
node. A node is defined as a vertex that starts or ends an arc segment. Point features are defined
by one coordinate pair, a vertex. Polygonal features are defined by a set of closed coordinate pairs.
In vector representation, the storage of the vertices for each feature is important, as well as the
connectivity between features, e.g. the sharing of common vertices where features connect.
Several different vector data models exist, however only two are commonly used in GIS
data storage.
2.10 Geographic Information System
The most popular method of retaining spatial relationships among features is to explicitly
record adjacency information in what is known as the topologic data model. Topology is a
mathematical concept that has its basis in the principles of feature adjacency and connectivity.
The topologic data structure is often referred to as an intelligent data structure because
spatial relationships between geographic features are easily derived when using them. Primarily
for this reason the topologic model is the dominant vector data structure currently used in GIS
technology. Many of the complex data analysis functions cannot effectively be undertaken
without a topologic vector data structure. Topology is reviewed in greater detail later on in the
book.
The secondary vector data structure that is common among GIS software is the computer-
aided drafting (CAD) data structure. This structure consists of listing elements, not features,
defined by strings of vertices, to define geographic features, e.g. points, lines, or areas. There is
considerable redundancy with this data model since the boundary segment between two polygons
can be stored twice, once for each feature. The CAD structure emerged from the development of
computer graphics systems without specific considerations of processing geographic features.
Accordingly, since features, e.g. polygons, are self-contained and independent, questions about
the adjacency of features can be difficult to answer. The CAD vector model lacks the definition of
spatial relationships between features that is defined by the topologic data model.
The size of cells in a tessellated data structure is selected on the basis of the data accuracy
and the resolution needed by the user. There is no explicit coding of geographic coordinates
required since that is implicit in the layout of the cells. A raster data structure is in fact a matrix
where any coordinate can be quickly calculated if the origin point is known, and the size of the
grid cells is known. Since grid-cells can be handled as two-dimensional arrays in computer
encoding many analytical operations are easy to program. This makes tessellated data structures a
popular choice for many GIS software. Topology is not a relevant concept with tessellated
Spatial Data Models 2.11
structures since adjacency and connectivity are implicit in the location of a particular cell in the
data matrix.
Several tessellated data structures exist, however only two are commonly used in GIS's.
The most popular cell structure is the regularly spaced matrix or raster structure. This data
structure involves a division of spatial data into regularly spaced cells. Each cell is of the same
shape and size. Squares are most commonly utilized.
Since geographic data is rarely distinguished by regularly spaced shapes, cells must be
classified as to the most common attribute for the cell. The problem of determining the proper
resolution for a particular data layer can be a concern. If one selects too coarse a cell size then data
may be overly generalized. If one selects too fine a cell size then too many cells may be created
resulting in a large data volume, slower processing times, and a more cumbersome data set. As
well, one can imply accuracy greater than that of the original data capture process and this may
result in some erroneous results during analysis.
As well, since most data is captured in a vector format, e.g. digitizing, data must be
converted to the raster data structure. This is called vector-raster conversion. Most GIS software
allows the user to define the raster grid (cell) size for vector-raster conversion. It is imperative that
the original scale, e.g. accuracy, of the data be known prior to conversion. The accuracy of the
data, often referred to as the resolution, should determine the cell size of the output raster map
during conversion.
Most raster based GIS software requires that the raster cell contain only a single discrete
value. Accordingly, a data layer, e.g. forest inventory stands, may be broken down into a series of
raster maps, each representing an attribute type, e.g. a species map, a height map, a density map,
etc. These are often referred to as one attribute maps. This is in contrast to most conventional
vector data models that maintain data as multiple attribute maps, e.g. forest inventory polygons
linked to a database table containing all attributes as columns. This basic distinction of raster data
storage provides the foundation for quantitative analysis techniques. This is often referred to as
raster or map algebra. The use of raster data structures allow for sophisticated mathematical
modelling processes while vector based systems are often constrained by the capabilities and
language of a relational DBMS.
The selection of a particular data model, vector or raster, is dependent on the source and
type of data, as well as the intended use of the data. Certain analytical procedures require raster
data while others are better suited to vector data.
Fig.2.10. Image data is most often used for remotely sensed imagery
such as satellite imagery or digital orthophotos.
Spatial Data Models 2.13
2.5.4. Vector and Raster – Advantages and Disadvantages
There are several advantages and disadvantages for using either the vector or raster data
model to store spatial data. These are summarized below.
Vector Data:
Advantages:
• Data can be represented at its original resolution and form without generalization.
• Graphic output is usually more aesthetically pleasing (traditional cartographic
representation);
• Since most data, e.g. hard copy maps, is in vector form no data conversion is required.
• Accurate geographic location of data is maintained.
• Allows for efficient encoding of topology, and as a result more efficient operations
that require topological information, e.g. proximity, network analysis.
Disadvantages:
• The location of each vertex needs to be stored explicitly. For effective analysis, vector
data must be converted into a topological structure. This is often processing intensive
and usually requires extensive data cleaning. As well, topology is static, and any
updating or editing of the vector data requires re-building of the topology. Algorithms
for manipulative and analysis functions are complex and may be processing intensive.
Often, this inherently limits the functionality for large data sets, e.g. a large number of
features.
• Continuous data, such as elevation data, is not effectively represented in vector form.
Usually substantial data generalization or interpolation is required for these data
layers.
• Spatial analysis and filtering within polygons is impossible
Raster Data
Advantages:
• The geographic location of each cell is implied by its position in the cell matrix.
Accordingly, other than an origin point, e.g. bottom left corner, no geographic
coordinates are stored.
• Due to the nature of the data storage technique data analysis is usually easy to program
and quick to perform.
• The inherent nature of raster maps, e.g. one attribute maps, is ideally suited for
mathematical modeling and quantitative analysis.
• Discrete data, e.g. forestry stands, is accommodated equally well as continuous data,
e.g. elevation data, and facilitates the integrating of the two data types.
• Grid-cell systems are very compatible with raster-based output devices, e.g.
electrostatic plotters, graphic terminals.
2.14 Geographic Information System
Disadvantages:
• The location of each vertex needs to be stored explicitly. For effective analysis, vector
data must be converted into a topological structure. This is often processing intensive
and usually requires extensive data cleaning. As well, topology is static, and any
updating or editing of the vector data requires re-building of the topology.
• Algorithms for manipulative and analysis functions are complex and may be
processing intensive. Often, this inherently limits the functionality for large data sets,
e.g. a large number of features.
• Continuous data; such as elevation data, is not effectively represented in vector form.
• Usually substantial data generalization or interpolation is required for these data
layers.
• Spatial analysis and filtering within polygons is impossible.
The huge size of the data is a major problem with raster data. An image consisting of
twenty different land-use classes takes the same storage space as a similar raster map showing the
location of a single forest. To address this problem many data compaction methods have been
developed which are discussed below:
2.6.4. Quadtree
• A raster is divided into a hierarchy of quadrants that are subdivided based on similar
value pixels.
• The division of the raster stops when a quadrant is made entirely from cells of the
same value.
• A quadrant that cannot be subdivided is called a leaf node.
A satellite or remote sensing image is a raster data where each cell has some value and
together these values create a layer. A raster may have a single layer or multiple layers. In a multi-
layer/ multi-band raster each layer is congruent with all other layers, have identical numbers of
rows and columns, and have same locations in the plane. Digital elevation model (DEM) is an
example of a single-band raster dataset each cell of which contains only one value representing
surface elevation.
Spatial Data Models 2.17
2.6.5. A single layer raster data can be represented using
(a) Two colors (binary):
The raster is represented as binary image with cell values as either 0 or 1 appearing black
and white respectively.
Gray-scale:
Typical remote sensing images are recorded in an 8 bit digital system. A grayscale image
is thus represented in 256 shades of gray which range from 0 (black) to 255 (white). However a
human eye can’t make distinction between the 255 different shades. It can only interpret 8 to 16
shades of gray.
A satellite image can have multiple bands, i.e. the scene/details are captured at different
wavelengths (Ultraviolet- visible- infrared portions) of the electromagnetic spectrum. While
creating a map we can choose to display a single band of data or form a color composite using
multiple bands. A combination of any three of the available bands can be used to create RGB
composites. These composites present a greater amount of information as compared to that
provided by a single band raster.
Compression ratio:
• The compression ratio (that is , the size of the compressed file compared to that of the
uncompressed file) of lossy video codec’s is nearly always far superior to that of the
audio and still-image equivalents Wavelet compression, used by raster formats such as
MrSID,JPEG2000,andER Map per’s ECW, takes time to decompress before drawing.
• Compression a series of techniques used for the reduction of space, bandwidth, cost,
transmission, generating time, and the storage of data.
• It’s a computer process using algorithms that reduces the size of electronic documents
so they occupy less digital storage space.
2.7.3. Quadtree
• Typical type of raster model is dividing area into equal-sized rectangular cells .
• However, many cases, variable sized grid cell size used for more compact raster
representation as shown figure.2.13.
• Larger cells used to represent large homogenous areas and smaller cells for finely
details.
• Process involves regularly subdividing a map into four equal sized quadrants.
Quadrant that has more than one class is again subdivided. Then; it is further
subdivided within each quadrant until a square is found to be so homogenous that it is
no longer needed to be divided.
• Then a Quadtree is prepared, resembling an inverted tree with “Root”, i.e., a point
from which all branches expand; Leaf is a lower most point and all other points in the
tree are nodes.
b) Topological features
A topology is a mathematical procedure that describes how features are spatially related
and ensures data quality of the spatial relationships. Topological relationships include following
three basic elements:
2.8.1. Connectivity
Arc node topology defines connectivity - arcs are connected to each other if they share a
common node. This is the basis for many network tracing and path finding operations.
Arcs represent linear features and the borders of area features. Every arc has a from-node
which is the first vertex in the arc and a to-node which is the last vertex. These two nodes define
the direction of the arc. Nodes indicate the endpoints and intersections of arcs. They do not exist
independently and therefore cannot be added or deleted except by adding and deleting arcs.
2.8.2. Contiguity
Polygon topology defines contiguity. The polygons are said to be contiguous if they share
a common arc. Contiguity allows the vector data model to determine adjacency.
Polygon A is outside the boundary of the area covered by polygons B, C and D. It is called
the external or universe polygon, and represents the world outside the study area. The universe
polygon ensures that each arc always has a left and right side defined.
2.8.3. Containment
Geographic features cover distinguishable area on the surface of the earth. An area is
represented by one or more boundaries defining a polygon.
The lake actually has two boundaries, one which defines its outer edge and the other
(island) which defines its inner edge. An island defines the inner boundary of a polygon. The
polygon D is made up of arc 5, 6 and 7. The 0 before the 7 indicates that the arc 7 creates an
island in the polygon.
Polygons are represented as an ordered list of arcs and not in terms of X, Y coordinates.
This is called Polygon-Arc topology. Since arcs define the boundary of polygon, arc coordinates
are stored only once, thereby reducing the amount of data and ensuring no overlap of boundaries
of the adjacent polygons.
Line entities:
Linear features made by tracing two or more XY coordinate pair.
• Simple line: It requires a start and an end point.
• Arc: A set of XY coordinate pairs describing a continuous complex line. The shorter
the line segment and the higher the number of coordinate pairs, the closer the chain
approximates a complex curve.
Simple Polygons:
Enclosed structures formed by joining set of XY coordinate pairs. The structure is simple
but it carries few disadvantages which are mentioned below:
• Lines between adjacent polygons must be digitized and stored twice, improper
digitization give rise to slivers and gaps
• Convey no information about neighbour
• Creating islands is not possible
Because points can be placed irregularly over a surface a TIN can have higher resolution
in areas where surface is highly variable. The model incorporates original sample points providing
a check on the accuracy of the model. The information related to TIN is stored in a file or a
database table. Calculation of elevation, slope, and aspect is easy with TIN but these are less
widely available than raster surface models and more time consuming in term of construction and
processing.
2.26 Geographic Information System
The TIN model is a vector data model which is stored using the relational attribute tables.
A TIN dataset contains three basic attribute tables: Arc attribute table that contains length, from
node and to node of all the edges of all the triangles.
• Node attribute table that contains x, y coordinates and z (elevation) of the vertices
• Polygon attribute table that contains the areas of the triangles, the identification
number of the edges and the identifier of the adjacent polygons.
Storing data in this manner eliminated redundancy as all the vertices and edges are stored
only once even if they are used for more than one triangle. As TIN stores topological
relationships, the datasets can be applied to vector based geo-processing such as automatic
contouring, 3D landscape visualization, volumetric design, surface characterization etc.
Data Model
The data model represents a set of guidelines to convert the real world (called entity) to
the digitally and logically represented spatial objects consisting of the attributes and geometry.
The attributes are managed by thematic or semantic structure while the geometry is represented by
geometric-topological structure.
Vector Data Model: [data models] A representation of the world using points, lines, and
polygons (show in the figure 2.21). Vector models are useful for storing data that has discrete
boundaries, such as country borders, land parcels, and streets.
Raster Data Model: [data models] A representation of the world as a surface divided into
a regular grid of cells. Raster models are useful for storing data that varies continuously, as in an
aerial photograph, a satellite image, a surface of chemical concentrations, or an elevation surface.
Spatial Data Models 2.27
Since the dawn of time, maps have been using symbols to represent real-world features. In
GIS terminology, real-world features are called spatial entities.
The cartographer decides how much data needs to be generalized in a map. This depends
on scale and how much detail will be displayed in the map. The decision to choose vector points,
lines or polygons is governed by the cartographer and scale of the map.
2.28 Geographic Information System
(1) Points
For Example: At a regional scale, city extents can be displayed as polygons because this
amount of detail can be seen when zoomed in. But at a global scale, cities can be represented
as points because the detail of city boundaries cannot be seen.
Vector data are stored as pairs of XY coordinates (latitude and longitude) represented as a
point. Complementary information like street name or date of construction could accompany it
in a table for its current use.
(2) Lines
Lines usually represent features that are linear in nature. Cartographers can use a different
thickness of line to show size of the feature. For Example, 500 meter Wide River may be
thicker than a 50 meter wide river. They can exist in the real-world such as roads or rivers. Or
they can also be artificial divisions such as regional borders or administrative boundaries.
Points are simply pairs of XY coordinates (latitude and longitude). When you connect
each point or vertex with a line in a particular order, they become a vector line feature.
Networks are line data sets but they are often considered to be different. This is because linear
networks are topologically connected elements. They consist of junctions and turns with
Spatial Data Models 2.29
connectivity. If you were to find an optimal route using a traffic line network, it would follow
one-way streets and turn restrictions to solve an analysis. Networks are just that smart.
(3) Polygons
Examples of polygons are buildings, agricultural fields and discrete administrative areas.
Cartographers use polygons when the map scale is large enough to be represented as polygons.
For example:
Each pixel value in a satellite image has a red, green and blue value. Alternatively, each
value in an elevation map represents a specific height. It could represent anything from
rainfall to land cover.
Raster models are useful for storing data that varies continuously. For example, elevation
surfaces, temperature and lead contamination.
In a discrete raster land cover/use map, you can distinguish each thematic class. Each class
can be discretely defined where it begins and ends. In other words, each land cover cell is
definable and it fills the entire area of the cell.
Discrete data usually consists of integers to represent classes. For example, the value 1
might represent urban areas; the value 2 represents forest and so on.
A continuous raster surface can be derived from a fixed registration point. For example,
digital elevation models use sea level as a registration point. Each cell represents a value above or
below sea level. As another example, aspect cell values have fixed directions such as north, east,
south or west.
Phenomena can gradually vary along a continuous raster from a specific source. In a raster
depicting an oil spill, it can show how the fluid moves from high concentration to low
concentration. At the source of the oil spill, concentration is higher and diffuses outwards with
diminishing values as a function of distance.
In the end, it really comes down to the way in which the cartographer conceptualizes the
feature in their map.
Spatial Data Models 2.31
• Do you want to work with pixels or coordinates? Raster data works with pixels.
Vector data consists of coordinates.
• What is your map scale? Vectors can scale objects up to the size of a billboard. But
you don’t get that type of flexibility with raster data
• Do you have restrictions for file size? Raster file size can result larger in comparison
with vector data sets with the same phenomenon and area.
Disadvantages:
• The location of each vertex needs to be stored explicitly. For effective analysis, vector
data must be converted into a topological structure. This is often processing intensive
and usually requires extensive data cleaning. As well, topology is static, and any
updating or editing of the vector data requires re-building of the topology.
• Algorithms for manipulative and analysis functions are complex and may be
processing intensive. Often, this inherently limits the functionality for large data sets,
e.g. a large number of features.
• Continuous data; such as elevation data, is not effectively represented in vector form.
• Usually substantial data generalization or interpolation is required for these data
layers.
• Spatial analysis and filtering within polygons is impossible.
A TIN is a vector based representation of the physical land surface or sea bottom, made up
of irregularly distributed nodes and lines with three dimensional coordinates (x,y, and z) that are
arranged in a network of non-overlapping triangles. TINs are often derived from the elevation data
of a rasterized digital elevation model (DEM).
Edges:
Every node is joined with its nearest neighbors by edges to form triangles, which satisfy
the Delaunay criterion. Each edge has two nodes, but a node may have two or more edges.
2.34 Geographic Information System
Because edges have a node with a z value at each end, it is possible to calculate a slope along the
edge from one node to the other.
TIN:
Advantages - ability to describe the surface at different level of resolution, efficiency in
storing data.
Disadvantages - in many cases require visual inspection and manual control of the
network.
The TIN creates triangles from a set of points called mass points, which always become
nodes. The user is not responsible for selecting; all the nodes are added according to a set of rules.
Mass points can be located anywhere, the more carefully selected, the more accurate the model of
the surface will be. Well-placed mass points occur when there is a major change in the shape of
the surface, for example, at the peak of a mountain, the floor of a valley, or at the edge (top and
bottom) of cliffs. By connecting points on a valley floor or along the edge of a cliff, a linear break
in the surface can be defined. These are called break lines. Break lines can control the shape of the
surface model.
They always form edges of triangles and, generally, cannot be moved. A triangle always
has three and only three straight sides, making their representation rather simple. A triangle is
assigned a unique identifier that defines by its three nodes and its two or three neighboring
triangles.
TIN is a vector-based topological data model that is used to represent terrain data. A TIN
represents the terrain surface as a set of interconnected triangular facets. For each of the three
vertices, the XY (geographic location) and the (elevation) Z values are encoded.
2.12. GRID/LUNR/MAGI
In this model each grid cell is referenced or addressed individually and is associated with
identically positioned grid cells in all other coverage’s, rather than like a vertical column of grid
cells, each dealing with a separate theme. Comparisons between coverage’s are therefore
performed on a single column at a time. Soil attributes in one coverage can be compared with
vegetation attributes in a second coverage. Each soil grid cell in one coverage can be compared
with a vegetation grid cell in the second coverage. The advantage of this data structure is that it
facilitates the multiple coverage analysis for single cells. However, this limits the examination of
spatial relationships between entire groups or themes in different coverage’s.
2.13.1. Standards
Most of the OGC standards depend on a generalized architecture captured in a set of
documents collectively called the Abstract Specification, which describes a basic data model for
representing geographic features. Atop the Abstract Specification members have developed and
continue to develop a growing number of specifications, or standards to serve specific needs for
interoperable location and geospatial technology, including GIS.
The OGC standards baseline comprises more than thirty standards, including:
Although the term "garbage in, garbage out" certainly applies to GIS data, there are other
important data quality issues besides the input data that need to be considered.
Position Accuracy
Position accuracy is the expected deviance in the geographical location of an object in the
data set (e.g. on a map) from its true ground position. Selecting a specified sample of points in a
prescribed manner and comparing the position coordinates with an independent and more accurate
Spatial Data Models 2.39
source of information usually test it. There are two components to position accuracy: the bias and
the precision.
Attribute Accuracy
Attributes may be discrete or continuous variables. A discrete variable can take on only a
finite number of values whereas a continuous variable can take on any number of values.
Categories like land use class, vegetation type, or administrative area are discrete variables. They
are, in effect, ordered categories where the order indicates the hierarchy of the attribute.
Logical Consistency
Logical consistency refers to how well logical relations among data elements are
maintained. It also refers to the fidelity of relationships encoded in the database, they may refer to
the geometric structure of the data model (e.g. topologic consistency) or to the encoded attribute
information e.g. semantic consistency).
(a) Completeness
Completeness refers to the exhaustiveness of the information in terms of spatial and
attribute properties encoded in the database. It may include information regarding feature
selection criteria, definition and mapping rules and the deviations from them. The tests on
spatial completeness may be obtained from topological test used for logical consistency
whereas the test for attribute completeness is done by comparison of a master list of geo-codes
to the codes actually appearing in the database.
There are several aspects to completeness as it pertains to data quality. They are grouped
here into three categories: completeness of coverage, classification and verification.
The completeness of coverage is the proportion of data available for the area of interest.
Example:
Demographic information is usually very time sensitive. It can change significantly
over a year. Land cover will change quickly in an area of rapid urbanization.
(c) Lineage
The lineage of a data set is its history, the source data and processing steps used to
produce it. The source data may include transaction records, field notes etc. Ideally, some
indication of lineage should be included with the data set since the internal documents are
rarely available and usually require considerable expertise to evaluate. Unfortunately, lineage
information most often exists as the personal experience of a few staff members and is not
readily available to most users.
Accessibility refers to the ease of obtaining and using the data. The accessibility of a data
set may be restricted because the data are privately held. Access to government-held information
may be restricted for reasons of national security or to protect citizen rights. Census data are
usually restricted in this way. Even when the right to use restricted data can be obtained, the time
and effort needed to actually receive the information may reduce its overall suitability.
The direct cost of a data set purchased from another organization is usually well known: it
is the price paid for the data. However, when the data are generated within the organization, the
true cost may be unknown. Assessing the true cost of these data is usually difficult because the
services and equipment used in their production support other activities as well.
The indirect costs include all the time and materials used to make use of the data. When
data are purchased from another organization, the indirect costs may actually be more significant
than the direct ones.
It may take longer for staff to handle data with which they are unfamiliar, or the data
may not be compatible with the other data sets to be used.
But during the last many years, GIS data most often have been digitized from several
sources, including hard copy maps, rectified aerial photography and satellite imagery. Hard-copy
maps (e.g. paper, vellum and plastic film) may contain unintended production errors as well as
unavoidable or even intended errors in presentation. The following are "errors" commonly found
in maps.
Indistinct Boundaries
Indistinct boundaries typically include the borders of vegetated areas, soil types, wetlands
and land use areas. In the real world, such features are characterized by gradual change, but
cartographers represent these boundaries with a distinct line. Some compromise is inevitable.
Map Scale
Cartographers and photogrammetrists work to accepted levels of accuracy for a given map
scale as per National Map Accuracy Standards. Locations of map features may disagree with
actual ground locations, although the error likely will fall within specified tolerances. Of course,
the problem is compounded by limitations in linear map measurements-typically about 1/100th of
an inch on a map scale.
Map Symbology
It is impossible to perfectly depict the real world using lines, colors, symbols and patterns.
Cartographers work with certain accepted conventions. As a result, facts and features represented
on maps often must be interpreted or interpolated, which can produce errors. For example, terrain
elevations typically are depicted using topographic contour lines and spot elevations. Elevations
of the ground between the lines and spots must be interpolated. Also, areas symbolized as "forest"
may not depict all open areas among the trees.
A digitizer must accurately discern the centre of a line or point as well as accurately trace
it with a cursor. This task is especially prone to error if the map scale is small and the lines or
symbols are relatively thick or large. The method of digitizing curvilinear lines also affects
accuracy. "Point-mode" digitizing, for example, places sample points at selected locations along a
line to best represent it in a GIS. The process is subject to judgment of the digitizer who selects
the number and placement of data points. "Stream-mode" digitizing collects data points at a pre-
set frequency, usually specified as the distance or time between data points. Every time an
operator strays from an intended line, a point digitized at that moment would be inaccurate. This
method also collects more data points than may be needed to faithfully represent a map feature.
Therefore, post-processing techniques often are used to "weed out" unneeded data points.
Heads-up digitizing often is preferred over table digitizing, because it typically yields
better results more efficiently. Keyed data entry of land parcel data is the most precise method.
Moreover, most errors are fairly obvious, because the source data usually are carefully computed
Spatial Data Models 2.43
and thoroughly checked. Most keyed data entry errors show as obvious mismatches in the parcel
"fabric."
GIS software usually includes functions that detect several types of database errors. These
error-checking routines can find mistakes in data topology, including gaps, overshoots, dangling
lines and unclosed polygons. An operator sets tolerances that the routine uses to search for errors,
and system effectiveness depends on setting correct tolerances. For example, tolerances too small
may pass over unintentional gaps, and tolerances too large may improperly remove short dangling
lines or small polygons that were intentionally digitized.
The phrasing of spatial and attribute queries also may lead to errors. In addition, the use of
Boolean operators can be complicated, and results can be decidedly different, depending on how a
data query is structured or a series of queries are executed. For example, the query, "Find all
structures within the 100 year flood zone," yields a different result than, "Find all structures
touching the 100 year flood zone." The former question will find only those structures entirely
within the flood zone, whereas the latter also will include structures that are partially within the
zone.
Dataset overlay is a powerful and commonly used GIS tool, but it can yield inaccurate
results. To determine areas suitable for a specific type of land development project, one may
overlay several data layers, including natural resources, wetlands, flood zones, land uses, land
ownership and zoning. The result usually will narrow the possible choices down to a few parcels
that would be investigated more carefully to make a final choice. The final result of the analysis
will reflect any errors in the original GIS data. Its accuracy only will be as good as the least
accurate GIS dataset used in the analysis.
It is also common to overlay and merge GIS data to form new layers. In certain
circumstances, this process introduces a new type of error: the polygon "sliver." Slivers often
appear when two GIS datasets with common boundary lines are merged. If the common elements
have been digitized separately, the usual result will be sliver polygons. Most GIS software
products offer routines that can find and fix such errors, but users must be careful in setting search
and correction tolerances.
Many errors can be avoided through proper selection and "scrubbing" of source data
before they are digitized. Data scrubbing includes organizing, reviewing and preparing the source
materials to be digitized. The data should be clean, legible and free of ambiguity. "Owners" of
source data should be consulted as needed to clear up questions that arise.
Data entry procedures should be thoroughly planned, organized and managed to produce
consistent, repeatable results. Nonetheless, a thorough, disciplined quality review and revision
process also is needed to catch and eliminate data entry errors. All production and quality control
Spatial Data Models 2.45
procedures should be documented, and all personnel should be trained in these procedures.
Moreover, the work itself should be documented, including a record of what was done, who did it,
when was it done, who checked it, what errors were found and how they were corrected.
To avoid misusing GIS data and the misapplication of analytical software, GIS analysts
including casual users need proper training. Moreover, GIS data should not be provided without
metadata indicating the source, accuracy and specifics of how the data were entered.
2.46 Geographic Information System
TWO MARK QUESTIONS WITH ANWERS
Vector, e.g,
• ArcInfo Coverage’s
• ArcGIS Shape Files
• CAD (AutoCAD DXF & DWG, or Micro Station DGN files)
• ASCII coordinate data
Raster, e.g,
• ArcInfo Grids
• Images
Spatial Data Models 2.47
• Digital Elevation Models (DEMs)
• generic raster datasets
The choice of data input method is governed largely by the application, the available
budget, and the type and the complexity of data being input.
There are at least four basic procedures for inputting spatial data into a GIS. These are:
• Manual digitizing;
• Automatic scanning;
• Entry of coordinates using coordinate geometry; and the
• Conversion of existing digital data.
3.2.3. Digitizing
While considerable work has been done with newer technologies, the overwhelming
majority of GIS spatial data entry is done by manual digitizing. A digitizer is an electronic device
consisting of a table upon which the map or drawing is placed. The user traces the spatial features
with a hand-held magnetic pen, often called a mouse or cursor. While tracing the features the
coordinates of selected points, e.g. vertices, are sent to the computer and stored. All points that are
recorded are registered against positional control points, usually the map corners that are keyed in
Data Input and Topology 3.5
by the user at the beginning of the digitizing session. The coordinates are recorded in a user
defined coordinate system or map projection. Latitude and longitude and UTM is most often used.
The ability to adjust or transform data during digitizing from one projection to another is a
desirable function of the GIS software. Numerous functional techniques exist to aid the operator
in the digitizing process.
Digitizing can be done in a point mode, where single points are recorded one at a time, or
in a stream mode, where a point is collected on regular intervals of time or distance, measured by
an X and Y movement, e.g. every 3 metres. Digitizing can also be done blindly or with a graphics
terminal. Blind digitizing infers that the graphic result is not immediately viewable to the person
digitizing. Most systems display the digitized linework as it is being digitized on an
accompanying graphics terminal.
Most GIS's use a spaghetti mode of digitizing. This allows the user to simply digitize lines
by indicating a start point and an end point. Data can be captured in point or stream mode.
However, some systems do allow the user to capture the data in an arc/node topological data
structure. The arc/node data structure requires that the digitizer identify nodes.
The sheer cost of scanning usually eliminates the possibility of using scanning methods
for data capture in most GIS implementations. Large data capture shops and government agencies
are those most likely to be using scanning technology.
Currently, general consensus is that the quality of data captured from scanning devices is
not substantial enough to justify the cost of using scanning technology. However, major
breakthroughs are being made in the field, with scanning techniques and with capabilities to
automatically clean and prepare scanned data for topological encoding. These include a variety of
line following and text recognition techniques. Users should be aware that this technology has
great potential in the years to come, particularly for larger GIS installations.
Most GIS software vendors also provide an ASCII data exchange format specific to their
product, and a programming subroutine library that will allow users to write their own data
conversion routines to fulfill their own specific needs. As digital data becomes more readily
available this capability becomes a necessity for any GIS. Data conversion from existing digital
data is not a problem for most technical persons in the GIS field. However, for smaller GIS
installations who have limited access to a GIS analyst this can be a major stumbling block in
getting a GIS operational. Government agencies are usually a good source for technical
information on data conversion requirements.
Some of the data formats common to the GIS marketplace are listed below. Please note
that most formats are only utilized for graphic data. Attribute data is usually handled as ASCII
text files. Vendor names are supplied where appropriate.
IGDS - Interactive Graphics This binary format is a standard in the turnkey CAD market
Design Software (Intergraph / and has become a de facto standard in Canada's mapping
Micro station) industry. It is a proprietary format, however most GIS
software vendors provide DGN translators.
DLG - Digital Line Graph (US This ASCII format is used by the USGS as a distribution
Geological Survey) standard and consequently is well utilized in the United
States. It is not used very much in Canada even though most
software vendors provide two way conversions to DLG.
DXF - Drawing Exchange This ASCII format is used primarily to convert to/from the
Format (Auto-cad) Auto-cad drawing format and is a standard in the
engineering discipline. Most GIS software vendors provide a
DXF translator.
GENERATE - ARC/INFO A generic ASCII format for spatial data used by the
Graphic Exchange Format ARC/INFO software to accommodate generic spatial data.
EXPORT - ARC/INFO Export An exchange format that includes both graphic and attribute
Format. data. This format is intended for transferring ARC/INFO
data from one hardware platform, or site, to another. It is
also often used for archiving.
3.8 Geographic Information System
ARC/INFO data. This is not a published data format,
however some GIS and desktop mapping vendors provide
translators. EXPORT format can come in either
uncompressed, partially compressed, or fully compressed
format
A wide variety of other vendor specific data formats exist within the mapping and GIS
industry. In particular, most GIS software vendors have their own proprietary formats. However,
almost all provide data conversion to/from the above formats. As well, most GIS software vendors
will develop data conversion programs dependant on specific requests by customers. Potential
purchasers of commercial GIS packages should determine and clearly identify their data
conversion needs, prior to purchase, to the software vendor.
3.3.1. Grids
Grids are an ESRI file format used to store both discrete features such as buildings, roads,
and parcels, and continuous phenomena such as elevation, temperature, and precipitation. Recall
that the basic unit of the raster data model is the cell. Cells store information about what things are
like at a particular location on the earth's surface. Depending on the type of data being stored, cell
values can be either integers (whole numbers) or floating points (numbers with decimals). There
are two types of grids: one store integers and the other stores floating points.
A discrete grid contains cells whose values are integers, often code numbers for a
particular category. Cells can have the same value in a discrete grid. For example, in a discrete
grid of land use, each land use type is coded by a different integer, but many cells may have the
same code. Discrete grids have an attribute table that stores the cell values and their associated
attributes.
Data Input and Topology 3.9
Continuous grid is used to represent continuous phenomena; its cell values are floating
points. Each cell in a continuous grid can have a different floating point value. For example, in a
continuous grid representing elevation, one cell might store an elevation value of 564.3 meters,
while the cell to the left might store an elevation value of 565.1 meters. Unlike discrete grids,
continuous grids don't have an attribute table.
Discrete grids represent discrete features such as land use categories with integer values.
Continuous grids represent continuous phenomena such as elevation with floating point values.
The attribute tables of discrete grids are INFO format, the same format in which coverage
feature class attribute tables are stored. As with coverage attribute tables, the INFO table of a
discrete grid is stored within an info folder, which is stored at the same level as the grid in a
workspace folder. Again like coverages, there is one info folder for all the grids in a workspace
folder. To avoid breaking or corrupting the connection between grid files and the info folder,
always use ArcCatalog to move, copy, rename, and delete grids.
The Grids workspace folder contains two grids: soils and vegetation. The attribute tables
for both grids are stored in the info folder. Auxiliary files called soils.aux and vegetation.aux link
the grids and their attribute tables.
3.3.2. Images
The term "image" is a collective term for raster’s whose cells, or pixels, store brightness
values of reflected visible light or other types of electromagnetic radiation, such as emitted heat
(infrared) or ultraviolet (UV). Aerial photos, satellite images, and scanned paper maps are
examples of images commonly used in a GIS.
Images can be displayed as layers in a map or they can be used as attributes for vector
features. For example, a real estate company might include photos of available houses as an
attribute of a home’s layer. To be displayed as a layer, however, images must be referenced to
real-world locations.
For example, an aerial photo as it comes from the camera is just a static picture, like a
picture of a house. There's no information about what part of the world the photo has captured,
and the photo may contain distortion and scale variations caused by the angle of the camera. To
display properly with other map layers, the aerial photo must be assigned a coordinate system and
some of its pixels must be linked to known geographic coordinates.
Raster images, such as aerial photographs and scanned maps, can be referenced to real-
world locations, then displayed as a layer in a GIS map.
3.10 Geographic Information System
There are many image file formats, which differ in the type of compression used to reduce
the file size. Some of the image formats supported by ArcGIS software.
3.5. DIGITIZERS
• Digitizers are the most common device for extracting spatial information from maps
and photographs
◦ the map, photo, or other document is placed on the flat surface of the digitizing
tablet
3.5.1. Hardware
• The position of an indicator as it is moved over the surface of the digitizing tablet is
detected by the computer and interpreted as pairs of x,y coordinates
◦ the indicator may be a pen-like stylus or a cursor (a small flat plate the size of a
hockey puck with a cross-hair)
• frequently, there are control buttons on the cursor which permit control of the system
without having to turn attention from the digitizing tablet to a computer terminal
• digitizing tablets can be purchased in sizes from 25x25 cm to 200x150 cm, at
approximate costs from $500 to $5,000
• early digitizers (ca. 1965) were backlit glass tables
Data Input and Topology 3.13
◦ a magnetic field generated by the cursor was tracked mechanically by an arm
located behind the table
◦ the arm's motion was encoded, coordinates computed and sent to a host processor
◦ some early low-cost systems had mechanically linked cursors - the free-cursor
digitizer was initially much more expensive
• the first solid-state systems used a spark generated by the cursor and detected by linear
microphones
◦ problems with errors generated by ambient noise
• contemporary tablets use a grid of wires embedded in the tablet to generate a magnetic
field which is detected by the cursor
◦ accuracies are typically better than 0.1 mm
◦ this is better than the accuracy with which the average operator can position the
cursor
◦ functions for transforming coordinates are sometimes built into the tablet and used
to process data before it is sent to the host
Dimensionality - the distinction between point, line, area, and volume, which are said to
have topological dimensions of 0, 1, 2, and 3 respectively.
3.6.1. Adjacency
Adjacency including the touching of land parcels, counties, and nation-states (They share
a common border).
3.6.2. Connectivity
Connectivity including junctions between streets, roads, railroads, and rivers (Very
common topological error. See diagrams about "Overshoot" below).
3.6.3. Containment
Containment when a point lies inside rather than outside an area.
Topology defines and enforces data integrity rules (there should be no gaps between
polygons). It supports topological relationship queries and navigation (navigating feature
adjacency or connectivity), sophisticated editing tools, and allows feature construction from
unstructured geometry (constructing polygons from lines).
3.16 Geographic Information System
Addressing topology is more than providing a data storage mechanism. In GIS, topology is
maintained by using some of the following aspects:
• The geo-database includes a topological data model using an open storage format for
simple features (i.e., feature classes of points, lines, and polygons), topology rules, and
topologically integrated coordinates among features with shared geometry. The data
model includes the ability to define the integrity rules and topological behaviour of the
feature classes that participate in a topology.
• Most GIS programs include a set of tools for query, editing, validation, and error
correction of topology.
• GIS software can navigate topological relationships, work with adjacency and
connectivity, and assemble features from these elements. It can identify the polygons that
share a specific common edge; list the edges that connect at a certain node; navigate along
connected edges from the current location; add a new line and "burn" it into the
topological graph; split lines at intersections; and create resulting edges, faces, and nodes.
Vector formats
• AutoCAD DXF – contour elevation plots in AutoCAD DXF format (by Autodesk)
• Cartesian coordinate system (XYZ) – simple point cloud
• Digital line graph (DLG) – a USGS format for vector data
• Esri TIN - proprietary binary format for triangulated irregular network data used
by Esri
• Geography Markup Language (GML) – XML based open standard (by OpenGIS)
for GIS data exchange
• GeoJSON – a lightweight format based on JSON, used by many open source GIS
packages
3.18 Geographic Information System
• GeoMedia – Intergraph's Microsoft Access based format for spatial vector storage
• ISFC – Intergraph's MicroStation based CAD solution attaching vector elements to a
relational Microsoft Access database
• Keyhole Markup Language (KML) – XML based open standard (by OpenGIS) for
GIS data exchange
• MapInfo TAB format – MapInfo's vector data format using TAB, DAT, ID and
MAP files
• National Transfer Format (NTF) – National Transfer Format (mostly used by the
UK Ordnance Survey)
• Spatialite – is a spatial extension to SQLite, providing vector geo-database
functionality. It is similar to Post-GIS, Oracle Spatial, and SQL Server with spatial
extensions
• Shapefile – a popular vector data GIS format, developed by Esri
• Simple Features – Open Geospatial Consortium specification for vector data
• SOSI – a spatial data format used for all public exchange of spatial data in Norway
• Spatial Data File – Autodesk's high-performance geo-database format, native
to MapGuide
• TIGER – Topologically Integrated Geographic Encoding and Referencing
• Vector Product Format (VPF) – National Geospatial-Intelligence Agency (NGA)'s
format of vectored data for large geographic databases
Grid formats
• USGS DEM – The USGS' Digital Elevation Model
• GTOPO30 – Large complete Earth elevation model at 30 arc seconds, delivered in the
USGS DEM format
• DTED – National Geospatial-Intelligence Agency (NGA)'s Digital Terrain Elevation
Data, the military standard for elevation data
• GeoTIFF – TIFF variant enriched with GIS relevant metadata
• SDTS – The USGS' successor to DEM
In the layer bar, right-click the COUNTY layer name to open the pop-up menu for the
COUNTY layer. Select Edit to open the GIS Layer window. In the definition for the COUNTY
Data Input and Topology 3.19
layer, select Thematic. The GIS Attribute Data Sets window appears for you to define the link to
the theme data set.
In the GIS Attribute Data Sets window, select New to define a new link. In the
resulting select a Member window, select MAPS.USAAC. You must next specify the values that
are common to both the attribute and spatial data, because the common values provide the
connection between the spatial data and the attribute data.
The spatial database and the MAPS.USAAC data set share compatible state and county
codes, so first select STATE in both the Data Set Vars and Compositeslists, and then select
COUNTY in both lists. Select Save to save the link definition to the Links list. Finally,
select Continue to close the GIS Attribute Data Setswindow.
After the GIS Attribute Data Sets window closes, the Var window automatically opens for
you. Select which variable in the attribute data provides the theme data for your theme. Select the
CHANGE variable to have the counties colored according to the level of change in the county
population. Select OK to close the Var window.
The counties in the spatial data are colored according to the demographic values in the
attribute data set, as shown in the following display.
If the external table has an existing spatial column that contains no data, the
ArcGIS Maps Connect workflow populates the column based on other location information in
the table (for example, address). If no spatial column exists, the ArcGIS Maps Connect
workflow creates a geography spatial type column named EsriShape with a Spatial Reference
Identifier (SRID) of 4326 (WGS 84). The EsriShape field supports all geometries including
points, lines, and polygons. In all scenarios, the external content can be enriched with
additional geographic data variables from ArcGIS.
3.7.2. Note
If the ArcGIS Maps Connect workflow fails, ensure the appropriate permissions for
Microsoft SQL Server have been set. You can view the error messages in the SharePoint site
workflow history to view exact details on the settings that need to be corrected.
When the ArcGIS Maps Connect workflow completes, the result is a regular SharePoint
list, not an external list. That said, the fields created from the SQL Server database are of an
external type, and edits made to these fields in SharePoint cannot be passed back to the database.
SharePoint can only pass back the fields it has created, such as for the ArcGIS Maps Locate
workflow and geoenrichment.
ODBC was originally developed by Microsoft and Simba Technologies during the early
1990s, and became the basis for the Call Level Interface (CLI) standardized by SQL Access
Group in the UNIX and mainframe field. ODBC retained several features that were removed as
part of the CLI effort. Full ODBC was later ported back to those platforms, and became a de facto
standard considerably better known than CLI. The CLI remains similar to ODBC, and
applications can be ported from one platform to the other with few changes.
Data Input and Topology 3.21
3.8.1. History of Before ODBC
The introduction of the mainframe-based relational database during the 1970s led to a
proliferation of data access methods. Generally these systems operated together with a simple
command processor that allowed users to type in English-like commands, and receive output. The
best-known examples are SQL from IBM and QUEL from the Ingres project. These systems may
or may not allow other applications to access the data directly, and those that did use a wide
variety of methodologies. The introduction of SQL aimed to solve the problem of language
standardization, although substantial differences in implementation remained.
Also, since the SQL language had only rudimentary programming features, users often
wanted to use SQL within a program written in another language, say Fortran or C. This led to the
concept of Embedded SQL, which allowed SQL code to be embedded within another language.
For instance, a SQL statement like SELECT * FROM city could be inserted as text within C
source code, and during compiling it would be converted into a custom format that directly called
a function within a library that would pass the statement into the SQL system. Results returned
from the statements would be interpreted back into C data formats like char * using similar library
code.
There were several problems with the Embedded SQL approach. Like the different
varieties of SQL, the Embedded SQLs that used them varied widely, not only from platform to
platform, but even across languages on one platform – a system that allowed calls into IBM's DB2
would look very different from one that called into their own SQL/DS. Another key problem to
the Embedded SQL concept was that the SQL code could only be changed in the program's source
code, so that even small changes to the query required considerable programmer effort to modify.
The SQL market referred to this as static SQL, versus dynamic SQL which could be changed at
any time, like the command-line interfaces that shipped with almost all SQL systems, or a
programming interface that left the SQL as plain text until it was called. Dynamic SQL systems
became a major focus for SQL vendors during the 1980s.
Older mainframe databases, and the newer microcomputer based systems that were based
on them, generally did not have a SQL-like command processor between the user and the database
engine. Instead, the data was accessed directly by the program – a programming library in the case
of large mainframe systems, or a command line interface or interactive forms system in the case
of dBASE and similar applications. Data from dBASE could not generally be accessed directly by
other programs running on the machine. Those programs may be given a way to access this data,
often through libraries, but it would not work with any other database engine, or even different
databases in the same engine. In effect, all such systems were static, which presented considerable
problems.
By the late 1980s there were several efforts underway to provide an abstraction layer for
this purpose. Some of these were mainframe related, designed to allow programs running on those
machines to translate between the variety of SQL's and provide a single common interface which
could then be called by other mainframe or microcomputer programs. These solutions included
IBM's Distributed Relational Database Architecture (DRDA) and Apple Computer's Data Access
Language. Much more common, however, were systems that ran entirely on microcomputers,
including a complete protocol stack that included any required networking or file translation
support.
One of the early examples of such a system was Lotus Development's DataLens, initially
known as Blueprint. Blueprint, developed for 1-2-3, supported a variety of data sources, including
SQL/DS, DB2, FOCUS and a variety of similar mainframe systems, as well as microcomputer
systems like dBase and the early Microsoft/Ashton-Tate efforts that would eventually develop into
Microsoft SQL Server. Unlike the later ODBC, Blueprint was a purely code-based system, lacking
anything approximating a command language like SQL. Instead, programmers used data
structures to store the query information, constructing a query by linking many of these structures
together. Lotus referred to these compound structures as query trees.
Around the same time, an industry team including members from Sybase (Tom Haggin),
Tandem Computers (Jim Gray & Rao Yendluri) and Microsoft (Kyle G) were working on a
standardized dynamic SQL concept. Much of the system was based on Sybase's DB-Library
system, with the Sybase-specific sections removed and several additions to support other
platforms. DB-Library was aided by an industry-wide move from library systems that were tightly
linked to a specific language, to library systems that were provided by the operating system and
required the languages on that platform to conform to its standards. This meant that a single
library could be used with (potentially) any programming language on a given platform.
The first draft of the Microsoft Data Access API was published in April 1989, about the
same time as Lotus' announcement of Blueprint. In spite of Blueprint's great lead – it was running
when MSDA was still a paper project – Lotus eventually joined the MSDA efforts as it became
clear that SQL would become the de facto database standard. After considerable industry input, in
the summer of 1989 the standard became SQL Connectivity (SQLC).
MS continued working with the original SQLC standard, retaining many of the advanced
features that were removed from the CLI version. These included features like scrollable cursors,
and metadata information queries. The commands in the API were split into groups; the Core
group was identical to the CLI, the Level 1 extensions were commands that would be easy to
implement in drivers, while Level 2 commands contained the more advanced features like cursors.
A proposed standard was released in December 1991, and industry input was gathered and worked
into the system through 1992, resulting in yet another name change to ODBC.
The SAG standardization efforts presented an opportunity for Microsoft to adapt their Jet
system to the new CLI standard. This would not only make Windows a premier platform for CLI
development, but also allow users to use SQL to access both Jet and other databases as well. What
was missing was the SQL parser that could convert those calls from their text form into the C-
interface used in Jet. To solve this, MS partnered with PageAhead Software to use their existing
query processor, SIMBA. SIMBA was used as a parser above Jet's C library, turning Jet into an
SQL database. And because Jet could forward those C-based calls to other databases, this also
allowed SIMBA to query other systems. Microsoft included drivers for Excel to turn its
spreadsheet documents into SQL-accessible database tables.
Meanwhile, the CLI standard effort dragged on, and it was not until March 1995 that the
definitive version was finalized. By then, Microsoft had already granted Visigenic Software a
source code license to develop ODBC on non-Windows platforms. Visigenic ported ODBC to a
wide variety of Unix platforms, where ODBC quickly became the de facto standard. "Real" CLI is
rare today. The two systems remain similar, and many applications can be ported from ODBC to
CLI with few or no changes.
Over time, database vendors took over the driver interfaces and provided direct links to
their products. Skipping the intermediate conversions to and from Jet or similar wrappers often
resulted in higher performance. However, by then Microsoft had changed focus to their OLE DB
concept (recently reinstated), which provided direct access to a wider variety of data sources from
address books to text files. Several new systems followed which further turned their attention
from ODBC, including ActiveX Data Objects (ADO) and ADO.net, which interacted more or less
with ODBC over their lifetimes.
As Microsoft turned its attention away from working directly on ODBC, the UNIX field
was increasingly embracing it. This was propelled by two changes within the market, the
introduction of graphical user interfaces (GUIs) like GNOME that provided a need to access these
sources in non-text form, and the emergence of open software database systems like PostgreSQL
and MySQL, initially under Unix. The later adoption of ODBC by Apple for using the standard
Unix-side iODBC package Mac OS X 10.2 (Jaguar) (which OpenLink Software had been
independently providing for Mac OS X 10.0 and even Mac OS 9 since 2001) further cemented
ODBC as the standard for cross-platform data access.
Sun Microsystems used the ODBC system as the basis for their own open standard, Java
Database Connectivity (JDBC). In most ways, JDBC can be considered a version of ODBC for
the programming language Java instead of C. JDBC-to-ODBC bridges allow Java-based programs
to access data sources through ODBC drivers on platforms lacking a native JDBC driver, although
these are now relatively rare. Inversely, ODBC-to-JDBC bridges allow C-based programs to
access data sources through JDBC drivers on platforms or from databases lacking suitable ODBC
drivers.
However, the rise of thin client computing using HTML as an intermediate format has
reduced the need for ODBC. Many web development platforms contain direct links to target
databases – MySQL being very common. In these scenarios, there is no direct client-side access
Data Input and Topology 3.25
nor multiple client software systems to support; everything goes through the programmer-supplied
HTML application. The virtualization that ODBC offers is no longer a strong requirement, and
development of ODBC is no longer as active as it once was.[citation needed]
3.9. GPS
Stands for "Global Positioning System." GPS is a satellite navigation system used to
determine the ground position of an object. GPS technology was first used by the United States
military in the 1960s and expanded into civilian use over the next few decades. Today, GPS
receivers are included in many commercial products, such as automobiles, Smartphone, exercise
watches, and GIS devices.
The GPS system includes 24 satellites deployed in space about 12,000 miles (19,300
kilometers) above the earth's surface. They orbit the earth once every 12 hours at an extremely fast
pace of roughly 7,000 miles per hour (11,200 kilometers per hour). The satellites are evenly
spread out so that four satellites are accessible via direct line-of-sight from anywhere on the globe.
Each GPS satellite broadcasts a message that includes the satellite's current position, orbit,
and exact time. A GPS receiver combines the broadcasts from multiple satellites to calculate its
exact position using a process called triangulation. Three satellites are required in order to
determine a receiver's location, though a connection to four satellites is ideal since it provides
greater accuracy.
In order for a GPS device to work correctly, it must first establish a connection to the
required number of satellites. This process can take anywhere from a few seconds to a few
minutes, depending on the strength of the receiver. For example, a car's GPS unit will typically
establish a GPS connection faster than the receiver in a watch or Smartphone. Most GPS devices
also use some type of location caching to speed up GPS detection. By memorizing its previous
location, a GPS device can quickly determine what satellites will be available the next time it
scans for a GPS signal.
Data: geospatial information (where things are located) and the details of objects such as
services, roads, buildings etc. are collected and entered into the GIS software
Software: GIS software analyses data and presents it in different combinations for the
user
Hardware: includes hand held devices for collecting data and computers with GIS
software
Positioning accuracy
Factors that trigger GPS position errors
Ionosphere
The ionosphere is a portion of the upper atmosphere, between the thermosphere and the
exosphere. When GPS signals pass through this layer, the propagation velocity of the GPS signal
goes slower, hence causing propagation error.
Troposphere
The troposphere is the lowest
portion of Earth's atmosphere. Radio
reflections caused by dry atmosphere
and water vapor within provoke GPS
position error.
Multipath propagation
GPS signal is not immune to
reflection when it hits on the ground,
structures and many others. This
phenomenon is called multipath
propagation, one of the causes of GPS
position errors. Fig.3.13. GPS position error
3.30 Geographic Information System
3.9.7. DOP (Dilution of Precision)
DOP is a value that shows the degree of degradation of the GPS positioning accuracy. The
smaller the value is, the higher the positioning accuracy is. This value depends upon the positions
of the GPS satellites tracked for positioning. If the tracked satellites spread evenly over the earth,
the positioning accuracy would become higher, and if the positions of tracked satellites are
disproportionate, the positioning accuracy would become lower.
If the number of the tracked satellites is great, GPS positioning becomes greater, but if
there were a fewer satellites tracked for positioning, it would be difficult to generate GPS position.
The Fig. 1-11 illustrates the occasion where the GPS receiver tracks a greater number of satellites
for positioning. The Fig. 1-12 illustrates the occasion where the GPS receiver tracks only a few
number of satellites for positioning.
4) What is overlying?
Map overlay is the process by which it is possible to take two or more different thematic map
layers of the same area and overlay them on top of the other and form a composite new layer
this techniques is used to overlay vector data on a raster image. In Vector base systems map
overlay is time consuming, complex and computationally expensive. In raster based systems
it is quick, straightforward and efficient
Flat bed scanner is a PC peripheral which is small and comparatively inaccurate. The rotating
drum scanners are accurate but they tend to be slow and expensive. Large format feed
scanner are the most suitable type for inputting GIS data as they are cheap, quick and
accurate.
Dimensionality - the distinction between point, line, area, and volume, which are said to
have topological dimensions of 0, 1, 2, and 3 respectively.
15) What is the difference between GPS remote sensing and GIS?
First do remote sensing than do GNSS (GPS) and at last GIS. GPS (global positioning
system) is a way to assign a location to a point on the Earth. Remote sensing is the use of
sensors on board either planes or satellites to collect data usually in a grid like pattern of
pixels called raster data.
DGPS uses pseudo-range corrections and range-rate corrections from the reference station
which has enough channels to track all satellites in view. The system calculates the difference
between the GPS calculated position for the ground reference station and the actual position
of the ground reference station.
Data Input and Topology 3.39
REVIEW QUESTIONS
1) Briefly explain the Scanners for Raster Data Input in GIS. (May/June - 2012)
2) Write detailed notes on raster data input system in GIS
3) Briefly explain the digitization processes in GIS. (May/June - 2012)
4) Write detailed notes on raster data file formats.
5) Describe the vector data input in GIS.
6) Briefly explain the digitization processes in GIS. (May/June - 2012)
7) Briefly explain the Topology in GIS.
8) Write detailed notes on linking the attribute data to the spatial data.
9) Briefly explain the Open Database Connectivity (ODBC)
10) Describe the GPS, Uses of GPS, structure of GPS and GPS positioning etc.
11) Briefly explain the Concept of GPS based mapping.
CHAPTER- 4
Spatial analysis is the process that turns raw input spatial data into value-added useful data
or information that ultimately supports decision making and reveals hidden or unknown patterns
Example:
• Identify potential customers within 3km of store
• Identify parks within 10km of Islamabad Highway
• Identify schools within 5 km of industrial zone
.
Fig.4.4. Overlay Analysis – Union
• Let’s say we are interested in knowing no potential zone for urban development. It is
clear that no construction can be done on a water body or land covered by agriculture
or forest. So, we can say union of areas under water, agriculture and forest would
present us the area having no potential for urban development.
• Intersect creates a new theme by overlaying a point, line or polygon theme with an
intersecting polygon theme. It is same as ‘and’ Boolean operator. The output theme
contains only the feature inside the intersecting polygons.
Append
Select
Clip
Data Quality and Standards 4.5
Erase
Split
Statistics The Statistics toolset contains tools that perform standard statistical analysis
(such as mean, minimum, maximum, and standard deviation) on attribute data as
well as tools that calculate area, length, and count statistics for overlapping and
neighboring features.
A new road should be aligned very carefully as improper alignment would result in one or
more of the following disadvantages:
1) Increase in construction cost
2) Increase in maintenance cost
3) Increase in vehicles operation cost
4) Increase in accident rate.
The basic requirements of an ideal alignment between two terminal stations are that it
should be:
1) Short 2) Easy 3) Safe 4) Economical
4.3.7. Methodology
• The base (study area) map, Drainage, Slope and Contour maps were prepared with the
help of SOI Topo-sheet (on 1:50,000 scale).
4.8 Geographic Information System
• IRS LISS III satellite data was used and by using Digital Image Processing techniques
the following thematic maps such as geomorphology, Land use/ Land Cover were
generated as shown in the figure.5.22.
• The DEM is used in order to understand the terrain condition, environmental factors
and social economic status in this study area.
• The factors considered are mainly related to the land use, geology, land value and soil.
The weights and ranks are assigned to each of the above themes, according to expert
opinions, for GIS analysis. After assigning weights and ranks these themes are
overlaid to get an overlaid map.
• Finally, possible/feasible route was identified based on various physical and cultural
parameters and their inherent properties.
• The cost reduction analysis was also done for substantiating the formation of highway.
Finally, possible/feasible route was identified based on various physical and cultural
parameters and their inherent properties. The cost reduction analysis was also done for
substantiating the formation of highway.
Figure.5.22. Methodology
• The main purpose of traffic survey are traffic monitoring, traffic control and
management, traffic enforcement, traffic forecasting, model calibration and validating
etc.
• The purpose of caring out traffic volume count are designing, improving traffic
system, planning, management.
• The traffic volume count study is carried out to get following useful information.
Magnitudes, classifications and the time and directional split of vehicular flows
Data Quality and Standards 4.9
• Proportion of vehicles in traffic stream - Hourly, daily, yearly and seasonal variation
of vehicular flows
• Flow fluctuation on different approaches at a junction or different parts of a road
network system.
• Network analysis is used to find the shortest alternated routes between origins to
destination Network Analyst provides network-based spatial analysis tools for solving
complex routing problems.
4.3.8. Conclusion
• The purpose of this study was to develop a tool to locate a suitable less time
consuming, shortest route between two points.
• The GIS approach using ground parameters and spatial analysis provided to achieve
this goal. Raster based map analysis provide a wealth of capabilities for incorporating
terrain information surrounding linear infrastructure.
• Costs resulting from terrain, geomorphology, land use, drainage and elevation
resulting the shortest routes for the study area.
• Results indicate that the route which was designed applying GIS method is avoid
traffic problems, less time consuming more environmentally effective, and cheaper.
• This proposed shortest route provides traffic free, pollution free, risk free, operating
for movement of vehicle passing from chettikullam to kottar.
• Time and consumption of fuel will also be reduced considerably. GIS method can also
be used for route determination for irrigation, drainage channels, power lines and
railways.
From the introduction of the World Wide Web in 1993 the young of the world have
experienced two models of digital education, that outside the school walls and that within.
Outside the young and the digitally connected families of the world employed – unseen –
the naturally evolving laissez faire model. Within the school the young worked within the
traditional, highly structured model.
4.10 Geographic Information System
It is time the difference is understood, the global success and benefits of the laissez faire
recognised and lauded, and the serious shortcomings of the highly structured understood and
addressed.
For much of the period the two models ran in parallel, with most schools showing little or
no interest in the out of school digital education.
Around 2010 – 2012 the scene began to change when a handful of digitally mature schools
began genuinely collaborating with their families in the 24/7/365 digital education of the children.
Those schools had reached the evolutionary stage where their teaching model and culture closely
mirrored that of the families. They revealed what was possible with collaboration.
That said it took time for that collaboration to take hold more widely and for the most part
the parallel models continue in operation today, with the difference between the in and out of
school teaching growing at pace.
It is surely time for schools and government to question the retention of the parallel modes
and to ask if taxpayers are getting value for the millions upon millions spent solely on schools
when the digitally connected families receive no support.
Might it be time to employ a more collaborative approach where the schools complement
and add value to the contribution of the families?
Without going into detail, it bears reflecting on the distinguishing features of the learning
environment and digital education model, of both the digitally connected family and the school,
and asking what is the best way forward,
Importantly the family as a small regulating unit, with direct responsibility for a small
number of children could readily trust each, and monitor, guide and value their learning from birth
onwards, assisting ensure each child had use of the current technology and that the use was wise
and balanced.
The learning occurred within a freewheeling, dynamic, market driven, naturally evolving
environment, anywhere, anytime, just in time and invariably in context. Those interested could
operate at the cutting edge and the depth desired.
Very early on the young’s use of the digital was normalized, with the learning occurring as
a natural part of life, totally integrated, with no regard for boundaries
Data Quality and Standards 4.11
The time available to the digitally connected family was – and continues to be – at least
four/five times greater than that in the school.
Very quickly the family learning environment became collaborative, socially networked,
global in its outlook, highly enjoyable and creative where the young believed anything was
possible.
By the latter 2000’s most families had created – largely unwittingly – their own
increasingly integrated and sophisticated digital ecosystem, operating in the main on the personal
mobile devices that connected all in the family to all manner of other ecosystems globally.
They chose, configured and controlled the use of both the hardware and software,
invariably opting for one device, one operating system and a standard suite of applications.
The students were taught within class groups, using highly structured, sequential, teacher
directed, regularly assessed instructional programs.
The school knew best. The clients – the parents and students – were expected to acquiesce.
There was little or no recognition of the out of school learning or technology or desire to
collaborate with the digitally connected families.
In reflecting on school’s teaching with the digital between 1993 and 2016 there was an all-
pervasive sense of constancy, continuity, with no real rush to change. There was little sense that
the schools were readying the total student body to thrive within in a rapidly evolving digitally
based world.
The reality was that while the learning environment of the digitally connected families
evolved naturally at pace that of most schools changed only little, with most schools struggling to
accommodate rapid digital evolution and transformation.
It wasn’t until the authors identified and documented the lead role of the digitally
connected families of the world did we appreciate all were using basically the same learning
approach. The pre-primary developments of the last few years affirmed the global application of
the model.
We saw at play a natural model that was embraced by the diverse families of the world.
All were using the same model – a naturally evolving model where the parents were
‘letting things take their own course ‘(OED).
The learning was highly individualized, with no controls other than the occasional parent
nudge. That said the learning was simultaneously highly collegial, with the young calling upon
and collaborating with their siblings, family members, peers and social networks when desired.
Interestingly from early on the young found themselves often knowing more about the
technology in some areas than their elders – experiencing what Tapscott (1998) termed an
‘inverted authority’ – being able to assist them use the technology.
Each child was free to learn how to use, and apply those aspects of the desired
technologies they wanted, and to draw upon any resources or people if needed.
In the process the children worldwide – from as young as two – directed their own
learning, opting usually for a discovery based approach, where the learning occurred anytime,
anywhere 24/7/365. Most of the learning was just in time, done in context and was current,
relevant, highly appealing and intrinsically motivating. Invariably it was highly integrated, with
no thought given to old boundaries – like was it educational, entertainment, communication, social
science or history.
In contrast the school digital teaching model has always been highly structured and
focused on what the school or education authority ‘experts’ believed to be appropriate.
Throughout the period the teaching has been unilaterally controlled, directed by the
classroom teacher, with the students disempowered, distrusted and obliged to do as told.
The teaching built upon linear, sequential instructional programs where the digital
education was invariably treated like all other subjects, shoehorned into an already crowded
curriculum and continually assessed. Some authorities made the ‘subject’ compulsory, others
made it optional.
The focus – in keeping with the other ‘subjects’ in the curriculum – was academic. There
was little interest in providing the young the digital understanding for everyday life.
The teaching took place within a cyber walled community, at the time determined by the
teaching program.
Data Quality and Standards 4.13
Increasingly the course taught and assessed became dated and irrelevant.
In considering why the young and the digitally connected families of the world have
embraced the laissez faire model of digital education aside from the young’s innate curiosity and
desire to learn we might do well to examine the model of digital learning we have used over the
last twenty plus years and reflect on how closely it approximates that adopted by the young.
Might they be following that ancient practice of modelling the behavior of their parents?
To minimize the negative impacts of urban sprawl and increased development, it’s
important for city planners to carefully determine the best way to grow urban areas. Urban
development needs to take into consideration today’s requirements, potential changes in demand
and the long-term effects of building upward and outward.
3D GIS software can help city planners visualize what their proposed changes will look
like and predict the outcomes for current residents and future generations. One example was the
2012 revitalization of the Mulheim Sud district in Cologne, Germany, located on the Rhine River.
The project set out to make the district, which included a mix of residential, commercial and
industrial buildings, more environmentally friendly over the course of two decades.
BIM and 3D GIS can also come together to support the preservation and restoration of
historical buildings. An effort to digitally record cultural heritage sites in Dublin, Ireland, drew on
Historic Building Information Modeling (HBIM) and 3D GIS to document and analyze selected
locations.
One project focused on restoring buildings along Henrietta Street, which dates its earliest
construction to 1730. By the 21st century, the street was lined with buildings in serious need of
care. HBIM technology made it possible to map the extent of the damage and visualize what the
area looked like when new. Researchers employed GIS tools to note attributes of individual
buildings like the years of construction and address information. In the process, they developed a
store of information that could be used to generate in-depth visualizations or guide tourists.
Its critical for planners to understand the factors that affect construction and maintenance
for shipping ports, fisheries, mineral mining operations and wilderness preservation areas.
Responsible coastal development must be informed by underwater topography, local vegetation
and predictions for the long-term environmental impact.
Resource planning systems that draw on GIS can provide insights into the economic,
environmental and cultural results of activities along the coast. The right data makes all the
difference in sustainably performing operations like construction or excavation. When preparing
for the extraction of resources on the coastline, organizations benefit from synthesizing
information like:
• Findings from 3D GIS mapping that suggest the likely outcomes of dredging material
in the water
• Lidar topographical surveys
• Data sets from past extraction activities
• Trends in coastal change
In Switzerland, developers wanted to find a way to accurately determine the noise that
would be generated from the installation of a new wind farm. A team developing a visual-acoustic
simulation tool decided to study Mont Crosin in the Canton of Bern, Switzerland, which is home
to 16 wind turbines. The researchers analyzed recordings taken on days with varying wind and
weather conditions and wind speed measurements taken with a 3D ultrasonic anemometer. They
generated 3D models representing vegetation, infrastructure and the wind turbines themselves.
The data allowed planners to predict the noise and environmental impact that would be produced
by the proposed wind farm.
Data Quality and Standards 4.17
TWO MARK QUESTIONS WITH ANWERS
3) What is buffering?
Buffering is the creation of polygons that surround other points, lines or polygons. Buffers
M/J be created either to exclude a certain amount of area around a point, line or polygon or to
include only the buffer area in a study
APPLICATIONS
Mapping locations:
GIS can be used to map locations. GIS allows the creation of maps through automated
mapping, data capture, and surveying analysis tools.
Mapping quantities:
People map quantities, like where the most and least are, to find places that meet their
criteria and take action, or to see the relationships between places. This gives an additional level
of information beyond simply mapping the locations of features.
Mapping densities:
While you can see concentrations by simply mapping the locations of features, in areas
with many features it may be difficult to see which areas have a higher concentration than others.
A density map lets you measure the number of features using a uniform areal unit, such as acres or
square miles, so you can clearly see the distribution.
Finding distances:
GIS can be used to find out what's occurring within a set distance of a feature.
Management of natural resources is the need of the hour and therefore commands due
importance in the GIS, CAD and Software industry as well. The technical specifications of
geographic information system (GIS) and computer aided designing and drafting services assist in
management of natural resources.
The GIS for Natural Resource Management maps are useful tool that help in fast and
effective decision making as they create comprehensive maps.
The services offered by a part of its GIS for Natural Resource Management are:
• Thematic Mapping for Natural Resource Management
• Geospatial Livelihood planning
• Spatial Data modelling
• Site suitability analysis
• Satellite image processing
• Land Use Land Cover Mapping
• Geological database creation
• Environmental management using GIS
• Forestry mapping services
GIS stands for Geographic Information system. It is a computer system that collects,
analyzes, stores and disseminates geographic information for use by different entities to facilitate
informed and smooth decision making. GIS has important uses in different fields, including in the
environmental field. Below are some of the importance of GIS in environment.
The GIS for mining and geology GIS services at IT can assist the industry by providing:
• Site suitability analysis
• Mineral mapping to identify potential mineral zones
• Lithological structural features detection
• Geological database creation
• Map updation for mineral exploration
• Capturing, gridding, contouring of features
• Georeferencing
• Topographical & Physiographical mapping
• Satellite image processing
• Image Classification
• Well-log data capturing
• Digitization and Attribution
• Theme Creation and Map Preparation
• Data Analysis and Report Generation
• Trench & Surface mapping
5.6 Geographic Information System
• Rock / Mineral mapping and Analysis
• Geological Thematic mapping
GIS for mining and geology GIS services use image analysis software packages such as
ENVI, ER-Mapper, ERDAS Imagine, MultiSpec, ArcGIS, ERDAS Imagine / ER Mapper, Arc-
GIS, MapInfo, AutoCAD Map and Micro-Station to easily build focused and detailed spectral
libraries tailored to the application requirements.
IT has been providing GIS for oil and gas to leading oil and petroleum exploration and
distribution companies globally. Their operational and research teams have been using GIS for oil
and gas for the following:
• Data index maps creation
• Exploration Survey
• Block ranking via quantitative analysis
• Land management
• Well log digitization
• Thematic maps
• Field operations using high resolution imagery
• Pipeline planning
• Environmental management
• Disaster management
GIS for oil and gas use industry standard software such as Arc-GIS, ERDAS Imagine / ER
Mapper, MapInfo and AutoCAD Map for projects pertaining to oil and gas industry.
5.2.13. Forestry
Forestry mapping services such as Forestry GIS are often required by government and
commercial agencies to perform value assessments for certification, funding, policy
recommendations and regulatory compliances. Often, the forest-cover data available with the
private and govt. agencies is dated and unreliable. This data is not helpful to the user, forestry GIS
Applications 5.7
can enable creation of updated data. The fast paced growth and the effect of multiple industries
have led to the need for a more precise and unique response to manage the forestry data.
IT‘s forestry mapping services use Forestry GIS and other techniques to provide
project specific solutions in the field of forest management, watershed management,
desertification combat, biodiversity management, environmental monitoring and wildlife
habitat encroachment.
The services provided a Forestry GIS sector to help in managing the forestry assets
can be summarized as:
• Forestry GIS management and suitable livelihood development
• Forest Spatial Data Information portal
• Forestry mapping services for corridor identification
• Species level classification and biodiversity forestry mapping services
• Change detection studies of forest management
• Environment impact assessment studies
• Identifying sites for eco-restoration
• Forestry mapping services of plantations and zonal mapping
• Ortho rectification, haze removal, NDVI for deforestation, forest management and fire
mapping
IT‘s Forestry GIS services use high end software such as Arc-GIS, ERDAS Imagine/ER
Mapper, MapInfo and AutoCAD Map, and Arc-FM.
5.2.14. Agriculture
GIS application in agriculture has been playing an increasingly important role in crop
production throughout the world by helping farmers in increasing production, reducing costs, and
managing their land resources more efficiently. GIS application in agriculture such as agricultural
mapping plays a vital role in monitoring and management of soil and irrigation of any given farm
land. GIS agriculture and agricultural mapping act as essential tools for management of
agricultural sector by acquiring and implementing the accurate information into a mapping
environment. GIS application in agriculture also helps in management and control of agricultural
resources. GIS agriculture technology helps in improvement of the present systems of acquiring
and generating GIS agriculture and resources data.
IT uses GIS application in agriculture sector such as GIS agriculture for improving present
method of acquiring and generating agricultural and resources data.
The company uses different spatial, spectral and temporal resolutions for:
• Crop mapping yield estimation
• Crop assessment and Crop health
• Irrigated landscape mapping
5.8 Geographic Information System
• Application development for GIS agriculture
• Soil and irrigation amendment analysis
• Suitability assessment studies
• Erosion identification and remediation
• Agricultural mapping for detailed vegetation cover and monitoring
• Change detection studies and developing crop models
• Damage and land degradation assessment studies
• Elevation models for efficient drainage
These studies and projects can be executed on any standard GIS formats using high end
software such as Arc-GIS, ERDAS Imagine/ER Mapper, MapInfo and Auto-CAD Map, and Arc-
FM.
5.3.1. Planning
In planning its major contribution is to give us with an organized set of data which can
help professionals to combat complex scenarios relating to the selection of site, environmental
impact, study of ecosystem, managing risk regarding the use of natural resources, sustainability
issues, managing traffic congestion, routing of roads and pipelines etc.
5.3.3. Analysis
Analysis is one of the major and most influential phases of infrastructure life cycle.
Analysis guides us about the validity or correctness of design or we can say that analysis is a
method which supports our design. Some of the analyses that can be performed by GIS are:
5.3.4. Construction
It is the stage when all layout plans and paper work design come into existence in the real
world. The GIS helps the professionals to understand the site conditions that affect the schedule
baseline and cost baseline. To keep the construction within budget and schedule GIS guides us
about how to utilize our resources on site efficiency by:
• Timely usage of construction equipment.
• Working Hours
• Effects of seasonal fluctuations.
• Optimizing routes for dumpers and concrete trucks
• Earth filling and cutting
• Calculation of volumes and areas of constructed phase thereby helping in Estimation
and Valuation.
5.3.5. Operations
Operations are controlled by modeling of site data and compared by the baselines prepared
in planning phase. Modeling of site may be in the form of raster images or CAD drawings. These
can help us to keep track of timely operations of activities.
GIS can help to make a record of work that has been completed and can give us
visualization in the form of thematic maps which will guide us about rate of operations, completed
operations and pending operations.
In short we can say that GIS will prove to be the foundation of next generation civil
engineering.
A GIS application needs to be explored even more and the near future is heavily relying
on GIS.
5.4. NAVIGATION
Navigation mapping assists the navigation content providers using GIS, GPS, and Remote
sensing technologies instead of traditional method of manual assimilation of data for generation of
navigation content. The accuracy of the new technology ensures creation of flawless maps. The
rampant increase in the number of people who use computerised navigational services for
everyday use. This trend has led to the development of new and innovative geodata content
services and products.
Its uses high end software such as Arc-GIS, ERDAS Imagine / ER Mapper, MapInfo and
AutoCAD Map to generate the final product in varied formats according to the project‘s
requirements.
5.4.1. Direction-Finding
The use of GIS in navigation industry has been growing at a steady rate. These days the
navigation industry uses new technologies such as GIS, GPS, Remote sensing instead of
Applications 5.11
traditional method of manual assimilation of data for generation of navigation content. The GIS
services assist the navigation industry in creating flawless maps.
This steep rise in use of GIS can be credited to the rampant increase in the number of
people who use computerised navigational services for everyday use. This trend has led to the
development of new and innovative geo-data content services and products. IT offers high quality
and cost effective navigational map data services.
The experts use high end software such as ArcGIS, ERDAS Imagine / ER Mapper,
MapInfo and AutoCAD Map to generate the final product in varied formats according to the
client‘s requirements.
This new technology, popularly called vehicle Tracking Systems has done wonders in
maintaining the security of the vehicle.
Vehicle Tracking System is one of the biggest technological advancements to track the
activities of the vehicle. The security system uses Global Positioning System GPS, to find the
location of the monitored or tracked vehicle and then uses satellite or radio systems to send to
send the coordinates and the location data to the monitoring center. At monitoring center various
software‘s are used to plot the Vehicle on a map. In this way the Vehicle owners are able to track
their vehicle on a real-time basis. Due to real-time tracking facility, vehicle tracking systems are
becoming increasingly popular among owners of expensive vehicles.
5.12 Geographic Information System
The vehicle tracking hardware is fitted on to the vehicle. It is fitted in such a manner that it
is not visible to anyone who is outside the vehicle. Thus it operates as a covert unit which
continuously sends the location data to the monitoring unit.
When the vehicle is stolen, the location data sent by tracking unit can be used to find the
location and coordinates can be sent to police for further action. Some Vehicle tracking System
can even detect unauthorized movements of the vehicle and then alert the owner. This gives an
edge over other pieces of technology for the same purpose
Monitoring center Software helps the vehicle owner with a view of the location at which
the vehicle stands. Browsing is easy and the owners can make use of any browser and connect to
the monitoring center software, to find and track his vehicle. This in turn saves a lot of effort to
find the vehicle's position by replacing the manual call to the driver.
As we have seen the vehicle tracking system is an exciting piece of technology for vehicle
security. It enables the owner to virtually keep an eye on his vehicle any time and from anywhere
in the world.
It was the shipping industry that first started using vehicle tracking systems. When large
fleet of vehicles were spread out over the vast expanses of ocean, the owner corporations often
found it difficult to keep track of what was happening. As the saying goes necessity is the mother
of invention; thus GPS tracking systems which originally were used in military operations found
their application here. The commercial application grew poular very soon and these days
consumer vehicles of all sorts use tracking systems as devices to prevent theft and enhance
retrieval.
Applications 5.13
The Technology
GPS (Global Positioning System) is the technology most commonly used for vehicle
tracking these days. There are also other variants of AVL (Automatic Vehicle Location) that
enable easy location of vehicles. The GPS modules with their satellite linked positioning
technique make easy and accurate location of the vehicle possible. The information can be viewed
on electronic maps that are connected to the Internet or otherwise supported by specialized
software. Advanced GPS modules may also have cellular or satellite transmitters that
communicate with remote users apart from the central station from where the tracking is done.
As we saw earlier, the GPS system uses satellite signals. These systems were originally
developed by the government for defense purposes. The satellite part is thus available to civilians
and commercial users free of cost. All the user needs to do is install the appropriate devices for
sending out and receiving signals. This makes GPS an inexpensive technology.
The other AVL systems like Loran and LoJack are terrestrial based and use radio
frequency (RF) transmitters. RF transmitters send out powerful signals that can pass through
walls, garages and other indoor barriers. Terrestrial or otherwise, most of these do not need
antenna to be in direct line of sight with the satellite. This is a major advantage of the
technology‘s progress.
These systems are also used in consumer vehicles as devices for preventing theft and
retrieving stolen/lost vehicles. The signal sent out by the installed device help the police to track
the vehicle. These tracking systems can be used as an alternative for traditional car alams or in
combination with it. Installing tracking systems can thus bring down the insurance costs for your
vehicle by reducing the risk factor.
Vehicle Tracking systems often have several alternatives, like sending automatic alerts to
a phone or email if the vehicle is moved without due authorization. They can also work as one
layer of several combined security measures.
Apart from security concerns, the tracking systems can also help users such as taxi
services to improve their customer service. The systems enable the operators to identify the empty
taxis and direct the nearest one to pick up the customer.
5.14 Geographic Information System
Vehicle tracking systems can also be applied for monitoring driving behavior for both
commercial and individual situations. Parents for instance can use tracking devices to keep an eye
on their teenage son‘s driving.
Tracking and Recovery of Stolen Vehicles: GPS, or better still RF, can help to trace and
recover stolen vehicles. LoJack devices can even be activated remotely by the police
System:
The system consists of modern hardware and software components enabling one to track
their vehicle online or offline. Any vehicle tracking system consists of mainly three parts mobile
vehicle unit, fixed based station and, database and software system.
• Vehicle Unit: It is the hardware component attached to the vehicle having either a
GPS/GSM modem. The unit is configured around a primary modem that functions
with the tracking software by receiving signals from GPS satellites or radio station
points with the help of antenna. The controller modem converts the data and sends the
vehicle location data to the server.
• Fixed Based Station: Consists of a wireless network to receive and forward the data
to the data center. Base stations are equipped with tracking software and geographic
map useful for determining the vehicle location. Maps of every city and landmarks are
available in the based station that has an in-built Web Server.
• Database and Software: The position information or the coordinates of each visiting
points are stored in a database, which later can be viewed in a display screen using
digital maps. However, the users have to connect themselves to the web server with
the respective vehicle ID stored in the database and only then s/he can view the
location of vehicle traveled.
Applications 5.15
Summation:
No doubt, Vehicle tracking system whether it is GPS based or any other wireless medium
has brought one of the most important technological advances in today‘s communication field.
Now one doesn‘t have to leave a place to know where a particular vehicle is at a given period of
time. The automatic vehicle locating system only with the help of a tinny electronic device and
tracking software can detect the real-time location of a vehicle by using the conventional cell
phone network and Internet.
It is also used as a defense mechanism which protects the vehicles such as construction
equipments, trucks, trailers etc. from theft. This is done with the help of some electronic
equipment attached to the vehicles so that the location of the missing vehicle can be identified as
early as possible.
Mainly two types of tracking are described: Active tracking and Passive tracking.
• Active tracking works on real time basis that is after collecting information about
location, speed, and destination, the device sends it to data centre. Radio frequency is
active technology.
5.16 Geographic Information System
• Passive tracking works as long as vehicle is on destination. It stores GPS data that can
be later downloaded to the computer. Here different tracking software is available for
storing and downloading data.
Benefits:
This technology is best used by fleet operators for better management. Everything routing,
dispatch, onboard information, security is managed efficiently. Driving behaviour can also be
known by this device.
Technology holds good for consumers‘ vehicles also. Device also makes the task of police
to recover the stolen vehicles easily. It also works as alert system and raise an alarm when vehicle
is removed from the place without authority. Information regarding its location, status and daily
reports can be taken on real time basis from the vehicle itself.
Field of uses:
Technology can be used for different purposes including service management, fleet
management, stolen vehicle recovery, field sales and trailer tracking.
The devices with various facilities help to keep a watch and control on them. You just
need to attach a small tracking device with your vehicle and the service providers do the rest with
tracking software that provides timely details of the vehicle. Some of the basic features are given
below.
Function:
Device identifies the location of the vehicle on a real time basis. Just through internet you
can easily track your exact position of the vehicle.
Whether passive or active the tracking devices store all location and speed details that you
can collect through your PC by using internet and a special software made for this purpose. Even
one can set the speed limitation in the device attached. If somebody drives the car beyond the
specified speed limit, it will alert you immediately.
Device helps not only to keep limit on its speed but also on its area. Similarly you can
create safety zones so that the moment your vehicle goes out of the designated area, the device
alerts you through SMS, email or by using any other wireless media.
Best feature of the system is its easiness with which it can be used. From anywhere one
can get the information about that particular vehicle by logging in the information required.
Details about the car are displayed within seconds.
Applications 5.17
Some devices work even in a smarter way by making a chart of the entire activities of the
vehicle over a period of time that you can analyse later on. Hence, these tracking devices have
really revolutionised the communication medium by providing various cool features at your
fingertips.
For tracing the time lag between the time code signal sent and the time it was actually
received on earth GPS makes use of theory of relativity propounded by Einstein and four signals
to provide accuracy.
Once the technology was confined to military domain, today it became a $15 billion
industry. Mostly GPS is being used by the public transportation agencies as they generally have to
deal with larger area. For example school buses in cities to keep the track of the buses to make
safety of the passengers sure. GPS gives drivers information about the other vehicle so that on the
basis of information received may align its own course avoiding traffic congestion. Thus keeping
passengers safe and also saving time.
This also gives parents peace of mind as they find that vehicle carrying their children are
regularly being monitored or kept watch.
In India, a fleet of one fifty GPS enabled metered taxis were launched in Mumbai. This
enhanced fleet provided better customer services as they now were able to book a taxi in advance
just by few strokes on internet or phone and call taxi at a preferred location. This added one more
to the list of benefits to the customer that now customer will be charged appropriately as all the
data regarding mileage will be sent back to the control centre. This also ensures the safety of the
passenger as there are reduced chances of kidnapping or harming the passenger by virtue of
monitoring of the vehicle.
In companies device is used to track especially supply trucks and trucks used by the
contractors going to job sites. This permits real time tracking of the mobile asset. This also
abridges the gap between the actual mileage and the bill paid for it. Transparency naturally comes
in the system.
Thus tracking of vehicles reduces the cost of labour, fuel and culminates into the efficient
service delivery system.
5.6. MARKETING
GIS provides more in-depth analysis, assessment, and understanding of your marketing
and promotions. Visualize where your customers are located by analyzing demographic,
psychographic, purchasing, and spending characteristics for accurate customer segmentation and
helping you find more like them. Use market potential to your advantage so you can optimize
your store network and maximize sales from every outlet by improving marketing outreach,
campaign performance, and brand interaction. Location analytics is a key element to the
5.18 Geographic Information System
successful development, evolution, and growth of all aspects of marketing, from social media to
display advertising, coupon optimization, and sentiment analysis.
Success stories
These stories show how business organizations are gaining benefits using GIS technology
and spatial analysis.
Banking:
Being market driven banks need to provide customer centric services around planning of
resources and marketing. GIS plays an important role providing planning, organizing and decision
making.
Assets Management:
GIS helps organizations to locate and store information about their assets. Operations and
maintenance staff can also deploy their enterprise and mobile workforce.
Dairy Industry:
Geographic Information Systems are used in the distribution of products, production rate,
location of shops and their selling rate. These can also be monitored by using a GIS system.
Tourism:
Tourists can get all the information they need on a click, measuring distance, finding
hotels, restaurants and even navigate to their respective links. This Information plays a vital role
to tourists in planning their travel from one place to another.
Business:
GIS is used for managing business information based on its location. GIS can keep track
of where customers are located, site business, target marketing campaigns, and optimize sales
territories and model retail spending patterns.
Market Share:
Examining branch locations, competitor locations and demographic characteristics to
identify areas worthy of expansion or determine market share in Maptitude.
Applications 5.19
ATM Machine:
Filling in market and service gaps by understanding where customers, facilities, and
competitors are with address locating, database management and query tools.
Mobile Banking:
Capturing locations where existing mobile transaction occur and assisting in mobile
security infrastructure.
Internet of Things:
Improving efficiency, accuracy and economic benefit through a network of physical
objects such as devices, vehicles, buildings and other items—embedded with electronics,
software, sensors, and network connectivity that enables these objects to collect and exchange
information with one another.
Remote Oversight:
Data is available on the desktop in real-time, subject to a suitable network connection in
the field, enabling decision makers to manage their time more effectively.
Data Security:
Video and pipe-wall condition data is held securely to prevent unauthorised access,
ensuring data quality and the audit trail are maintained. Only authorised users can view the data,
accessing the system using a secure user name and password.
High Availability:
Private cloud hosting ensures data is available when required. Data is backed-up to a
separate storage location to ensure business continuity in the unlikely event of data loss on the live
service.
Extensible Service:
The service is designed to allow new data sources to be incorporated as and when they are
required, such as on the introduction of new network inspection technologies.
5.22 Geographic Information System
5.7.3. Benefits to Utility Companies
An increasing number of pipe condition assessment and camera systems provide asset
management information in a variety of formats. The SCISYS Network Intelligence Portal will
combine all of the data to provide a comprehensive view of this information, enabling focused and
astute solution design and delivery.
By combining the above data, asset managers and capital planners are able to take more
effective investment decisions, resulting in better service outcomes for their customers. This will
provide a robust, long term and sustainable solution that delivers a more targeted solution with
lower CAPEX, whilst driving progress towards water efficiency and leakage reduction targets.
5.7.4. Digiright
Digiright is a solution from SCISYS that allows users to quickly and efficiently capture
and edit real world features (point, line and polygon) through their web browser. Digiright can be
deployed on premise or provided as a fully managed cloud service, including the provision of base
mapping from Ordnance Survey or other mapping providers.
Digiright has integrated security and is scalable. It will integrate with mainstream
Geographic Information data feeds as well as a customer‘s own spatial data.
Functional Overview
• Easily customised and extensible
• Lightweight and simple to deploy
• Easily integrated
• Inbuilt validation rules
Applications 5.23
Digiright is ideal for
• Capturing point, line and polygon data
• Editing in the office or the field
• Maintaining clean geometries
• snapping to existing features, guide lines or features from other data sources
• Overlaying raster / vector data
• Dragging Geometries
5.7.5. Cartosys
Cartosys delivers all of the essential tools an organisation needs to successfully deploy
location-based services.
The Cartosys toolkit, built on open standards, integrates proven and supported open source
tools into a single out-of-the-box solution that is quick and easy to deploy, configure and manage,
without the need for software coding.
• Built on open standards with an open API.
• Integrates with your existing open source and COTS products.
• Provides a rapid, cost effective way to deliver your geospatial needs.
Background
Ash dieback is a tree disease that has hit the headlines in recent years, threatening to
almost completely destroy Britain‘s estimated 130 million ash trees, but it‘s not the only threat to
native tree species. An increasing number of invasive pests and diseases threaten British trees. In
response, the Forestry Commission needed to have efficient processes to address the threats in a
proactive and consistent way.
In July 2013, EU LIFE+ approved funding for the 4 year Observatree project lead by the
Forestry Commission. As part of this project the opportunity was seized to take advantage of an
engaged public as ‗citizen scientists‘ as well as the 25,000 or so people who work in the arboreal
industry. What was needed was an easy to use application that would allow them to make
scientifically useful reports about their observations.
5.7.7. Solution
With limited time and budget available to the project team, an innovative and efficient
approach was required. Any solution needed to be ready for deployment prior to the start of the
next growing season, when most of the reports are made.
To achieve this goal the Forestry Commission engaged SCISYS to help them design and
deliver Tree Alert. Tree Alert is a browser based application that guides users through a series of
intelligent questions, which reflect previous answers, to help to identify the issue, including
confirmation of the species as well as the key symptoms.
Using an Open Architecture combining a number of Open Source and existing SCISYS
technologies, the project team were able to take advantage of functionality already available. The
basis of the site uses the content management system Joomla to provide an application framework.
This has been integrated with the SCISYS Digiright component to capture the location
information about the observation.
Applications 5.25
Benefits
Tree Alert has significantly increased the reach of the Forestry Commission in collecting
observation about potential tree diseases and pests. Commenting on the solution Roger Coppock,
Head of Corporate and Forestry Support, Forestry Commission said ―The information gathered on
the Tree Alert site doesn‘t just allow us to investigate individual cases reported by the general
public, as we can now set priorities based on risk and track the incidence of new and invasive
pests and diseases. This information will prove invaluable in our understanding of how tree pests
and diseases spread in Britain, and help us to protect our woodlands for generations to come.‖
Benefits
The benefits to Natural Resources Wales include:
• A data comparison tool that can be used to detect change within the flood zones. The
tool compares newly generated data to previous polygons to assist in the preparation
of data for delivery into external data products.
• Improved data management procedures and processes which are aligned to and
integrated with their data management lifecycle workflow.
• Gemini 2.1 INSPIRE compliant metadata integrated into the data management
lifecycle.
5.26 Geographic Information System
• A set of recommendations and an associated roadmap to enable Natural Resources
Wales to streamline and develop its fly-tipping photograph geo-tagging process.
Background
In undertakings these operations the Authority collects, uses and maintains a vast array of
spatial data in some cases dating back hundreds of years. TCA were using a wide variety of end-
of-life point solutions to deliver services from islands of data across their enterprise. Apart from
the considerable financial implications of maintaining this position it was identified as limiting the
business as it was unable to adapt to changing market conditions and predicted volumes of
business. Corporate knowledge was ―siloed‖ making it time consuming to provide the highest
quality safety advice.
Solution
TCA commissioned SCISYS to supply an enterprise information service: known as
"Inferis", this facilitates smarter working via corporate data sharing and service re-use. Services
supplied span the entire software life cycle from analysis to support. A centralised spatially
enabled database serves all functions. SCISYS worked closely with TCA to understand their
corporate business processes and policies, considering each process in isolation. These were then
translated into the minimum number of generic capabilities and provided as re-useable services,
including the innovative approach of embedding suitably configured mapping (GIS) services
within workflow steps. Previously isolated pools of data are now accessible via the re-useable
services from a spatially enabled database. This significantly empowers the business as all
information and opinion on a site is available in one system: making the most of TCA assets. The
same services are re-used in all aspects of the business, from web based customer self service to
forensic geospatial analysis of corporate data by TCA specialists.
Inferis uses a hybrid architecture combining Esri components for back-end data processing
combined with open source components (Open Layers and Geo-server) to visualise and capture
data via the internal corporate GIS browser, cartosys, and external feature capture tool (the
SCISYS Digiright component).
"Up to date and accurate operational information is key to every decision we make in the
business and to our customers who rely on our coal mining report when buying and selling
property. With Inferis, for the first time, we have a single integrated system which provides the
solutions we need across the business and access to the most up to date information including
Ordnance Survey products." Says Steve Pennell, The Coal Authority Director of Mining
Information and Services.
Applications 5.27
Benefits
• All spatial data assets are now available across the whole enterprise
• 100% increase in mining report generation productivity
• 98% of reports ordered and returned electronically via self service
• >70% of property boundaries digitised by the customer
Commenting on the delivery Steve Pennell, The Coal Authority Director of Mining
Information and Services said "Up to date and accurate operational information is key to every
decision we make in the business and to our customers who rely on our coal mining report when
buying and selling property. With Inferis, for the first time, we have a single integrated system
which provides the solutions we need across the business and access to the most up to date
information including Ordnance Survey products."
5.28 Geographic Information System
PART–B (2 × 13 = 26 MARKS)
11. (a) Describe in detail, the Basic spatial concepts. Add a note on introduction to GIS.
(or)
(b) Write an essay on attribute data model.
App-A.2 Geographic Information System
12. (a) Compare and contrast raster and vector data representation with suitable examples.
(or)
(b) Explain Triangular Irregular Network (TIN)
13. (a) Briefly explain the Scanners for Raster Data Input in GIS
(or)
(b) Briefly explain the Open Database Connectivity (ODBC)
(or)
14. (a) Explain
a) Buffering techniques
b) Overlaying techniques
c) Feature Manipulation
(or)
(b) Explain the methods of modeling in GIS highway alignment studies
15. (a) Briefly Application of GIS in forestry
(or)
(b) Briefly explain applications of GIS case studies
PART-C (1 × 16 = 16 MARKS)
16. (a) Write detailed notes on Raster Data Compression.
(or)
(b) Briefly explain the Data Analysis tools.
Model Question Papers App-A.3
MODEL QUESTION PAPERS - II
B.E/B.Tech. DEGREE MODEL EXAMINATION
Third Semester
Civil Engineering
OCE552 GEOGRAPHIC INFORMATION SYSTEM
(Regulations 2013)
Answer ALL questions
PART-A (10 × 2 = 20 Marks)
1. What is the difference between datum and ellipsoid?
2. What is GIS datum?
3. Distinguish between spatial & non spatial data.
4. Explain TIN.
5. What is overlying?
6. Write short notes on topographical in GIS
7. What is buffering?
8. Where does GIS fit within BIM?
9. How is GIS used in agriculture?
10. Significance of SLD EDITOR in GIS
PART–B (2 × 13 = 26 MARKS)
16. (a) What is data model? Add notes on different types of data
(or)
(b) Describe the GPS, Uses of GPS, structure of GPS and GPS positioning etc.