High precision - can represent exact locations and complex geometric shapes accurately. Ideal for detailed mapping.
Lower precision - limited by pixel size, with larger cells leading to coarser resolution. Precision depends on grid size and resolution.
Ideal for spatial queries, overlay analysis, proximity analysis, network analysis, and operations where precision in location or shape is critical.
Suitable for spatial modeling, surface analysis (e.g., terrain modeling), and when analyzing data that varies continuously across a region.
Used for mapping infrastructure, land parcels, roads, administrative boundaries, and point features like schools or wells.
Widely used for remote sensing, climate modeling, land use classification, and any analysis requiring continuous surface data.
The electromagnetic (EM) spectrum is the range of all types of EM radiation. It consists of oscillating electric and magnetic fields, perpendicular to each other, traveling through space. The electric field oscillates along one axis (X), while the magnetic field oscillates along another perpendicular axis (Y).
The relationship between frequency (f) and wavelength (λ) is inversely proportional, described by the equation:
c = λ × f
Where: c is the speed of EM wave (approximately 3×10^8 m/s), λ is the wavelength, and f is the frequency.
As wavelength increases, frequency decreases, and vice versa.
UV radiation (major part) is mostly absorbed by Earth’s atmosphere, specifically the ozone layer, which limits its effectiveness for remote sensing.
These high-energy wavelengths are absorbed by both the atmosphere and Earth’s surface. It also have destructive effects making them impractical for environmental remote sensing.
A small part of UV escapes atmosphere used in coastal aerosol, Visible light is used to generate natural color images, Infrared used for Near and thermal IR images in multispectral and hyperspectral imaging
Used in radar and synthetic aperture radar (SAR) for all-weather imaging
Different materials reflect and absorb different wavelengths of electromagnetic radiation.
The reflected wavelengths detected by a sensor and determine the type of material it reflected from. This is known as a spectral signature .
Spectral signatures are crucial in remote sensing because they allow us to differentiate between various materials or surfaces, such as water, vegetation, soil, or urban areas.
By analyzing the spectral signature of an object, we can identify its composition and condition, making it useful for applications like land cover classification, vegetation health monitoring, and mineral exploration.
Have you ever wondered how the bands in multispectral or hyperspectral images are decided? The answer lies in spectral signatures.
In multispectral imaging limited number of broad bands used and in hyperspectral imaging captures hundreds of narrow bands across the spectrum. This band selection is based on the spectral characteristics of objects, allowing for more accurate identification and analysis.
Surface characteristics such as texture, moisture content, and physical structure of an object.
Atmospheric Conditions like the presence of clouds, humidity, and atmospheric particles.
Sun Angle and Illumination affects the amount and direction of light reflected.
Sensor Characteristics such as resolution, sensitivity, and wavelength range.
Time of Day and Season like seasonal variations (e.g., vegetation growth) and time of day (lighting conditions).
Spatial resolution in raster data refers to the size of each pixel or cell in the grid that makes up the raster image. It represents the area on the ground covered by a single pixel.
For example, a spatial resolution of 10 meters means each pixel represents a 10x10 meter area on the Earth’s surface.
In this graph, it compare the relationship between area covered in ground and spatial resolutions.
Higher Resolution (smaller pixel size): Provides more detail, as smaller features on the ground can be captured. However, this also leads to larger file sizes and increased processing time.
Lower Resolution (larger pixel size): Generalizes the data over a larger area, capturing less detail. This is suitable for large-scale, regional studies where fine details are not necessary.
Purpose of the Analysis: The type of analysis you are conducting will determine the necessary level of detail. For example, urban planning or infrastructure mapping requires high resolution, while large-scale environmental studies may work with lower resolution.
Size of the Study Area: A large study area analyzed at high resolution can result in very large datasets, requiring more storage and computational power. Balancing resolution with the size of the area is crucial.
Data Availability and Costs: High-resolution data may be expensive or difficult to acquire, while lower-resolution datasets (like those from satellites) are often more readily available and free of charge.
Processing Power and Time: Higher resolution data requires more time and resources to process. It’s important to choose a resolution that your hardware and software can handle efficiently.
Indian Sensors (ISRO): Cartosat-3 Panchromatic: 0.25 m Resourcesat-2 LISS-IV (Multispectral): 5.8 m RISAT-2B SAR (X-band): 1-25 m
US Sensors (NASA/USGS): Landsat-8/9 (Multispectral R, G, B, NIR, SWIR): 30 m MODIS (Terra/Aqua R, G, B): 250 m VIIRS (Visible Infrared Imaging Radiometer Suite, R, G, B): 375 m
European Sensors (ESA): Sentinel-1 SAR (C-band): 10 m Sentinel-2 Multispectral (R, G, B, NIR): 10 m Sentinel-3 OLCI (Ocean and Land Color Instrument): 300 m
Temporal resolution refers to how frequently a satellite or sensor captures data of the same location over time. It is the interval between consecutive observations of the same area.
For example, if a satellite revisits the same location every 5 days, its temporal resolution is 5 days.
Due to its much wider imaging swath, MODIS provides global coverage every 1-2 days versus 16 days for the Landsat OLI.
Red dots - Center point of each Landsat data Blue boxes - Orbital swath of MODIS
Higher Temporal Resolution (frequent revisits): Enables more accurate monitoring of dynamic changes over time, such as vegetation growth, weather patterns, or urban development. It is crucial for time-sensitive analyses like disaster monitoring (e.g., floods, wildfires) or agricultural assessment.
Lower Temporal Resolution (longer intervals): Reduces the ability to capture short-term changes, making it less useful for real-time monitoring. However, it may be sufficient for slow-moving processes like long-term land use change or deforestation.
Purpose of Analysis: For fast-changing phenomena (e.g., weather events or crop growth), high temporal resolution is essential to capture accurate trends. For slower processes (e.g., yearly land cover changes), lower temporal resolution may suffice.
Sensor Availability: Not all sensors have the same revisit frequency. Some satellites, like geostationary satellites, provide continuous coverage, while others revisit only every few days or weeks.
Data Volume and Processing: Higher temporal resolution generates more data over short periods, increasing storage and processing requirements. For large-scale or long-term projects, balancing temporal resolution with data management capabilities is crucial.
Cost and Accessibility: High-temporal-resolution data (e.g., daily observations) may come at a higher cost or be harder to obtain. Lower temporal resolution data may be more accessible and sufficient for certain analyses.
Spectral resolution refers to the ability of a sensor to detect and differentiate between different wavelengths of electromagnetic radiation.
Higher spectral resolution means the sensor can detect finer differences between closely spaced wavelengths, while lower spectral resolution captures broader, less distinct bands.
The comparison between the number and width of the spectral bands captured by the sensor.
Multispectral: 3-10 wider bands. Covers Large Area
Hyperspectral: Hundreds of narrow bands. Covers Small Area
Spectral resolution plays a key role in distinguishing different surface materials because each material reflects or absorbs light differently at various wavelengths, creating a unique spectral signature.
A sensor with high spectral resolution can capture more detailed information across many narrow bands, making it easier to differentiate between materials with similar reflectance properties, such as various types of vegetation or minerals.
In contrast, low spectral resolution may combine wavelengths into broad bands, potentially missing subtle differences, which can lead to less accurate identification of materials.
For example, high spectral resolution is critical in applications like vegetation classification or mineral exploration, where precise differentiation between similar surface types is required.
Radiometric resolution is the amount of information in each pixel, that is, the number of bits representing the energy recorded. Each bit records an exponent of power 2.
For example, an 8 bit resolution is 2^8, which indicates that the sensor has 256 potential digital values (0-255) to store information.
The higher the radiometric resolution, the more values are available to store information, providing better discrimination between even the slightest differences in energy.
For example, when assessing water quality, radiometric resolution is necessary to distinguish between subtle differences in ocean color.
India: 8-bit, 10-bit, 12-bit (IRS, Cartosat)
US: 8-bit, 12-bit, 16-bit (Landsat, MODIS)
Europe: 8-bit, 10-bit, 12-bit, 16-bit (Sentinel)
Combining all types of resolutions-spatial, spectral, temporal, and radiometric into a single system is challenging due to technical and practical trade-offs.
Each resolution requires specific sensor configurations, and optimizing one often comes at the expense of others.
Spatial vs. Temporal: More detail requires longer revisit intervals.
Spectral vs. Spatial: Finer spectral data reduces spatial detail.
Radiometric vs. Temporal/Spectral: More frequent or spectrally rich data usually requires less intensity precision.
Spatial resolution - A sensor with a small Instantaneous Field of View (IFOV) has a higher spatial resolution, but it can’t detect as much energy.
Radiometric resolution - To increase the amount of energy detected, the radiometric sensitivity can be improved, but broadening the wavelength range detected reduces spectral resolution.
Temporal resolution - narrower swath is required for higher spatial resolution, which means more time between observations of the same area, resulting in lower temporal resolution
For example,
When researching weather, which changes over time, a high temporal resolution is critical.
When researching seasonal vegetation changes, a higher spectral or spatial resolution may be more important than a high temporal resolution.
Passive sensors detect energy emitted or reflected from an object, and include different types of radiometers and spectrometers.
Most passive systems used in remote sensing applications operate in the visible, infrared, and thermal infrared portions of the electromagnetic spectrum.
The passive sensor uses energy only for movement of sensors, sending and receiving the data and does not use energy for capturing the data.
Accelerometer- An instrument that measures acceleration (change in velocity per unit time). There are two general types of accelerometers. One measures translational accelerations (changes in linear motions in one or more dimensions), and the other measures angular accelerations (changes in rotation rate per unit time).
Sounder- An instrument that measures vertical distributions of atmospheric parameters such as temperature, pressure, and composition from multispectral information.
Spectrometer- A device that is designed to detect, measure, and analyze the spectral content of incident electromagnetic radiation. Conventional imaging spectrometers use gratings or prisms to disperse the radiation for spectral discrimination.
Radiometer- An instrument that quantitatively measures the intensity of electromagnetic radiation in some bands within the spectrum. Usually, a radiometer is further identified by the portion of the spectrum it covers; for example, visible, infrared, or microwave.
Spectroradiometer- A radiometer that measures the intensity of radiation in multiple wavelength bands (i.e., multispectral). Many times the bands are of high-spectral resolution, designed for remotely sensing specific geophysical parameters
Imaging radiometer- A radiometer that has a scanning capability to provide a two-dimensional array of pixels from which an image may be produced. Scanning can be performed mechanically or electronically by using an array of detectors.
Hyperspectral radiometer- An advanced multispectral sensor that detects hundreds of very narrow spectral bands throughout the visible, near-infrared, and mid-infrared portions of the electromagnetic spectrum. This sensor’s very high spectral resolution facilitates fine discrimination between different targets based on their spectral response in each of the narrow bands.
Active sensors send out a pulse of energy and detect the changes in the return signal.
Most active sensors operate in the microwave portion of the electromagnetic spectrum, which makes them able to penetrate the atmosphere under most conditions.
The active sensor uses energy also for capturing the data apart from the energy used for movement of sensors, receiving the commands and sending the data.
Laser altimeter- An instrument that uses lidar to measure the height of the platform (spacecraft or aircraft) above the surface. The height of the platform with respect to the mean Earth’s surface is used to determine the topography of the underlying surface.
LiDAR- A Light Detection and Ranging sensor that uses a laser (light amplification by stimulated emission of radiation) radar to transmit a light pulse and a receiver with sensitive detectors to measure the backscattered or reflected light. Distance to the object is determined by recording the time between transmitted and backscattered pulses and by using the speed of light to calculate the distance traveled.
Radar- An active radio detection and ranging sensor that provides its own source of electromagnetic energy. An active radar sensor, whether airborne or spaceborne, emits microwave radiation in a series of pulses from an antenna. When the energy reaches the target, some of the energy is reflected back toward the sensor. This backscattered microwave radiation is detected, measured, and timed. The time required for the energy to travel to the target and return back to the sensor determines the distance or range to the target. By recording the range and magnitude of the energy reflected from all targets as the system passes by, a two-dimensional image of the surface can be produced. Synthetic aperture radar (SAR) is a radar technique.
Ranging Instrument- A device that measures the distance between the instrument and a target object. Radars and altimeters work by determining the time a transmitted pulse (microwaves or light) takes to reflect from a target and return to the instrument.
Another technique employs identical microwave instruments on a pair of platforms. Signals are transmitted from each instrument to the other, with the distance between the two determined from the difference between the received signal phase and transmitted (reference) phase. These are examples of active techniques.
An active technique views the target from either end of a baseline of known length. The change in apparent view direction (parallax) is related to the absolute distance between the instrument and target.
Scatterometer- A high-frequency microwave radar designed specifically to measure backscattered radiation. Over ocean surfaces, measurements of backscattered radiation in the microwave spectral region can be used to derive maps of surface wind speed and direction.
Sounder- An instrument that measures vertical distribution of precipitation and other atmospheric characteristics such as temperature, humidity, and cloud composition.
Designing optimal bands for passive sensors involves understanding the electromagnetic spectrum and how different wavelengths interact with various materials.
By analyzing spectral signatures, we can determine which wavelengths are most effective for detecting specific features or phenomena on the Earth’s surface.
This process requires selecting bands that maximize the sensor’s ability to distinguish between different surface types while ensuring sufficient coverage of relevant spectral regions.
Band Selection: Each band should be chosen based on the specific applications and the unique spectral signatures of the materials being studied.
Resolution: The width of each band affects the sensor’s ability to discriminate between similar materials; narrower bands can provide more detail but may increase complexity.
Coverage: Ensuring that the selected bands cover critical wavelengths for the target applications is essential for effective data capture.
Designing optimal bands for active sensors, like radar systems, involves selecting frequency ranges that suit specific applications based on how different wavelengths interact with target surfaces.
Since, active sensors emit their own signals and measure the return signal, the choice of frequency and wavelength critical for achieving the desired results.
Target Interaction: Higher frequencies (like X or Ku bands) have lower penetration and are suited for applications requiring surface detail (urban areas, ice). Lower frequencies (like L or P bands) penetrate deeper, making them ideal for biomass or subsurface monitoring.
Coherence and Penetration: Active sensors in higher frequencies tend to have lower coherence in vegetated areas, while lower-frequency bands maintain signal stability and allow deeper penetration, useful in dense vegetation or soil.
Application-Specific Band Selection: The choice of band is driven by the research focus, such as agriculture monitoring, urban infrastructure analysis, or ice sheet dynamics.
GNSS stands for global navigation satellite system.
A Global Navigation Satellite System (GNSS) consists of a constellation of satellites orbiting the Earth in very specific trajectories.
For global coverage, it is estimated that a constellation requires 18 to 30 satellites.
GNSS is often generically referred to as GPS (Global Positioning System) but that acronym actually refers specifically to the United States constellation.
There are several GNSS constellations provided by governments around the world.
RNSS stands for regional navigation satellite system.
RNSS systems are engineered to service specific regions only,
rather than offering a global service.
RNSS Systems
Navigation satellites provide orbit information and accurate timing (and other services) to radio receivers specifically designed to receive those satellite signals and decode the signal message contents.
With the contents of the messages from at least four “visible” satellites, the position on or near most of the Earth’s surface can be calculated using a mathematical process known as trilateration.
Image depicting how trilateration works. The GNSS receiver now has a three-dimensional position fix; that is, X-Y coordinates plus altitude/elevation (Z). The more satellites that can be seen, the easier it is to resolve position with enhanced precision.
The location of the GNSS receiver is somewhere on the perimeter of the first satellite coverage area.
When a second satellite can be seen, the location must be at one of the two points where the perimeters intersect.
When a third satellite is visible, the intersection of all three coverage perimeters is the two-dimensional location and the remaining point of intersection is ignored.
A fourth visible satellite enables elevation/altitude to be calculated.
There are two primary uses for GNSS:
The GNSS receiver can perform trilateration and provide accurate position only if it knows: Where the satellite is, - Exactly when the signal was sent from the satellite, and - Exactly what time the signal is received.
The signals sent from the satellites in space to the GNSS receivers are complex and vary in structure and frequency.
Different frequencies are used to improve signal reliability, signal accuracy, and system redundancy.
Some signal frequencies are better suited to pass through trees due to the different signal wavelengths.
In addition, by using multiple frequencies simultaneously, modern multi-frequency GNSS receivers can improve the position accuracy by measuring the difference in signal propagation through the atmosphere, and effectively remove it as a source of error.
An orbit is the curved path that an object in space (such as a star, planet, moon, asteroid or spacecraft) takes around another object due to gravity.
Gravity causes objects in space that have mass to be attracted to other nearby objects. If this attraction brings them together with enough momentum, they can sometimes begin to orbit each other.
Geostationary Transfer Orbit (GTO) • Orbit Time: Varies (transfer phase) • Altitude: From low Earth orbit to 35,786 km (geostationary) • Property: An elliptical orbit used to transfer satellites from low orbit to geostationary orbit by gradually adjusting altitude.
There are a number of factors that limit (or potentially limit) the use of GNSS. The primary factor is the imperfect reception of the signal as received by the GNSS receiver from the individual satellites.
Malicious radio interference, known as spoofing or jamming, of GNSS to corrupt the signal so that it cannot be received coherently happens – this requires a strategic approach to overcome that is outside the scope of this article.
Range: ± 5 m Source: Signal propagation delay – upper atmosphere is loaded with electrons caused by ionizing solar radiation that can “bend” and reflect radio waves.
Range: ± 2.5 m Source: Position drift – as with clocks, miniscule errors
in satellite orbit position become much larger when used for position
calculation on Earth.
Range: ± 2 m Source: Timing drift – due to the distances, tiny timing errors in satellite clock accuracy become much larger errors on Earth.
Range: ± 1 m Source: Signal replication due to reflection off objects such as buildings and terrain.
Range: ± 0.5 m Source: Signal propagation delay – lower atmosphere is far denser than other atmospheric layers and can refract radio waves.
Range: ± 0.3 m Source: GNSS receiver hardware and software induced signal noise that affects accuracy of perceived signal.
The goal of base stations and GNSS error correction services is to establish the true path of the GNSS receiver, or as close to its true path as possible, in relation to an absolute position on the Earth’s surface.
The base station receives GNSS signals and uses sophisticated measurement techniques to precisely calculate distances to observable satellites and thus to calculate GNSS signal errors.
All measurement and error correction data made at each base station is logged and archived.
This data is used for various error correction solutions, some of which are outlined as follows.
-Satellite-based augmentation system is a network of ground reference stations that provide satellite clock, ephemeris and signal propagation corrections via geostationary satellites, based on satellite observation from multiple reference locations.
Accuracy Refers to the closeness of a measurement or representation to its actual value in the real world.
In simpler terms, it signifies how correct the data is.
For example, if a map claims a lake is located at a specific coordinate, the accuracy reflects how close that coordinate is to the lake’s true location on Earth.
Precision presents the level of detail or reproducibility of a measurement.
It essentially tells you how consistent the measurements are.
In GIS, precision often relates to the resolution of the data. High-resolution data, like aerial imagery collected by drones, allows for more precise measurements and captures finer details compared to low-resolution data.
• In this case, a GIS system might generate points that are far off from their true location and scattered widely. For example, if you’re mapping a city’s roads, but the coordinates are both incorrectly located (inaccurate) and randomly scattered (imprecise), it fails in both aspects.
The core components of GIS, are • Hardware • Software • Data • People • Method
Hardware, which includes physical devices like computers and GPS units needed to run GIS applications.
Software, such as QGIS, provides tools for spatial analysis, mapping, and data management.
Data is the core, comprising spatial data (geographic coordinates) and attribute data (details about geographic features), stored in vector or raster formats.
People are the users who interpret GIS data, make decisions, and apply it in fields like urban planning and environmental management.
Methods involve procedures and workflows that ensure consistent and accurate processing of spatial data for analysis.
In GIS, levels of measurement define how spatial data can be categorized, measured, and analyzed.
These include nominal, ordinal, interval, and ratio levels, each determining the type of analysis possible.
Understanding these levels ensures accurate data interpretation and appropriate application of analytical tools.
Nominal data represent categories without a specific order, like land use types or vegetation classes.
In GIS, nominal data are used to classify features (e.g., forests, urban areas), and are typically stored as attribute data in vector layers.
No mathematical operations can be performed on nominal data.
Ordinal data have a ranked order but lack consistent intervals between categories.
Examples include ranking areas by flood risk (low, medium, high).
In GIS, ordinal data allow for simple analysis like sorting, but cannot accurately calculate differences between ranks.
Interval data have ordered categories with known, equal distances between them but no true zero.
Temperature data in degrees Celsius is a common example.
In GIS, interval data enable more detailed analysis, such as identifying spatial patterns, but comparisons of ratios are not possible.
Ratio data have all the properties of interval data, but with a true zero point, allowing for meaningful comparisons of ratios.
Elevation and population density are examples.
In GIS, ratio data enable advanced quantitative analysis, such as calculating growth rates and spatial modeling.
Georeferencing is the process of aligning spatial data (maps, images, or datasets) to a specific coordinate system so that they correspond to real-world locations.
It is useful in GIS for integrating various datasets, ensuring spatial accuracy, and enabling analysis by overlaying different layers, such as satellite imagery and vector maps, with real-world coordinates.
Georeferencing raster data involves using spatial data, such as georeferenced rasters or vector features, to identify control points (known x,y coordinates) on the raster and real-world locations.
Various features, like roads or streams, can serve as control points, which align the raster with target data for spatial accuracy.
Once enough control points are created, various transformations (e.g., polynomial, spline, adjust) can be applied to convert the raster to map coordinates.
A polynomial transformation, optimized for global accuracy, uses a least-squares fitting algorithm and requires a specific number of non-correlated control points based on the transformation order.
Image data usually comprises an array of numbers. Spatial data is similar, but it also includes numerical information (coordinates) that allows us to position it on earth.
These numbers are part of a coordinate system that provides a frame of reference for the data to locate features on the surface of the earth, to align the data relative to other data, to perform spatially accurate analysis, to add and to edit the data, and to create maps.
Data is defined in both horizontal and vertical coordinate systems (also referred as coordinate reference systems).
Horizontal coordinate systems locate data across the surface of the earth
Vertical coordinate systems locate the relative height or depth of data.
Horizontal coordinate systems locate data across Earth’s surface using geographic, projected, or local coordinates.
Geographic systems use angular measurements (latitude and longitude), while projected systems translate these into planar coordinates like meters and local systems apply a false origin for smaller-scale mapping, typically expressed in linear units like meters or feet.
Vertical coordinate systems provide reference points for z-coordinates (height or depth).
They measure elevations relative to mean sea level (gravity-based) or a mathematically derived ellipsoid.
Gravity-based systems are more accurate for terrain, while ellipsoidal systems are simpler but may lack precision for large-scale applications like topography.
Horizontal coordinate systems locate data across the surface of the earth
Horizontal coordinate systems can be of three types: (i)Geographic (ii)Projected (iii)Local
GCS are based on a three-dimensional ellipsoidal or spherical surface, and locations are defined using angular measurements, usually in decimal degrees, measuring degrees of longitude (x-coordinates) and degrees of latitude (y-coordinates).
The location of data is expressed as positive or negative numbers: positive x- and y-values for north of the equator and east of the prime meridian and negative values for south of the equator and west of the prime meridian.
PCS are planar systems that use linear measurements for the coordinates rather than angular units. A PCS is composed of a GCS and a map projection together.
A map projection contains the mathematical calculations that convert the angular geodetic coordinates of the GCS to Cartesian coordinates of the planar projected coordinate system.
LCS uses a false origin (0, 0 or other values) in an arbitrary location anywhere on earth. LCS are often used for large-scale (small area) mapping.
The false origin may or may not be aligned to a known real-world coordinate, but for the purpose of data capture, bearings and distances can be measured using the LCS rather than global coordinates. LCS are usually expressed in meters or feet.
Vertical coordinate systems provide a reference for z-coordinates, which are measurements of the height or depth of features.
They are always measured in linear units such as meters or feet.
Using a vertical coordinate system improves locational accuracy in analysis and editing.
The difference between the center and reference of gravity and ellipsoidal vertical coordinate systems.
Gravity-based vertical coordinate systems are more commonly used.
They reference a mean sea level calculation (or in some cases, derived from the level of a single point).
They can be gravity based only if they cover a full-world extent. EGM2008 Geoid and EGM96 Geoid are examples of global gravity-based vertical coordinate systems.
Ellipsoidal coordinate systems reference a mathematically derived spheroidal or ellipsoidal surface.
Since they are calculated on a mathematical model, ellipsoidal coordinate systems are simpler than gravity-based vertical coordinate systems, but they may lack significant accuracy, especially in large-scale applications.
Mean sea level is used as the zero level for height values and Mean low water is used as the zero level for depth values.
Projections are a mathematical transformation that take spherical coordinates (latitude and longitude) and transform them to an XY (planar) coordinate system.
This enables to create a map that accurately shows distances, areas, or directions.
Left - 3D Globe Right - Projected 2D map
Map projections are necessary to create flat maps from the spherical Earth.
Since we work with 2D surfaces like paper or computer screens, projections allow us to represent large areas, make measurements, and analyze geographic relationships effectively despite distortions.
Different map projections handle the distortions of shape, area, distance, or direction differently.
For example, cylindrical projections like Mercator preserve direction but distort size, while equal-area projections preserve area but may distort shapes. The choice of projection depends on the map’s intended use.
Projection Surface (3 types)
Cylindrical: A projection onto a cylinder, often used for world maps.
Conical: A projection onto a cone, commonly used for mid-latitude regions.
Azimuthal (Planar): A projection onto a flat plane, usually for polar areas or small regions.
Projection Surface (3 types)
Planar - Conical - Cylindrical
Normal: The projection surface is aligned with the Earth’s axis (e.g., for a cylindrical projection, the cylinder touches the equator).
Transverse: The surface is rotated 90 degrees (e.g., a cylinder touching along a meridian).
Oblique: The surface is rotated at an angle between normal and transverse, not aligned with the poles or the equator.
Projection Aspect (3 types)
Normal - Transverse - Oblique
Conformal: Preserves angles, so shapes are correct locally, but area is distorted.
Equal-area: Preserves area, but shapes are distorted.
Equidistant: Preserves distances from certain points or along certain lines.
Azimuthal: Preserves directions from a central point.
Distortion Property (4 types)
Conformal - Equal area Equidistant - Azimuthal
Secant: The projection surface intersects the Earth, resulting in two standard parallels where distortion is minimized.
Tangent: The projection surface is tangent to the Earth at one point or line, minimizing distortion at that location.
Projection Mode (2 types)
Tangent Secant
Projection Surface (3 types) × Projection Aspect (3 types) × Distortion Property (4 types) × Projection Mode (2 types) =
72 combinations
• The Open Geospatial Consortium (OGC) is an international, voluntary consensus-based organization that develops and promotes open standards for geospatial content, services, and data sharing.
• OGC is comprised of hundreds of member organizations, including government agencies, academic institutions, private companies, and individuals, all working together to ensure the interoperability of geospatial systems across different platforms and technologies.
Development of Standards: Creating open standards for geospatial and location-based services.
Testing and Certification: Providing a framework for testing the compliance of software products with OGC standards.
Collaboration: Working with other standards bodies, such as ISO (International Organization for Standardization) and W3C (World Wide Web Consortium), to ensure global interoperability.
Interoperability: Enables seamless sharing and use of spatial data across different systems and platforms.
Cost Efficiency: Reduces vendor lock-in and supports data reuse.
Innovation: Fosters new geospatial technologies like 3D mapping and IoT.
Global Collaboration: Facilitates worldwide data sharing for disaster management, urban planning, and environmental monitoring.
OGC-approved data formats are standardized formats established by the Open Geospatial Consortium to ensure interoperability and consistency in Geographic Information Systems (GIS).
These formats support various types of geospatial data, including vector, raster, and 3D data, enabling seamless data sharing, analysis, and visualization across diverse platforms and applications.
GeoTIFF (Georeferenced Tagged Image File Format)
NetCDF (Network Common Data Form)
HDF (Hierarchical Data Format)
WCS (Web Coverage Service)
GeoPackage (GPKG, supports raster data)
GeoJSON (Geographic JSON)
GML (Geography Markup Language)
KML (Keyhole Markup Language)
WFS (Web Feature Service)
GeoPackage (GPKG, supports vector data)
WMS (Web Map Service)
WMTS (Web Map Tile Service)
3D Tiles (OGC format for streaming 3D geospatial data)
CityGML (for 3D city models)
IndoorGML (for indoor space representation)
COLLADA (OGC format for 3D models, interoperable with many 3D applications)
Geodesy is the science of accurately measuring and understanding the three fundamental properties of the Earth such as,
While Earth is often approximated as an ellipsoid or spheroid, its true shape is a geoid - an irregularly shaped ellipsoid that reflects variations in gravitational force due to Earth’s uneven mass distribution.
Earth’s average diameter is about 12,742 kilometers. The equatorial circumference is approximately 40,075 kilometers, making Earth slightly wider at the equator due to its rotation.
Earth is tilted at an angle of about 23.5 degrees relative to its orbital plane around the Sun. This axial tilt causes seasonal variations and influences climate patterns.
Gravity anomalies, variations from standard gravity, are measured in gravity units (g.u.) or milligals (mGal), where 1 mGal equals 10 g.u. One g.u. is approximately one ten-millionth of Earth’s surface gravity.
Geodesy is essential for defining Earth’s shape, size, and gravitational field, enabling accurate mapping and spatial analysis.
It supports the development of horizontal and vertical coordinate systems, crucial for navigation, GPS, and geographic information systems (GIS).
In GIS, layers are used for the representation different types of spatial data.
Each layer corresponds to specific geographic features, such as roads, rivers, land use, or population density.
Layers can be in vector (points, lines, polygons) or raster (grids) formats and are stacked to form a composite map.
-x By overlaying different layers, users can examine relationships between datasets, make informed decisions, and conduct spatial analysis, such as identifying patterns or predicting trends across regions.
Interpolation predicts values for cells in a raster from a limited number of sample data points.
It can be used to predict unknown values for any geographic point data, such as rainfall, elevation, chemical concentrations, etc.
The input here is a point dataset of known rainfall-level values, shown by the illustration on the left. The illustration on the right shows a raster interpolated from these points. The unknown values are predicted with a mathematical formula that uses the values of nearby known points.
A typical use for point interpolation is to create an elevation surface from a set of sample measurements. In the following graphic, each symbol in the point layer represents a location where the elevation has been measured. By interpolating, the values for each cell between these input points will be predicted.
In the example below, the interpolation tools were used to study the correlation of the ozone concentration on lung disease. The image on the left shows the locations of the ozone monitoring stations. The image on the right displays the interpolated surface, providing predictions for each location.
Deterministic interpolation methods rely on direct mathematical calculations without assuming any spatial statistical model.
They estimate unknown values from nearby known points, based on the assumption that locations close to one another have similar values.
This method assigns the value of the nearest known point to unsampled locations, creating tessellated polygons.
It is best used for categorical data or situations where a sudden change between zones is acceptable, such as assigning rainfall stations to regions.
Figure: The tessellated surface shown illustrates the boundaries created by Thiessen polygons. Each polygon encloses one sample point, and every unsampled location inside a polygon is assigned the value of the nearest point. Abrupt value changes occur at the polygon edges.
IDW estimates unknown values by averaging nearby points, with closer points having more influence.
It’s used when spatial relationships are known to decrease with distance.
IDW is effective for continuous data with moderate variation, such as temperature or precipitation measurements.
Figure: The IDW interpolation of a precipitation illustrates how values are influenced by nearby points. An IDW power coefficient of 2 creates a smooth surface, while a higher coefficient (15) increases the influence of nearby points, resembling a Thiessen interpolation output.
The statistical interpolation methods include trend surfaces and Kriging.
These methods rely on statistical models to account for spatial relationships and predict unknown values.
They offer more flexibility for complex surfaces and are effective for continuous data.
Trend surfaces model spatial patterns using polynomial equations.
A 0th-order surface is flat and represents the mean value of all points.
A 1st-order surface adds a slope, capturing linear trends across space.
The 2nd-order surface introduces curvature, better reflecting more complex spatial variations.
Figure: Illustrates the progression from a flat surface (0th order) to a sloped plane (1st order) and a curved, parabolic surface (2nd order). Each figure highlights increasing complexity in how spatial data trends are captured.
Kriging is a geostatistical method that uses spatial autocorrelation to interpolate values at unsampled locations.
It first removes any spatial trends, computes a variogram, fits a model, and uses this model to assign weights for interpolation.
Kriging adapts these weights locally, enhancing accuracy.
Figure 1: Shows the Kriging interpolation of de-trended residual precipitation values. Figure 2: The final combined surface integrates trend and kriging outputs. Figure 3: Displays a variance map, illustrating interpolation uncertainty, with lower variance indicating higher confidence in predictions.
Overlay analysis is a GIS technique used to combine multiple spatial layers to identify relationships, patterns, or areas of overlap between them.
It helps in decision-making by evaluating how different spatial features interact.
We can classify overlays as raster or vector based on the data, but there are tools that can be used for both types.
It helps to answer questions like,
What agricultural zones overlap with protected wildlife areas?
What rivers flow through which national parks?
Which forests are located within specific watershed boundaries?
What transportation routes are inside urban growth boundaries?
Which industrial zones are near public water supply sources?
Proximity analysis assesses how close features are to each other, allowing for the measurement of distances, identification of nearby features, and the evaluation of spatial patterns.
Common methods include buffering, nearest neighbor analysis, and distance calculations.
This involves conducting a buffer analysis around a point, line, or polygon, which can be categorized into the following combinations:
Figure: What is the suitable extent around a residential area for establishing a noise buffer zone?
Figure: What variable buffer distances should be applied around a residential area to account for different types of land use impacts?
Figure: Which areas near major roads are most suitable for industrial development based on proximity analysis?
Figure: How far can floodwaters potentially spread from a river, affecting the surrounding residential areas?
Surface analysis involves examining the characteristics of a surface represented in a geographic information system (GIS) to understand spatial relationships and patterns.
It utilizes digital elevation models (DEMs) to derive various attributes, providing valuable insights for terrain modeling and analysis.
Figure: Slope measures the steepness or degree of incline of a surface. It is essential for applications like assessing landslide risks, planning infrastructure, and determining suitable locations for agriculture or forestry.
Figure: Aspect indicates the direction a slope faces. Understanding aspect is crucial for applications like solar radiation analysis, habitat suitability modeling, and microclimate studies, as it influences temperature and moisture conditions.
Figure: Hillshade creates a shaded relief representation of the terrain based on the angle of sunlight. This visualization helps in understanding the topography, enhancing landscape features, and improving the visual interpretation of the terrain.
Figure: Contour analysis involves drawing lines that connect points of equal elevation on a map. Contours are used for mapping terrain, calculating elevation changes, and analyzing watershed boundaries.
Figure: Viewshed analysis determines the visible area from a specific viewpoint based on the terrain. This analysis is vital for urban planning, tourism development or defense use cases.
Center and dispersion analysis techniques help identify the central tendency and dispersion patterns of geographic features or points of interest within a given area.
These techniques provide valuable insights into the distribution patterns, concentration, and spatial relationships among various elements
The mean center is the average geographical location of a set of points, calculated by averaging the coordinates of each point. It represents the “central” point of a dataset.
Example Question: What is the average location of all the schools in a city?
The median center is the point that minimizes the distance to all the locations in a dataset, essentially the median of the coordinates. It is particularly useful in skewed datasets.
Example Question: Where is the central point of population distribution in a city, ensuring equal access to resources?
Standard distance is a measure of dispersion that indicates how far away points are from the mean center on average. It helps understand the spread of a dataset.
Example Question: How concentrated or dispersed are the locations of new businesses within a neighborhood?
Network analysis is a set of methods used in GIS to analyze the relationships and flow patterns along a network of interconnected lines, such as roads, railways, or pipelines.
It helps solve problems related to route optimization, service area coverage, and connectivity.
Network analysis can be used to calculate the shortest path between two or more points along a network of roads, helping users find the quickest or shortest travel route. What is the shortest path between these two locations?
Network analysis allows for route optimization by applying restrictions, such as avoiding toll roads, highways, to find a path that meets the user’s preferences. What is the best route to take from point A to point B while avoiding toll roads?
Watershed analysis in GIS involves identifying the land area (watershed) that drains into a specific body of water, such as a river or reservoir.
This analysis helps in understanding how water flows across a landscape, managing water resources, preventing floods, and controlling pollution. These are used to delineate watersheds, assess water flow direction, and calculate runoff.
Stream Order assigns a number to stream segments based on their tributary connections, helping to classify streams. First-order streams have no tributaries and are highly sensitive to pollution.
It supports better water management and conservation strategies in a watershed, using two key methods: the Strahler method and the Shreve method.
In the Strahler method, all streams without tributaries are assigned an order of 1 (first-order). Stream order increases only when streams of the same order converge. For example, two first-order streams create a second-order stream. This method is widely used but does not account for all links in a network, making it sensitive to network changes.
In the Shreve method, all links are counted, and the order (or “magnitude”) is additive. Each first-order link contributes to the total order, making it more inclusive of stream network complexity. For example, a first- and second-order intersection results in a third-order link. This method provides a higher order count than Strahler and is especially useful for detailed analysis.
Thematic mapping and visualization involve creating maps focused on specific themes, such as population density, literacy rates, or employment statistics, rather than geographic features alone.
This approach makes it easier to understand and interpret spatial patterns and insights within datasets, supporting data-driven decision-making in fields like urban planning and environmental management.
Uses color gradients to represent data intensity across areas, useful for visualizing variations like population density or income levels. Question: What is the population density of the area?
Represents quantity using dots, where each dot equals a specific value, effective for showing distribution patterns like literacy rates. Question: What is the population literacy rate in the area?
Displays pie charts over regions, using segment sizes to represent parts of a whole, ideal for workforce distributions in regions. Question: What is the total working and non-working population in each region?
Uses bar charts on map regions to represent comparative quantities, like female workforce distribution across different areas. Question: What is the distribution of working and non-working female population in each area?
GIS technologies encompass various tools and software that support spatial data analysis, mapping, and visualization.
Open-source options enable customization, scalability, and community-driven improvements, making them widely used for accessible, flexible GIS solutions.
Open-source GIS software provides complete platforms for spatial analysis, mapping, and data management.
Examples: QGIS, SNAP
GIS platforms and application structures are the foundation for building, deploying, and running GIS applications.
These applications rely on technologies, including programming languages, frameworks, libraries, software, and databases, all of which must be compatible with specific platforms and application structures to ensure optimal performance, accessibility, and user experience.
Open-source GIS software provides accessible, customizable, and cost-effective solutions for managing and analyzing geospatial data, making GIS tools available to a broader audience.
It enables collaboration and community-driven improvements, fostering innovation in geospatial technology.
(i)We can install and use the publicly available plugins and, (ii)We can also create our own plugins.
GIS development requires various programming languages, each suited to specific platforms and tasks.
The choice of language depends on the application’s platform - web, mobile, or desktop and the type of geospatial processing required.