Skip to main content

Image Classification → Steps


Assembling the Training Data

Training data (also called training samples or signature sets) are the foundation of supervised image classification in remote sensing.
This is where the analyst selects representative examples of each land-cover class—such as water, vegetation, urban, soil, etc.—from the satellite image.

To prepare training data properly, several analytical and interactive steps are used. These help ensure that the classes are well separated and that the classifier receives the correct spectral information.

1. Graphical Representation of Spectral Response Patterns

✔ What it means

For each class (e.g., water, forest, built-up), the training pixels have a spectral signature—a pattern of reflectance values across the image's spectral bands.

This pattern is visualized using:

  • Spectral reflectance curves

  • Band-by-band scatter plots

  • Histograms for each band

✔ Purpose

  • To understand how different classes behave in different bands

  • To check if the selected training pixels are spectrally consistent

  • To identify overlaps between classes (e.g., dark soil and turbid water)

✔ Key terminology

  • Spectral profile / spectral signature

  • Spectral separability

  • Spectral scatterplot

  • Feature space

2. Quantitative Expressions of Category Separation

This step uses mathematical measures to check if classes are well-separated in spectral space.

✔ Why it matters

Classification accuracy depends on how distinct one class is from another.
If training classes overlap too much, classification errors will occur.

✔ Common quantitative measures

  • Transformed Divergence (TD)

  • Jeffries–Matusita Distance (JM)

  • Bhattacharyya Distance (BD)

✔ What these values indicate

  • Values close to 2.0 (JM scale) → excellent class separability

  • Values close to 0.0 → poor separability; classes overlap

  • Helps decide whether to:

    • combine classes

    • redefine training samples

    • collect more samples

    • split a mixed class

✔ Key terminology

  • Separability index

  • Statistical distance

  • Cluster separation

  • Spectral overlap

3. Self-Classification of the Training Data Set

✔ Concept

Before performing classification on the full image, the classifier is run only on the training pixels themselves.

✔ Purpose

  • To check if the algorithm correctly "recognizes" the classes it was trained on.

  • If the classifier mislabels the training samples, the training data need to be corrected.

✔ What it reveals

  • Misclassified pixels → inaccurate training sets

  • Mixed or overlapping classes

  • Inconsistencies in attribute statistics (means, variances)

  • Too much variability within a class

✔ Key terminology

  • Internal accuracy check

  • Confusion among training classes

  • Spectral homogeneity

4. Interactive Preliminary Classification

✔ What it is

A rough or temporary classification is generated on the image using preliminary training samples.

✔ Purpose

  • To visually inspect how the training data behave when applied to the entire image

  • To refine training sites

  • To identify new sub-classes or remove misidentified ones

✔ What the analyst checks

  • Are water bodies correctly classified?

  • Are vegetation areas split properly (forest vs cropland)?

  • Are built-up areas being confused with dry soil?

✔ Why "interactive"?

The analyst reviews the output and actively adjusts:

  • training polygons

  • class definitions

  • band combinations

  • class separability

✔ Key terminology

  • Pre-classification map

  • Trial classification

  • Interactive refinement

5. Representative Subscene Classification

✔ Concept

Instead of classifying the whole image, a small but representative subscene is used.

A subscene:

  • contains all major land-cover types

  • captures geographic and spectral variability

  • is easier to evaluate and test

✔ Purpose

  • To test classifier performance on a manageable area

  • To refine spectral signatures before final classification

  • To avoid wasting processing time on the full image if training data are weak

✔ What it helps detect

  • Class confusion in specific regions

  • Spectral variability across the scene

  • Need for more training samples

  • Problems with similar classes (e.g., shallow water vs wet soil)

✔ Key terminology

  • Subscene

  • Training refinement

  • Pilot classification

  • Signature validation


Assembling training data for supervised image classification involves:

  1. Graphical representation of spectral response patterns – using spectral curves, histograms, and scatter plots to visualize class behavior.

  2. Quantitative expressions of category separation – using statistical measures (JM, TD, BD) to evaluate how distinct classes are.

  3. Self-classification of training data – testing if the classifier correctly labels its own training samples.

  4. Interactive preliminary classification – producing a trial classification to visually refine training sites.

  5. Representative subscene classification – testing the classifier on a smaller, diverse image subset to check accuracy and refine signatures.


Comments

Popular posts from this blog

Platforms in Remote Sensing

In remote sensing, a platform is the physical structure or vehicle that carries a sensor (camera, scanner, radar, etc.) to observe and collect information about the Earth's surface. Platforms are classified mainly by their altitude and mobility : Ground-Based Platforms Definition : Sensors mounted on the Earth's surface or very close to it. Examples : Tripods, towers, ground vehicles, handheld instruments. Applications : Calibration and validation of satellite data Detailed local studies (e.g., soil properties, vegetation health, air quality) Strength : High spatial detail but limited coverage. Airborne Platforms Definition : Sensors carried by aircraft, balloons, or drones (UAVs). Altitude : A few hundred meters to ~20 km. Examples : Airplanes with multispectral scanners UAVs with high-resolution cameras or LiDAR High-altitude balloons (stratospheric platforms) Applications : Local-to-regional mapping ...

Types of Remote Sensing

Remote Sensing means collecting information about the Earth's surface without touching it , usually using satellites, aircraft, or drones . There are different types of remote sensing based on the energy source and the wavelength region used. 🛰️ 1. Active Remote Sensing 📘 Concept: In active remote sensing , the sensor sends out its own energy (like a signal or pulse) to the Earth's surface. The sensor then records the reflected or backscattered energy that comes back from the surface. ⚙️ Key Terminology: Transmitter: sends energy (like a radar pulse or laser beam). Receiver: detects the energy that bounces back. Backscatter: energy that is reflected back to the sensor. 📊 Examples of Active Sensors: RADAR (Radio Detection and Ranging): Uses microwave signals to detect surface roughness, soil moisture, or ocean waves. LiDAR (Light Detection and Ranging): Uses laser light (near-infrared) to measure elevation, vegetation...

Resolution of Sensors in Remote Sensing

Spatial Resolution 🗺️ Definition : The smallest size of an object on the ground that a sensor can detect. Measured as : The size of a pixel on the ground (in meters). Example : Landsat → 30 m (each pixel = 30 × 30 m on Earth). WorldView-3 → 0.31 m (very detailed, you can see cars). Fact : Higher spatial resolution = finer details, but smaller coverage. Spectral Resolution 🌈 Definition : The ability of a sensor to capture information in different parts (bands) of the electromagnetic spectrum . Measured as : The number and width of spectral bands. Types : Panchromatic (1 broad band, e.g., black & white image). Multispectral (several broad bands, e.g., Landsat with 7–13 bands). Hyperspectral (hundreds of very narrow bands, e.g., AVIRIS). Fact : Higher spectral resolution = better identification of materials (e.g., minerals, vegetation types). Radiometric Resolution 📊 Definition : The ability of a sensor to ...

geostationary and sun-synchronous

Orbital characteristics of Remote sensing satellite geostationary and sun-synchronous  Orbits in Remote Sensing Orbit = the path a satellite follows around the Earth. The orbit determines what part of Earth the satellite can see , how often it revisits , and what applications it is good for . Remote sensing satellites mainly use two standard orbits : Geostationary Orbit (GEO) Sun-Synchronous Orbit (SSO)  Geostationary Satellites (GEO) Characteristics Altitude : ~35,786 km above the equator. Period : 24 hours → same as Earth's rotation. Orbit type : Circular, directly above the equator . Appears "stationary" over one fixed point on Earth. Concepts & Terminologies Geosynchronous = orbit period matches Earth's rotation (24h). Geostationary = special type of geosynchronous orbit directly above equator → looks fixed. Continuous coverage : Can monitor the same area all the time. Applications Weather...

Man-Made Disasters

  A man-made disaster (also called a technological disaster or anthropogenic disaster ) is a catastrophic event caused directly or indirectly by human actions , rather than natural processes. These disasters arise due to negligence, error, industrial activity, conflict, or misuse of technology , and often result in loss of life, property damage, and environmental degradation . Terminology: Anthropogenic = originating from human activity. Technological hazard = hazard caused by failure or misuse of technology or industry. 🔹 Conceptual Understanding Man-made disasters are part of the Disaster Management Cycle , which includes: Prevention – avoiding unsafe practices. Mitigation – reducing disaster impact (e.g., safety regulations). Preparedness – training and planning. Response – emergency actions after the disaster. Recovery – long-term rebuilding and policy correction. These disasters are predictable and preventable through strong...