Skip to main content

Image Classification → Steps


Assembling the Training Data

Training data (also called training samples or signature sets) are the foundation of supervised image classification in remote sensing.
This is where the analyst selects representative examples of each land-cover class—such as water, vegetation, urban, soil, etc.—from the satellite image.

To prepare training data properly, several analytical and interactive steps are used. These help ensure that the classes are well separated and that the classifier receives the correct spectral information.

1. Graphical Representation of Spectral Response Patterns

✔ What it means

For each class (e.g., water, forest, built-up), the training pixels have a spectral signature—a pattern of reflectance values across the image's spectral bands.

This pattern is visualized using:

  • Spectral reflectance curves

  • Band-by-band scatter plots

  • Histograms for each band

✔ Purpose

  • To understand how different classes behave in different bands

  • To check if the selected training pixels are spectrally consistent

  • To identify overlaps between classes (e.g., dark soil and turbid water)

✔ Key terminology

  • Spectral profile / spectral signature

  • Spectral separability

  • Spectral scatterplot

  • Feature space

2. Quantitative Expressions of Category Separation

This step uses mathematical measures to check if classes are well-separated in spectral space.

✔ Why it matters

Classification accuracy depends on how distinct one class is from another.
If training classes overlap too much, classification errors will occur.

✔ Common quantitative measures

  • Transformed Divergence (TD)

  • Jeffries–Matusita Distance (JM)

  • Bhattacharyya Distance (BD)

✔ What these values indicate

  • Values close to 2.0 (JM scale) → excellent class separability

  • Values close to 0.0 → poor separability; classes overlap

  • Helps decide whether to:

    • combine classes

    • redefine training samples

    • collect more samples

    • split a mixed class

✔ Key terminology

  • Separability index

  • Statistical distance

  • Cluster separation

  • Spectral overlap

3. Self-Classification of the Training Data Set

✔ Concept

Before performing classification on the full image, the classifier is run only on the training pixels themselves.

✔ Purpose

  • To check if the algorithm correctly "recognizes" the classes it was trained on.

  • If the classifier mislabels the training samples, the training data need to be corrected.

✔ What it reveals

  • Misclassified pixels → inaccurate training sets

  • Mixed or overlapping classes

  • Inconsistencies in attribute statistics (means, variances)

  • Too much variability within a class

✔ Key terminology

  • Internal accuracy check

  • Confusion among training classes

  • Spectral homogeneity

4. Interactive Preliminary Classification

✔ What it is

A rough or temporary classification is generated on the image using preliminary training samples.

✔ Purpose

  • To visually inspect how the training data behave when applied to the entire image

  • To refine training sites

  • To identify new sub-classes or remove misidentified ones

✔ What the analyst checks

  • Are water bodies correctly classified?

  • Are vegetation areas split properly (forest vs cropland)?

  • Are built-up areas being confused with dry soil?

✔ Why "interactive"?

The analyst reviews the output and actively adjusts:

  • training polygons

  • class definitions

  • band combinations

  • class separability

✔ Key terminology

  • Pre-classification map

  • Trial classification

  • Interactive refinement

5. Representative Subscene Classification

✔ Concept

Instead of classifying the whole image, a small but representative subscene is used.

A subscene:

  • contains all major land-cover types

  • captures geographic and spectral variability

  • is easier to evaluate and test

✔ Purpose

  • To test classifier performance on a manageable area

  • To refine spectral signatures before final classification

  • To avoid wasting processing time on the full image if training data are weak

✔ What it helps detect

  • Class confusion in specific regions

  • Spectral variability across the scene

  • Need for more training samples

  • Problems with similar classes (e.g., shallow water vs wet soil)

✔ Key terminology

  • Subscene

  • Training refinement

  • Pilot classification

  • Signature validation


Assembling training data for supervised image classification involves:

  1. Graphical representation of spectral response patterns – using spectral curves, histograms, and scatter plots to visualize class behavior.

  2. Quantitative expressions of category separation – using statistical measures (JM, TD, BD) to evaluate how distinct classes are.

  3. Self-classification of training data – testing if the classifier correctly labels its own training samples.

  4. Interactive preliminary classification – producing a trial classification to visually refine training sites.

  5. Representative subscene classification – testing the classifier on a smaller, diverse image subset to check accuracy and refine signatures.


Comments

Popular posts from this blog

Accuracy Assessment

Accuracy assessment is the process of checking how correct your classified satellite image is . 👉 After supervised classification, the satellite image is divided into classes like: Water Forest Agriculture Built-up land Barren land But classification is done using computer algorithms, so some areas may be wrongly classified . 👉 Accuracy assessment helps to answer this question: ✔ "How much of my classified map is correct compared to real ground conditions?"  Goal The main goal is to: Measure reliability of classified maps Identify classification errors Improve classification results Provide scientific validity to research 👉 Without accuracy assessment, a classified map is not considered scientifically reliable . Reference Data (Ground Truth Data) Reference data is real-world information used to check classification accuracy. It can be collected from: ✔ Field survey using GPS ✔ High-resolution satellite images (Google Earth etc.) ✔ Existing maps or survey reports 🧭 Exampl...

Landsat 8 Band designation and Band Combination.

Landsat 8 Band designation and Band Combination.  Landsat 8-9 Operational Land Imager (OLI) and Thermal Infrared Sensor (TIRS) Bands Wavelength (micrometers) Resolution (meters) Band 1 - Coastal aerosol 0.43-0.45 30 Band 2 - Blue 0.45-0.51 30 Band 3 - Green 0.53-0.59 30 Band 4 - Red 0.64-0.67 30 Band 5 - Near Infrared (NIR) 0.85-0.88 30 Band 6 - SWIR 1 1.57-1.65 30 Band 7 - SWIR 2 2.11-2.29 30 Band 8 - Panchromatic 0.50-0.68 15 Band 9 - Cirrus 1.36-1.38 30 Band 10 - Thermal Infrared (TIRS) 1 10.6-11.19 100 Band 11 - Thermal Infrared (TIRS) 2 11.50-12.51 100 Vineesh V Assistant Professor of Geography, Directorate of Education, Government of Kerala. https://www.facebook.com/Applied.Geography http://geogisgeo.blogspot.com

Change Detection

Change detection is the process of finding differences on the Earth's surface over time by comparing satellite images of the same area taken on different dates . After supervised classification , two classified maps (e.g., Year-1 and Year-2) are compared to identify land use / land cover changes .  Goal To detect where , what , and how much change has occurred To monitor urban growth, deforestation, floods, agriculture, etc.  Basic Concept Forest → Forest = No change Forest → Urban = Change detected Key Terminologies Multi-temporal images : Images of the same area at different times Post-classification comparison : Comparing two classified maps Change matrix : Table showing class-to-class change Change / No-change : Whether land cover remains same or different Main Methods Post-classification comparison – Most common and easy Image differencing – Subtract pixel values Image ratioing – Divide pixel values Deep learning methods – Advanced AI-based detection Examples Agricult...

Landsat band composition

Short-Wave Infrared (7, 6 4) The short-wave infrared band combination uses SWIR-2 (7), SWIR-1 (6), and red (4). This composite displays vegetation in shades of green. While darker shades of green indicate denser vegetation, sparse vegetation has lighter shades. Urban areas are blue and soils have various shades of brown. Agriculture (6, 5, 2) This band combination uses SWIR-1 (6), near-infrared (5), and blue (2). It's commonly used for crop monitoring because of the use of short-wave and near-infrared. Healthy vegetation appears dark green. But bare earth has a magenta hue. Geology (7, 6, 2) The geology band combination uses SWIR-2 (7), SWIR-1 (6), and blue (2). This band combination is particularly useful for identifying geological formations, lithology features, and faults. Bathymetric (4, 3, 1) The bathymetric band combination (4,3,1) uses the red (4), green (3), and coastal bands to peak into water. The coastal band is useful in coastal, bathymetric, and aerosol studies because...

Development and scope of Environmental Geography and Recent concepts in environmental Geography

Environmental Geography studies the relationship between humans and nature in a spatial (place-based) way. It combines Physical Geography (natural processes) and Human Geography (human activities). A. Early Stage 🔹 Environmental Determinism Concept: Nature controls human life. Meaning: Climate, landforms, and soil decide how people live. Example: People in deserts (like Sahara Desert) live differently from people in fertile river valleys. 🔹 Possibilism Concept: Humans can modify nature. Meaning: Environment gives options, but humans make choices. Example: In dry areas like Rajasthan, people use irrigation to grow crops. 👉 In this stage, geography was mostly descriptive (explaining what exists). B. Evolution Stage (Mid-20th Century) Environmental problems increased due to: Industrialization Urbanization Deforestation Pollution Geographers started studying: Environmental degradation Resource management Human impact on ecosystems The field became analytical and problem-solving...