Skip to main content

RADIOMETRIC CORRECTION

 


Radiometric correction is the process of removing sensor and environmental errors from satellite images so that the measured brightness values (Digital Numbers or DNs) truly represent the Earth's surface reflectance or radiance.

In other words, it corrects for sensor defects, illumination differences, and atmospheric effects.


1. Detector Response Calibration

Satellite sensors use multiple detectors to scan the Earth's surface. Sometimes, each detector responds slightly differently, causing distortions in the image. Calibration adjusts all detectors to respond uniformly.

This includes:

(a) De-Striping

  • Problem: Sometimes images show light and dark vertical or horizontal stripes (banding).

    • Caused by one or more detectors drifting away from their normal calibration — they record higher or lower values than others.

    • Common in early Landsat MSS data.

  • Effect: Every few lines (e.g., every 6th line) appear consistently brighter or darker.

  • Solution (De-Striping):

    • Compare histograms of scan lines (e.g., 1,7,13 or 2,8,14) for mean and standard deviation.

    • Adjust the detector's response to match neighboring detectors.

    • Methods:

      • Histogram equalization and normalization

      • Fourier transformation (removes periodic striping patterns)


(b) Missing Scan Line Removal

  • Problem: Sometimes a detector stops working or becomes temporarily saturated, creating blank lines or missing data in the image.

  • Solution:

    • Replace missing lines with estimated pixel values based on the lines above and below using interpolation techniques.

    • Example: Affected Landsat 7 ETM+ (Scan Line Corrector failure).


(c) Random Noise Removal

  • Problem: Some pixels show random bright or dark spots known as "salt-and-pepper noise" or "snowy noise."

    • Caused by random electronic interference or transmission errors.

  • Solution:

    • Spatial filtering: Replace noisy pixels with average values from neighboring pixels.

    • Convolution filtering: Smooths image by using a moving filter (kernel) to reduce random pixel variation.


(d) Vignetting Removal

  • Problem: In images taken with lenses, the corners often appear darker than the center — this is vignetting.

  • Cause: Uneven illumination across the sensor array or lens curvature.

  • Solution:

    • Use sensor calibration data that describes how brightness varies from center to edges.

    • Apply Fourier Transform or other normalization methods to equalize brightness.


2. Sun Angle and Topographic Correction

(a) Sun Angle Correction

  • The sun's position changes with time of day and season, affecting image brightness.

  • Higher solar angle (summer) → more direct sunlight → brighter image.

  • Lower solar angle (winter) → less sunlight → darker image.

  • Correction Method:

    • Adjust each pixel's brightness (DN) by dividing it with the sine of the solar elevation angle:
      [
      DN_{corrected} = \frac{DN_{original}}{\sin(\text{solar elevation angle})}
      ]

    • Solar elevation data is given in the image metadata or header file.


(b) Topographic Correction

  • Problem: In hilly or mountainous areas, slopes facing the sun appear brighter, while those facing away appear darker due to uneven solar illumination.

  • Cause:

    • Slope and aspect of terrain

    • Shadowing effects

    • Bidirectional Reflectance Distribution Function (BRDF) differences

  • Solution: Adjust radiance based on slope orientation and sun angle using models such as:

    Minnaert Correction:
    [
    L_n = L \cdot (\cos e)^{k-1} \cdot \cos i
    ]
    Where:

    • (L_n): normalized radiance

    • (L): measured radiance

    • (e): slope angle (from DEM)

    • (i): solar incidence angle

    • (k): Minnaert constant (depends on land cover and illumination conditions)

    This correction helps produce uniform brightness across slopes.


3. Atmospheric Correction

  • Problem: Before reaching the sensor, sunlight interacts with the atmosphere, where gases, dust, and water vapor scatter and absorb radiation.

    • Causes haze, color distortion, and lower contrast in the image.

  • Goal: Remove the effects of atmosphere to obtain true surface reflectance.

  • Methods:

    • Dark Object Subtraction (DOS): Assumes that dark pixels (like water) should have near-zero reflectance; subtracts atmospheric haze values.

    • Radiative Transfer Models: e.g., 6S, MODTRAN, FLAASH, or QUAC to simulate atmospheric scattering and absorption effects accurately.


Type of CorrectionProblem FixedExample of ErrorCommon Methods
Detector CalibrationUneven sensor responseStriping, noiseHistogram matching, Fourier transform
Missing LineLost data linesLandsat 7 SLC failureInterpolation
Random NoiseSalt-and-pepper noiseBright/dark spotsSpatial/convolution filtering
VignettingDark cornersLens-based imagesFourier normalization
Sun AngleSeasonal/diurnal illuminationWinter images darkerDivide by sin(solar angle)
TopographicSlope illumination differenceBright/dark slopesMinnaert correction
AtmosphericScattering, absorptionHazy imagesDOS, FLAASH, MODTRAN


Comments

Popular posts from this blog

Natural Disasters

A natural disaster is a catastrophic event caused by natural processes of the Earth that results in significant loss of life, property, and environmental resources. It occurs when a hazard (potentially damaging physical event) interacts with a vulnerable population and leads to disruption of normal life . Key terms: Hazard → A potential natural event (e.g., cyclone, earthquake). Disaster → When the hazard causes widespread damage due to vulnerability. Risk → Probability of harmful consequences from interaction of hazard and vulnerability. Vulnerability → Degree to which a community or system is exposed and unable to cope with the hazard. Resilience → Ability of a system or society to recover from the disaster impact. 👉 Example: An earthquake in an uninhabited desert is a hazard , but not a disaster unless people or infrastructure are affected. Types Natural disasters can be classified into geophysical, hydrological, meteorological, clim...

geostationary and sun-synchronous

Orbital characteristics of Remote sensing satellite geostationary and sun-synchronous  Orbits in Remote Sensing Orbit = the path a satellite follows around the Earth. The orbit determines what part of Earth the satellite can see , how often it revisits , and what applications it is good for . Remote sensing satellites mainly use two standard orbits : Geostationary Orbit (GEO) Sun-Synchronous Orbit (SSO)  Geostationary Satellites (GEO) Characteristics Altitude : ~35,786 km above the equator. Period : 24 hours → same as Earth's rotation. Orbit type : Circular, directly above the equator . Appears "stationary" over one fixed point on Earth. Concepts & Terminologies Geosynchronous = orbit period matches Earth's rotation (24h). Geostationary = special type of geosynchronous orbit directly above equator → looks fixed. Continuous coverage : Can monitor the same area all the time. Applications Weather...

Types of Remote Sensing

Remote Sensing means collecting information about the Earth's surface without touching it , usually using satellites, aircraft, or drones . There are different types of remote sensing based on the energy source and the wavelength region used. 🛰️ 1. Active Remote Sensing 📘 Concept: In active remote sensing , the sensor sends out its own energy (like a signal or pulse) to the Earth's surface. The sensor then records the reflected or backscattered energy that comes back from the surface. ⚙️ Key Terminology: Transmitter: sends energy (like a radar pulse or laser beam). Receiver: detects the energy that bounces back. Backscatter: energy that is reflected back to the sensor. 📊 Examples of Active Sensors: RADAR (Radio Detection and Ranging): Uses microwave signals to detect surface roughness, soil moisture, or ocean waves. LiDAR (Light Detection and Ranging): Uses laser light (near-infrared) to measure elevation, vegetation...

India remote sensing

1. Foundational Phase (Early 1970s – Early 1980s) Objective: To explore the potential of space-based observation for national development. 1972: The Space Applications Programme (SAP) was initiated by the Indian Space Research Organisation (ISRO), focusing on applying space technology for societal benefits. 1975: The Department of Space (DoS) was established, providing an institutional base for space applications, including remote sensing. 1977: India began aerial and balloon-borne experiments to study Earth resources and assess how remote sensing data could aid in agriculture, forestry, and hydrology. 1978 (June 7): Bhaskara-I launched by the Soviet Union — India's first experimental Earth Observation satellite . Payloads: TV cameras (for land and ocean surface observation) and a Microwave Radiometer. Significance: Proved that satellite-based Earth observation was feasible for India's needs. 1981 (November 20): Bhaskara-II launche...

Linear Arrays Along-Track Scanners or Pushbroom Scanners

Multispectral Imaging Using Linear Arrays (Along-Track Scanners or Pushbroom Scanners) Multispectral Imaging: As previously defined, this involves capturing images using multiple sensors that are sensitive to different wavelengths of electromagnetic radiation. Linear Array of Detectors (A): This refers to a row of discrete detectors arranged in a straight line. Each detector is responsible for measuring the radiation within a specific wavelength band. Focal Plane (B): This is the plane where the image is formed by the lens system. It is the location where the detectors are placed to capture the focused image. Formed by Lens Systems (C): The lens system is responsible for collecting and focusing the incoming radiation onto the focal plane. It acts like a camera lens, creating a sharp image of the scene. Ground Resolution Cell (D): As previously defined, this is the smallest area on the ground that can be resolved by a remote sensing sensor. In the case of linear array scanne...