1 / 20

Lecture 12: Image Processing

Thursday 11 February 2010. Lecture 12: Image Processing. Reading Ch 7.1 - 7.6. Last lecture: Earth-orbiting satellites. Image Processing Because of the way most remote-sensing texts are organized, what strikes most students is the vast array of algorithms with odd names

dawson
Download Presentation

Lecture 12: Image Processing

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Thursday 11 February 2010 Lecture 12: Image Processing Reading Ch 7.1 - 7.6 Last lecture: Earth-orbiting satellites

  2. Image Processing Because of the way most remote-sensing texts are organized, what strikes most students is the vast array of algorithms with odd names and obscure functions What is elusive is the underlying simplicity. Many algorithms are substantially the same – they have similar purposes and similar results

  3. Image Processing • There are basically five families of • algorithms that do things to images: • Radiometric algorithms • change the DNs • Calibration • Contrast enhancement • 2) Geometric algorithms • change the spatial arrangement of pixels or adjust DN’s based on their neighbors’ values • Registration • “Visualization” • Spatial-spectral transformation • Spatial filtering

  4. Contrast stretching & calibration Enhancement: Imagine a DN histogram centered at 75 DN and running from 50 to 100. In lab, you would move sliders to 50 and 100 DN to display it well. Mathematically, you are saying that (100-50)=50 DN’s are going to be packed into 256 gray levels, DN’. Furthermore, the center of the distribution will be 128 DN’. DN’=gain *DN+offset So the amplification factor or gain will be 256/(100-50)=5.12: DN’=5.12*DN+offset Now if we take 75 DN, the central value that we want to be 128, and multiply it by 5.12, we get 384 DN’, so we need to subtract 256 to get the right answer: DN’=5.12*DN-256. Check: DN’=5.12*50-256 = 0; DN’=5.12*75-256=128; DN’=5.12*100-256=156 Calibration: We measure radiance in DNs, but we want to know reflectance. So we can take a known target (say, black and white cardboard with reflectances measured in the lab of 5 and 25%) and image them to find out what radiance DN’s they give (say, 13 and 47, respectively). Then we can do a controlled contrast stretch to give the image in reflectance units: Now, the gain will be DDN /Drefl = (25-5)/(47-13)=0.59 (That is, refl=0.59*DN+offs, and we find offset by Knowing 0.59*13=5, or offset = 5-0.59*13=5-7.67=-2.67, so refl=0.59*DN-2.67. Check: 25=0.59*47-2.67=25.06 (roundoff) Calibration is just a special kind of contrast stretch

  5. Geometric registration Pixel locations in original and corrected images DN values in corrected image are found by interpolation from the nearest neighbors in the acquired image Acquired image, distorted Map with locations of control points

  6. Image Processing 3) Spectral analysis algorithms are based on the relationship of DNs within a given pixel Color enhancement Spectral transformations (e.g., PCA) Spectral Mixture Analysis 4) Statistical algorithms characterize or compare groups of radiance data Estimate geophysical parameters Spectral similarity (classification, spectral matching) Input to GIS

  7. Image Processing 5) Modeling calculate non-radiance parameters from the radiance and other data Estimate geophysical parameters Make thematic maps Input to GIS

  8. Image Processing There is a dazzling array of things for the future professional to become familiar with I’m trying to over-simplify it to begin with Most algorithms are handled pretty well in most remote-sensing texts. Spectral Mixture Analysis is an exception, so… - we’ll look at Spectral Mixture Analysis next lecture

  9. Image Processing Sequence(single image) Raw image data 1. Image display/inspection 2. Instrument calibration Image rectification, cartographic projection, registration, geocoding 3. Pre-processing 4. Atmospheric compensation Pixel illumination-viewing geometry (topographic compensation) 5. Working image data

  10. Working image data Product Image Processing Sequence(single image) 6. Further image processing 7. Spectral analysis Selection of training data/endmembers 8. Processing Initial classification or other type of analysis 9. Interpretation/verification or further analysis 10.

  11. Ratios in 2-space TM4 TM3 TM4 Ratio – 11 sunlit Ratio – 1.5 Ratio - 1.1 shadowed shadow TM3

  12. Ratios The Vegetation Index (VI) = DN4/DN3 is a ratio. Ratios suppress topographic shading because the cos(i) term appears in both numerator and denominator.

  13. NDVINormalized Difference Vegetation Index DN4-DN3 is a measure of how much chlorophyll absorption is present, but it is sensitive to cos(i) unless the difference is divided by the sum DN4+DN3.

  14. Dimension rotation y y’ x’ 0.7x,0.7y -0.7x, 0.7y + x y 0.5x,0.87y -087x,0.5y + x y x’ 0x,1y + x y’ -1x,0y

  15. Principal Component Analysis (PCA) • Designed to reduce redundancy in multispectral bands • Topography - shading • Spectral correlation from band to band • Either enhancement prior to visual interpretation or pre-processing for classification or other analysis • Compress all info originally in many bands into fewer bands • http://en.wikipedia.org/wiki/Principal_component_analysis

  16. Principal Component Analysis (PCA) - The math behind the button [ ] PC1 PC2’ = In the simple case of 45º axis rotation, Finding q [ ] [ ] [ ] [ ] cos q sin q -sin q cos q DN1’ DN2’ DN1 DN2 n11 n12 n21 n22 = cov = q = 45º Cov’=RTcovR; cov’ is the matrix having eigenvalues as diagonal elements and RT is the transpose of R. Eigenvalues can be found by diagonalizing cov. R has eigenvectors as column vectors http://www.cs.otago.ac.nz/cosc453/student_tutorials/principal_components.pdf http://en.wikipedia.org/wiki/Principal_component_analysis

  17. Principal Component Analysis In the simple case of 45º axis rotation, The rotation in PCA depends on the data. In the top case, all the image data have similar DN2/DN1 ratios but different intensities, and PC1 passes through the elongated cluster. In the bottom example, vegetation causes there to be 2 mixing lines (different DN4/DN3 ratios (and the “tasseled cap” distribution such that PC1 still passes through the centroid of the data, but is a different rotation that in the top case. PC1 PC2

  18. Tasseled Cap Transformation • Transforms (rotates) the data so that the majority of the information is contained in 3 bands that are directly related to physical scene characteristics • Brightness (weighted sum of all bands – principal variation in soil reflectance) • Greenness (contrast between NIR and VIS bands • Wetness (canopy and soil moisture)

  19. Tasseled Cap Transformation (TCT) TCT is a fixed rotation that is designed so that the mixing line connecting shadow and sunlit green vegetation parallels one axis and shadow-soil another. It is similar to the PCT. Soil Green

  20. Next lecture – Spectral Mixture Analysis

More Related