Unsupervised Landcover Classification In SNAP Using Sentinel 1 Imagery

From CUOSGwiki
Jump to navigationJump to search

Introduction

Synthetic Aperture Radar, or SAR, is a method of RADAR imaging that propagates a microwave signal to create an image over an area. The image is generated based on the type of backscatter that occurs when the signal hits a target. Unlike optical remote sensing, SAR is not affected by clouds or time of day; however, it can be adversely affected by heavy precipitation. Sentinel 1 is a constellation of two polar orbiting satellites that use C-band (3.8-7.5 cm) wavelengths to generate imagery of the earth's surface. This tutorial will use SAR data and image processing software to classify a RADAR image using an unsupervised classification method.

Unsupervised Classification

Unsupervised classification aggregates spectral classes (pixel values), or clusters, in a multiband image based on pixel value alone. This differs from supervised classification, where the user provides training data that the algorithm can use to assess the class to which a pixel belongs. For unsupervised classification, the user instead tells the image classifier how many “clusters”, or classes, are desired, and what method to use.

In this tutorial, we will be using a K-means classifier for the clustering algorithm.

K-means is an algorithm that combines “observations” (in this case, pixels) into discrete groups. It does this by creating nodes, which represent the centre of a data cluster. This node is positioned so that the distance between it and the nearest points is on average smaller than the distance between those points and the next node (Hunt-Walker & Ewing, 2014).

SNAP

The Sentinel Application Platform, or SNAP, is an open source imagery analysis software provided by the European Space Agency. Along with the base tools, there are a number of other toolboxes available for more complex analyses. It can be found here: http://step.esa.int/main/download/snap-download/previous-versions/

SNAP 7.0 has a bug in the software that effects a component of classification; until this is patched, SNAP 6.0 should be used. The European Space Agency recommends SNAP be run on computers with at least 4GB of memory; however, for the classification, 16 GB would be preferred.

This tutorial was built on Windows 10, with 32GB of RAM and a 64-bit operating system. The classification component of the tutorial took over an hour with these specifications.

Other optional software

Google Earth (v7.3) https://www.google.com/earth/versions/#download-pro

Operating System: Windows 7 or higher CPU: 2GHz dual-core or faster System Memory (RAM): 4GB Hard Disk: 4GB free space High-Speed Internet Connection Graphics Processor: DirectX 11 or OpenGL 2.0 compatible

Data

To acquire data, you will need to register an account with the Alaska Data Search, found here: https://search.asf.alaska.edu/#/?flightDirs. To do so, go to the link and select “sign in” then “Register”. Follow the instructions. This example uses the Bosnian and Herzegovinian and Croatian coastline. If desired, you can download several images and “stack” them. This process will be shown as well. Although SAR is purported to be "all weather", precipitation and sandstorms can adversely effect RADAR signals, so you may wish to check the weather of your AOI.

Asf images ajl.png


Find the area of interest that you wish to classify.

There is no need to use the exact image used in this tutorial, although the steps do require Sentinel 1 GRD format. GRD format can be input in the "Filters" window. If you do wish to use the image in the tutorial, parameters can be found below.


Example data Parameters

Search parameters:

Search type: Geographic

Dataset: Sentinel-1

Area of interest: POLYGON((17.6743 42.9283,18.3897 42.9283,18.3897 43.3711,17.6743 43.3711,17.6743 42.9283))

Start date:5/5/2019

End Date:6/13/2019

Under filters, select "L1 Detected High res Dual Pol(GRD)"

Frames: 140


Asf ajl.png

Methods

The steps to complete the classification are as follows:

Set up your workspace by creating a folder on the drive you wish to work in. Save your images to this folder.

Open SNAP.

Import your data into SNAP using the “Import Product” button, and navigate to your working folder.

To get a feel for what your data looks like, double click the imported product to expand, then the “Bands” folder. Double click any band, and an image will appear of that band's backscatter imagery.


Hv ajl.png


The band you open should look somewhat similar in composition to the image above.

As well, you can open the metadata folder in the product menu and examine characteristics relating to geographic location,satellite orbit, satellite incidence angle, and others. As you preprocess the image, this data will change.

Preprocessing

1.Subsetting

The next step will be to subset your image. Because of the time it takes to do unsupervised classification, you may wish to make a very small image.

Navigate to Raster->subset

As shown in the images below, it is easiest to make a subset by dragging each side of the blue box on the left side of the Specified Product Subset window to the size you want, then selecting the middle and moving it to the desired area within the thumbnail image. Once you are satisfied, select the “OK” button.


Subset AJL.png


Subset2 AJL.png


If you wish to incorporate other images, you will need to copy either the Scene Start and End pixel coordinates, or the Geo-Coordinates, in order for your data to stack properly.

After you have completed the subset, right click the subset product in the Product Explorer pane and select "save product as" and save to your working directory. Repeat on all your chosen images. An RGB version of the tutorial subset image can be seen below. To view an RGB image, right click on the product and select "open RBG window".


Once this is complete, it is time to preprocess your image. Preprocessing is necessary because a raw SAR image can contain missing or incorrect geometric data.

Ensure you are writing to the working directory for all preprocessing and classification steps, and performing each step on each subsequent image.


2. Application of Orbit File

Over time, satellites can drift, and their orbit must be corrected. This can effect the recorded location of the image. To account for this orbital drift, SNAP can download precise satellite location information. To apply an orbit file, navigate to the Radar->Apply Orbit File menu. The default of "Sentinel Precise" setting in the orbit state vectors dropdown menu does not need to be changed, but "Do not fail if new orbit file is not found" must be toggled.


Orbit AJL.png


3. Backscatter Conversion

Radiometric calibration normalizes variations in RADAR system responses.

Navigate to the Radar->Radiometric->Calibrate menu.

Select both polarization bands on the processing parameters tab, and ensure the "sigma0" button is toggled.

Press Run.

Calibration AJL.png


4. Speckle Filter

Speckle is "noise" in a SAR image. Filtering suppresses noise effects and makes an image clearer. The type and size of filter depends on application. We will use a 7x7 Gamma Map, as this is what Agriculture and Agri- Food Canada uses for image classification.

Navigate to the Radar->Speckle Filtering-> Single Product Speckle Filter menu.

Select all processing parameter bands.

find and select "Gamma Map" in the filter dropdown menu.

Type 7 in both the X and Y fields.

Press Run.

Speckle ajl.png


5. Geometric Correction

Terrain correction corrects for satellite tilt so that the geometric representation of the image is approximate to real-world topography.

Navigate to the Radar->Terrain Correction-> Range-Doppler Terrain Correction menu.

Under the "Processing Parameters" tab, select all source bands.

Toggle the "Apply radiometric normalization" button.

Leave all other values as default.

Press run.

Terrain ajl.PNG

If you only intend to run the classification on one image, skip the next step. If not, continue on to Coregistration.


6i. Coregistration

If you intend to run the classification on multiple images, you will need to "stack" them. To do this, navigate to the Radar->Coregistration->Coregistration menu.

Add all your subset and preprocessed images by selecting either "add" (symbolized in 6.0 as a + sign), which will open a "Open Product" window, allowing you to navigate to your working folder and add the products of interest by selecting them and clicking "ok"; or "add opened" (symbolized as a + with a square background), which will add all open products to the ProductSet-Reader tab.

Coreg ajl.PNG

In the Coregistation window, select the CreateStack tab next, and select "Bilinear Interpolation" in the resampling type dropdown menu, and press "find Optimal Master".

Create stack ajl.PNG

Do not make changes on the Cross-Correlation and Warp tabs. In the "Write" tab, save to your working directory.

Cross-correlation ajl.PNG

Warp ajl.PNG

Write ajl.PNG

Press Run.

6ii. Texture Analysis

Due to the nature of SAR data, texture is critical to accurate classification. Therefore, we will use the Grey Level Co-occurrence Matrix texture analysis to generate texture bands. This can be a time consuming process, depending on the size of your image.

To apply the GLCM, navigate to Raster->Image Analysis->Texture Analysis-> Grey Level Co-occurrence Matrix

In the processing parameters tab, ensure all your bands are selected, and choose 7x7 in the windows dropdown menu. All other variables can remain as default.

Press Run.


Texture ajl.PNG


The RGB result can be seen below.

Glcm ajl.PNG

Once the GLCM is complete, you can apply the classification.


7. Classification

To run the classification, navigate to the Raster->Classification->Unsupervised Classification-> K-Means menu.

Select the desired number of clusters; for this example on the Bosnian and Herzegovinan and Croatian coast, there were three clusters of interest, but depending on your area of interest there can be less or more. Both number of iterations and random seed can be left as their default values. Select all the source band names you wish the classifier to use. Press Run.

Note that this can be quite time consuming; however, if it has no progress after several hours, you may need to subset your image smaller, or reduce the amount of bands used for classification. The result should look similar to the image below; however, colours may differ depending on your AOI and the cluster parameters.

Classified ajl.PNG

Colours and cluster names can be changed in the Colour Manipulation tab.

Colour ajl.PNG


8. Validation

Note that, if this were an optical image, the water would all be a single colour. However, since SAR interacts with textures as well as water content, this will effect the classification. As well, there is the chance of miss-classified pixels. To validate the image:


1. export as kml by right clicking the classified image and select "export view as Google Earth KMZ", and save. If you open the kml file, it will open in Google earth and you can visually assess the classes.

Googleearth ajl.jpg

The image above shows that the teal is shallow water, yellow and dark blue are deeper water, and purple is vegetation. The grey "null" values are human infrastructure, such as buildings.


Conclusion and Citations

You can now preprocess GRD Sentinel 1 data, and perform an unsupervised K-means classification on dual-pol SAR imagery in SNAP.

Citations: Special Topics in Geomatics. Hosseini, M., & Mcnairn, H. (n.d.). Ottawa.