Rational forest management in today’s changing world is an important element of sustainable development. Forest ecosystems comprise a significant carbon pool and play a substantial role in the carbon exchange between the land and the atmosphere through the processes of photosynthesis, respiration and decomposition. For this reason, forests have an important resource, recreation and conservation function, and they maintain and stabilize the climate system. This is why the REDD Programme was launched in 2008 under the auspices of the United Nations (
In recent years, there have been numerous works undertaken to study forest stand structure and their bioproductivity using synthetic aperture radar (SAR) systems (Champion et al. 2013, 2014; Balzter et al. 2007; Neumann et al. 2010; Santoro et al. 2011, 2013; Stankevich et al. 2017). The results have shown that radar surveys can be used successfully to improve forest inventory. Unlike optical systems, radar ones are able to penetrate the forest cover and allow us to estimate the forest stand structure. Therefore, combining these two types of systems is very promising for improving the estimation of forest biomass (Cartus et al. 2012; Hyde et al. 2006; Attarchi and Gloaguen 2014). Based on the integration of multi-dimensional models of forest ecosystems, multi-sensor remote sensing concepts and ground data, a comprehensive quantification of forest cover and its parametrization is provided with uncertainties acceptable for policy-making (Schepaschenko et al. 2014).
This research aimed to study and comparexx1 various approaches of optical and SAR images merging for aboveground forest biomass modelling.
The main source of data for ground-based estimation of Ukrainian forest biomass parameters is the forest inventory carried out by forest enterprises. Previous assessments of Ukrainian forest live biomass have been undertaken by various researchers (Lakyda 2002; Lakyda et al. 2011, 2012). In these cases, the aboveground forest biomass has not been measured directly during the forest inventory but computed using allometric models (Shvidenko et al. 1987, 2007; Lakyda 2002; Lakyda et al. 2012).
Algorithms for modelling aboveground forest biomass have been developed by Lakyda et al. (2011). The models for the assessment of live aboveground biomass and the carbon content of different components of the forest stand are based on the relations between forest productivity and the main inventory parameters of the forest stand (e.g. age, diameter and height) (Lakyda et al. 2012; Cortés et al. 2014; Lu et al. 2004; Singh et al. 2014). It has been shown that the relation between live biomass and the main biometric indicators of forest stands can be expressed as follows:
i - a component of the forest stand (stem, branches, leaves, etc.),
For example, aboveground biomass of deciduous forests of the Ukrainian Polissya was estimated using the following equation (Lakyda et al. 2012):
The biomass for the main fractions of trees (stem (st), branches (br) and leaves (lv)) has been calculated using equation (2), while the biomass of the crown (
To convert this biomass into carbon, coefficients of carbon content in the corresponding fraction are utilized. From the literature, Lakyda et al. (2012) have assumed values of 0.5 and 0.45 for wood and leaves respectively. Thus, the equations for calculating carbon stock in the forest stand are given as follows:
Optical remote sensing of forests is based on developing relationships between spectral signatures and measured parameters. Multispectral satellite images provide the spatial distribution of land cover reflectance in the visible, near-infrared and infrared spectral ranges. These reflectance data can be used directly as well as for VI calculations. Depending on the complexity, such indices can be divided into simple ratios (SR), normalized differences (ND) and complex vegetation indices (Lu et al. 2004). A wide set of
Some spectral vegetation indices used for vegetation cover estimation
Index | Equation | Reference |
---|---|---|
Simple ratio | ||
TM4/3 | TM4/TM3 | Jordan, 1969 |
TM5/3 | TM5/TM3 | Lu et al., 2004 |
TM5/4 | TM5/TM4 | Lu et al., 2004 |
TM5/7 | TM5/TM7 | Lu et al., 2004 |
Normalized different vegetation indices | ||
NDVI | (TM4 – TM3)/(TM4 + TM3) | Rouse et al., 1974 |
NDVI53 | (TM5 – TM3)/(TM5 + TM3) | Lu et al., 2004 |
NDVI54 | (TM5 – TM4)/(TM5 + TM4) | Lu et al., 2004 |
NDVI57 | (TM5 – TM7)/(TM5 + TM7) | Lu et al., 2004 |
NDVI32 | (TM3 – TM2)/(TM3 + TM2) | Lu et al., 2004 |
GNDVI | (TM4 – TM2)/(TM4 + TM2) | Gitelson et al., 1996 |
NDII | (TM4 – TM5)/(TM4 + TM5) | Hardisky et al, 1983 |
NDII7 | (TM4 – TM7)/(TM4 + TM7) | Hardisky et al, 1983 |
NDWI | (TM2 – TM5)/(TM2 + TM5) | Lacaux et al., 2007 |
NDWI7 | (TM2 – TM7)/(TM2 + TM7) | Lacaux et al., 2007 |
Complex vegetation indices | ||
SAVI | (TM4 – TM3)/(TM4 + TM3 + L)(1 + L) | Huete, 1988 |
EVI | G(TM4 – TM3)/(TM4 + C1TM3 – C2TM1)+L | Huete, 1997 |
ARVI | (TM4 – 2TM3 + TM1)/(TM4 + 2TM3 – TM1) | Lu et al., 2004 |
GEMI | ξ(1 – 0.25ξ) – ((TM3 – 0.125)/(1 – TM3)) | Lu et al., 2004 |
ASVI | ((2TM4 + 1) – √((2TM4 + 1)2 – 8(TM4 – 2TM3 + TM1)))/2 | Lu et al., 2004 |
MSAVI | ((2TM4 + 1) – √((2TM4 + 1)2 – 8(TM4 – 2TM3)))/2 | Lu et al., 2004 |
Note: TM(X) is corresponding Landsat TM/ETM+ spectral band
The enhanced vegetation index (EVI) incorporates this atmospheric resistance concept in the atmospheric resistant index (ARVI), along with the removal of soilbrightness-induced variations in VI as in the soil adjusted vegetation index (SAVI). The EVI additionally decouples the soil and atmospheric influences from the vegetation signal by including a feedback term for simultaneous correction.
Lu et al. (2004) showed that the relationships between TM reflectance and forest stand parameters vary depending on the characteristics of the study areas. However, not all vegetation indices are significantly related to forest stand parameters. What is crucial is the selection of suitable TM band (s) and
Unlike optical remote sensing, which allows for estimating only the top of the forest cover and does not provide any information about forest structure (due to impermeability of the optical waves), radar remote sensing uses the microwave portion of the electromagnetic spectrum. Canopy penetration varies with different wavelengths. Shorter wavelengths (e.g. X-band imagery at 3 cm) are reflected from the top of the canopy, while longer wavelengths (e.g. L-band imagery at 24 cm) go down to the ground and are reflected from there. Using these properties of different wavelengths makes it possible to discern information about canopy structure of a forested area from a multi-wavelength image and thus estimate the different component of AGB.
One more useful feature of radar data for studying forest structure is polarization. Transmitted and received radar signals are propagated in a certain plane. The propagation planes are usually horizontal (H) and vertical (V). Vertically polarized waves interact with the vertical stems of the forest cover, while horizontally polarized waves penetrate through the canopy. Thus, the combination of images with different channels can provide additional information about forest structure.
As with optical remote sensing data, two sets of texture features can be calculated using SAR data. The first set of calculations can be derived from the backscatter (sigma nought, σ0) distribution and the second one from the grey-level co-occurrence matrix. The intensity scenes of SAR images are converted in their corresponding backscattering coefficient (σ°) values using the following equation (Attarchi et al. 2014; Shimada et al. 2009):
To take into account the effects of relief, the topographic normalized backscattering coefficient (the corrected backscatter in gamma-nought
The exponent
Previous studies have shown that the forest backscatter can be described as a function of the growing stock volume,
The backscatter model in Eq. (9) contains three unknowns that need to be estimated:
Data fusion is a general multi-discipline approach. It combines data from multiple sources to improve the potential values and interpretation performances of the source data and to produce a high-quality visible representation of the data.
In general, remote sensing fusion techniques can be classified into three different levels: the pixel/data level, the feature level and the decision level (Pohl and van Genderen 1998; Zhang 2010). Pixel-level fusion is the combination of raw data from multiple sources into single resolution data, which is expected to be more accurate than either of the individual input data or it may reveal the changes between data sets acquired at different times (Zhang 2010).
Feature-level fusion extracts various features, e.g. edges, corners, lines and texture parameters, from different data sources and then combines them into one or more feature maps that may be used instead of the original data for further processing. This is particularly important when the number of available spectral bands becomes so large that it is impossible to analyse each band separately. Typically, in image processing, such fusion requires a precise (pixel-level) registration of the available images. Feature maps thus obtained are then used as inputs to pre-processing for image segmentation or change detection (Zhang 2010).
Decision-level fusion combines the results from multiple algorithms to yield a final fused decision. When the results from different algorithms are expressed as confidences (or scores) rather than decisions, it is called soft fusion; otherwise, it is called hard fusion. Methods of decision fusion include voting methods, statistical methods and fuzzy logic-based methods (Zhang 2010).
There are a set of multi-source remote sensing data fusion methods, which are based on different techniques: Markov random field (MRF) (Solberg et al. 1996), support vector machines (SVM) (Waske and Benediktsson 2007) and the decision fusion approach for multi-temporal classification (Jeon and Landgrebe 1999). However, the different nature and content of remote sensing imagery and GIS data prevent a direct comparison. Therefore, the integration of data from different applications must address the differences in the object model and the semantics of the objects themselves (Zhang 2010). Images are usually composed of a raster of pixels representing the intensities, whereas GIS data contain artificial objects (points, lines and polygons) with label forms, representing the objects or region affiliations. To combine segmented objects or primitives from remote sensing images and GIS data at the feature level or decision level, traditional pattern recognition methods can be used and have demonstrated their potential capabilities, e.g. knowledge-based techniques (Amarsaikhan and Douglas 2004), neural network and statistical approaches (Benediktsson and Kanellopoulos 1999), fuzzy set theories (Fauvel et al. 2006), Bayesian techniques and Dempster-Shafer-based methods (Zhang 2010).
The field data were collected during detailed forest inventory supported by the National University of Life and Environmental Sciences of Ukraine. The study area is located in Chernihiv region (Ukrainian Polissya) (lat., long. of the area centre are 52.070556 N, 31.839722 E), and it covers an area of over 60 sq. km. Within this site, the measurements of forest parameters by ground-based methods have been performed. The dataset contains the results of the measurements of forest live biomass structure. The data are presented in the same units and organized as a unified structure. The dataset is designed for studying the biological production of Ukrainian forests under global change. It has been used for (1) modelling of fractional structure of forest live biomass in Ukraine based on data from the forest inventory (State Forest Account) and (2) the development of models and tables of dynamics of biological productivity of forests. For these reasons, the dataset contains detailed biometric indicators of forest stands within the sample plots.
The dataset contains the following information for each test plot:
Geographical location (administrative region, forestry, plot number, plot area and geographical coordinates (if available)) Forest type Biometric characteristics of stands (dominant species, species composition, number of model trees, age, average height and average stem diameter, number of trees per 1 ha, absolute and relative stocking, growing stock volume and AGB of different components of the forest stand (stem, branches, leaves, bark and undergrowth)
There are more than 150 and more than 200 samples for coniferous forest and for softwood species respectively. This provides an adequate sample for building the models and their validation.
The RapidEye multispectral image from July 1, 2010, and the PALSAR radar image from August 2, 2009, were selected as sources of remote sensing data.
PALSAR is L-band SAR sensor. We used imagery of SLC Fine Beam Double polarization HH/HV product (level 1.1) with a 12-meter resolution. In order to use SAR data in a quantitative fashion, we applied a range of pre-processing steps. First, radiometric calibration has been done to convert radiometry default value of amplitude (the pixel values in the image are raw digital numbers) into sigma value. Second, terrain correction has been done to remove geometry-induced distortions and geocoding has been performed to transform the image from the SAR geometry into UTM projection. Finally, speckle filtering has been applied using Lee-Sigma filter.
RapidEye optical system provides five spectral bands (440–510 nm (blue), 520–590 nm (green), 630–685 nm (red), 690–730 nm (red edge) and 760–850 nm (near infrared)) with a 6.5-meter spatial resolution. We used RapidEye Ortho Tile Product (Level 3A). In this product, radiometric and sensor corrections applied to the data as well as imagery are orthorectified using the RPCs and an elevation model. Therefore, we additionally performed only atmospheric correction of the image and converted its spatial resolution to adapt it to SAR image.
Using the RapidEye multispectral image, the average reflectance for each spectral band, as well as the average amplitude value for HH and HV polarimetry modes in case of the PALSAR image, was calculated for each forest sample plot. The obtained values were compared with the aboveground forest biomass data from field data set. This preliminary analysis shows that forest AGB values have either exponential or power relation with spectral reflectance and SAR amplitude value. It also reveals that the saturation level of informative signal for the optical spectral bands is achieved at the value of forest AGB about 50 t/ha. This is due to the above-mentioned limitations of optical systems. It means that at the forest AGB value of 50 t/ha and above, the forest canopy is too dense to be penetrated by optical system. In case of the PALSAR image, such saturation level of radar signal is achieved at the value of forest AGB about 100–150 t/ha.
However, one of the disadvantages of radar data is that
HPF is one of the first developed image merge methods (Schowengerdt 1980, 2007a, b; Chavez et al. 1991; Pohl and van Genderen 2016), which has been used for image fusion for more than 30 years. It was primarily designed to improve the resolution of multispectral images by transferring spatial details derived from a higher resolution PAN or SAR image to a lower resolution multispectral image. This method is performed in three stages (Pohl and van Genderen 2016):
High-pass filtering the high-resolution SAR image Add the high-pass filtered image to each multispectral band using individual weights depending on the standard deviation of the MS bands Match the histograms of the fused image to the original MS bands
Thus, this method extracts high-frequency information from the SAR image, which is then added to each spectral channel of the multispectral image.
Figure 1a shows the result of merging the PALSAR image (HH mode) with RapidEye multispectral image using HPF method.
IHS transform is one of the most popular methods of merging remote sensing images. It uses a mathematical colour model based on a cylindrical or spherical coordinate system. This method effectively separates spatial (I) and spectral (H, S) information from a standard RGB image (Pohl and van Genderen 2016).
There are two ways to use IHS transformation to merge images: direct and substitutional. The first way is to directly convert the three spectral channels of the image into I, H and S. The second method involves converting the three spectral channels from the RGB into IHS colour model, where colour aspects are separated by its average brightness. The hue and saturation in this case are related to the surface reflectivity or composition. The SAR image then replaces one of the components, and reverse transformation from IHS to RGB converts the data into their original colour model to produce a new synthesized image.
Figure 1b shows the result of merging the PALSAR image (HH mode) with a multispectral RapidEye image using IHS method.
Wavelet transform is another powerful mathematical tool used for signal processing. It is used to decompose a function or signal into several components (coefficients). When using this method to merge images, the main idea is to decompose the original images into a number of fragments using direct wavelet transform. The information is merged on the basis of the obtained coefficients of these fragments, and the inverse wavelet transform is applied to synthesize new image. This method is suitable for merging images from different sources with varied physical backgrounds (such as radar and optical images) because it decomposes images into different types of coefficients, preserving the source information. Based on these coefficients, a new image can be synthesized using inverse wavelet transform (Pajares and Cruz 2004).
Figure 1c shows the result of merging the PALSAR image (HH mode) with RapidEye multispectral image using WT method.
Hybrid data merging methods are widely used to compromise between spatial and spectral optimization. WHIS combines the IHS transformation approach with methods used to merge data with different spatial resolution, such as WT (Chibani and Houacine 2002; Gonzalez-Audicana et al. 2004; Zhang and Hong 2005). During wavelet-IHS transformation (WIHS), the multispectral data are converted to IHS colour model. Then, the intensity component I is decomposed using WT method. The SAR image is matched to I and then also decomposed by the WT method. The decomposed component I is replaced by the SAR decompositions, and the inverse IHS transformation is performed (Pohl and van Genderen 2016).
Figure 1d shows the result of merging the PALSAR image (HH mode) with RapidEye multispectral image using WHIS method.
The field data set of forest AGB was divided into two samples with the proportion of 70/30. The first sample (70% of data) was used to build models, which accuracy was further assessed using the second sample (30% of data). Regression analysis methods were used to find a set of relationships between forest AGB and varied combinations of remote sensing data for each approach. As a final result, one model that has the highest accuracy was selected for each approach.
In case of using HPF approach, the best result was achieved for merged HH PALSAR band with 4th RapidEye spectral band. It is described by the following equation (Fig. 2):
HPF(HH + B4) – the signal from fused HH PALSAR band with 4th RapidEye spectral band using HPF approach.
In case of using IHS approach, the best result was achieved for merged HH PALSAR band with 4th RapidEye spectral band. It is described by the following equation (Fig. 3):
IHS(HH+B4) – the signal from fused HH PALSAR band with 4th RapidEye spectral band using IHS approach.
In case of using WT approach, the best result was achieved for merged HH PALSAR band with 4th RapidEye spectral band. It is described by the following equation (Fig. 4):
WT (HH+B4) – the signal from fused HH PALSAR band with 4th RapidEye spectral band using WT approach.
In case of using hybrid WIHS approach, the best result was achieved for merged HH PALSAR band with 4th RapidEye spectral band. It is described by the following equation (Fig. 5):
WIHS (HH+B4) – the signal from fused HH PALSAR band with 4th RapidEye spectral band using WIHS approach.
In case of using MLR approach, the best result was achieved for merged HH PALSAR band with 4th RapidEye spectral band. It is described by the following equation:
B4 is the reflection coefficient of the 4th spectral band of the RapidEye image and HH is the amplitude of the reflected radar signal in the HH mode.
The accuracy of the obtained models was assessed with a test data set, which was not used for the models building. The modelled data were compared with the data of field measurements, and the correlations between measured and modelled data were calculated for each model (Tab. 2).
Parameters of the models for the accuracy estimation
Fusion techniques | R2 | Correlation | MMA |
---|---|---|---|
MLR (RapidEye, 4th band ++ PALSAR, HH band) | 0.698 | 0.837 | 0.575 |
IHS (RapidEye, 4st band + PALSAR, HH band) | 0.530 | 0.731 | 0.591 |
HPF (RapidEye, 4th band + + PALSAR, HH band) | 0.643 | 0.804 | 0.652 |
WT (RapidEye, 4th band + PALSAR, HH band) | 0.370 | 0.614 | 0.623 |
WIHS (RapidEye, 4th band + +PALSAR, HH band) | 0.476 | 0.694 | 0.604 |
The correlation shows the relationship between the measured and modelled values. The index lays in a range from −1 to 1. Higher correlation means better model performance. A correlation of 0 indicates no relationship between the measured and modelled values.
Also, to assess the accuracy of each model, the min–max accuracy (MMA) was calculated:
This parameter shows how close the predicted values are to the actual ones. The MMA value ranges from 0 to 1, where a value of 1 indicates a perfect match between actual and predicted values. So, higher MMA score means better model performance.
Among the different methods of data fusion used in the research, the HPF method (correlation 0.804, MMA 0.652) showed the best result with a significant advantage over other methods.
Nowadays, data from both passive and active remote sensing are becoming more and more available. They have different nature, and accordingly, they have different advantages and disadvantages. Therefore, their joint using for the land cover studying and vegetation parameters assessing (including forest cover) can significantly improve the results. This research aimed to study and compare various approaches of optical and SAR images merging for forest AGB modelling. Five models for estimating forest AGB were built and analysed using data from test area in Chernihiv region (Ukrainian Polissya). Obtained results confirm conclusions from previous studies (Santoro et al. 2013) about low accuracy of aboveground biomass modelling using SAR data due to the speckle effect. The merging of optical and SAR data significantly increases the accuracy of the simulation, and among all the data fusion approaches used in the study, the method of high-pass filtering (HPF) has shown the greatest efficiency.