1. Introduction
The Brazilian Rainforest in the Amazon was nearly intact until 1970, when the construction of Trans-Amazonian Highway triggered the high level of deforestation rates. Since then, the rate of deforestation in the Legal Amazon [
2] has oscillated. However, the numbers have always been elevated [
3]. According to Instituto Nacional de Pesquisas Espaciais—INPE (National Institute for Space Research), approximately 16% of the Forest was destroyed, that is, out of its 3.5 million km², over 550,000 km² had been deforested. In 1978, INPE carried out a survey of the forest for the first time using a satellite data. They found out that as much as 140,000 km² had been cleared. In the following years, no other survey was done, because the government’s agenda did not prioritize environmental preservation issues. Not until 1988 did INPE started to carry out annual surveys, due to international concern regarding the way the Amazon Forest was being taken care of [
4].
Considering the elevated rates of deforestation, it becomes clear that some action to help monitor the region was necessary, which would increase the authorities’ control over the area. Some tools have been developed, aiming towards improvement in monitoring. Nevertheless, none of them were used daily. As an example of the main projects that have been developed and used by INPE, we can name PRODES and DETER.
PRODES methodology aims to produce estimates of deforestation levels of annual clear cut in Legal Amazon, through digital classification of images. Such methodology was described by [
5]. It is an important tool for obtaining estimates. However, it is inefficient as surveillance is regarded, because the estimates are produced annually.
The DETER Project creates maps with the location of the areas that are being cleared, using photo interpretation of images acquired by MODIS sensor (Moderate Resolution Imaging Spectroradiometer) aboard the Terra satellite. These maps have information about the dynamics of the deforestation in Legal Amazon. Such information is sent to entities and departments like IBAMA (Instituto Brasileiro do Meio Ambiente e Recursos Naturais Renováveis) [
6] fortnightly and supports the surveillance and control of deforestation.
Ideally, as far as surveillance is regarded, there should be a detection system running daily, which would make it possible to forecast any changes in the Forest, increasing the surveillance and reducing loss in the areas surrounding the deforested location. Within this framework, an artificial intelligence technique that seems to present great potential for fast classification of satellite images is Artificial Neural Network, which is applied to the images of MODIS/TERRA sensor and might contribute to the daily detection of deforestation zones [
7,
8,
9]. According to [
10] an Artificial Neural Network (ANN) can be seen as a group of artificial neurons that have the capacity to process data locally. A connection topology and a learning rule define how these neurons are connected. [
11] adds that an ANN is a processor distributed in parallel through simple processing units (neurons), which have the capacity to store knowledge and use it to solve complex problems. In this sense, we propose the creation of a tool for the training and use of artificial neural networks, so that a faster classification of images from MODIS/TERRA sensor can be achieved.
2. Data and Methodology
Images of tile H11V09 of MODIS/TERRA sensor from 20th May–15th July 2003 were used for the testing and training of this tool. Fifty-seven images were used, which were numbered 140 to 196, according to the Julian calendar. The images were chosen because they present a ground truth and have been used in previous works, e.g., [
8,
9]. This tile includes the area of study in the State of Rondônia, according to the following geographic coordinates: 64° 16’ 24.19” and 62° 26’ 59.41” of longitude W and 9° 30’ 27.69” and 7° 50’ 28.62” of latitude S (see
Figure 1). To the pre-processing of the images, a tool called SPRING 5.1.5 (Sistema de Processamento de Informações Georeferenciadas-Georeferenced Information Processing System) was used [
12].
Figure 1.
Legal Amazon, outlined in bold the area of study. Source: Adapted from [
6].
Figure 1.
Legal Amazon, outlined in bold the area of study. Source: Adapted from [
6].
The software was implemented using the Java programming language in addition to AWT (Abstract Windowing Toolkit) and Swing components for the elaboration of graphical interface. Framework Encog [
13] was incorporated for the development of the neural network module, and the database management system MySQL Server was used for the storage of data related to the processed images.
The neural tool for deforestation detection was developed according to the methodology shown by [
8], aiming to detect daily degradation based on Terra Satellite sensor data. Firstly, the SPRING program produced three gray scale images divided into soil (red component), vegetation (green component) and shade fractions (blue component) from the 57 available images using the Spectral Linear Mixture Model. The fraction images were then turned into GeoTIFF [
14] format, which characterized the database used in this research.
With the georeferenced fraction images in GeoTIFF format, it became possible to develop a neural module and test it. The neural module takes in the same pixel of each one of the fraction images: soil, shade and vegetation. The outcome is expected to be one of the following five results: water/shade, savanna, deforestation, vegetation or clouds.
When the neural network classifies the area as deforestation, a visual alarm is fired, showing the geographic coordinates that the pixel represents. In order to avoid processing areas that have already been completely cleared out or areas that are known to be out of vegetation, a digital mask was created. Pixels that can be identified as savanna or completely cleared out are removed from this mask. It is then possible to analyze the pixels which are expected to show vegetation.
Due to the neural networks’ difficulty in determining the best architecture for solving the problem in question, freedom of parameterization was considered the best option. That is, it is possible to adjust the number of intermediate layers, through graphical interface, as well as the number of neurons in each layer, the number of epochs, expected error, and training algorithm (Back-Propagation and a variation of it, the Resilient Back-Propagation [
15]).
Within this scenario, the network will be configured at the time of training, but the number of input and output neurons will already be pre-configured due to the fact that they model the problem being addressed. The three input neurons represent the values of soil, shade and vegetation of each pixel of the gray scale generated fraction images. While the five output neurons map each class of interest, which are: water, savanna, deforestation, vegetation and clouds.
In order to find the best architecture for the problem in question, in each training program there was a different number of neurons in the intermediate layer. In addition, the number of epochs and the number of intermediate layers were also altered. As there is a possibility of there being performance differences between the two networks, even if the same training parameters are applied, all the networks were double-trained and the best result was used.
An automatic generator of confusion matrix was applied in order to determine the quality of the neural response. The confusion matrix shows the extent to which the image classifier mixes up each mapped class.
The data used for training were obtained randomly out of 240 pixels of the images dating 20 May and 14 July, first and last images of the data collection, which have the best quality among the available images. The data are divided into: 30 points of vegetation, 60 points of savanna, 60 points of deforestation, 30 points of shade/water and 60 points of clouds determining,ds.
The test data collection, used to certify neural response correction, is made up of 30 samples of each class, which totalize 150. It is necessary to mention that the selection of points for training and testing were collected randomly, at distinct times and are two disjoint sets.
3. Results and Discussion
According to what was previously explained in
Section 2, various training sections were conducted in order to obtain the best architecture for a problem. It should be pointed out, however, that it is impossible to guarantee that an artificial neural network is totally reliable for solving the given problem. A criterion commonly used in neural networks with supervised training’s algorithms is the Mean Square Error-MSE of the network output over the expected response to determine when the neural network is responding well and the training can be stopped. During the training processes that were conducted, the Mean Squared Error (MSE) set was 1%. However, it was not possible to achieve this number in any of the training sections. The fact that the network has not achieved the expected MSE percentage does not mean that the training has been faulty. The hit rates in
Table 1—whose value is as much as or over 90% for all training processes—show that the Network is responding correctly, although the stipulated MSE has not been achieved. The hit rates are calculated by the equation below.
In which nhits is the number of points that the Network has associated to correct categories and nsum is the total amount of samples used for the test with the trained Neural Network.
Table 1.
Set of Training Procedures and Parameters for the Neural Network.
Table 1.
Set of Training Procedures and Parameters for the Neural Network.
Training | Neurons (hidden) | Epochs | Achieved MSE | Hit Rates |
---|
1 | 5 | 1.000 | 29,47% | 92,67% |
2 | 5 | 2.000 | 29,63% | 94,00% |
3 | 6 | 1.000 | 29,60% | 92,67% |
4 | 6 | 2.000 | 28,81% | 93,33% |
5 | 7 | 1.000 | 30,90% | 93,33% |
6 | 7 | 2.000 | 29,60% | 93,33% |
7 | 8 | 1.000 | 30,83% | 92,67% |
8 | 8 | 2.000 | 26,41% | 93,33% |
9 | 9 | 1.000 | 27,78% | 92,67% |
10 | 9 | 2.000 | 27,48% | 94,67% |
11 | 9 | 3.000 | 26,44% | 94,00% |
12 | 9 | 10.000 | 25,07% | 92,67% |
13 | 9 | 20.000 | 20,58% | 92,00% |
14 | 10 | 1.000 | 27,74% | 94,00% |
15 | 10 | 2.000 | 27,65% | 92,67% |
16 | 11 | 1.000 | 29,63% | 93,33% |
17 | 11 | 2.000 | 27,00% | 93,33% |
18 | 15 | 1.000 | 27,93% | 94,67% |
19 | 15 | 2.000 | 28,49% | 92,00% |
20 | 30 | 1.000 | 24,72% | 93,33% |
21 | 30 | 2.000 | 22,43% | 90,00% |
22 | 50 | 1.000 | 23,95% | 92,67% |
23 | 50 | 2.000 | 19,31% | 92,00% |
24 | 50 | 20.000 | 14,95% | 90,67% |
The Artificial Neural Network which has classified the set of samples with the highest number of hits was represented by training process numbers 10 and 18 on the
Table 1. Training process number 18 obtained the same rate, but it shows a more complex structure. According to [
10] the complexity of the neural model defines the scope for possible solutions for a determined problem, to which the use of a greater number of neurons than necessary affects its power of generalization. For that reason, the least complex neural network, which also solves the problem satisfactorily, is considered the best. In order to evaluate Neural Network 10, a confusion matrix was produced (
Table 2). Confusion matrixes make it possible to verify confusion mapped classes. Another category called Undefined was added to cases in which there was excitation of more than one neuron or the inhibition of all of them, allowing us to define a particular mapped category.
The confusion matrix shows that the neural network successfully classified 142 pixels out of the 150. When there was confusion, it only involved pixels of deforestation and savanna. Out of a total of 30 deforestation samples, 25 were correctly classified by the network and 5 were confused with Savanna.
Table 2.
Confusion Matrix for the Neural Network 10.
Table 2.
Confusion Matrix for the Neural Network 10.
Category | Water | Cloud | Savanna | Deforestation | Vegetation | Undefined | Sum |
---|
Water | 30 | 0 | 0 | 0 | 0 | 0 | 30 |
Cloud | 0 | 30 | 0 | 0 | 0 | 0 | 30 |
Savanna | 0 | 0 | 27 | 3 | 0 | 0 | 30 |
Deforestation | 0 | 0 | 5 | 25 | 0 | 0 | 30 |
Vegetation | 0 | 0 | 0 | 0 | 30 | 0 | 30 |
Sum | 30 | 30 | 32 | 28 | 30 | 0 | 94.67% |
It is believed that confusion is due to the fact that savanna and deforestation have a similar spectral signature, which makes the classification process harder. Difficulty in category sorting similar spectral signature categories was also observed by [
7] and by [
16]. The latter conducted a comparative study between statistics technique and ANN, and found that both methods show difficulties in classification.
In order to demonstrate the difficulty in sorting out two categories with similar spectral signature, the same network was trained with exactly the same parameters, without, however, the points related to savanna. Three training sections were conducted and in all of them the network converged to the expected MSE before iteration number 500; the confusion between categories was nullified. It is important to emphasize that, due to fast convergence, the Mean Square Error stipulated was 0.1%, and not 1% as in previous training processes.
The confusion matrix for the neuron network produced with the same training data set, except for the savanna points, allows us to infer that the Savanna category causes confusion in the classification process. In that case, the hit rate of the artificial neural network was 100%, and there was no confusion in association with any other categories, as can be seen in
Table 3. It should be highlighted that this neural network was able to classify the five pixels which had previously been wrongly classified as savanna. Nonetheless, a neural network containing only savanna and deforestation samples was trained. It was observed that the MSE found was similar to those achieved in the training processes of
Table 1. The tests show that savanna and deforestation are two spectrally similar categories, which interferes with the neural process of classification by the neural network. It is possible to point out that, in addition to architecture, the sets of data and the mapped categories affect the good performance of the neural network.
The capacity of generalization of the neural network was, firstly, determined with the application of a set of tests and construction of the confusion matrix. However, it is necessary to determine the capacity that the neural network has to completely identify an image of MODIS/TERRA sensor. The first image of the set, dated from 20th May, was classified by the tool and compared with a thematic map (without clouds) of the same area generated by INPE [
4]. This image was chosen due to the fact it is on the NADIR and has no clouds in it, which is considered appropriate for the comparison of the produced maps. Overall, the classification obtained was satisfactory, and all the categories, including water were distinguished in the whole image.
Deforestation, however, was mistaken as savanna in some points, due to their spectral similarity. Also, there were some areas of savanna on the image that were misclassified as vegetation, possibly due to the lack of training samples of those areas.
Table 3.
Confusion Matrix for the Neural Network without Savanna Category.
Table 3.
Confusion Matrix for the Neural Network without Savanna Category.
Category | Water | Cloud | Deforestation | Vegetation | Undefined | Sum |
---|
Water | 30 | 0 | 0 | 0 | 0 | 30 |
Cloud | 0 | 30 | 0 | 0 | 0 | 30 |
Savanna | 0 | 0 | 0 | 0 | 0 | 30 |
Deforestation | 0 | 0 | 30 | 0 | 0 | 30 |
Vegetation | 0 | 0 | 0 | 30 | 0 | 30 |
Sum | 30 | 30 | 30 | 30 | 0 | 100% |
Some pixels were categorized as undefined when there was the excitation of more than one neuron of the network or when none of them were activated. In case there is no activation of a neuron, the use of that pixel becomes impossible for the triggering of alarms to the processed scene. Nevertheless, if there is more than one active neuron, it is still possible to use the estimates for the calculation of alarm triggering. It should be emphasized that, as far as producing a thematic map for the scene is concerned, the use of a filter is possible, in order to determine the category of the undefined pixels. One such option could be to consider the undefined category point according to the majority of points around it.
4. Conclusions
According to the results obtained, it is possible to conclude that Neural Networks can be used as orbital images categorizers to help to quickly detect deforestation. This work also shows that the spectral similarity of the mapped classes, such as deforestation and savanna, influences the outcome of the neural network. In this sense it is believed that using more samples from different soil types can contribute to a better classification. As such, the developed tool emerges as a classifier of satellite images, as much as a general purpose tool, including application in teaching and for the free creation and testing of neural networks applicable in many areas.
In view of the inherent difficulty of the process of efficient deforestation alarms, that is, when false alarms are not fired and all areas are considered deforestation points, more studies are necessary. The developed tool was able to detect deforestation points. However, it is still unstable, and produces false alarms. Potentially, better results might be found if the data used are obtained by remote sensing specialists for the training of neural networks. Nonetheless, the software must be enhanced in order to standardize the alarm triggering.