The aim of this paper is to 1)conduct a comparative analysis of the SMILE Google Earth Engine machine learning classifiers and analyze their performance in crop type classification over a semi-arid irrigated heterogeneous landscape; 2) classify cannabis fields in the Bekaa region for the years 2016, 2017 and 2018 using multi-temporal multispectral imagery and GEE built-in classifiers; 3) evaluate the added value of the increase in temporal frequency due to the use of Landsat-Sentinel-2-Sentinel-1 in crop type classification. We use data from Sentinel 2, Sentinel 1, and Landsat 8. We evaluate the accuracy of four different SMILE classifiers: Random Forest, Support Vector Machine, Classification and Regression Trees, and Gradient Tree Boosting, which was recently introduced to GEE classifiers. We perform multi-temporal image classification using three different imagery combinations. Each combination consists of two or three sensors. The first combination consists of Sentinel 2 and Sentinel 1 imagery , the second combination consists of Landsat 8 and Sentinel 2 imagery , and the third combination consists of Landsat 8, Sentinel 2, and Sentinel 1 imagery . Sentinel-1 mission is composed of two polar-orbiting sun-synchronized satellites, each carrying an imaging C-band synthetic aperture radar instrument. The Sentinel-1 mission supports four imaging modes providing different resolutions: Interferometric Wide Swath , Extra Wide Swath , Strip Map , and Wave . We use the IW mode which supports operation in dual-polarization and has a ground range resolution of 5 × 20 m. Sentinel-2 mission provides high spatial resolution multispectral satellite imagery with a high revisit frequency and a wide field of view. The spatial resolution varies from 10 m to 60 m depending on the spectral band. Sentinel-2 carries the Multispectral Imager sensor with 13 spectral bands in the visible, near-infrared, and shortwave infrared part of the electromagnetic spectrum.
Landsat 8 , a NASA and USGS collaboration, provides global moderate-resolution imagery in the visible, near-infrared, shortwave infrared, and thermal infrared part of the electromagnetic spectrum. Fig. 2 represents the flow of work and the three main stages for crop classification. Sentinel 2 TOA reflectance dataset was obtained from GEE platform, it was then refined to remove cloudy pixels by cloud masking using QA band and pre-filtered at 5% or fewer clouds. We used the Normalized Difference Vegetation Index and the Enhanced Difference Vegetation Index , the two most widely used vegetation indices as additional bands in the classification process. EVI provides improved sensitivity in high biomass regions while minimizing soil and atmosphere influences but requires a blue band. EVI2 performs similar to EVI but without the blue band . Other vegetation indices such as the soil adjusted vegetation index were tested but did not offer an advantage. The images were filtered over six months, from March to August and one image from each month was chosen. This resulted in six images from each dataset each containing 13 bands . A median, maximum, and minimum composite were created from the image collection. All bands were then combined to form a single image containing median, maximum, and minimum values for each band . The calibrated and ortho-corrected Sentinel-1 product was accessed from within the GEE platform. It was then filtered to get images with VV and VH polarization. The images were filtered over the same period as Sentinel 2 images. A mean composite was obtained of all Sentinel 1 images which resulted in a total number of 12 bands . Three texture metrics were also added for both bands . The images were converted to bands and added to the Sentinel 2 bands which resulted in a total of 282 raster features for the six months. Finally, cannabis grow system the GEE Landsat 8 Surface Reflectance dataset was filtered to mask clouds and cloud shadows over the study area over the same period as the previous two datasets. A median and variance composite of all Landsat 8 images was created. The result is six Landsat 8 images with 8 bands each . We combine Landsat 8 bands with Sentinel 2 bands for the L8S2 dataset, resulting in a total of 330 raster features. The third and final dataset combination contains a total of 378 raster features for S1, S2 and L8 combined over the full-time series. Table 1 summarizes the datasets used. The training data used includes ground-truth crop fields collected in 2016 & 2018 as polygons. After delineating surveyed crops, they were imported to GEE as assets and grouped into six crop groups. For the 2016 classification: wheat and barley fields were combined; beans, broccoli, cabbage, celery, coriander, corn, cucumber, lentil, lettuce, melon, mint, onions, parsley, pepper, radish, tomato, carrot, watermelon fields were combined into one crop group; and stone fruits and vineyards were combined.
For the 2018 classification: wheat, barley, alfalfa, and vetch fields were grouped into one crop group; onion, cucumber, eggplant, squash, pepper, tomato, lettuce, parsley, mint, cabbage, carrot, chickpea, peas, beans, corn, and melon fields were grouped into another; and apple, apricot, peach, plum trees, and vineyards were grouped. Cannabis fields were filtered into irrigated and rain-fed or not irrigated, and only the irrigated fields were used in this classification. We use irrigated cannabis fields for the training of the classifiers because tobacco and cannabis share a similar growing cycle . Since tobacco fields are rain-fed or non-irrigated, confusion between detection of tobacco and cannabis occurred. Thus, we decided to use only irrigated cannabis fields to avoid this type of confusion. The other two crop groups were potato and tobacco. Urban and fallow polygons were selected based on their NDVI being less than 0.18 over the whole year. Finally, each crop group and the urban/fallow group were given a label starting from 0 until 6. Fig. 3 shows the spatial distribution of surveyed crops over the Bekaa plain in 2016. We use the GEE “sample Regions” as the sampling method to select training data. The method requires inputs that are a Feature Collection, a class property, and a scale. We provide the polygons in a form of a Feature Collection as an input. The class property in our classification is ‘Croplabel’. Hence, a ‘Croplabel’ is assigned to each class of the sampling data, and the scale size used is 50 m. The number of training pixels used for the 2016 classification and 2018 classification is shown in Table 3. The number of pixels used was obtained by taking the average number of pixels of 10 GEE permutations. The total number of pixels was divided such that and 70% of the pixels were used for training and 30% of the pixels were used for validation. The number of training and testing pixels for each crop type is also shown in Table 3. Classifying different crop types requires knowledge of their spectral characteristics. Spectral characteristics of different crops are an important factor that affects the accuracy of crop identification. Each crop type has a unique spectral signature that changes with time depending on crop phenology and crop growth stage . Thus, making the temporal feature essential in crop identification. Defining the best “identification time” provides opportunities to identify spectral differences during specific phenological stages. To enhance spectral discrimination among different crops, vegetation indices can be used to take advantage of the difference in reflectance between different wavelengths . In this study, we aimed to identify a period where the spectral differences are suitable to maximize classification accuracy. To achieve this, we analyzed the spectral characteristics of the crops in the study during a whole year. We calculated the mean NDVI, using Sentinel 2 imagery, over 15 different surveyed fields from each crop group for the year 2016.
The most spectral differences among these crop groups were observed between April to August in both years. However, we applied the classification on different periods and found that the highest classification accuracy was between March to September, which was adopted in this analysis. Random Forest classifier is an ensemble-based tree-structured classification method that grows several classifiers instead of one classifier to perform better classification . It creates several decision trees and aggregates their results, thus predicting a response from several decision trees. The variation between individual trees results in more diversification and better accuracy. The classifier requires two user-defined parameters: the number of decision trees to be grown and the number of features used at each node. In our classification, the number of trees is chosen to be 20 trees based on parameter tuning to minimize computational timeouts, reduce run times, and avoid over-fitting . We performed hyper-parameter tuning by training the classifier using a different number of trees and plotting the overall accuracy results for each number of trees. The number of splits at each node is set by default to the square root of several variables. Gradient tree boosting , similar to the random forest, is a tree-based ensemble learning algorithm that uses gradient descent as its optimization algorithm to minimize the loss function. It works by sequentially training several weak learners, which are shallow decision trees. During the training process, the mis-classified pixels – due to the weak learners that classified them earlier – are assigned to stronger weights and thus classified correctly . A loss function is associated with the model such that, each time a new tree is added it minimizes the loss function. With every addition of a new tree, the overall prediction error will decrease . Classification performance of gradient boosting is highly influenced by parameter tuning. Parameters include the number of trees, shrinkage, learning rate, loss function, sampling rate, and maximum node size. In this classification, we use 15 trees and the default parameters set by GEE: “Least Absolute Deviation” as loss function, shrinkage of 0.005, and sampling rate of 0.007. Classification and regression tree is a tree-based machine learning algorithm . It can be used to classify numerical and categorical data. The algorithm behind CART classifier builds a decision tree starting from the roots, which will split at each node. The training data will pass down the tree through the splits and a decision is made at each node to decide the next direction of the data. The decision is made to reduce the impurity at each node, which depends on the splitting rule. The splitting rule metrics for classification trees include, but are not limited to, misclassification error, Gini index, Entropy index, and Twoing . The splitting will continue until there is only one sample left, and a final decision is made at the terminal node . The inputs for CART classifier are a feature collection which is the 70% training data, a class property which is ‘Croplabel’, and input properties which are all the bands being used for classification. We classify crops in 2017 using the previously created models from 2016. We prepare a composite, similar to Section 2.2, for the year 2017. We extract the spectral, spatial, and textural characteristics of the 2016 training polygons using different dataset combinations. We train the classifiers using the 2016 training data and we classify the 2017 composite using the trained classifiers in 2016. We choose the year 2017 because we have obtained marijuana grow system fields’ survey data during 2017. We perform an accuracy assessment as described earlier. The surveyed fields were all used as a testing data set since this data was not used for training. We perform several permutations, using the most robust classifiers, in order to decide which model classifies cannabis in 2017 better than the other.
The most accurate model in classifying cannabis is presented below. The application of machine learning algorithms in crop-type and land cover classification has gained popularity in the past decade. However, the identification of cannabis plantation areas has not received enough attention. Our study provides a crop classification method to identify cannabis fields using freely available medium resolution imagery and machine learning algorithms. We perform several permutations using different dataset combinations and different metrics for each satellite data. The highest accuracy was achieved using the mean composite of Sentinel-1 and median and variance composite of Landsat 8, however, for Sentinel-2, using min, max and median composite achieved better results. Our results are in agreement with Carrasco et al. who evaluated the use of different metrics derived from Sentinel-1, Sentinel-2, and Landsat 8 as a time series for land cover mapping, and classification was achieved using the mean composite of Sentinel-1, the median composite of Sentinel-2, and the median and variance composite of Landsat 8. Previous studies have reported the improvement in classification accuracy after combining radar sensors such as Sentinel-1 to optical sensors . Our classification results show a slight improvement in classification accuracy when Sentinel-1 is added using RF and SVM classifiers.This marginal improvement in accuracy was also observed earlier by Boryan et al. when identifying winter wheat in two different locations using optical and SAR data. The authors suggest that texture features from SAR are not necessarily important when cloudless optical data is available.