Introduction

Despite their relatively low area coverage, riparian forests are central landscape features that provide several ecosystem services: stream bank stabilization, reduction of fine sediment inputs and nutrient contamination, aquatic and terrestrial habitat improvement, and recreational and educational opportunities (Guo et al. 2011; Vidal-Abarca Gutiérrez and Suárez Alonso 2013; Clerici et al. 2014). Nevertheless, riparian forests are critically endangered in European countries by human pressures, such as livestock grazing, land use pressures, canalizations and wastewater, and also by natural hazards, such as the black alder (Alnus glutinosa) disease caused by Phytophthora alni (Cech 1998; Gibbs et al. 2003; De Merlier et al. 2005). The importance of riparian zones has been acknowledged by European directives (notably the Habitats Directive and the Water Framework Directive), involving multi-scale monitoring (local to network scales). Because of the typical linear and narrow shape of the riparian zones, field-based monitoring involves sampling, high labor costs, and time-consuming travels (Debruxelles et al. 2009; Myers 1989). The continuous improvement of the spatial resolution of remote sensing data combined with more powerful computer capacity and new geomatic procedures to extract information make the remote sensing approach more competitive (Alber and Piégay 2011; Carbonneau and Piégay 2012; Johansen et al. 2010; Michez et al. 2013; Roux et al. 2014). The use of this very-high-resolution (VHR) imagery in a multi-temporal approach is an emerging topic (Ardila et al. 2012; Champion 2012; Lasaponara et al. 2014). The trend is reinforced by the recent and rapid growth of the use of the unmanned aerial system (UAS), which has prompted the development of innovative methodology.

The use of the UAS in environmental science has become more common since the late 2000s. UAS imagery can be an effective alternative for cost-effectively describing riparian zones at the local scale (Dunford et al. 2009; Husson et al. 2014). In the ecological scope, the scientific community is very enthusiastic about the use of the UAS, proclaiming the “dawn of drone ecology” (Koh and Wich 2012) and that the UAS “will revolutionize spatial ecology” (Anderson and Gaston 2013). The two most important characteristics of UAS imagery are said to be the very high spatial and temporal resolution, allowing description of events occurring at a very local scale in a finite time window (e.g., the flowering of or the attack of a pathogen on a given tree species). Because of its very high spatial resolution (<0.1-m ground sampling distance (GSD)), UAS imagery is regularly characterized as hyperspatial imagery (Carbonneau and Piégay 2012; Greenberg et al. 2005; Laliberte et al. 2007; Strecha et al. 2012). Many studies have taken advantage of these two characteristics, for a broad range of environmental applications, such as landslide mapping (Lucieer et al. 2014), forest fire mapping (de Dios et al. 2011; Merino et al. 2012; Urbahs et al. 2013), precision farming (Bendig et al. 2013), wildlife census (Lisein et al. 2013a; Vermeulen et al. 2013), tree and forest characterization (Lisein et al. 2013b; Zarco-Tejada et al. 2014), forest biodiversity assessment (Getzin et al. 2012), and forest species composition (Dunford et al. 2009; Gini et al. 2014). Recent studies have demonstrated the potential of UAS imagery to finely describe the forest canopy (Dandois and Ellis 2010, 2013). When a good-quality digital terrain model is available, UAS-derived photogrammetric point clouds (>10 points/m2) can provide a canopy height model (CHM) with a quality comparable to light detection and ranging (LiDAR) CHM but with significant cost differences (Lisein et al. 2013b). References on classification of forest species by use of UAS imagery are still rare in the literature and include only the single-date approach (Dunford et al. 2009; Gini et al. 2014). However, useful pioneering studies are available on vegetation mapping projects based on UAS imagery (Knoth et al. 2013; Zaman et al. 2011).

Classification of forest species using remote-sensing data is relatively well documented. Successful classifications have been carried out using multi-spectral aerial/satellite imagery (Immitzer et al. 2012; Ke et al. 2010; Key et al. 2001; Korpela et al. 2011; Lucas et al. 2008; Sesnie et al. 2008) and hyperspectral data (Dalponte et al. 2014; 2012). The scale of analysis varies from tree crown pixel or object (object-based image analysis) to the crown itself. When available, 3D data (ALS LiDAR, SRTM DEM) can improve the classification accuracy (Dalponte et al. 2014; Ke et al. 2010; Sesnie et al. 2008). Forest species classification has also been studied through the use of multi-temporal remote sensing dataset which allows to highlight phenological differences between forest species (Key et al. 2001; Hill et al. 2010; Zhu and Liu 2014). These multi-temporal studies were intended to identify discriminant phenological stages (and related date) in order to schedule future imagery acquisition at lower cost within the identified time window. These studies identified the late growing season as the most relevant for these purposes. As UAS imagery allows acquisition of cost-effective dense time series of very-high-resolution imagery, new approaches can be considered.

Few studies have taken advantage of the temporal resolution of UAS imagery and the subsequent possibility of redundancy. Interesting case studies can be found, but they are outside the proper ecological scope. In precision farming, Bendig et al. (2013) computed UAS-derived digital surface model time series to measure the growth rate of experimental field crops while Torres-Sánchez et al. (2014) used UAS orthophotos to differentiate weeds in wheat fields in the early season. Multi-temporal data from UAS imagery have also been used to evaluate landslide and rockslide evolution (Lucieer et al. 2014; Ragg and Fey 2013). Most of these previous contributions used the temporal resolution of UAS imagery to fit the acquisition period to the occurrence of the events to be mapped, because UAS allow the final user to perform (at least to really get involved in) the survey. Subsequently, a methodological gap remains to be filled in terms of approaches that utilize dense (UAS) time series for a unified and multi-temporal classification purpose.

We propose a methodological framework to explore multi-temporal hyperspatial UAS imagery acquired during a single vegetative period in order to characterize riparian forest species and health condition. More specifically, we intend to identify which type of variables derived from hyperspatial imagery and which scale of image analysis are the most relevant to reach our objectives. Another major ambition is to highlight the best time window and the configuration of off-the-shelf camera.

The methodological framework is applied to describe the riparian forest through two fundamental characteristics: the species composition and the health condition. These characteristics were selected because of their applicability as proxies for ecological integrity of the riparian zones (Innis et al. 2000; Naiman et al. 2005) and for their usefulness in targeting management strategies (Debruxelles et al. 2009).

Methods

Study sites

Both study sites are riparian forests located in two different landscapes in Wallonia, southern Belgium (Fig. 1) in the Ardennes eco-region. To ensure the riparian character of the studied forest, forest patches located along the river bank (<20 m from the wetted channel) were selected.

Fig. 1
figure 1

Study sites location. Site 1 (Salm River) is located in an agricultural landscape, with a globally managed riparian forest strip; site 2 is located in an undisturbed forested landscape, with an unmanaged riparian forest strip

Riparian forests of site 1 (Vielsalm, Salm River) are located in an agricultural landscape (pastures). They are highly pressured by farm activities and remain mostly in a managed forest strip along the riverbanks (<10 m). The population of black alders in site 1 is known to be infested by P. alni (Di Prinzio et al. 2013a). Riparian forests of site 2 (Felenne—Houille River) are located in a relatively undisturbed forest landscape. These strips are mainly unmanaged and have a regular presence of standing dead woods.

In terms of species composition, riparian forests of both sites are dominated by A. glutinosa with non-native coniferous stands (Picea abies and Larix decidua) in site 2. A. glutinosa is the typical overstory tree species in the riparian zones of Wallonia (Claessens et al. 2010; Debruxelles et al. 2009).

Acquisition of UAS imagery

Our workflow is based on a time series of 25 UAS flights (Table 1) allowing production of 18 and 7 orthophotos in site 1 and site 2, respectively, using a Gatewing X100 (wingspan 100 cm, weight 2 kg, cruising speed 80 km/h, flight height 100 to 750 m). Two pre-calibrated Ricoh GR3 cameras (10-megapixel sensor, focal length 6 mm) were used. One camera was adapted for near-infrared acquisition (Fig. 2). The UAS can only carry one camera per flight. The flight duration never exceeded 45 min, and the overlap between images was set to 75 %.

Table 1 Eighteen (site 1) and seven (site 2) orthophotos (0.1-m GSD) generated
Fig. 2
figure 2

Idealized filter profiles (lines) and channel sensitivity (surface) for the RGB (top) camera and the RGNIR modified camera (bottom). The RGNIR camera was modified by removing the internal hot-mirror filter and adding a blue-light-blocking filter (i.e., a yellow long-pass filter). Adapted from Nijland et al. (2014)

Field staff performed the maximum number of flight surveys during approximately 2 h around solar noon. The number of flights per day was directly influenced by the meteorological conditions and logistical constraints (i.e., period of air space restrictions, UAS equipment failures, etc.). These circumstances resulted in a varying number of flights per day (Table 1). The weather conditions induced varying lighting conditions within and between the orthophotos. Although orthophotos which are the closest to solar noon should theoretically present the best quality to reach our objectives, the varying lighting conditions seriously complicated the selection of a single orthophoto per flight date. With the willingness to build a computer-based approach, all available orthophotos were therefore used to avoid an operator-mediated selection.

Because one of the main objectives of the study was the characterization of riparian forest health condition, the aerial data survey was planned for the late growing season, which is the best time window (Gibbs et al. 2003) for catching the leaf symptoms of the most common riparian forest pathogen in Wallonia: P. alni on the black alder (Di Prinzio et al. 2013b; Pintos Varela et al. 2010). This time window also matches well with the time windows found in other multi-temporal remote sensing studies for differentiating forest species (Key et al. 2001; Hill et al. 2010; Zhu and Liu 2014).

General workflow

We propose an innovative and replicable multi-platform workflow to handle large and complex datasets induced by the spatial and temporal resolution of the UAS time series in a computer-based perspective to accurately map riparian forest species and health conditions (see summary in Fig. 3).

Fig. 3
figure 3

General overview of the method applied to each classification process. Steps before the “generation of sample database” are executed once for the 25 UAS surveys; the other is specific for each classification scenario (species composition and health condition)

Photogrammetric process

General photogrammetric workflow

Agisoft Photoscan 1.0 Professional, which is now relatively common in the UAS community, was used to perform photogrammetric surveys with proven efficiency (Dandois and Ellis 2013; 2010; Sona et al. 2014). Every flight dataset was processed by following the workflow described in Fig. 4.

Fig. 4
figure 4

Photogrammetric workflow

Our photogrammetric workflow was based on the work of Lisein et al. (2013b) in the Photoscan environment. On the basis of the information of the inertial measurement unit (IMU), the GPS positions recorded by the UAS, and a set of ground control points (GCPs), an initial aerotriangulation, is performed (number 1 in Fig. 4). The georeferencing process is based on a set of particular points (road crossings, roofs, pedestrian crossings, etc.) that are used as GCPs in Photoscan. For site 1, the GCPs have been measured with a real-time kinematic (RTK) GPS (Leica 200+). For site 2 (low mobile network coverage), the GCP planimetric coordinates were defined on the regional orthophoto coverage (0.25-m GSD), and the altimetric positions were extracted from LiDAR digital terrain model (DTM). Bare ground areas of the low-resolution 3D model (number 2 in Fig. 4) are aligned on a LiDAR DTM within CloudCompare (http://www.danielgm.net/cc/), by use of the iterative closest-point matching strategy. The matching process results in a global transformation matrix (number 3 in Fig. 4) of six parameters (translation and rotation), which is used to perform a rigid transformation of the GCP coordinates. The registered GCP positions are used to finely georeference the initial aerotriangulation (number 4 in Fig. 4) and compute a high-resolution dense matching (number 5 in Fig. 4).

Orthophoto generation

The photogrammetric workflow (Fig. 4) allows generation of orthophotos for every flight dataset. The orthophotos were resampled to 0.1-m GSD. The orthophoto generation process used the “mosaic” blending mode with the color correction option enabled in Photoscan 1.0. Table 1 presents the list of the orthophotos, with the reconstruction error provided by Photoscan.

Canopy height model computation

For each site, we computed a single CHM (0.1-m GSD) by the combination of a LiDAR DTM and a photogrammetric digital surface model (DSM) (DSM—DTM). The two “hybrid” CHMs have been used as the sole 3D reference for the segmentation/classification process, under the condition that the surface remained unchanged during the time window of the surveys (<6 months).

In each site, a reference flight date was selected to compute the two site reference DSM (see Table 1). The flights were selected manually on the basis of the visual quality of the individual raw images (ground visibility, luminosity, and sharpness). For each site, a new photogrammetric project (Table 2) was conducted with all images of the selected date, by following the same photogrammetric workflow (Fig. 4).

Table 2 Key characteristics of the photogrammetric process to compute the reference photogrammetric DSM

Reference dataset

A georeferenced database of 1186 deciduous trees (434 in site 1, 752 in site 2) was compiled. The position of the stem was measured with a field-map system (http://www.fieldmap.cz), initially positioned using a DGPS (GNSS Leica GPS1200+) on site 1 and with a post-processed SXBlue II GPS (http://sxbluegps.com) on site 2.

All overstory trees were mapped and identified (species level) within the two study sites. Because site 1 has been previously identified as infested by P. alni, the health condition of the black alders was estimated by a skilled operator according to two classes: asymptomatic tree and tree with P. alni foliar symptoms (defoliation and/or foliage yellowing). The field survey was conducted in August 2012 in site 1, at the time the symptoms of the P. alni disease were the most distinctive (Gibbs et al. 2003; Di Prinzio et al. 2013b). In site 2, the field survey was conducted in October 2012 and was only dedicated to identifying species.

Crowns of georeferenced trees were delineated by the use of the generated orthophotos and the two site-reference CHMs by the operator who performed the field survey. Respectively, A. glutinosa represent 67 and 48 % of the trees in sites 1 and 2 (see details in Table 3).

Table 3 Tree crown database

Multi-resolution image segmentation

The segmentation was performed by the use of eCognition developer, with a multi-resolution image segmentation algorithm (Blaschke 2010). This region-growing segmentation algorithm starts from individual pixels that are merged to most-similar adjacent regions. The size and the shape of the resulting objects are defined by a scale parameter (related to the object size), shape, and compactness criterion (object homogeneity). To define which object size is the most relevant, we tested six different scale parameters: 5, 10, 15, 30, 45, and 60 (Fig. 5). The mean size of the different objects of homogeneity criteria (shape and compactness) was equally balanced and set to 0.5.

Fig. 5
figure 5

Manual tree crown delineation and six values of the OBIA scale parameter. The values of the OBIA scale parameter result in object size meanly varying from ca. 50 to 1 m2

The segmentation process is performed on selected layers to which a pre-defined weight has been assigned. To identify an object that can be considered as constant during the time window of the multi-temporal dataset, the segmentation process used in this study was based on the two canopy height models.

We extracted from each of the image objects a set of metrics (Table 4) at eight scales of analysis. In addition to the six OBIA scales, we tested one pixel-based scale (4 pixels square objects) and the manually delineated tree crown scale. To exploit the whole time series in a single automated classification process on the basis of a random forest machine-learning algorithm, various spectral, texture, and vertical structure metrics were derived from the UAS imagery.

Table 4 Metrics computed and related example case study

We choose five gray level co-occurrence matrix (GLCM) derivatives based on the work of Haralick et al. (1973) and computed with eCognition: entropy, standard deviation, correlation, mean, and contrast (see Table 4). Those derivatives were chosen, most notably, because of their proven efficiency in similar approaches (Laliberte and Rango 2009; Stumpf and Kerle 2011). Those GLCM derivatives are typically computed through symmetric matrices for pixels neighboring directly at 0°, 45°, 90°, or 135°. In our case, we select an omnidirectional approach, in which the sums of all four directions of GLCM derivatives are calculated before texture calculation.

We test several simple metrics (see Table 4 for details), derived from the spectral layers’ mean itself and from different band ratios. Our approach was based on normalized band ratios (such as NDVI or GNDVI) and individual band ratios, such as the green ratio vegetation index.

On the basis of the two canopy height models, we compute three metrics (mean, SD, and skewness) that describe the vertical structure of a riparian forest, considered as constant during the time series (see Table 4).

Classification process

Generation of sample database

Objects with at least 75 % of relative overlay with delineated tree crowns were selected to compile the tree crowns’ object database (Table 5). This threshold is relatively restrictive regarding to other comparable studies (60 % for Ma et al. (2015) and even 50 % Stumpf and Kerle (2011)). This choice was justified because of the relatively high similarity between classes compared to these two studies (land cover classes for Ma et al. (2015) and “landslide” vs “other objects” for Stumpf and Kerle (2011)).

Table 5 Key information about sample database linked to the scale of analysis

We test two classification scenarios for the riparian forest species and health condition (Table 6). For each of the sites, we established a classification with five classes. The first four classes group the spatial units of each of the four most common species from the field survey, whereas the fifth class (“other riparian species”) regroups the spatial units corresponding to the other species of the field survey. The purpose of this selection was to ensure a sufficient number of objects at every scale of image analysis. Because the black alders of the riparian forest of site 1 were known to be highly contaminated by P. alni (De Merlier et al. 2005; Di Prinzio et al. 2013b),. a focus was placed on the distinction between black alders with and without foliage symptoms.

Table 6 Classification approaches

Selection of variables by the use of random forest

Random forests’ machine-learning algorithms have proved their ability to handle very complex and high-dimensional remote-sensing datasets (Cutler et al. 2007; Lawrence et al. 2006; Stumpf and Kerle 2011; Watts et al. 2009). Most of the previously cited studies used the initial implementation of the random forest algorithm (Breiman 2001) in the R software. Moreover, different studies have provided some improvement to this first package, notably to use the random forest algorithm for variable selection (Diaz-Uriarte 2007; Genuer et al. 2010). To identify which variable is the most relevant to differentiate species or tree health conditions, we use an implementation of the random forest algorithm by Genuer et al. (2010) and implemented within the R package “VSURF” ( http://cran.r-project.org/ ). This process of variable selection was developed to handle high-dimensional data (number of variables exceeding number of observations) by following three steps (Genuer et al. 2010). The first step (“thresholding step”) runs a set of classic random forest processes (ntree parameter 2000) and eliminates irrelevant variables from the dataset through the study of variable importance (Gini index), computed during the random forest processes (see Breiman (2001) for details). The second step (“interpretation step”) selects the best model through successive calls of random forest processes, starting with the random forest build utilizing only the most important variable and ending with all variables. The last step (“prediction step”) refines the selected variables by eliminating redundancy in the set of previously selected variables. The first model is based only on the most important variable, and other variables are added to the model in a stepwise manner. To remain in the final model, a variable must have a bigger impact on the model accuracy than a noise variable (variables left out by interpretation). This last concern is tuned into VSURF through the “number of mean jump (nmj)” parameter, which was set to five with a trial and error approach (default value 1).

Classification and accuracy assessment

Once the selected variables have been identified, the original sample database is randomly resampled (50 %) to create training and evaluation sets. The training set is used to run a new RF process (ntree parameter 2000). The model is then applied to the evaluation set for the accuracy assessment. The entire classification and accuracy process has been repeated 50 times to compute mean ± SD values of accuracies

Results

Riparian forest species composition

The classification process was run on the entire sample database, classifying riparian tree crown objects into five classes (Table 6). The accuracy of the classification varies with the scale of analysis (Table 7), suggesting that the overall accuracy is improving while the size of the analyzed objects is decreasing. Pixel-based or close to pixel OBIA objects (scale 5) perform the best (approximately 80 % of overall accuracies).

Table 7 Accuracy assessment and selected variables

The study of the selected variables (Table 7) at various scales reveals that the data derived from flights in the late growing season (site 1) or in partially leaf-off conditions (site 2) are the most appropriate. Phenological differences are enhanced during the late season, especially in the riparian forest foliage. In very late survey dates, various states of phenology can be found from fully photosynthetically active foliage in coniferous stands to completely “leaf-off” deciduous riparian trees.

Compared to selected variables of the site 1, variables derived from the modified RGNIR camera are more frequent in selected variables of site 2 (Table 7). The variables derived from the unmodified RGB camera (regardless the scale of analysis) are far more used on site 1 (Fig. 6).

Fig. 6
figure 6

Frequency of variables derived from the RGNIR and RGB camera in selected variables for the species composition classification approach (site 1). RGB-derived variables are more frequent in selected variables than variables derived from the modified off-the-shelf RGNIR camera

Results presented in Fig. 7 show that most of the selected variables are derived from the spectral information. The results also highlight the advantage of adding additional information to the classical spectral information from 3D layers or texture information. The 3D information, from derivatives of the CHM, is used for the finest scales of analysis, whereas the texture information, from GLCM derivatives, is used at coarser scale of analysis.

Fig. 7
figure 7

Scale of analysis and variable type frequency in selected variables for the species composition classification approach (site 1). Height: variables derived from a canopy height model. Texture: GLCM derivative variable. Spectral: variables derived from the spectral layers. Spectral variables are the most used. Texture variables are used at coarse scales, whereas 3D variables are used at fine scales

The study of the confusion matrices for the best performing classification models (species composition approach) for sites 1 and 2 (Tables 8 and 9) reveal contrasted individual classification errors.

Table 8 Individual confusion matrix of the best performing scale of analysis (OBIA scale 5) for the species composition classification approach (site 1)
Table 9 Individual confusion matrix of the best performing scale of analysis (4 pixels) for the species composition classification approach (site 2)

Classification of alder health conditions

The variables used to classify the health condition of the black alders differ from those used for the species composition classification in two important ways. The health condition classification presents better overall accuracy value (>90 %) and requires fewer variables to construct the optimal classification models (12 vs 9 for site 1) (Table 10).

Table 10 Accuracy assessment and selected variables (species composition classification)

The classification of the health condition of A. glutinosa in site 1 is based more on variables derived from NIR imagery, which allow healthy alders to be distinguished from those that presented defoliation symptoms during the field survey.

Discussion

Riparian forest species and health condition

Overall accuracies of our tree species classifications reached 79.5 % (site 1) and 84.1 % (site 2), which are comparable with those raised by recent researchers who used much more expensive hyperspectral imagery (79 %, Dalponte et al. 2014) or multi-spectral imagery (83 %, Waser et al. 2011). High overall accuracy results were obtained for the classification of the health condition of the typical overstory tree species of Walloon riparian forest (A. glutinosa, 90.6 %).

The study of individual classification errors (see Table 8 and Table 9) for the best performing species classification models can be analyzed through the phenological traits of the tree species. In site 2 (Table 9), Acer pseudoplatanus and A. glutinosa present similar and relatively high classification error. These higher classification errors can be explained by high intraspecific phenological variations for Acer species (reported by Hill et al. 2010) and by variation in the health condition of A. glutinosa (recorded presence of P. alni). P. abies or Quercus species present very low classification errors in site 2. The phenological characteristics of P. abies are quite obvious because this species forms distinct even-aged and evergreen stands along site 2. In site 1, the highest classification error is related to the class of Fraxinus excelsior and can be explained by the presence of the highly pathogenic Chalaria fraxinea (recorded during field survey).

Operational recommendations

Time window

The most appropriate time windows to perform the flight surveys were found to be the late season of vegetation, when phenological differences of riparian forest trees are the most enhanced. These results are in line with previous studies such as Key et al. (2001) and Hill et al. (2010), who used aerial multi-spectral imagery, and Zhu and Liu (2014), who used satellite imagery.

Number of surveys

Respectively, 6 and 5 days of surveys were required to achieve the best classification results for sites 1 and 2, whereas 6 days of flight surveys were required to assess the health condition of black alders of site 1. These results highlight the importance of the multi-temporal character of the dataset. However, the health condition of black alders in site 1 can be evaluated with a sufficient accuracy (81 %) using two variables derived from a single flight survey with the modified RGNIR camera.

Scale of analysis

The relation between the scale of the analysis and the accuracy of the classification shows evidence of higher global accuracies for the classification scenarios performed with smaller objects (from segmented objects <1 m2 to 4 pixel objects, 0.04 m2). The trend is more obvious for the “species composition” scenario.

This result highlights the importance of the tree crown object segmentation process. The closer the size of the segmented object is to the tree crown scale, the more accurate the segmentation process must be to avoid objects overlapping several individual trees. The oversegmentation induced by the reduction of the OBIA scale parameters limits this overlapping. However, tree crown delineation is a research topic in itself, and the multi-resolution segmentation process was, in our approach, intentionally simple, without scale-related fine-tuning. Fine-tuning of the segmentation process in relation to the scale of image analysis would especially improve the performance of coarser scales of image analysis but would turn the approach into a more time-consuming and less computer-mediated perspective.

Besides the quality of segmented objects, larger objects present higher spectral heterogeneity. This fact combined to a lower number of training objects (due to the fixed sampling ratio) induces the decrease of the overall accuracy as the size of the objects increases. These findings are in line with recent publication of Ma et al. (2015) who also used UAS imagery. The lower overall accuracy values of the manually delineated tree crown scale (despite a lower mean area value than the OBIA 60 scale objects) can be further explained by the fact that a major part of delineated individuals in the reference dataset is actually clump of trees. In this case, a single individual polygon can regroup up to ten crowns, with shadow and blurry effect between and around them, inducing even higher spatial heterogeneity in these objects (compared to larger “OBIA 60” objects).

Relevant type of variable

The analysis of the selected variables also confirms the value of the use of 3D information (derived from a canopy height model) at fine scale and texture information at a coarser scale. The selection of the “mean height” at fine scale can be explained by height differences between “healthy” and “unhealthy” riparian trees. At the local scale, the dense matching process can fail to reconstruct the canopy of trees with extreme defoliation symptom, leading to an underestimation of the height. The low efficiency of texture metrics (and thus the high performance of spectral metrics) can also be explained by the orthophoto generation process. Because of the very high complexity of the surface of the canopy, tree crowns on orthophotos can present a blurry aspect. Considering that the best results were achieved at finer scale (<1-m2 objects), we do not recommend the use of texture metrics, which are time-consuming and were unused by classifiers at best performing scales.

Use of modified RGNIR off-the-shelf camera

The site 1 being less contrasted in terms of forest type and presenting the “densest” time series, it can be used as a model for comparing the performance of the modified RGNIR camera. The results (see Fig. 6) highlight the higher performance of the RGB camera, which can be explained by the spectral overlap between the spectral bands of the modified RGNIR camera (see Fig. 2).

Compared to site 1, the more intense use of RGNIR imagery in site 2 can be explained by the forest species composition. As the appropriate time windows to perform the flight surveys were found to be the late season of vegetation, the simultaneous presence in site 2 of photosynthetically active (evergreen coniferous stands) and non-active species (riparian leaf-off deciduous trees) induces higher contrast in NIR. This fact can also explain the selection of the blue band capture in the late season of vegetation in site 2, because the soil (and subsequently leaf-off trees) has a specific response within the blue wavelength.

Conclusions

The quality of the classification models developed with a UAS-based, cost-effective, and semi-automatic approach competes successfully with those developed using more expensive imagery, such as multi-spectral and hyperspectral airborne imagery. Our study uses the temporal resolution of the UAS to compile a redundant dataset of UAS imagery to highlight riparian forest phenological differences. This innovative approach is the first forest case study that takes advantage of both spatial and temporal resolution of UAS imagery to successfully describe subtle features, such as individual tree species or health condition.

The accuracy of the classification of the diseased alders opens the door to applications dedicated to monitoring of the health conditions of riparian forests. Our approach is directly applicable in other context, notably to characterize other emerging dieback diseases that also induce defoliation symptoms (e.g., C. fraxinea on F. excelsior in Europe).

Regarding the forest characterization, most of the previous research on time series imagery aimed to identify a specific time window when the classification would be the most reliable. The final operational application was subsequently to schedule imagery acquisition at a lower cost within the identified time window. With very low operational costs induced by UAS imagery, applications can integrate dense time series imagery on an operational basis. Our methodological framework will provide UAS users with a solution to manage large imagery metric datasets derived from those dense time series. In temperate climate areas, our results confirm the relevance of imagery acquired in the late season of vegetation. Our results also highlight the importance of time series to characterize riparian forest species while the health condition of A.glutinosa can be assessed through a single flight survey with satisfactory accuracy.

Future work should focus on the use of denser multi-temporal datasets, covering several entire vegetation seasons linked with the climatic conditions. This effort will limit the influence of early phenological disturbances and allow identification of more general phenological traits to develop better classification models.