Introduction

Natural disasters cause the loss of many lives and property in the world. Some of the most occurring phenomena like earthquakes, floods, landslides, and avalanche are often observed in Turkey due to geological, geomorphological, and climatic characteristics. Landslides are the most devastating disasters after the earthquakes in Turkey (Gökçe et al. 2008). Therefore, it is crucial to predict landslide-prone areas. For this purpose, the landslide susceptibility map (LSM) is prepared. The main objective is to make predictions for the future using existing data with various methods (statistical or machine learning) (Dikshit et al. 2020).

Spatial prediction of landslides is the probability of potential instability of slopes related to a set of causal factors (Guzzetti et al. 2005). It can be carried out by analyzing the spatial relationship between past landslide events and a set of geo-environmental factors (Pham et al. 2016). It assumes that future landslides will occur under the same conditions as previous landslides (Ermini et al. 2005). Previous landslides are identified, and a landslide inventory map is prepared (Guzzetti et al. 2012). Dependent and independent variables are determined. Dependent variables are landslide occurrences, and independent variables are determined by considering the geographical and geological conditions of the region. LSM is an outcome of the spatial prediction of landslides.

Geographic information system (GIS) based studies are used to produce LSM. Many methods have been applied for preparing LSM using GIS in recent years. These methods are expert opinion-based and data mining-based approaches (Song et al. 2012). The expert opinion-based approach is based on the perspective of experts in selecting variables and assigning weights to variables. This susceptibility modeling is one of the most suitable approaches for some regions with insufficient data in terms of spatial resolution and coverage (Ozer et al. 2020; Polat and Erik 2020). Data mining-based approach uses machine learning algorithms to determine factors leading to landslide occurrences and calculate weights of the factors during the learning of models. Out of these approaches, the data mining-based approach is more commonly used for creating LSM. Commonly used methods are logistic regression (LR) (Lee et al. 2007; Nefeslioglu et al. 2008; Xu et al. 2012; Devkota et al. 2013), decision tree (DT) (Nefeslioglu et al. 2010; Tien et al. 2012; Pradhan 2013; Hong et al. 2015), artificial neural network (ANN) (Das et al. 2013; Zare et al. 2013; Nourani et al. 2014; Dou et al. 2015) and support vector machines (SVM) (Tien et al. 2015; Hong et al. 2016; Chen et al. 2017). Frequency ratio (FR) and information value (IV) methods, which are bivariate statistical analysis methods, are frequently used in LSM studies (Pourghasemi et al. 2012a, b; Umar et al. 2014; Aghdam et al. 2016; Cui et al. 2016).

ArcGIS (ESRI 2011) is a widely used GIS software in many fields. It supports Python and Visual Basic scripting. Jimenez-Peralvarez et al. (2009), have used Model Builder in ArcGIS (ESRI) for landslide-susceptibility assessment. Zhu (2010) developed a risk evaluation model for Earthquake-induced landslide with the help of a Model Builder. Zhang et al. (2014) proposed a scripting model called GIScript, which presents a conceptual framework of directing access to geospatial data and parallel processing. Dobesova (2011) has discussed the relevance of the Python programming language for data processing in ArcGIS. Luo et al. (2011) have taken up the application of Python language and ArcGIS software into analyzing the Gorges Reservoir Area Remote Sensing data. Akgün et al. (2012) developed an easy-to-use program “MamLand” for the construction of LSM. This program was employed in Matlab based on expert opinion creates LSM using Mamdani fuzzy inference system. In Torizin (2012), an ArcGIS-toolbox ‘Landslide Susceptibility Assessment Tools’ (LSAT) was presented. This toolbox includes the FR and weight of evidence (WofE) as well as the multivariate LR. A single and standalone application for constructing LSM was developed by Osna et al. (2014). Jebur et al. (2015) proposed a tool applied in the ArcMap. This tool constructs LSM using FR, WofE, and evidential belief function methods. A toolbar was created thanks to the software development kit available with ArcGIS v.10 by Palamakumbure et al. (2015). This tool includes data preparation, model optimizing, derivation of decision trees, calculating predictions, and validation tasks. Sezer et al. (2016) developed an expert-based LSM module for Netcad Architect Software. This module uses the methods of modified analytical hierarchy process (M-AHP) and Mamdani type fuzzy inference system (FIS). In Arca et al. (2018), a Python script, integrated within the ArcGIS, was developed to perform all the calculations and plot drawings of LSM. Narayanan and Sivakumar (2018) created a Python-based customized toolbox as an extension tool to ArcGIS that helps to understand the spatial subsurface events through the development of the seismic information system. Bragagnolo et al. (2019) created a tool written in the python language named r.landslide and was integrated within Qgis. Köse and Turk (2019) developed user interface programs creating LSM and calculating accuracy for methods of FR and WofE in the GIS environment.

In this study, landslide susceptibility maps related to the Akıncılar region were generated by using five methods (FR, IV, RF, MLP, LR) on the LSAT. The Weka tool (ver.3.8) (Eibe et al. 2016) was used to compare results (for LR method) and test for dataset produced by LSAT.

The main contributions of this paper are as follows:

  • LSM of Akıncılar region is created automatically with high performance.

  • A novel toolbox called LSAT is developed.

  • LSM is done quickly and easily thanks to LSAT.

  • Best model parameters are selected with the use of tuning scripts.

Materials

Study area

The study area is located 150 km northeast of Sivas city (Fig. 1). This region covers an area of approximately 319.886 km\(^2\). Middle of the basin has an approximately flat area, but south margin and north margin have about 60\(^{\circ }\) slope values. The elevation is low in the middle sections. Contrary, there are high mountainous areas in the north and south.

The North Anatolian Fault Zone (NAFZ) is the primary continental strike-slip fault system in Turkey. The study area is located on NAFZ. Due to this configuration of the region, landslides are observed frequently. The study area is also a part of the Suşehri pull-apart basin formed by NAFZ (Polat et al. 2014). NAF passes through the middle of the basin (Fig. 2). The 1939 Erzincan Earthquake (M = 7.8) occurred on the NAFZ and created a 360-km-long surface rupture. The recent landslide that happened in this fault zone is the Kuzulu landslide (Gökçeoğlu et al. 2005; Tatar et al. 2005; Ulusay et al. 2007; Polat and Gürsoy 2014). 15 people died due to occurred this landslide in the northwest of the Koyulhisar on 17 March 2005.

Fig. 1
figure 1

Location map of the study area

According to the geological map prepared by the General Directory of Mineral Research and Exploration (Akbaş et al. 1991), a large part of the study area consists of Upper Miocene–Pliocene age conglomerate–sandstone–mudstone. Quaternary aged alluvium and Eocene aged andesite–basalt are the other most observed units in the region. Younger units are observed in the middle section of the study area, while older units are seen in the north and south edges (Fig. 2). As shown in Fig. 2, the great majority of landslides are observed in units of conglomerate-sandstone and mudstones.

Fig. 2
figure 2

Geological map of the study area and training landslides and test landslides

Installing required libraries

LSAT works on ArcGIS software with the Windows platform. Before using the LSAT, the required installations must be done. Python 2.7 is already installed with ArcGIS 10.4. The libraries must be compatible with the version of Python 2.7. Pip is recommended for easy installation. If a different version of python is installed, pip2.7 must be used. Scikit-learn (Pedregosa et al. 2011) is the main library for our tasks. Numpy (version of 1.8.2 or higher) and SciPy (version of 0.13.3 or higher) must be installed before installing Scikit-learn. Python 2.7 supports the versions of Scikit-learn 0.20 and earlier. In this study, the version of 0.20.2 was installed. Also, Numpy 1.15.4, Pandas 0.16.1, and Matplotlib 1.4.3 were installed. In addition to these, the C++ compiler must be installed for windows.

Preparing data

Area boundary data, controlling factors (parameters) data, and landslide inventory data are required as input parameters to create an LSM. All required parameter data were created and reclassified in GIS environment. Various sampling methods are applied to create LSM (Dagdelenler et al. 2016). In this study, scarp of landslides was used as landslide data. These data were created manually in polygon type. A Python script was written to do the next steps. The name of the script is preparing data. Preparing data script processes the data and prepares it for analysis. Figure 3 shows the flow diagram of these processes. This script needs to landslide inventory data, area boundary data, and reclassified controlling factors data, as mentioned above. The outputs of the script are landslide validation data, landslide training data, and all parameter data. Training data and all data will be used in the analysis processes. Landslide validation data will be used for validating the performance of prediction.

Fig. 3
figure 3

Data preparation flow chart

The area, first input parameter, was used to clip input raster data. Also, the values of controlling factors were extracted through this parameter. Users may select any desired region for analysis. In this study, the Akıncılar county boundary was used as area data. Area data was converted to point. This data consists of 511,828 pixels. All parameter pixel values were extracted and assigned to this point data.

Data preparation script splits landslide inventory data as training and test. A total of 100 landslide data were obtained in the study area. Random selection is made based on polygon-type landslide. There is no standard value for splitting data as a training set and test set. 70\(\%\) of landslides data were selected as training, and 30\(\%\) of landslides data were selected as validation randomly (Chen et al. 2017; Arabameri et al. 2019; Nohani et al. 2019). Users can also select the splitting ratio on the user interface.

All training pixel values were converted to point data. “1” value was assigned to landslide pixels, and “0” value assigned to the no-landslide pixels. Training landslide pixel count is 24,015 for this study area. It was preferred to use the same amount of no-landslide pixels. Therefore, 24,015 no-landslide pixels were selected randomly. So, 48,030 pixels were generated as training data. All parameter pixel values were extracted from this point data. Then, FR and IV values of all parameters were calculated. In this study, IV values of parameters were used as output data. This data can be used in two ways. External tools or our tool can be used to predict landslide probability. If an external tool or software was used, LSM could be constructed, and ROC values can be calculated with the “Create LSM and Calculate ROC” script in our tool.

Using the accurate data is very important in this type of analysis. Therefore, detailed field studies were carried out for defining landslide locations. In the study area, mostly complex type landslides were observed. Large-scale landslides and small-scale secondary landslides developed within them were seen in the study area (Fig. 4). The movement started as a rotational slip in areas higher elevation and close to the ridge. Then, it turned into a flow type due to the high slope. Scarps of all landslides were used, and toes of the landslides were ignored in the analysis.

Fig. 4
figure 4

Landslide image from study area: (a) main landslide body, (b), (c) and (d) secondary landslides

Another important criterion is to select landslide controlling factors. The most used controlling factors in the literature have been selected for the study area considering the geographic and geological location of the region (Hasekioğulları and Ercanoğlu 2012). Ten controlling factors were used in this study. These are elevation, aspect, curvature, slope, geology, distance to ridge, Normalized Difference Vegetation Index (NDVI), Topographic Ruggedness Index (TRI), Topographic Wetness Index (TWI), Slope Length and Steepness Index (ls factor) (Fig. 5). Distance to faults parameter was reclassified as <500 m, 500–1000 m, 1000–1500 m, 1500–2000 m, 2000–2500 m, 2500–3000 m, >3000 m. Analyses were made with these values. However, the performance decreased, and it caused the overfitting problem. So, this parameter was not used in analyses.

Fig. 5
figure 5figure 5

Landslide controlling factors: a curvature, b aspect, c lithology, d elevation, e Normalized Difference Vegetation Index (NDVI), f Slope Length and Steepness Index (ls factor), g slope, h distance to ridge, i Topographic Ruggedness Index (TRI), j Topographic Wetness Index (TWI)

In this study, ArcGIS and Saga-GIS (Conrad et al. 2015) software were used for GIS studies. Firstly, \(25 \times 25\) m resolution DEM data was created from 1/25,000 scaled topographical map in ArcGIS. Slope angle, aspect angle, curvature data were derived from DEM by ArcGIS and TWI, TRI, LS factor, ridge data were obtained from Saga-GIS. NDVI was calculated by using Landsat 8 satellite image in ArcGIS. The parameters used in the analyses are explained in detail below.

Curvature: The curvature is a parameter that defines the morphology of the topography, and the terrains are classified as concave, convex, and flat. Curvature is a useful factor in the landslide occurrences. For example, flow-type landslides are more likely to occur in convex areas. Curve data were reclassified into three classes. It is assumed that the curve values lower than \(-0.2\) are concave, curve values higher than 0.2 are convex, and curvature values between these values are flat (Maggioni and Gruber 2003).

Slope aspect: The slope aspect describes the direction of the slope. The slope direction is related to the amount of precipitation and sun exposure. These two properties affect slope stability. In this study, this parameter was reclassified into nine classes. The FR of the flat area is minimum. The maximum FR is observed in the west direction.

Lithology: Lithology can provide valuable information for landslide prediction studies. Lithological properties are directly related to many properties of materials such as strength, permeability, and hardness (Baeza and Corominas 2001). Therefore, lithological features of the area should be appropriately evaluated, and the convenient classification value should be selected. Geological units were simplified as ten groups in the study area (Table 1). Lithological units with similar characteristics were placed in the same groups. The final decision was made by studying the relationship between landslides and lithological units in the geological map.

Table 1 Simplified geological units

Elevation: Elevation value is frequently used in this type of studies. The minimum height is 832 m, and the maximum height is 2230 m in the study area. These values were classified as six classes, as in Table 2. A linear relationship was observed between the elevation and possibility of landslide until 2000 m. The FR value is zero over 2000 m.

Normalized Difference Vegetation Index (NDVI): Plants have the effect of stopping or increasing movement on slopes. Stopping effect is observed, especially in shallow landslides, and the increasing effect is observed in a deep landslide (Eker and Aydın 2014). The Normalized Difference Vegetation Index (NDVI) is one of the frequently used environmental factors in the assessment of landslide susceptibility. The red band and the Near-Infrared band (NIR) are used for calculating NDVI (http://www.esri.com). Landsat 8 image acquired from USGS Earth Explorer was used for calculating NDVI. NDVI is calculated as follows:

$$\begin{aligned} {\text {NDVI}} = \frac{{{\text {NIR}} - {\text {Red}}}}{{{\text {NIR}} + {\text {Red}}}}, \end{aligned}$$
(1)

where NDVI is Normalized Difference Vegetation Index, NIR is Near Infrared Band (band 5), and Red is the red band (band 4) of Landsat 8 image. NDVI data were classified into four classes, as shown in Table 2.

Slope length and slope angle (LS-factor): LS-factor has a considerable effect on soil loss. The L-factor is the impact of slope length, and the S-factor is the effect of slope steepness. These two factors are significant for landslide occurrence. The LS-factor calculation was performed using the equation proposed by Boehner and Selige (2006) and implemented using the System for Automated Geoscientific Analyses (SAGA) software. LS data were classified into six classes. The values with the highest landslide density are between 5 and 15, and as shown in Table 2.

Slope angle: Slopes have a significant influence on landslide occurrence as they affect the flow of water and the soil (Jones et al. 1983). The slope angle was reclassified into ten classes. The maximum FR was obtained between 15\(^{\circ }\)–20\(^{\circ }\).

Distance to ridge: It has been determined that landslides occur more frequently in areas close to the ridges. Therefore, the distance to the ridge parameter was used as a landslide controlling factor. The landslide scarps are also commonly observed in areas close to the ridge. Ridge data were classified at intervals of 150 meters. As shown in Table 2, the maximum FR is between 150 and 300 m.

Topographic Wetness Index (TWI): TWI is a tool indicating areas accumulating water flow, often with the seasonally and permanently waterlogged ground. As such, it is beneficial to show the geomorphic complexity of a landslide terrain, including the pattern of topographic highs (‘dry’ areas) and lows (‘wet’ areas) (Rozycka et al. 2016). TWI is defined by the following equation:

$$\begin{aligned} {\text {TWI}} = \ln A/\tan \beta , \end{aligned}$$
(2)

where A is the upslope contributing area, and tan\(\beta\) is the local slope (Moore et al. 1991; Sorensen et al. 2006; Gruber and Peckham 2009). TWI data were classified into six classes, as shown in Table 2.

Topographic Ruggedness Index (TRI): TRI is a measurement developed by Riley et al. (1999). TRI corresponds to average elevation differences between any point on a grid and its surrounding area (Rozycka et al. 2016). TRI is expressed as follow:

$${\text {TRI}} = \left(\sum {(Z_{{\text {c}}} - Z_{i} )^{2} } \right)^{2},$$
(3)

where \(Z_{{\text {c}}}\) is the elevation of a central cell, and \(Z_i\) is the elevation of one of the eight neighbouring cells (i = 1, 2, ..., 8). TRI data were classified into five classes, as shown in Table 2.

Methods

In this study, GIS, statistics, and machine learning methods were used. GIS-based studies were used for preparing data. Model results were processed in a GIS environment to display landslide susceptibility index (LSI). Weka tool and Python (Sklearn library) were used for creating a classification model and analysis. LSAT contains 10 Python scripts that were written by the author and integrated into the ArcGIS. The first script is used to prepare data for modeling. The second script is used to create LSM and to calculate the area under curve (AUC) values. The other scripts are used to construct LSM with the methods of FR, IV, LR, Random Forest (RF), and multi-layer perceptron (MLP). Data preparation script was tested to predict landslide probability in Weka software with the LR algorithm. Then classification results were processed with create LSM and calculate ROC script in GIS and susceptibility map was created.

Frequency ratio method

FR method proposed by Lee and Talib (2005) is based on density analyses. FR is the ratio of the area where landslides occurred in the total study area and the ratio of the probabilities of a landslide occurrence to a non-occurrence for a given attribute (Bonham-Carter 1994). The proportion of landslide occurrence to non-occurrence (Regmi et al. 2014) was calculated for each parameter. The following equation defines the FR:

$$\begin{aligned} {\text {FR}} = \frac{{N_{{{\text {pix(si)}}}} /N_{{{\text {pix(Ni)}}}} }}{{\sum {N_{{{\text {pix(si)}}}} } /\sum {N_{{{\text {pix(Ni)}}}} } }} . \end{aligned}$$
(4)
N\(_{\text {pix(si)}}\)::

The number of pixels containing landslide in class (i).

N\(_{\text {pix(Ni)}}\)::

Total number of pixels having class (i) in the whole area.

\(\sum N_{\text {pix(si)}}\)::

Total number of pixels containing landslide.

\(\sum N_{\text {pix(Ni)}}\)::

Total number of pixels in the whole area.

In this method, it is stated that FR values greater than 1 are relatively more effective in landslide occurrence, and FR values less than 1 have less effect in landslide occurrence (Lee and Talib 2005). The FR method was implemented with the FR script in a GIS environment. The obtained FR values are assigned to parameters and subgroups to construct LSM in the GIS environment. The FR value was calculated, as shown in Table 2. This script requires data created with the Data Preparation script.

Information value method

The IV method, a statistical approach, was used for calculating the weight of parameters like FR. Yin and Yan (1988) suggested this method. The weight value is defined as the natural logarithm of the ratio of the landslide density in the relevant parameters and subgroups to the landslide density in the total area (Van Westen et al. 1997). The following equation defines the weight:

$$\begin{aligned} W_{i} = \ln \frac{{\frac{{N_{{{\text {pix}}}} Si}}{{N_{{{\text {pix}}}} Ni}}}}{{\frac{{\sum {N_{{{\text {pix}}}} } Si}}{{\sum {N_{{{\text {pix}}}} } Ni}}}}. \end{aligned}$$
(5)
W\(_i\)::

weight of parameter

N\(_{\text {pix}}Si\)::

number of pixels of the landslide within class i

N\(_{\text {pix}}Ni\)::

number of pixels of class i

\(\sum N_\text {pix}Si\)::

number of pixels of the landslide within the all study area

\(\sum N_{\text {pix}}Ni\)::

number of pixels of the all study area.

Table 2 Frequency ratios and information values of landslide conditioning parameters

Negative values of weight show lower density, and the positive values of weight show a higher density of landslide relatively (Van Westen 1997).

Any GIS software can determine the pixel number of parameter and their distribution in landslide pixels. Data preparation script was used to calculate weight with the IV method. Then IV script was used to create LSM and calculate AUC value. The AUC value of the IV method is higher than the FR method. So, the weights of the IV method were used as the input parameter for the LR, RF, and MLP algorithms.

Logistic regression method

LR is a method frequently used in the literature in recent years (Chen et al. 2018, 2019; Benchelha et al. 2019; Sevgen et al. 2019; Sahin et al. 2020). The dependent variable is categorical. Landslide pixels are represented by 1, and non-landslide pixels are represented by 0. Parameters are independent variables. LR is a method used to determine the cause-effect relationship with independent variables. The most important feature of the LR method is that it provides a direct calculation of the estimated probability values of the dependent variables, regardless of the assumptions made in factor analysis or discriminant analysis. LR formula is given as follows:

$$\begin{aligned} \Pr = \frac{{e^{u} }}{{1 + e^{u} }}, \end{aligned}$$
(6)

where Pr is the probability of a landslide occurrence and u is the independent variable.

\(\beta\)1, \(\beta\)2, \(\beta\)3 are corresponding coefficients to each of the respective contributing factors.

$$\begin{aligned} u=\beta 0+\beta 1x1+\beta 2x2+\beta 3x3. \end{aligned}$$
(7)

LR script and Weka software were used for constructing the susceptibility map by LR. Weka was used because of the need for external software to test the scripts of data preparation and create LSM and calculate ROC.

Data preparation script saves the weight values as CSV format. This file is easily used in Weka. After the required data is imported, tenfold cross-validation was applied for the logistic regression model in Weka. Coefficient values were calculated as 0.002, 0.0012, 0.0106, 0.0281, 0.0035, 0.0098, 0.0083, 0.0117, 0.0028, 0.0094 for Aspect, Curve, Elevation, Lithology, Ls-factor, NDVI, distance to ridge, Slope, TRI and TWI, respectively. Intercept value was calculated as \(-6.5421\) for this method by Weka. After this process, the model was evaluated on all data set. Class distribution values were used for prediction. Distributions of class-1 (probability of landslide occurrence) values were exported to CSV format, and the coordinate of a pixel was added. This file was used as an input parameter of create LSM and calculate ROC script. Both scripts worked successfully. Success rate and predict rate were calculated as 75.40 and 72.57. These results obtained with the Weka software are the same as the results of the LR script in the LSAT tool.

Tuning LR method

Parameters of C, max_iter, and solver, are used as tuning parameters in tuning LR script. C is inverse of regularization strength.

$$\begin{aligned} C=1/\lambda . \end{aligned}$$
(8)

Lambda (\(\lambda\)) controls the trade-off between allowing the model to increase its complexity as much as it wants with trying to keep it simple. For example, if \(\lambda\) is very low or 0, the model will have enough power to increase its complexity (overfit) by assigning high values to the weights for each parameter. If, on the other hand, the value of \(\lambda\) is increased, the model will tend to underfit, as the model will become too simple.

Scikit-learn offers several techniques for training LR, called solvers. These are Newton-cg (Newton’s method), lbfgs (Limited-memory Broyden–Fletcher–Goldfarb–Shanno Algorithm), liblinear (a Library for Large Linear Classification—Fan et al. 2008), SAG (Stochastic Average Gradient—Schmidt et al. 2017), SAGA (a variant of SAG—Defazio et al. 2014).

The last parameter is max_iter, the maximum number of iterations taken for the solvers to converge. These three parameters can be tuned individually, or the best parameters can be selected automatically. GridsearchCV method was used to select the best three parameters for the LR algorithm. GridsearchCV is the method that finds the best parameter values using the defined parameter values for an estimator. This method uses cross-validation. Fivefold cross-validation was used in the tuning script. LSM was created successfully with these parameter values of the GridsearchCV method. Tuning parameters and their values are shown in Table 3.

Table 3 Tuning parameters of LR

Random forest method

RF is one of the machine learning algorithms developed by Breiman (2001). It is an ensemble machine learning algorithm used the method of Bootstrap Aggregation or bagging. RF creates multiple decision trees and combines them to achieve a more accurate and stable prediction. The samples in each tree are randomly selected. RF uses the majority voting method for the result. It can be used for both classification and regression problems. RF has the capability of robust and accurate performance on complex and big datasets with little need for fine-tuning. Another significant feature of RF is that it is more resistant to over-fitting. RF method has been frequently used recent LSM studies (Taalab et al. 2018; Kim et al. 2018; Dou et al. 2019; Park and Kim 2019).

Tuning RF algorithm

Machine learning algorithms need to be tune of parameters for more accurate predictions. The complexity and size of the trees should be controlled by setting hyperparameter values to reduce memory consumption.

The tuning script includes parameters of n_estimators, max_depth, min_samples_splits, and min_samples_leaf for tuning. Also, the user can use the RandomizedSearchCV (RSCV) method to choose the best value of the parameters. It is recommended to use the RSCV method first.

RSCV implements a fit and a score method. It also implements predict, predict_proba, decision_function, transform, and inverse_transform if they are implemented in the estimator used. RSCV uses a cross-validation method for searching. The tuning parameters are shown in Table 4.

N_estimators is the number of trees in the forest. Mostly, the higher the number of trees, the better the learning rate. Max_depth is the depth of each tree in the forest. The greater the depth, the more information is obtained about data. Min_samples_split is the minimum number of samples required to split. Min_samples_leaf is the minimum number of samples required to be at a leaf node (Ben Fraj 2017).

Table 4 Tuning parameters of RF

Also, each parameter can be tuned individually. All tuning results could be saved as a png file for visualization.

RandomizedSearchCV results were used as parameter values, and LSM was created. But LSM was not reliable, and the difference between the success rate and prediction rate was too much. Therefore, parameters were tuned individually. Integer and float values were tested for min_samples_splits and min_samples_leaf. It was observed that integer values were better than float. After tuning process, n_estimators=2000, max_depth=4, min_samples_splits=2, criterion=entropy and min_samples_leaf=100 values are selected as the best values of parameters.

Multi-layer perceptron method

MLP is a supervised learning method used in many areas. It is also a preferred method in LSM studies due to its computational simplicity, finite parameterization, and stability. The simple perceptron model has only the input and output layer. The input layer sends incoming data to the hidden layer(s). The incoming information is transferred to the next layer. The number of hidden layers varies according to the problem, at least one, and adjusted according to the need. The number of neurons in the layer is also determined by the problem. The output layer determines the output of the network by processing the data from the previous layers. Any mathematical function can be used as an activation function in the model. The learning method of the system consists of two stages in general. The first part is a forward calculation (feedforward). The second part is a backward calculation (backpropagation).

MLP script uses an MLP classifier algorithm that trains using backpropagation. Landslide factors are the input feature, and landslide probability is the output layer. One hidden layer or more hidden layer can be used. Using multiple layers often does not increase accuracy (Rumelhart et al. 1986; Lippman 1987).

Tuning MLP algorithm

MLP classifier has many parameters (see scikit-learn.org). In this study, parameters of hidden_layer_size, activation, solver, alpha, learning_rate, learning_rate_init, max_iter, and momentum were used as tuning parameters. The methods of RSCV was used for the best parameter selection. After assessment of tuning results with RandomizedSearchCV method, the values of parameters were selected as shown in Table 5.

Tol parameter is used to stop training when the loss or score is not improving. The default value of tol is \(1e-4\). The values of \(1e-5\) and \(1e-6\) were tested with 20,000 iterations, and MLP results show overfitting with \(1e-6\) value. MLP script was executed with the above values of parameters, and LSM was created successfully.

Table 5 Tuning parameters of MLP

Results and discussion

There are many studies in literature related to the assessment of landslide susceptibility, as described in the introduction. These studies will continue in the future to create a more accurate result. One of the most critical processes in these studies is the preparation of data. It takes a long time and requires attention. Pixel values of the raw data need to be extracted in this stage. This process creates a large data set.

Moreover, if the number of parameters increases, the data size increases too. Evaluation of prepared data could be done with different algorithms in the analysis phase. Machine learning methods are mostly used to predict landslide areas. The results of the analysis need to be visualized in the GIS environment. The other process is the validation of LSM. This step is necessary to test the accuracy of the analyses.

In this study, a novel tool was developed to produce LSM. LSAT automatically prepares data, analyses, calculates the success and prediction rate, and creates result maps. LSM of the study area was prepared by using LSAT with five different models. All LSM were reclassified into five classes, such as very low, low, moderate, high, and very high using the natural break method (Fig. 6). ROC analysis was used to assess the accuracy of the models. ROC Curves plot the true positive rate (sensitivity) against the false positive rate (1-specificity). AUC values are generally evaluated between 0.5 and 1. Yeşilnacar and Topal (2005) mention the relationship between the prediction accuracy and the AUC value that can be classified as follows: 0.5–0.6 (poor), 0.6–0.7 (average), 0.7–0.8 (good), 0.8–0.9 (very good), and 0.9–1 (excellent).

Fig. 6
figure 6

LSM of study area constructed with the methods of a FR, b IV, c LR, d LR (Weka), e MLP and f RF

ROC calculation script accepts susceptibility map data (in raster format), train data (polygon type), and test data (polygon type) as the input parameters. The outputs are the ROC curve and AUC values of test and train data. Landslide validation data was used to calculate the prediction rate. Landslide training data was used to calculate the success rate of LSM. For this method, the LSM was reclassified into 100 equal zones in GIS environment. Train and validation landslide pixel counts were defined for every class. Cumulative pixel values of zones were used to calculate the true positive rate (TPR) and false positive rate (FPR). The TPR formula is as follows:

$$\begin{aligned} {\text {TPR}} = \frac{{{\text {TP}}}}{{{\text {TP}} + {\text {FN}}}}, \end{aligned}$$
(9)

where TPR is true positive rate (sensitivity), TP is true-positive, and FN is false-negative. The false positive rate (FPR) is equal to 1-specificity. FPR formula is as follows:

$$\begin{aligned} {\text {FPR}} = \frac{{{\text {FP}}}}{{{\text {FP}} + {\text {TN}}}}, \end{aligned}$$
(10)

where FPR is false positive rate, FP is false-positive, and TN is true-negative. Then the trapezoidal formula was used to calculate the AUC (Bernard and Liengme 2002). This formula is as follows:

$$\begin{aligned} \begin{aligned} \begin{aligned} {\text {AUC}}&= \left( {\left( {\frac{{y_{1} - y_{2} }}{2}} \right) \times (x_{1} - x_{2} )} \right) \\&\quad + \left( {\left( {\frac{{y_{2} - y_{3} }}{2}} \right) \times (x_{2} - x_{3} )} \right) \\&\quad + \cdots \left( {\left( {\frac{{y_{{n - 1}} - y_{n} }}{2}} \right) \times (x_{{n - 1}} - x_{n} )} \right) , \\ \end{aligned} \end{aligned} \end{aligned}$$
(11)

where y0 is sensitivity values (TPR), and x is 1-specificity (FPR) values.

The prediction and the success rates of the models are shown in Fig. 7. AUC values were calculated as 70.95 \(\%\), 71.85 \(\%\), 72.57 \(\%\), 72.67 \(\%\), 73.93\(\%\) for prediction rate of FR, IV, LR, MLP and RF methods, respectively. AUC values also were calculated as 72.71 \(\%\), 72.94 \(\%\), 75.40 \(\%\), 75.39 \(\%\), 77.39 \(\%\) for success rate of FR, IV, LR, MLP and RF methods, respectively. The success and prediction capability of the RF algorithm are better than the other algorithms. The model derived using FR has the lowest success and prediction capability (Fig. 7).

Fig. 7
figure 7

ROC curves of models a FR, b MLP, c LR (Weka), d RF, e IV, f LR

Elevation, aspect, curvature, slope, geology, distance to ridge, NDVI, TRI, TWI, ls-factor were used as landslide controlling factors. Fault data were not used in this study. When the fault distances data were used, it was observed that the value of the success rate increased, and the value of the prediction rate decreased. Different training and test data sets experimented; different class ranges were tested, but the problem could not be solved. This issue could be about the accumulation of faults in a flat area in the middle of the basin. The parameter of distance to road was not used because it did not show a proper distribution for the whole area. LS factor parameter was preferred instead of the proximity to streams.

The main purpose of this study is to make the landslide susceptibility assessment quickly and accurately with a GIS tool. The other tools designed for this purpose use statistical methods such as FR, WofE. Most of them are prepared with a model builder. Our tool uses both statistical and machine learning methods. In this study, only five methods were used. Different methods can also be added to this tool if desired.

Interfaces of all scripts, analysis processes, and processing times are given at https://github.com/apolat2018/LSAT. Also, the elapsed time for scripts is given in Table 6. The elapsed time varies according to the size of the study area, the number of used parameters, and the computer speed. In this study, ten raster data (controlling factors) covering an area of 319.886 km\(^2\) were used. One raster data corresponds to 511,828 pixels. As seen in Table 6, the most time-consuming processes are MLP script, data preparation script, and create LSM and calculate ROC script, respectively. Other scripts give results in seconds.

Table 6 Elapsed time for scripts (m:minutes, s:seconds)

Conclusions

The main objective of this study is to prepare a tool for creating LSM automatically. In literature, the LSM tools were created by bivariate statistical analysis methods. The tool developed in this study also includes machine learning methods. LSAT includes five methods (FR, IV, LR, RF, MLP). Also, data preparation, create LSM and calculate ROC, and tuning (LR, RF, and MLP) scripts were created by Python. Data preparation script makes the data ready for analysis using the weights of FR and IV. Create LSM and calculate ROC script visualizes the LSM and calculates the AUC value if external software is used for analysis. Weka software was used to create a data set with the LR method for this purpose. Tuning scripts are used to select the best values of parameters with the methods of GridSearcCv and RandomizedSearchCV. Tuning scripts were supported with graphics.

In conclusion, these scripts have greatly accelerated and simplified the data preparation and analysis processes. The best values of model parameters can be selected thanks easily to tuning scripts with machine learning methods. LSAT is useful for automatically creating LSM with five different methods, tuning of the LR, RF, MLP models, and comparing models for any area. LSAT was created with the Python language. It has very comprehensive machine learning libraries. LSAT was developed by selecting methods that are frequently used in LSM studies. In future work, different methods can be used, such as deep learning algorithms, and LSAT can be developed. Moreover, LSM can be created with various python libraries without using a GIS environment.