1 Introduction

Any forecast is characterised by its accuracy. The forecasting accuracy depends on extrinsic factors (i.e., factors depending on the data set subjected to the analysis, such as data accuracy and data availability) and intrinsic factors (i.e., factors depending on the forecasting method). While a person performing a forecast can not control the extrinsic factors, the forecasting method could be chosen and improved in order to obtain higher levels of forecasting accuracy. This is achieved by using modern tools, such as artificial intelligence, which is today widely recognised as a popular tool for economic forecasting.

Artificial intelligence (AI) is a relatively new branch of computer science. Its origins lie in the mid-1950s, when the growing desire to obtain a machine able to reproduce human intelligence was equalled and surpassed by the development of new technologies. The most challenging AI task was creating an artificial version of the human central nervous system. This was achieved by artificial neural networks (ANNs). An ANN is strongly related to the mathematical optimisation process, i.e., finding a functional minimised by a given dataset. The obtained functional was claimed to extrapolate any dataset related to the initial data. With the development of computers and computer software, ANNs became increasingly complex and efficient and found applicability in various domains.

The learning capacity is probably the fundamental feature of a neural network, as modern algorithms plan to emulate this characteristic within computers. The fundamental ANN features can be divided into two categories: architecture and properties. An ANN’s architecture defines the structure, giving the number of neurons and their connections; other features are the Input/Output (I/O), synapse intensity, deviations and activations. An ANN’s properties concern the learning method, synapse reactivation, continuous associations and comparison of the new information with the existing information and how the new information is categorised.

Fausett (1994) classified ANNs and highlighted their existing applications in signal processing, car control, pattern recognition (text recognition), medicine (diagnosis and treatment suggestions), speech production and recognition and business (mortgage assessment). Even more applications of ANNs were found (Taylor 1996), including, but not limited to, the automotive and aerospace industry, digital data processing and transmission, texture recognition, clinical psychiatry, business and finance. These applications underlined the advantages of ANNs in solving complex problems, even in real-time. The financial applications considered (estimating means and medians of financial data, an on-line system for such data analysis, an S&P 500 predictor and a more general time series predictor) again demonstrate the power of forecasting and analysis of ANNs in economics. Russell and Norvig (2002) highlighted AI applications used until then: autonomous planning and scheduling, game playing (chess), autonomous car controls, medical diagnosis, logistic planning, robotics (including microsurgery), language understanding and problem solving. By now, AI has applications in almost every field of human endeavor. All of these applications involve algorithms for processing information, decision making and acting optimally based on the feedback of the environment.

Due to its higher accuracy level, AI has been recognised as a resourceful approach in economic forecasting. Balan and Morariu (2008) underline differences between AI forecast methods, including genetic algorithms and ANNs. Their accuracy levels differ slightly, with the main element of comparison being application complexity and their target groups.

The use of ANNs as an economic forecast method regarding companies’ evolution dates from 1997. The study developed by Wu (1997) suggested using an ANN for identifying firms that need more complex auditing investigations and underlined the superiority of AI methods over traditional methods in forecast accuracy.

From an investment perspective, Pinson used AI in 2006 to create a multi-expert approach based on a meta-model for business risk assessment. The advantage compared to a traditional method is an adaptable system capable of giving dynamic resolution strategies to the problems. From a similar business forecast approach, AI methods have been used to forecast corporate dividends.

In 2008, Kim et al. characterised the viability of Small Manufacturing Enterprises (SMEs) by employing methods such as adaptive learning networks. The main product and management characteristics of SMEs were investigated to establish the main influence factors for their long-term survival. Their overall accuracy was 61.31 %.

In 2011, Chen proved the superior accuracy of AI methods over traditional methods in forecasting corporate financial distress. The AI methods used were principle component analysis, decision trees and logistic regression.

The Eastern European economic climate is still influenced by the political development of the region in the second half of the twentieth century. Immediately after 1990, there was a widespread enthusiasm for starting new private companies. This enthusiasm is measured at an economical level through entrepreneurial indicators. In the short term, the number of private companies increased, but, as the enthusiasm was not always a good substitute for managerial skills and experience, many new-founded companies faced bankruptcy (Pirtea 2003). Moreover, the actual state of economic crisis severely affected the young Eastern European private companies. The uncertainty of the medium- and short-term situation of a company caused unwanted market blockages. The need for a good forecasting tool for the bankruptcy of Eastern European companies thus arises. Several authors used neural networks to meet this need, including Darvasi (2010), Dorneanu et al. (2011) and proposed strategies for local use (Mnerie and Mnerie 2011).

Within the framework of European competitiveness policies, developing specific instruments for firm viability is a necessity, especially in the context of economic crisis. There is currently a push to rebuild the trust of economic agents to re-establish normal business trends (Dogaru et al. 2011). The present economic crisis led to the bankruptcy of an important part of the SMEs. Moreover, this decline led to a recession of the global economy, especially in Eastern Europe, as shown by Kim et al. (2010), Dorneanu and Untaru (2011) and also by surveys of Directorate General Economic and Financial Affairs (2011). Such a negative economic event reflected upon less-developed regions, influencing entrepreneurial behaviour (Mnerie and Mnerie 2011). According to the Directorate General of Economic and Financial Affairs, the economic sentiment indicator in the EU has decreased from a value of 106.4 in December 2010 to 92.8 in November 2011. In this view, elaborating more accurate instruments for firm viability and risk assessment gained importance (Kim et al. 2008).

In response to the need for an efficient instrument that could forecast firms’ viability, we create an optimised neural network. The network considers four relevant factors regarding companies’ financial states and returns a forecast of their bankruptcy state. Our work thus offers valuable information applicable to Eastern European SMEs and potentially beyond.

2 Methods

2.1 Data set

The dataset used for the computations refers to the financial state of 55 Eastern European companies. The NACE codes of the companies, together with their net annual turnover, the value of the assets and the medium number of employees classify them as small manufacturing enterprises. The data was collected from official reports and refers to the period 1994–1998. Based on the collected data, several classes of indicators were further computed. These indicators can be grouped in five main categories:

  • Activity rates (5 indicators): stock rotation rate, debt collection rate, payment period of obligations, payment period of suppliers, asset leverage rate

  • Cash flow rates (2 indicators): immediate cash flow, current cash flow

  • Leverage rates (4 indicators): equity leverage, coverage rate of interests, coverage rate of liabilities with cash-flow, global leverage rate

  • Return rates (4 indicators): operation return rate, financial return rate, assets return rate, return revenue rate

  • Other economic and financial information (3 indicators): ratio of arrears over actives, ratio of arrears over turnover, dividend rate

A further stage in our work was to determine the indicators most relevant to the financial state of the companies. This stage was best described by Tucu and Rotarescu (2006) and will not be detailed here. The 4 indicators most relevant to the financial state of a company (revenue return rate, coverage rate of liabilities with cash flow, asset leverage rate and payment period of obligations) are computed as:

Next, an ANN is constructed and trained to recognise the state of bankruptcy of a company.

As the period covered by the collected data is prior to the period in which our initial studies were performed, it was possible to check the accuracy of the ANN forecast against the reality, i.e., for each of the companies considered, the actual state of bankruptcy or non-bankruptcy was checked in 2004.

The influence of the network parameters is next studied and the ANN is optimised to obtain a high level of forecast accuracy (over 95 %).

2.2 Structure of the ANN

The structure of the ANN refers to the network topology and the algorithms used (see Fig. 1). The neural network we developed is a classification (pattern recognition) network and is implemented using the neural network toolbox of Matlab (2011).

Fig. 1
figure 1

Neural network (Source authors’ elaboration with Matlab Program)

The network topology contains one input layer of four units; one, two or three hidden layers of 10 or 20 units; and one output layer with a single output (Slavici 2009). The four units of the input layer are financial rates, chosen from the classical forecast methods: revenue return rate (X1), coverage rate of liabilities with cash flow (X2), asset leverage (X3) and the payment period of obligations (X4) (Tucu and Rotarescu 2006).

The algorithms used are presented in the following.

The initial available data were divided into 3 categories - training, validation and test. Data division could be performed in various modes: randomly, in contiguous blocks, using an interleaved selection or simply by index. However, for our purposes, we choose a random division of the available data, with the ratios of 90 % for the training set, 5 % for the validation set and 5 % for the test set.

It should be underlined that the validation of the ANN was possible, due to the time lapsed between the period studied and the study taken. This time lapse was necessary to verify the bankruptcy state of the firms. Thus, it was possible to automatically check if the validation results were accurate or not.

The training of a neural network involves tuning the values of the weights and biases of the network to optimize network performance. Different algorithms could be used, (e.g., Levenberg-Marquardt, Bayesian regularization, BFGS quasi-Newton, resilient backpropagation, different types of conjugate gradients, one step secant, variable learning rate gradient descent, etc). All these algorithms compute the current weights and biases \((\mathrm x_{k+1})\) based on previous known weights, biases \((\mathrm x_{k})\), learning rates \((\alpha _k)\) and gradients \((\mathrm g_{k})\):

$$\begin{aligned} \mathrm x_{k+1}=\mathrm x_{k}-\alpha _k\,\mathrm g_{k} \end{aligned}$$
(1)

For our purpose, a scaled conjugated gradient back-propagation algorithm is used for the training phase, and the maximum number of epochs to train is set to 1,000. To improve network performance, early stopping of the training algorithm is also used.

The training algorithm ends in one of the following situations: the maximum number of epochs is reached (in our case, this value is set to 1,000), the maximum time is exceeded (which is not our case, as the process is static), performance is minimised to the goal (in our case, the default goal is 0), the performance gradient falls below a predefined value (in our case, the default value is \(10^{-6}\)) or validation performance has increased more than a predefined number of times since the last time it decreased (in our case, the default value 5). In our case, the training algorithm ends because the validation performance increased above the maximum value.

The performance of the training algorithm is measured by the mean squared normalised error function, i.e.,

$$\begin{aligned} F=\frac{1}{N}\sum _{i=1}^N(e_i-a_i)^2 \end{aligned}$$
(2)

where \(N\) is the dimension of the training set, \(e_i\) is the normalised value for estimated output corresponding to the variable \(i\) (network output), and \(a_i\) is the actual (apriori known) normalised output corresponding to the variable \(i\) (target output).

The derivative is the default derivative and in our case, it returns the performance gradient needed for the training algorithm.

2.3 Optimisation of the ANN

Using specific statistical tools, like Statgraphics Centurion (2011), one could find the most relevant factors influencing a certain characteristic, as well as the distribution of the required characteristic, as a function of these relevant factors. In the following, we describe this process, applied to the accuracy of forecast.

After performing a statistical analysis, the conclusion was that the factors with most influence on the forecast accuracy are the number of training epochs (e) the number of layers (s) and the number of neurons per layer (n). It should be underlined that the random data division assures the generality of the obtained results.

Next, the forecasting accuracy was computed for various values of the influencing factors:

  • For the number of epochs (e): 50, 100, 200 or 500;

  • For the number of layers (s): 3, 4 or 5;

  • For the number of neurons per layer (n): 10 or 20.

The early stopping condition was not imposed, so the experiments ended when the maximum number of training epochs was reached. Based on the obtained values, the distribution of the variable “forecasting accuracy” was computed and, hence, the optimal values for e, s and n for which the maximum of accuracy should be obtained.

3 Results and discussion

3.1 Training, validating and testing the ANN

To evaluate network performance, the confusion matrices and error rates were studied, for the two phases of the ANN construction:

Phase 1 Training the network using 49 companies.

Phase 2 Checking the network forecast accuracy using another 6 companies (3 for the validation phase and 3 for the test phase).

In the following, we describe the confusion matrices corresponding to the training and validation testing phases of the neural network build-up procedure, respectively, as well as a global confusion matrix embedding the three matrices. The matrices are structured such that the rows contain the network’s output and the columns contain the desired targets (Slavici et al. 2008):

$$\begin{aligned} \left( \begin{array}{ccc} true &{}&{} false \\ positive &{}&{} positive \\ &{}&{} \\ false &{}&{} true \\ negative &{}&{} negative \end{array}\right) \end{aligned}$$

For the training phase, from 49 total subjects, 24 (49.0 %) were marked true positives, 1 (2.0 %) as a false negative, 0 as false positives and 24 (49.0 %) as true negatives, leading to an overall accuracy of 98.0 %.

For the validation phase, from 3 total subjects, 2 (66.7 %) were marked true positive and 1 (33.3 %) true negative, leading to an overall accuracy of 100 %.

The test phase considered 3 subjects, 2 (66.7 %) marked as true positive and 1 (33.3 %) as true negative, with an overall accuracy of 100 %.

The overall accuracy is thus 98.2 %, corresponding to 28 (50.9 %) true positive, 1 (1.8 %) false negative, 0 false positive and 26 (47.3 %) true negative marks.

This result is compared to the prediction marks obtained with classical methods, which has an overall accuracy of 92.0 %, corresponding to 48 (42.8 %) true positive, 1 (0.9 %) false negative, 8 (7.1 %) false positive and 55 (49.1 %) true negative marks.

Figure 2 shows the overall evolution of the training phase. Two aspects were monitored: the performance gradient (i.e., the backward derivative of the performance with respect to weights and biases) and the validation test (i.e., the number of times the error on the validation set increased since the last decrease). In our case, the error on the validation set increased during the last six epochs. This situation means that for the last six epochs, the network begins to overfit the data, memorising the training data rather than learning to generalise from the training set, leading to the necessity of stopping the algorithm early to obtain a network with good generalisation properties. The early stopping technique states that the parameters assigned to the network are those for which the validation error is minimised. In our case, these parameters are those corresponding to the 38th epoch (Fig. 3).

Fig. 2
figure 2

Evolution of the training phase

Fig. 3
figure 3

Error evolution during all training epochs

Figure 3 displays the error evolution during all 44 training epochs. The training error (shown in blue) and, even more important, the test error (shown in red) and the validation error (displayed in green) are represented. A small training error shows that the model fits the training data well, while a lower testing error signifies that the neural network displays good performance on real-life data that differs from the training set. In our case, the best validation error was obtained for the parameters corresponding to the 38th training epoch.

The results obtained using the ANN were further compared to the results obtained using the Score A method on the same input data. Tables 1, 2 and 3 gather the obtained results.

Table 1 Training data (sample)
Table 2 Validation data
Table 3 Test data

Table 1 presents training data, Table 2 presents validation data and Table 3 presents test data. The first four columns contain the first four input variables. The fifth column represents the forecast given by the network, using the following coding: 0 for bankruptcy and 1 for non-bankruptcy. The sixth column is the Score A function computed by classical methods, for which a negative value signifies that the firm is in bankruptcy, a value over 2.05 signals a non-bankruptcy state and a value between 0 and 2.05 signifies uncertainty regarding the bankruptcy state. The seventh column is the real situation of the (non-) existence of the bankruptcy and the eighth column is the accuracy of the forecast.

3.2 Optimising the ANN

The network developed in the previous section can be optimised to obtain forecasts of a higher accuracy. The optimisation can be made using two classes of methods: empirical and statistical methods.

According to Slavici (2009), six relevant factors influence the network’s architecture: network topology, learning method, the nonlinearity function used, number of hidden layers, neuron number of each layer and number of training epochs. At this stage, all six factors are considered to be determined a priori. If only four different values are considered for each factor and no additional lines are made for a given set of values, \(4^6 = 4,096\) experiments would be required. Using a statistical method for experiment planning is thus necessary.

A first quasi-empirical experiment plan was elaborated and presented with the preliminary conclusions by Slavici (2009). From the variety of existent neural networks typologies (architectures), the achievement of four topologies has been made using Matlab functions (in other cases, conflicts in network exploitation have occurred). The best results obtained corresponded to a feedforward type of neural network. This topology will be used below for the statistical study of the obtained results.

Tables 4 and 5 present the intermediate results of the dispersion bifactorial analysis performed successively for the two factors (epochs and layers). The Fischer criterion value - one of the strongest tools in appreciating the significance of a variable - is thus established. The computed value is compared with the value in the lookup tables. If the computed value is higher than the corresponding lookup table one, the correspondent variable is considered significant for the analysed objective function.

Table 4 Intermediate results for the first factor (number of epochs)
Table 5 Intermediate results for the second factor (number of layers)

Figure 4 shows the forecast accuracy depending on the number of training epochs, the number of neurons per layer being considered a parameter (10 or 20 neurons per layer). It can be observed that a better accuracy is achieved using 10 neurons per layer than using 20 neurons per layer. The accuracy obtained for 500 training epochs is also slightly higher than that obtained for 200 training epochs.

Fig. 4
figure 4

Forecast accuracy versus number of neurons per layer

Figure 5 illustrates the forecast accuracy as a function of the number of training epochs, with the number of layers considered a parameter (3, 4 or 5 layers). These values are computed based on the data included in Table 6. Table 6 reflects the forecast accuracy obtained for a specified number of training epochs and layers. For each pair of values (s,e), three repetitions of the experiment were performed and the mean values were computed. While a lower number of training epochs (50 or 100) requires 4 layers for better accuracy, 3 or 4 layers are required for better accuracy for a higher number of training epochs (over 200). The best accuracy obtained is for 200 training epochs and 3 layers.

Fig. 5
figure 5

Forecast accuracy versus number of layers

Table 6 Forecast accuracy as a function of e and s

Figures 4 and 5 refer to the discrete dependence of the forecast accuracy on two given sets of data: the number of training epochs and either the number of layers or the number of neurons per layer. These dependences give information about the forecast accuracy value for specific values of the parameters (n,e), respectively (s,e). To obtain more accurate answers about these dependences, two continuous functions describing the evolution of the forecast accuracy on the parameters (n,e), respectively (s,e), are required. The maximum value of these functions is then computed to obtain the values of the parameters for which the forecast accuracy is highest.

Based on the least squares method, two objective functions have been constructed, describing the dependence of the forecast accuracy on its most influencing factors. The first objective function describes the dependence of the forecast accuracy on the number of neurons per layer n and epochs e and it has the following expression:

$$\begin{aligned} \mathrm {P(n,e) = 21.648 + 5.92 \times n + 0.194 \times e - 0.201 \times n^2 + 0.0001 \times n \times e } \end{aligned}$$

The second objective function describes the dependence of the forecast accuracy on the number of layers s and epochs e and it is given as:

$$\begin{aligned} \mathrm {P(s,e) = 43.543 + 11.501 \times s + 0.182 \times e - 1.542 \times s^2 + 0.002 \times s \times e } \end{aligned}$$

Figure 6 shows the forecast accuracy for the first objective function. Good forecast accuracy for the first function (over 97 %) can be achieved using a number of neurons per layer close to 15 and a number of training epochs close to 350. Figure 7 shows the forecast accuracy for the second objective function. In this case, the best accuracy (over 95 %) can be achieved for a number of training epochs close to 350 and the number of layers has little influence on this maximum value. A good forecast accuracy, over 95 %, can be achieved with a neural network having the following structure: close to 350 training epochs, close to 15 neurons per layer and close to 4 layers.

Fig. 6
figure 6

The forecast accuracy of the first objective function

Fig. 7
figure 7

The forecast accuracy of the second objective function

4 Conclusions

Using ANNs to predict company bankruptcy is extremely efficient, as the forecast accuracy is higher than in traditional methods. A data volume higher than the network training set can thus be considered. The data processing has been made using AI methods and not mathematical modelling (with statistical tools) for the following reasons:

  1. (a)

    A mathematical model of the process is either unknown or too complex and is associated with insufficient accuracy; it cannot even be determined in some cases.

  2. (b)

    The available data are incomplete or subject to noise in some cases.

  3. (c)

    There are a number of constraints applied to the process that require being simultaneously optimised.

By studying the dependence of the forecast accuracy on the network topology, an optimised structure with a forecast accuracy over 95 % has been developed. This forecast, obtained by AI methods, is superior to those obtained by traditional methods and does not require statistical tests or research planning with factorial type tests.

The benefits brought by ANN and expert systems compared to the classical economic forecast and decision making methods include: higher forecast accuracy, faster decision making and the possibility of dynamical update of the wide databases constructed. The increasing use of AI methods is thus justified and represents an added value for the financial environment.