Introduction

Soundscapes (Pijanowski et al. 2011), their functional sub-units, sonotopes (Farina 2014, p. 16–17), and acoustic communities (Farina and James 2016), are highly dynamic emergent patterns of the acoustic activity of physical (geophonic), biological (biophonic), and anthropogenic (technophonic) agents. These patterns respond to several environmental variables (Farina 2014), resulting in important indicators of the degree of ecosystem complexity and functioning. In particular, sound is used by many organisms as a carrier of meaning to assure intra- and interspecific communication, to memorize their acoustic umwelt (sensu von Uexküll 1982), and to navigate across their individual-based cognitive landscape (sensu Farina 2010, p. 19).

Ecoacoustics, a recent field of ecology, investigates the ecological role of sounds from individual to landscape scale using a passive noninvasive recording methodology and has been defined by Sueur and Farina (2015) as “a theoretical and applied discipline that studies the sounds emanating from natural systems along a broad range of spatial and temporal scales in order to tackle biodiversity and ecological questions.”

Ecoacoustics offers a broad spectrum of theoretical and applied possibilities for ecological research. For instance, it may detect community diversity and daily dynamics of individual species, populations, and communities (Sueur et al. 2008a; Depraetere et al. 2012; Gasc et al. 2013). It represents an efficient tool to explore the relationship between environmental variables like temperature, humidity, light, and the acoustic activity of species, and it reports the functioning of animal aggregations under different disturbance regimes like climate change, land use change, or landscape degradation (Llusia et al. 2013; Krause and Farina 2016). The acoustic behavior that requires conspicuous energetic investment for individual species (Gillooly and Ophir 2010) is strictly related to the health status of ecosystems and for this it can be used as proxy for landscape assessment, planning and design (Joo et al. 2011; Bormpoudakis et al. 2013; Tucker et al. 2014), and long-term monitoring (Bardeli et al. 2010). In the conservation field, ecoacoustics can help detect the first stress signals of species as a consequence of human effects on ecosystems (Brandes 2008; Monacchi 2014; Harris et al. 2016).

To process the acoustic information according to an ecological perspective, new metrics have been successfully tested and applied in different environmental conditions (Gage et al. 2001; Sueur et al. 2008a; Gage and Axel 2014; Sueur et al. 2014; Towsey et al. 2014; Fuller et al. 2015).

However, the difficulty to assign appropriate meaning of the acoustic information when processed by current ecoacoustic metrics and the prohibitive time required to process big data remain problematic and unresolved topics. Unlike bioacoustics analysis, where the scale of investigation favors individual sounds and where investigators often do not pay attention to the sonic scene around an audio sensor, in ecoacoustic analysis, individual sounds are considered part of a sonic context, and the sources of sound are not analyzed in isolation but as part of emergent acoustic patterns that fulfill specific functions. These patterns, although partially intercepted by the ecoacoustic metrics, often remain difficult to interpret. To find a solution to this important aspect, we introduce the concept of the ecoacoustic event, an ecosemiotic tool to interpret the patterns emerging from sonic context.

Furthermore, in the ecoacoustic investigation, time is expanded, covering seasonal or yearly periods. The exigency to investigate for long periods of time causes a dramatic accumulation of acoustic data, which represents a challenge for researchers that intend to relate acoustic patterns to environmental processes. The processing of such big data requires a great amount of computational time that often forces investigators to adopt a sampling strategy instead of continuous surveys (Winner et al. 2013; Pieretti et al. 2015).

The aim of this paper is to face both the problems of introducing the concept of an ecoacoustic event as an agent of an ecosemiotic process (sensu Maran and Kull 2014) and to describe in detail the Acoustic Complexity Index (ACI) as a tool to detect and identify ecoacoustic events.

The ACI (Farina and Morri 2008; Pieretti et al. 2011), now renamed ACItf, is powered with the new index ACIft (formally presented here for the first time), which includes the metrics proposed to process acoustic files and execute the ecoacoustic event detection and identification procedure (EEDI). At the same time, we propose a dedicated software (SoundscapeMeter 2.0) (Farina and Salutari 2016) to accelerate the computational processes.

Materials

To test and validate ACI metrics and describe the EEDI procedure with empirical cases, we have selected four recording stations located in the northern part of Tuscany (Italy): Carpaneta, Madonna Colli (1), Madonna Colli (2), and Agnino, characterized by sub-Mediterranean shrub lands. More details about the localities and recording sessions are reported in Table 1.

Table 1 Geographic position and altitude of the four localities and number of days of 240 daily recording sessions

Four Soundscape Explorer (terrestrial) SET digital recorders (Lunilettronik, Fivizzano, Italy) were used to collect acoustic information at a sampling frequency of 48 kHz. On-board ACI metrics, applied in real time, were set with a background filter of 8 amplitude and a clumping of 1 s. The fast Fourier transform (FFT) window was fixed at 1024 points, obtaining 512 frequency bins. Acoustic files were collected from the 4th to the 16th of October 2015, for one minute during every five minutes (240 files a day), with a total of 8640 files processed (all localities included). An additional day from the Agnino location (December 21, 2015) has been used to present a further example of the EEDI procedure.

The beta release of the Soundscape Meter 2.0 (Farina and Salutari 2016) after the incorporation of the EEDI model has been applied to process the data.

The Ecoacoustic Events: Definition

Ecoacoustic events are emergent sonic patterns that are recognized by individual species during the completion of a specific function. Ecoacoustic events are incorporated into the eco-field theory (Farina and Belgrano 2004; Farina and Belgrano 2006). Their detection is a common practice in different fields of theoretical and applied acoustics (e.g., Mesaros et al. 2010; Zhuang et al. 2010; Heittola et al. 2011), although difficulties remain to complete the event identification due to the polyphonic nature of the signals present.

Definitively, categorizing acoustic events is not a simple anatomy of sounds or a descriptive exercise, but assuming that events have meaning not only for humans but also for other organisms, the detection and identification of functional units has become central in ecoacoustic analysis as well as a true challenge in ecosemiotic research. According to this perspective, acoustic events are not simply a description of some distinct characteristics that can be isolated and categorized from an indistinct background (Ma et al. 2006) but are functional agents actively used by organisms to fulfill specific functions. Therefore, in our narrative, we consider such events to be delimited by ecological roles and name them ecoacoustic events.

This vision is coincident with the ecosemiotic model of the (acoustic) eco-field (Farina and Belgrano 2004; Farina and Belgrano 2006), which has been effectively described by formally accepting the triadic mechanism of the Peirce semiosis. This model considers a sound collection (summation of individual sounds) as a spatial configuration carrier of meaning, the representamen that is used by a representant (the function, sensu Farina and Belgrano 2004; Farina and Belgrano 2006) to recognize the object that is requested to locate resources (e.g., territory, nesting site, roosting, presence of predators, and location of food resources) (Farina 2008).

According to this perspective, ecoacoustic events assume a central role inside the ecoacoustic narrative and justify the efforts to open a specific field of investigation inside ecoacoustics, providing a robust quantitative tool for developing ecosemiotic studies. Ecoacoustic events are scale dependent; they exist along a broad range of temporal scales. For instance, Ma et al. (2006) argued that a background noise is simply the summation of events that happen at finer temporal granularities.

Ecoacoustic events may be represented by broad categories of sounds, such as thunders, blasts, rifle shots, horns, different typologies of wind and rain, alarm calls of birds, morning and dusk animal choruses, silence, human voices, road traffic, or their combination, that result in some meaning for a listener and that produce an interruption or an invasion of the soundscape. The basic difference between the concept of events in social sciences and in ecological sciences consists of the typology of the sounds that are considered events, although some cases (e.g., the human voice), at least formally, are coincident to both epistemologies. The category to which we can assign an event depends on the goal of the classification. Consequently, the categories are variable and can change accordingly. In the ecological analysis of sounds, detection and identification of ecoacoustic events represent an important step forward.

The Acoustic Complexity Index

The ACI metrics, like other metrics operating in ecoacoustics, are based on the assumption that there is a strict relationship between the complexity of animal assemblages in the landscape or inside a community and the spectral and temporal complexity of a soundscape or, at a more detailed spatial scale, of an acoustic community. This means that more individuals and/or more species are present in a location and the acoustic information expressed by ACI metrics is greater (Farina et al. 2011).

The ACI algorithm was written in the 2008 by one of us (FA) and applied by Davide Morri a graduate student of Urbino University (Institute of Biomatematics) in a PhD project (Morri 2008; Farina and Morri 2008), later improved and tested by Pieretti et al. (2011) and successively applied both in terrestrial habitats (Farina et al. 2011; Farina et al. 2013; Pieretti and Farina 2013; Lozano et al. 2014; Bobryk et al. 2015; Duarte et al. 2015; Farina 2015) and in marine system (McWilliam and Hawkins 2013; Staaterman et al. 2013; Kaplan et al. 2015). Recently has been also applied in neurophysiological responses to musical stimuli (Jancke et al. 2015).

The ACI has been used to process data from different ecoregions: Madagascar (Farina in prep.), Brazil (Duarte et al. 2015; Pieretti et al. 2015), Mediterranean Europe (Farina et al. 2011; Farina et al. 2013), California (Krause and Farina 2016), Temperate Neartic (Bobryk et al. 2015), and Australia (Towsey et al. 2014).

The ACI has been included in WaveSurfer software (Sjolander and Beskow 2000) as SoundscapeMeter 1.0 plug-in (Farina et al. 2012) and later the ACI algorithm was included in the SeeWave software (Sueur et al. 2008b). More recently, it has been used for automated processing inside Soundscape Explorer [Terrestrial] (SET) (Lunilettronik), a new device that belongs to the category of low cost recorders (sensu Farina et al. 2014), that automatically analyzes soundscapes and acoustic communities powered by SoundscapeMeter 2.0 software (Farina and Salutari 2016).

The ACI that is applied without a logarithm conversion to the numerical expression of a FFT conversion that is utilized when the FFT output is expressed in decibels is a measure of the variation of the amplitude between the following:

  1. a).

    a couple of pulses along a specific frequency bin (ACI, here named ACItf) and

  2. b).

    a couple of pulses across frequencies along the same temporal interval (ACIft) (Fig. 1a).

Fig. 1
figure 1

a Numerical matrix as FFT output: x-axis frequency bins, y-axis time; b Confrontation between pulses along a frequency bin (ACItf); C) Clumping aggregation categories to calculate ACItf

Both measures are computed simultaneously along a temporal interval chosen according to the requirements of the research (e.g., 1 min, 5 min, and so on). The size of the frequency bin depends on the sampling frequency and the adopted FFT window.

The ACItf is considered a reasonable proxy to assess the acoustic complexity/information embedded in an acoustic file (Pieretti et al. 2011) and ACIft should a further important companion to describe the acoustic complexity (Fig. 1b). Definitively, when an acoustic sequence of pulses has low differences in amplitude, ACItf and ACIft are expected to be low (small information), reaching zero when adjacent pulses have the same amplitude.

To be applied in a more appropriate way (e.g., searching for patterns), ACItf requires an aggregation of data (clumping). In fact, in the analysis of complex patterns of biological or ecological origin, often it is necessary to operate at a higher scale of temporal resolution (O’Neill et al. 1986). More recurrent amplitudes will dominate in higher clumping. On the other hand, more singularities will emerge with smaller clumping. The temporal scaling of the clumping (Fig. 1c) consists of the repartition of an entire temporal sequence of pulses (e.g., 1 min), in temporal segments of equal length (e.g., 1 s). The ACItf values are different according to the clumping size; ACItf decreases with the increase of the clumping size (higher clump size and more data in the denominator lower the value of the ACItf). The complete (explicit) equation for ACItf for each frequency bin f, where the clumping option is considered, can be represented as follows:

$$ {\mathrm{ACIt}}_{\mathrm{f}}={\displaystyle \sum_{\mathrm{k}=1}^{\mathrm{c}}\left(\frac{{\displaystyle \sum_{\mathrm{i}=1}^{\frac{\mathrm{t}}{\mathrm{c}}}\left|{\mathrm{a}}_{\mathrm{i},\mathrm{j}}-{\mathrm{a}}_{\mathrm{i}+1,\mathrm{j}}\right|}}{{\displaystyle \sum_{\mathrm{i}-1}^{\frac{\mathrm{t}}{\mathrm{c}}}{\mathrm{a}}_{\mathrm{i},\mathrm{j}}}}\right)} $$

where a i,j is the FFT numerical output, t is the number of temporal steps in which a file is subdivided after FFT, f is the frequency bin, c is the number of clumps in the recording, and t/c is the number of elements composing a clump.

The clumping has nonlinear behavior along the frequency spectrum because it is affected by the structure of the acoustic matrix. When the acoustic matrix has more zero values along the frequency bins, the variation of ACItf due to a different clumping size is more modest (see Fig. 2a in which four clumping categories were confronted for the same spectrogram). In a real spectrogram, low frequencies, at least below 1 kHz, result in a more continuous distribution of pulses ≠ 0 than the higher frequencies, which may result in several gaps due to a higher number of zero values inside a complex spectrum like the one created by biophonies. Consequently, the ACItf value changes more at low frequencies for the different clumping sizes than at higher frequencies, where the clumping size is biased (buffered) by zero values that reduce the numerator and denominator as well.

Fig. 2
figure 2

The spectrogram obtained by one minute of recording (Blackcap (Sylvia atricapilla) song) at Carpaneta location on 0619 am of 12 April 2016 is used to illustrate the clumpling procedure and the filter effect. a The increase of clumping aggregation reduces the value of ACItf. This effect is particularly evident for the first frequencies that usually saturate the spectrogram. b The application of a filter during the ACIt process allows the elimination of every masking effect due to signal noise. The filter threshold must be selected empirically and is largely due to the quality of microphone capsules

The clumping operation is not mandatory for ACI metrics, but in our experience, it has demonstrated a more realistic approximation of the acoustic information. The comparison between two ACItf calculated with a different clumping value may introduce important biases. To compare two ACItf with a different temporal interval (e.g., 1 min versus 2 min), ACItf must be averaged resulting in ACItf/t.

ACIft a New Component of the ACI Metric

The ACIft measures the information that occurs between two successive frequency bins inside the same temporal step t, such that:

$$ \mathrm{ACIft}={\displaystyle \sum_{\mathrm{i}=1}^{\mathrm{t}}{\displaystyle \sum_{\mathrm{j}=1}^{\mathrm{f}-1}\frac{\left|{\mathrm{a}}_{\mathrm{i},\mathrm{j}}-{\mathrm{a}}_{\mathrm{i},\mathrm{j}+1}\right|}{\left({\mathrm{a}}_{\mathrm{i},\mathrm{j}}+{\mathrm{a}}_{\mathrm{i},\mathrm{j}+1}\right)}}} $$

where a i,j is the amplitude of each pulse, t is the number of temporal intervals, and f is the frequency bin. The ACIft varies according to the temporal steps. The length of the temporal interval may be fixed equal to the clumping dimension of ACItf; however, this option is not mandatory. To compare two ACIft with a different temporal interval (e.g., 1 min versus 2 min), the ACIft must be averaged, resulting in ACIft/t.

The ACIft is the measure of information present inside an interval of time empirically chosen according to specific purposes. In our experience, 1 min is a good temporal interval to evaluate the instantaneous complexity of a terrestrial acoustic community of birds in a Mediterranean region (see Pieretti et al. 2015 for further discussion). In one minute, operating at a frequency of sampling of 48 kHz and applying a FFT window of 1024 points x 512 frequency bins (46.875 Hz each), there are 2812 elements (lasting 0.021337 s each) on which to make 512 confrontations of the amplitude from which ACIft is extracted.

In every spectrogram, due to the low quality of the built-in microphone capsules used in commercial field digital recorders, the errors introduced in analog/digital conversion, and the electronic or electrical noise of circuits, the signal-to-noise ratio (SNR) may be quite low. To work appropriately, both metrics (ACItf, ACIft) require the application of a filter to exclude pulses below an empirically-fixed amplitude threshold attributed to non-environmental sounds. Such a filter is applied to the entire spectrogram, considering every pulse with an amplitude lower than a specific threshold to be zero. This amplitude threshold (represented by the real component of the FFT) corresponds to the mVolt2/Hz of the analogic device and is evaluated empirically by an experienced operator. Figures 2a, b illustrate an example of the application of ACI to wave files with and without the filter.

After the application of an amplitude filter to eliminate the background noise present in the FFT matrix, zero values will appear beside the zero values due to the absence of pulses in the unmodified file. Based on the difference between the two elements of the spectral matrix in the cases in which there is a pulse distribution such as … 0, 5, 0, 9, 0 … (where zero may be the absence of a signal or the presence of a value below the threshold) every difference from zero produces a ratio equal to one (e.g., |(0–5)|/(0 + 5) = 1, |(0–9)|/(0 + 9) = 1) that could be expressed as the max novelty, but this is extraneous to our goal, which is focused on the assessment of the difference of amplitude inside a signal structure and not between this structure and its background. To eliminate this possible unwanted information, the model automatically explores the FFT numerical matrix; it detects these cases in which the comparison between two pulses of which one has a value of zero and excludes them from the numerator of the algorithm. This means that in a sequence like this: … 2, 5, 0, 4, 0, 9, 0, 6, 8, 0… (for simplicity, we have used the value of a hypothetical amplitude from zero to nine without overlap between two successive clumps), after a clumping fixed equal to five, for the number of data to be considered in each of the two resulting clumps (… [2, 5, 0, 4, 0] [9, 0, 6, 8, 0]…), the algorithm will be ACItf = [|2–5|/(2 + 5 + 4)] + [|6–8|/(9 + 6 + 8)], and at the numerator [4] (in the first clump) and [9] (in the second clump) values have not been considered at all because they have no value to be confronted on at least one of their sides.

ACItf versus ACIft

The ACItf and ACIft are weakly correlated (Pearson’s r) when sorted according to their values, and the correlation becomes dramatically weak with the decrease of the ACItf or of the ACIft values, for instance in n = 8392 cases of our data set, sorted by ACItf (first quartile r = .91, p < 0.05; second quartile r = .12, p < 0.05; third quartile r = .02, ns; fourth quartile r = −.03, ns). This may be interpreted by a similar amount of information that ACItf and ACIft have in presence of saturated spectra (due to geophonies or technophonies), but these two indices diverge when the spectrum is less dense. This generally happens when there are biophonic signals that usually have different patterns according to time or frequency orientation. This confirms that the two indices extract different aspects of acoustic information. In this way, ACItf and ACIft are used with their specific explicit attributes (e.g., ACItf: distribution of frequencies to build the acoustic signature and ACIft: distribution of acoustic information along time).

The Evenness of ACI Metrics

Evenness, equidistribution, or equitability is a measure of the repartition of the number of animals, resources, or energy categories in a collection (Lloyd and Ghelardi 1964; Hill 1973; Whittaker 1975). Several metrics have calculated this index, which has an important application in population and community ecology (e.g., Levins 1968; Hill 1973; Heip 1974; Hurlbert 1978; Peet 1974; Alatalo 1981; Molinari 1989; Bulla 1994; Hill 1997). The application of the evenness metric to the ACI distribution allows evaluation of how frequency classes are distributed along a frequency set (ACItf) and how the acoustic information is distributed along a specific time lag (ACIft). The level of evenness of ACItf and ACIft has been calculated using the Levins evenness B (Levins 1968; Hurlbert 1978):

$$ \mathrm{B}=\frac{1}{{{\displaystyle \sum_{\mathrm{i}-1}^{\mathrm{t}}{\mathrm{P}}_{\mathrm{i}}^2}}_{\left|\right.}} $$

where p i is the importance of ACI in each frequency bin f (ACItf) or in each temporal Step t (ACIft). The standardized measure is

$$ {ACI}_{\mathrm{e}}=\frac{\mathrm{B}-1}{\mathrm{t}-1} $$

This measure ranges from zero to one. When information is equally distributed in all the classes of frequencies according to the FFT setting in which ACItfe max is 1/512 (if the FFT window size is fixed at 1024), ACItfe is one.

The ACIfte is low when only a few acoustic events are present along the temporal step considered (e.g., 1 min). The ACIfte is equal to one when ACIft is equally distributed in all the temporal steps. For instance, operating at a sampling frequency of 48 kHz for 1 min with an FFT window of 1024 points, we obtain 2812 temporal steps. In this case, the ACIfte max is 1/2812.

The ACItf is well correlated with ACItfe (Pearson’s r = 0.90, p < 0.05, n = 7920), but ACIft and ACIfte show weak dependence (Pearson’s r = 0.25, p < 0.05, n = 7920). These last two metrics seem good candidates to be used for the detection of the acoustic events, as better explained in the next section.

Ecoacoustic Event Detection and Identification (EEDI)

The philosophy of the EEDI is based on the possibility of restricting the analysis of the acoustics files, excluding all the conditions about which we have no interest to investigate, reducing the computational effort by selecting the threshold to apply environmental variables, such as temperature, light, humidity, or daily time to ACI metrics.

The EEDI model is based on three steps. The first step selects the environmental variables (time, temperature, light, or humidity) according to an empirical threshold. The second step detects an event after the confrontation between ACIft and ACIfte according an empirical threshold, and the third step identifies an event after the computation of the level of correlation/similarity between the detected event(s) and the acoustic signatures of the library (Fig. 3). The use of thresholds is necessary to filter out conditions from which it is difficult to extract meaningful information or simply when we decide to consider or exclude a specific category of events, such as geophonies (e.g., wind or rain) or technophonies (e.g., road traffic or airplanes).

Fig. 3
figure 3

The EEDI comprises three steps: a Recording and on-board FFT and ACI processing, b Ecoacoustic event detection, and c Ecoacoustic event identification

After the selection of files associated with certain environmental conditions, such as temperature and light, the plot of ACIft x ACIfte creates an ecoacoustic event space (EES). This space may be divided into four nominal quadrants or regions, where each quadrant is characterized by a different distribution of ACIft and ACIfte. To simplify, the quadrants have been indicated according to a geographical nomenclature: first quadrant (NW), second quadrant (NE), third quadrant (SE), and fourth quadrant (SW) (Table 2, Fig. 4).

Table 2 Tentative distribution in the acoustic event space (EES) of the major categories of acoustic events
Fig. 4
figure 4

Example of an EEDI applied on October 4, 2015 (Agnino, 44° 14′ 12.32″ N, 10° 0416.17″ E). The threshold of ACIft, fixed at >10, ACIfte < 0.7, has detected 127 cases on 240 when these conditions are respected separately. If the two thresholds are respected at the same time, only four cases are found. Three cases selected with the criterion to evaluate the two extremes (Heavy rain, robin alarm call) and an event in the middle (Robin call & light rain) are reported. At the side of each spectrogram the acoustic signature of each event is reported

The first quadrant (NW) is characterized by low ACIft and high ACIfte (e.g., biophonies at low amplitude but at high evenness, such as a cricket song). The second quadrant (NE) is characterized by high ACIft and high ACIfte (e.g., geophonies, such as strong wind and heavy rain that are distributed regularly along the time axis). The third quadrant (SE) is characterized by high ACIft and low ACIfte (e.g., short technophonies at high amplitude, such as a jet at low altitude, or a geophony, such as thunder). The fourth quadrant (SW) is characterized by low ACIft and low ACIfte (e.g., biophonies, such as isolated alarm calls). Each combination of ACIft and ACIfte potentially corresponds to a real case of an ecoacoustic event.

After the events have been detected according to an established threshold (in this case: ACIft >10 or ACIfte < 0.7), they must be identified. This further step requires the use of suitable libraries of known events. Such libraries are created from aural identifications of a specific event, and the acoustic signature (ACItf) of the event is compared with the acoustic signature of every detected event. The Pearson correlation, chord distance (Orloci 1967), and Whittaker’s index of association (Whittaker 1952) are some of the possible metrics to test the relationship between the detected events and the library.

Figure 4 reports an example from the Agnino location collected on October 4, 2015. In this case, three events have been selected by way of example. The first corresponds to heavy rain, and a second event corresponds to the European robin (Erithacus rubecula) call in the presence of light rain, while a third event is characterized by the European robin alarm calls without background noise. The acoustic signature shows very well the different distributions of ACItf along the frequencies.

The identification of the selected patterns is made by comparing different groups of detected events with a library of selected events. In Fig. 5, examples show an aggregation of events of the NE and SW quadrants, which are identified as rain and the European robin call.

Fig. 5
figure 5

From the example of Fig. 4, the acoustic signature of a group of events ACIfte values (rain gradient) a and with similar ACIft (robin calls) b and the value of the r correlation inside each group are reported

A second example from the same location is reported in Fig. 6. In this last case, few events are discriminated in the EES of which the two extremes are moderate rain (NE quadrant) and an isolated alarm call of the European blackbird (Turdus merula) (SW quadrant), respectively. The value of ACIft is a quarter of the ACIft found in October in the same location, and this is due to the reduction of biophonic activity of birds and less rain in December 2015.

Fig. 6
figure 6

Example of an EEDI applied on December 21, 2015 (Agnino, 44° 14′ 12.32″ N, 10° 0416.17″ E). The threshold of ACIft was fixed at >10, ACIfte < 0.7, these conditions are respected separately. This day is characterized by low bird activity and light rain. In a moderate rain event) and in b European blackbird (Turdus merula) alarm call event. On their side the acoustic signature (ACItf)

We have applied the same conditions (ACIft >10 or ACIfte < 0.7) to the 36 days of total recording (all locations) obtaining 3000 (34 %) events in total of the possible 8640 events. The 36 EES obtained using these criteria are reported in the supplementary material for the four locations (Fig. 1 suppl.). From these data, it emerges that the majority of biophonies in this period of the year are confined below a value of ACIft of 10. The highest value of ACIft (100) has been obtained only during heavy rains. Rain or wind events range from 10 to 100 ACIft according to intensity, but they are discriminated by a value of ACIfte close its max, which is one.

Discussion and Conclusions

The acoustic environment is rich in underexploited sources of information (Ma et al. 2006), and EEDI offers a new opportunity to model the acoustic information decoded by ACIft and ACItf metrics with the use of environmental parameters that are automatically associated with the acoustic data (e.g., light intensity, temperature, humidity, atmospheric pressure, and time of day). In this way, we can select the files according to one variable (e.g., light versus dark or cold versus warm) or according to a combination (e.g., warm morning, cold night, etc.) (Table 3).

Table 3 Possible applications of the EEDI procedure to ecological research, nature conservation, climate change, and human intrusion in natural environments. Small size, optional variables to put in the EEDI model

After this selection, we can select ACIft and ACIfte thresholds. The threshold applied to ACIft may distinguish an EES, occupied by low or high information. The threshold applied to ACIfte may distinguish how ACIft information is aggregated or evenly distributed along time (in this study, for one minute). Finally, the information expressed by ACItf allows the production of the acoustic signature of a detected event.

The analysis can be conducted at the soundscape level or at the level of acoustic communities or individual species. At the soundscape level, we can detect geophonies, biophonies, technophonies, or their combinations. In this way, the analysis allows distinguishing the different typologies of sonotopes present around the recording site (Farina 2014, p. 17–18), although it remains problematic to produce good libraries of these components of a soundscape. At the level of acoustic community, the model may consider special events, such as the dawn or dusk chorus. To have a more precise idea of how active an acoustic community is, we can simply measure the number of events that the EEDI model detects as biophonies. At the level of single species, it is possible to investigate the daily dynamics or the presence of an isolated sound as in the example reported in Fig. 6.

The EEDI model enables the analysis of long-term monitoring data regarding biophonies produced by vocal animals (fishes, frogs, birds, mammals, and arthropods), but the possible applications of EEDI are several, and they are largely dependent on the target of the research. For instance, the EEDI model can detect the amount of ice calving and ice melting in the polar seas.

In the examples illustrated here, the application of a model, where ACIft ranges from 10 to 100 and where ACIfte ranges from .7 to 1, makes it possible to find rainy or windy events that could be excluded in the case of a bioacoustics target. To search for acoustically rare species, we must select files with extremely low ACIfte associated with a very low ACIft.

If we want to discover how many civil aircrafts fly over the sampled area, we can select events characterized by low ACIft (usually airplanes far from an airport have a low amplitude sound and a high ACIfte because the sound is constant along the interval of time considered, e.g., one minute). The EEDI model can be used to test the efficiency of soundscape categorization (classification) in different landscapes, exploring the event diversity as a proxy of the complexity of the land mosaic used by organisms and people.

The choice of the ACIft and ACIfte thresholds depends on the geographic areas, the habitats, and the goals. The EEDI model integrates the information of the ACI metrics, moving from the crude data of ACItf or ACIft to the identification of the possible combinations with which these metrics can express meaningful information, and acts as a new and innovative tool of ecosemiotic investigation.

The model is simple, can be easily applied using a free open-access routine (SoundscapeMeter 2.0, Farina and Salutari 2016), and is particularly adept at analyzing a large amount of acoustic data. In fact, the model can concentrate on focal events, economizing work energy and processing time, excluding all the conditions that we have no interest to investigate. In the future, the entire procedure could be implemented by incorporating other ecoacoustic indices, like the evenness of ACItf, allowing a more discriminant analysis.