Introduction

An in-depth understanding of volcanic activity requires a combination of modelling and field observations. Models make it possible to determine natural properties that are not accessible for direct measurement and can be used to predict the future evolution of a natural system. To do this, field observations are used in parallel with models, to ensure that the models correctly reproduce observable parts of the natural phenomena, thus giving confidence that they can also be used for the unobservable parts and for predictions (for modelling ballistic fields see, for example Waythomas and Mastin 2020). For Strombolian and Vulcanian activity, bomb velocities and directions are essential parameters for calculating the distance affected by bomb impacts (Mastin 2001), and for the understanding of the superficial activity of the magmatic system and the nature of the interactions between gas and conduit magma to generate the ensuing explosion (Wilson 1980; Mastin 1995).

Velocity estimations for volcanic bombs have been previously performed using image analysis (e.g. Chouet et al. 1974; Formenti et al. 2003; Zanon et al. 2009; Vanderkluysen et al. 2012; Taddeucci et al. 2015). Manual tracking methods, suitable for a limited number of bombs, have been recently improved by automated tracking to capture a large number of bombs (Bombrun et al. 2014; Gaudin et al. 2014). However, analyses have usually been 2D, meaning that the velocities can only be estimated correctly if the bombs move solely in a plane, vertical or perpendicular to the line of sight of the camera. Volcanic ejecta generally move in diverse directions and 2D approaches can thus induce error in the estimation of bomb absolute velocities. Moreover, 2D methods cannot estimate the direction of motion of the ejecta.

More recently, 3D calculations have been performed by focusing on selected bombs using stereoscopic reconstructions derived from two synchronised cameras (Gaudin et al. 2016). The main limitations of their method were that manual tracking of pyroclasts limited the number of bombs analysed. As a further step, we present here a method for the automated recognition of hundreds of bombs and for the reconstruction of their trajectories in time and space by stereophotogrammetry.

During the course of an eruption, many thousands of bombs can be ejected (Bombrun et al. 2014). The more of these that can be tracked, the better the statistical analysis of the trajectory variations, and the greater the insight into eruptive dynamics. Tracking and identifying the bombs manually is complex, since they are launched at a range of times, angles and velocities so that at any one time thousands of different trajectories will be active. This means so bombs will be ascending while others are descending with intersecting paths, and frequently masking each other. Bombs are present over a large volume of space rather than on a planar surface, making the identification of any given bomb within images acquired from different camera positions even more difficult. To facilitate trajectory identification and 3D reconstruction, we have developed a new methodology using stereoscopic calculation based on the automated recognition of bombs. Three-dimensions reconstruction of topography has made substantial progress in recent years and has been used to follow the dynamics of natural phenomena such as glaciers, landslides and lavas flows (e.g. Eiken and Sund 2012; James and Robson 2014; James et al. 2014; Eltner et al. 2017). Bundle adjustment (Granshaw 1980) and recent developments that include structure-from-motion (SfM) and multi-view stereo (MVS) algorithms has widened the use of the techniques and facilitates automatic estimation of camera positions, orientations and lens characteristics. Difficulties associated with reconstructing volcanic bomb trajectories come from (1) the bomb locations that cover a large volume of space and are not restricted to a known surface, (2) the high bomb velocities that require synchronised cameras and prevent the use of a single moving cameras and (3) the high number of bombs with similar size, shape and luminosity making bomb recognition challenging in multiple images. This explains why the bundle adjustment method has not been used for this study. Instead, we here use a custom-built 3D tracking routine.

The target chosen for the study is Stromboli volcano (Aeolian Island, Italy) because of its permanent activity characterised by discrete bursts every 5–20 min (Chouet et al. 1974; Patrick et al. 2007), relatively easy access to the summit and the complexity of the trajectory reconstruction related to the high number of bombs. Moreover, despite the activity of Stromboli being widely studied (e.g. Ripepe et al. 2001, 2002; Rosi et al. 2006; Harris and Ripepe 2007; Pistolesi et al. 2011; Gurioli et al. 2014), there remain open questions regarding conduit ascent dynamics and explosion mechanisms. The most frequent activity consists of repeated explosive events related to slow magma ascent in the upper conduit and/or ascent of bubbles in a low viscosity magma (Wilson 1980). Several models have been proposed based on experimental and geophysical data but a number important characteristics remain poorly constrained, such as the conduit geometry, the depth at which the explosion is triggered and the origin of the different activity types observed at the vents (e.g. Chouet et al. 1997, 2008b; Harris et al. 2008; Gurioli et al. 2014; Gaudin et al. 2017). A multidisciplinary field campaign took place during September and October 2012 to increase understanding of the normal explosive activity at Stromboli, as well as to compare the results obtained using a large set of observational techniques (Harris et al. 2013; Bombrun et al. 2015; Chevalier and Donnadieu 2015). Among the spectrum of techniques used, a system of stereoscopic cameras forming the basis of the new 3D reconstruction methodology described in this article was deployed to improve the identification of ejecta trajectories in space and time.

Methodology

System characteristics

The system comprises two digital cameras recording greyscale images from visible wavelengths (Basler 1300-30, lens Edmund Optics 8 mm F1.4). These acquire 1280- by 960-pixel images, 8 bytes per pixel, at frame rates of up to 30 frames per second, so that data acquisition rates were around 37 Mb/s for each camera. They were connected by Ethernet to the same computer to ensure that the acquisition time reference was identical for both cameras, an essential requirement for the spatial reconstruction processing stage. The positions of the two cameras were located with centimetre-accuracy using differential GPS. One camera was set to film the vent continuously and the images were analysed in real time to detect eruption occurrence (using the software Genika Trigger by Airylab). Recording was made at night to facilitate the bomb recognition: bombs appear as white on a black background. When the eruption is detected, the observing camera triggered a recording of the images from both cameras onto the computer over a period of 30 s. The triggering analysis is sufficiently fast (< 20 ms) that data can be acquired from the onset periods of eruptions without an image buffer. If bombs continued to be detected, the acquisition was automatically extended by successive periods of 30 s.

The principles of a 3D reconstruction

The location of an object projection on the sensor array (i.e. the charge-coupled device, CCD) within a camera depends on the position of the object in space, the camera position, the camera orientation and lens characteristics. To calculate the 3D-position of the object in real-space, we thus invert this imaging model. To do this, we first need to locate the object on two images taken at the same time by the two cameras. Using a correction for the lens distortion and misalignment, we can then calculate the undistorted position of each projection (p1 and p2 in Fig. 1). From the orientations of the cameras and their positions, we next calculate two lines (L1 and L2 in Fig. 1) that pass through each point and the optical centre of the lens (O1 and O2 in Fig. 1). The 3D-position of the object corresponds to the intersection of the two lines. Identifying the object on a sequence of images taken during an eruption allows a reconstruction of the trajectory to be made, i.e. in both 3D space and time.

Fig. 1
figure 1

Principle of position calculation by stereophotogrammetry. The 3D position of an object is calculated by the intersection of the lines L1 and L2 which pass by the object projections p1 and p2 and the optical centres of the lenses

Stereo-system geometry

At a given distance from the vent, the distance between the cameras (i.e. the baseline) is chosen based on a compromise between bomb identification, practical aspects and precision of the 3D reconstruction. The greater the baseline, the larger the angle camera 1–bomb–camera 2 (angle θ in Fig. 1). If θ is very small, errors in camera orientation or subpixel uncertainty in the object detection can induce large errors in the calculation of the 3D position. The system was installed around 430 m east of Stromboli’s NE crater (Fig. 2) chosen for the study. In this configuration, with an angle θ of 2°, an error of 1 pixel induces an error in the bomb position of up to 6 m. The most accurate calculation would be obtained for θ = 90° (an error of 1 pixel induces a 20-cm error in the bomb position). However, because an eruption can emit thousands of bombs per second, a strong difference in the view angles (i.e. large values of θ) gives very different images on the two cameras. This makes bombs impossible to recognise visually and makes some steps of our bomb identification method more complex (e.g. the auto-calibration of camera orientations—see next section—is simpler if supervised). Thus, the smaller the distance between the two cameras, the easier it is to match the bombs. Given the above considerations, we estimated the ideal angle θ of about 10°. For the observation distance of about 430 m and for installation convenience, we chose a distance of 81.6 m that corresponds to an angle between the positions of the cameras and of the crater of about 11°. The system is convergent (Fig. 1) and the angle between the optical axes of the cameras (i.e. their lines of sight) is 9°. In this configuration, an error of 1 pixel in the bomb location on the image induced an error of about 1 m in the East-West direction (called x in the following) and less than 30 cm in North-South (y-direction) and in elevation (z-direction). The error is higher in the East-West direction because of the positions and the orientations of the cameras (to the West).

Fig. 2
figure 2

Location of the NE crater, of the two cameras used for observation of ejecta trajectories and of the photograph of Fig. 10, Stomboli, Italy, 2012. Lidar-derived DEM courtesy of INGV-Pisa (Favalli et al. 2009)

Due to the distance chosen to observe the crater and the camera characteristics, each pixel in an individual image (also called the ground sample distance) corresponds to an area about 20 cm in width and height at the crater location. This, in principle, determines the minimum detectable bomb size but hot bombs much smaller than the pixel size could still illuminate the pixel and hence be detected at night.

Lens distortions

In order to calculate bomb positions accurately, the image distortions induced by the camera lenses need to be corrected. The corrections are done with the Brown–Conrady equation (Brown 1966):

$$ \left\{\begin{array}{l}{x}_{\mathrm{cor}}={x}_n\left(1+{k}_1\times r+{k}_2\times {r}^2+{k}_3\times {r}^4\right)+2\times {q}_1\times {x}_n\times {y}_n+{q}_2\times \left({r}^2+2\times {x_n}^2\right)\\ {}{y}_{\mathrm{cor}}={y}_n\left(1+{k}_1\times r+{k}_2\times {r}^2+{k}_3\times {r}^4\right)+2\times {q}_2\times {x}_n\times {y}_n+{q}_1\times \left({r}^2+2\times {y_n}^2\right)\end{array}\right. $$
(1)

where xcor and ycor is the corrected position, xn and yn is the uncorrected normalised position calculated by xn = (xp − xc)/f and yn = (yp − yc)/f. The variables xp and yp are the uncorrected image position in pixels, f is the calibrated focal length (Fig. 1; Granshaw 2016), \( r=\sqrt{x_n^2+{y}_n^2} \) is the distance between the position to be corrected and the position (xc, yc) of the principal point (i.e. the intersection between the optical axis and the CCD, c1 and c2 Fig. 1) and q1, q2, k1, k2 and k3 are the distortion correction parameters. Note that the principal point is not generally located at the geometric centre of the image due to lens distortion and lens misalignment. The distortion parameters, the position of the principal point and the calibrated focal length were calculated using adapted grids and protocols of the Institut Pascal based on the work of Lavest et al. (1999). The calibration was done in laboratory with a distance of the calibration target of 3 m and the same focus as used in the field. Lens distortions caused differences of up to 30 pixels at the edges of the images between the corrected and uncorrected positions. The standard deviation of the correction has been calculated by comparing the parameters of Eq. 1 to parameters obtained with other sets of images. It is of less than 0.3 pixel. The codes used for the reconstruction, including that for lens distortions, are available in Online Resource 1.

Orientation of the cameras

We used three angles to define our cameras (Granshaw 2016): φ is the pitch, i.e. the angle between the horizontal and the optical axis of the camera, where φ > 0 means that the camera is pointing upwards (Fig. 1). The angle κ defines the azimuth, i.e. the orientation of the optical axis in the horizontal plan, where κ = 0means that the camera is oriented to the North, and κ > 0 is rotated to the east. The angle ω defines the roll, i.e. the rotation around the optical axis, with ω > 0 for a counter-clockwise tilt of the camera.

If φ = 0, κ = 0 and ω = 0, the orientation of the camera is defined by \( {X}_{\mathrm{i}\mathrm{ni}}=\left(\begin{array}{c}{u}_{\mathrm{i}}\\ {}{v}_i\\ {}{w}_i\end{array}\right)=\left(\begin{array}{ccc}1& 0& 0\\ {}0& 0& -1\\ {}0& 1& 0\end{array}\right) \), which conveys that the camera is horizontal and pointing northward, ui and vi defining respectively the orientation of the horizontal and the vertical for the images (CCD sensor is vertical if the camera is horizontal, Fig. 1) and wi defining the orientation of the optical axis of the camera. The rotation matrices are

$$ {R}_{\varphi }=\left(\begin{array}{ccc}1& 0& 0\\ {}0& \cos \varphi & -\sin \varphi \\ {}0& \sin \varphi & \cos \varphi \end{array}\right),{R}_{\kappa }=\left(\begin{array}{ccc}\cos \kappa & 0& \sin \kappa \\ {}0& 1& 0\\ {}-\sin \kappa & 0& \cos \kappa \end{array}\right)\ \mathrm{and}\ {R}_{\omega }=\left(\begin{array}{ccc}\cos \omega & -\sin \omega & 0\\ {}\sin \omega & \cos \omega & 0\\ {}0& 0& 1\end{array}\right) $$
(2)

Using these matrices and these references, the orientation of the camera can be calculated by

$$ X={R}_{\omega}\times {R}_{\varphi}\times {R}_{\kappa}\times {X}_{\mathrm{ini}} $$
(3)

Calculating the camera orientation on an active volcano can be challenging. The common method consists of placing ground control points (GCP), with accurately measured positions, in the camera’s field of view (Diefenbach et al. 2012). Using the images of the GCP, in combination with the lens characteristics, allows the camera orientations to be calculated. However, placing GCPs on active volcanoes is often impossible for security and accessibility reasons. To solve this problem, we used a two-step process, first calculating the approximate orientation of the camera using the surface of the sea and the volcano topography, then improving the relative orientation of the cameras using the positions of selected bombs on the images.

For the first step, we used high-resolution topography (lidar topography, resolution 1 m, Favalli et al. 2009) and calculated its projection on the CCD sensors as well as the projection of the sea surface. The projection depends on the values of φ, κ and ω, and the angles were adjusted by fitting the projections on the real images (the sea and the topography were visible on the first images taken after sunset). For details, the readers may refer to Online Resource 1 and 2 for numerical codes and an explanation of the methodology respectively. The angles φ and ω could be estimated using the sea surface, which appeared clearly and sharply on images, and can be located with a precision of less than 2 pixels. Such a precision of the sea surface gives a precision better than 0.2° for φ and ω. The calculation of κ was carried out using the digital topography (except for crater area which has evolved over time due to volcanic activity). The precision is better than 1°. Such precision in the angles (φ ± 0.2°, ω ± 0.2° and κ ± 1°) induces uncertainties in the bomb locations of less than ± 10 m in y and z and about 100 m in x (i.e. East-West direction). In the second step, to improve the precision of the relative orientations of the cameras and of the bomb locations, we manually matched 30 volcanic bombs which could be recognised unambiguously in the two cameras. The image of each bomb projected on the first camera must be located on the second camera along a line, called epipolar, which is defined by the orientations and positions of the cameras (Fig. 1). It was then possible to find the best relative orientation of the two cameras by using a least squares minimisation of the distances between all 30 bombs on the second camera with their corresponding epipolar lines. The precision of the angles is better than 0.1°, the method mainly improving the precision of κ.

Trajectory reconstruction for a single camera

For trajectory reconstruction, the first step was to identify the maximum number of bombs on all images corresponding to a given eruption. To locate the maximum of bombs and minimise false detections, we used the algorithm written by Crocker and Grier (1996), and Blair and Dufresne (see Online Resource 1). It suppresses pixel noise and long-wavelength image variations to detect circular objects and locate their centres from their luminosity. This was made easier when image acquisition was carried out at night so that bombs appeared as bright pixels on a black background. More than 1000 bombs can be detected unambiguously on an image.

The second step was to follow all the identified bombs on the successive images of each camera. We developed our own algorithm for counting and tracking the rise and fall of bombs because no tracking algorithm available was found to be adapted to the complexity of bomb trajectories. The values of the arbitrary parameters given here have been estimated by several tests, retaining the values that give the greatest number of correct trajectories. Starting from each detected bomb in the first image of a given camera, the algorithm first initialises the trajectory by detecting all bombs in the next image from the same camera that are not ‘too far’ from its initial position. A distance of 100 pixels that can detect all bombs of velocities slower than 300 m s−1 gives the best results for the explosions studied (see detect_trajectories in Online Resource 1 for details). At this stage, hundreds of trajectories are possible for a given bomb. Then, for each trajectory, the algorithm estimates the position of the bomb on the third image by extrapolation of its position on the previous two images. It detects if a bomb with similar brightness (i.e. the bomb luminosity must be greater than 20% of the luminosity of the bomb initially detected) exists in the third image in the neighbourhood of the estimated position (5 pixels, i.e. ~ 1.5 m between the estimated and the real positions). If no bomb exists, the trajectory is deleted. If a bomb is detected, it extrapolates the position onto the fourth image and so on until the end of the sequence. Once a trajectory has been identified by more than 10 consecutive positions, it is fixed and the algorithm will ignore the absence of a bomb within a subsequent image to take into account the fact that bombs can be hidden by others or by an ash cloud. At this stage, the algorithm has only detected the trajectories of the bombs visible on the first image. To allow the detection of trajectories of bombs that were not present or were hidden on the first image, the algorithm starts again from the second image, then from the third and so on. Trajectories that have already been detected are not recorded again. An example of a trajectory reconstruction is shown in Fig. 3 for more than 1600 unambiguous bomb trajectories. A movie of the eruption with superimposed trajectories and the related code are provided in Online Resource 1 and 3.

Fig. 3
figure 3

Example of a trajectory reconstruction. More than 1600 bomb trajectories are identified unambiguously. The coloured trajectories A and B are used as examples in the text. Eruption of 17:20 UT, on October 5, 2012, taken by camera 2 (location shown in Fig. 2). Axes are the image distance in pixels

Trajectory matching between the two cameras

In order to calculate the real position of bomb trajectories in space, we need to match all the pairs of trajectories that correspond to similar bombs. Some bombs are easy to identify due to their size, brightness or position, but overall matching is very difficult (and often impossible to do manually).

The first step is to account for the short delay (< 20 ms, i.e. 2 m for a bomb moving at 100 m s−1) between image acquisition at the two cameras due to the second being triggered by the first. With the cameras connected to the same computer, the delay value is known and can be used to interpolate the trajectories of the second camera to estimate the bomb positions at the same time as the images of the first camera.

The automated method we have developed for trajectories matching is illustrated by the animation and a tutorial code in Online Resource 1 and 4. It first selects a bomb in the first image. Using its epipolar line, it is then possible to estimate which of the bombs observed in the second image it may be. To allow for errors in the camera orientations, in distortion corrections and in trajectory fluctuations, an uncertainty of 5 pixels around the epipolar line is accepted for the bomb matching. Continuing this process for all the images over the time sequence, we could identify the matching trajectory, i.e. the trajectory of camera 2 whose bombs are always on the epipolar lines of the selected trajectory of camera 1. However, due to the high number of trajectories, this method can give non-unique solutions, particularly for short trajectories close to the crater. We then filter the solution by calculating the 3D shape of the trajectory in space (see next section). A match is rejected if the calculated trajectory does not originate from the crater area and if the estimated value for gravity is outside an acceptable range. A range of 8 to 12 m s−2 is chosen so as not to eliminate trajectories that could have been influenced by wind, blast explosion, collisions, etc. If several trajectories still match the chosen trajectory, we discard them even though it might still have been possible to identify the most probable corresponding trajectory by analysing errors in the epipolar calculation and the gravity value. We also discard trajectories that are too short. As a conservative result, only 10 to 20% of the total trajectories is retained in order to improve the chance to delete all the false matching trajectories. For example, at 17:20, about 1600 2D trajectories have been reconstructed for each camera (Fig. 3) but only 230 have been retained for 3D reconstruction and hundreds of correlations, being non-unique with the accuracy chosen, have been discarded.

3D reconstruction

The last step is based on classic 3D reconstruction. Using camera positions and orientations, lens characteristics and the matching trajectories, it is possible to calculate the position of each bomb at each time step following the principle of Fig. 1 (see also Gaudin et al. 2016). Using the recorded time of each image, we reconstructed the 3D position—in time and space—of the bombs. Due to the uncertainties in camera orientations and lens correction and in pixel locations, the lines L1 and L2 do not intersect in 3D but their closest distance of approach is generally less than 50 cm. The accuracy of the bomb positions is not easy to estimate because it is not possible to identify objects in the images whose positions have been measured by independent methods. From a comparison of the trajectories of the bombs that roll around the crater with the lidar topography, accuracy appears better than a few meters.

The precision is estimated by selecting all the matching trajectories and by calculating 10,000 variations of each trajectory by adding small random offsets to the image positions, the camera orientations and distortion parameters of the lenses. The bomb images are circular and some pixels of diameter for the largest and their centres are located quite accurately. Their offsets are assumed to follow a normal distribution law with a standard deviation σ = 1 pixel in row and column. For the orientation of the cameras, the three angles of each camera follow a standard deviation σ = 0.1°. A standard deviation of 0.3 pixel is used for the uncertainties related to the distortion parameters q1, q2, k1, k2 and k3 and in the calibrated focal length. The effect on the estimated positions of the bombs in space is σ ~ 4.5 m. The standard deviation is smaller than 2 m in y and z, and of 4 m in x (i.e. east-west direction) due to the system geometry (Fig. 2). If the random offset is only applied to the image positions of the bomb centre, the standard deviation on the bombs in space is less than 1 m. If the random offset is only applied to the lens distortion parameters, the standard deviation on the bombs in space is less than 0.3 m. The main uncertainty is caused by the camera orientations, due to the calibration method used and the inaccessibility of the crater area that prevent the installation of ground control points.

Results

Estimation of bomb velocity using 2D and 3D trajectory reconstruction

Figure 4 compares velocities estimated by the 2D method (assuming that the bombs move in a vertical plane passing through the crater) with those measured by our 3D reconstruction method. Two bombs from the October 5, 2012 17:20 UT eruption are used as examples (throughout the article, time is given in Universal Time, UT, which corresponds to the local time minus 2 h). One bomb moved approximately perpendicular to the lines of sight of the cameras (bomb A, Fig. 3), the other moved approximately parallel (bomb B, Fig. 3). Both methods show a deceleration of the bomb during the rising phase followed by a downward acceleration that decreased towards a constant terminal velocity when the air drag (related to the bomb velocity) equilibrates the bomb weight. On the curves shown, the bombs impacted the ground before a constant fall velocity was reached. Figure 4 shows that the curves are not superimposed and that the error on bomb velocity incurred using the 2D assumption leads to systematic underestimations of the true velocity between 20 and 95%, the maximum error being associated with the trajectory high point. Bomb B, which moved parallel to the camera’s line of sight, had a very small horizontal component of displacement on the images. For example, where it reached the highest elevation (t = 1217.15 s, Fig. 4), the 2D velocity was near to zero while the 3D analysis shows that the bomb was actually moving away from the cameras with a horizontal velocity of 5 m s−1.

Fig. 4
figure 4

Comparison of the bomb velocities estimated using the 2D method (vertical plan passing by the crater) and those calculated with the 3D reconstruction (eruption of 17:20, October 5, 2012). A and B (and their respective colours) correspond to the bombs of Fig. 3

Characteristics of the explosions using 3D-analysis

Figure 5 shows a 3D view of the bomb trajectories through time for six eruptions that occurred on October 5, 2012. The trajectories reveal the highly asymmetric ejection of the largest pyroclasts as a function of time, the strong variability of jet directivity and of the maximum bomb heights among eruptions over a period of 2 h. Figure 6 is a plot of the maximum elevation reached by the bombs according to the azimuthal direction of their ejection. It illustrates the differences in direction and intensity between the recorded explosions. Maximal ejection heights above the vent (about 755 m a.s.l.) vary between 105 m (860 m in elevation) for the weakest eruption (i.e. 17:47) to 165 m (920 m in elevation) for the strongest (i.e. 17:20). Maximum heights tend to be reached in the predominant ejection direction (Fig. 6a, c) but exceptions occur (Fig. 6b).

Fig. 5
figure 5

af 3D reconstruction of the trajectories of 6 explosions on October 5, 2012. The colours indicate the time in seconds from the onset of the explosion. Black lines are parabolic fits of the trajectories

Fig. 6
figure 6

Maximal elevation reached by the bombs and azimuthal direction of ejections. The elevation of the NE crater in October 2012 was about 755 m a.s.l. The graphics show that the bombs were essentially ejected in a SW direction. The eruption at 19:12 was characterised by a significant amount of ejecta to the NW, a direction that is rare or absent from the other eruptions

Figure 7 links the azimuth of the ejection, the ejection time, the initial velocity (calculated in a further section) and the trajectory duration. In Figs. 5d and 7d, it can be seen that during the small eruption at 19:00, bombs were ejected for a period of 2 s with no preferential azimuth. For the more energetic eruption at 17:47 (Fig. 7c), the bombs were ejected over a wide azimuth range from south-east to north-west, although with an overall predominant direction towards the south-west (azimuth, N200). This predominant direction of ejection was also visible during other high energy explosions (17:04, 17:20, 19:12 and 19:22). For example, bombs were ejected continuously for a period of 8 s at 17:04 and again at 17:20 (Fig. 7a, b). These two eruptions were also characterised by an initial ejection of bombs with no preferential azimuth for about 1 s, before becoming predominantly oriented to the south-west. The eruption at 19:22 (Fig. 7f) began with an emission predominantly oriented towards the south-west for the first 7 s. Two seconds after the beginning of the eruption, a number of bombs were ejected in all directions. When the bomb trajectories exhibit a high directivity, the highest bomb velocities tend to be recorded in the predominant ejection direction. The eruption of 19:12 had two predominant directions of ejection (Figs. 5e, 6c and 7e). The eruption began with an ejection of bombs in the same direction (SW) as the other eruptions for 5 s. However, 2 s after the beginning of the eruption, another preferential direction of ejection joined the first, in which bombs were ejected towards the north-west for 2 s. This direction of emission was detected too for other eruptions but is less clear. It can be seen at 17:04 in Fig. 5a, with the two short pulses recorded at 17:47 in Fig. 7c (at t = 2841 s and t = 2843 s) and with few bombs ejected at 17:20 (Figs. 5b and 7b).

Fig. 7
figure 7

Azimuth of the bomb ejection against time for the six eruptions studied on Oct. 5, 2012. The horizontal axis is the time in seconds (ac from 17:00 UT, and df from 19:00 UT). The black dots indicate the ejection time of each bomb and their size indicates the initial ejection velocity (see inset in a). The grey lines represent the time period of the reconstructed trajectories (up to the point the bombs reach the ground, are masked by the topography or are too cooled to be detected)

Discussion

Bomb modelling

To illustrate the potential of our 3D-reconstruction method, this section gives an example of the parameters that can be deduced by comparing a simple ballistic model with our 3D-trajectories reconstruction. Details of the model are presented in Online Resource 2. The model is simple and more complex models have been developed (Taddeucci et al. 2017, and references therein). It is presented here only to illustrate the possibilities given by the 3D reconstruction of trajectories. Our model calculates the bomb velocity from the Newton’s First Law and the drag force exerted by the atmosphere on a bomb:

$$ \frac{d\mathbf{v}}{dt}=\mathbf{g}-c\times \left(\mathbf{v}-\mathbf{w}\right)\times u $$
(4)

where t is time, v = (vx, vy, vz) is the bomb velocity, w = (wx, wy, wz), the wind velocity, u = ‖v − w‖the relative velocity between the bomb and the air and g = (0, 0, −9.81) is gravity in m s−2. For a spherical bomb, the parameter c is defined by

$$ c=\frac{3}{8\times r}\frac{\rho_a}{\rho}\times {C}_d $$
(5)

with ρa the atmosphere density, ρ the bomb density, r its radius and Cd the drag coefficient (Alatorre-Ibargüengoitia et al. 2012; Konstantinou 2015).

By fitting the model to our 3D trajectories, we can estimate the three components of the bomb velocity (vx, vy, vz) at the first detection (black dot in Fig. 8), the horizontal wind velocity (wx and wy) and its orientation (βw, azimuthal origin), and the coefficient c related to the atmospheric drag on the bomb. Best fit estimations of these parameters were obtained by systematically varying the six parameters and minimising the standard deviation between the model and the observed trajectories in both space and time. Figure 8 shows a simulation of bomb B (see Fig. 3). For this bomb, the set of parameters that reproduces the observed data (grey line, Fig. 8) are as follows: vx = − 7 m s−1, vy = − 1.5 m s−1, vz = 25.5 m s−1, c = 0.0055 m−1, w = 3.5 m s−1, βw = − 55° (wind from NW). The effects of misestimating the projection positions of the bombs and the camera orientations are dominantly accommodated by translations of the trajectories. This is why, even in the worst cases, the errors are low for the above values. They are estimated to be less than 10% for velocities and the drag coefficient, by varying the bomb positions, the camera angles and the lens distortion parameters with the normal laws of the 3D reconstruction section.

Fig. 8
figure 8

Measurement of the 3D trajectory of bomb B (in grey; cf. Figure 3) and simulations (in black) for 3 values of c (see Eq. 5). The data are fitted in space (a, d) and time (b, c). The dashed line corresponds to the extrapolation of the bomb trajectory from its first detected position back to the vent using the best-fit values

By using the values obtained for the parameters vx, vy, vz, c, w and βw, and inverting the time, we can determine the initial velocity and the initial direction of each bomb at the vent. The fitting between the measured trajectories and the model can be done automatically for all bombs detected during an eruptive phase. Figure 9a is a histogram of the initial bomb velocities of the 17:20 eruption. The majority of the bombs were ejected at a velocity of 50 m s−1, with the velocities ranging from 30 to 130 m s−1. At the crater where the bombs are ejected nearly vertically, the 3D characterisation improves only the velocity estimation of less than 10% compared to 2D methods. The 3D method is, however, a powerful tool for the estimation of ejection velocities if the bomb cannot be observed directly at the vent.

Fig. 9
figure 9

a Histogram of calculated initial ejection velocities at the vent for the 230 bombs trajectories reconstructed during the 17:20 eruption using the model and its extrapolation. b Apparent drag coefficient c calculated using the model. The vertical lines indicate the radii corresponding to the values of c for Cd = 0.7 and a bomb density of 1800 kg m−3

Modelling of the 3D trajectories is also a powerful tool for estimating the drag of the bomb through the air. To illustrate the sensitivity of the model to the parameter c, two other curves are added to the initial curve (c = 0.0055 m−1) in Fig. 8 with values of c = 0.01 m−1 and c = 0.0025 m−1. Figure 9b is a histogram of the coefficient c calculated for all the bombs of the 17:20 eruption. The coefficient ranges between 5 × 10−4 (10–3.25) and 0.25 m−1 (10–0.6). To first order, the histogram can be used to estimate the size distribution of the bombs from Eq. 5. If, for example, we assume a value of Cd = 0.7, which has been estimated for natural bombs (Alatorre-Ibargüengoitia and Delgado-Granados 2006; de’Michieli Vitturi et al. 2010) and a bomb density of 1800 kg m−3 (e.g. Gurioli et al. 2013; Bombrun et al. 2015; Lautze and Houghton 2007; Harris et al. 2013), a value of c = 10−3 m−1 corresponds to a radius of 14.6 cm. The bomb radii corresponding to Cd = 0.7 and ρ = 1800 kg m−3 are plotted in Fig. 9b. They range between < 1 mm and > 30 cm. Note that, below 64 mm down to 2 mm, the term lapilli must be used instead of bomb and ash below 2 mm. It should also be pointed out that it seems unlikely for the smallest size particles of the histogram (in particular ash < 1 mm) to be sufficiently radiative to be detectable with a pixel size (i.e. a ground sample distance) of 20 cm. The sizes given are dependent on the assumptions about the bomb/lapilli density and the drag coefficient. A particle with a low density, a complex shape and roughness might show the same coefficient c as a smaller particle with the density chosen and the spherical shape used for the example. A strong deceleration, comparable to that of a small particle, can also be observed for a larger particle leaving a gas jet above the crater. Our method of 3D reconstruction can be used with more complete numerical models that would take into account these parameters.

Limitations and improvements for the method

In the future, the number of trajectories detected by our methodology can be improved by improving the camera resolution and frame rates. For example, more than 3000 trajectories were detected for each camera, but fewer than 300 bombs were identified unambiguously. The use of higher resolution cameras would improve the precision of the bomb locations, would give more precise estimation of the bomb size and would reduce bomb identification uncertainties. Higher frame rates, using high-speed cameras, would also reduce uncertainties of bomb recognition and facilitate trajectory reconstruction for each camera because bomb positions would be very close on successive images. With more confidence in the recognition, selection criteria that reject some trajectories could be lowered. This would allow complex trajectories, such as those induced by bomb collisions, to be detected unambiguously (Vanderkluysen et al. 2012). Another parameter that could be improved is the dynamics of the sensor. The challenge is to be able to see the bombs in the first phase of the eruption, when the number of hot bombs is high and tends to saturate the sensor. However, if the sensitivity is too low, the bombs that cool rapidly are not detectable. With the cameras used, it was possible to acquire images with a higher dynamic range, for example 12 or 16 bytes. However, this would have produced a very high flux of data to be transmitted and recorded (55 to 73 Mb/s), which exceeded the capability of the laptop used during fieldwork. Finally, the system could be improved by using more than two synchronised cameras around the area under observation. This would combine the advantages of easier recognition of similar bombs with the accuracy related to a wider angle of view and would highly improve bomb recognition.

Contribution of 3D reconstruction for conduit processes

The computed trajectories for six eruptions show a preferential direction of ejection to the south-west. It could be argued that such a direction, away from the camera, could be an artefact. Bombs ejected parallel to the line between the vent and the cameras could be more difficult to identify because of their location in the brightest area above the vent, potentially causing larger errors on their trajectories. To check if the calculated direction is correct, we have also taken photos of the same vent from the summit of Stromboli (Fig. 2). From this view angle perpendicular to the view angles of the cameras, the direction of ejection calculated, towards the south-west, is clearly confirmed (Fig. 10).

Fig. 10
figure 10

Photograph of the eruption at 12:56 UT (Oct. 5, 2012); taken from the south-east (Pizzo) showing that the ejection is oriented to the SW (location of the photo in Fig. 2)

Our method might also be a useful tool for understanding conduit processes. It shows that, during the period studied, the direction of bomb ejection was either uni-directional or multi-directional and that it varied over time during an eruption and among eruptions within timescales of tens of minutes. For the eruptions at 17:04 and 17:20, the first bombs were ejected with relatively slow velocities (50–100 m s−1) and in all directions (Fig. 7a, b). The emissions then evolved to higher velocities with a predominant direction of ejection (N200). This is compatible with a slug of gas that reached the surface (James et al. 2004; Leduc et al. 2015). The explosions initially occurred close to the surface and they ejected bombs radially. Afterwards, the successive explosions might have become progressively deeper, and more influenced by the orientation of the magma conduit, which seems to have been oriented N200 with a dip of about 75° (95% of the trajectories lie between 60° and 85°) during our field campaign. These values are compatible with the inclinations obtained from the locations of VLP seismic events (Chouet et al. 2008a) even if the comparison is limited by the strong morphological changes that have affected the crater area within 15 years separating our field campaigns. The smallest eruptions at 17:47 and 19:00 correspond to superficial explosions, compatible with the radial directions of the bombs. However, it appears that even during small and superficial eruptions, the gas pressure can be high, based on the high velocities recorded at 19:00.

The eruption at 19:12 provides a more complete view of the superficial geometry. It began by two directions of ejection followed by a radial emission. The two directions of ejection, recorded to a lesser extent at 17:04 (Fig. 5a), can hardly be explained by a simple conduit geometry. The explanation could be that the directions of the explosions are controlled by the rheology of the superficial magma and its spatial distribution. The upper part of the conduit may be clogged by pyroclasts (Capponi et al. 2016) and is probably filled with a mixture of vesicular and denser, partially crystallised and degassed magma (e.g. Lautze and Houghton 2007; Burton et al. 2007; Bai et al. 2011; Gurioli et al. 2014). The latter can form static zones representing more difficult pathways, so that the preferential path of the rising slugs would be around the edges of the degassed magma zones. An upper conduit that broadened towards the surface, the top of which is partially obstructed by lava clots recycled from previous explosions in variable amount and conditions, might explain how explosions could occur simultaneously in many different orientations (Fig. 11). Perturbations of the static magma, rising of other slugs, or variations in the slug properties (viscosity, gas content) downwards could explain how a radial explosion occurred 3 s after the beginning of the directed explosions.

Fig. 11
figure 11

A possible scheme of the upper conduit of Stromboli compatible with the bomb trajectories observed. Radial ejections (red) can be explained by explosions near the surface. With time, the slug bursts at increasing depth, focusing the directions of ejections. The various directions observed may indicate heterogeneities in the magma rheology that can form more than a single path for the slugs

Conclusion

We have developed a system of synchronised cameras for the automated reconstruction of hundreds of volcanic bomb trajectories in 3D in space plus time. The synchronisation, done by connecting the cameras to the same computer, is essential for the study of high-velocity phenomena. The reconstructed trajectories, coupled with a ballistic model, allows deduction of bomb particle sizes (given shape and density assumption) from their drag coefficients and to calculate their initial velocities at the vents as well as their directions of ejection. Alternatively, if particle size is known, the drag coefficient can be used to solve for shape and/or density. The trajectories reveal the time and space variations in velocities and directions within single explosive events as well as between successive explosions. Their interpretation is compatible with preferential paths of slugs, which can become focused at the edges of the upper part of the conduit probably due to formation of a central viscous cap. Our method thus represents a tool allowing insights into superficial magmatic conditions and their relation with particle dynamics. It also provides calibration data for future techniques developed for emission dynamic characterisation, such as the Doppler radar method (Gouhier and Donnadieu 2008).