1 Introduction

In recent years, deep learning-based methods have become the mainstream in target detection. The existing detection algorithms can be divided into two broad categories: the two-stage detection algorithm, represented by the Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation (R-CNN) series [1,2,3], and the so-called one-stage detection algorithm, such as single-shot multibox detector (SSD) [4] and You Only Look Once: Unified, Real-Time Object Detection (YOLO) [5]. The two-stage model presents a relatively high precision but slow speed because of its preparation and the subsequent detection of the region proposal [6]. Furthermore, due to the feature-extraction ability of the convolution networks, the accuracy of the detection model is improved with an increase in the convolution layers [7]. Typically, fast R-CNN enhances R-CNN using ROI Pooling and softmax for classification, which further improves the detection performance [8]. The mask R-CNN adds a branch for segmentation tasks based on the faster R-CNN, which can enhance precise segmentation while detecting the target. On the contrary, the one-stage detection does not require a region proposal; thus, it achieves faster detection by direct regression. Typically, YOLO divides the input image into \(N \times N\) regions responsible for object detection. The detection results are derived directly after the IOU mapping. The SSD achieves end-to-end detection [9, 10] via multi-scale detection. It detects the image at a higher resolution and introduces an anchor mechanism similar to the faster R-CNN to predict the offset value and the confidence of the anchor box. It provides higher accuracy compared to YOLO. However, for practical applications, the SSD consumes much computation time to achieve high precision.

Many researchers resorted to truncating channels to improve the detection speed [11, 12]. The deep compression mehod compressed deep neural networks through channel pruning and Huffman coding [13]. In Ref. [14], the less weighted channels within convolutional networks were pruned to exploit a linear structure for efficient evaluation. For the lightweight of the SSD model, a Pelee-SSD model was proposed, where a Stem Block was adopted to improve the feature extraction capability of the network [15]. However, the Pelee-SSD model experienced a precision decline compared to the original SSD [4].

In this paper, unlike the previous speed improvement algorithms, we minimize the computation cost by removing the convolution layers and channels within the SSD model, where the the single convolution computation is reduced by the convolution separation. At the same time, we introduce a light multiple dilated convolution (LMDC) operator as a filter at the feature extraction layer to improve the detection accuracy. Then, we propose a combined LMDC-SSD model to improve the detection speed without compromising the accuracy.

The rest of this paper is organized as follows. Section 1 summarizes the relevant literature and highlights relevant issues. Section 2 details the construction of the LMDC operator. Section 3 describes the procedure for the parameter reduction and convolution separation. Then, the LMDC-SSD model is presented in Sects. 4, and 5 gives the detection evaluation results on a new apple dataset. Finally, Sect. 6 presents the conclusions and discusses possible future work.

2 Related work and problem statement

The vast majority of target-detection neural networks continuously deepen the network level in an attempt to achieve higher accuracy. A large number of convolution layers have been added to the network, resulting in a huge increase in computation. Typically, the detection performance of a CNN model, such as SSD, is determined by the complex convolution operation. The increase of convolution layers gives a rise of the computation amount, leading to a low detection speed and redundant parameters. A quantity of parameters contributes less to the detection precision, and there exist many channels within convolution layers providing negligibly extra features for the object description [12]. From a point of improving the computation speed of the network, we delete part of the convolutional layer and channels to reduce directly the computation amount firstly. Secondly, after calculating the number of operational parameters at each layer, the convolutional layers with a large amount of computation are split to further improve the speed of the network. At last, a LMDC operator is adopted to improve the feature extraction ability, so that the network derives a high detection speed and precision.

The computation cost is determined by the number of convolutional layers and the convolution operation of a single convolutional layer, i.e.,

$$N = \sum\limits_{k = 1}^{Q} {g(x,y)}$$
(1)

where Q is the number of convolution layers and g (x, y) the number of operating parameters for a convolution, which is defined as;

$$g\left( {x_{i} ,y_{j} } \right) = \sum\limits_{i = 1}^{P} {\sum\limits_{j = 1}^{C} {f(H,W,m,n,s)} }$$
(2)

where C and P are the number of convolution channels and layers, respectively; H and W denote the size of the feature map under the current channel; m and n are the size of the convolution kernel; s is the step length of the convolution. Then, the amount of a single computation under the current channel is governed by

$$f(H,W,m,n,s) = \left( {H - s} \right) \times \left( {W - s} \right) \times m \times n$$
(3)

Considering that the convolution operation can be regarded as a process of feature extraction and fusion, there is an amount of repeated calculation that contributes less to performance improvement. Thus, it is expected to have a considerable decrease in the computation complexity if the feature extraction and fusion are separated. Then, the operation of Eq. (2) is alternatively expressed as;

$$g_{1} (x_{i} ,y_{i} ) = r(x_{i} ,y_{i} )$$
(4)
$$g_{2} (x_{i} ,y_{i} ) = z(x_{i} ,y_{i} )$$
(5)

where g1 and g2 represent the feature extraction and fusion, respectively. They extract features firstly, and then, conduct feature fusion.

For the reinforcement of the future extraction-convolution operation, a novel LMDC operator is adopted as a filter at the feature extraction layer. Thus, the removal of the redundant convolution channels and layers can directly reduce the convolution parameters with no performance loss at a decreased network depth.

3 LMDC operator and dimension splicing

The performance of the detection neural network model depends on the completeness of the feature extracted from the convolution layers therein. To retain the information of the original feature map, which attenuates with an increase in the convolution layers, an LMDC operator is introduced into the convolution layers. It assumes the residual network structure, as shown in Fig. 1. The input feature map channel is reduced to 1/8 of the original size after the convolution. Subsequently, diverse information is perceived through the specially shaped convolution and the dilated convolution. The information of the feature map is obtained by the dilated convolution with different receptive fields. Then the features are fused by dimension splicing module with those from the original feature image to form a new feature map providing more comprehensive information.

Fig. 1
figure 1

LMDC operator

The dilated convolution and the ordinary convolution are compared in Fig. 2. The dilated convolution can obtain various ranges of information without changing the amount of computation [16, 17]. Hence, it improves the ability to acquire global information. The number of receptive fields is determined by the range of feature map on which the dilated convolution kernels operate. For the number limitations of the dilated convolution layers to be less than six, five receptive fields are adopted. Then, the LMDC is constructed by superpositioning parallelly multiple dilated convolution layers. Each receptive field of the dilated convolution operator uses a fraction of the original feature map channel. After the fusion of five groups of different receptive field features, the information becomes more comprehensive and richer.

Fig. 2
figure 2

Receptive field of dilated convolution

LMDC operators reduce the calculation amount by transforming the serial computation into parallel computation during future extractions. A five-layered parallel feature extraction structure uses 1/8 channel of the original feature map of size H × W × C. Since the number of channels in the network is a multiple of 8, the five groups of channel parameters are set to 1/8 of the original channel. Therefore, the number of arguments is reduced by 7/8 of the original channel.

Within the LMDC operator, following the feature fusion, the dilated convolution with smaller receptive fields operate directly on the feature map fraction. For the three groups with larger receptive fields, the 3 × 1 and 1 × 3 specialized convolution kernels are respectively used for operations before the dilated convolution. Besides, an asymmetric convolution pair with horizontal and vertical convolutions is employed to improve the multidirectional features. Figure 3 shows the asymmetric convolution. The ordinary convolution is replaced by two specially shaped convolutions. The feature map A undergoes a 1 × 3 convolution to generate the feature map B and subsequently a 3 × 1 convolution to generate the feature map C’. The new feature map C’ takes on an identical size to the ordinary convolution.

Fig. 3
figure 3

Asymmetric convolution

Finally, five groups of feature maps are spliced and fused to generate a new feature map identical to the size of the original.

4 Parameters reduction and convolution separation

The operation parameters are reduced by deleting the convolution layers and channels therein. Figure 4a and b show the deletion procedure. Based on the original SSD structure depicted in the left column of Fig. 4a, the LMDC-SSD is derived by the channel deletion and feature enhancement, as shown in the right column of the figure. For the ith convolution layer conv-i, it is composed of the ordinary convolution, the transposition convolution and the LMDC operator, as shown in Fig. 4b. Where n × n is the size of the convolution kernel and k is the number of the involved convolution channels. After the deletion of the convolutional layers and channels, the LMDC operator and transposition convolution are applied to obtain more detailed information and realize the feature fusion, respectively.

Fig. 4
figure 4

Channel deletion diagram of LMDC-SSD

The size of the channels is set to 1024. The ‘conv’ stage of each layer indicates the convolution operation on the input; convW-V represents the Wth convolution layer and its Vth stage. The number of the parameters is reduced by the following procedures. The functional layers are slimed by deleting five layers in the original SSD network, i.e., conv3-3, conv4-3, conv5-3, conv6-2, and conv7-2, and halving the outlet number of conv1 to conv4. Consequently, the channel size of conv6 decreases from 1024 to 256, whereas the number of channels for conv8-1, conv9-1, and conv10-1 increases from 256 to 128. Simultaneously, feature fusion layers are added at the end of conv4, conv6, conv7, and conv8, whereas conv4-3, conv6-2, and conv7-1 are replaced by LMDC operators, respectively. Finally, the transposition convolution is adopted at conv6-3, conv7-3, conv8-3, and conv9-3 to raise the feature map to the same dimension as the upper layer. The size of the ends of conv6, conv7, and conv8 is reduced to half of the dimension of the original feature map.

Then, the number of calculation parameters is expressed as:

$$g_{l} = \sum\limits_{i = 1}^{P - m} {\sum\limits_{j = 1}^{C/2} {f(x,y)} }$$
(6)

The calculated parameters are reduced compared to the original model expressed by Eq. (1).

To accelerate the speed, the separation of the feature extraction and fusion in the convolution layer were investigated. The numbers of computational parameters at each layer are listed in Table 1. According to the complexity in the distribution of computations in Table 1, the convolution separation is performed on the convolution layer with a large amount of computation. Figure 5 shows the separation diagram. In the original network model, as shown in Fig. 5a, the feature fusion and extraction are carried out at the same time. However, via the convolution separation, the feature extraction is conducted first for each channel and subsequently the feature fusion is completed, as depicted in Fig. 5b.

Table 1 Number of calculation parameters in each convolution layer
Fig. 5
figure 5

Schematic diagram of convolution separation

The original total number of parameters q is calculated as:

$$g = N_{1} \times N_{2} \times \Gamma \left[ {\sum\limits_{i = 1}^{{C_{1} }} {\Upsilon (n_{1} ,n_{2} )} } \right] \times K$$
(7)
$$\Upsilon (n_{1} ,n_{2} ) = \Gamma \left[ {\sum\limits_{i = 1}^{{n_{1} }} {\sum\limits_{j = 1}^{{n_{2} }} {\left( {z(i,j) \times n(i,j)} \right)} } } \right]$$
(8)
$$q = g,$$
(9)

where \(n_{1} \times n_{2}\) is the kernel size of the original convolution layer, \(N_{1} \times N_{2} \times C_{1}\) the number of the input channel, and C2 the output channel.

When the convolution layer is split, the number of parameters is calculated as:

$$g_{1} = N_{1} \times N_{2} \times \Gamma \left[ {\Upsilon (n_{1} ,n_{2} )} \right]$$
(10)
$$g_{2} = N_{1} \times N_{2} \times \Gamma \left[ {\sum\limits_{i = 1}^{{C_{1} }} {\Upsilon (1,1)} } \right] \times K$$
(11)
$$q = g_{1} { + }g_{2} ,$$
(12)

where \(\Gamma\) is the convolution operation for the matrix, and \(\Upsilon\) the single-point convolution of the feature map. This yields a computation relief by transforming from the compound computation to an explicit addition operation.

5 LMDC-SSD model

5.1 Network architecture of LMDC-SSD

Figure 6 shows the LMDC-SSD network. The first five layers are subsampled using pooling layers and other layers using convolution layers. The output of the LMDC operator and the adjacent feature layer are spliced to obtain information features. A total of 11,620 prior boxes are generated. Finally, target detection is completed by non-maximum suppression. However, only the accurate and single boxes are left, and the redundant invalid boxes are removed.

Fig. 6
figure 6

Network-structure diagram of LMDC-SSD

The LMDC operator is adopted to extract features within the original feature detection layers. Therefore, Conv3 and Conv5 irresponsible for the feature extraction, Conv8 and Conv9 with a smaller size of the feature map than the kernels in the LMDC are excluded from the use of the LMDC. While Conv4, Conv6 and Conv7 are suspended with a LMDC operator. The feature map obtained by the LMDC operator is then concatenated with the next layer through the transposition convolution. Figure 7 shows the operation of the transposition convolution. The conversion of a feature map from a low dimension to a high dimension is achieved, as shown in Fig. 7. Thus, the cascaded feature map captures the global information obtained by the LMDC operator and the semantic information of the two-layer feature map.

Fig. 7
figure 7

Transposition convolution

5.2 Configuration of prior boxes

The prior boxes are set based on the single-point multi-box detection method. According to the size of the feature map, six prior boxes of size [4, 4, 6, 6, 6, 6] were constructed for six layers in the feature map. The scale of the prior boxes is computed as;

$$s_{k} = s_{\min } + \frac{{s_{\max } - s_{\min } }}{m - 1}(k - 1),k \in [1,m]$$
(13)

where smin and smax (the scales of the lower and highest layers) are 0.2 and is 0.9, respectively, and all layers therein are regularly spaced. Different aspect ratios are assigned to the default boxes as follows:

$$a_{r} \in \left\{ {1,2,3,\frac{1}{2},\frac{1}{3}} \right\}$$
(14)

The width is computed as

$$w_{k}^{a} = s_{k} \sqrt {a_{r} }$$
(15)

and the height as

$$h_{k}^{a} = \frac{{s_{a} }}{{\sqrt {a_{r} } }}$$
(16)

for each default box indexed by k.

For an aspect ratio of 1, the default box is scaled by a factor,

$$s_{k}^{\prime } = \sqrt {s_{k} s_{k + 1} }$$
(17)

For the prior box specified by

$$d = (d^{cx} ,d^{cy} ,d^{w} ,d^{h} )$$
(18)

its corresponding bounding box is denoted by

$$b = (b^{cx} ,b^{cy} ,b^{w} ,b^{h} )$$
(19)

With

$$b^{{{\text{cx}}}} = d^{w} \left( {{\text{variance}}[0]*l^{{{\text{cx}}}} } \right) + d^{{{\text{cx}}}}$$
(20)
$$b^{{{\text{cx}}}} = d^{w} \left( {{\text{variance}}[0]*l^{{{\text{cx}}}} } \right) + d^{{{\text{cx}}}}$$
(21)
$$b^{w} = d^{w} \exp \left( {{\text{variance}}[2]*l^{w} } \right)$$
(22)
$$b^{h} = d^{h} \exp \left( {{\text{variance}}[3]*l^{h} } \right)$$
(23)

where \(l = (l^{{{\text{cx}}}} ,l^{{{\text{cy}}}} ,l^{w} ,l^{h} )\) is the predicted value of the bounding box. It is a representation of b relative to d and scaled by a hyperparameter variance.

During the object detection, the matching degree between the prior box and the real target is determined by the intersection over union [18] as follows:

$$J\left( {A,B} \right) = \frac{{\left| {A \cap B} \right|}}{{\left| {A \cup B} \right|}} = \frac{{\left| {A \cap B} \right|}}{{\left| A \right| + \left| B \right| - \left| {A \cap B} \right|}} \in \left[ {0,1} \right]$$
(24)

A large value of intersection over union indicates a high correlation. A threshold of 0.5 for the intersection over union is used to classify objects from the background. The prior box with an intersection over union larger than the threshold is assigned as the object labels.

5.3 Loss function formulation

The loss function is governed by the label and coordinate deviation of each prior box to tune the parameters in the network through back propagation algorithm. Specifically, the deviation is calculated by comparing the prediction box coordinates, category confidence and prior boxes in the network operation results with the target results. The target loss of the improved SSD model is defined as the weighted sum of the confidence loss Lconf and position loss Lloc, i.e.,

$$L\left( {x,c,l,g} \right) = \frac{1}{N}L_{{{\text{conf}}}} (x,c) + \alpha L_{{{\text{loc}}}} (x,l,g)$$
(25)

where α is the weight parameter and N the number of matches.

Lconf and Lloc are expressed as:

$$L_{{{\text{conf}}}} (x,c) = - \sum\limits_{{i \in {\text{Pos}}}}^{N} {x_{ij}^{p} \log (\hat{c}_{i}^{p} ) - \sum\limits_{{i \in {\text{Neg}}}} {\log (\hat{c}_{i}^{o} )} }$$
(26)
$$L_{{{\text{loc}}}} (x,l,g) = \sum\limits_{{i \in {\text{Pos}}}}^{N} {\sum\limits_{{m \in \left\{ {{\text{cx,cy}},w,h} \right\}}} {x_{{{\text{ij}}}}^{k} {\text{smooth}}_{L1} (l_{i}^{m} - \hat{g}_{i}^{m} )} } ,$$
(27)

where

$$\hat{g}_{j}^{{{\text{cx}}}} = \frac{{(g_{j}^{{{\text{cx}}}} - d_{i}^{{{\text{cx}}}} )}}{{d_{i}^{w} }}\;\hat{g}_{j}^{{{\text{cy}}}} = \frac{{(g_{j}^{{{\text{cy}}}} - d_{i}^{{{\text{cy}}}} )}}{{d_{i}^{h} }}$$
(28)
$$\hat{g}_{j}^{w} = \log \left( {\frac{{g_{j}^{w} }}{{d_{i}^{w} }}} \right)\,\;\hat{g}_{j}^{h} = \log \left( {\frac{{g_{j}^{h} }}{{d_{i}^{h} }}} \right)$$
(29)
$${\text{and}}\,{\text{smooth}}_{L1} \left( x \right) = \left\{ {\begin{array}{*{20}c} {0.5x^{2} } & {\left| x \right| < 1} \\ {\left| x \right| - 0.5} & {{\text{others}}} \\ \end{array} } \right.$$
(30)

l is the prior box position, c the confidence of the prior frame,\(g = (g^{{{\text{cx}}}} ,g^{{{\text{cy}}}} ,g^{w} ,g^{h} )\) the position parameter of the real box, and \(x_{{{\text{ij}}}}^{k}\) ∈ {0, 1} the matching degree between the predicted box i and the real box j. The superscript p and o are the probability of \(x_{{{\text{ij}}}}^{k}\) in the positive and negative samples, respectively.

6 Experimental results

The SSD detection model in Ref. [4] has shown a competent superiority in precision and speed performance. The proposed LMDC-SSD model derives from the SSD model by the parameters reduction, LMDC operator and feature cascade. Therefore, the effectiveness of the LMDC-SSD model was validated by the comparison to the SSD model with the same experimental conditions. Experimental verification of the LMDC-SSD was performed on three datasets according to the requirement of PASCAL VOC type. The validation datasets contain the sample information from different environments including apple datasets, bicycle and person datasets and vehicle datasets. Diverse background conditions like overlap, occlusion, strong light, etc., are involved in the samples for the model evaluation. Table 2 present the experimental configurations.

Table 2 Experimental configuration

The stochastic gradient descent algorithm was adopted for training the LMDC-SSD model. The input size was set to 300 × 300 and batch-size to 32. A step-wisely attenuated learning rate was adopted to improve the learning ability of the model. The initial learning rate was 0.05. A damping coefficient was used to attenuate the learning rate of the training process to a smaller value. The damping coefficient was 0.1 for the first 30 and 100 epochs, and the subsequent damping coefficient was chosen as 0.1 per 50 epochs. Besides, a batch normalization layer (BN) [19] was used to improve the training speed and the generalization ability of the network. The BN layer can train the network at a large learning rate, accelerate the convergence of the network, and control the overfitting problem.

During the training, the pretraining model was removed, and the training process started from 0. After 80,000 iterations, a stable result was obtained. Figure 8, which was generated by the visualization tool TensorboardX, shows that the LOSS curve of LMDC-SSD took on a similar changing trend as SSD. LMDC-SSD maintained good stability with the introduction of the LMDC module and the reduction of the BN layer. The training loss gradually declined during the first 2500 iterations. After about 10,000 iterations, the loss attained a stable value.

Fig. 8
figure 8

Training loss curve

Figure 9a shows the detection of the yellow apples with the object concealed by similar-colored background leaves. Figure 9b and c show the detection results of the red apples, which were partially occluded by the leaves or overlapped each other. Figure 9d shows the detection result of the apples with a simple and uniform background. The validation results showed that the LMDC-SSD provided excellent robustness in detecting objects against background interference.

Fig. 9
figure 9

Object detection samples of apple datasets

Table 3 compares the LMDC-SSD and the original SSD. The average precision (mAP) [20] and the frames per second (FPS) [21] were used as evaluation indexes to demonstrate the detection accuracy and speed of the algorithm, respectively. Except that the SSD ran with its original parameters, other experiments were conducted with a learning rate of 0.05. It means that the models were trained based on a randomly initialized values for the parameters. Test 5 was the result of the original SSD operation, and Test 1 was the operation result of LMDC-SSD. The first four columns show that in the case of the pretraining removal while using BN, the derived SSD model effectively solved the non-convergence problem of the training, which occurred when BN was not added, as in Test 6. Furthermore, the modified SSD model with merely reduced parameters, as in Test 3, obviously improved the detection speed by 75.4%, and the mAP increased by 4.2% compared to the SSD model in Test 4. Introducing LMDC into the SSD, as in Test 2, improved the mAP by 10.8% and greatly increased the FPS by 33.3%, compared to the SSD model in Test 4. The convolution separation further improved the performance of the LMDC-SSD model. As listed in Test 1, it yielded a rather satisfactory result; mAP was improved by 10.4% and FPS by 49%. The mAP of LMDC-SSD was 98.99%, which is 8.1% higher than that of SSD (90.89%). The FPS of LMDC-SSD was 85, which was much higher than that of SSD (59). Compared with the original SSD model, the FPS was 85, the accuracy was 8.1% higher than that of SSD (90.89%), and the speed was 44% higher than that of SSD (59).

Table 3 Performance comparison between LMDC-SSD and SSD

Further, to evaluate the influence of the modifications in conv3 to conv7, a series of stepwise experiments involving the parameters reduction, adoption of the LMDC and the feature cascade operation. The corresponding testing results are shown in Table 4. In Test 0, the original structure and components in SSD remained unchanged. It yielded a 90.89% mAP and 59 FPS on the apple dataset. In Test 1 and Test 2, the detection models took a parameter reduction in channel deletion and convolution reduction, respectively. They showed a better detection performance of 95.33% mAP with Test 1 and 93.49% mAP with Test 2, 89 FPS with Test 1 and 100 FPS with Test 2. A LMDC operator introduced in Conv4 for the Test 3. It leads to an abnormal degradation with the mAP and FPS. The similar results appeared to the Test 4, where Conv 4 adopted a LMDC operator while Conv6 were suspended by the LMDC operator and the successive transposition convolution. Test 4 gave a 1.5% mAP and 51.7 FPS. Based on the model in Test 4, Test 5 employed a LMDC operator in Conv7 and successive transposition convolution. It showed an improvement on the detection precision and speed with 98.79% mAP and 85.8 FPS. Test 6 and Test 7 adopted the transposition convolutions at conv 8, both conv 8 and conv 9, respectively. They provided a gradual improvement with the mAP compared to the Test 5, from 98.79% mAP, 98.87% mAP, to 98.99% mAP, while a negligible decline at the FPS, from 85.8 FPS, 85.5 FPS to 85 FPS. Thus, the LMDC-SSD model as depicted in Fig. 6 provided a satisfactory detection performance both in precision and speed.

Table 4 A stepwise experiments results in Conv3 to Conv7

Table 5 presents the detection results of LMDC-SSD on the vehicle, bicycle, and person datasets. The mAP of LMDC-SSD increased by 0.99% in the bicycle and person datasets, whereas the FPS increased by 23. For the vehicle datasets, the mAP increased by 0.26% and the detection speed by 35 FPS. Compared to SSD, LMDC-SSD improved by 0.62% in the average detection accuracy and improved the average detection speed by 29 FPS.

Table 5 Experimental detection results

Figure 10 shows the detection results of LMDC-SSD on the bicycle and people’s datasets. As we can see, since most bicycles are in the cycling, there is overlap between bicycles and people, and people also have the problem of covering with each other. However, LMDC-SSD can successfully detect the target in the case of overlapping and shielding.

Fig. 10
figure 10

Detection results on bicycle and person datasets

Figure 11 shows the detection results of LMDC-SSD for the vehicle datasets. The vehicles have an incomplete exposure because of the overlap in Fig. 11a, while the vehicles in Fig. 11b encounter a partly capture due to the viewpoint of the image acquisition equipment. Thus their appearance is in poor integrity, and some vehicles are even less than 50%. However, LMDC-SSD can overcome this difficulty and successfully identify the vehicles. In addition, some vxiangehicles are in remote locations and overlap in large areas, and LMDC-SSD also can detect vehicles in a certain extent.

Fig. 11
figure 11

Detection results on vehicle datasets

The LMDC-SSD proposed herein is different from the network compression or network acceleration schemes. The feature extraction ability was improved from the network itself by compensating for the deficiency caused by network compression. The detection speed was improved without reducing the detection accuracy.

7 Conclusions

This paper presents a fast target detection method by parameter reduction. Convolution layers deletion and channel pruning were realized based on the SSD network model. The LMDC operator was introduced for the feature extraction. The separation between the feature extraction and feature fusion was adopted to improve the detection speed. The experimental results confirmed the effectiveness of the proposed LMDC-SSD model. It greatly improves the speed of network detection, and the LMDC operator effectively compensates for the precision loss caused by parameter reduction.

The application of shallow level features in this paper is still insufficient, and the detection performance in the overlapping part of small targets is weak. Shallow level features contain much detailed information, but the network itself is less applied. Subsequent research will focus on the fusion of the front and rear level features of the network. The scheme of shallow and deep levels fusion could effectively improve the detection performance of the network.