Abstract
Traffic sign detection is a challenging task for the unmanned driving system, especially for the detection of multi-scale targets and the real-time problem of detection. In the traffic sign detection process, the scale of the targets changes greatly, which will have a certain impact on the detection accuracy. Feature pyramid is widely used to solve this problem, but due to the diversity of traffic sign sizes, it cannot accurately extract multi-size feature maps, thus destroying the feature consistency between traffic signs. Moreover, in practical application, it is difficult for common methods to improve the detection accuracy of multi-scale traffic signs while ensuring real-time detection. In this paper, we propose an improved feature pyramid model, named AF-FPN, which utilizes the adaptive attention module (AAM) and feature enhancement module (FEM) to reduce the information loss in the process of feature map generation and enhance the representation ability of the feature pyramid. We replaced the original feature pyramid network in YOLOv5 with AF-FPN, which improves the detection performance for multi-scale targets of the YOLOv5 network under the premise of ensuring real-time detection. Furthermore, a new automatic learning data augmentation method is proposed to enrich the dataset and improve the robustness of the model to make it more suitable for practical scenarios. Extensive experimental results on the Tsinghua-Tencent 100 K (TT100K) dataset demonstrate that compared with several state-of-the-art methods, our method is more universal and superior.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
The traffic sign recognition system is the data foundation of intelligent transportation systems (ITS) and unmanned driving system, balancing the accuracy and real-time performance of the traffic sign detection and recognition technology, which plays an important role in the subsequent decision-making of ITS and unmanned driving system [1].
In recent years, most of the state-of-the-art object-detection algorithms have used convolutional neural networks (CNNs) and have achieved fruitful results in target detection tasks, such as the two-stage detectors Faster R-CNN [2], R-FCN [3], the one-stage detectors SSD [4], and YOLO [5]. However, directly applying these methods to traffic sign recognition is hard to achieve satisfactory results in practical application. The target recognition and detection of the vehicle-mounted mobile terminal require high accuracy for targets of different scales, and high requirements for recognition speed, which means to meet the two requirements of accuracy and real-time [6, 7].
Traditional CNNs use a large number of parameters and floating-point operations per second (FLOP) to achieve better detection performance. For example, VGG-16 [8] has about 138 M parameters and requires 14.9B FLOPs to process an image of size 608 × 608. However, mobile devices (e.g., smartphones and self-driving cars) with limited memory and computation resources cannot be used for deployment and inference for larger networks. As a one-stage detector, the YOLOv5 [9] is used in this paper because of the advantages of low computation and fast recognition speed.
In this paper, an improved YOLOv5 network is proposed, which not only ensures that the model size can meet the requirements of deployment on the vehicle side but also improve the ability of multi-scale targets and meet the real-time requirement.
The main contributions of our work are summarized as follows:
-
A novel feature pyramid network is proposed in this paper. Through adaptive feature fusion and receptive field enhancement, it retains the channel information in the feature transfer process to a large extent and learns different receptive fields in each feature map adaptively to enhance the representations of feature pyramids, effectively improving the accuracy of multi-scale targets recognition.
-
A new automatic learning data augmentation strategy is proposed. Inspired by AutoAugment [10], the latest data augmentation operations have been added. The improved data augmentation method effectively improves the model training effect and the robustness of the training model, which has more practical significance.
-
Unlike the existing YOLOv5 network, the current version is improved to reduce the impact of scale invariance. Meanwhile, it can be deployed on the mobile terminal of the vehicle to detect and recognize traffic signs in real-time.
The rest of this paper is organized as follows: Sect. 2 introduces related works about CNN-based traffic sign detection and data augmentation. Section 3 introduces the details of the proposed method to detect and recognize the traffic signs efficiently in real-time. The experimental results and analysis are presented in Sect. 4. Finally, the conclusion is described in Sect. 5.
2 Related works
2.1 CNN-based traffic sign detection
At present, CNN as a popular algorithm for deep learning has a wide range of applications in computer vision, natural language processing, visual-semantic alignments, and other fields [11,12,13,14]. According to whether region proposal is required, it can be divided into two categories: single-stage detection and two-stage detection. Single-stage detection is often used in traffic detection due to its fast detection performance.
Shao et al. [15, 16] proposed an improved Faster R-CNN to traffic sign detection. They simplified the Gabor wavelet through a regional suggestion algorithm to improve the recognition speed of the network. Zhang et al. [17] modified the number of convolutional layers in the network based on YOLOv2, proposed an improved one-stage traffic sign detector, and used the China Traffic Sign Dataset for training to make it better adapted to Chinese traffic road scenes. A novel perceptual generative adversarial network was developed for the recognition of small-sized traffic signs [18], which boosts detection performance by generating super-resolve representations for small traffic signs. Aiming the scale variety problems in traffic sign detection, SADANet [19] combines a domain adaptive network with a multi-scale prediction network to improve the ability of the network to extract multi-scale features.
Most of the above-mentioned networks use single-stage detection and only use single-scale depth features, so it is difficult for them to have superior detection and recognition performance in sophisticated traffic scenes. Traffic sign instances of different scales have great differences in visual features, and the proportion of traffic signs in the entire traffic scene image is very small. Therefore, the scale variety problem has become a major challenge in traffic sign detection and recognition. And learning scale-invariant representation is critical for target recognition and location [20]. At present, this challenge is handled mainly from two aspects: network architecture modification and data augmentation [21].
At present, multi-scale features are widely used in high-level object recognition to improve the recognition performance of multi-scale targets [22]. Feature Pyramid Network (FPN) [23], as a commonly used multi-layer feature fusion method, uses its multi-scale expression ability to derive many networks with high detection accuracies, such as Mask R-CNN [24] and RetinaNet [25]. It is worth noting that the feature maps will suffer from information loss due to the reduced feature channels and only contain some less relevant context information in the feature maps of other levels. Moreover, FPN pays too much attention to the extraction and optimization of low-level features. As the number of channels decreases, high-level features will lose a lot of information, resulting in a decrease in the detection accuracy of large-scale targets [22]. In response to this problem, a simple yet effective method named receptive field pyramid (RFP) [26] is proposed to enhance the representation ability of feature pyramids and drive the network to learn the optimal feature fusion pattern [14].
2.2 Data augmentation
Data augmentation has been widely utilized for network optimization and proven to be beneficial in vision tasks [2, 8, 27], which can improve the performance of CNN, prevent over-fitting [28], and is easy to implement [29].
Data augmentation methods could be roughly divided into color transformation (e.g., noise, blur, contrast, and color casting) and geometric transformation (e.g., rotation, random cropping, translation, and zoom) [30]. These augmentation operations artificially inflate the training dataset size by either data warping or oversampling. Lv et al. [31] proposed five data augmentation methods dedicated to face detection, including landmark perturbation and a synthesis method for four features of hairstyles, glasses, poses, illuminations. Nair et al. [32] performed geometric transformation and color transformation on the dataset. The training dataset is expanded and enriched by random crop and horizontal reflections and applying PCA on color space to change the intensity of the RGB channel. These frequently used methods just do simple transformations and cannot simulate the complex reality. Dwibedi et al. [33] improved detection performance with the cut-and-paste strategy. Furthermore, the method of annotated instance masks with a location probability map is utilized to augment the training dataset [34], which can effectively enhance the generalization of the dataset. YOLOv4 [13] and Stitcher [35] introduce mosaic inputs that contain rescaled sub-images, which are also used in YOLOv5. However, these data augmentation implementations are manually designed and the best augmentation strategies are dataset-specific.
The effect of data augmentation strategies is related to the characteristics of the dataset itself, so the focus of recent work has shifted to learning data augmentation strategies directly from the data itself. Tran et al. [36] generated augmented data, using the Bayesian approach, based on the distribution learned from the training set. Cubuk et al. [10] proposed a new data augmentation method that can automatically search for improved data augmentation policies, named AutoAugment.
3 Proposed method
3.1 The improved YOLOv5 network
As the latest model in the current YOLO series, the superior flexibility of YOLOv5 makes it convenient for rapid deployment on the vehicle hardware side [37]. YOLOv5 contains four models, namely YOLOv5s, YOLOv5m, YOLOv5l, and YOLOv5x. YOLOv5s is the smallest model of the YOLO series and is more suitable for deployment on a vehicle-mounted mobile hardware platform due to its memory size of 14.10 M, but the recognition accuracy cannot meet the requirements of accurate and efficient recognition, especially for the recognition of small-scale targets.
The basic framework YOLOv5 can be divided into four parts: Input, Backbone, Neck, and Prediction [37]. The Input part enriches the dataset with mosaic data augmentation, which has low requirements for hardware devices and low computational cost. However, it will cause the original small targets in the dataset to become smaller, resulting in the deterioration of the generalization performance of the model. The Backbone part is mainly composed of CSP modules, which perform feature extraction through the CSPDarknet53 [13]. FPN and Path Aggregation Network (PANet) [38] are used to aggregate the image feature at this stage in Neck. As mentioned earlier, in the processing of generating high-level feature maps in the FPN, the number of channels is reduced, which may lead to the loss of key information, especially for small-sized target recognition. Finally, the network performs target prediction and output through the Prediction.
In this paper, the AF-FPN and the automatic learning data augmentation are introduced to solve the problem of incompatibility between model size and recognition accuracy, and further improve the recognition performance of the model. The original FPN structure is replaced by AF-FPN to improve the ability to recognize multi-scale targets and make an effective trade-off between recognition speed and accuracy [26]. Moreover, we remove the mosaic augmentation in the original network and use the best data augmentation methods according to the automatic learning data augmentation policy to enrich the dataset and improve the training effect. The improved YOLOv5s network structure is shown in Fig. 1.
In Prediction, generalized IoU (GIoU) [39] loss is used as the loss function of the bounding box and the weighted non-maximum suppression (NMS) [40] method is used for NMS. The loss function is as follows:
where C is the smallest box covering B and Bgt. Bgt = (xgt, ygt, wgt, hgt) is the ground-truth box, and B = (x, y, w, h) is the predicted box.
However, when the predicted box is inside the ground-truth box and the size of the predicted box is the same, the relative positions of the predicted box and the ground-truth box cannot be distinguished.
In this paper, the GIoU is replaced by complete IoU (CIoU) [41] loss. Based on GIoU loss, the CIoU loss considers the overlap area, central point distance of bounding boxes, and the consistency of aspect ratios for bounding boxes. The loss function can be defined as:
where RCIoU is the penalty term, which is defined by minimizing the normalized distance between central points of two bounding boxes. b and bgt denote the central points of B and Bgt, ρ(·) is the Euclidean distance, and c is the diagonal length of the smallest enclosing box covering the two boxes. α is a positive trade-off parameter, and ν measures the consistency of the aspect ratio.
And the trade-off parameter α is defined as:
which the overlap area factor is given higher priority for regression, especially for non-overlapping cases.
3.2 AF-FPN structure
Based on the traditional feature pyramid network, AF-FPN adds the adaptive attention module (AAM) and the feature enhancement module (FEM). The former part is used to alleviate the problem of context loss in high-level feature maps due to feature channel reduction. The latter part enhances the representation of feature pyramids and accelerates the inference speed while achieving state-of-the-art performance. The structure of the AF-FPN is shown in Fig. 2.
The input image generates feature maps {C1, C2, C3, C4, C5} through multiple convolutions. C5 generates the feature map M6 through AAM. And M6 is combined with M5 by summation and propagated to fuse with other features at lower levels through a top-down path, and the receptive field is expanded through FEM after each fusion. PANet shortens the information path between lower layers and the topmost feature.
The operation of the adaptive attention module can be performed in two steps. First of all, the multiple context features with different scales are obtained through the adaptive average pooling layer. The pooling coefficient β is [0.1, 0.5], and it adaptively changes according to the target size in the dataset. Secondly, a spatial weight map is generated for each feature map through the spatial attention mechanism. Through the weight map, context features are fused to generate a new feature map, which contains multi-scale context information. The new feature map is combined with the original high-level feature map and propagated to fuse with other features at lower levels.
The specific structure of the AAM is shown in Fig. 3. As the input of the adaptive attention module, the size of C5 is S = h × w. It first obtains context features with different scales of (β1 × S, β2 × S, β3 × S) through the adaptive pooling layer. Then each context feature undergoes a 1 × 1 convolution to obtain the same channel dimension 256. Bilinear interpolation is used to upsample them to the scale of S for subsequent fusion. The spatial attention mechanism merges the channels of the three context features through a Concat layer, and then the feature map sequentially passes 1 × 1 convolution layer, ReLU activation layer, 3 × 3 convolution layer, and sigmoid activation layer to generate corresponding spatial weights for each feature map. The generated weight map and the feature map after the merged channel are subjected to the Hadamard product operation, which is separated and added to the input feature map M5 to aggregate context features into M6. The final feature map has rich multi-scale context information, which to a certain extent alleviates the loss of information due to the reduction in the number of channels.
FEM mainly uses the dilated convolution to learn the different receptive fields in each feature map adaptively based on the varying scales of detected traffic signs, thereby improving the accuracy of multi-scale target detection and recognition. As shown in Fig. 4, it can be divided into two components: the multi-branch convolution layer and the branch pooling layer. The multi-branch convolution layer is used to provide different sizes of receptive fields for the input feature map through the dilated convolution. And the average pooling layer is used to fuse the traffic information from the three branch receptive fields to improve the accuracy of multi-scale prediction.
The multi-branch convolution layer consists of dilated convolution, BN layer, and ReLU activation layer. The dilated convolutions in the three parallel branches have the same kernel size but different dilation rates. Specifically, the kernel of each dilated convolution is 3 × 3 and the dilation rates d is 1, 3, and 5 for different branches.
Dilated convolutions support exponentially expanding receptive fields without losing resolution or coverage [42]. However, in the convolution operation of dilated convolution, the elements of the convolution kernel are spaced, and the size of the space depends on the dilation rates, which is different from the elements of the convolution kernel that are all adjacent in the standard convolution operation.
The convolution kernel changed from 3 × 3 to 7 × 7 and the receptive field of this layer is 7 × 7. The formula for the receptive field of dilated convolution is as follows:
where k and ri denote the kernel size and dilation rate, respectively. And d denotes the stride of the convolution.
The branch pooling layer [43] is proposed to fuse information from different parallel branches and avoid introducing additional parameters. The averaging operation is utilized to balance the representation of different parallels branches during training, which enables a single branch to implement inference during the test. The expression is as follows:
where yp denotes the output of the branch pooling layer. B represents the number of parallel branches and we set B = 3.
3.3 Data augmentation
The augmentation policy consists of two parts: search space and search algorithm [44]. The search space contains 5 sub-policies, each of which consists of two simple image enhancement operations applied in sequence. One of the sub-policies are be chosen at random and applied to the current image. In addition, each operation is also associated with two hyperparameters: the probability of applying the operation and the magnitude of the operation [10]. The operations we used in the experiment include the latest data augmentation methods such as Mosaic [13], SnapMix [45], CutMix, Mixup, Gaussian blur [46] and Translate X/Y. In total, we have 15 operations in our search space. Each operation also comes with a default range of magnitude. We discretize the range of magnitude into D = 11 values that follow the uniform distribution so that we can use a discrete search algorithm to find them. Similarly, we also discretize the probability of applying one of all operation into P = 10 values (also following a uniform distribution). Finding each sub-policy becomes a search problem in a space of (19 × D × P)2 possibilities. Therefore, the search space with 5 sub-policies then has roughly (19 × D × P)2×5 possibilities and requires an efficient search algorithm to navigate this space [47]. Figure 5 shows the policy with 5 sub-policies in the search space.
Through the search space, the problem of searching for a learned augmentation policy into a discrete optimization problem. Reinforcement Learning [48] is used as the search algorithm, which has two components: a controller RNN and the training algorithm. The controller RNN is a recurrent neural network, and the proximal policy optimization (PPO) [47] with a learning rate of 0.00035 is used as the network training algorithm. It is trained by a reward signal, which is the degree to which the found policy improves the generalization ability of the proposed model [49, 50]. The proposed model is trained on augmented data produced by the five sub-policies in the training set. For the images in each batch, one of the five sub-policies is randomly selected for data augmentation, and then the accuracy of the network on the validation set is measured. This accuracy is used as a reward signal to train the RNN. The training process is shown in Fig. 6.
The controller RNN predicts a decision produced by a softmax at each step and the prediction is then fed into the next step as an embedding from the search space. Totally, the controller has 30 softmax predictions to predict 5 sub-policies, each of which has two operations, and each operation requires an operation type, magnitude, and probability. We applied the automatic learning data augmentation method to the TT100K dataset, and then used the best data augmentation policy obtained through training.
3.4 Experiments and analysis
In this section, we comprehensively trained the improved YOLOv5 model through the TT100K [51, 52] dataset, which includes 182 types of traffic signs instances with detailed annotations and covers different actual traffic environments. And about 42.5% of traffic signs in TT100K are small objects, which means it is more suitable for actual vehicle-mounted target recognition. The size distribution of traffic sign instances in the dataset is displayed in Fig. 7.
3.5 Experimental setting
Considering the fixed size of input demanded by the YOLOv5 network, we resized the images to uniform dimensions of 608 × 608. The training and validation datasets include 9146 images and the test dataset include 1121 image from TT100K. In addition, the CCTSDB [53] dataset is used to verify the generalization ability of the model. We selected 2000 images in the dataset as the test set for the experiments. In the process of training, the initial value of the learning rate was 0.01, and use the cosine annealing strategy to reduce the learning rate. The epochs and the batch size are set to 500 and 32, respectively. Our experiments were performed on a Linux4.15.0–142-generic Ubuntu 18.04 with Intel(R) Xeon(R) Silver 4210R CPU @ 2.40GH, 8 × 32 GB DDR4 and 8 × TITAN Xp, 12 GB memory. The mobile device used in the experiment is Jetson Xavier NX, and an external USB3.0 industrial camera.
3.6 Experimental analysis
To demonstrate the advantages of the proposed method in traffic sign detection, we evaluated our method on TT100K and compared it with the original YOLOv5, YOLOv5-Lite [54], EfficientDet [55], YOLOv5-face [56], M2det [57], SSD, and YOLOv3 [58]. We evaluated performance using metrics including model size, parameters, floating-point operations per second (FLOPs), mean average precision (mAP), average precision of large, medium, and small size targets (APL, APM, APS), and frames per second (FPS). FPS refers to the number of frames per second transmitted by the screen [59]. The videos we usually see are continuous pictures composed of pictures, and each picture is each frame of the video. Therefore, the higher the FPS, the smoother the picture. The specific results are shown in Table 1.
Firstly, it can be seen that the model size of the proposed method is 16.3 M, which is easy to be deployed on a mobile platform so that it can be used for real-time shooting and recognition on the vehicle side. The amount of parameters in the training process is slightly higher than that of EfficientDet and YOLOX. FLOPs is 17.9G which is only 3.3G larger than the optimal YOLOv5-Lite. It can be seen from these two indicators that our method has a faster training speed and requires less hardware equipment, which is convenient for popularization. Excessive reduction in the number of parameters and calculations will lead to a decrease in the detection effect of the final training model. Secondly, our method achieves the mAP of 65.14% on all 182 traffic sign classes, which is second only to YOLOX. Although the APL of our method is lower than M2det and YOLOX, the recognition accuracy on small targets is 41.46%, which is significantly higher than other methods. Finally, FPS is used as a metric to evaluate the speed of target detection, indicating that our method can meet the real-time requirements of detection on the mobile terminal. In general, our method has high accuracy for multi-scale target detection and can achieve a balance between recognition accuracy and recognition speed. The model size is suitable for deployment on the mobile terminal and has practical application significance.
In addition, traffic sign detection is a multi-category and multi-target recognition task, and the false detection rate and missed detection rate are also important metrics to measure the detection network. In order to verify the missed detection requirements of the proposed method in real-time traffic sign detection, Log-Average Miss RATE (LAMR) [60] is selected as the evaluation index. LAMR reflects the relationship between the false negative (FP) of each image and the missed detection rate. The lower the FP, the better the detection performance of traffic signs. We selected the top 19 traffic sign categories in the dataset, and compared the missed detections of each method on these types of traffic signs, as shown in Fig. 8. It can be seen that the missed detection rate of our method for traffic sign recognition is significantly lower than other methods, and it has practical application significance. However, the missed detection rate of several types of traffic signs such as ip, w57, and po is still high, and we will make further improvements in future research. Further, we count the average recall rate of the proposed method on these 19 categories, as shown in Fig. 9. The average recall rate is the ratio of the number of correctly identified samples to the number of all positive samples in the test set, which can reflect the accuracy and missed detection of the model to a certain extent.
We visualize the method proposed in this paper, as shown in Fig. 10a. It can be observed that our method successfully recognized the small-sized traffic signs on the actual traffic scene with high recognition accuracy, and there are almost no missed detections and false detections. Figure 10b shows the test performance of the trained model on the CCTSDB dataset. As can be seen from the figure, our model still has good training effect on other datasets, and can accurately and quickly recognize traffic signs in different environments, and has strong generalization ability.
We transplanted the trained model to the mobile platform and connected an external camera to shoot the actual road scene. Real-time traffic sign detection and recognition are performed on the captured images, and the recognition results are displayed on the LED screen, as shown in Fig. 11.
3.7 Ablation study
To more intuitively demonstrate the better performance of the proposed method for traffic sign detection and recognition, we conduct the ablation study, and the results are shown in Table 2. It shows the ablation result of incrementally adding the components training on the YOLOv5s model.
As observed from the results, the standard YOLOv5s provides a detection mAP of 60.18%. Integrating the data augmentation and the AF-FPN improves the mAP to 61.31 and 62.67%, respectively. The mean average precision (mAP) of our method on the TT100K dataset is 4.96% higher than that of the YOLOv5s, which means the proposed method achieves impressive performance in target and recognition. At the same time, the model size and parameters amount only slightly increase, and the FLOPs do not change, which means that the training speed of the improved network and the requirements for training equipment are basically unchanged. These ensure that our method can be easily deployed on the vehicle side. Although FPS decreases by 10, it still meets the requirements of real-time detection on the vehicle side.
4 Conclusion
In this paper, we proposed a real-time traffic sign detection network based on modified YOLOv5s, which achieves better detection performance than state-of-the-art one-stage detectors. In this work, the proposed AF-FPN structure improves the information extraction ability of feature maps and its representation ability for detecting multi-scale objects. And the new data augmentation strategy enriches the traffic sign dataset by adding Noise, Mosaic, and other methods to improve the training effect of the model. The empirical results verified that the proposed method could achieve state-of-the-art performance with a fast inference speed, the detection speed on the vehicle side is 95 FPS. The proposed method provides the input feature map of different receptive fields and fuses the receptive field pyramids for the target traffic signs. Therefore, the improved network can enhance the recognition accuracy of multi-scale targets without introducing additional calculations, the mAP has increased by 4.96% compared to the original network on the TT100K. Due to the size of the trained model being small, it is easy to deploy on the mobile device of the vehicle and perform real-time recognition and detection of the road scene. However, in practical applications, a faster vehicle speed will cause the motion blur of the image, which will affect the recognition result. In future, we plan to explore a better performance detection model for high-speed moving targets.
Availability of data and material
Data are available upon request.
Code availability
Code is available upon request.
References
Timofte R, Zimmermann K, Van Gool L (2009) Multi-view traffic sign detection, recognition, and 3D localisation. 2009 Workshop Appl Comput Vision (WACV). https://doi.org/10.1109/WACV.2009.5403121
Shaoqing Ren KH, Girshick Ross, Sun Jian (2017) Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Machine Intell 39:1137–49. https://doi.org/10.1109/TPAMI.2016.2577031
Dai J, Li Y, He K, Sun J (2016) R-FCN: Object detection via region-based fully convolutional networks. In: 30th conference on neural information processing systems (NIPS 2016), Barcelona, Spain
Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY, et al. (2016) SSD: single shot multibox detector. In: Computer vision–ECCV 2016 ECCV 2016 lecture notes in computer science. vol 9905 pp 21–37 https://doi.org/10.1007/978-3-319-46448-0_2
Redmon J, Farhadi A (2017) YOLO9000: better, faster, stronger. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 6517–25.https://doi.org/10.1109/Cvpr.2017.690
Pramanik A, Sarkar S, Maiti J (2021) A real-time video surveillance system for traffic pre-events detection. Accident Anal Prev. https://doi.org/10.1016/j.aap.2021.106019
Shen L, You L, Peng B, Zhang C (2021) Group multi-scale attention pyramid network for traffic sign detection. Neurocomputing 452:1–14. https://doi.org/10.1016/j.neucom.2021.04.083
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. CoRR
Ultralytics (2020) YOLOv5 2020 Available from: https://github.com/ultralytics/yolov5
Cubuk ED, Zoph B, Mane D, Vasudevan V, Le QV (2019) AutoAugment: learning augmentation strategies from data. 2019 IEEE/CVF conference on computer vision and pattern recognition (CVPR 2019). pp 113–23. https://doi.org/10.1109/Cvpr.2019.00020.
Ning X, Gong K, Li W, Zhang L, Bai X, Tian S (2020) Feature refinement and filter network for person re-identification. IEEE Trans Circuits Syst Video Technol. https://doi.org/10.1109/tcsvt.2020.3043026
Ning X, Duan PF, Li WJ, Zhang SL (2020) Real-time 3D face alignment using an encoder-decoder network with an efficient deconvolution layer. IEEE Signal Proc Let 27:1944–1948. https://doi.org/10.1109/Lsp.2020.3032277
Bochkovskiy A, Wang C-Y, Mark Liao H-Y (2020) Yolov4: optimal speed and accuracy of object detection. Computer vision and pattern recognition
Ouyang WL, Wang XG, Zeng XY, Qiu S, Luo P, Tian YL et al (2015) DeepID-Net: deformable deep convolutional neural networks for object detection. IEEE Conf Comput Vision Pattern Recognition (CVPR) 2015:2403–2412. https://doi.org/10.1109/CVPR.2015.7298854
Shao FM, Wang XQ, Meng FJ, Rui T, Wang D, Tang J (2018) Real-time traffic sign detection and recognition method based on simplified gabor wavelets and CNNs. Sens Basel. https://doi.org/10.3390/s18103192
Shao FM, Wang XQ, Meng FJ, Zhu JW, Wang D, Dai JY (2019) Improved faster R-CNN traffic sign detection based on a second region of interest and highly possible regions proposal network. Sens Basel. https://doi.org/10.3390/s19102288
Zhang J, Huang M, Jin X, Li X (2017) A real-time chinese traffic sign detection algorithm based on modified YOLOv2. Algorithms. https://doi.org/10.3390/a10040127
Li JA, Liang XD, Wei Y, Xu TF, Feng JS, Yan SC (2017) Perceptual generative adversarial networks for small object detection. Proc Cvpr IEEE. https://doi.org/10.1109/Cvpr.2017.211
Liu ZW, Shen C, Qi MY, Fan X (2020) SADANet: integrating scale-aware and domain adaptive for traffic sign detection. IEEE Access 8:77920–77933. https://doi.org/10.1109/Access.2020.2989758
Singh B, Davis LS (2018) An analysis of scale invariance in object detection-SNIP. arXiv:171108189 [csCV]
Yukang Chen YL, Tao Kong, Lu Qi, Ruihang Chu, Lei Li, Jiaya Jia (2021) Scale-aware automatic augmentation for object detection. arXiv:210317220
Luo J-Q, Fang H-S, Shao F-M, Zhong Y, Hua X (2020) Multi-scale traffic vehicle detection based on faster R-CNN with NAS optimization and feature enrichment. Def Technol. https://doi.org/10.1016/j.dt.2020.10.006
Lin TY, Dollar P, Girshick R, He KM, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. Proc CVPR IEEE. https://doi.org/10.1109/Cvpr.2017.106
He KM, Gkioxari G, Dollar P, Girshick R (2017) Mask R-CNN. IEEE I Conf Comp Vis. https://doi.org/10.1109/Iccv.2017.322
Lin TY, Goyal P, Girshick R, He KM, Dollar P (2017) Focal loss for dense object detection. IEEE I Conf Comp Vis. https://doi.org/10.1109/Iccv.2017.324
Cao L, Xiao Y, Xu L (2021) EMface detecting hard faces by exploring receptive field pyraminds. Comput Vision Pattern Recogn
Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L (2009) ImageNet: a large-scale hierarchical image database. CVPR 2009 IEEE Conf Comput Vision Pattern Recogn 14:248–55. https://doi.org/10.1109/cvpr.2009.5206848
Shorten C, Khoshgoftaar TM (2019) A survey on image data augmentation for deep learning. J Big Data. https://doi.org/10.1186/s40537-019-0197-0
Taylor L, Nitschke G (2018) Improving deep learning with generic data augmentation. IEEE Sympos Ser Comput Intell (IEEE Ssci) 2018:1542–1547
Zhang H, Wu QMJ (2011) Pattern recognition by affine legendre moment invariants. IEEE Image Proc 797–800
Lv JJ, Cheng C, Tian GD, Zhou XD, Zhou X (2016) Landmark perturbation-based data augmentation for unconstrained face recognition. Signal Proc Image 47:465–475. https://doi.org/10.1016/j.image.2016.03.011
Nair V, Hinton GE (2010) Rectified linear units improve restricted boltzmann machines. In: International conference on international conference on machine learning omnipress
Dwibedi D, Misra I, Hebert M (2017) Cut, paste and learn: surprisingly easy synthesis for instance detection. IEEE I Conf Comp Vis. https://doi.org/10.1109/Iccv.2017.146
Fang HS, Sun JH, Wang RZ, Gou MH, Li YL, Lu CW (2019) InstaBoost: boosting instance segmentation via probability map guided Copy-pasting. 2019 IEEE CVF Int Conf Comput Vision (ICCV 2019). https://doi.org/10.1109/Iccv.2019.00077
Singh B, Najibi M, Davis LS (2018) SNIPER: efficient multi-scale training. Adv Neur 31
Tran T, Pham T, Carneiro G, Palmer L, Reid I (2017) A bayesian data augmentation approach for learning deep models. Adv Neural Inform Process Syst 30 (Nips 2017). 30
Shi X, Hu J, Lei X, Xu S (2021) Detection of flying birds in airport monitoring based on improved YOLOv5. In: 2021 6th International Conference on Intelligent Computing and Signal Processing (ICSP)2021. p 1446–1451 https://doi.org/10.1109/icsp51882.2021.9408797.
Liu S, Qi L, Qin H, Shi J, Jia J (2018) Path aggregation network for instance segmentation. 2018 IEEE/CVF Conf Comput Vision Pattern Recogn. https://doi.org/10.1109/cvpr.2018.00913
Rezatofighi H, Tsoi N, Gwak J, Sadeghian A, Reid I, Savarese S (2019) Generalized intersection over union: a metric and a loss for bounding box regression. 2019 IEEE/CVF Conf Comput Vision Pattern Recogn (CVPR 2019). https://doi.org/10.1109/Cvpr.2019.00075
He YH, Zhu CC, Wang JR, Savvides M, Zhang XY (2019) Bounding box regression with uncertainty for accurate object detection. 2019 IEEE/Cvf Conf Comput Vision Pattern Recogn (CVPR). https://doi.org/10.1109/Cvpr.2019.00300
Zheng Z, Wang P, Liu W, Li J, Ye R, Ren D (2020) Distance-IoU loss: faster and better learning for bounding box regression. AAAI Conf on Aritif Intell. https://doi.org/10.1609/aaai.v34i07.6999
Kim M, Park C, Kim S, Hong T, Ro WW (2019) Efficient dilated-winograd convolutional neural networks. IEEE Int Conf Image Process (ICIP) 2019:2711–2715
He KM, Zhang XY, Ren SQ, Sun J (2015) Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Anal Mach Intell 37(9):1904–1916. https://doi.org/10.1109/Tpami.2015.2389824
Zoph B, Cubuk ED, Ghiasi G, Lin T-Y, Shlens J, Le QV (2019) Learning data augmentation strategies for object detection. arXiv:190611172 [csCV]
Huang S, Wang X, Tao D (2020) SnapMix: semantically proportional mixing for augmenting fine-grained data
Zhou W, Hao X, Cui J, Yu Y, Cao X, Kuijper A (2021) A self-adaptive learning method for motion blur kernel estimation of the single image. Optik. https://doi.org/10.1016/j.ijleo.2021.168023
Wang Z, Li H, Wu ZX, Wu HL (2021) A pretrained proximal policy optimization algorithm with reward shaping for aircraft guidance to a moving destination in three-dimensional continuous space. Int J Adv Robot Syst. https://doi.org/10.1177/1729881421989546
Zoph B, Vasudevan V, Shlens J, Le QV (2018) Learning transferable architectures for scalable image recognition. IEEE/CVF Conf Comput Vision Pattern Recogn (CVPR) 2018:8697–8710. https://doi.org/10.1109/Cvpr.2018.00907
Zoph B, Shlens J, Le QV (2017) Neural Architecture Search With Reinforcement Learning. arXiv:170707012 [csCV]
Dong Z, Lai CS, Zhang Z, Qi D, Gao M, Duan S (2021) Neuromorphic extreme learning machines with bimodal memristive synapses. Neurocomputing 453:38–49. https://doi.org/10.1016/j.neucom.2021.04.049
Zhu Y, Zhang C, Zhou D, Wang X, Bai X, Liu W (2016) Traffic sign detection and recognition using fully convolutional network guided proposals. Neurocomputing 214:758–766. https://doi.org/10.1016/j.neucom.2016.07.009
Zhu Z, Liang D, Zhang SH, Huang XL, Li BL, Hu SM (2016) Traffic-sign detection and classification in the wild. IEEE Conf Comput Vision Pattern Recogn (CVPR) 2016:2110–2118. https://doi.org/10.1109/Cvpr.2016.232
Zhang J, Xie Z, Sun J, Zou X, Wang J (2020) A Cascaded R-CNN with multiscale attention and imbalanced samples for traffic sign detection. IEEE Access 8:29742–29754. https://doi.org/10.1109/access.2020.2972338
YOLOv5-Lite (2021) Available from: https://github.com/ppogg/YOLOv5-Lite
Tan M, Pang R, Le QV (2020) EfficientDet: scalable and efficient object detection. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp 10781–10790
Qi D, Tan W, Yao Q, Liu J (2021) YOLO5Face: why reinventing a face detector
Zhang Q, Sheng T, Wang Y, Tang Z, Chen Y, Cai L, Ling H (2019) M2Det a single-shot object detector based on multi-level feature pyramid network. Proc AAAI Conf Artif Intell 33:9259–66
Redmon J, Farhadi A (2018) YOLOv3: an incremental improvement. arXiv preprint arXiv:180402767
Vaquero L, Brea VM, Mucientes M (2022) Tracking more than 100 arbitrary objects at 25 FPS through deep learning. Pattern Recogn. https://doi.org/10.1016/j.patcog.2021.108205
Dollar P, Wojek C, Schiele B, Perona P (2012) Pedestrian detection: an evaluation of the state of the art. IEEE Trans Pattern Anal Mach Intell 34(4):743–761. https://doi.org/10.1109/Tpami.2011.155
Acknowledgments
The authors would like to thank the editorial board and reviewers for the improvement of this paper.
Funding
This research was funded by the Zhejiang Provincial Key Lab of Equipment Electronics (No. 2019E10009), and the Key Research and Development Program of Zhejiang Province (No. 2020C01110).
Author information
Authors and Affiliations
Contributions
Conceptualization was contributed by Junfan Wang, Mingyu Gao; methodology was contributed by Yi Chen, Zhekang Dong; formal analysis and investigation were contributed by Junfan Wang, Yi Chen; writing—original draft preparation, was contributed by Junfan Wang, Yi Chen; writing—review and editing, was contributed by Junfan Wang; funding acquisition was contributed by Mingyu Gao; resources were contributed by Zhekang Dong; supervision was contributed by Mingyu Gao.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wang, J., Chen, Y., Dong, Z. et al. Improved YOLOv5 network for real-time multi-scale traffic sign detection. Neural Comput & Applic 35, 7853–7865 (2023). https://doi.org/10.1007/s00521-022-08077-5
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-022-08077-5