Abstract
Object detection plays a vital role in autonomous driving systems, and the accurate detection of surrounding objects can ensure the safe driving of vehicles. This paper proposes a category-assisted transformer object detector called DetectFormer for autonomous driving. The proposed object detector can achieve better accuracy compared with the baseline. Specifically, ClassDecoder is assisted by proposal categories and global information from the Global Extract Encoder (GEE) to improve the category sensitivity and detection performance. This fits the distribution of object categories in specific scene backgrounds and the connection between objects and the image context. Data augmentation is used to improve robustness and attention mechanism added in backbone network to extract channel-wise spatial features and direction information. The results obtained by benchmark experiment reveal that the proposed method can achieve higher real-time detection performance in traffic scenes compared with RetinaNet and FCOS. The proposed method achieved a detection performance of 97.6% and 91.4% in AP50 and AP75 on the BCTSDB dataset, respectively.
Keywords: autonomous driving, deep learning, object detection, transformer
1. Introduction
Vision-based object detection in traffic scenes plays a crucial role in autonomous driving systems. With the rapid development of autonomous driving, the performance of object detection has made significant progress. The traffic object (e.g., traffic signs, vehicles, and pedestrians) can be detected automatically by extracting the features. The result of perceiving the traffic scenario can ensure the safety of the autonomous vehicle. This kind of method can be divided into anchor-based and anchor-free.
Deep-learning-based object detection can be divided into single-stage and multi-stage object detection. The multi-stage algorithms extract the region of interest first, and then the location of the object is determined in these candidate areas. The single-stage algorithm’s output the location and category with dense bounding boxes directly on the original image. These detection algorithms classify each anchor box or key point and detect different categories independently, while ignoring the relationships between categories. There exists a specific relationship between other objects, such as probability, location, and scale of different objects in a particular environment, which is essential for object detection and can improve object detection accuracy.
This relationship between categories exists in many cases in traffic scenarios. For example, pedestrians appearing in highway scenes and vehicles appearing on the pedestrian path are low-probability events, which indicates the connection between object categories and scenarios. Secondly, the signs “Passing” and “No Passing” should not appear in the same scene, which indicates the connection between different object categories. There exist specific implicit relationships between object categories and the background of traffic scenes. Existing object detection methods do not consider this relationship in scenes, and their classification subnetwork is trained to independently classify different objects as individuals without the objects knowing each other, which results in the model underperforming in terms of fitting the distribution of objects and the scene background. Additionally, the model does not thoroughly learn the features required by the detection task and will cause a gap in the classification confidence between categories, which influences the detection performance.
Based on the above-mentioned assumptions, this paper proposes a category-assisted transformer object detector to learn the relationships between different objects called DetectFormer, based on the single-stage method. The motivation of this study was to allow the classification subnetwork to fit better the distribution of object categories with specific scene backgrounds and ensure that the network model is more focused on this relationship.
Transformer [1] is widely used in natural language processing, machine translation, and computer vision because of its ability to perceive global information. Specifically, the vision transformer (ViT) [2] and DETR [3] have been proposed and applied to computer vision. Previous studies have used transformers to capture global feature information and reallocate network attention to features, which is called self-attention. In this study, DetectFormer was built based on the transformer concept. Still the inputs and structure of the multi-head attention mechanism are different because the purpose of DetectFormer is to improve the detection accuracy with the assistance of category information.
The contributions of this study are as follows:
-
(1)
The Global Extract Encoder (GEE) is proposed to extract the global information of the image features output by the backbone network, enhancing the model’s global perception ability.
-
(2)
A novel category-assisted transformer called ClassDecoder is proposed. It can learn the object category relationships and improve the model’s sensitivity by implicitly learning the relationships between objects.
-
(3)
The attention mechanism is added to the backbone network to capture cross-channel, direction-aware and position-sensitive information during feature extraction.
-
(4)
Efficient data augmentation methods are proposed to enhance the diversity of the dataset and improve the robustness of model detection.
The rest of this paper is organized as follows. In Section 2, we introduce object detection algorithms and transformer structure. Details of the proposed DetectFormer are presented in Section 3. In Section 4, the model’s implementation is discussed, and the model is compared with previous methods. The conclusions and direction of future work are discussed in Section 5.
2. Related Work
2.1. Object Detection
Traditional object detection uses HOG [4] or DPM [5] to extract the image features, and then feed them into a classifier such as SVM [6]. Chen et al. [7] use SVM for traffic light detection. In recent years, deep learning based object detection algorithms have achieved better performance in terms of accuracy compared with traditional methods and have become a research hotspot. Generally, there are two types of object detection based on deep convolutional networks: (1) multi-stage detection, such as R-CNN series [8,9,10], and Cascade R-CNN [11]; (2) one-stage detection, which is also known as the dense detector and can be divided into anchor-based methods (for example, the You Only Look Once series [12,13,14] and RetinaNet [15]) and anchor-free methods (for example, FCOS [16], CenterNet [17], and CornerNet [18]). Multi-stage detection methods extract features of the foreground area using region proposal algorithms from preset dense candidates in the first stage. The bounding boxes of objects are regressed in the subsequent steps. The limitation of this structure is that it reduces the detection speed and cannot satisfy the real-time requirements of autonomous driving tasks. Single-stage detection methods directly detect the object and regress the bounding boxes different from multi-stage methods, which can avoid the repeated calculation of the feature map and obtains the anchor boxes directly on the feature map. He et al. [19] proposed a detection method using CapsNet [20] based on visual inspection of traffic scenes. Li et al. [21] proposed improved Faster R-CNN for multi-object detection in a complex traffic environments. Lian et al. [22] proposed attention fusion for small traffic object detection. Liang et al. [23] proposed a light-weight anchor-free detector for traffic scene object detection. However, their models cannot capture global information limited by the size of the receptive field. The above-mentioned approaches obtain local information when extracting image features, and enlarge the receptive field by increasing the size of the convolution kernel or stacking the number of convolution layers. In recent years, transformers have been introduced as new attention-based building blocks applied to computer vision, they have achieved superior performance because they can obtain the global information of the image without increasing the receptive field.
2.2. Transformers Structure
The transformer is a new encoder–decoder architecture introduced by Vaswani et al. [1] first used in machine translation and has better performance than LSTM [24], GRU [25], RNNs [26] (MoE [27], GNMT [28]) in translation tasks. Transformer extracts features by aggregating global information, making it suited for long sequence prediction tasks and other information-heavy tasks, which has better performance than other RNN-based models in natural language processing [29,30], speech processing [31], transfer learning [32]. It is comparable to the performance of CNN in computer vision as a new framework. Alexey et al. [2] proposed a vision transformer, which applied a transformer to computer vision and image classification tasks. Nicolas et al. [3] proposed DETR, which applied a transformer to object detection task. Yan et al. use a transformer to predict long-term traffic flow [33]. Cai et al. [34] use a transformer to capture the spatial dependency for continuity and periodicity time series.
Although the transformer structure shows strong performance, the training based on the transformer takes a long time, and requires a large amount of data sets and ideal pre-training. This paper proposes a learnable object relationship module based on a transformer with self-attention, and a single-stage detector was designed to complete the task of traffic scene object detection. Compared with other methods, the proposed method achieves better detection performance in a shorter training time.
3. Proposed Method
The overall pipeline of our proposed method is shown in Figure 1. The main contributions of the proposed method are the following three parts: (1) attention mechanism in backbone network based on position information; (2) the Global Extract Encoder can enhance the model’s global perception ability; (3) a novel learnable object relationship module called ClassDecoder. Finally, efficient data augmentation was used to improve the robustness of the model.
3.1. Global Extract Encoder
The convolutional neural network is usually affected by the kernel size, network depth, and other factors, causing the receptive field cannot cover the whole area of the image, which is challenging to learn the relationship between long-distant regions or pixels. When extracting the features of the object, the network cannot obtain global information.
Inspired by the transformer architecture and the vision transformer, this study designed the Global Extract Encoder (GEE) to enhance the model’s global perception ability. As shown in Figure 1, the GEE accepts the image features extracted from the backbone network, performs global information perception on , and sends to the following Decoder for object detection. The typical values used in this study are and , where are the height and width of the original image . The structure of GEE is shown in Figure 2 and consists of two primary modules. The first module is the multi-head self-attention layer, and the second one is the feedforward network (FFN). Residual connections are used between each sub-layer.
We split the feature maps into patches, and collapsed the spatial dimensions of from to a one dimension sequence . Then, a fixed position embedding is added to the feature sequence owing to permutation invariance and fed into GEE. The obtained information from different subspaces and positions by adding multi-head self-attention .
(1) |
(2) |
(3) |
where projection matrix Additionally, , and n donates the number of heads. The feedforward network (FFN) enables GEE the ability of nonlinear fitting. After global feature extraction, expands the spatial dimension into . Thus, the dimensions of the GEE module output are consistent with the input dimensions, and the model can obtain long distance regional relationships and global information rather than local information when extracting object features.
3.2. Class Decoder
To learn the object category relationships and improve the model’s sensitivity to the categories by implicitly learning the relationships between objects, a novel learnable object relationship module called ClassDecoder is proposed. The structure of ClassDecoder is shown in Figure 3 and is similar to the transformer architecture. However, this study disregarded the self-attention mechanism, the core of transformer blocks, and designed a module from the perspective of object categories to implicitly learn the relationship between categories, including the foreground and background. Here, 1 × 1 convolution was used to reduce the channel dimension of the global feature map from C to a smaller dimension m, and the spatial dimensions were collapsed to create a new feature sequence .
(4) |
where the means 1 × 1 convolutional operation to reduce the channel dimension of , and F(.) means collapse operator, which transforms two-dimensional feature matrices into feature sequences.
ClassDecoder block requires two inputs: the feature sequence G and the proposal categories P. The proposed ClassDecoder is to detect different categories of objects, using proposal categories to predict the confidence vector of each category, and the depth n of ClassDecoder represents the number of categories. Then, the convolution operation is used to generate the global descriptor of each vector. Finally, the softmax function is used to output the prediction result of the category.
(5) |
(6) |
where the global information G (), the proposal categories P (), and is the same as the first dimension of . In this study, the dimensions of and were set to be the same and equal to the feature channels ; denotes various learnable sequences that are referred to as proposal categories and are independently decoded into class labels, resulting in final class predictions, where denotes the total number of dataset categories in anchor-free methods and is the product of the number of categories and number of anchor boxes in anchor-based methods.
There are many ways to initialize the proposal categories. Transformer architecture does not contain any inductive bias; this study attempted to feed prior knowledge into ClassDecoder, and proposal categories were initialized as follows. A 1 × 1 convolution was used to reduce the dimension of and reduce the original dimension to the dimension (generally, ), where represents the total number of categories in the dataset of the detection task based on the anchor-free method. ClassDecoder globally reasons about all categories simultaneously using the pair-wise relationships between objects while learning the relationship between categories, including the foreground and background.
3.3. Attention Mechanism in the Backbone Network
The attention mechanisms in computer vision can enhance the objects in the feature maps. CBAM [35] attempts to utilize position information by reducing the channel dimension of the input tensor and using convolution to compute spatial attention. Different from CBAM, our proposed method adds a location attention feature to build the direction-aware information, which can improve the network more accurately locate objects, by capturing precise location information in two different spatial directions. A global encoding for channel-wise spatial information is added based on Coordinate Attention [36]. Specifically, the features are aggregated along W and H spatial directions to obtain feature maps of perception in two directions. These two features and allow the attention module to obtain long-term dependencies along with different spatial directions. The concatenate operation F is performed with the channel descriptor with global spatial information. Then, the convolution function is used to transform them and obtain the output , as shown in Figure 4.
, and are defined as follows:
(7) |
(8) |
(9) |
(10) |
where is the input from the features extracted from the previous layer associated with the c-channel, is the convolutional operation, and is concatenate operation. After the output of different information through their respective convolution layer , the normalization is activated by sigmoid activation function . The final output is the multiply of the original feature map and information weights.
(11) |
(12) |
(13) |
(14) |
The proposed attention mechanism in the backbone could be applied to different kinds of networks. As shown in the following experimental part, the improved attention mechanism can be plugged into lightweight backbone networks and improve the network detection capability.
3.4. Data Augmentation
Traffic scene object detection is usually affected by light, weather, and other factors. The data-driven deep neural networks require a large number of labeled images to train the model. Most traffic scene datasets cannot cover all complex environmental conditions. In this paper, we use three types of data augmentation methods global pixel level, spatial level, and object level, as shown in Figure 5. Specifically, we use Brightness Contrast, Blur, and Channel Dropout for illumination transformation; we use Rain, Sun Flare, and Cutout [37] for the spatial level data augmentation, Mixup, CutMix [38] for the object level augmentation. The data augmented by these methods can simulate complex traffic scenarios, which can improve the detection robustness of the model.
4. Experience and Results
4.1. Evaluation Metrics
The average precision (AP) metrics were used to evaluate the detection performance, including AP at different IoU thresholds (AP, AP50, AP75) and AP for different scale objects (APS, APM, APL), which consider both recall and precision. The top-n accuracy was used to evaluate the classification ability of different methods. Top-n represents the truth value of the object in the first n confidence results of the model. We also use parameters and FLOPs (floating-point operations per second) to measure the volume and computation of different models.
4.2. Datasets
Detection performance in traffic scenes is evaluated using the BCTSDB [39], KITTI [40], and COCO [41] datasets to evaluate the generalization ability. The KITTI dataset contains 7481 training images and 7518 test images, totaling 80,256 labeled objects with three categories (e.g., vehicle, pedestrian, and cyclist). The BCTSDB dataset contains 15,690 traffic sign images, including 25,243 labeled traffic signs. The COCO dataset is used to test the generalization ability of the model including 80 object categories and more than 220 K labeled images.
4.3. Implementation and Training Details
The network structure constructed by PyTorch and the default hyperparameters used were the same as those for MMDetection [42] unless otherwise stated. Two NVIDIA TITAN V graphics cards with 24 GB VRAM were used to train the model. The linear warming up policy was used to start the training, where the warm-up ratio was set to 0.1. The optimizer of DetectFormer is AdamW [43]; the initial learning rate is set to 10−4, and the weight decay is set to 10−4. The backbone network is established using pre-trained weights from ImageNet [44], and other layers used Xavier [45] for parameter initialization except for the proposal categories. The input images are scaled to a full scale of 640 × 640, while maintaining the aspect ratio.
4.4. Performances
We first evaluate the effectiveness of the different proposed units. The ClassDecoder head, Global Extract Encoder, Attention, Anchor-free head, and Data augmentation are gradually added to the RetinaNet baseline on COCO and BCTSDB dataset to test the generalization ability of the proposed method and the detection ability in the traffic scene, as shown in Table 1 and Table 2, respectively.
Table 1.
Methods | Parameters (M) | FLOPs (G) | AP (%) | AP50 (%) | AP75 (%) |
---|---|---|---|---|---|
RetinaNet baseline | 37.74 | 95.66 | 32.5 | 50.9 | 34.8 |
+ClassDecoder | 35.03 (−2.71) | 70.30 (−25.36) | 34.6 (+2.1) | 53.5 (+2.6) | 36.1 (+1.3) |
+Global Extract Encoder | 36.95 (+1.92) | 90.45 (+20.15) | 36.2 (+1.6) | 55.7 (+2.2) | 37.8 (+1.7) |
+Attention | 37.45 (+0.5) | 90.65 (+0.2) | 38.3 (+2.1) | 58.3 (+2.6) | 39.3 (+1.5) |
+Anchor-free | 37.31 (−0.14) | 89.95 (−0.7) | 38.9 (+0.6) | 59.1 (+0.8) | 39.6 (+0.3) |
+Data augmentation | 37.31 (+0) | 89.95 (+0) | 41.3 (+2.4) | 61.8 (+2.7) | 41.5 (+1.9) |
Table 2.
Methods | Parameters (M) | FLOPs (G) | AP (%) | AP50 (%) | AP75 (%) |
---|---|---|---|---|---|
RetinaNet baseline | 37.74 | 95.66 | 59.7 | 89.4 | 71.2 |
+ClassDecoder | 35.03 (−2.71) | 70.30 (−25.36) | 61.6 (+3.7) | 91.8 (+2.4) | 75.8 (+4.6) |
+Global Extract Encoder | 36.95 (+1.92) | 90.45 (+20.15) | 63.4 (+3.4) | 93.9 (+2.1) | 80.6 (+4.8) |
+Attention | 37.45 (+0.5) | 90.65 (+0.2) | 65.2 (+3.1) | 95.1 (+1.2) | 84.2 (+3.6) |
+Anchor-free | 37.31 (−0.14) | 89.95 (−0.7) | 65.8 (+2.1) | 95.7 (+0.6) | 87.4 (+3.2) |
+Data augmentation | 37.31 (+0) | 89.95 (+0) | 76.1 (+4.1) | 97.6 (+1.9) | 91.4 (+4.0) |
We further compare the different performances of anchor-based and anchor-free methods on KITTI dataset. As shown in Table 3, the detection performance of an anchor-free detector with Feature Pyramid Network (FPN) [46] is better than the anchor-based detector. FPN plays a crucial role in improving detection accuracy based on the anchor-free method.
Table 3.
Methods | Detector | Car (%) | Pedestrian (%) | Cyclist (%) |
---|---|---|---|---|
DetectFormer | Anchor-based | 83.24 | 70.11 | 73.54 |
Anchor-free | 69.45 | 61.15 | 62.24 | |
Anchor-free w/. FPN | 86.59 | 79.45 | 81.71 |
For the initialization method of proposal categories, we compare different methods, as shown in Figure 6. The experiment shows that the orthogonalized initial parameter method better than the random initialization method in the early stage of training. The advantage becomes less obvious as the training continue.
The efficiency of attention and detection results of DetectFormer with different number of parameter backbone networks, from light-weight backbone network (MobileNetv3 [47]) to high-performance backbone network (ResNet101 [48]) are shown in Table 4, which shows that it can improve the detection performance of the model by inserting attention mechanism into the backbone network, especially in the lightweight backbone network, our method is competitive in lightweight networks.
Table 4.
Backbone | Params. | FLOPs | Head | Attention | AP (%) |
---|---|---|---|---|---|
MobileNetv3 × 1.0 | 5.4 M | 220 M | RetinaNet | w/o. | 51.2 |
ResNet50 | 25 M | 3.8 G | 59.7 | ||
ResNet101 | 46.3 M | 7.6 G | 64.8 | ||
MobileNetv3 × 1.0 | 5.9 M | 231 M | RetinaNet | w/. | 54.1 |
ResNet50 | 25 M | 3.8 G | 62.5 | ||
ResNet101 | 46.3 M | 7.6 G | 66.3 |
Table 5 presents the classification performance of baseline methods and that of the proposed method on the BCTSDB dataset. Anchor-based and anchor-free methods were used to compare RetinaNet and FCOS, respectively. The experimental results reveal that DetectFormer is helpful in improving the classification ability of the model. Remarkably, DetectFormer can reduce the computation and parameter number of the detection networks.
Table 5.
The convergence curves among the DetectFormer and other SOTA (state-of-the-art) methods, including RetinaNet, DETR, Faster R-CNN, FCOS, and YOLOv5, are shown in Figure 7, which illustrates that DetectFormer achieves better performance with efficient training and accurate detection. The vertical axis is the detection accuracy.
Table 6 shows the detection results on BCTSDB dataset produced by multi-stage methods (e.g., Faster R-CNN, Cascade R-CNN) and single-stage methods, including anchor-based methods (e.g., YOLOv3, RetinaNet) and the anchor-free method FCOS. DetectFormer shows high detection accuracy and more competitive performance. The AP, AP50, and AP75 are 76.1%, 97.6%, and 84.3%, respectively. DetectFormer can suit the distribution of object categories and boost detection confidence in the field of autonomous driving better than other networks.
Table 6.
Model | Backbone | Head | AP | AP50 | AP75 | APS | APM | APL | FPS |
---|---|---|---|---|---|---|---|---|---|
Faster R-CNN [10] | ResNet50 | Anchor-based | 70.2 | 94.7 | 86.0 | 65.3 | 76.5 | 84.5 | 28 |
Cascade R-CNN [11] | ResNet50 | Anchor-based | 75.8 | 96.7 | 92.5 | 72.9 | 79.3 | 89.2 | 23 |
YOLOv3 [14] | Darknet53 | Anchor-based | 59.5 | 92.7 | 70.4 | 54.2 | 70.1 | 83.8 | 56 |
RetinaNet [15] | ResNet50 | Anchor-based | 59.7 | 89.4 | 71.2 | 47.2 | 72.5 | 83.3 | 52 |
FCOS [16] | ResNet50 | Anchor-free | 68.6 | 95.8 | 83.9 | 62.7 | 75.7 | 83.9 | 61 |
Ours. | ResNet50 | Anchor-free | 76.1 | 97.6 | 91.4 | 63.1 | 77.4 | 84.5 | 60 |
The proposed method was also evaluated on the KITTI dataset. As shown in Table 7, compared with other methods, DetectFormer shows better detection results.
Table 7.
Methods | Car | Pedestrian | Cyclist | Time (ms) |
||||||
---|---|---|---|---|---|---|---|---|---|---|
Easy (%) | Moderate (%) | Hard (%) | Easy (%) | Moderate (%) | Hard (%) | Easy (%) | Moderate (%) | Hard (%) | ||
Regionlets [49] | 84.75 | 76.45 | 59.70 | 73.14 | 61.15 | 55.21 | 70.41 | 58.72 | 51.83 | - |
Faster R-CNN [10] | 87.97 | 79.11 | 70.62 | 78.97 | 65.24 | 60.09 | 71.40 | 61.86 | 53.97 | 142 |
Mono3D [50] | 84.52 | 89.37 | 79.15 | 80.30 | 67.29 | 62.23 | 77.19 | 65.15 | 57.88 | - |
MS-CNN [51] | 93.87 | 88.68 | 76.11 | 85.71 | 74.89 | 68.99 | 84.88 | 75.30 | 65.27 | - |
SSD [52] | 87.34 | 87.74 | 77.27 | 50.38 | 48.41 | 43.46 | 48.25 | 52.31 | 52.13 | 30 |
ASSD [53] | 89.28 | 89.95 | 82.11 | 69.07 | 62.49 | 60.18 | 75.23 | 76.16 | 72.83 | 30 |
RFBNet [54] | 87.31 | 87.27 | 84.44 | 66.16 | 61.77 | 58.04 | 74.89 | 72.05 | 71.01 | 23 |
Ours. | 90.48 | 88.03 | 81.25 | 83.32 | 79.35 | 75.67 | 85.04 | 82.33 | 77.76 | 22 |
Figure 8 and Figure 9 and show that DetectFormer can improve the model’s sensitivity to categories by implicitly learning the relationships between objects.
The detection results are shown in Figure 10 and Figure 11 on the KITTI and BCTSDB datasets, respectively. The results demonstrate the proposed method’s effectiveness in traffic scenarios. Three types of traffic signs on the BCTSDB dataset, including warning, prohibitory, mandatory, and three types of traffic objects on the KITTI dataset, including car, pedestrian, cyclist were detected. The detection result does not include other types of traffic objects such as a motorcycle in Figure 10, but the proposed model can detect those kinds of objects.
5. Discussion
Why can ClassDecoder improve the classification ability of models? In this paper, we propose ClassDecoder to improve the classification ability, which is designed based on the transformer architecture without any convolution operations. The model interacts with different background feature maps in scaled dot-product attention and multi-head attention by using proposal categories, and learns the implicit relationship between the background and the category by using the key-value pair idea in the Transformer. The number of proposal categories is equal to the number of object categories, and the parameters of proposal categories are learnable. The input of ClassDecoder is the feature maps, and proposal categories, and the output is the prediction category of the current bounding box. The output dimensions are the same as those of the proposal categories, and the proposal categories are associated with the output in the role of Query (Query-Key-Value relationship in transformer architecture). It can be understood that the proposal categories are vectors that can be learned, and their quantity represents the confidence vectors corresponding to different categories of the current bounding box. Then, the model converts the confidence vector into category confidence through feed-forward network. The category with the highest confidence is the category of the predicted bounding box.
6. Conclusions
This paper proposes a novel object detector called DetectFormer, which is assisted by a transformer to learn the relationship between objects in traffic scenes. By introducing the GEE and ClassDecoder, this study focused on fitting the distribution of object categories to specific scene backgrounds and implicitly learning the object category relationships to improve the sensitivity of the model to the categories. The results obtained by experiments on the KITTI and BCTSDB datasets reveal that the proposed method can improve the classification ability and achieve outstanding performance in complex traffic scenes. The AP50 and AP75 of the proposed method are 97.6% and 91.4% on BCTSDB, and the average accuracies of car, pedestrian, and cyclist are 86.6%, 79.5%, and 81.7% on KITTI, respectively, which indicates that the proposed method achieves better results compared to other methods. The proposed method improved detection accuracy, but it still encountered many challenges when applied to natural traffic scenarios. The experiment in this paper is trained on public datasets and real traffic scenes facing challenges with complex lighting and weather factors. Our future work is focused on object detection in an open environment and the deployment of models to vehicles.
Abbreviations
AP | Averaged AP at IoUs from 0.5 to 0.95 with an interval of 0.05 |
AP50 | AP at IoU threshold 0.5 |
AP75 | AP at IoU threshold 0.75 |
APL | AP for objects of large scales (area > 962) |
APM | AP for objects of medium scales (322 < area < 962) |
APS | AP for objects of small scales (area < 322) |
BCTSDB | BUU Chinese Traffic Sign Detection Benchmark |
CNN | Convolutional Neural Network |
FLOPs | Floating-point operations per second |
FPN | Feature pyramid network |
FPS | Frames Per Second |
GEE | Global Extract Encoder |
HOG | Histogram of Oriented Gradients |
IoU | Intersection over union |
LSTM | Long Short-Term Memory |
NMS | Non-Maximum Suppression |
RNN | Recurrent Neural Network |
SOTA | State-of-the-art |
SSD | Single Shot MultiBox Detector |
SVM | Support Vector Machine |
VRAM | Video random access memory |
YOLO | You Only Look Once |
Author Contributions
Conceptualization, T.L. and W.P.; methodology, H.B.; software, T.L. and W.P.; validation, X.F. and H.L.; writing—original draft preparation, T.L.; writing—review and editing, T.L. and W.P. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Conflicts of Interest
The authors declare no conflict of interest.
Funding Statement
This research was funded by the National Natural Science Foundation of China (Nos. 61802019, 61932012, 61871039, 61906017 and 62006020) and the Beijing Municipal Education Commission Science and Technology Program (Nos. KM201911417003, KM201911417009 and KM201911417001). Beijing Union University Research and Innovation Projects for Postgraduates (No.YZ2020K001). By the Premium Funding Project for Academic Human Resources Development in Beijing Union University under Grant BPHR2020DZ02.
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Vaswani A., Shazeer N., Parmar N., Uszkoreit J., Jones L., Gomez A.N., Kaiser Ł., Polosukhin I. Attention is all you need; Proceedings of the 31st International Conference on Neural Information Processing Systems; Long Beach, CA, USA. 4–9 December 2017; Long Beach, CA, USA: Curran Associates Inc.; 2017. pp. 6000–6010. [Google Scholar]
- 2.Dosovitskiy A., Beyer L., Kolesnikov A., Weissenborn D., Zhai X., Unterthiner T., Dehghani M., Minderer M., Heigold G., Gelly S., et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv. 20212010.11929 [Google Scholar]
- 3.Carion N., Massa F., Synnaeve G., Usunier N., Kirillov A., Zagoruyko S. End-to-End Object Detection with Transformers. In: Vedaldi A., Bischof H., Brox T., Frahm J.-M., editors. Computer Vision—ECCV 2020. Volume 12346. Springer International Publishing; Cham, Switzerland: 2020. pp. 213–229. Lecture Notes in Computer Science. [Google Scholar]
- 4.Dalal N., Triggs B. Histograms of Oriented Gradients for Human Detection; Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05); San Diego, CA, USA. 20–25 June 2005; San Diego, CA, USA: IEEE; 2005. pp. 886–893. [Google Scholar]
- 5.Felzenszwalb P.F., Girshick R.B., McAllester D., Ramanan D. Object Detection with Discriminatively Trained Part-Based Models. IEEE Trans. Pattern Anal. Mach. Intell. 2010;32:1627–1645. doi: 10.1109/TPAMI.2009.167. [DOI] [PubMed] [Google Scholar]
- 6.Hearst M.A., Dumais S.T., Osuna E., Platt J., Scholkopf B. Support vector machines. IEEE Intell. Syst. Their Appl. 1998;13:18–28. doi: 10.1109/5254.708428. [DOI] [Google Scholar]
- 7.Chen Z., Shi Q., Huang X. Automatic detection of traffic lights using support vector machine; Proceedings of the 2015 IEEE Intelligent Vehicles Symposium (IV); Seoul, Korea. 28 June–1 July 2015; Seoul, Korea: IEEE; 2015. pp. 37–40. [Google Scholar]
- 8.Girshick R., Donahue J., Darrell T., Malik J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation; Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition; Columbus, OH, USA. 23–28 June 2014; Columbus, OH, USA: IEEE; 2014. pp. 580–587. [Google Scholar]
- 9.Girshick R. Fast R-CNN; Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV); Santiago, Chile. 7–13 December 2015; Santiago, Chile: IEEE; 2015. pp. 1440–1448. [Google Scholar]
- 10.Ren S., He K., Girshick R., Sun J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017;39:1137–1149. doi: 10.1109/TPAMI.2016.2577031. [DOI] [PubMed] [Google Scholar]
- 11.Cai Z., Vasconcelos N. Cascade R-CNN: Delving Into High Quality Object Detection; Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition; Salt Lake City, UT, USA. 18–23 June 2018; Salt Lake City, UT, USA: IEEE; 2018. pp. 6154–6162. [Google Scholar]
- 12.Redmon J., Divvala S., Girshick R., Farhadi A. You Only Look Once: Unified, Real-Time Object Detection; Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Las Vegas, NV, USA. 27–30 June 2016; Las Vegas, NV, USA: IEEE; 2016. pp. 779–788. [Google Scholar]
- 13.Redmon J., Farhadi A. YOLO9000: Better, Faster, Stronger; Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Honolulu, HI, USA. 21–26 July 2017; Honolulu, HI, USA: IEEE; 2017. pp. 6517–6525. [Google Scholar]
- 14.Redmon J., Farhadi A. YOLOv3: An Incremental Improvement. arXiv. 20181804.02767 [Google Scholar]
- 15.Lin T.-Y., Goyal P., Girshick R., He K., Dollar P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020;42:318–327. doi: 10.1109/TPAMI.2018.2858826. [DOI] [PubMed] [Google Scholar]
- 16.Tian Z., Shen C., Chen H., He T. FCOS: A Simple and Strong Anchor-free Object Detector. IEEE Trans. Pattern Anal. Mach. Intell. 2020;44:1922–1933. doi: 10.1109/TPAMI.2020.3032166. [DOI] [PubMed] [Google Scholar]
- 17.Zhou X., Wang D., Krähenbühl P. Objects as Points. arXiv. 20191904.07850 [Google Scholar]
- 18.Law H., Deng J. CornerNet: Detecting Objects as Paired Keypoints. Int. J. Comput. Vis. 2020;128:642–656. doi: 10.1007/s11263-019-01204-1. [DOI] [Google Scholar]
- 19.He S., Chen L., Zhang S., Guo Z., Sun P., Liu H., Liu H. Automatic Recognition of Traffic Signs Based on Visual Inspection. IEEE Access. 2021;9:43253–43261. doi: 10.1109/ACCESS.2021.3059052. [DOI] [Google Scholar]
- 20.Sabour S., Frosst N., Hinton G.E. Dynamic Routing Between Capsules; Proceedings of the NIPS; Long Beach, CA, USA. 4–9 December 2017; pp. 3859–3869. [Google Scholar]
- 21.Li C., Qu Z., Wang S., Liu L. A method of cross-layer fusion multi-object detection and recognition based on improved faster R-CNN model in complex traffic environment. Pattern Recognit. Lett. 2021;145:127–134. doi: 10.1016/j.patrec.2021.02.003. [DOI] [Google Scholar]
- 22.Lian J., Yin Y., Li L., Wang Z., Zhou Y. Small Object Detection in Traffic Scenes Based on Attention Feature Fusion. Sensors. 2021;21:3031. doi: 10.3390/s21093031. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Liang T., Bao H., Pan W., Pan F. ALODAD: An Anchor-Free Lightweight Object Detector for Autonomous Driving. IEEE Access. 2022;10:40701–40714. doi: 10.1109/ACCESS.2022.3166923. [DOI] [Google Scholar]
- 24.Graves A. Supervised Sequence Labelling with Recurrent Neural Networks. Volume 385. Springer; Berlin/Heidelberg, Germany: 2012. Long Short-Term Memory; pp. 37–45. Studies in Computational Intelligence. [Google Scholar]
- 25.Cho K., van Merrienboer B., Gulcehre C., Bahdanau D., Bougares F., Schwenk H., Bengio Y. Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. arXiv. 2014 doi: 10.48550/ARXIV.1406.1078.1406.1078 [DOI] [Google Scholar]
- 26.Zaremba W., Sutskever I., Vinyals O. Recurrent Neural Network Regularization. arXiv. 2014 doi: 10.48550/ARXIV.1409.2329.1409.2329 [DOI] [Google Scholar]
- 27.Shazeer N., Mirhoseini A., Maziarz K., Davis A., Le Q., Hinton G., Dean J. Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer. arXiv. 2017 doi: 10.48550/ARXIV.1701.06538.1701.06538 [DOI] [Google Scholar]
- 28.Wu Y., Schuster M., Chen Z., Le Q.V., Norouzi M., Macherey W., Krikun M., Cao Y., Gao Q., Macherey K., et al. Google’s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation. arXiv. 2016 doi: 10.48550/ARXIV.1609.08144.1609.08144 [DOI] [Google Scholar]
- 29.Wang H., Wu Z., Liu Z., Cai H., Zhu L., Gan C., Han S. HAT: Hardware-Aware Transformers for Efficient Natural Language Processing; Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; Online. 5–10 July 2020; Online: Association for Computational Linguistics; 2020. pp. 7675–7688. [Google Scholar]
- 30.Floridi L., Chiriatti M. GPT-3: Its Nature, Scope, Limits, and Consequences. Minds Mach. 2020;30:681–694. doi: 10.1007/s11023-020-09548-1. [DOI] [Google Scholar]
- 31.Dong L., Xu S., Xu B. Speech-Transformer: A No-Recurrence Sequence-to-Sequence Model for Speech Recognition; Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); Calgary, AB, Canada. 5–20 April 2018; Calgary, AB, Canada: IEEE; 2018. pp. 5884–5888. [Google Scholar]
- 32.Devlin J., Chang M.-W., Lee K., Toutanova K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv. 2016 doi: 10.48550/ARXIV.1810.04805.1810.04805 [DOI] [Google Scholar]
- 33.Yan H., Ma X., Pu Z. Learning Dynamic and Hierarchical Traffic Spatiotemporal Features With Transformer. IEEE Trans. Intell. Transport. Syst. 2021:1–14. doi: 10.1109/TITS.2021.3102983. [DOI] [Google Scholar]
- 34.Cai L., Janowicz K., Mai G., Yan B., Zhu R. Traffic transformer: Capturing the continuity and periodicity of time series for traffic forecasting. Trans. GIS. 2020;24:736–755. doi: 10.1111/tgis.12644. [DOI] [Google Scholar]
- 35.Woo S., Park J., Lee J.-Y., Kweon I.S. CBAM: Convolutional Block Attention Module. In: Ferrari V., Hebert M., Sminchisescu C., Weiss Y., editors. Computer Vision—ECCV 2018. Volume 11211. Springer International Publishing; Cham, Switzerland: 2018. pp. 3–19. Lecture Notes in Computer Science. [Google Scholar]
- 36.Hou Q., Zhou D., Feng J. Coordinate Attention for Efficient Mobile Network Design; Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); Nashville, TN, USA. 20–25 June 2021; Nashville, TN, USA: IEEE; 2021. pp. 13708–13717. [Google Scholar]
- 37.DeVries T., Taylor G.W. Improved Regularization of Convolutional Neural Networks with Cutout. arXiv. 20171708.04552 [Google Scholar]
- 38.Yun S., Han D., Chun S., Oh S.J., Yoo Y., Choe J. CutMix: Regularization Strategy to Train Strong Classifiers With Localizable Features; Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV); Seoul, Korea. 27 October–2 November 2019; Seoul, Korea: IEEE; 2019. pp. 6022–6031. [Google Scholar]
- 39.Liang T., Bao H., Pan W., Pan F. Traffic Sign Detection via Improved Sparse R-CNN for Autonomous Vehicles. J. Adv. Transp. 2022;2022:3825532. doi: 10.1155/2022/3825532. [DOI] [Google Scholar]
- 40.Geiger A., Lenz P., Stiller C., Urtasun R. Vision meets robotics: The KITTI dataset. Int. J. Robot. Res. 2013;32:1231–1237. doi: 10.1177/0278364913491297. [DOI] [Google Scholar]
- 41.Lin T.-Y., Maire M., Belongie S., Hays J., Perona P., Ramanan D., Dollár P., Zitnick C.L. Microsoft COCO: Common Objects in Context. In: Fleet D., Pajdla T., Schiele B., Tuytelaars T., editors. Computer Vision—ECCV 2014. Volume 8693. Springer International Publishing; Cham, Switzerland: 2014. pp. 740–755. Lecture Notes in Computer Science. [Google Scholar]
- 42.Chen K., Wang J., Pang J., Cao Y., Xiong Y., Li X., Sun S., Feng W., Liu Z., Xu J., et al. MMDetection: Open MMLab Detection Toolbox and Benchmark. arXiv. 20191906.07155 [Google Scholar]
- 43.Loshchilov I., Hutter F. Decoupled Weight Decay Regularization. arXiv. 20191711.05101 [Google Scholar]
- 44.Krizhevsky A., Sutskever I., Hinton G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM. 2017;60:84–90. doi: 10.1145/3065386. [DOI] [Google Scholar]
- 45.Glorot X., Bengio Y. Understanding the difficulty of training deep feedforward neural networks; Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics; Sardinia, Italy. 13–15 May 2010; pp. 249–256. JMLR Workshop and Conference Proceedings. [Google Scholar]
- 46.Lin T.-Y., Dollar P., Girshick R., He K., Hariharan B., Belongie S. Feature Pyramid Networks for Object Detection; Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Honolulu, HI, USA. 21–26 July 2017; Honolulu, HI, USA: IEEE; 2017. pp. 936–944. [Google Scholar]
- 47.Koonce B. Convolutional Neural Networks with Swift for Tensorflow. Apress; Berkeley, CA, USA: 2021. MobileNetV3; pp. 125–144. [Google Scholar]
- 48.He K., Zhang X., Ren S., Sun J. Deep Residual Learning for Image Recognition; Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Las Vegas, NV, USA. 27–30 June 2016; Las Vegas, NV, USA: IEEE; 2016. pp. 770–778. [Google Scholar]
- 49.Wang X., Yang M., Zhu S., Lin Y. Regionlets for Generic Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2015;37:2071–2084. doi: 10.1109/TPAMI.2015.2389830. [DOI] [PubMed] [Google Scholar]
- 50.Chen X., Kundu K., Zhang Z., Ma H., Fidler S., Urtasun R. Monocular 3D Object Detection for Autonomous Driving; Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Las Vegas, NV, USA. 27–30 June 2016; Las Vegas, NV, USA: IEEE; 2016. pp. 2147–2156. [Google Scholar]
- 51.Cai Z., Fan Q., Feris R.S., Vasconcelos N. A Unified Multi-scale Deep Convolutional Neural Network for Fast Object Detection. In: Leibe B., Matas J., Sebe N., Welling M., editors. Computer Vision—ECCV 2016. Volume 9908. Springer International Publishing; Cham, Switzerland: 2016. pp. 354–370. Lecture Notes in Computer Science. [Google Scholar]
- 52.Liu W., Anguelov D., Erhan D., Szegedy C., Reed S., Fu C.-Y., Berg A.C. SSD: Single Shot MultiBox Detector. In: Leibe B., Matas J., Sebe N., Welling M., editors. Computer Vision—ECCV 2016. Volume 9905. Springer International Publishing; Cham, Switzerland: 2016. pp. 21–37. Lecture Notes in Computer Science. [Google Scholar]
- 53.Yi J., Wu P., Metaxas D.N. ASSD: Attentive single shot multibox detector. Comput. Vis. Image Underst. 2019;189:102827. doi: 10.1016/j.cviu.2019.102827. [DOI] [Google Scholar]
- 54.Liu S., Huang D., Wang Y. Receptive Field Block Net for Accurate and Fast Object Detection. In: Ferrari V., Hebert M., Sminchisescu C., Weiss Y., editors. Computer Vision—ECCV 2018. Volume 11215. Springer International Publishing; Cham, Switzerland: 2018. pp. 404–419. Lecture Notes in Computer Science. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Not applicable.