An Automatic Car Accident Detection Method Based On Cooperative Vehicle Infrastructure Systems
An Automatic Car Accident Detection Method Based On Cooperative Vehicle Infrastructure Systems
ABSTRACT Car accidents cause a large number of deaths and disabilities every day, a certain proportion
of which result from untimely treatment and secondary accidents. To some extent, automatic car accident
detection can shorten response time of rescue agencies and vehicles around accidents to improve rescue
efficiency and traffic safety level. In this paper, we proposed an automatic car accident detection method
based on Cooperative Vehicle Infrastructure Systems (CVIS) and machine vision. First of all, a novel image
dataset CAD-CVIS is established to improve accuracy of accident detection based on intelligent roadside
devices in CVIS. Especially, CAD-CVIS is consisted of various kinds of accident types, weather conditions
and accident location, which can improve self-adaptability of accident detection methods among different
traffic situations. Secondly, we develop a deep neural network model YOLO-CA based on CAD-CVIS and
deep learning algorithms to detect accident. In the model, we utilize Multi-Scale Feature Fusion (MSFF)
and loss function with dynamic weights to enhance performance of detecting small objects. Finally, our
experiment study evaluates performance of YOLO-CA for detecting car accidents, and the results show that
our proposed method can detect car accident in 0.0461 seconds (21.6FPS) with 90.02% average precision
(AP). In additionally, we compare YOLO-CA with other object detection models, and the results demonstrate
the comprehensive performance improvement on the accuracy and real-time over other models.
INDEX TERMS Car accident detection, CVIS, machine vision, deep learning.
VOLUME 7, 2019 This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see http://creativecommons.org/licenses/by/4.0/ 127453
D. Tian et al.: Automatic Car Accident Detection Method Based on CVIS
deep-learning into methods of car accident detection. These which monitors the change of acceleration to detect accident
methods extract and process complex image features instead and utilize V2X communication technologies to report it.
of single vehicle motion parameter, which improves the accu- To a certain extent, these methods can detect and report car
racy of detecting car accidents. However, the datasets of these accidents in short time, and improve the efficiency of car acci-
methods are mostly captured by car cameras or cell phones of dents warning. However, the vehicle running condition before
pedestrian, which is not suitable for roadside devices in CVIS. car accidents is complex and unpredictable, and the accuracy
In additionally, the reliability and real-time performance of of accident detection only based on speed and acceleration
these methods need to be improved to meet the requirements may be low. In addition, they rely too heavily on vehicu-
of car accident detection. lar monitoring and communication equipment, which may
In this paper, we propose a data-driven car accident detec- be unreliable or damaged in some extreme circumstances,
tion method based on CVIS, whose goal is improving effi- such as heavy canopy, underground tunnel, and serious car
ciency and accuracy of car accident response. With the goal, accidents.
we focus on such a general application scenario when there is
an accident on the road, roadside intelligent devices recognize B. METHOD BASED VIDEO FEATURES
and locate it efficiently. First, we build a novel dataset, Car With the development of machine vision and artificial neural
Accident Detection for Cooperative Vehicle Infrastructure network technology, more and more applications based on
System dataset (CAD-CVIS), which is more suitable for car video processing have been applied in transportation and
accident detection based on roadside intelligent devices in vehicle fields. Under this background, some researchers uti-
CVIS. Then, a deep learning model YOLO-CA based on lized video features of the car accident to detect it. Refer-
CAD-CVIS is developed to detect car accident. Especially, ence [25] presented a Dynamic-Spatial-Attention Recurrent
we optimize the network of traditional deep learning models Neural Network (RNN) for anticipating accidents in dashcam
YOLO [21] to build network of YOLO-CA, which is more videos, which can predict accidents about 2 seconds before
accurate and fast in detecting car accident. In additionally, they occur with 80% recall and 56.14% precision. Refer-
considering of wide shooting scope of roadside cameras in ence [26] proposed a car accident detection system based on
CVIS, multi-scale feature fusion method and loss function first-person videos, which detected anomalies by predicting
with dynamic weights are utilized to improve performance the future locations of car participants and then monitoring
of detecting small objects. the prediction accuracy and consistency metrics. These meth-
The rest of this paper is organized as follows: ods also have some limitations because of low penetration
Section 2 gives an overviews of related work. We present the of vehicular intelligent devices and shielding effects between
details of our proposed method in Section 3. The performance vehicles.
evaluation is discussed in Section 4. Finally, Section 5 con- There are also some other methods which use roadside
clude this paper. devices instead of vehicular equipments to obtain and process
video. Reference [27] proposed a novel accident detection
II. RELATED WORK system at intersection, which composed background images
The car accident detection and notification method is a from image sequence and detected accidents by using Hid-
challenging issue and has attracted a lot of attention from den Markov Model. Reference [28] outlined a novel method
researchers. They have proposed and applied various car for modeling of interaction among multiple moving objects,
accident detection methods. In generally, car accident detec- and used the Motion Interaction Field to detect and localize
tion methods are mainly divided into the following two car accidents. Reference [29] proposed a novel approach
kinds: vehicle running condition-based and accident video for automatic road accident detection, which was based
features-based. on detecting damaged vehicles from footage received from
surveillance cameras installed in roads. In this method, His-
A. METHOD BASED ON VEHICLE RUNNING CONDITION togram of gradients (HOG) and Gray level co-occurrence
When an accident occurs, the motion state of the vehicle matrix features were used to train support vector machines.
will change dramatically. Therefore, many researchers pro- Reference [30] presented a novel dataset for car accidents
posed the accident detection method by monitoring motion analysis based on traffic Closed-Circuit Television (CCTV)
parameters, such as acceleration, velocity and so on. footage, and combined Faster Regions-Convolutional Neural
Reference [22] used On Board Diagnosis (OBD) system to Network (R-CNN) and Context Mining to detect and predict
monitor speed and engine status to detect a crash, and uti- car accidents. The method in [30] achieved 1.68 seconds in
lized smart-phone to report the accident by Wi-Fi or cellular terms of Time-To-Accident measure with an Average Preci-
network. Reference [23] developed an accident detection and sion of 47.25%. Reference [8] proposed a novel framework
reporting system using GPS, GPRS, and GSM. The speed of for automatic car accident detection, which learned feature
vehicle obtained from High Sensitive GPS receiver is consid- representation from the spatio-temporal volumes of raw pixel
ered as the index for detecting accidents, and the GSM/GPRS intensity instead of traditional hand-crafted features. The
modem is utilized to send the location of the accident. Ref- experiments of method in [8] demonstrated it can detect on
erence [24] presented a prototype system called e-NOTIFY, average 77.5% accidents correctly with 22.5% false alarms.
TABLE 1. Comparison between CAD-CVIS and related datasets. TABLE 2. Composition of YOLO-CA network.
For each image in training set, the total loss is defined as (6).
Especially, because that the multi-scale feature fusion is used
in YOLO-CA, the loss is the sum of conditions under S = 13
and S = 26. In additionally, the loss of each batch of images
is defined as (7).
Loss_img = Lossxy + Losswh + LossCS + Lossp (6) FIGURE 7. The training results of YOLO-CA. (a) Precision (b) Recall (c) IoU
(d) Loss.
b
1 X
Loss = Loss_imgk (7)
b TABLE 3. Distribution map of prediction results.
k=1
IV. EXPERIMENT
In this section, we evaluate our proposed model YOLO-CA
on the CAD-CVIS dataset. First, we give the training results
of YOLO-CA, which include the change process of sev- in iteration process. In the training process of YOLO-CA,
eral performance indexes. Then, we show the results of we regard the prediction result with IoU over 0.5 and right
some comparative experiments between YOLO-CA and other classification as true result, and other predictions are all false
detection models. Finally, the visual results are demonstrated results. As shown in Table. 3, the prediction results can be
among various types and scales of car accident objects. divided into four parts: (1) TP: Truth Positive. (2) FP: False
Positive. (3) FN: False Negative. (4) TN: True Negative. The
TP
A. IMPLEMENTATION DETAILS precision is defined as precision = TP+FP and recall is
We implement our model in TensorFlow [41] under the oper- TP
defined as recall = TP+FN .
ating system Ubuntu 18.04 and perform experiments on a As shown in Fig. 7a, with the increasing of iterations,
system with Nvidia Titan Xp GPU. We divide the CAD-CVIS the precision of YOLO-CA is increasing gradually and
dataset into three parts: (1) Training set (80%), which is used converge over 90%. Moreover, recall eventually converges
to train the parameter weight of network. (2) Validation set to more than 95%. In terms of locating performance of
(5%), which is utilized to adjust hyperparameters, such as YOLO-CA in training set, IoU finally stabilizes above
learning rate and drop out rate. (3) Test set (15%), which is 0.8. The Fig. 7d shows the decreasing process of loss of
used to evaluate the performance of different algorithms for YOLO-CA in (7), and the final convergence of loss is less
detecting car accident. In additionally, each part of dataset than 0.2.
contains all types of accident in Fig. 3. The batch size is set to
64, and the models are trained for up to 30000 iterations. The 2) COMPARATIVE EXPERIMENTS AND VISUAL RESULTS
initial learning rate is set to 0.001, and updating with iteration The comparative experiments are conducted for comparing
parameter of 0.1/10000 iterations. The SGD optimizer with a seven detection models: (1) One-stage models: SSD, our
momentum of 0.9 is utilized to adjust parameters of network. proposed YOLO-CA, traditional YOLO-v3 and YOLO-v3
Moreover, we use a weight decay of 0.0005 to prevent model without MSFF (Multi-Scale Feature Fusion). (2) Two-stage
overfitting. models: Fast R-CNN, Faster R-CNN and Faster R-CNN with
FPN. In order to comparatively demonstrate the validation
B. RESULTS AND ANALYSIS of YOLO-CA as well as confirm its strength in terms of the
1) TRAINING RESULTS OF YOLO-CA comprehensive performance on the accuracy and real-time,
Fig. 7 shows the training results of YOLO-CA, including the following indexes are selected for comparison among the
the changes of precision, recall, IoU and loss of each batch seven models:
TABLE 4. AP and IoU results of different models among different scales of object.
FIGURE 10. Some visual results of the seven models among different scales of objects.
of YOLOv3 without MSFF, especially for small scales of so comparing them may not be fair at this stage. But still,
object (76.51% vs 58.89%). Compared with Faster R-CNN we list the performance achieved by these methods on their
with FPN, YOLO-CA can approach the AP of it (90.66% individual datasets. ARRS [3] achieve about 63% AP with
vs 90.03%) with an obvious speed advantage. Faster R-CNN 6% false alarms. The method of [27] achieve 89.50% AP.
cost about 277ms on average to detect one frame, whereas DSA-RNN [25] achieve about 80% recall and 56.14% AP.
YOLO-CA only need 46 ms, which illustrates the speed of The method in [30] achieve about 47.25% AP. The method
YOLO-CA is about 6× faster than Faster R-CNN with FPN. of [8] achieve 77.5% AP and 22.5% false alarms. Moreover,
Car accident detection in CVIS requires high real-time per- the number of accident scenes of the datasets utilized in these
formance because of the high dynamics of vehicles. To sum- methods is limited, which will result in poor adaptability for
marize, our proposed YOLO-CA have higher practicality and new scenarios.
comprehensive performance on accuracy and real-time. V. CONCLUSION
In this paper, we have proposed an automatic car accident
4) COMPARISON WITH OTHER CAR ACCIDENT detection method based on CVIS. First of all, we present
DETECTION METHODS the application principles of our proposed method in the
Although other car accident detection methods utilize a small CVIS. Secondly, we build a novel image dataset CAD-CVIS,
private collection of datasets and do not make them public which is more suitable for car accident detection method
based on intelligent roadside devices in CVIS. Then we [18] D. Dooley, B. McGinley, C. Hughes, L. Kilmartin, E. Jones, and M. Glavin,
develop the car accident detection model YOLO-CA based ‘‘A blind-zone detection method using a rear-mounted fisheye camera with
combination of vehicle detection methods,’’ IEEE Trans. Intell. Transp.
on CAD-CVIS and deep learning algorithms. In the model, Syst., vol. 17, no. 1, pp. 264–278, Jan. 2016.
we combine the multi-scale feature fusion and loss function [19] X. Changzhen, W. Cong, M. Weixin, and S. Yanmei, ‘‘A traffic sign
with dynamic weights to improve real-time and accuracy detection algorithm based on deep convolutional neural network,’’ in Proc.
IEEE Int. Conf. Signal Image Process. (ICSIP), Aug. 2016, pp. 676–679.
of YOLO-CA. Finally, we show the simulation experiments [20] S. Zhang, C. Bauckhage, and A. B. Cremers, ‘‘Efficient pedestrian detec-
results of our method, which demonstrates our proposed tion via rectangular features based on a statistical shape model,’’ IEEE
methods can detect car accident in 0.0461 seconds with Trans. Intell. Transp. Syst., vol. 16, no. 2, pp. 763–775, Apr. 2015.
[21] J. Redmon and A. Farhadi, ‘‘YOLOv3: An incremental improvement,’’
90.02% AP. Moreover, the comparative experiments results Aug. 2018, arXiv:1804.02767. [Online]. Available: https://arxiv.org/abs/
show that YOLO-CA has comprehensive performance advan- 1804.02767
tages of detecting car accident than other detection models, [22] J. Zaldivar, C. T. Calafate, J. C. Cano, and P. Manzoni, ‘‘Providing
accident detection in vehicular networks through OBD-II devices and
in terms of accuracy and real-time. Android-based smartphones,’’ in Proc. IEEE 36th Conf. Local Comput.
Netw. (LCN), Oct. 2011, pp. 813–819.
REFERENCES [23] M. S. Amin, J. Jalil, and M. B. I. Reaz, ‘‘Accident detection and reporting
system using GPS, GPRS and GSM technology,’’ in Proc. Int. Conf.
[1] WHO. Global Status Report on Road Safety 2018. Accessed: Dec. 2018. Inform., Electron. Vis., 2012, pp. 640–643.
[Online]. Available: https://www.who.int/violence_injury_prevention/ [24] M. Fogue, P. Garrido, F. J. Martinez, J.-C. Cano, C. T. Calafate, and
road_safety_status/2018/en/ P. Manzoni, ‘‘Automatic accident detection: Assistance through communi-
[2] H. L. Wang and M. A. Jia-Liang, ‘‘A design of smart car accident rescue cation technologies and vehicles,’’ IEEE Veh. Technol. Mag., vol. 7, no. 3,
system combined with WeChat platform,’’ J. Transp. Eng., vol. 17, no. 2, pp. 90–100, Sep. 2012.
pp. 48–52, Apr. 2017. [25] F.-H. Chan, Y.-T. Chen, Y. Xiang, and M. Sun, ‘‘Anticipating accidents
[3] Y. K. Ki and D. Y. Lee, ‘‘A traffic accident recording and reporting model at in dashcam videos,’’ in Computer Vision-ACCV 2016. Springer, 2017,
intersections,’’ IEEE Trans. Intell. Transp. Syst., vol. 8, no. 2, pp. 188–194, pp. 136–153.
Jun. 2007. [26] Y. Yao, M. Xu, Y. Wang, D. J. Crandall, and E. M. Atkins, ‘‘Unsu-
[4] W. Hao and J. Daniel, ‘‘Motor vehicle driver injury severity study under pervised traffic accident detection in first-person videos,’’ Mar. 2019,
various traffic control at highway-rail grade crossings in the United States,’’ arXiv:1903.00618. [Online]. Available: https://arxiv.org/abs/1903.00618
J. Saf. Res., vol. 51, pp. 41–48, Dec. 2014. [27] S. Kamijo, Y. Matsushita, K. Ikeuchi, and M. Sakauchi, ‘‘Traffic monitor-
[5] J. White, C. Thompson, H. Turner, H. Turner, and D. C. Schmidt, ing and accident detection at intersections,’’ IEEE Trans. Intell. Transp.
‘‘Wreckwatch: Automatic traffic accident detection and notification Syst., vol. 1, no. 2, pp. 108–118, Jun. 2000.
with smartphones,’’ Mobile Netw. Appl., vol. 16, no. 3, pp. 285–303, [28] K. Yun, H. Jeong, K. M. Yi, S. W. Kim, and J. Y. Choi, ‘‘Motion interaction
Jun. 2011. field for accident detection in traffic surveillance video,’’ in Proc. 22nd Int.
[6] S. Sadek, A. Al-Hamadi, B. Michaelis, and U. Sayed, ‘‘Real-time auto- Conf. Pattern Recognit., Aug. 2014, pp. 3062–3067.
matic traffic accident recognition using HFG,’’ in Proc. 20th Int. Conf. [29] V. Ravindran, L. Viswanathan, and S. Rangaswamy, ‘‘A novel approach to
Pattern Recognit., Aug. 2010, pp. 3348–3351. automatic road-accident detection using machine vision techniques,’’ Int.
[7] A. Shaik, N. Bowen, J. Bole, G. Kunzi, D. Bruce, A. Abdelgawad, and J. Adv. Comput. Sci. Appl., vol. 7, no. 11, pp. 235–242, 2016.
K. Yelamarthi, ‘‘Smart car: An IoT based accident detection system,’’ in [30] A. P. Shah, J.-B. Lamare, T. Nguyen-Anh, and A. Hauptmann, ‘‘CADP:
Proc. IEEE Global Conf. Internet Things (GCIoT), Dec. 2018, pp. 1–5. A novel dataset for CCTV traffic camera based accident analysis,’’ in
[8] D. Singh and C. K. Mohan, ‘‘Deep spatio-temporal representation for Proc. 15th IEEE Int. Conf. Adv. Video Signal Based Surveill., Nov. 2018,
detection of road accidents using stacked autoencoder,’’ IEEE Trans. Intell. pp. 1–9.
Transp. Syst., vol. 20, no. 3, pp. 879–887, Mar. 2019. [31] Tzutalin. (2015). Labelimg. Git Code. [Online]. Available: https://github.
[9] M. Zheng, T. Li, R. Zhu, J. Chen, Z. Ma, M. Tang, Z. Cui, and Z. Wang, com/tzutalin/labelImg
‘‘Traffic accident’s severity prediction: A deep-learning approach-based [32] C. Lu, J. Shi, and J. Jia, ‘‘Abnormal event detection at 150 FPS in MAT-
CNN network,’’ IEEE Access, vol. 7, pp. 39897–39910, 2019. LAB,’’ in Proc. IEEE Int. Conf. Comput. Vis., Dec. 2013, pp. 2720–2727.
[10] L. Zheng, Z. Peng, J. Yan, and W. Han, ‘‘An online learning and unsu- [33] V. Mahadevan, W. Li, V. Bhalodia, and N. Vasconcelos, ‘‘Anomaly detec-
pervised traffic anomaly detection system,’’ Adv. Sci. Lett., vol. 7, no. 1, tion in crowded scenes,’’ in Proc. IEEE Comput. Soc. Conf. Comput.
pp. 449–455, 2012. Vis. Pattern Recognit., Jun. 2010, pp. 1975–1981. [Online]. Available:
[11] Y. Fangchun, W. Shangguang, L. Jinglin, L. Zhihan, and S. Qibo, http://ieeexplore.ieee.org/document/5539872/
‘‘An overview of Internet of vehicles,’’ China Commun., vol. 11, no. 10, [34] R. Girshick, J. Donahue, T. Darrell, and J. Malik, ‘‘Rich feature hierarchies
pp. 1–15, Oct. 2014. for accurate object detection and semantic segmentation,’’ in Proc. IEEE
[12] C. Ma, W. Hao, A. Wang, and H. Zhao, ‘‘Developing a coordinated Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2014, pp. 580–587.
signal control system for urban ring road under the vehicle-infrastructure [35] R. Girshick, ‘‘Fast R-CNN,’’ in Proc. IEEE Int. Conf. Comput. Vis.,
connected environment,’’ IEEE Access, vol. 6, pp. 52471–52478, 2018. Dec. 2015, pp. 1440–1448.
[13] S. Zhang, J. Chen, F. Lyu, N. Cheng, W. Shi, and X. Shen, ‘‘Vehicular [36] S. Ren, K. He, R. Girshick, and J. Sun, ‘‘Faster R-CNN: Towards real-time
communication networks in the automated driving era,’’ IEEE Commun. object detection with region proposal networks,’’ in Proc. Int. Conf. Neural
Mag., vol. 56, no. 9, pp. 26–32, Sep. 2018. Inf. Process. Syst., 2015, pp. 91–99.
[14] Y. Wang, D. Zhang, Y. Liu, B. Dai, and L. H. Lee, ‘‘Enhancing transporta- [37] T.-Y. Lin, P. Dollár, R. Girshick, K. He, B. Hariharan, and S. Belongie,
tion systems via deep learning: A survey,’’ Transp. Res. C, Emerg. Technol., ‘‘Feature pyramid networks for object detection,’’ in Proc. IEEE Conf.
2018. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2016, pp. 936–944.
[15] G. Wu, F. Chen, X. Pan, M. Xu, and X. Zhu, ‘‘Using the visual intervention [38] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.-Y. Fu, and
influence of pavement markings for rutting mitigation—Part I: Prelimi- A. C. Berg, ‘‘SSD: Single shot multibox detector,’’ in Proc. Eur. Conf.
nary experiments and field tests,’’ Int. J. Pavement Eng., vol. 20, no. 6, Comput. Vis., 2016, pp. 21–37.
pp. 734–746, 2019. [39] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, ‘‘You only look once:
[16] S. Ramos, S. Gehrig, P. Pinggera, U. Franke, and C. Rother, ‘‘Detecting Unified, real-time object detection,’’ in Proc. IEEE Conf. Comput. Vis.
unexpected obstacles for self-driving cars: Fusing deep learning and geo- Pattern Recognit., Jun. 2016, pp. 779–788.
metric modeling,’’ in Proc. IEEE Intell. Vehicles Symp. (IV), Jun. 2017, [40] K. He, X. Zhang, S. Ren, and J. Sun, ‘‘Deep residual learning for image
pp. 1025–1032. recognition,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR),
[17] T. Qu, Q. Zhang, and S. Sun, ‘‘Vehicle detection from high-resolution aerial Jun. 2016, pp. 770–778.
images using spatial pyramid pooling-based deep convolutional neural [41] M. Abadi et al., ‘‘TensorFlow: Large-scale machine learning on hetero-
networks,’’ Multimedia Tools Appl., vol. 76, no. 20, pp. 21651–21663, geneous distributed systems,’’ Mar. 2015, arXiv:1603.04467. [Online].
2017. Available: https://arxiv.org/abs/1603.04467
DAXIN TIAN is currently a Professor with the XUTING DUAN is currently a Lecturer with the
School of Transportation Science and Engineer- School of Transportation Science and Engineer-
ing, Beihang University, Beijing, China. His cur- ing, Beihang University, Beijing, China. His cur-
rent research interests include mobile computing, rent research interests include connected vehicles,
intelligent transportation systems, vehicular ad hoc vehicular ad hoc networks, and vehicular
networks, and swarm intelligent. localization.
CHUANG ZHANG is currently pursuing the XIXIAN WANG is currently pursuing the
master’s degree with the School of Transporta- master’s degree with the School of Transporta-
tion Science and Engineering, Beihang University, tion Science and Engineering, Beihang University,
Beijing, China. His current research interests Beijing, China. His current research interests
include multimedia communications and process- include image processing and machine learning.
ing and machine learning.