Testing and benchmarking of microscopic traffic flow simulation
Testing and benchmarking of microscopic traffic flow simulation models Elmar Brockfeld, Peter Wagner Elmar. Brockfeld@dlr. de, Peter. Wagner@dlr. de Institute of Transport Research German Aerospace Center (DLR) Rutherfordstrasse 2 12489 Berlin, Germany 10 th WCTR, Istanbul, 06. 07. 2004 DLR-Institute of Transport Research
„State of the art“ The situation in microscopic traffic flow modelling today: » A very large number of models exists describing the traffic flow. » If they are tested, this is done separately with special data sets. » By now the microscopic models are quantitatively not comparable. DLR-Institute of Transport Research 2
Motivation Idea » » Calibrate and validate microscopic traffic flow models with the same data sets. ( quantitative comparibility, benchmark possible ? ) Calibration and validation in a microscopic way by analysing any timeseries produced by single cars. In the following » » Calibration and validation of ten car-following models with data recorded on a test track in Hokkaido, Japan. Comparison with results of other approaches. DLR-Institute of Transport Research 3
Test track Hokkaido, Japan curve 300 m 1200 m » 10 cars equipped with DGPS driving on a 3 km test track » Delivery of positions in intervals of 0. 1 second DLR-Institute of Transport Research 4
Test track Hokkaido, Japan Impressions DLR-Institute of Transport Research 5
Hokkaido, Japan – The data » Data recorded by Nakatsuji et al. in 2001 » Data from 4 out of 8 experiments are used for the analyses: Experiment Duration [min] Full loops „ 11“ 26 6 „ 12“ 25 7 „ 13“ 18 6 „ 21“ 14 4 » Exchange of drivers between the cars after each experiment » Leading car performed certain “driving patterns” on the straight sections: » driving with constant speeds of 20, 40, 60 and 80 km/h » driving in waves varying from about 30 to 70 km/h DLR-Institute of Transport Research 6
Hokkaido, Japan – Speed development of the leading car in all four experiments DLR-Institute of Transport Research 7
The models The following existing models have been analysed: follower leader » 4 parameters, CA 0. 1 („cellular automaton model“) » 4 p, OVM (“Optimal Velocity Model” by Bando) » 6 p, GIPPSLIKE (basic model by P. G. Gipps) » 6 p, AERDE (used in the software INTEGRATION) » 6 p, IDM (“Intelligent Driver Model” by D. Helbing) » 7 p, IDMM (“Intelligent Driver Model with Memory”) » 7 p, SK_STAR (based on the model by S. Krauss) » 7 p, NEWELL (CA-variant of the model with more variable acceleration and deceleration by G. Newell) » 13 p, FRITZSCHE (used in the british software PARAMICS; similar to what is used in the german software VISSIM by PTV) » 15 p, MITSIM (used in the software Mit. Sim) DLR-Institute of Transport Research 8
The model‘s parameters Parameters used by all models: V_max Maximum velocity l Vehicle length a acceleration Most models: b deceleration tau reaction time Models with different driving regimes: MITSIM and FRITZSCHE Java Applet for testing the models DLR-Institute of Transport Research 9
Hokkaido, Japan - Simulation setup V_sim V_data gap » For each simulation run one vehicle pair is under consideration » Movement of leading car: as recorded in the data » Movement of following car: following the rules of a traffic model » Error measurement: » e percentage error » T time series of experiment » g(obs) observed gaps/headways » g(sim) simulated gaps/headways » Objective of calibration: Minimize the error e ! DLR-Institute of Transport Research 10
Hokkaido, Japan – gaps time series DLR-Institute of Transport Research 11
Hokkaido, Japan – Calibration and Validation Calibration (“Adjust parameters of a model to real data”) » Find the optimal parameter sets for each vehicle pair in each experiment (9*4 = 36 calibrations for each model): » Minimize the error e as defined before » Minimization with a gradient free (direct search) optimisation algorithm (“downhill simplex” or “Nelder-Mead”) » To avoid local minima: about 100 simulations with random initializations Validation (“Apply calibrated model to other real data sets”) » For each model all optimal parameter results are transferred to data sets of three other driver pairs (in total 108 validations for each model) DLR-Institute of Transport Research 12
Hokkaido, Japan - Calibration Results (1/2) Results of the first experiment “ 11” » Errors between 9 and 19 %, mostly between 13 and 17 % » No model appears to be the best DLR-Institute of Transport Research 13
Hokkaido, Japan - Calibration Results (2/2) Calibration of 10 models with 36 data sets ( 4 experiments „ 11“, „ 12“, „ 13“ and „ 21“, each with 9 driver pairs): » Calibration error: mostly 12 - 18 % (range 9 - 23 %) » All models share the same problems with the same data sets -> Which is the best model? » Average error of best model: 15. 14 % » Average error of worst model: 16. 20 % » Models with more parameters do not produce better results » Average difference of the models per data set is 2. 5 percentage points Diversity in driver behaviour (6 %) DLR-Institute of Transport Research > Diversity of models (2. 5 %) 14
Hokkaido, Japan - Validation Results (1/2) Validation of each calibration result with three other driver pairs (->108 validations for each model) » Validation error: mostly 17 - 27 % » “Overfitting 1”: Special driver behaviour may produce high errors up to 40 or 50 % for all models » “Overfitting 2”: For some driver pairs some models produce singular high errors of more than 100 % Sample plots for 2*9 validations DLR-Institute of Transport Research 15
Hokkaido, Japan - Validation Results (2/2) Distribution functions of the errors » Calibration errors: 12 - 18 % (peak 15 -17) » Validation errors: 17 - 27 % (peak 21 -23) » Additional validation error to calibration: 6 percentage points DLR-Institute of Transport Research » Calibration: MEDIAN best/worst model: 14. 84 % / 16. 04 % » Validation: MEDIAN best/worst model: 21. 60 % / 22. 58 % » Additional validation error to calibration: 5. 66 pp / 7. 23 pp 16
Comparison with other calibration approaches Approach Recorded on; Device Data volume Hokkaido (Brockfeld et al) Single lane test track; DGPS 40 traces 15 -30 minutes Hokkaido (Ranjitkar, Japan) ICC FOT (Schober, USA) Single lane test track; DGPS Multilane highway; Radar 47 traces 1 -2 minutes 300 traces One to a few minutes San Pablo Dam (Brockfeld et al) Single lane rural road; Humans I-80, Berkeley (Wagner et al) DLR-Institute of Transport Research Passing times of 2300 vehicles at 8 positions on two days Multilane highway; 24 hours Loop detectors highway data Calibration errors headways Main range: 13 -19 % 10 models: 15 -16 % (Validation: 21. 5 -22. 5 %) Main range: 9 -17 % 6 models: 12 -13 %; 15 %; 18 %; 21 % 3 models: 18 -20 % Speed class <35 mph: 9 -10 % Speed class 35 -55 mph: 13 -16 % Speed class > 55 mph: 16 -17 % Travel times 10 models: 15 -17 % (6), 23% (Validation: 17 -27%) Speed, flow 2 models: 18% 17
Conclusions Essential results: » Minimum reachable levels for calibration: » Short traces or special situations: 9 to 11 % » Simulating more than a few minutes: 15 to 20 % » Minimum reachable levels for validation: » > 20 % ; about 3 to 7 percentage points higher than calibration case. » The analysed models do not differ so much » The diversity in the driver behaviour is bigger than the diversity of the models. » Models with more parameters must not necessarily produce better results than simple ones. » Preliminary advice: Take the simplest model or the one you know best! DLR-Institute of Transport Research 18
Perspectives and future research » Testing more models and more data sets » Test some other calibration techniques and measurements (speeds, accelerations, …) » Sensitivity analyses of the parameters (robustness of the models) » What are the problems of the models? Analysis of parameter results. Development of better models. » Finally development of a benchmark for microscopic traffic flow models. DLR-Institute of Transport Research 19
THANK YOU FOR YOUR ATTENTION ! DLR-Institute of Transport Research 20
- Slides: 20