notesum.ai
Published at October 21Reinforced Imitative Trajectory Planning for Urban Automated Driving
cs.LG
cs.AI
Released Date: October 21, 2024
Authors: Di Zeng1, Ling Zheng1, Xiantong Yang1, Yinong Li1
Aff.: 1College of Mechanical and Vehicle Engineering, Chongqing University, Shazheng Street, Chongqing, 40044, Chongqing, China.

| Hyperparameters | AVRL | QCMAE | (H)RITP |
| Dropout probability | 0.1 | 0.1 | 0.1 |
| Planning horizon | 80 | 80 | 80 |
| Historical time horizon | - | 50 | 50 |
| Hidden feature dimension | 64 | 64 | 64 |
| Number of modes | - | 6 | 6 |
| Exploration strength | - | - | 0.1 |
| Policy noise strength | - | - | 0.2 |
| Clipping boundary | - | - | 0.5 |
| Optimizer | Adam [45] | AdamW [46] | AdamW |
| Learning rate | 3e-7 | 5e-55e-45e-5 | 5e-5 |
| Learning rate scheduler | - | OneCycle [47] | - |
| Batch size | 1 | 4 | 4, 1111Footnote for Data 2 |
| Discount factor | - | - | 0.99 |
| Target update rate | - | - | 0.005 |
| Update delay | - | - | 2 |
| Uncertainty-penalization strength | - | - | 1.5 |
| Number of stochastic forward passes | 10 | - | - |
| Max number of experiences in replay buffer | - | - | 1e4 |
| Total training epochs | 10 | 60 | - |
| Total training steps | - | - | 1e5 |
| \botrule |