This study develops a synchromodal transport model with real-time replanning using a Double-Dueling Deep Q-Network (D3QN) agent. The hinterland port-hinterland leg dominates supply-chain costs and remains vulnerable to disruptions. We simulate realistic disruptions using varying
...
This study develops a synchromodal transport model with real-time replanning using a Double-Dueling Deep Q-Network (D3QN) agent. The hinterland port-hinterland leg dominates supply-chain costs and remains vulnerable to disruptions. We simulate realistic disruptions using varying occurrence and severity levels and train a D3QN agent to decide whether an affected shipment should wait or reassign to another mode. Model performance is evaluated in a Rhine-Alpine corridor case study against two baselines strategies (Always Wait, Always Reassign) and a tabular Q-learning benchmark. The results show that the D3QN policy achieves the lowest costs on a combined set of disruptions, outperforming alternatives overall. In particular, D3QN excels under high-occurrence, low-severity disruptions but appears less effective under rare, high-severity events. The integration of the D3QN with parallel simulation greatly sped up training (roughly 4 times faster convergence) without degrading policy quality. It is discussed how classifying disruptions by occurrence and severity helps tractability, and show that the D3QN learns to reassign shipments more effectively. The findings indicate that deep reinforcement learning can improve synchromodal resilience, particularly in large-scale settings where conventional RL or heuristics fail.