An overview of the STR2-CPKS model which has a sequence of context, proposal, key points, and future states for the MoE backbone to model. For STR2-CKS, proposals are removed in the sequence for better efficiency. The context part has rasterized environment information encoded by scalable ViT encoders and past ego states.
The planning results, in red, from PDM-Hybrid and STR2 at the pickup area at the top, and an illustration of the MoE model learning and balancing different explicit rewards at the bottom.
Performance comparison on testHard Set with details of the closed-loop reactive simulations. Higher scores indicate better performance
Scaling results with the size of the training dataset, counted as the number of tokens D in the left and scaling results with model parameters N in the right. All axes are logarithmically scaled.