Adversarially Robust Decision Transformer

Neural Information Processing Systems 

However, in adversarial environments, these methods can be non-robust, since the return is dependent on the strategies of both the decision-maker and adversary. Training a probabilistic model conditioned on observed return to predict action can fail to generalize, as the trajectories that achieve a return in the dataset might have done so due to a suboptimal behavior adversary.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found