sequential game
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > United States > Texas (0.04)
- North America > Canada (0.04)
- Europe > Switzerland > Zürich > Zürich (0.05)
- North America > United States > California > Los Angeles County > Los Angeles (0.04)
- North America > Canada (0.04)
- (2 more...)
- Transportation > Infrastructure & Services (0.68)
- Leisure & Entertainment > Games (0.47)
Adversarially Robust Decision Transformer
Decision Transformer (DT), as one of the representative Reinforcement Learning via Supervised Learning (RvS) methods, has achieved strong performance in offline learning tasks by leveraging the powerful Transformer architecture for sequential decision-making. However, in adversarial environments, these methods can be non-robust, since the return is dependent on the strategies of both the decision-maker and adversary. Training a probabilistic model conditioned on observed return to predict action can fail to generalize, as the trajectories that achieve a return in the dataset might have done so due to a suboptimal behavior adversary. To address this, we propose a worst-case-aware RvS algorithm, the Adversarially Robust Decision Transformer (ARDT), which learns and conditions the policy on in-sample minimax returns-to-go. ARDT aligns the target return with the worst-case return learned through minimax expectile regression, thereby enhancing robustness against powerful test-time adversaries. In experiments conducted on sequential games with full data coverage, ARDT can generate a maximin (Nash Equilibrium) strategy, the solution with the largest adversarial robustness. In large-scale sequential games and continuous adversarial RL environments with partial data coverage, ARDT demonstrates significantly superior robustness to powerful test-time adversaries and attains higher worst-case returns compared to contemporary DT methods.
Exploiting Opponents Under Utility Constraints in Sequential Games
Recently, game-playing agents based on AI techniques have demonstrated super-human performance in several sequential games, such as chess, Go, and poker. Surprisingly, the multi-agent learning techniques that allowed to reach these achievements do not take into account the actual behavior of the human player, potentially leading to an impressive gap in performances. In this paper, we address the problem of designing artificial agents that learn how to effectively exploit unknown human opponents while playing repeatedly against them in an online fashion. We study the case in which the agent's strategy during each repetition of the game is subject to constraints ensuring that the human's expected utility is within some lower and upper thresholds. Our framework encompasses several real-world problems, such as human engagement in repeated game playing and human education by means of serious games. As a first result, we formalize a set of linear inequalities encoding the conditions that the agent's strategy must satisfy at each iteration in order to do not violate the given bounds for the human's expected utility. Then, we use such formulation in an upper confidence bound algorithm, and we prove that the resulting procedure suffers from sublinear regret and guarantees that the constraints are satisfied with high probability at each iteration. Finally, we empirically evaluate the convergence of our algorithm on standard testbeds of sequential games.
- Europe > Switzerland > Zürich > Zürich (0.05)
- North America > United States > California > Los Angeles County > Los Angeles (0.04)
- North America > Canada (0.04)
- (2 more...)
- Transportation > Infrastructure & Services (0.68)
- Leisure & Entertainment > Games (0.47)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > United States > Texas (0.04)
- North America > Canada (0.04)
Appendix of the paper Exploiting Opponents under Utility Constraints in Sequential Games
Appendix A provides the proofs omitted from Section 4.1, describing the method adopted Appendix B provides the proofs omitted from Section 4.2, describing the method adopted Appendix D provides some additional experimental results. A Proofs omitted from Section 4.1 B Proofs omitted from Section 4.2 Theorem 2. Let t [T ] and δ (0, 1). The proof follows the reasoning outlined in Section 4.2. Before proving Theorem 3, we need to show the following technical lemma. Lemma 4. Let f (τ):= This is reasonable in practice, since a new player can always be profiled according to a number of user classes.
Adversarially Robust Decision Transformer
Decision Transformer (DT), as one of the representative Reinforcement Learning via Supervised Learning (RvS) methods, has achieved strong performance in offline learning tasks by leveraging the powerful Transformer architecture for sequential decision-making. However, in adversarial environments, these methods can be non-robust, since the return is dependent on the strategies of both the decision-maker and adversary. Training a probabilistic model conditioned on observed return to predict action can fail to generalize, as the trajectories that achieve a return in the dataset might have done so due to a suboptimal behavior adversary. To address this, we propose a worst-case-aware RvS algorithm, the Adversarially Robust Decision Transformer (ARDT), which learns and conditions the policy on in-sample minimax returns-to-go. ARDT aligns the target return with the worst-case return learned through minimax expectile regression, thereby enhancing robustness against powerful test-time adversaries.