Goto

Collaborating Authors

 action set


Exploring Large Action Sets with Hyperspherical Embeddings using von Mises-Fisher Sampling

Bendada, Walid, Salha-Galvan, Guillaume, Hennequin, Romain, Bontempelli, Théo, Bouabça, Thomas, Cazenave, Tristan

arXiv.org Artificial Intelligence

This paper introduces von Mises-Fisher exploration (vMF-exp), a scalable method for exploring large action sets in reinforcement learning problems where hyperspherical embedding vectors represent these actions. vMF-exp involves initially sampling a state embedding representation using a von Mises-Fisher distribution, then exploring this representation's nearest neighbors, which scales to virtually unlimited numbers of candidate actions. We show that, under theoretical assumptions, vMF-exp asymptotically maintains the same probability of exploring each action as Boltzmann Exploration (B-exp), a popular alternative that, nonetheless, suffers from scalability issues as it requires computing softmax values for each action. Consequently, vMF-exp serves as a scalable alternative to B-exp for exploring large action sets with hyperspherical embeddings. Experiments on simulated data, real-world public data, and the successful large-scale deployment of vMF-exp on the recommender system of a global music streaming service empirically validate the key properties of the proposed method.


Zero-determinant strategies in repeated continuously-relaxed games

Ueda, Masahiko, Fujita, Ayaka

arXiv.org Artificial Intelligence

Mixed extension has played an important role in game theory, especially in the proof of the existence of Nash equilibria in strategic form games. Mixed extension can be regarded as continuous relaxation of a strategic form game. Recently, in repeated games, a class of behavior strategies, called zero-determinant strategies, was introduced. Zero-determinant strategies unilaterally enforce linear relations between payoffs, and are used to control payoffs of players. There are many attempts to extend zero-determinant strategies so as to apply them to broader situations. Here, we extend zero-determinant strategies to repeated games where action sets of players in stage game are continuously relaxed. We see that continuous relaxation broadens the range of possible zero-determinant strategies, compared to the original repeated games. Furthermore, we introduce a special type of zero-determinant strategies, called one-point zero-determinant strategies, which repeat only one continuously-relaxed action in all rounds. By investigating several examples, we show that some property of mixed-strategy Nash equilibria can be reinterpreted as a payoff-control property of one-point zero-determinant strategies.


Action Set Based Policy Optimization for Safe Power Grid Management

Zhou, Bo, Zeng, Hongsheng, Liu, Yuecheng, Li, Kejiao, Wang, Fan, Tian, Hao

arXiv.org Artificial Intelligence

Maintaining the stability of the modern power grid is becoming increasingly difficult due to fluctuating power consumption, unstable power supply coming from renewable energies, and unpredictable accidents such as man-made and natural disasters. As the operation on the power grid must consider its impact on future stability, reinforcement learning (RL) has been employed to provide sequential decision-making in power grid management. However, existing methods have not considered the environmental constraints. As a result, the learned policy has risk of selecting actions that violate the constraints in emergencies, which will escalate the issue of overloaded power lines and lead to large-scale blackouts. In this work, we propose a novel method for this problem, which builds on top of the search-based planning algorithm. At the planning stage, the search space is limited to the action set produced by the policy. The selected action strictly follows the constraints by testing its outcome with the simulation function provided by the system. At the learning stage, to address the problem that gradients cannot be propagated to the policy, we introduce Evolutionary Strategies (ES) with black-box policy optimization to improve the policy directly, maximizing the returns of the long run. In NeurIPS 2020 Learning to Run Power Network (L2RPN) competition, our solution safely managed the power grid and ranked first in both tracks.


SAS and R Integration for Machine Learning

#artificialintelligence

R first appeared in 1993 and has gained a steady and fiercely loyal fan base. But as data sets become both longer and wider, storage and processing speeds become an issue. Having spent weeks whipping an extremely wide and messy data set into shape using only R, I am so grateful for SAS Viya and not having to go through that again. SAS Viya is a cloud-enabled, in-memory analytics engine which allows for rapid analytics insights. SAS Viya utilizes the SAS Cloud Analytics Services (CAS) to perform various actions and tasks.