Goto

Collaborating Authors

 cleanup



ad7ed5d47b9baceb12045a929e7e2f66-Supplemental.pdf

Neural Information Processing Systems

A.1 Costforincentivization We justify the way in which LIO accounts for the cost of incentivization as follows. However, both the reward-giverand recipients require sufficient time tolearn the effect ofincentives,which means that too large anα would lead to the degenerate result ofrηi = 0. On the other extreme, α = 0means there isno penalty and may result inprofligate incentivization that serves no useful purpose. Let θi for i {1,2} denote each agent's probability of taking the cooperative action. Each plot has afixed value for the incentive givenfortheotheraction. Each agent observesallagents' positions andcanmoveamong thethree available states: lever, start, and door.



This 30% off Black Friday deal on CleanMyMac software will make your life easier all year

Popular Science

CleanMyMac itself hooks into macOS's "Allow in the Background" framework here, so it's playing by Apple's rules rather than working around them. You could do most of this via System Settings and a lot of manual digging, but the point here is visibility: you see what's running, how heavy it is, and you can trim without spelunking through multiple folders.






A tiny grain of nuclear fuel is pulled from ruined Japanese nuclear plant, in a step toward cleanup

FOX News

Fox News Flash top headlines are here. Check out what's clicking on Foxnews.com. A robot that has spent months inside the ruins of a nuclear reactor at the tsunami-hit Fukushima Daiichi plant delivered a tiny sample of melted nuclear fuel on Thursday, in what plant officials said was a step toward beginning the cleanup of hundreds of tons of melted fuel debris. The sample, the size of a grain of rice, was placed into a secure container, marking the end of the mission, according to Tokyo Electric Power Company Holdings, which manages the plant. It is being transported to a glove box for size and weight measurements before being sent to outside laboratories for detailed analyses over the coming months.


Learning to Balance Altruism and Self-interest Based on Empathy in Mixed-Motive Games

Kong, Fanqi, Huang, Yizhe, Zhu, Song-Chun, Qi, Siyuan, Feng, Xue

arXiv.org Artificial Intelligence

Real-world multi-agent scenarios often involve mixed motives, demanding altruistic agents capable of self-protection against potential exploitation. However, existing approaches often struggle to achieve both objectives. In this paper, based on that empathic responses are modulated by inferred social relationships between agents, we propose LASE Learning to balance Altruism and Self-interest based on Empathy), a distributed multi-agent reinforcement learning algorithm that fosters altruistic cooperation through gifting while avoiding exploitation by other agents in mixed-motive games. LASE allocates a portion of its rewards to co-players as gifts, with this allocation adapting dynamically based on the social relationship -- a metric evaluating the friendliness of co-players estimated by counterfactual reasoning. In particular, social relationship measures each co-player by comparing the estimated $Q$-function of current joint action to a counterfactual baseline which marginalizes the co-player's action, with its action distribution inferred by a perspective-taking module. Comprehensive experiments are performed in spatially and temporally extended mixed-motive games, demonstrating LASE's ability to promote group collaboration without compromising fairness and its capacity to adapt policies to various types of interactive co-players.