Goto

Collaborating Authors

 aaai


Counterfactual Evaluation of Peer-Review Assignment Policies Supplemental Material Martin Saveski, Steven Jecmen, Nihar B. Shah, Johan Ugander A Linear Programs for Peer-Review Assignment

Neural Information Processing Systems

Our estimators assume that there is no interference between the units, i.e., that the treatment of one The first assumption is quite realistic as in most peer review systems the reviewers cannot see other reviews until they submit their own. The second assumption is important to understand, as there could be "batch effects": a Monte Carlo methods to tightly estimate these covariances. AAAI datasets, we sampled 1 million assignments and computed the empirical covariance. In our setting, small amounts of attrition (relative to the number of policy-induced positivity violations) mean that the fraction of data that is missing is not exactly known before assignment, but almost. To get more robust estimates of the performance, we repeat this process 10 times.










Robohub highlights 2025

Robohub

Over the course of the year, we've had the pleasure of working with many talented researchers from across the globe. As 2025 draws to a close, we take a look back at some of the excellent blog posts, interviews and podcasts from our contributors. Jiahui Zhang and Jesse Zhang to tell us about their framework for learning robot manipulation tasks solely from language instructions without per-task demonstrations. Hui Zhang writes about work presented at CoRL2025 on RobustDexGrasp, a novel framework that tackles different grasping challenges with targeted solutions. In this podcast from AAAI, host Ella Lan asked Professor Marynel Vázquez about what inspired her research direction, how her perspective on human-robot interactions has changed over time, robots navigating the social world, and more.