Goto

Collaborating Authors

Maximum Entropy Reinforcement Learning via Energy-Based Normalizing Flow Chen-Hao Chao 1,2 Wei-Fang Sun 2

Neural Information Processing Systems

Existing Maximum-Entropy (MaxEnt) Reinforcement Learning (RL) methods for continuous action spaces are typically formulated based on actor-critic frameworks and optimized through alternating steps of policy evaluation and policy improvement. In the policy evaluation steps, the critic is updated to capture the soft Q-function. In the policy improvement steps, the actor is adjusted in accordance with the updated soft Q-function. In this paper, we introduce a new MaxEnt RL framework modeled using Energy-Based Normalizing Flows (EBFlow).


When to Act and When to Ask: Policy Learning With Deferral Under Hidden Confounding

Neural Information Processing Systems

We consider the task of learning how to act in collaboration with a human expert based on observational data. The task is motivated by high-stake scenarios such as healthcare and welfare, where algorithmic action recommendations are made to a human expert, opening the option of deferring recommendation in cases where the human might act better on their own. This task is especially challenging when dealing with observational data, as using such data runs the risk of hidden confounders whose existence can lead to biased and harmful policies. However, unlike standard policy learning, the presence of a human expert can mitigate some of these risks. We build on the work of Mozannar and Sontag [2020] on consistent surrogate loss for learning with the option of deferral to an expert, where they solve a cost-sensitive supervised classification problem. Since we are solving a causal problem, where labels do not exist, we use a causal model to learn costs which are robust to a bounded degree of hidden confounding. We prove that our approach can take advantage of the strengths of both the model and the expert to obtain a better policy than either. We demonstrate our results by conducting experiments on synthetic and semi-synthetic data and show the advantages of our method compared to baselines.


Y, where X and Y are topological spaces, guaranteeing that if F is dense in C(R

Neural Information Processing Systems

Modifications to a neural network's input and output layers are often required to accommodate the specificities of most practical learning tasks. However, the impact of such changes on architecture's approximation capabilities is largely not understood. We present general conditions describing feature and readout maps that preserve an architecture's ability to approximate any continuous functions uniformly on compacts. As an application, we show that if an architecture is capable of universal approximation, then modifying its final layer to produce binary values creates a new architecture capable of deterministically approximating any classifier. In particular, we obtain guarantees for deep CNNs and deep feed-forward networks. Our results also have consequences within the scope of geometric deep learning.


We would like to thank the reviewers for taking the time to carefully read, evaluate, and give feedback on our submission

Neural Information Processing Systems

We would like to thank the reviewers for taking the time to carefully read, evaluate, and give feedback on our submission. Assumptions 3.1 and 3.2-Reviewer 2: "[T]his paper... assumes that the feature map is regular (Assumption 3.1)... Therefore, for a DNN to be universal, we only need the middle two layers, described by f, to be fully-connected. In particular, this gives us the flexibility of encoding many "inductive biases" into the architecture since only the two Therefore, Theorem 3.3 implies that if dropout is used to improve Illustration of Stakes - Reviewer 4: "One can hope that something as simple as the softmax function... does not spoil It is not surprising that the softmax function preserve's the ability for an architecture Similar issues arise with the other mentioned examples and we would be happy to add a brief discussion outlining each.


Bileve: Securing Text Provenance in Large Language Models Against Spoofing with Bi-level Signature

Neural Information Processing Systems

Text watermarks for large language models (LLMs) have been commonly used to identify the origins of machine-generated content, which is promising for assessing liability when combating deepfake or harmful content. While existing watermarking techniques typically prioritize robustness against removal attacks, unfortunately, they are vulnerable to spoofing attacks: malicious actors can subtly alter the meanings of LLM-generated responses or even forge harmful content, potentially leading to the wrongful attribution of blame to the LLM developer. To overcome this, we introduce a bi-level signature scheme, Bileve, which embeds fine-grained signature bits for integrity checks (mitigating spoofing attacks) as well as a coarse-grained signal to trace text sources when the signature is invalid (enhancing detectability) via a novel rank-based sampling strategy. Compared to conventional watermark detectors that only output binary results, Bileve can differentiate 5 scenarios during detection, reliably tracing text provenance and regulating LLMs. The experiments conducted on OPT-1.3B and LLaMA-7B demonstrate the effectiveness of Bileve in defeating spoofing attacks with enhanced detectability.


Supplementary Material I2DFormer: Learning Image to Document Attention for Zero-Shot Image Classification Yongqin Xian 1 Federico Tombari

Neural Information Processing Systems

In this supplementary, we additionally report the following: Section 1.1: Additional ablation results over the two modules of our model. Section 2: Examples of collected documents. Our comparisons in the main paper use the global score for inference as it is computationally fast and can additionally provide semantic embeddings for other ZSL methods. Comparing row 1 and row 2, we see that only training the individual block already results in a competitive model. However, I2D Global achieves better performance as learning cross-modal attention is a harder task than matching global embeddings. We see that the two modules of I2DFormer have a symbiotic relationship where both greatly benefit from joint training and achieve a boost in performance.


A Proof of Theorem 4.1

Neural Information Processing Systems

Note the first and third equalities are driven by Bayes' theorem, the second one is from the independence between bY and X The inequality in Equation A.8 is driven by Jensen's inequality. While our datasets and analyses reveal the relationship between CF and GF in image classification, we clarify our study's limitations. First of all, our study uses sex as a sensitive attribute based on visually perceived biological traits. However, as mentioned in Section 2, this simplification does not capture the full spectrum of sexual traits, which is more complex and nuanced. Therefore, we emphasize again that practitioners should use our data with these considerations in mind; they should not utilize our datasets for gender categorization but rather for investigating the unfairness in terms of CF and GF and enhancing fairness in AI systems. Second, our data generation process relies on IP2P to create CTF samples. The counterfactual (CTF) images regarding the "sex attribute" are shown. The top row shows the original image, while the bottom row displays the CTF image generated by IP2P. Original and CTF samples are shown when age or skin color is considered as the sensitive attribute. Third, while we assume the structural causal model (SCM) for images as Figure 2, specifying an SCM in the real world is often infeasible. This difficulty also makes it challenging to apply some of our experiments, such as analyzing the attribute G or using a robust teacher model to G. User Interface shown to five annotators for image filtering.


Sangwon Jung

Neural Information Processing Systems

The notion of algorithmic fairness has been actively explored from various aspects of fairness, such as counterfactual fairness (CF) and group fairness (GF). However, the exact relationship between CF and GF remains to be unclear, especially in image classification tasks; the reason is because we often cannot collect counterfactual samples regarding a sensitive attribute, essential for evaluating CF, from the existing images (e.g., a photo of the same person but with different secondary sex characteristics). In this paper, we construct new image datasets for evaluating CF by using a high-quality image editing method and carefully labeling with human annotators. Our datasets, CelebA-CF and LFW-CF, build upon the popular image GF benchmarks; hence, we can evaluate CF and GF simultaneously. We empirically observe that CF does not imply GF in image classification, whereas previous studies on tabular datasets observed the opposite. We theoretically show that it could be due to the existence of a latent attribute G that is correlated with, but not caused by, the sensitive attribute (e.g., secondary sex characteristics are highly correlated with hair length). From this observation, we propose a simple baseline, Counterfactual Knowledge Distillation (CKD), to mitigate such correlation with the sensitive attributes. Extensive experimental results on CelebA-CF and LFW-CF demonstrate that CF-achieving models satisfy GF if we successfully reduce the reliance on G (e.g., using CKD).


Unbalanced Optimal Transport through Non-negative Penalized Linear Regression

Neural Information Processing Systems

This paper addresses the problem of Unbalanced Optimal Transport (UOT) in which the marginal conditions are relaxed (using weighted penalties in lieu of equality) and no additional regularization is enforced on the OT plan. In this context, we show that the corresponding optimization problem can be reformulated as a non-negative penalized linear regression problem. This reformulation allows us to propose novel algorithms inspired from inverse problems and nonnegative matrix factorization. In particular, we consider majorization-minimization which leads in our setting to efficient multiplicative updates for a variety of penalties. Furthermore, we derive for the first time an efficient algorithm to compute the regularization path of UOT with quadratic penalties. The proposed algorithm provides a continuity of piece-wise linear OT plans converging to the solution of balanced OT (corresponding to infinite penalty weights). We perform several numerical experiments on simulated and real data illustrating the new algorithms, and provide a detailed discussion about more sophisticated optimization tools that can further be used to solve OT problems thanks to our reformulation.


Trump's Crackdown on Foreign Student Visas Could Derail Critical AI Research

WIRED

Secretary of State Marco Rubio said Wednesday that the US plans to "aggressively revoke" the visas of Chinese students, including those working in critical fields or with ties to the Chinese Communist Party. Experts warn the move--along with the Trump administration's broader crackdown on international students--could drain American scientific labs of top STEM talent and upend cutting-edge research in areas like artificial intelligence. "If you were aiming to help China beat the US at AI, the first thing you would do is disrupt the flow of top talent from all around the world into the US," says Helen Toner, director of strategy and foundational research grants at Georgetown University's Center for Security and Emerging Technology. While it has a population only about a quarter the size of China, "the US has had a huge asymmetric advantage in attracting the cream of the global crop," she adds. Several close Trump allies, including Elon Musk, have argued that attracting the best engineers from around the world is essential for the US to maintain its technological dominance.