Plotting


Discrete Modeling via Boundary Conditional Diffusion Processes

Neural Information Processing Systems

We present an novel framework for efficiently and effectively extending the powerful continuous diffusion processes to discrete modeling. Previous approaches have suffered from the discrepancy between discrete data and continuous modeling. Our study reveals that the absence of guidance from discrete boundaries in learning probability contours is one of the main reasons. To address this issue, we propose a two-step forward process that first estimates the boundary as a prior distribution and then rescales the forward trajectory to construct a boundary conditional diffusion model. The reverse process is proportionally adjusted to guarantee that the learned contours yield more precise discrete data. Experimental results indicate that our approach achieves strong performance in both language modeling and discrete image generation tasks. In language modeling, our approach surpasses previous state-of-the-art continuous diffusion language models in three translation tasks and a summarization task, while also demonstrating competitive performance compared to auto-regressive transformers.


Rotation-Equivariant Conditional Spherical Neural Fields for Learning a Natural Illumination Prior-Supplementary Material-James A. D. Gardner

Neural Information Processing Systems

Here we provide further results, additional training information, and explore the how RENI performs when fit to illumination maps that contain unnatural illumination conditions. A.1 Cosine Loss When optimising just the latent codes to unseen environment maps, we found improved performance when including a cosine similarity loss. This loss is used again during the inverse rendering task. However as the images used in that loss are not equirectangular the sin(θ(d)) term is not included. A.2 Gamma Correction For display, all linear HDR images I had their gamma adjusted using the following process: 1. Adjust exposure to set the white level to the p-th percentile (p = 98) I I percentile(I, p) 2. Clamp between [0, 1] I clamp(I, 0, 1) 3. Apply gamma correction using the standard sRGB gamma curve: { We include additional qualitative results of the RENI model.


Israeli drone strikes kill four people in southern Lebanon

Al Jazeera

Lebanon's Ministry of Public Health has said that at least four people have been killed in two separate Israeli strikes in south Lebanon, as Israel claimed it struck Hezbollah operatives. Thursday's strikes were the latest in a series of deadly attacks in south Lebanon, despite a November ceasefire between Israel and Hezbollah after more than a year of hostilities, including two months of open war. An "Israeli enemy strike on a car in Yohmor al-Shaqeef led to the death of three people", said a Health Ministry statement reported by the National News Agency (NNA) on Thursday. The NNA said an "enemy drone" targeted a vehicle near the town, in a strike that came at the same time as artillery shelling. Elsewhere, the Israeli military said in a statement that "several Hezbollah terrorists were identified transferring weapons in the area of Yohmor in southern Lebanon", adding that the army "struck the terrorists".


Rotation-Equivariant Conditional Spherical Neural Fields for Learning a Natural Illumination Prior

Neural Information Processing Systems

Inverse rendering is an ill-posed problem. Previous work has sought to resolve this by focussing on priors for object or scene shape or appearance. In this work, we instead focus on a prior for natural illuminations. Current methods rely on spherical harmonic lighting or other generic representations and, at best, a simplistic prior on the parameters. We propose a conditional neural field representation based on a variational auto-decoder with a SIREN network and, extending Vector Neurons, build equivariance directly into the network. Using this, we develop a rotationequivariant, high dynamic range (HDR) neural illumination model that is compact and able to express complex, high-frequency features of natural environment maps. Training our model on a curated dataset of 1.6K HDR environment maps of natural scenes, we compare it against traditional representations, demonstrate its applicability for an inverse rendering task and show environment map completion from partial observations. A PyTorch implementation, our dataset and trained models can be found at jadgardner.github.io/RENI.


Convolutional Differentiable Logic Gate Networks Hilde Kuehne Stanford University Tuebingen AI Center InftyLabs Research MIT-IBM Watson AI Lab

Neural Information Processing Systems

With the increasing inference cost of machine learning models, there is a growing interest in models with fast and efficient inference. Recently, an approach for learning logic gate networks directly via a differentiable relaxation was proposed. Logic gate networks are faster than conventional neural network approaches because their inference only requires logic gate operators such as NAND, OR, and XOR, which are the underlying building blocks of current hardware and can be efficiently executed. We build on this idea, extending it by deep logic gate tree convolutions, logical OR pooling, and residual initializations. This allows scaling logic gate networks up by over one order of magnitude and utilizing the paradigm of convolution. On CIFAR-10, we achieve an accuracy of 86.29% using only 61 million logic gates, which improves over the SOTA while being 29 smaller.


Modulated Neural ODEs

Neural Information Processing Systems

Neural ordinary differential equations (NODEs) have been proven useful for learning non-linear dynamics of arbitrary trajectories. However, current NODE methods capture variations across trajectories only via the initial state value or by autoregressive encoder updates. In this work, we introduce Modulated Neural ODEs (MoNODEs), a novel framework that sets apart dynamics states from underlying static factors of variation and improves the existing NODE methods. In particular, we introduce time-invariant modulator variables that are learned from the data. We incorporate our proposed framework into four existing NODE variants. We test MoNODE on oscillating systems, videos and human walking trajectories, where each trajectory has trajectory-specific modulation. Our framework consistently improves the existing model ability to generalize to new dynamic parameterizations and to perform far-horizon forecasting.


RankRAG: Unifying Context Ranking with Retrieval-Augmented Generation in LLMs

Neural Information Processing Systems

Large language models (LLMs) typically utilize the top-k contexts from a retriever in retrieval-augmented generation (RAG). In this work, we propose a novel instruction fine-tuning framework RankRAG, which instruction-tunes a single LLM for the dual purpose of context ranking and answer generation in RAG. In particular, the instruction-tuned LLMs work surprisingly well by adding a small fraction of ranking data into the training blend, and outperform existing expert ranking models, including the same LLM exclusively fine-tuned on a large amount of ranking data. For generation, we compare our model with many strong baselines, including GPT-4-0613, GPT-4-turbo-2024-0409, and ChatQA-1.5,



Suitable is the Best: Task-Oriented Knowledge Fusion in Vulnerability Detection

Neural Information Processing Systems

Deep learning technologies have demonstrated remarkable performance in vulnerability detection. Existing works primarily adopt a uniform and consistent feature learning pattern across the entire target set. While designed for general-purpose detection tasks, they lack sensitivity towards target code comprising multiple functional modules or diverse vulnerability subtypes. In this paper, we present a knowledge fusion-based vulnerability detection method (KF-GVD) that integrates specific vulnerability knowledge into the Graph Neural Network feature learning process. KF-GVD achieves accurate vulnerability detection across different functional modules of the Linux kernel and vulnerability subtypes without compromising general task performance. Extensive experiments demonstrate that KF-GVD outperforms SOTAs on function-level and statement-level vulnerability detection across various target tasks, with an average increase of 40.9% in precision and 26.1% in recall. Notably, KF-GVD discovered 9 undisclosed vulnerabilities when employing on C/C++ open-source projects without ground truth.