Government


under the water A global multi-temporal satellite dataset for rapid flood mapping Maria Sdraka

Neural Information Processing Systems

Global flash floods, exacerbated by climate change, pose severe threats to human life, infrastructure, and the environment. Recent catastrophic events in Pakistan and New Zealand underscore the urgent need for precise flood mapping to guide restoration efforts, understand vulnerabilities, and prepare for future occurrences. While Synthetic Aperture Radar (SAR) remote sensing offers day-and-night, all-weather imaging capabilities, its application in deep learning for flood segmentation is limited by the lack of large annotated datasets. To address this, we introduce Kuro Siwo, a manually annotated multi-temporal dataset, spanning 43 flood events globally.


I'm a Public-School English Teacher. The Most Vocal Defenders of Kโ€“12 Liberal Arts Are Not Who You'd Expect.

Slate

Sign up for the Slatest to get the most insightful analysis, criticism, and advice out there, delivered to your inbox daily. On May 6, the Texas House Committee on Public Education discussed S.B. 13, a bill seeking to remove from public school libraries and classrooms all "profane" and "indecent content." At the hearing, Republican Rep. Terri Leo-Wilson focused on the concern that the legislation could harm the transmission of cultural heritage by depriving students of "classics." She explained, using an adjective that in our current culture wars has come to describe a type of humanities education favored by conservatives, that her "kids were classically trained, so they had their graduation picture with all sorts of books โ€ฆ classic works of literature." When an activist commenting during the hearing remarked that among renowned writers, Toni Morrison's work is singularly "very sexualized," Leo-Wilson replied, without reference to any one book, "She might be famous, but that's not considered, I don't think, a classic."


Separations in the Representational Capabilities of Transformers and Recurrent Architectures Michael Hahn 2 Phil Blunsom 1,3

Neural Information Processing Systems

Transformer architectures have been widely adopted in foundation models. Due to their high inference costs, there is renewed interest in exploring the potential of efficient recurrent architectures (RNNs). In this paper, we analyze the differences in the representational capabilities of Transformers and RNNs across several tasks of practical relevance, including index lookup, nearest neighbor, recognizing bounded Dyck languages, and string equality. For the tasks considered, our results show separations based on the size of the model required for different architectures. For example, we show that a one-layer Transformer of logarithmic width can perform index lookup, whereas an RNN requires a hidden state of linear size. Conversely, while constant-size RNNs can recognize bounded Dyck languages, we show that one-layer Transformers require a linear size for this task. Furthermore, we show that two-layer Transformers of logarithmic size can perform decision tasks such as string equality or disjointness, whereas both one-layer Transformers and recurrent models require linear size for these tasks. We also show that a log-size two-layer Transformer can implement the nearest neighbor algorithm in its forward pass; on the other hand recurrent models require linear size. Our constructions are based on the existence of N nearly orthogonal vectors in O(log N) dimensional space and our lower bounds are based on reductions from communication complexity problems.


CoSy: Evaluating Textual Explanations of Neurons

Neural Information Processing Systems

A crucial aspect of understanding the complex nature of Deep Neural Networks (DNNs) is the ability to explain learned concepts within their latent representations. While methods exist to connect neurons to human-understandable textual descriptions, evaluating the quality of these explanations is challenging due to the lack of a unified quantitative approach.


Logical Activation Functions: Logit-space equivalents of Probabilistic Boolean Operators, Jason d'Eon

Neural Information Processing Systems

The choice of activation functions and their motivation is a long-standing issue within the neural network community. Neuronal representations within artificial neural networks are commonly understood as logits, representing the log-odds score of presence of features within the stimulus. We derive logit-space operators equivalent to probabilistic Boolean logic-gates AND, OR, and XNOR for independent probabilities. Such theories are important to formalize more complex dendritic operations in real neurons, and these operations can be used as activation functions within a neural network, introducing probabilistic Boolean-logic as the core operation of the neural network. Since these functions involve taking multiple exponents and logarithms, they are computationally expensive and not well suited to be directly used within neural networks.


Assessing Social and Intersectional Biases in Contextualized Word Representations

Neural Information Processing Systems

Social bias in machine learning has drawn significant attention, with work ranging from demonstrations of bias in a multitude of applications, curating definitions of fairness for different contexts, to developing algorithms to mitigate bias. In natural language processing, gender bias has been shown to exist in context-free word embeddings. Recently, contextual word representations have outperformed word embeddings in several downstream NLP tasks. These word representations are conditioned on their context within a sentence, and can also be used to encode the entire sentence. In this paper, we analyze the extent to which state-of-the-art models for contextual word representations, such as BERT and GPT-2, encode biases with respect to gender, race, and intersectional identities. Towards this, we propose assessing bias at the contextual word level. This novel approach captures the contextual effects of bias missing in context-free word embeddings, yet avoids confounding effects that underestimate bias at the sentence encoding level. We demonstrate evidence of bias at the corpus level, find varying evidence of bias in embedding association tests, show in particular that racial bias is strongly encoded in contextual word models, and observe that bias effects for intersectional minorities are exacerbated beyond their constituent minority identities. Further, evaluating bias effects at the contextual word level captures biases that are not captured at the sentence level, confirming the need for our novel approach.


Trump reverses course on Middle East tech policy, but will it be enough to counter China?

FOX News

National security and military analyst Dr. Rebecca Grant joins'Fox & Friends First' to discuss President Donald Trump's historic business-focused trip to the Middle East and why a Trump-Putin meeting could be essential for peace in Ukraine. President Donald Trump secured 2 trillion worth of deals with Saudi Arabia, Qatar and the UAE during his trip to the Middle East last week in what some have argued is a move to counter China's influence in the region. While China has increasingly bolstered its commercial ties with top Middle Eastern nations who have remained steadfast in their refusal to pick sides amid growing geopolitical tension between Washington and Beijing, Trump may have taken steps to give the U.S. an edge over its chief competitor. But concern has mounted after Trump reversed a Biden-era policy โ€“ which banned the sale of AI-capable chips to the UAE and Saudi Arabia โ€“ that highly coveted U.S. technologies could potentially fall into the hands of Chinese companies, and in extension, the Chinese Communist Party (CCP). U.S. President Donald Trump walks with Saudi Crown Prince Mohammed Bin Salman during a welcoming ceremony in Riyadh, Saudi Arabia, May 13, 2025.


Bringing Image Structure to Video via Frame-Clip Consistency of Object Tokens

Neural Information Processing Systems

Recent action recognition models have achieved impressive results by integrating objects, their locations and interactions. However, obtaining dense structured annotations for each frame is tedious and time-consuming, making these methods expensive to train and less scalable. On the other hand, one does often have access to a small set of annotated images, either within or outside the domain of interest. Here we ask how such images can be leveraged for downstream video understanding tasks. We propose a learning framework StructureViT (SViT for short), which demonstrates how utilizing the structure of a small number of images only available during training can improve a video model.


MACD: Multilingual Abusive Comment Detection at Scale for Indic Languages

Neural Information Processing Systems

Social media platforms were conceived to act as online'town squares' where people could get together, share information and communicate with each other peacefully. However, harmful content borne out of bad actors are constantly plaguing these platforms slowly converting them into'mosh pits' where the bad actors take the liberty to extensively abuse various marginalised groups. Accurate and timely detection of abusive content on social media platforms is therefore very important for facilitating safe interactions between users. However, due to the small scale and sparse linguistic coverage of Indic abusive speech datasets, development of such algorithms for Indic social media users (one-sixth of global population) is severely impeded.


Congress Passed a Sweeping Free-Speech Crackdown--and No One's Talking About It

Slate

Sign up for the Slatest to get the most insightful analysis, criticism, and advice out there, delivered to your inbox daily. Had you scanned any of the latest headlines around the TAKE IT DOWN Act, legislation that President Donald Trump signed into law Monday, you would have come away with a deeply mistaken impression of the bill and its true purpose. The surface-level pitch is that this is a necessary law for addressing nonconsensual intimate images--known more widely as revenge porn. Obfuscating its intent with a classic congressional acronym (Tools to Address Known Exploitation by Immobilizing Technological Deepfakes on Websites and Networks), the TAKE IT DOWN Act purports to help scrub the internet of exploitative, nonconsensual sexual media, whether real or digitally mocked up, at a time when artificial intelligence tools and automated image generators have supercharged its spread. Enforcement is delegated to the Federal Trade Commission, which will give online communities that specialize primarily in user-generated content (e.g., social media, message boards) a heads-up and a 48-hour takedown deadline whenever an appropriate example is reported.