red list
StealthInk: A Multi-bit and Stealthy Watermark for Large Language Models
Jiang, Ya, Wu, Chuxiong, Boroujeny, Massieh Kordi, Mark, Brian, Zeng, Kai
Watermarking for large language models (LLMs) offers a promising approach to identifying AI-generated text. Existing approaches, however, either compromise the distribution of original generated text by LLMs or are limited to embedding zero-bit information that only allows for watermark detection but ignores identification. We present StealthInk, a stealthy multi-bit watermarking scheme that preserves the original text distribution while enabling the embedding of provenance data, such as userID, TimeStamp, and modelID, within LLM-generated text. This enhances fast traceability without requiring access to the language model's API or prompts. We derive a lower bound on the number of tokens necessary for watermark detection at a fixed equal error rate, which provides insights on how to enhance the capacity. Comprehensive empirical evaluations across diverse tasks highlight the stealthiness, detectability, and resilience of StealthInk, establishing it as an effective solution for LLM watermarking applications.
- Pacific Ocean (0.04)
- North America > United States > Virginia (0.04)
- North America > United States > Maryland > Baltimore (0.04)
- (3 more...)
A Watermark for Low-entropy and Unbiased Generation in Large Language Models
Mao, Minjia, Wei, Dongjun, Chen, Zeyu, Fang, Xiao, Chau, Michael
Recent advancements in large language models (LLMs) have highlighted the risk of misuse, raising concerns about accurately detecting LLM-generated content. A viable solution for the detection problem is to inject imperceptible identifiers into LLMs, known as watermarks. Previous work demonstrates that unbiased watermarks ensure unforgeability and preserve text quality by maintaining the expectation of the LLM output probability distribution. However, previous unbiased watermarking methods are impractical for local deployment because they rely on accesses to white-box LLMs and input prompts during detection. Moreover, these methods fail to provide statistical guarantees for the type II error of watermark detection. This study proposes the Sampling One Then Accepting (STA-1) method, an unbiased watermark that does not require access to LLMs nor prompts during detection and has statistical guarantees for the type II error. Moreover, we propose a novel tradeoff between watermark strength and text quality in unbiased watermarks. We show that in low-entropy scenarios, unbiased watermarks face a tradeoff between watermark strength and the risk of unsatisfactory outputs. Experimental results on low-entropy and high-entropy datasets demonstrate that STA-1 achieves text quality and watermark strength comparable to existing unbiased watermarks, with a low risk of unsatisfactory outputs. Implementation codes for this study are available online.
- North America > United States > Colorado (0.04)
- North America > United States > New York (0.04)
- North America > United States > Louisiana > Orleans Parish > New Orleans (0.04)
- Asia > China > Hong Kong (0.04)
- Information Technology > Security & Privacy (0.90)
- Leisure & Entertainment > Sports > Baseball (0.46)
X-Mark: Towards Lossless Watermarking Through Lexical Redundancy
Chen, Liang, Bian, Yatao, Deng, Yang, Li, Shuaiyi, Wu, Bingzhe, Zhao, Peilin, Wong, Kam-fai
Text watermarking has emerged as an important technique for detecting machine-generated text. However, existing methods can severely degrade text quality due to arbitrary vocabulary partitioning, which disrupts the language model's expressiveness and impedes textual coherence. To mitigate this, we introduce XMark, a novel approach that capitalizes on text redundancy within the lexical space. Specifically, XMark incorporates a mutually exclusive rule for synonyms during the language model decoding process, thereby integrating prior knowledge into vocabulary partitioning and preserving the capabilities of language generation. We present theoretical analyses and empirical evidence demonstrating that XMark substantially enhances text generation fluency while maintaining watermark detectability. Furthermore, we investigate watermarking's impact on the emergent abilities of large language models, including zero-shot and few-shot knowledge recall, logical reasoning, and instruction following. Our comprehensive experiments confirm that XMark consistently outperforms existing methods in retaining these crucial capabilities of LLMs.
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > New York (0.04)
- Asia > Singapore (0.04)
- Asia > China > Hong Kong (0.04)
A Watermark for Large Language Models
Kirchenbauer, John, Geiping, Jonas, Wen, Yuxin, Katz, Jonathan, Miers, Ian, Goldstein, Tom
Potential harms of large language models can be mitigated by watermarking model output, i.e., embedding signals into generated text that are invisible to humans but algorithmically detectable from a short span of tokens. We propose a watermarking framework for proprietary language models. The watermark can be embedded with negligible impact on text quality, and can be detected using an efficient open-source algorithm without access to the language model API or parameters. The watermark works by selecting a randomized set of "green" tokens before a word is generated, and then softly promoting use of green tokens during sampling. We propose a statistical test for detecting the watermark with interpretable p-values, and derive an information-theoretic framework for analyzing the sensitivity of the watermark. We test the watermark using a multi-billion parameter model from the Open Pretrained Transformer (OPT) family, and discuss robustness and security.
- Africa > Middle East > Egypt (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- Asia > Middle East > UAE (0.04)
- (10 more...)
- Information Technology > Security & Privacy (1.00)
- Education (1.00)
- Leisure & Entertainment > Sports > Skiing (0.67)
- Government > Regional Government > North America Government > United States Government (0.45)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.96)
How machine learning could help save threatened species from extinction
There are thousands of species on Earth that we still don't know much about -- but we now know that they are already teetering on the edge of extinction. A new study used machine learning to figure out just how threatened these lesser-known species are, and the results were grim. Some species of animals and plants are labeled "data deficient" because conservationists haven't been able to gather enough information about them to understand how they live or how many of them are left. It turns out that those "data deficient" species are unfortunately even more threatened than other species that are more well known (to scientists, at least). The data from this study came from the International Union for Conservation of Nature (IUCN), which maintains a global "Red List" that ranks species based on how threatened they are.
Machine learning could improve plant conservation efforts
A new paper published on 3 December in Proceedings of the National Academy of Sciences claims that a large number of currently unassessed plant species are likely at risk (1). The researchers also identified several geographic regions with the highest need for conservation efforts. Moreover, several of these regions are not currently recognized as areas of global concern. According to the authors, 10 per cent of plant species should be categorised as "at risk" on the Red List of Threatened Species, a comprehensive inventory of the global conservation status of biological species maintained by the International Union for Conservation of Nature (IUCN). This equates to nearly 15,000 additional species.
- Oceania > Australia (0.06)
- North America > United States > Ohio (0.06)
- North America > United States > Maryland (0.06)
- Information Technology > Artificial Intelligence > Machine Learning (0.71)
- Information Technology > Data Science > Data Mining > Big Data (0.51)
Dozens Of Polar Bears Feast On Whale Carcass In Unusual Group Behavior
As climate change continues to cause a reduction in Arctic sea ice and overall ice cover in the polar region, the already threatened polar bears are beginning to display highly unusual behavior. Largely solitary animals in their adult life, dozens of them were seen together recently on an island in northeast Russia. A tourist boat passing by Wrangel Island, off the coast of Chukotka in Russia's Far East, saw over 200 polar bears on a mountain slope on the island. Dozens of the animals were seen at the bottom of the slope, eating the carcass of a bowhead whale that had washed ashore. The incident took place in September, but wasn't widely reported at the time.