Goto

Collaborating Authors

 Sevastopol


Russia-Ukraine war: List of key events, day 1,457

Al Jazeera

How the US left Ukraine exposed to Russia's winter war Will Europe use frozen Russian assets to fund war? How can Ukraine rebuild China ties? Russian forces launched 448 attacks on 34 settlements in Ukraine's front-line Zaporizhia region in a single day, injuring a six-year-old child and damaging homes, cars and other infrastructure, regional governor Ivan Fedorov wrote on the Telegram app. Russian drone, missile and artillery attacks on Ukraine's Kherson region injured five people and damaged homes, including seven high-rise buildings, the local military administration said on Telegram. Russian attacks also continued in Ukraine's Dnipropetrovsk and Sumy regions, but local officials there noted that "fortunately, no people were injured".


Ukraine says it carried out first-ever underwater drone strike on Russian submarine in Novorossiysk

FOX News

This material may not be published, broadcast, rewritten, or redistributed. Quotes displayed in real-time or delayed by at least 15 minutes. Market data provided by Factset . Powered and implemented by FactSet Digital Solutions . Mutual Fund and ETF data provided by Refinitiv Lipper .


Democratizing Diplomacy: A Harness for Evaluating Any Large Language Model on Full-Press Diplomacy

Duffy, Alexander, Paech, Samuel J, Shastri, Ishana, Karpinski, Elizabeth, Alloui-Cros, Baptiste, Marques, Tyler, Olson, Matthew Lyle

arXiv.org Artificial Intelligence

We present the first evaluation harness that enables any out-of-the-box, local, Large Language Models (LLMs) to play full-press Diplomacy without fine-tuning or specialized training. Previous work required frontier LLMs, or fine-tuning, due to the high complexity and information density of Diplomacy's game state. Combined with the high variance of matches, these factors made Diplomacy prohibitive for study. In this work, we used data-driven iteration to optimize a textual game state representation such that a 24B model can reliably complete matches without any fine tuning. We develop tooling to facilitate hypothesis testing and statistical analysis, and we present case studies on persuasion, aggressive playstyles, and performance across a range of models. We conduct a variety of experiments across many popular LLMs, finding the larger models perform the best, but the smaller models still play adequately. We also introduce Critical State Analysis: an experimental protocol for rapidly iterating and analyzing key moments in a game at depth. Our harness democratizes the evaluation of strategic reasoning in LLMs by eliminating the need for fine-tuning, and it provides insights into how these capabilities emerge naturally from widely used LLMs. Our code is available in the supplement and will be open sourced.


DipLLM: Fine-Tuning LLM for Strategic Decision-making in Diplomacy

Xu, Kaixuan, Chai, Jiajun, Li, Sicheng, Fu, Yuqian, Zhu, Yuanheng, Zhao, Dongbin

arXiv.org Artificial Intelligence

Diplomacy is a complex multiplayer game that requires both cooperation and competition, posing significant challenges for AI systems. Traditional methods rely on equilibrium search to generate extensive game data for training, which demands substantial computational resources. Large Language Models (LLMs) offer a promising alternative, leveraging pre-trained knowledge to achieve strong performance with relatively small-scale fine-tuning. However, applying LLMs to Diplomacy remains challenging due to the exponential growth of possible action combinations and the intricate strategic interactions among players. To address this challenge, we propose DipLLM, a fine-tuned LLM-based agent that learns equilibrium policies for Diplomacy. DipLLM employs an autoregressive factorization framework to simplify the complex task of multi-unit action assignment into a sequence of unit-level decisions. By defining an equilibrium policy within this framework as the learning objective, we fine-tune the model using only 1.5% of the data required by the state-of-the-art Cicero model, surpassing its performance. Our results demonstrate the potential of fine-tuned LLMs for tackling complex strategic decision-making in multiplayer games.


Ukraine's 'Spiderweb' drone assault forces Russia to shelter, move aircraft

Al Jazeera

Russia's increased sense of vulnerability may be the most important result of a recent large-scale Ukrainian drone attack named Operation Spiderweb, experts tell Al Jazeera. The operation destroyed as much as a third of Russia's strategic bomber fleet on the tarmac of four airfields deep inside Russia on June 1. Days later, Russia started to build shelters for its bombers and relocate them. An open source intelligence (OSINT) researcher nicknamed Def Mon posted time-lapse satellite photographs on social media showing major excavations at the Kirovskoe airfield in annexed Crimea as well as in Sevastopol, Gvardiyskoye and Saki, where Russia was constructing shelters for military aircraft. They reported similar work at several airbases in Russia, including the Engels base, which was targeted in Ukraine's attacks on June 1.


Ukraine bombs Russian bases: Here are some of Kyiv's most audacious attacks

Al Jazeera

Ukrainian drones struck multiple military airbases deep inside Russia on Sunday in a major operation a day before the neighbours held peace talks in Istanbul. The Russian Defence Ministry said Ukraine had launched drone strikes targeting Russian military airfields across five regions, causing several aircraft to catch fire. The attacks occurred in the Murmansk, Irkutsk, Ivanovo, Ryazan, and Amur regions. Air defences repelled the assaults in all but two regions – Murmansk and Irkutsk, the ministry said. "In the Murmansk and Irkutsk regions, the launch of FPV drones from an area in close proximity to airfields resulted in several aircraft catching fire," the Defence Ministry said.


SPIN-Bench: How Well Do LLMs Plan Strategically and Reason Socially?

Yao, Jianzhu, Wang, Kevin, Hsieh, Ryan, Zhou, Haisu, Zou, Tianqing, Cheng, Zerui, Wang, Zhangyang, Viswanath, Pramod

arXiv.org Artificial Intelligence

Reasoning and strategic behavior in social interactions is a hallmark of intelligence. This form of reasoning is significantly more sophisticated than isolated planning or reasoning tasks in static settings (e.g., math problem solving). In this paper, we present Strategic Planning, Interaction, and Negotiation (SPIN-Bench), a new multi-domain evaluation designed to measure the intelligence of strategic planning and social reasoning. While many existing benchmarks focus on narrow planning or single-agent reasoning, SPIN-Bench combines classical PDDL tasks, competitive board games, cooperative card games, and multi-agent negotiation scenarios in one unified framework. The framework includes both a benchmark as well as an arena to simulate and evaluate the variety of social settings to test reasoning and strategic behavior of AI agents. We formulate the benchmark SPIN-Bench by systematically varying action spaces, state complexity, and the number of interacting agents to simulate a variety of social settings where success depends on not only methodical and step-wise decision making, but also conceptual inference of other (adversarial or cooperative) participants. Our experiments reveal that while contemporary LLMs handle basic fact retrieval and short-range planning reasonably well, they encounter significant performance bottlenecks in tasks requiring deep multi-hop reasoning over large state spaces and socially adept coordination under uncertainty. We envision SPIN-Bench as a catalyst for future research on robust multi-agent planning, social reasoning, and human--AI teaming. Project Website: https://spinbench.github.io/


DSGBench: A Diverse Strategic Game Benchmark for Evaluating LLM-based Agents in Complex Decision-Making Environments

Tang, Wenjie, Zhou, Yuan, Xu, Erqiang, Cheng, Keyan, Li, Minne, Xiao, Liquan

arXiv.org Artificial Intelligence

Large Language Model~(LLM) based agents have been increasingly popular in solving complex and dynamic tasks, which requires proper evaluation systems to assess their capabilities. Nevertheless, existing benchmarks usually either focus on single-objective tasks or use overly broad assessing metrics, failing to provide a comprehensive inspection of the actual capabilities of LLM-based agents in complicated decision-making tasks. To address these issues, we introduce DSGBench, a more rigorous evaluation platform for strategic decision-making. Firstly, it incorporates six complex strategic games which serve as ideal testbeds due to their long-term and multi-dimensional decision-making demands and flexibility in customizing tasks of various difficulty levels or multiple targets. Secondly, DSGBench employs a fine-grained evaluation scoring system which examines the decision-making capabilities by looking into the performance in five specific dimensions and offering a comprehensive assessment in a well-designed way. Furthermore, DSGBench also incorporates an automated decision-tracking mechanism which enables in-depth analysis of agent behaviour patterns and the changes in their strategies. We demonstrate the advances of DSGBench by applying it to multiple popular LLM-based agents and our results suggest that DSGBench provides valuable insights in choosing LLM-based agents as well as improving their future development. DSGBench is available at https://github.com/DeciBrain-Group/DSGBench.


HoT: Highlighted Chain of Thought for Referencing Supporting Facts from Inputs

Nguyen, Tin, Bolton, Logan, Taesiri, Mohammad Reza, Nguyen, Anh Totti

arXiv.org Artificial Intelligence

An Achilles heel of Large Language Models (LLMs) is their tendency to hallucinate non-factual statements. A response mixed of factual and non-factual statements poses a challenge for humans to verify and accurately base their decisions on. To combat this problem, we propose Highlighted Chain-of-Thought Prompting (HoT), a technique for prompting LLMs to generate responses with XML tags that ground facts to those provided in the query. That is, given an input question, LLMs would first re-format the question to add XML tags highlighting key facts, and then, generate a response with highlights over the facts referenced from the input. Interestingly, in few-shot settings, HoT outperforms vanilla chain of thought prompting (CoT) on a wide range of 17 tasks from arithmetic, reading comprehension to logical reasoning. When asking humans to verify LLM responses, highlights help time-limited participants to more accurately and efficiently recognize when LLMs are correct. Yet, surprisingly, when LLMs are wrong, HoTs tend to make users believe that an answer is correct.


License Plate Images Generation with Diffusion Models

Shpir, Mariia, Shvai, Nadiya, Nakib, Amir

arXiv.org Artificial Intelligence

Despite the evident practical importance of license plate recognition (LPR), corresponding research is limited by the volume of publicly available datasets due to privacy regulations such as the General Data Protection Regulation (GDPR). To address this challenge, synthetic data generation has emerged as a promising approach. In this paper, we propose to synthesize realistic license plates (LPs) using diffusion models, inspired by recent advances in image and video generation. In our experiments a diffusion model was successfully trained on a Ukrainian LP dataset, and 1000 synthetic images were generated for detailed analysis. Through manual classification and annotation of the generated images, we performed a thorough study of the model output, such as success rate, character distributions, and type of failures. Our contributions include experimental validation of the efficacy of diffusion models for LP synthesis, along with insights into the characteristics of the generated data. Furthermore, we have prepared a synthetic dataset consisting of 10,000 LP images, publicly available at https://zenodo.org/doi/10.5281/zenodo.13342102. Conducted experiments empirically confirm the usefulness of synthetic data for the LPR task. Despite the initial performance gap between the model trained with real and synthetic data, the expansion of the training data set with pseudolabeled synthetic data leads to an improvement in LPR accuracy by 3% compared to baseline.