Goto

Collaborating Authors

 Nepal


Pentagon seeks 75 billion for drones in record budget ask

The Japan Times

A soldier carries a drone during a military parade in Washington on June 14, 2025. The Pentagon's largest-ever budget request earmarks $75 billion for drones and technologies to counter them, mainly for a massive increase for a little-known office working with U.S. commandos to test and evaluate various systems, according to defense officials. The drone-funding proposal includes $54.6 billion for the Defense Autonomous Working Group, or DAWG, from just $225.9 million this year. That would appear to be the largest single year-over-year boost of any defense program or office, meaning it's likely to draw particular congressional and public scrutiny in an already eye-catching $1.5 trillion request that's 42% larger than this year's budget. The big boost for the Pentagon's little-known drone unit comes as the U.S. and Israeli war against Iran illustrates how drones can help level the playing field against even the world's most well-funded armed forces.


Enhancing Online Support Group Formation Using Topic Modeling Techniques

Barman, Pronob Kumar, Reynolds, Tera L., Foulds, James

arXiv.org Machine Learning

Online health communities (OHCs) are vital for fostering peer support and improving health outcomes. Support groups within these platforms can provide more personalized and cohesive peer support, yet traditional support group formation methods face challenges related to scalability, static categorization, and insufficient personalization. To overcome these limitations, we propose two novel machine learning models for automated support group formation: the Group specific Dirichlet Multinomial Regression (gDMR) and the Group specific Structured Topic Model (gSTM). These models integrate user generated textual content, demographic profiles, and interaction data represented through node embeddings derived from user networks to systematically automate personalized, semantically coherent support group formation. We evaluate the models on a large scale dataset from MedHelp, comprising over 2 million user posts. Both models substantially outperform baseline methods including LDA, DMR, and STM in predictive accuracy (held out log likelihood), semantic coherence (UMass metric), and internal group consistency. The gDMR model yields group covariates that facilitate practical implementation by leveraging relational patterns from network structures and demographic data. In contrast, gSTM emphasizes sparsity constraints to generate more distinct and thematically specific groups. Qualitative analysis further validates the alignment between model generated groups and manually coded themes, showing the practical relevance of the models in informing groups that address diverse health concerns such as chronic illness management, diagnostic uncertainty, and mental health. By reducing reliance on manual curation, these frameworks provide scalable solutions that enhance peer interactions within OHCs, with implications for patient engagement, community resilience, and health outcomes.



Breaking the False Sense of Security in Backdoor Defense through Re-Activation Attack

Neural Information Processing Systems

To further verify this finding, we empirically show that these dormant backdoors can be easily re-activated during inference stage, by manipulating the original trigger with well-designed tiny perturbation using universal adversarial attack.






Label Poisoning is All You Need

Neural Information Processing Systems

In a backdoor attack, an adversary injects corrupted data into a model's training dataset in order to gain control over its predictions on images with a specific attacker-defined trigger. A typical corrupted training example requires altering both the image, by applying the trigger, and the label. Models trained on clean images, therefore, were considered safe from backdoor attacks. However, in some common machine learning scenarios, the training labels are provided by potentially malicious third-parties. This includes crowd-sourced annotation and knowledge distillation. We, hence, investigate a fundamental question: can we launch a successful backdoor attack by only corrupting labels?