Goto

Collaborating Authors

 iffusion





X-Diffusion: Training Diffusion Policies on Cross-Embodiment Human Demonstrations

Pace, Maximus A., Dan, Prithwish, Ning, Chuanruo, Bhardwaj, Atiksh, Du, Audrey, Duan, Edward W., Ma, Wei-Chiu, Kedia, Kushal

arXiv.org Artificial Intelligence

Human videos can be recorded quickly and at scale, making them an appealing source of training data for robot learning. However, humans and robots differ fundamentally in embodiment, resulting in mismatched action execution. Direct kinematic retargeting of human hand motion can therefore produce actions that are physically infeasible for robots. Despite these low-level differences, human demonstrations provide valuable motion cues about how to manipulate and interact with objects. Our key idea is to exploit the forward diffusion process: as noise is added to actions, low-level execution differences fade while high-level task guidance is preserved. We present X-Diffusion, a principled framework for training diffusion policies that maximally leverages human data without learning dynamically infeasible motions. X-Diffusion first trains a classifier to predict whether a noisy action is executed by a human or robot. Then, a human action is incorporated into policy training only after adding sufficient noise such that the classifier cannot discern its embodiment. Actions consistent with robot execution supervise fine-grained denoising at low noise levels, while mismatched human actions provide only coarse guidance at higher noise levels. Our experiments show that naive co-training under execution mismatches degrades policy performance, while X-Diffusion consistently improves it. Across five manipulation tasks, X-Diffusion achieves a 16% higher average success rate than the best baseline. The project website is available at https://portal-cornell.github.io/X-Diffusion/.


SyncDiffusion: Coherent Montage via Synchronized Joint Diffusions -- Supplementary Material Y useung Lee Kunho Kim Hyunjin Kim Minhyuk Sung KAIST {phillip0701,kaist984,rlaguswls98,mhsung }@kaist.ac.kr

Neural Information Processing Systems

S.2 includes a detailed quantitative evaluation of our method with different gradient S.3 shows quantitative evaluation of our method on S.4, we show the comparisons of our method S.5 shows an ablation study result substituting Sec. S.7 explains the details of our user study. More qualitative results with various prompts are shown in the figures below. S2 shows the detailed quantitative results of S Intra-Style-L displayed in Fig. S2. Fig. S1 shows the qualitative comparison of the panorama images We show the quantitative results on different resolutions in Tab. Figure S2: Line plots of the quantitative results shown in Tab.




Could AI Trace and Explain the Origins of AI-Generated Images and Text?

Fang, Hongchao, Liu, Yixin, Du, Jiangshu, Qin, Can, Xu, Ran, Liu, Feng, Sun, Lichao, Lee, Dongwon, Huang, Lifu, Yin, Wenpeng

arXiv.org Artificial Intelligence

AI-generated content is becoming increasingly prevalent in the real world, leading to serious ethical and societal concerns. For instance, adversaries might exploit large multimodal models (LMMs) to create images that violate ethical or legal standards, while paper reviewers may misuse large language models (LLMs) to generate reviews without genuine intellectual effort. While prior work has explored detecting AI-generated images and texts, and occasionally tracing their source models, there is a lack of a systematic and fine-grained comparative study. Important dimensions--such as AI-generated images vs. text, fully vs. partially AI-generated images, and general vs. malicious use cases--remain underexplored. Furthermore, whether AI systems like GPT-4o can explain why certain forged content is attributed to specific generative models is still an open question, with no existing benchmark addressing this. To fill this gap, we introduce AI-FAKER, a comprehensive multimodal dataset with over 280,000 samples spanning multiple LLMs and LMMs, covering both general and malicious use cases for AI-generated images and texts. Our experiments reveal two key findings: (i) AI authorship detection depends not only on the generated output but also on the model's original training intent; and (ii) GPT-4o provides highly consistent but less specific explanations when analyzing content produced by OpenAI's own models, such as DALL-E and GPT-4o itself.


Unlocking Point Processes through Point Set Diffusion

Lüdke, David, Raventós, Enric Rabasseda, Kollovieh, Marcel, Günnemann, Stephan

arXiv.org Machine Learning

Point processes model the distribution of random point sets in mathematical spaces, such as spatial and temporal domains, with applications in fields like seismology, neuroscience, and economics. Existing statistical and machine learning models for point processes are predominantly constrained by their reliance on the characteristic intensity function, introducing an inherent trade-off between efficiency and flexibility. In this paper, we introduce Point Set Diffusion, a diffusion-based latent variable model that can represent arbitrary point processes on general metric spaces without relying on the intensity function. By directly learning to stochastically interpolate between noise and data point sets, our approach enables efficient, parallel sampling and flexible generation for complex conditional tasks defined on the metric space. Experiments on synthetic and real-world datasets demonstrate that Point Set Diffusion achieves state-of-the-art performance in unconditional and conditional generation of spatial and spatiotemporal point processes while providing up to orders of magnitude faster sampling than autoregressive baselines.


ClickDiffusion: Harnessing LLMs for Interactive Precise Image Editing

Helbling, Alec, Lee, Seongmin, Chau, Polo

arXiv.org Artificial Intelligence

Recently, researchers have proposed powerful systems for generating and manipulating images using natural language instructions. However, it is difficult to precisely specify many common classes of image transformations with text alone. For example, a user may wish to change the location and breed of a particular dog in an image with several similar dogs. This task is quite difficult with natural language alone, and would require a user to write a laboriously complex prompt that both disambiguates the target dog and describes the destination. We propose ClickDiffusion, a system for precise image manipulation and generation that combines natural language instructions with visual feedback provided by the user through a direct manipulation interface. We demonstrate that by serializing both an image and a multi-modal instruction into a textual representation it is possible to leverage LLMs to perform precise transformations of the layout and appearance of an image. Code available at https://github.com/poloclub/ClickDiffusion.