Goto

Collaborating Authors

 dmin


OptimizingGeneralizedPageRankMethodsfor Seed-ExpansionCommunityDetection

Neural Information Processing Systems

PageRank (PR), an algorithm originally proposed by Page et al. for ranking web-pages [1] has found manysuccessful applications, including community detection [2,3],linkprediction [4]and recommendersystemdesign[5,6].


NeurIPS2021_ImperfectCommmunicationBandits

Madhu

Neural Information Processing Systems

Networked1).LetG=(V,E)beaconnected, thecommunicationE contains (i, j) ifagentsi and j can directlyviamessagest, eachagentj broadcastsmj(t) toalltheirneighbors. times G, after discarded.



Asymptotically optimal regret in communicating Markov decision processes

Boone, Victor

arXiv.org Machine Learning

In this paper, we present a learning algorithm that achieves asymptotically optimal regret for Markov decision processes in average reward under a communicating assumption. That is, given a communicating Markov decision process $M$, our algorithm has regret $K(M) \log(T) + \mathrm{o}(\log(T))$ where $T$ is the number of learning steps and $K(M)$ is the best possible constant. This algorithm works by explicitly tracking the constant $K(M)$ to learn optimally, then balances the trade-off between exploration (playing sub-optimally to gain information), co-exploration (playing optimally to gain information) and exploitation (playing optimally to score maximally). We further show that the function $K(M)$ is discontinuous, which is a consequence challenge for our approach. To that end, we describe a regularization mechanism to estimate $K(M)$ with arbitrary precision from empirical data.


DMin: Scalable Training Data Influence Estimation for Diffusion Models

Lin, Huawei, Lao, Yingjie, Zhao, Weijie

arXiv.org Artificial Intelligence

Identifying the training data samples that most influence a generated image is a critical task in understanding diffusion models, yet existing influence estimation methods are constrained to small-scale or LoRA-tuned models due to computational limitations. As diffusion models scale up, these methods become impractical. To address this challenge, we propose DMin (Diffusion Model influence), a scalable framework for estimating the influence of each training data sample on a given generated image. By leveraging efficient gradient compression and retrieval techniques, DMin reduces storage requirements from 339.39 TB to only 726 MB and retrieves the top-k most influential training samples in under 1 second, all while maintaining performance. Our empirical results demonstrate DMin is both effective in identifying influential training samples and efficient in terms of computational and storage requirements.