Goto

Collaborating Authors

 Shanghai


Young Chinese use AI to launch one-person firms over job anxiety

The Japan Times

One-person company SoloNest sounder Karen Dai preparing for a coffee chat at a conference room in Shanghai on April 12. | AFP-JIJI Shanghai - Young Chinese, many who fear age discrimination in their workplace after turning 35, are increasingly starting one-person companies that have artificial intelligence do most of the work. Smaller startups are already in vogue in Silicon Valley and elsewhere, with rapidly advancing AI tools seen as a welcome teammate even as they threaten layoffs at existing firms. More young people in China are subscribing to the model, as cities pledge millions of dollars in funding and rent subsidies for such ventures, in alignment with Beijing's political goal of technological self-reliance. In a time of both misinformation and too much information, quality journalism is more crucial than ever. By subscribing, you can help us get the story right.


Interview with Xinwei Song: strategic interactions in networked multi-agent systems

AIHub

In this interview series, we're meeting some of the AAAI/SIGAI Doctoral Consortium participants to find out more about their research. We hear from Xinwei Song about the two main research threads she's worked on so far, plans to expand her investigations, and what inspired her to study AI. Could you start with a quick introduction - where are you studying, and what is the topic of your research? My research primarily focuses on strategic interactions in networked multi-agent systems. Could you give us an overview of the research you've carried out so far during your PhD? My research to date consists of two main threads, which complement each other in exploring strategic interactions from different perspectives.

  Country:
  Genre: Personal > Interview (0.68)

Distributional Off-Policy Evaluation with Deep Quantile Process Regression

Kuang, Qi, Wang, Chao, Jiao, Yuling, Zhou, Fan

arXiv.org Machine Learning

This paper investigates the off-policy evaluation (OPE) problem from a distributional perspective. Rather than focusing solely on the expectation of the total return, as in most existing OPE methods, we aim to estimate the entire return distribution. To this end, we introduce a quantile-based approach for OPE using deep quantile process regression, presenting a novel algorithm called Deep Quantile Process regression-based Off-Policy Evaluation (DQPOPE). We provide new theoretical insights into the deep quantile process regression technique, extending existing approaches that estimate discrete quantiles to estimate a continuous quantile function. A key contribution of our work is the rigorous sample complexity analysis for distributional OPE with deep neural networks, bridging theoretical analysis with practical algorithmic implementations. We show that DQPOPE achieves statistical advantages by estimating the full return distribution using the same sample size required to estimate a single policy value using conventional methods. Empirical studies further show that DQPOPE provides significantly more precise and robust policy value estimates than standard methods, thereby enhancing the practical applicability and effectiveness of distributional reinforcement learning approaches.


Paula R-AI-dcliffe! Watch the moment a robot wins the Beijing half marathon - beating the human record by almost 7 minutes

Daily Mail - Science & tech

Ritzy Bay Area town torn apart after teacher's daughter, 16, was behind wheel when four friends died in high-speed crash... then she posted a TikTok video that poured fuel on the flames Two CIA officers killed in Mexico when their car skidded off ravine and exploded after meeting about bust of'largest ever drug lab' Nancy Guthrie sheriff's appalling past revealed: Beat handcuffed suspect so badly he needed intensive care, used VILE language about woman and lied in sworn statement Trump confronts Xi as US forces seize Chinese ship carrying mysterious'gift' to Iran New'Hollywood dose' pill: A-listers hooked on'youth elixir' that dermatologists say is anti-ageing, shrinks pores, smooths wrinkles... and even banishes rosacea Days after we got engaged, the love of my life told me he'd killed a man and buried him in a bog. I reported him to police... but then I made this irreversible mistake Ark of the Covenant's final resting place pinpointed by archaeologists as fresh search begins Fury as murderer marries pen pal behind bars... as teenage victim's mom says: 'I'm serving a life sentence without my son' Insiders claim failed AI rollout could be to blame for Tim Cook's departure from Apple - as one says'the AI era requires a different kind of leadership' Life-threatening cantaloupe recall in four states upgraded to FDA's highest risk level... 'reasonable probability of death' AMANDA PLATELL: Why Sarah Ferguson - with the ghost of Princess Diana at her side - is ready to sensationally blow up the Royal Family. She knows ALL their secrets... Team USA Olympics star Noah Lyles slammed for'horrible' reaction to his wife's wedding dress reveal In honour of the Queen's (purple!) reign: Kate mirrors late monarch's colourful wardrobe and wears her pearl earrings and necklace US troops board second tanker as Iran is accused of breaking ceasefire'numerous times' How to lose weight when perimenopause sabotages your metabolism: I'm a trainer but when I hit 46, I piled on the pounds overnight. The new'posh' drug that's easier to order than Uber Eats - and why all my middle-class friends have ditched booze and cocaine for it: JANA HOCKING Autistic woman, 24, worked hard to build independent life for herself... now she's PARALYZED thanks to selfishness of stranger READ MORE: McDonald's is testing humanoid ROBOTS in Shanghai During last year's shambolic Beijing robot half marathon, humanoid machines tripped, shuffled, and occasionally shattered into pieces as they collapsed under the strain. But 12 months later, supporters looked on in awe as a new generation of speedy robotic racers left the human athletes in the dust.


Universality of Gaussian-Mixture Reverse Kernels in Conditional Diffusion

Ishtiaque, Nafiz, Haque, Syed Arefinul, Alam, Kazi Ashraful, Jahara, Fatima

arXiv.org Machine Learning

We prove that conditional diffusion models whose reverse kernels are finite Gaussian mixtures with ReLU-network logits can approximate suitably regular target distributions arbitrarily well in context-averaged conditional KL divergence, up to an irreducible terminal mismatch that typically vanishes with increasing diffusion horizon. A path-space decomposition reduces the output error to this mismatch plus per-step reverse-kernel errors; assuming each reverse kernel factors through a finite-dimensional feature map, each step becomes a static conditional density approximation problem, solved by composing Norets' Gaussian-mixture theory with quantitative ReLU bounds. Under exact terminal matching the resulting neural reverse-kernel class is dense in conditional KL.


Adaptive Budget Allocation in LLM-Augmented Surveys

Ye, Zikun, Lyu, Jiameng, Tao, Rui

arXiv.org Machine Learning

Large language models (LLMs) can generate survey responses at low cost, but their reliability varies substantially across questions and is unknown before data collection. Deploying LLMs in surveys still requires costly human responses for verification and correction. How should a limited human-labeling budget be allocated across questions in real time? We propose an adaptive allocation algorithm that learns which questions are hardest for the LLM while simultaneously collecting human responses. Each human label serves a dual role: it improves the estimate for that question and reveals how well the LLM predicts human responses on it. The algorithm directs more budget to questions where the LLM is least reliable, without requiring any prior knowledge of question-level LLM accuracy. We prove that the allocation gap relative to the best possible allocation vanishes as the budget grows, and validate the approach on both synthetic data and a real survey dataset with 68 questions and over 2000 respondents. On real survey data, the standard practice of allocating human labels uniformly across questions wastes 10--12% of the budget relative to the optimal; our algorithm reduces this waste to 2--6%, and the advantage grows as questions become more heterogeneous in LLM prediction quality. The algorithm achieves the same estimation quality as traditional uniform sampling with fewer human samples, requires no pilot study, and is backed by formal performance guarantees validated on real survey data. More broadly, the framework applies whenever scarce human oversight must be allocated across tasks where LLM reliability is unknown.


Unbounded Density Ratio Estimation and Its Application to Covariate Shift Adaptation

Liu, Ren-Rui, Fan, Jun, Shi, Lei, Guo, Zheng-Chu

arXiv.org Machine Learning

This paper focuses on the problem of unbounded density ratio estimation -- an understudied yet critical challenge in statistical learning -- and its application to covariate shift adaptation. Much of the existing literature assumes that the density ratio is either uniformly bounded or unbounded but known exactly. These conditions are often violated in practice, creating a gap between theoretical guarantees and real-world applicability. In contrast, this work directly addresses unbounded density ratios and integrates them into importance weighting for effective covariate shift adaptation. We propose a three-step estimation method that leverages unlabeled data from both the source and target distributions: (1) estimating a relative density ratio; (2) applying a truncation operation to control its unboundedness; and (3) transforming the truncated estimate back into the standard density ratio. The estimated density ratio is then employed as importance weights for regression under covariate shift. We establish rigorous, non-asymptotic convergence guarantees for both the proposed density ratio estimator and the resulting regression function estimator, demonstrating optimal or near-optimal convergence rates. Our findings offer new theoretical insights into density ratio estimation and learning under covariate shift, extending classical learning theory to more practical and challenging scenarios.


Few Batches or Little Memory, But Not Both: Simultaneous Space and Adaptivity Constraints in Stochastic Bandits

Huang, Ruiyuan, Lyu, Zicheng, Zhu, Xiaoyi, Huang, Zengfeng

arXiv.org Machine Learning

We study stochastic multi-armed bandits under simultaneous constraints on space and adaptivity: the learner interacts with the environment in $B$ batches and has only $W$ bits of persistent memory. Prior work shows that each constraint alone is surprisingly mild: near-minimax regret $\widetilde{O}(\sqrt{KT})$ is achievable with $O(\log T)$ bits of memory under fully adaptive interaction, and with a $K$-independent $O(\log\log T)$-type number of batches when memory is unrestricted. We show that this picture breaks down in the simultaneously constrained regime. We prove that any algorithm with a $W$-bit memory constraint must use at least $Ω(K/W)$ batches to achieve near-minimax regret $\widetilde{O}(\sqrt{KT})$, even under adaptive grids. In particular, logarithmic memory rules out $O(K^{1-\varepsilon})$ batch complexity. Our proof is based on an information bottleneck. We show that near-minimax regret forces the learner to acquire $Ω(K)$ bits of information about the hidden set of good arms under a suitable hard prior, whereas an algorithm with $B$ batches and $W$ bits of memory allows only $O(BW)$ bits of information. A key ingredient is a localized change-of-measure lemma that yields probability-level arm exploration guarantees, which is of independent interest. We also give an algorithm that, for any bit budget $W$ with $Ω(\log T) \le W \le O(K\log T)$, uses at most $W$ bits of memory and $\widetilde{O}(K/W)$ batches while achieving regret $\widetilde{O}(\sqrt{KT})$, nearly matching our lower bound up to polylogarithmic factors.


A Distribution-to-Distribution Neural Probabilistic Forecasting Framework for Dynamical Systems

Yang, Tianlin, Du, Hailiang, Aslett, Louis

arXiv.org Machine Learning

Probabilistic forecasting provides a principled framework for uncertainty quantification in dynamical systems by representing predictions as probability distributions rather than deterministic trajectories. However, existing forecasting approaches, whether physics-based or neural-network-based, remain fundamentally trajectory-oriented: predictive distributions are usually accessed through ensembles or sampling, rather than evolved directly as dynamical objects. A distribution-to-distribution (D2D) neural probabilistic forecasting framework is developed to operate directly on predictive distributions. The framework introduces a distributional encoding and decoding structure around a replaceable neural forecasting module, using kernel mean embeddings to represent input distributions and mixture density networks to parameterise output predictive distributions. This design enables recursive propagation of predictive uncertainty within a unified end-to-end neural architecture, with model training and evaluation carried out directly in terms of probabilistic forecast skill. The framework is demonstrated on the Lorenz63 chaotic dynamical system. Results show that the D2D model captures nontrivial distributional evolution under nonlinear dynamics, produces skillful probabilistic forecasts without explicit ensemble simulation, and remains competitive with, and in some cases outperforms, a simplified perfect model benchmark. These findings point to a new paradigm for probabilistic forecasting, in which predictive distributions are learned and evolved directly rather than reconstructed indirectly through ensemble-based uncertainty propagation.


Shape-Adaptive Conditional Calibration for Conformal Prediction via Minimax Optimization

Bao, Yajie, Zhang, Chuchen, Wang, Zhaojun, Ren, Haojie, Zou, Changliang

arXiv.org Machine Learning

Achieving valid conditional coverage in conformal prediction is challenging due to the theoretical difficulty of satisfying pointwise constraints in finite samples. Building upon the characterization of conditional coverage through marginal moment restrictions, we introduce Minimax Optimization Predictive Inference (MOPI), a framework that generalizes prior work by optimizing over a flexible class of set-valued mappings during the calibration phase, rather than simply calibrating a fixed sublevel set. This minimax formulation effectively circumvents the structural constraints of predefined score functions, achieving superior shape adaptivity while maintaining a principled connection to the minimization of mean squared coverage error. Theoretically, we provide non-asymptotic oracle inequalities and show that the convergence rate of the coverage error attains the optimal order under regular conditions. The MOPI also enables valid inference conditional on sensitive attributes that are available during calibration but unobserved at test time. Empirical results on complex, non-standard conditional distributions demonstrate that MOPI produces more efficient prediction sets than existing baselines.