Goto

Collaborating Authors

 Zhu, Jialiang


Disentangle Estimation of Causal Effects from Cross-Silo Data

arXiv.org Artificial Intelligence

Estimating causal effects among different events is of great importance to critical fields such as drug development. Nevertheless, Related works Recent years have seen the emergence of machine the data features associated with events may be distributed learning methods for estimating various causal effects across various silos and remain private within respective [11-15]. In single domains, these methods often rely on extensive parties, impeding direct information exchange between experimentation and observations with similar spatial them. This, in turn, can result in biased estimations distribution of data dimensions [16-18]. Inductive approaches of local causal effects, which rely on the characteristics of such as FlextNet [19] leverage structural similarities only a subset of the covariates. To tackle this challenge, we among latent outcomes for causal effect estimation. HTCE introduce an innovative disentangle architecture designed to [20] aids in estimating causal effects in the target domain facilitate the seamless cross-silo transmission of model parameters, with assistance from source domain data, but it is limited to enriched with causal mechanisms, through a combination specific source and target domains. FedCI [21] and Causal-of shared and private branches. Besides, we introduce RFF [22] primarily focus on scenarios where different parties global constraints into the equation to effectively mitigate have the same data feature dimensions.


Multiple View Geometry Transformers for 3D Human Pose Estimation

arXiv.org Artificial Intelligence

In this work, we aim to improve the 3D reasoning ability of Transformers in multi-view 3D human pose estimation. Recent works have focused on end-to-end learning-based transformer designs, which struggle to resolve geometric information accurately, particularly during occlusion. Instead, we propose a novel hybrid model, MVGFormer, which has a series of geometric and appearance modules organized in an iterative manner. The geometry modules are learning-free and handle all viewpoint-dependent 3D tasks geometrically which notably improves the model's generalization ability. The appearance modules are learnable and are dedicated to estimating 2D poses from image signals end-to-end which enables them to achieve accurate estimates even when occlusion occurs, leading to a model that is both accurate and generalizable to new cameras and geometries. We evaluate our approach for both in-domain and out-of-domain settings, where our model consistently outperforms state-of-the-art methods, and especially does so by a significant margin in the out-of-domain setting. We will release the code and models: https://github.com/XunshanMan/MVGFormer.