Si, Zhongwei
NeRFCom: Feature Transform Coding Meets Neural Radiance Field for Free-View 3D Scene Semantic Transmission
Yue, Weijie, Si, Zhongwei, Wu, Bolin, Wang, Sixian, Qin, Xiaoqi, Niu, Kai, Dai, Jincheng, Zhang, Ping
Abstract--We introduce NeRFCom, a novel communication system designed for end-to-end 3D scene transmission. Comp ared to traditional systems relying on handcrafted NeRF semanti c feature decomposition for compression and well-adaptive c hannel coding for transmission error correction, our NeRFCom empl oys a nonlinear transform and learned probabilistic models, en abling flexible variable-rate joint source-channel coding and effi cient bandwidth allocation aligned with the NeRF semantic featur e's different contribution to the 3D scene synthesis fidelity. E xperi-mental results demonstrate that NeRFCom achieves free-vie w 3D scene efficient transmission while maintaining robustness under adverse channel conditions. Index T erms --Neural radiance field (NeRF), 3D scene transmission, semantic features, nonlinear transform coding. IRTUAL reality (VR) and augmented reality (AR) construct 3D scenes to provide users with immersive experiences [ 1 ]. However, traditional 3D scene synthesis techniques often rely on manual scene modeling, and the complex workflow increases the cost of deploying 3D technologies.
Toward Adaptive Semantic Communications: Efficient Data Transmission via Online Learned Nonlinear Transform Source-Channel Coding
Dai, Jincheng, Wang, Sixian, Yang, Ke, Tan, Kailin, Qin, Xiaoqi, Si, Zhongwei, Niu, Kai, Zhang, Ping
The emerging field semantic communication is driving the research of end-to-end data transmission. By utilizing the powerful representation ability of deep learning models, learned data transmission schemes have exhibited superior performance than the established source and channel coding methods. While, so far, research efforts mainly concentrated on architecture and model improvements toward a static target domain. Despite their successes, such learned models are still suboptimal due to the limitations in model capacity and imperfect optimization and generalization, particularly when the testing data distribution or channel response is different from that adopted for model training, as is likely to be the case in real-world. To tackle this, we propose a novel online learned joint source and channel coding approach that leverages the deep learning model's overfitting property. Specifically, we update the off-the-shelf pre-trained models after deployment in a lightweight online fashion to adapt to the distribution shifts in source data and environment domain. We take the overfitting concept to the extreme, proposing a series of implementation-friendly methods to adapt the codec model or representations to an individual data or channel state instance, which can further lead to substantial gains in terms of the bandwidth ratio-distortion performance. The proposed methods enable the communication-efficient adaptation for all parameters in the network without sacrificing decoding speed. Our experiments, including user study, on continually changing target source data and wireless channel environments, demonstrate the effectiveness and efficiency of our approach, on which we outperform existing state-of-the-art engineered transmission scheme (VVC combined with 5G LDPC coded transmission).