Goto

Collaborating Authors

 msssim


aae3ff05a5638ce4e2ef2fbc04229797-Supplemental-Conference.pdf

Neural Information Processing Systems

The total loss of the model is a combination of both regularization terms and a reconstructionloss. Herexr refers to reference image,xa to adversarial image and xr, xa to their corresponding reconstructions. The maximum input noise perturbation levelλ is limited to1,3 and 5. However, it should be also noted that with PGD-based training, the computational time is two times more expensive than our original method. These attacks are more successful when the adversarial reconstructions are less similar in appearance to the clean reconstructions.





Discouraging posterior collapse in hierarchical Variational Autoencoders using context

Kuzina, Anna, Tomczak, Jakub M.

arXiv.org Artificial Intelligence

Hierarchical Variational Autoencoders (VAEs) are among the most popular likelihood-based generative models. There is a consensus that the top-down hierarchical VAEs allow effective learning of deep latent structures and avoid problems like posterior collapse. Here, we show that this is not necessarily the case, and the problem of collapsing posteriors remains. To discourage this issue, we propose a deep hierarchical VAE with a context on top. Specifically, we use a Discrete Cosine Transform to obtain the last latent variable. In a series of experiments, we observe that the proposed modification allows us to achieve better utilization of the latent space and does not harm the model's generative abilities.


Photon Field Networks for Dynamic Real-Time Volumetric Global Illumination

Bauer, David, Wu, Qi, Ma, Kwan-Liu

arXiv.org Artificial Intelligence

Volume data is commonly found in many scientific disciplines, like medicine, physics, and biology. Experts rely on robust scientific visualization techniques to extract valuable insights from the data. Recent years have shown path tracing to be the preferred approach for volumetric rendering, given its high levels of realism. However, real-time volumetric path tracing often suffers from stochastic noise and long convergence times, limiting interactive exploration. In this paper, we present a novel method to enable real-time global illumination for volume data visualization. We develop Photon Field Networks -- a phase-function-aware, multi-light neural representation of indirect volumetric global illumination. The fields are trained on multi-phase photon caches that we compute a priori. Training can be done within seconds, after which the fields can be used in various rendering tasks. To showcase their potential, we develop a custom neural path tracer, with which our photon fields achieve interactive framerates even on large datasets. We conduct in-depth evaluations of the method's performance, including visual quality, stochastic noise, inference and rendering speeds, and accuracy regarding illumination and phase function awareness. Results are compared to ray marching, path tracing and photon mapping. Our findings show that Photon Field Networks can faithfully represent indirect global illumination across the phase spectrum while exhibiting less stochastic noise and rendering at a significantly faster rate than traditional methods.