Disentangled Interleaving Variational Encoding

Wong, Noelle Y. L., Cheu, Eng Yeow, Chiam, Zhonglin, Srinivasan, Dipti

arXiv.org Machine Learning 

Conflicting objectives present a considerable challenge in interleaving multi-task learning, necessitating the need for meticulous design and balance to ensure effective learning of a representative latent data space across all tasks without mutual negative impact. Our proposed model, Deep Disentangled Interleaving Variational Encoding (Deep-DIVE) learns disentangled features from the original input to form clusters in the embedding space and unifies these features via the cross-attention mechanism in the fusion stage. We theoretically prove that combining the objectives for reconstruction and forecasting fully captures the lower bound and mathematically derive a loss function for disentanglement using Naïve Bayes. Experiments on two public datasets show that DeepDIVE disentangles the original input and yields forecast accuracies better than the original VAE and comparable to existing state-of-the-art baselines. In multi-objective deep learning, gradients from different objectives can conflict, when the different loss terms induce competing gradient directions during training of the network. Balancing these gradients to ensure stable and effective learning is a significant challenge prompting the development of methods to mitigate this issue, such as Liu et al. (2021); Yu et al. (2020); Sener & Koltun (2018) which solve an additional optmization problem before each gradient update step, to manipulate conflicting gradients before the update.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found