Goto

Collaborating Authors

 prob


1 Appendix 1 Bayes-by-backprop The Bayesian posterior neural network distribution P (w |D) is approximated

Neural Information Processing Systems

In Algorithm 1 we give the full clustering algorithm used for each of the T fixing iterations. In Figure 1 we show how the layers' In Figure 2 we show the impact of increasing the regularisation strength.




DropBlock: A regularization method for convolutional networks

Golnaz Ghiasi, Tsung-Yi Lin, Quoc V. Le

Neural Information Processing Systems

Deep neural networks often work well when they are over-parameterized and trained withamassiveamount ofnoiseandregularization, suchasweight decay and dropout. Although dropout is widely used as a regularization technique for fully connected layers, it is often less effective for convolutional layers.



Appendix ABroaderSocietalImpact

Neural Information Processing Systems

Our intention is for this algorithm to be used in a real-world setting where humans can provide natural language instructions to robots that can carry them out. For the LOReL baseline, we used the numbers from the paper for LOReL Images. Unless otherwise specified, we use the following settings. D.4 MICalculation We calculate mutual information between the language instructions (L) and the skill codes (z) by writing MI(L,z)=H(L) H(L|z). RephrasalType Flat LISA seen 15 40 unseennoun 13.33 33.33 unseenverb 28.33 30 unseennoun+verb 6.7 20 human 26.98 27.35 Thismeansweare using a manual (human) skill predictor as opposed to our trained skill predictor.