Goto

Collaborating Authors

 inthiscase


Optimal Subsampling with Influence Functions

Daniel Ting, Eric Brochu

Neural Information Processing Systems

As the amount of data increases, the question arises as to how best to deal with the large datasets. While computational platforms such as Spark [28] and Ray [23] help process large datasets once a desired model is chosen, simply using smaller data can be a faster solution for exploratory data modeling, rapid prototyping, or other tasks where the accuracy obtainable from the full dataset is notneeded.




Attention in Convolutional LSTM for Gesture Recognition

Liang Zhang, Guangming Zhu, Lin Mei, Peiyi Shen, Syed Afaq Ali Shah, Mohammed Bennamoun

Neural Information Processing Systems

In the preliminary "Res3D+ConvLSTM+MobileNet" architecture, the blocks 1-4 of Res3D [16] are used first to learn the local short-term spatiotemporal feature maps which have a relativelylargespatialsize.



kcur kcurX i=1

Neural Information Processing Systems

Out of the box, these models take as input a sequence of vectors in embedding space and output asequence ofvectors inthe same space. We treat the prediction of the model at the position corresponding toxi (that is absolute position 2i 1)asthepredictionof f(xi). A.2 Training Each training prompt is produced by sampling a random functionf from the function class we are training on, then sampling inputsxi from the isotropic Gaussian distributionN(0,Id) and constructing apromptas(x1,f(x1),...,xk,f(xk)). For the class of decision trees, the random functionf is represented by a decision tree of depth4 (with16leafnodes),with20dimensionalinputs. Minimum norm least squares is the optimal estimator for the linear regression problem.