Goto

Collaborating Authors

 confidenceinterval


Appendix

Neural Information Processing Systems

We have shown experimentally that our method is effective in a variety of domains; however, other problem domains may require additional hyperparameter tuning, which can be expensive.


Debiased Estimators in High-Dimensional Regression: A Review and Replication of Javanmard and Montanari (2014)

Smith, Benjamin

arXiv.org Machine Learning

High-dimensional statistical settings ($p \gg n$) pose fundamental challenges for classical inference, largely due to bias introduced by regularized estimators such as the LASSO. To address this, Javanmard and Montanari (2014) propose a debiased estimator that enables valid hypothesis testing and confidence interval construction. This report examines their debiased LASSO framework, which yields asymptotically normal estimators in high-dimensional settings. The key theoretical results underlying this approach are presented. Specifically, the construction of an optimized debiased estimator that restores asymptotic normality, which enables the computation of valid confidence intervals and $p$-values. To evaluate the claims of Javanmard and Montanari, a subset of the original simulation study and the real-data analysis is presented. The original empirical analysis is extended to the desparsified LASSO, which is referenced but not implemented in the original study. The results demonstrate that while the debiased LASSO achieves reliable coverage and controls Type I error, the LASSO projection estimator can offer improved power in idealized low-signal settings without compromising error rates. The results reveal a trade-off: the LASSO projection estimator performs well in low-signal settings, while Javanmard and Montanari's method is more robust to complex correlations, improving precision and signal detection in real data.


Model-FreeActiveExploration inReinforcementLearning

Neural Information Processing Systems

We study the problem of exploration in Reinforcement Learning and present a novel model-free solution. We adopt an information-theoretical viewpoint and start from the instance-specific lower bound ofthe number ofsamples that have to be collected to identify a nearly-optimal policy.



Explaining Deep Learning Models -- A Bayesian Non-parametric Approach

Wenbo Guo, Sui Huang, Yunzhe Tao, Xinyu Xing, Lin Lin

Neural Information Processing Systems

While recent research hasproposed various technical approaches to provide some clues as to how an ML model makes individual predictions, they cannot provide users with an ability to inspect a model as a completeentity.



c97e7a5153badb6576d8939469f58336-Supplemental.pdf

Neural Information Processing Systems

Our initial experiments (implementation, debugging, hyperparameter tuning, etc.) required about 5000CPUhoursofcompute. Due to these rules, it is recommended to group together in order to attack simultaneously. In Warehouse[4], QTRAN makes slightly faster progress than VAST(η = 12). The results forWarehouse[16], Battle[80], and GaussianSqueeze[800] are shown in Figure 1. Figure 10: Visualizations of the generated sub-teams ofXMetaGrad with η = 14 and XSpatial with k-means clustering using 10 centroids at different stages (early, middle, late) inBattle[80] after training. Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments.



Appendix

Neural Information Processing Systems

We have shown experimentally that our method is effective in a variety of domains; however, other problem domains may require additional hyperparameter tuning, which can be expensive.


41a6fd31aa2e75c3c6d427db3d17ea80-Supplemental.pdf

Neural Information Processing Systems

In order to accelerate the NES search phase, we generated the pool using the weight sharing schemes proposed by Random Search with WeightSharing[37]andDARTS[39]. Specifically, we trained one-shot weight-sharing models usingeachof these two algorithms, then we sampled architectures from the weightshared models uniformly at random to build the pool.