Goto

Collaborating Authors

 nlayer


b9523d484af624986c2e0c630ac44ecb-Supplemental-Conference.pdf

Neural Information Processing Systems

Lemma B.4. (Lemma 2.1.8 in [4]) For any diffeomorphismf Diffkc Rd and any δ > 0, there exists a finite sequence of(δ,k)-near-identity diffeomorphismsg1,,gs such that f = gs gs 1 g1. Let πi: Rd R denote the projection onto theith coordinate. Supposef: Rd Rd is compactly supported and sufficientlyCk-close to the identity. In this section, we analysis how to make the affine coupling flow with dimension-augmentation invertible. Tohandle this problem, we need to makesure thatRange(F)is tractable for easy sampling.




Efficient Bayesian Learning Curve Extrapolation using Prior-Data Fitted Networks

Neural Information Processing Systems

Learning curve extrapolation aims to predict model performance in later epochs of training, based on the performance in earlier epochs. In this work, we argue that, while the inherent uncertainty in the extrapolation of learning curves warrants a Bayesian approach, existing methods are (i) overly restrictive, and/or (ii) computationally expensive.



Do Neural Scaling Laws Exist on Graph Self-Supervised Learning?

Ma, Qian, Mao, Haitao, Liu, Jingzhe, Zhang, Zhehua, Feng, Chunlin, Song, Yu, Shao, Yihan, Ma, Yao

arXiv.org Artificial Intelligence

Self-supervised learning~(SSL) is essential to obtain foundation models in NLP and CV domains via effectively leveraging knowledge in large-scale unlabeled data. The reason for its success is that a suitable SSL design can help the model to follow the neural scaling law, i.e., the performance consistently improves with increasing model and dataset sizes. However, it remains a mystery whether existing SSL in the graph domain can follow the scaling behavior toward building Graph Foundation Models~(GFMs) with large-scale pre-training. In this study, we examine whether existing graph SSL techniques can follow the neural scaling behavior with the potential to serve as the essential component for GFMs. Our benchmark includes comprehensive SSL technique implementations with analysis conducted on both the conventional SSL setting and many new settings adopted in other domains. Surprisingly, despite the SSL loss continuously decreasing, no existing graph SSL techniques follow the neural scaling behavior on the downstream performance. The model performance only merely fluctuates on different data scales and model scales. Instead of the scales, the key factors influencing the performance are the choices of model architecture and pretext task design. This paper examines existing SSL techniques for the feasibility of Graph SSL techniques in developing GFMs and opens a new direction for graph SSL design with the new evaluation prototype. Our code implementation is available online to ease reproducibility on https://github.com/GraphSSLScaling/GraphSSLScaling.


Fully-fused Multi-Layer Perceptrons on Intel Data Center GPUs

Yuan, Kai, Bauinger, Christoph, Zhang, Xiangyi, Baehr, Pascal, Kirchhart, Matthias, Dabert, Darius, Tousnakhoff, Adrien, Boudier, Pierre, Paulitsch, Michael

arXiv.org Artificial Intelligence

This paper presents a SYCL implementation of Multi-Layer Perceptrons (MLPs), which targets and is optimized for the Intel Data Center GPU Max 1550. To increase the performance, our implementation minimizes the slow global memory accesses by maximizing the data reuse within the general register file and the shared local memory by fusing the operations in each layer of the MLP. We show with a simple roofline model that this results in a significant increase in the arithmetic intensity, leading to improved performance, especially for inference. We compare our approach to a similar CUDA implementation for MLPs and show that our implementation on the Intel Data Center GPU outperforms the CUDA implementation on Nvidia's H100 GPU by a factor up to 2.84 in inference and 1.75 in training. The paper also showcases the efficiency of our SYCL implementation in three significant areas: Image Compression, Neural Radiance Fields, and Physics-Informed Machine Learning. In all cases, our implementation outperforms the off-the-shelf Intel Extension for PyTorch (IPEX) implementation on the same Intel GPU by up to a factor of 30 and the CUDA PyTorch version on Nvidia's H100 GPU by up to a factor 19. The code can be found at https://github.com/intel/tiny-dpcpp-nn.


Efficient Bayesian Learning Curve Extrapolation using Prior-Data Fitted Networks

Adriaensen, Steven, Rakotoarison, Herilalaina, Müller, Samuel, Hutter, Frank

arXiv.org Machine Learning

Learning curve extrapolation aims to predict model performance in later epochs of training, based on the performance in earlier epochs. In this work, we argue that, while the inherent uncertainty in the extrapolation of learning curves warrants a Bayesian approach, existing methods are (i) overly restrictive, and/or (ii) computationally expensive. We describe the first application of prior-data fitted neural networks (PFNs) in this context. A PFN is a transformer, pre-trained on data generated from a prior, to perform approximate Bayesian inference in a single forward pass. We propose LC-PFN, a PFN trained to extrapolate 10 million artificial right-censored learning curves generated from a parametric prior proposed in prior art using MCMC. We demonstrate that LC-PFN can approximate the posterior predictive distribution more accurately than MCMC, while being over 10 000 times faster. We also show that the same LC-PFN achieves competitive performance extrapolating a total of 20 000 real learning curves from four learning curve benchmarks (LCBench, NAS-Bench-201, Taskset, and PD1) that stem from training a wide range of model architectures (MLPs, CNNs, RNNs, and Transformers) on 53 different datasets with varying input modalities (tabular, image, text, and protein data). Finally, we investigate its potential in the context of model selection and find that a simple LC-PFN based predictive early stopping criterion obtains 2 - 6x speed-ups on 45 of these datasets, at virtually no overhead.