Goto

Collaborating Authors

 fourier transform



Improving the Learning Capability of Small-size Image Restoration Network by Deep Fourier Shifting

Neural Information Processing Systems

State-of-the-art image restoration methods currently face challenges in terms of computational requirements and performance, making them impractical for deployment on edge devices such as phones and resource-limited devices.


Utilizing Image Transforms and Diffusion Models for Generative Modeling of Short and Long Time Series

Neural Information Processing Systems

Lately, there has been a surge in interest surrounding generative modeling of time series data. Most existing approaches are designed either to process short sequences or to handle long-range sequences. This dichotomy can be attributed to gradient issues with recurrent networks, computational costs associated with transformers, and limited expressiveness of state space models. Towards a unified generative model for varying-length time series, we propose in this work to transform sequences into images.


Learning Functional Transduction: S.I. Contents

Neural Information Processing Systems

We propose below the proofs of the results presented in the main text. RKBS developed in (Zhang et al., 2009; Song et al., 2013) to develop the notion of vector-valued (Giles, 1967). " 0, @ j ď n, @ u P U (9) which allows us to say that O P RKBS (Corollary 3.2 of Zhang (2013)) that we recall hereafter: We first define for any linear operator We show our result in the case J=1 and can be directly extended to any cardinality J. Specifically, we tested three expressions: Exp. The two first expressions yield similar result in the ADR experiment at an equal compute cost. We also tried a'branch' and'trunk' networks formulation of the model as in DeepONet (Lu T able S.2: Summary of the architectural hyperparameters used to build the Transducer in the four experiments. 'Depth' corresponds to network number of layers, 'MLP dim' to the dimensionality of the hidden layer As stated, we used for all experiments, the same meta-training procedure. T able S.3: Summary of the meta-learning hyperparameters used to meta-train the Transducer in our four Figure S.1: Examples of sampled functions δ p xq and ν px q used to build operators O We train Tranducers for 200K gradient steps. Flow library (Holl et al., 2020) that allows for batched and differentiable simulations of fluid dynamics Figure S.5: Magnitude of the complex coefficients of the Fourier transform of an exemple pair of input and In order to tackle the high-resolution climate modeling experiment, we take inspiration from Pathak et al. (2022), which combines neural operators with the patch splitting L " 12, in order to match number of trainable parameters.







A Another universality result for neural oscillators

Neural Information Processing Systems

The universal approximation Theorem 3.1 immediately implies another universal approximation Thus y (t) solves the ODE (2.6), with initial condition y (0) = y (0) = 0 . Reconstruction of a continuous signal from its sine transform. Step 0: (Equicontinuity) We recall the following fact from topology. F (τ):= null f (τ), for τ 0, f ( τ), for τ 0. Since F is odd, the Fourier transform of F is given by We provide the details below. The next step in the proof of the fundamental Lemma 3.5 needs the following preliminary result in By (B.3), this implies that It follows from Lemma 3.4 that for any input By the sine transform reconstruction Lemma B.1, there exists It follows from Lemma 3.6, that there exists Indeed, Lemma 3.7 shows that time-delays of any given input signal can be approximated with any Step 1: By the Fundamental Lemma 3.5, there exist It follows from Lemma 3.6, that there exists an oscillator Step 3: Finally, by Lemma 3.8, there exists an oscillator network,