Goto

Collaborating Authors

 mindeye


A Appendix

Neural Information Processing Systems

KAN oversaw the project and contributed valuable feedback. MindEye was developed using a training and validation set of Subject 1's data, with the test set (and other subjects' data) untouched until final PyTorch code for the MLP backbone and projector is depicted in Algorithm 1. Specifics on how we DALL-E 2. This makes our prior much faster at inference time. For simplicity we use bidirectional attention in our final model. To map to Stable Diffusion's V AE latent space we use a low-level pipeline with the same architecture as the high level pipeline. Recent works in low-level vision (super-resolution, denoising, deblurring, etc.) have observed that This performs worse than only applying the loss in latent space and also requires significantly more GPU memory.



Reconstructing the Mind's Eye: fMRI-to-Image with Contrastive Learning and Diffusion Priors

Neural Information Processing Systems

We present MindEye, a novel fMRI-to-image approach to retrieve and reconstruct viewed images from brain activity. Our model comprises two parallel submodules that are specialized for retrieval (using contrastive learning) and reconstruction (using a diffusion prior). MindEye can map fMRI brain activity to any high dimensional multimodal latent space, like CLIP image space, enabling image reconstruction using generative models that accept embeddings from this latent space. We comprehensively compare our approach with other existing methods, using both qualitative side-by-side comparisons and quantitative evaluations, and show that MindEye achieves state-of-the-art performance in both reconstruction and retrieval tasks. In particular, MindEye can retrieve the exact original image even among highly similar candidates indicating that its brain embeddings retain fine-grained image-specific information. This allows us to accurately retrieve images even from large-scale databases like LAION-5B. We demonstrate through ablations that MindEye's performance improvements over previous methods result from specialized submodules for retrieval and reconstruction, improved training techniques, and training models with orders of magnitude more parameters. Furthermore, we show that MindEye can better preserve low-level image features in the reconstructions by using img2img, with outputs from a separate autoencoder. All code is available on GitHub.



A Appendix

Neural Information Processing Systems

KAN oversaw the project and contributed valuable feedback. MindEye was developed using a training and validation set of Subject 1's data, with the test set (and other subjects' data) untouched until final PyTorch code for the MLP backbone and projector is depicted in Algorithm 1. Specifics on how we DALL-E 2. This makes our prior much faster at inference time. For simplicity we use bidirectional attention in our final model. To map to Stable Diffusion's V AE latent space we use a low-level pipeline with the same architecture as the high level pipeline. Recent works in low-level vision (super-resolution, denoising, deblurring, etc.) have observed that This performs worse than only applying the loss in latent space and also requires significantly more GPU memory.


Reconstructing the Mind's Eye: fMRI-to-Image with Contrastive Learning and Diffusion Priors

Neural Information Processing Systems

We present MindEye, a novel fMRI-to-image approach to retrieve and reconstruct viewed images from brain activity. Our model comprises two parallel submodules that are specialized for retrieval (using contrastive learning) and reconstruction (using a diffusion prior). MindEye can map fMRI brain activity to any high dimensional multimodal latent space, like CLIP image space, enabling image reconstruction using generative models that accept embeddings from this latent space. We comprehensively compare our approach with other existing methods, using both qualitative side-by-side comparisons and quantitative evaluations, and show that MindEye achieves state-of-the-art performance in both reconstruction and retrieval tasks. In particular, MindEye can retrieve the exact original image even among highly similar candidates indicating that its brain embeddings retain fine-grained image-specific information.


Reconstructing the Mind's Eye: fMRI-to-Image with Contrastive Learning and Diffusion Priors

Neural Information Processing Systems

We present MindEye, a novel fMRI-to-image approach to retrieve and reconstruct viewed images from brain activity. Our model comprises two parallel submodules that are specialized for retrieval (using contrastive learning) and reconstruction (using a diffusion prior). MindEye can map fMRI brain activity to any high dimensional multimodal latent space, like CLIP image space, enabling image reconstruction using generative models that accept embeddings from this latent space. We comprehensively compare our approach with other existing methods, using both qualitative side-by-side comparisons and quantitative evaluations, and show that MindEye achieves state-of-the-art performance in both reconstruction and retrieval tasks. In particular, MindEye can retrieve the exact original image even among highly similar candidates indicating that its brain embeddings retain fine-grained image-specific information.


Lite-Mind: Towards Efficient and Versatile Brain Representation Network

Gong, Zixuan, Zhang, Qi, Miao, Duoqian, Bao, Guangyin, Hu, Liang

arXiv.org Artificial Intelligence

Research in decoding visual information from the brain, particularly through the non-invasive fMRI method, is rapidly progressing. The challenge arises from the limited data availability and the low signal-to-noise ratio of fMRI signals, leading to a low-precision task of fMRI-to-image retrieval. State-of-the-art MindEye remarkably improves fMRI-to-image retrieval performance by leveraging a deep MLP with a high parameter count orders of magnitude, i.e., a 996M MLP Backbone per subject, to align fMRI embeddings to the final hidden layer of CLIP's vision transformer. However, significant individual variations exist among subjects, even within identical experimental setups, mandating the training of subject-specific models. The substantial parameters pose significant challenges in deploying fMRI decoding on practical devices, especially with the necessitating of specific models for each subject. To this end, we propose Lite-Mind, a lightweight, efficient, and versatile brain representation network based on discrete Fourier transform, that efficiently aligns fMRI voxels to fine-grained information of CLIP. Our experiments demonstrate that Lite-Mind achieves an impressive 94.3% fMRI-to-image retrieval accuracy on the NSD dataset for Subject 1, with 98.7% fewer parameters than MindEye. Lite-Mind is also proven to be able to be migrated to smaller brain datasets and establishes a new state-of-the-art for zero-shot classification on the GOD dataset. The code is available at https://github.com/gongzix/Lite-Mind.


Reconstructing the Mind's Eye: fMRI-to-Image with Contrastive Learning and Diffusion Priors

Scotti, Paul S., Banerjee, Atmadeep, Goode, Jimmie, Shabalin, Stepan, Nguyen, Alex, Cohen, Ethan, Dempster, Aidan J., Verlinde, Nathalie, Yundler, Elad, Weisberg, David, Norman, Kenneth A., Abraham, Tanishq Mathew

arXiv.org Artificial Intelligence

We present MindEye, a novel fMRI-to-image approach to retrieve and reconstruct viewed images from brain activity. Our model comprises two parallel submodules that are specialized for retrieval (using contrastive learning) and reconstruction (using a diffusion prior). MindEye can map fMRI brain activity to any high dimensional multimodal latent space, like CLIP image space, enabling image reconstruction using generative models that accept embeddings from this latent space. We comprehensively compare our approach with other existing methods, using both qualitative side-by-side comparisons and quantitative evaluations, and show that MindEye achieves state-of-the-art performance in both reconstruction and retrieval tasks. In particular, MindEye can retrieve the exact original image even among highly similar candidates indicating that its brain embeddings retain fine-grained image-specific information. This allows us to accurately retrieve images even from large-scale databases like LAION-5B. We demonstrate through ablations that MindEye's performance improvements over previous methods result from specialized submodules for retrieval and reconstruction, improved training techniques, and training models with orders of magnitude more parameters. Furthermore, we show that MindEye can better preserve low-level image features in the reconstructions by using img2img, with outputs from a separate autoencoder. All code is available on GitHub.