Goto

Collaborating Authors

 potential application



existence of multiple representations of the same environment for a few sample neurons, we performed hypothesis tests for multiple

Neural Information Processing Systems

We thank all reviewers for their careful reviews and many positive comments. We feel that the typos and minor issues are easily addressable and will be corrected. We will incorporate this analysis into a revision of the paper. We thank R1 for bringing this highly related work to our attention. That work focuses on environments for which mice have previously developed spatial maps.



Reviews: Nonparametric learning from Bayesian models with randomized objective functions

Neural Information Processing Systems

The idea: You want to do Bayesian inference on a parameter theta, with prior pi(theta) and parametric likelihood f_theta, but you're not sure if the likelihood is correctly specified. So put a nonparametric prior on the sampling distribution: a mixture of Dirichlet processes centered at f_theta with mixing distribution pi(theta). The concentration parameter of the DP provides a sliding scale between vanilla Bayesian inference (total confidence in the parametric model) and Bayesian bootstrap (no confidence at all, use the empirical distribution). This is a simple idea, but the paper presents it lucidly and compellingly, beginning with a diverse list of potential applications: the method may be viewed as regularization of a nonparametric Bayesian model towards a parametric one; as robustification of a parametric Bayesian model to misspecification; as a means of correcting a variational approximation; or as nonparametric decision theory, when the log-likelihood is swapped out for an arbitrary utility function. As for implementation, the procedure requires (1) sampling from the parametric Bayesian posterior distribution and (2) performing a p-dimensional maximization, where p is the dimension of theta.


Interview with Sherry Yang: Learning interactive real-world simulators

AIHub

Sherry Yang, Yilun Du, Kamyar Ghasemipour, Jonathan Tompson, Leslie Kaelbling, Dale Schuurmans and Pieter Abbeel won an outstanding paper award at ICLR2024 for their work Learning Interactive Real-World Simulators. In the paper, they introduce a universal simulator (called UniSim) which takes image and text input to train a robot simulator. We spoke to Sherry about this work, some of the challenges, and potential applications. There are two components – there is the universal component and then there is a simulator component. Looking at the simulator component first – typically when people build a simulator, they do this based on an understanding of the real world, using physics equations. Researchers will build a simulator to study how things work, such as how cars move, for example.


MechGPT, a language-based strategy for mechanics and materials modeling that connects knowledge across scales, disciplines and modalities

Buehler, Markus J.

arXiv.org Artificial Intelligence

For centuries, researchers have sought out ways to connect disparate areas of knowledge. While early scholars (Galileo, da Vinci, etc.) were experts across fields, specialization has taken hold later. With the advent of Artificial Intelligence, we can now explore relationships across areas (e.g., mechanics-biology) or disparate domains (e.g., failure mechanics-art). To achieve this, we use a fine-tuned Large Language Model (LLM), here for a subset of knowledge in multiscale materials failure. The approach includes the use of a general-purpose LLM to distill question-answer pairs from raw sources followed by LLM fine-tuning. The resulting MechGPT LLM foundation model is used in a series of computational experiments to explore its capacity for knowledge retrieval, various language tasks, hypothesis generation, and connecting knowledge across disparate areas. While the model has some ability to recall knowledge from training, we find that LLMs are particularly useful to extract structural insights through Ontological Knowledge Graphs. These interpretable graph structures provide explanatory insights, frameworks for new research questions, and visual representations of knowledge that also can be used in retrieval-augmented generation. Three versions of MechGPT are discussed, featuring different sizes from 13 billion to 70 billion parameters, and reaching context lengths of more than 10,000 tokens. This provides ample capacity for sophisticated retrieval augmented strategies, as well as agent-based modeling where multiple LLMs interact collaboratively and/or adversarially, the incorporation of new data from the literature or web searches, as well as multimodality.


Gpt-4: A Review on Advancements and Opportunities in Natural Language Processing

Baktash, Jawid Ahmad, Dawodi, Mursal

arXiv.org Artificial Intelligence

Generative Pre-trained Transformer 4 (GPT-4) is the fourth-generation language model in the GPT series, developed by OpenAI, which promises significant advancements in the field of natural language processing (NLP). In this research article, we have discussed the features of GPT-4, its potential applications, and the challenges that it might face. We have also compared GPT-4 with its predecessor, GPT-3. GPT-4 has a larger model size (more than one trillion), better multilingual capabilities, improved contextual understanding, and reasoning capabilities than GPT-3. Some of the potential applications of GPT-4 include chatbots, personal assistants, language translation, text summarization, and question-answering. However, GPT-4 poses several challenges and limitations such as computational requirements, data requirements, and ethical concerns.


dall-e-2-revolutionary-ai-model.html

#artificialintelligence

DALL-E 2 is an advanced artificial intelligence (AI) model developed by OpenAI that can generate high-quality images from textual descriptions. The model builds upon the success of the original DALL-E, which was first introduced in January 2021. DALL-E 2 is a significant improvement over its predecessor, with the ability to generate images that are larger, more complex, and more realistic than before. In this article, we will explore the technology behind DALL-E 2, its potential applications, and its impact on the field of AI and computer vision. DALL-E 2 is a generative model based on the GPT-3 architecture, which is a state-of-the-art language model developed by OpenAI.


AI Can Create Deepfakes for You Now

#artificialintelligence

The field of artificial intelligence (AI) is advancing at an unprecedented pace, and the latest breakthrough comes in the form of video generation systems. These systems can instantly create videos from a short description, making video creation as easy as typing a quick note. The technology has been developed by several companies, including giants like Google and Microsoft, as well as smaller start-ups like Runway. One of the most impressive aspects of these video-generation systems is their ability to produce highly realistic videos in just a few minutes. For example, if you type "a tranquil river in the forest" into the system, it will quickly generate a short video of a river flowing through a forest, complete with sunlight streaming through the trees and birds chirping in the background.


From Text to Conversation: How ChatGPT is Changing the Way We Communicate

#artificialintelligence

In recent years, artificial intelligence (AI) has seen significant advancements, particularly in the field of natural language processing (NLP). One of the most notable developments in this field is ChatGPT, an advanced language model developed by OpenAI that uses the GPT architecture. ChatGPT is designed to generate human-like responses to text-based inputs, making it a useful tool for a variety of applications. With its ability to understand and generate natural language, ChatGPT has the potential to revolutionize the way we interact with technology and each other. But what exactly is ChatGPT, and why is it significant?