Goto

Collaborating Authors

 yu wang


Can AI Understand Our Universe? Test of Fine-Tuning GPT by Astrophysical Data

Wang, Yu, Zhang, Shu-Rui, Momtaz, Aidin, Moradi, Rahim, Rastegarnia, Fatemeh, Sahakyan, Narek, Shakeri, Soroush, Li, Liang

arXiv.org Artificial Intelligence

ChatGPT has been the most talked-about concept in recent months, captivating both professionals and the general public alike, and has sparked discussions about the changes that artificial intelligence (AI) will bring to the world. As physicists and astrophysicists, we are curious about if scientific data can be correctly analyzed by large language models (LLMs) and yield accurate physics. In this article, we fine-tune the generative pre-trained transformer (GPT) model by the astronomical data from the observations of galaxies, quasars, stars, gamma-ray bursts (GRBs), and the simulations of black holes (BHs), the fine-tuned model demonstrates its capability to classify astrophysical phenomena, distinguish between two types of GRBs, deduce the redshift of quasars, and estimate BH parameters. We regard this as a successful test, marking the LLM's proven efficacy in scientific research. With the ever-growing volume of multidisciplinary data and the advancement of AI technology, we look forward to the emergence of a more fundamental and comprehensive understanding of our universe. This article also shares some interesting thoughts on data collection and AI design. Using the approach of understanding the universe - looking outward at data and inward for fundamental building blocks - as a guideline, we propose a method of series expansion for AI, suggesting ways to train and control AI that is smarter than humans.


Hybrid Rule-Neural Coreference Resolution System based on Actor-Critic Learning

Wang, Yu, Jin, Hongxia

arXiv.org Artificial Intelligence

A coreference resolution system is to cluster all mentions that refer to the same entity in a given context. All coreference resolution systems need to tackle two main tasks: one task is to detect all of the potential mentions, and the other is to learn the linking of an antecedent for each possible mention. In this paper, we propose a hybrid rule-neural coreference resolution system based on actor-critic learning, such that it can achieve better coreference performance by leveraging the advantages from both the heuristic rules and a neural conference model. This end-to-end system can also perform both mention detection and resolution by leveraging a joint training algorithm. We experiment on the BERT model to generate input span representations. Our model with the BERT span representation achieves the state-of-the-art performance among the models on the CoNLL-2012 Shared Task English Test Set.


Deep Learning Project in NLP – Yu Wang's Personal Page

#artificialintelligence

Natural language processing is one of the most popular research area in machine learning field. During this year, I have got a chance to collaborate with a Google invested AI company, and work with them on building a deep learning based Question and Answering (QA) platform by applying my Neural Network expertise. The major work involved is the query classification of various questions from the user end. The implementation is using the Google open source Deep Learning platform: Tensorflow. In this tutorial, I will mainly discuss two different deep learning algorithms that are applied in our project: feed-forward neural network (MLP based deep learning) and recurrent neural network (like LSTM and GRU).